diff --git a/20250120235238/rank1.log b/20250120235238/rank1.log new file mode 100644 index 0000000000000000000000000000000000000000..d784035c7109b0317d8fe209d5b05e1ded2cd859 --- /dev/null +++ b/20250120235238/rank1.log @@ -0,0 +1,395 @@ +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.19s +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:07:54][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 149.58 seconds, peak gpu memory 13.4G +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.7GB text_tokens: 31548.0 tgs: 57 data_time: 2.14s time: 546.74s eta: 3 days, 18:03:37 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32129.0 tgs: 61 data_time: 0.80s time: 523.23s eta: 3 days, 14:02:31 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.9GB text_tokens: 31947.0 tgs: 61 data_time: 0.86s time: 522.93s eta: 3 days, 13:50:53 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31471.0 tgs: 60 data_time: 0.76s time: 520.27s eta: 3 days, 13:16:02 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 32503.0 tgs: 62 data_time: 0.78s time: 520.97s eta: 3 days, 13:14:13 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.7GB text_tokens: 31133.0 tgs: 59 data_time: 0.90s time: 520.90s eta: 3 days, 13:04:49 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 31847.0 tgs: 60 data_time: 1.21s time: 523.37s eta: 3 days, 13:20:16 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.362 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 32030.0 tgs: 61 data_time: 0.96s time: 520.65s eta: 3 days, 12:44:58 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 31786.0 tgs: 61 data_time: 0.75s time: 520.16s eta: 3 days, 12:31:35 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 32078.0 tgs: 61 data_time: 1.09s time: 520.48s eta: 3 days, 12:25:58 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32240.0 tgs: 61 data_time: 0.89s time: 524.51s eta: 3 days, 12:56:31 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 31632.0 tgs: 60 data_time: 0.86s time: 520.63s eta: 3 days, 12:10:08 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.1GB text_tokens: 30691.0 tgs: 59 data_time: 0.79s time: 519.96s eta: 3 days, 11:54:57 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32119.0 tgs: 61 data_time: 0.88s time: 521.27s eta: 3 days, 11:58:57 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 31154.0 tgs: 59 data_time: 0.67s time: 524.11s eta: 3 days, 12:17:39 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.9GB text_tokens: 31896.0 tgs: 61 data_time: 0.65s time: 520.52s eta: 3 days, 11:34:22 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 31357.0 tgs: 60 data_time: 0.93s time: 518.89s eta: 3 days, 11:09:59 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 30112.0 tgs: 57 data_time: 0.71s time: 522.09s eta: 3 days, 11:32:03 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.1GB text_tokens: 31775.0 tgs: 60 data_time: 0.88s time: 523.81s eta: 3 days, 11:39:52 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32263.0 tgs: 61 data_time: 0.80s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31621.0 tgs: 60 data_time: 1.09s time: 518.42s eta: 3 days, 10:30:55 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.400 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 32165.0 tgs: 61 data_time: 1.11s time: 522.82s eta: 3 days, 11:04:12 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31731.0 tgs: 60 data_time: 0.60s time: 523.52s eta: 3 days, 11:02:09 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31953.0 tgs: 61 data_time: 0.74s time: 521.01s eta: 3 days, 10:29:36 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.0GB text_tokens: 32071.0 tgs: 61 data_time: 0.84s time: 519.09s eta: 3 days, 10:02:39 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 32232.0 tgs: 61 data_time: 0.64s time: 523.28s eta: 3 days, 10:33:44 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.196 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31872.0 tgs: 60 data_time: 0.70s time: 522.96s eta: 3 days, 10:21:58 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 31715.0 tgs: 60 data_time: 0.81s time: 521.30s eta: 3 days, 9:57:35 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31142.0 tgs: 59 data_time: 0.69s time: 520.62s eta: 3 days, 9:42:30 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.7GB text_tokens: 31246.0 tgs: 59 data_time: 1.11s time: 522.42s eta: 3 days, 9:50:43 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.8GB text_tokens: 31427.0 tgs: 60 data_time: 0.89s time: 523.77s eta: 3 days, 9:54:43 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.319 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 31009.0 tgs: 59 data_time: 0.83s time: 520.41s eta: 3 days, 9:14:30 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.6GB text_tokens: 31125.0 tgs: 59 data_time: 0.70s time: 520.66s eta: 3 days, 9:08:10 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 32138.0 tgs: 61 data_time: 0.92s time: 521.32s eta: 3 days, 9:05:40 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.8GB text_tokens: 31588.0 tgs: 60 data_time: 0.95s time: 524.09s eta: 3 days, 9:22:48 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.8GB text_tokens: 31748.0 tgs: 60 data_time: 0.90s time: 520.94s eta: 3 days, 8:44:45 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 32016.0 tgs: 61 data_time: 0.93s time: 520.21s eta: 3 days, 8:29:19 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.9GB text_tokens: 31540.0 tgs: 60 data_time: 0.79s time: 522.55s eta: 3 days, 8:42:18 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.8GB text_tokens: 30666.0 tgs: 58 data_time: 0.84s time: 524.31s eta: 3 days, 8:49:54 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31939.0 tgs: 61 data_time: 0.79s time: 519.87s eta: 3 days, 8:00:07 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.8GB text_tokens: 30286.0 tgs: 58 data_time: 0.84s time: 520.12s eta: 3 days, 7:53:45 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32181.0 tgs: 61 data_time: 0.92s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32188.0 tgs: 61 data_time: 0.75s time: 524.34s eta: 3 days, 8:15:09 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.370 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 32373.0 tgs: 62 data_time: 0.71s time: 520.56s eta: 3 days, 7:31:46 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.7GB text_tokens: 30468.0 tgs: 58 data_time: 0.67s time: 518.55s eta: 3 days, 7:04:42 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31398.0 tgs: 60 data_time: 0.77s time: 523.14s eta: 3 days, 7:37:59 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.9GB text_tokens: 30639.0 tgs: 58 data_time: 0.59s time: 523.61s eta: 3 days, 7:33:33 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 32054.0 tgs: 61 data_time: 0.94s time: 520.20s eta: 3 days, 6:53:50 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.8GB text_tokens: 31329.0 tgs: 60 data_time: 0.91s time: 520.04s eta: 3 days, 6:43:41 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 30789.0 tgs: 58 data_time: 0.79s time: 523.74s eta: 3 days, 7:08:37 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31630.0 tgs: 60 data_time: 0.80s time: 523.82s eta: 3 days, 7:00:35 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.221 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 31039.0 tgs: 59 data_time: 0.80s time: 520.22s eta: 3 days, 6:19:20 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 31956.0 tgs: 61 data_time: 1.05s time: 518.76s eta: 3 days, 5:57:29 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 32128.0 tgs: 61 data_time: 0.72s time: 522.67s eta: 3 days, 6:24:02 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 31665.0 tgs: 60 data_time: 0.69s time: 522.76s eta: 3 days, 6:16:10 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.8GB text_tokens: 31563.0 tgs: 60 data_time: 0.74s time: 520.51s eta: 3 days, 5:47:14 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32095.0 tgs: 61 data_time: 0.70s time: 519.64s eta: 3 days, 5:30:48 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32082.0 tgs: 61 data_time: 1.08s time: 521.33s eta: 3 days, 5:37:15 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.1GB text_tokens: 32250.0 tgs: 61 data_time: 0.70s time: 523.57s eta: 3 days, 5:48:32 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 31468.0 tgs: 60 data_time: 0.82s time: 519.29s eta: 3 days, 5:01:41 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 31281.0 tgs: 60 data_time: 0.84s time: 520.67s eta: 3 days, 5:05:15 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.1GB text_tokens: 31987.0 tgs: 61 data_time: 0.54s time: 521.47s eta: 3 days, 5:03:40 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.3GB text_tokens: 30618.0 tgs: 58 data_time: 0.61s time: 523.93s eta: 3 days, 5:16:46 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 32.4GB text_tokens: 30691.0 tgs: 59 data_time: 0.67s time: 518.83s eta: 3 days, 4:23:02 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.9GB text_tokens: 30766.0 tgs: 59 data_time: 0.64s time: 519.27s eta: 3 days, 4:18:14 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 30736.0 tgs: 59 data_time: 0.78s time: 520.70s eta: 3 days, 4:22:08 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 32451.0 tgs: 61 data_time: 0.74s time: 524.24s eta: 3 days, 4:44:33 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.0GB text_tokens: 30934.0 tgs: 59 data_time: 0.93s time: 520.34s eta: 3 days, 4:01:38 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.174 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32071.0 tgs: 61 data_time: 0.77s time: 519.52s eta: 3 days, 3:45:47 +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 1][DP 0][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 31757.0 tgs: 60 data_time: 0.94s time: 522.30s eta: 3 days, 4:01:25 diff --git a/20250120235238/rank11.log b/20250120235238/rank11.log new file mode 100644 index 0000000000000000000000000000000000000000..40b6c91e86c5acf915a2dd5e32cde29f75171a6d --- /dev/null +++ b/20250120235238/rank11.log @@ -0,0 +1,395 @@ +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.13s +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.88 seconds, peak gpu memory 13.4G +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31696.0 tgs: 57 data_time: 1.88s time: 548.03s eta: 3 days, 18:16:21 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31732.0 tgs: 60 data_time: 1.03s time: 523.24s eta: 3 days, 14:02:39 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 32529.0 tgs: 62 data_time: 0.89s time: 522.88s eta: 3 days, 13:50:21 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.9GB text_tokens: 31543.0 tgs: 60 data_time: 0.85s time: 520.29s eta: 3 days, 13:16:11 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31681.0 tgs: 60 data_time: 0.94s time: 520.98s eta: 3 days, 13:14:18 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.4GB text_tokens: 30335.0 tgs: 58 data_time: 0.82s time: 520.90s eta: 3 days, 13:04:46 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.221 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.8GB text_tokens: 30779.0 tgs: 58 data_time: 0.68s time: 523.38s eta: 3 days, 13:20:26 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31645.0 tgs: 60 data_time: 0.91s time: 520.66s eta: 3 days, 12:45:07 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32125.0 tgs: 61 data_time: 0.66s time: 520.18s eta: 3 days, 12:31:44 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.7GB text_tokens: 31629.0 tgs: 60 data_time: 0.91s time: 520.41s eta: 3 days, 12:25:22 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.347 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 32.9GB text_tokens: 31480.0 tgs: 60 data_time: 0.93s time: 524.53s eta: 3 days, 12:56:42 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 32172.0 tgs: 61 data_time: 0.70s time: 520.65s eta: 3 days, 12:10:17 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.7GB text_tokens: 31315.0 tgs: 60 data_time: 1.22s time: 519.93s eta: 3 days, 11:54:40 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32182.0 tgs: 61 data_time: 0.63s time: 521.29s eta: 3 days, 11:59:07 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 31862.0 tgs: 60 data_time: 0.74s time: 524.13s eta: 3 days, 12:17:48 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.0GB text_tokens: 31743.0 tgs: 60 data_time: 0.88s time: 520.54s eta: 3 days, 11:34:31 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32304.0 tgs: 62 data_time: 0.84s time: 518.83s eta: 3 days, 11:09:22 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31895.0 tgs: 61 data_time: 0.89s time: 522.11s eta: 3 days, 11:32:12 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.6GB text_tokens: 30695.0 tgs: 58 data_time: 0.88s time: 523.83s eta: 3 days, 11:40:01 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.2GB text_tokens: 30721.0 tgs: 59 data_time: 0.97s time: 520.46s eta: 3 days, 10:59:04 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 29162.0 tgs: 56 data_time: 0.58s time: 518.43s eta: 3 days, 10:31:01 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 31808.0 tgs: 60 data_time: 0.81s time: 522.84s eta: 3 days, 11:04:24 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 32435.0 tgs: 61 data_time: 1.02s time: 523.54s eta: 3 days, 11:02:20 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32461.0 tgs: 62 data_time: 0.79s time: 520.96s eta: 3 days, 10:29:09 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.8GB text_tokens: 31514.0 tgs: 60 data_time: 0.99s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 30176.0 tgs: 57 data_time: 0.56s time: 523.29s eta: 3 days, 10:33:50 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32145.0 tgs: 61 data_time: 1.00s time: 522.93s eta: 3 days, 10:21:43 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 32029.0 tgs: 61 data_time: 0.58s time: 521.32s eta: 3 days, 9:57:44 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31383.0 tgs: 60 data_time: 0.64s time: 520.64s eta: 3 days, 9:42:39 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.4GB text_tokens: 30895.0 tgs: 59 data_time: 0.77s time: 522.41s eta: 3 days, 9:50:39 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 31598.0 tgs: 60 data_time: 1.06s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 32406.0 tgs: 62 data_time: 0.78s time: 520.43s eta: 3 days, 9:14:39 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 32322.0 tgs: 62 data_time: 0.83s time: 520.68s eta: 3 days, 9:08:19 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 30969.0 tgs: 59 data_time: 0.85s time: 521.30s eta: 3 days, 9:05:29 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.8GB text_tokens: 31334.0 tgs: 59 data_time: 0.89s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31952.0 tgs: 61 data_time: 0.83s time: 520.96s eta: 3 days, 8:44:55 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.212 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 32127.0 tgs: 61 data_time: 0.90s time: 520.15s eta: 3 days, 8:28:43 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.9GB text_tokens: 31858.0 tgs: 60 data_time: 0.82s time: 522.56s eta: 3 days, 8:42:25 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.8GB text_tokens: 31386.0 tgs: 59 data_time: 0.84s time: 524.34s eta: 3 days, 8:50:07 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.355 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 32250.0 tgs: 62 data_time: 0.70s time: 519.88s eta: 3 days, 8:00:12 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.308 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32507.0 tgs: 62 data_time: 1.06s time: 520.11s eta: 3 days, 7:53:41 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.9GB text_tokens: 31757.0 tgs: 60 data_time: 0.67s time: 522.00s eta: 3 days, 8:02:22 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.8GB text_tokens: 31888.0 tgs: 60 data_time: 0.97s time: 524.35s eta: 3 days, 8:15:17 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31891.0 tgs: 61 data_time: 0.80s time: 520.51s eta: 3 days, 7:31:18 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.9GB text_tokens: 31410.0 tgs: 60 data_time: 0.69s time: 518.56s eta: 3 days, 7:04:51 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31763.0 tgs: 60 data_time: 0.97s time: 523.15s eta: 3 days, 7:38:08 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.6GB text_tokens: 30226.0 tgs: 57 data_time: 0.89s time: 523.62s eta: 3 days, 7:33:41 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31914.0 tgs: 61 data_time: 0.86s time: 520.17s eta: 3 days, 6:53:35 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 31209.0 tgs: 60 data_time: 0.49s time: 520.06s eta: 3 days, 6:43:49 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.9GB text_tokens: 31712.0 tgs: 60 data_time: 0.75s time: 523.76s eta: 3 days, 7:08:44 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31541.0 tgs: 60 data_time: 0.79s time: 523.78s eta: 3 days, 7:00:13 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32580.0 tgs: 62 data_time: 0.79s time: 520.24s eta: 3 days, 6:19:29 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.7GB text_tokens: 31793.0 tgs: 61 data_time: 0.57s time: 518.78s eta: 3 days, 5:57:37 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.8GB text_tokens: 31726.0 tgs: 60 data_time: 0.64s time: 522.68s eta: 3 days, 6:24:09 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.380 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 32291.0 tgs: 61 data_time: 0.85s time: 522.73s eta: 3 days, 6:15:51 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.0GB text_tokens: 31197.0 tgs: 59 data_time: 0.78s time: 520.53s eta: 3 days, 5:47:23 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.7GB text_tokens: 31650.0 tgs: 60 data_time: 0.90s time: 519.66s eta: 3 days, 5:30:56 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.0GB text_tokens: 31693.0 tgs: 60 data_time: 0.78s time: 521.30s eta: 3 days, 5:36:54 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.0GB text_tokens: 30935.0 tgs: 59 data_time: 0.82s time: 523.59s eta: 3 days, 5:48:40 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31373.0 tgs: 60 data_time: 0.75s time: 519.32s eta: 3 days, 5:01:58 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31466.0 tgs: 60 data_time: 0.72s time: 520.65s eta: 3 days, 5:05:08 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31582.0 tgs: 60 data_time: 0.75s time: 521.45s eta: 3 days, 5:03:31 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31379.0 tgs: 59 data_time: 0.79s time: 523.95s eta: 3 days, 5:16:55 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 31834.0 tgs: 61 data_time: 0.50s time: 518.85s eta: 3 days, 4:23:10 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31863.0 tgs: 61 data_time: 0.64s time: 519.24s eta: 3 days, 4:17:58 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31055.0 tgs: 59 data_time: 0.65s time: 520.71s eta: 3 days, 4:22:17 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 31953.0 tgs: 60 data_time: 0.75s time: 524.25s eta: 3 days, 4:44:41 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32157.0 tgs: 61 data_time: 0.60s time: 520.33s eta: 3 days, 4:01:33 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 32.7GB text_tokens: 31301.0 tgs: 60 data_time: 0.73s time: 519.50s eta: 3 days, 3:45:35 +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 11][DP 2][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 31423.0 tgs: 60 data_time: 0.46s time: 522.32s eta: 3 days, 4:01:34 diff --git a/20250120235238/rank13.log b/20250120235238/rank13.log new file mode 100644 index 0000000000000000000000000000000000000000..1425932e7ce9d8ed03cbb952473d2f8a6f95f104 --- /dev/null +++ b/20250120235238/rank13.log @@ -0,0 +1,395 @@ +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.14s +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.91 seconds, peak gpu memory 13.4G +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.3GB text_tokens: 30603.0 tgs: 55 data_time: 1.93s time: 547.92s eta: 3 days, 18:15:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.7GB text_tokens: 31849.0 tgs: 60 data_time: 0.70s time: 523.24s eta: 3 days, 14:02:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.5GB text_tokens: 31059.0 tgs: 59 data_time: 0.91s time: 522.88s eta: 3 days, 13:50:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 32411.0 tgs: 62 data_time: 0.83s time: 520.29s eta: 3 days, 13:16:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 31204.0 tgs: 59 data_time: 0.65s time: 520.98s eta: 3 days, 13:14:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.9GB text_tokens: 31449.0 tgs: 60 data_time: 0.83s time: 520.90s eta: 3 days, 13:04:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.9GB text_tokens: 31590.0 tgs: 60 data_time: 0.80s time: 523.38s eta: 3 days, 13:20:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 32.9GB text_tokens: 31967.0 tgs: 61 data_time: 0.77s time: 520.66s eta: 3 days, 12:45:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 31084.0 tgs: 59 data_time: 0.90s time: 520.18s eta: 3 days, 12:31:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.7GB text_tokens: 31694.0 tgs: 60 data_time: 0.85s time: 520.42s eta: 3 days, 12:25:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32132.0 tgs: 61 data_time: 0.93s time: 524.53s eta: 3 days, 12:56:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31935.0 tgs: 61 data_time: 0.76s time: 520.65s eta: 3 days, 12:10:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 31576.0 tgs: 60 data_time: 0.81s time: 519.93s eta: 3 days, 11:54:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.0GB text_tokens: 31631.0 tgs: 60 data_time: 0.78s time: 521.29s eta: 3 days, 11:59:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 31904.0 tgs: 60 data_time: 0.82s time: 524.12s eta: 3 days, 12:17:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32061.0 tgs: 61 data_time: 0.81s time: 520.54s eta: 3 days, 11:34:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.6GB text_tokens: 30810.0 tgs: 59 data_time: 0.70s time: 518.83s eta: 3 days, 11:09:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31204.0 tgs: 59 data_time: 0.89s time: 522.11s eta: 3 days, 11:32:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.336 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.6GB text_tokens: 31557.0 tgs: 60 data_time: 1.16s time: 523.83s eta: 3 days, 11:40:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.9GB text_tokens: 31964.0 tgs: 61 data_time: 0.71s time: 520.46s eta: 3 days, 10:59:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.1GB text_tokens: 31859.0 tgs: 61 data_time: 0.67s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 32340.0 tgs: 61 data_time: 0.87s time: 522.84s eta: 3 days, 11:04:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 32.7GB text_tokens: 31453.0 tgs: 60 data_time: 0.47s time: 523.54s eta: 3 days, 11:02:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31779.0 tgs: 61 data_time: 0.60s time: 520.96s eta: 3 days, 10:29:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.4GB text_tokens: 31169.0 tgs: 60 data_time: 0.79s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 29708.0 tgs: 56 data_time: 0.90s time: 523.29s eta: 3 days, 10:33:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.5GB text_tokens: 31114.0 tgs: 59 data_time: 1.01s time: 522.93s eta: 3 days, 10:21:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 32372.0 tgs: 62 data_time: 0.94s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.0GB text_tokens: 31554.0 tgs: 60 data_time: 0.94s time: 520.64s eta: 3 days, 9:42:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.5GB text_tokens: 31042.0 tgs: 59 data_time: 1.05s time: 522.41s eta: 3 days, 9:50:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.335 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 31728.0 tgs: 60 data_time: 0.81s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.349 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.7GB text_tokens: 31816.0 tgs: 61 data_time: 0.73s time: 520.43s eta: 3 days, 9:14:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 30775.0 tgs: 59 data_time: 0.64s time: 520.68s eta: 3 days, 9:08:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31569.0 tgs: 60 data_time: 0.48s time: 521.30s eta: 3 days, 9:05:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.7GB text_tokens: 31161.0 tgs: 59 data_time: 0.68s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.1GB text_tokens: 31655.0 tgs: 60 data_time: 0.97s time: 520.96s eta: 3 days, 8:44:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31369.0 tgs: 60 data_time: 0.79s time: 520.15s eta: 3 days, 8:28:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32447.0 tgs: 62 data_time: 0.83s time: 522.56s eta: 3 days, 8:42:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 31783.0 tgs: 60 data_time: 0.91s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.7GB text_tokens: 30915.0 tgs: 59 data_time: 0.73s time: 519.88s eta: 3 days, 8:00:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.8GB text_tokens: 30941.0 tgs: 59 data_time: 0.75s time: 520.11s eta: 3 days, 7:53:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 31815.0 tgs: 60 data_time: 0.54s time: 522.00s eta: 3 days, 8:02:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32295.0 tgs: 61 data_time: 0.66s time: 524.35s eta: 3 days, 8:15:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.9GB text_tokens: 32051.0 tgs: 61 data_time: 0.62s time: 520.51s eta: 3 days, 7:31:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.340 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.1GB text_tokens: 32370.0 tgs: 62 data_time: 0.69s time: 518.56s eta: 3 days, 7:04:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31953.0 tgs: 61 data_time: 0.53s time: 523.15s eta: 3 days, 7:38:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31767.0 tgs: 60 data_time: 0.67s time: 523.62s eta: 3 days, 7:33:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 31980.0 tgs: 61 data_time: 0.60s time: 520.17s eta: 3 days, 6:53:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31917.0 tgs: 61 data_time: 0.85s time: 520.05s eta: 3 days, 6:43:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31824.0 tgs: 60 data_time: 0.67s time: 523.76s eta: 3 days, 7:08:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.6GB text_tokens: 30092.0 tgs: 57 data_time: 0.76s time: 523.78s eta: 3 days, 7:00:13 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31783.0 tgs: 61 data_time: 0.71s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.9GB text_tokens: 30978.0 tgs: 59 data_time: 0.72s time: 518.78s eta: 3 days, 5:57:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 31772.0 tgs: 60 data_time: 0.78s time: 522.69s eta: 3 days, 6:24:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31322.0 tgs: 59 data_time: 1.04s time: 522.73s eta: 3 days, 6:15:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.6GB text_tokens: 31440.0 tgs: 60 data_time: 0.72s time: 520.53s eta: 3 days, 5:47:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31790.0 tgs: 61 data_time: 0.84s time: 519.66s eta: 3 days, 5:30:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32257.0 tgs: 61 data_time: 0.76s time: 521.29s eta: 3 days, 5:36:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.7GB text_tokens: 31778.0 tgs: 60 data_time: 0.71s time: 523.59s eta: 3 days, 5:48:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.0GB text_tokens: 31886.0 tgs: 61 data_time: 0.98s time: 519.32s eta: 3 days, 5:01:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 32.9GB text_tokens: 31703.0 tgs: 60 data_time: 0.82s time: 520.65s eta: 3 days, 5:05:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.1GB text_tokens: 31729.0 tgs: 60 data_time: 0.96s time: 521.45s eta: 3 days, 5:03:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.1GB text_tokens: 32208.0 tgs: 61 data_time: 0.79s time: 523.94s eta: 3 days, 5:16:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 31951.0 tgs: 61 data_time: 0.92s time: 518.85s eta: 3 days, 4:23:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.9GB text_tokens: 31537.0 tgs: 60 data_time: 0.56s time: 519.24s eta: 3 days, 4:17:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31597.0 tgs: 60 data_time: 0.75s time: 520.71s eta: 3 days, 4:22:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.8GB text_tokens: 31472.0 tgs: 60 data_time: 0.79s time: 524.25s eta: 3 days, 4:44:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 31847.0 tgs: 61 data_time: 0.89s time: 520.33s eta: 3 days, 4:01:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32423.0 tgs: 62 data_time: 0.91s time: 519.50s eta: 3 days, 3:45:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 32055.0 tgs: 61 data_time: 0.96s time: 522.32s eta: 3 days, 4:01:33 diff --git a/20250120235238/rank15.log b/20250120235238/rank15.log new file mode 100644 index 0000000000000000000000000000000000000000..3c535a8fc939c6043264c6da34b0ec058b161bb3 --- /dev/null +++ b/20250120235238/rank15.log @@ -0,0 +1,395 @@ +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.22s +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.91 seconds, peak gpu memory 13.4G +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.3GB text_tokens: 30603.0 tgs: 55 data_time: 1.99s time: 547.92s eta: 3 days, 18:15:15 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.7GB text_tokens: 31849.0 tgs: 60 data_time: 0.71s time: 523.25s eta: 3 days, 14:02:41 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.5GB text_tokens: 31059.0 tgs: 59 data_time: 0.94s time: 522.88s eta: 3 days, 13:50:21 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 32411.0 tgs: 62 data_time: 0.86s time: 520.29s eta: 3 days, 13:16:11 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 31204.0 tgs: 59 data_time: 0.69s time: 520.98s eta: 3 days, 13:14:19 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.9GB text_tokens: 31449.0 tgs: 60 data_time: 0.86s time: 520.90s eta: 3 days, 13:04:46 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.9GB text_tokens: 31590.0 tgs: 60 data_time: 0.83s time: 523.38s eta: 3 days, 13:20:26 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 32.9GB text_tokens: 31967.0 tgs: 61 data_time: 0.79s time: 520.66s eta: 3 days, 12:45:08 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.394 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 31084.0 tgs: 59 data_time: 0.92s time: 520.18s eta: 3 days, 12:31:43 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.215 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.7GB text_tokens: 31694.0 tgs: 60 data_time: 0.87s time: 520.41s eta: 3 days, 12:25:22 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32132.0 tgs: 61 data_time: 0.94s time: 524.53s eta: 3 days, 12:56:42 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31935.0 tgs: 61 data_time: 0.80s time: 520.65s eta: 3 days, 12:10:18 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 31576.0 tgs: 60 data_time: 0.83s time: 519.93s eta: 3 days, 11:54:36 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.0GB text_tokens: 31631.0 tgs: 60 data_time: 0.82s time: 521.29s eta: 3 days, 11:59:06 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 31904.0 tgs: 60 data_time: 0.85s time: 524.13s eta: 3 days, 12:17:49 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32061.0 tgs: 61 data_time: 0.84s time: 520.54s eta: 3 days, 11:34:31 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.6GB text_tokens: 30810.0 tgs: 59 data_time: 0.73s time: 518.83s eta: 3 days, 11:09:22 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31204.0 tgs: 59 data_time: 0.96s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.6GB text_tokens: 31557.0 tgs: 60 data_time: 1.18s time: 523.83s eta: 3 days, 11:40:01 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.9GB text_tokens: 31964.0 tgs: 61 data_time: 0.73s time: 520.46s eta: 3 days, 10:59:04 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.1GB text_tokens: 31859.0 tgs: 61 data_time: 0.69s time: 518.43s eta: 3 days, 10:31:01 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 32340.0 tgs: 61 data_time: 0.89s time: 522.84s eta: 3 days, 11:04:24 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 32.7GB text_tokens: 31453.0 tgs: 60 data_time: 0.48s time: 523.54s eta: 3 days, 11:02:20 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31779.0 tgs: 61 data_time: 0.61s time: 520.96s eta: 3 days, 10:29:09 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.4GB text_tokens: 31169.0 tgs: 60 data_time: 0.82s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 29708.0 tgs: 56 data_time: 0.96s time: 523.29s eta: 3 days, 10:33:50 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.5GB text_tokens: 31114.0 tgs: 59 data_time: 1.03s time: 522.93s eta: 3 days, 10:21:43 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.317 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 32372.0 tgs: 62 data_time: 0.94s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.0GB text_tokens: 31554.0 tgs: 60 data_time: 0.97s time: 520.64s eta: 3 days, 9:42:39 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.5GB text_tokens: 31042.0 tgs: 59 data_time: 1.07s time: 522.41s eta: 3 days, 9:50:39 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 31728.0 tgs: 60 data_time: 0.84s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.7GB text_tokens: 31816.0 tgs: 61 data_time: 0.73s time: 520.43s eta: 3 days, 9:14:39 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 30775.0 tgs: 59 data_time: 0.66s time: 520.68s eta: 3 days, 9:08:19 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31569.0 tgs: 60 data_time: 0.50s time: 521.30s eta: 3 days, 9:05:28 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.7GB text_tokens: 31161.0 tgs: 59 data_time: 0.70s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.1GB text_tokens: 31655.0 tgs: 60 data_time: 0.99s time: 520.96s eta: 3 days, 8:44:54 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31369.0 tgs: 60 data_time: 0.82s time: 520.15s eta: 3 days, 8:28:43 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.208 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32447.0 tgs: 62 data_time: 0.84s time: 522.56s eta: 3 days, 8:42:24 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 31783.0 tgs: 60 data_time: 0.93s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.7GB text_tokens: 30915.0 tgs: 59 data_time: 0.76s time: 519.89s eta: 3 days, 8:00:16 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.8GB text_tokens: 30941.0 tgs: 59 data_time: 0.75s time: 520.11s eta: 3 days, 7:53:42 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 31815.0 tgs: 60 data_time: 0.58s time: 522.00s eta: 3 days, 8:02:22 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32295.0 tgs: 61 data_time: 0.68s time: 524.35s eta: 3 days, 8:15:17 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.9GB text_tokens: 32051.0 tgs: 61 data_time: 0.63s time: 520.50s eta: 3 days, 7:31:17 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.1GB text_tokens: 32370.0 tgs: 62 data_time: 0.72s time: 518.56s eta: 3 days, 7:04:52 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31953.0 tgs: 61 data_time: 0.55s time: 523.15s eta: 3 days, 7:38:08 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31767.0 tgs: 60 data_time: 0.68s time: 523.62s eta: 3 days, 7:33:41 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 31980.0 tgs: 61 data_time: 0.62s time: 520.18s eta: 3 days, 6:53:35 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31917.0 tgs: 61 data_time: 0.86s time: 520.05s eta: 3 days, 6:43:49 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31824.0 tgs: 60 data_time: 0.65s time: 523.76s eta: 3 days, 7:08:44 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.6GB text_tokens: 30092.0 tgs: 57 data_time: 0.77s time: 523.78s eta: 3 days, 7:00:12 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.334 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31783.0 tgs: 61 data_time: 0.73s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.9GB text_tokens: 30978.0 tgs: 59 data_time: 0.73s time: 518.78s eta: 3 days, 5:57:37 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 31772.0 tgs: 60 data_time: 0.79s time: 522.69s eta: 3 days, 6:24:09 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31322.0 tgs: 59 data_time: 1.04s time: 522.73s eta: 3 days, 6:15:51 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.333 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.6GB text_tokens: 31440.0 tgs: 60 data_time: 0.73s time: 520.53s eta: 3 days, 5:47:22 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31790.0 tgs: 61 data_time: 0.86s time: 519.66s eta: 3 days, 5:30:56 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32257.0 tgs: 61 data_time: 0.77s time: 521.30s eta: 3 days, 5:36:54 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.7GB text_tokens: 31778.0 tgs: 60 data_time: 0.73s time: 523.59s eta: 3 days, 5:48:40 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.316 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.0GB text_tokens: 31886.0 tgs: 61 data_time: 1.01s time: 519.32s eta: 3 days, 5:01:58 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 32.9GB text_tokens: 31703.0 tgs: 60 data_time: 0.84s time: 520.65s eta: 3 days, 5:05:07 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.1GB text_tokens: 31729.0 tgs: 60 data_time: 0.97s time: 521.45s eta: 3 days, 5:03:31 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.342 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.1GB text_tokens: 32208.0 tgs: 61 data_time: 0.80s time: 523.94s eta: 3 days, 5:16:54 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 31951.0 tgs: 61 data_time: 0.94s time: 518.85s eta: 3 days, 4:23:11 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.9GB text_tokens: 31537.0 tgs: 60 data_time: 0.58s time: 519.24s eta: 3 days, 4:17:57 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31597.0 tgs: 60 data_time: 0.78s time: 520.71s eta: 3 days, 4:22:17 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.8GB text_tokens: 31472.0 tgs: 60 data_time: 0.80s time: 524.25s eta: 3 days, 4:44:41 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 31847.0 tgs: 61 data_time: 0.91s time: 520.33s eta: 3 days, 4:01:33 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32423.0 tgs: 62 data_time: 0.93s time: 519.50s eta: 3 days, 3:45:35 +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 15][DP 3][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 32055.0 tgs: 61 data_time: 0.97s time: 522.32s eta: 3 days, 4:01:34 diff --git a/20250120235238/rank17.log b/20250120235238/rank17.log new file mode 100644 index 0000000000000000000000000000000000000000..aec46d8a3f17298d4230dc82a67a1d34e858634e --- /dev/null +++ b/20250120235238/rank17.log @@ -0,0 +1,395 @@ +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.17s +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.27 seconds, peak gpu memory 13.4G +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31743.0 tgs: 58 data_time: 1.81s time: 547.02s eta: 3 days, 18:06:20 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.0GB text_tokens: 32110.0 tgs: 61 data_time: 0.92s time: 523.24s eta: 3 days, 14:02:39 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.6GB text_tokens: 31536.0 tgs: 60 data_time: 0.84s time: 522.89s eta: 3 days, 13:50:27 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31549.0 tgs: 60 data_time: 1.03s time: 520.29s eta: 3 days, 13:16:11 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31125.0 tgs: 59 data_time: 0.85s time: 520.99s eta: 3 days, 13:14:20 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.7GB text_tokens: 31488.0 tgs: 60 data_time: 1.13s time: 520.87s eta: 3 days, 13:04:34 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 30800.0 tgs: 58 data_time: 0.76s time: 523.38s eta: 3 days, 13:20:26 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.308 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 32301.0 tgs: 62 data_time: 0.94s time: 520.66s eta: 3 days, 12:45:08 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31885.0 tgs: 61 data_time: 0.92s time: 520.18s eta: 3 days, 12:31:45 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.337 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.0GB text_tokens: 32246.0 tgs: 61 data_time: 0.83s time: 520.43s eta: 3 days, 12:25:31 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 32255.0 tgs: 61 data_time: 0.79s time: 524.53s eta: 3 days, 12:56:43 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31976.0 tgs: 61 data_time: 0.86s time: 520.65s eta: 3 days, 12:10:18 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 32057.0 tgs: 61 data_time: 0.72s time: 519.92s eta: 3 days, 11:54:35 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.0GB text_tokens: 31783.0 tgs: 60 data_time: 0.79s time: 521.29s eta: 3 days, 11:59:09 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.6GB text_tokens: 30276.0 tgs: 57 data_time: 0.68s time: 524.12s eta: 3 days, 12:17:48 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.0GB text_tokens: 32240.0 tgs: 61 data_time: 0.74s time: 520.54s eta: 3 days, 11:34:31 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32111.0 tgs: 61 data_time: 0.89s time: 518.81s eta: 3 days, 11:09:11 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 30872.0 tgs: 59 data_time: 0.75s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31286.0 tgs: 59 data_time: 0.96s time: 523.83s eta: 3 days, 11:40:02 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.9GB text_tokens: 31340.0 tgs: 60 data_time: 0.68s time: 520.47s eta: 3 days, 10:59:11 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31187.0 tgs: 60 data_time: 0.86s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.7GB text_tokens: 31884.0 tgs: 60 data_time: 0.89s time: 522.84s eta: 3 days, 11:04:24 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 32216.0 tgs: 61 data_time: 0.86s time: 523.51s eta: 3 days, 11:02:04 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.0GB text_tokens: 31832.0 tgs: 61 data_time: 0.64s time: 520.96s eta: 3 days, 10:29:07 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 31974.0 tgs: 61 data_time: 0.93s time: 519.10s eta: 3 days, 10:02:49 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31429.0 tgs: 60 data_time: 0.74s time: 523.29s eta: 3 days, 10:33:50 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31130.0 tgs: 59 data_time: 0.93s time: 522.95s eta: 3 days, 10:21:50 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.6GB text_tokens: 30936.0 tgs: 59 data_time: 0.57s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.0GB text_tokens: 31633.0 tgs: 60 data_time: 0.68s time: 520.64s eta: 3 days, 9:42:39 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.1GB text_tokens: 31970.0 tgs: 61 data_time: 0.97s time: 522.40s eta: 3 days, 9:50:32 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.210 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.8GB text_tokens: 25656.0 tgs: 48 data_time: 0.79s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31702.0 tgs: 60 data_time: 0.87s time: 520.43s eta: 3 days, 9:14:39 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.8GB text_tokens: 31282.0 tgs: 60 data_time: 0.65s time: 520.68s eta: 3 days, 9:08:20 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31475.0 tgs: 60 data_time: 0.54s time: 521.29s eta: 3 days, 9:05:20 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 31004.0 tgs: 59 data_time: 0.90s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.9GB text_tokens: 31063.0 tgs: 59 data_time: 0.88s time: 520.96s eta: 3 days, 8:44:56 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.8GB text_tokens: 31013.0 tgs: 59 data_time: 0.89s time: 520.16s eta: 3 days, 8:28:50 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.355 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32430.0 tgs: 62 data_time: 0.75s time: 522.56s eta: 3 days, 8:42:25 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32259.0 tgs: 61 data_time: 0.56s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.9GB text_tokens: 31282.0 tgs: 60 data_time: 0.81s time: 519.89s eta: 3 days, 8:00:16 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.0GB text_tokens: 31847.0 tgs: 61 data_time: 0.72s time: 520.10s eta: 3 days, 7:53:33 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32064.0 tgs: 61 data_time: 0.55s time: 522.00s eta: 3 days, 8:02:22 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 30877.0 tgs: 58 data_time: 0.57s time: 524.35s eta: 3 days, 8:15:18 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.0GB text_tokens: 31532.0 tgs: 60 data_time: 0.66s time: 520.51s eta: 3 days, 7:31:20 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.8GB text_tokens: 31384.0 tgs: 60 data_time: 1.06s time: 518.57s eta: 3 days, 7:04:52 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31932.0 tgs: 61 data_time: 0.90s time: 523.15s eta: 3 days, 7:38:08 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.0GB text_tokens: 31103.0 tgs: 59 data_time: 0.67s time: 523.62s eta: 3 days, 7:33:42 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.0GB text_tokens: 32490.0 tgs: 62 data_time: 0.93s time: 520.16s eta: 3 days, 6:53:27 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 32230.0 tgs: 61 data_time: 0.92s time: 520.05s eta: 3 days, 6:43:49 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31721.0 tgs: 60 data_time: 0.93s time: 523.76s eta: 3 days, 7:08:45 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.9GB text_tokens: 31830.0 tgs: 60 data_time: 0.81s time: 523.77s eta: 3 days, 7:00:04 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.9GB text_tokens: 30994.0 tgs: 59 data_time: 0.73s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 31322.0 tgs: 60 data_time: 0.86s time: 518.78s eta: 3 days, 5:57:38 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.9GB text_tokens: 32082.0 tgs: 61 data_time: 0.77s time: 522.69s eta: 3 days, 6:24:11 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31833.0 tgs: 60 data_time: 0.77s time: 522.74s eta: 3 days, 6:15:57 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.0GB text_tokens: 31319.0 tgs: 60 data_time: 0.85s time: 520.53s eta: 3 days, 5:47:24 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.8GB text_tokens: 29936.0 tgs: 57 data_time: 0.94s time: 519.66s eta: 3 days, 5:30:55 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.0GB text_tokens: 32364.0 tgs: 62 data_time: 0.87s time: 521.28s eta: 3 days, 5:36:47 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.8GB text_tokens: 31346.0 tgs: 59 data_time: 0.68s time: 523.59s eta: 3 days, 5:48:41 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31934.0 tgs: 61 data_time: 0.78s time: 519.32s eta: 3 days, 5:01:58 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32223.0 tgs: 61 data_time: 0.82s time: 520.65s eta: 3 days, 5:05:07 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31650.0 tgs: 60 data_time: 0.80s time: 521.45s eta: 3 days, 5:03:31 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 32139.0 tgs: 61 data_time: 0.75s time: 523.95s eta: 3 days, 5:16:55 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 30916.0 tgs: 59 data_time: 0.63s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31094.0 tgs: 59 data_time: 0.95s time: 519.25s eta: 3 days, 4:18:02 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31963.0 tgs: 61 data_time: 0.83s time: 520.71s eta: 3 days, 4:22:17 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 31710.0 tgs: 60 data_time: 0.88s time: 524.26s eta: 3 days, 4:44:42 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32386.0 tgs: 62 data_time: 0.93s time: 520.29s eta: 3 days, 4:01:13 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32507.0 tgs: 62 data_time: 0.75s time: 519.50s eta: 3 days, 3:45:35 +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 17][DP 4][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31766.0 tgs: 60 data_time: 0.59s time: 522.32s eta: 3 days, 4:01:35 diff --git a/20250120235238/rank18.log b/20250120235238/rank18.log new file mode 100644 index 0000000000000000000000000000000000000000..b41afd895d38bb9b5d91a0f1e4e12d118a89e233 --- /dev/null +++ b/20250120235238/rank18.log @@ -0,0 +1,395 @@ +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.12s +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.71 seconds, peak gpu memory 13.4G +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31743.0 tgs: 58 data_time: 1.77s time: 547.16s eta: 3 days, 18:07:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.335 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.0GB text_tokens: 32110.0 tgs: 61 data_time: 0.88s time: 523.25s eta: 3 days, 14:02:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.6GB text_tokens: 31536.0 tgs: 60 data_time: 0.82s time: 522.89s eta: 3 days, 13:50:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31549.0 tgs: 60 data_time: 1.03s time: 520.29s eta: 3 days, 13:16:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31125.0 tgs: 59 data_time: 0.84s time: 520.99s eta: 3 days, 13:14:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.7GB text_tokens: 31488.0 tgs: 60 data_time: 1.09s time: 520.87s eta: 3 days, 13:04:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 30800.0 tgs: 58 data_time: 0.74s time: 523.39s eta: 3 days, 13:20:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 32301.0 tgs: 62 data_time: 0.91s time: 520.66s eta: 3 days, 12:45:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31885.0 tgs: 61 data_time: 0.90s time: 520.18s eta: 3 days, 12:31:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.0GB text_tokens: 32246.0 tgs: 61 data_time: 0.78s time: 520.43s eta: 3 days, 12:25:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 32255.0 tgs: 61 data_time: 0.76s time: 524.53s eta: 3 days, 12:56:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31976.0 tgs: 61 data_time: 0.83s time: 520.65s eta: 3 days, 12:10:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 32057.0 tgs: 61 data_time: 0.71s time: 519.92s eta: 3 days, 11:54:36 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.0GB text_tokens: 31783.0 tgs: 60 data_time: 0.76s time: 521.29s eta: 3 days, 11:59:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.6GB text_tokens: 30276.0 tgs: 57 data_time: 0.67s time: 524.13s eta: 3 days, 12:17:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.0GB text_tokens: 32240.0 tgs: 61 data_time: 0.72s time: 520.54s eta: 3 days, 11:34:33 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32111.0 tgs: 61 data_time: 0.86s time: 518.81s eta: 3 days, 11:09:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 30872.0 tgs: 59 data_time: 0.73s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31286.0 tgs: 59 data_time: 0.94s time: 523.83s eta: 3 days, 11:40:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.9GB text_tokens: 31340.0 tgs: 60 data_time: 0.67s time: 520.47s eta: 3 days, 10:59:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31187.0 tgs: 60 data_time: 0.83s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.7GB text_tokens: 31884.0 tgs: 60 data_time: 0.87s time: 522.84s eta: 3 days, 11:04:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 32216.0 tgs: 61 data_time: 0.82s time: 523.51s eta: 3 days, 11:02:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.0GB text_tokens: 31832.0 tgs: 61 data_time: 0.61s time: 520.97s eta: 3 days, 10:29:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 31974.0 tgs: 61 data_time: 0.91s time: 519.09s eta: 3 days, 10:02:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31429.0 tgs: 60 data_time: 0.72s time: 523.29s eta: 3 days, 10:33:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31130.0 tgs: 59 data_time: 0.92s time: 522.95s eta: 3 days, 10:21:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.6GB text_tokens: 30936.0 tgs: 59 data_time: 0.56s time: 521.32s eta: 3 days, 9:57:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.0GB text_tokens: 31633.0 tgs: 60 data_time: 0.66s time: 520.64s eta: 3 days, 9:42:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.1GB text_tokens: 31970.0 tgs: 61 data_time: 0.94s time: 522.40s eta: 3 days, 9:50:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.8GB text_tokens: 25656.0 tgs: 48 data_time: 0.77s time: 523.75s eta: 3 days, 9:54:33 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31702.0 tgs: 60 data_time: 0.84s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.8GB text_tokens: 31282.0 tgs: 60 data_time: 0.63s time: 520.68s eta: 3 days, 9:08:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31475.0 tgs: 60 data_time: 0.52s time: 521.29s eta: 3 days, 9:05:21 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 31004.0 tgs: 59 data_time: 0.85s time: 524.11s eta: 3 days, 9:22:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.9GB text_tokens: 31063.0 tgs: 59 data_time: 0.86s time: 520.96s eta: 3 days, 8:44:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.8GB text_tokens: 31013.0 tgs: 59 data_time: 0.88s time: 520.17s eta: 3 days, 8:28:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32430.0 tgs: 62 data_time: 0.73s time: 522.56s eta: 3 days, 8:42:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32259.0 tgs: 61 data_time: 0.55s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.9GB text_tokens: 31282.0 tgs: 60 data_time: 0.79s time: 519.89s eta: 3 days, 8:00:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.0GB text_tokens: 31847.0 tgs: 61 data_time: 0.70s time: 520.10s eta: 3 days, 7:53:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32064.0 tgs: 61 data_time: 0.54s time: 522.00s eta: 3 days, 8:02:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 30877.0 tgs: 58 data_time: 0.54s time: 524.35s eta: 3 days, 8:15:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.0GB text_tokens: 31532.0 tgs: 60 data_time: 0.65s time: 520.51s eta: 3 days, 7:31:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.8GB text_tokens: 31384.0 tgs: 60 data_time: 1.04s time: 518.57s eta: 3 days, 7:04:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31932.0 tgs: 61 data_time: 0.89s time: 523.15s eta: 3 days, 7:38:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.0GB text_tokens: 31103.0 tgs: 59 data_time: 0.65s time: 523.63s eta: 3 days, 7:33:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.0GB text_tokens: 32490.0 tgs: 62 data_time: 0.91s time: 520.16s eta: 3 days, 6:53:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 32230.0 tgs: 61 data_time: 0.91s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.339 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31721.0 tgs: 60 data_time: 0.92s time: 523.76s eta: 3 days, 7:08:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.9GB text_tokens: 31830.0 tgs: 60 data_time: 0.81s time: 523.77s eta: 3 days, 7:00:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.337 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.9GB text_tokens: 30994.0 tgs: 59 data_time: 0.72s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 31322.0 tgs: 60 data_time: 0.89s time: 518.78s eta: 3 days, 5:57:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.9GB text_tokens: 32082.0 tgs: 61 data_time: 0.75s time: 522.69s eta: 3 days, 6:24:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.315 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31833.0 tgs: 60 data_time: 0.75s time: 522.74s eta: 3 days, 6:15:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.0GB text_tokens: 31319.0 tgs: 60 data_time: 0.84s time: 520.53s eta: 3 days, 5:47:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.8GB text_tokens: 29936.0 tgs: 57 data_time: 0.92s time: 519.66s eta: 3 days, 5:30:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.0GB text_tokens: 32364.0 tgs: 62 data_time: 0.86s time: 521.28s eta: 3 days, 5:36:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.8GB text_tokens: 31346.0 tgs: 59 data_time: 0.66s time: 523.59s eta: 3 days, 5:48:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31934.0 tgs: 61 data_time: 0.77s time: 519.32s eta: 3 days, 5:01:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32223.0 tgs: 61 data_time: 0.81s time: 520.65s eta: 3 days, 5:05:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31650.0 tgs: 60 data_time: 0.79s time: 521.45s eta: 3 days, 5:03:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.206 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 32139.0 tgs: 61 data_time: 0.75s time: 523.95s eta: 3 days, 5:16:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 30916.0 tgs: 59 data_time: 0.62s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31094.0 tgs: 59 data_time: 0.91s time: 519.25s eta: 3 days, 4:18:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31963.0 tgs: 61 data_time: 0.83s time: 520.72s eta: 3 days, 4:22:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 31710.0 tgs: 60 data_time: 0.86s time: 524.25s eta: 3 days, 4:44:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32386.0 tgs: 62 data_time: 0.92s time: 520.29s eta: 3 days, 4:01:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32507.0 tgs: 62 data_time: 0.74s time: 519.50s eta: 3 days, 3:45:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31766.0 tgs: 60 data_time: 0.59s time: 522.32s eta: 3 days, 4:01:35 diff --git a/20250120235238/rank21.log b/20250120235238/rank21.log new file mode 100644 index 0000000000000000000000000000000000000000..52136d4bffca8549eed110bf8cb30a0b9088831c --- /dev/null +++ b/20250120235238/rank21.log @@ -0,0 +1,395 @@ +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.15s +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.70 seconds, peak gpu memory 13.4G +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.3GB text_tokens: 29620.0 tgs: 54 data_time: 1.87s time: 547.04s eta: 3 days, 18:06:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32369.0 tgs: 61 data_time: 0.80s time: 523.24s eta: 3 days, 14:02:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 32546.0 tgs: 62 data_time: 0.79s time: 522.89s eta: 3 days, 13:50:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 32411.0 tgs: 62 data_time: 1.04s time: 520.29s eta: 3 days, 13:16:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 30927.0 tgs: 59 data_time: 0.90s time: 520.99s eta: 3 days, 13:14:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.0GB text_tokens: 32427.0 tgs: 62 data_time: 1.18s time: 520.88s eta: 3 days, 13:04:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 31527.0 tgs: 60 data_time: 1.10s time: 523.38s eta: 3 days, 13:20:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.331 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31799.0 tgs: 61 data_time: 0.68s time: 520.66s eta: 3 days, 12:45:08 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 30776.0 tgs: 59 data_time: 0.71s time: 520.18s eta: 3 days, 12:31:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.0GB text_tokens: 31699.0 tgs: 60 data_time: 0.72s time: 520.43s eta: 3 days, 12:25:31 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 32.8GB text_tokens: 31472.0 tgs: 59 data_time: 0.98s time: 524.53s eta: 3 days, 12:56:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 31345.0 tgs: 60 data_time: 0.67s time: 520.65s eta: 3 days, 12:10:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.332 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 30795.0 tgs: 59 data_time: 0.76s time: 519.92s eta: 3 days, 11:54:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 31328.0 tgs: 60 data_time: 0.73s time: 521.29s eta: 3 days, 11:59:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 32265.0 tgs: 61 data_time: 0.83s time: 524.13s eta: 3 days, 12:17:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.8GB text_tokens: 29937.0 tgs: 57 data_time: 0.77s time: 520.54s eta: 3 days, 11:34:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 30786.0 tgs: 59 data_time: 0.74s time: 518.81s eta: 3 days, 11:09:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31670.0 tgs: 60 data_time: 0.82s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31374.0 tgs: 59 data_time: 0.61s time: 523.83s eta: 3 days, 11:40:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32080.0 tgs: 61 data_time: 0.74s time: 520.47s eta: 3 days, 10:59:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 30100.0 tgs: 58 data_time: 0.76s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.8GB text_tokens: 31525.0 tgs: 60 data_time: 0.87s time: 522.84s eta: 3 days, 11:04:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.383 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32200.0 tgs: 61 data_time: 0.89s time: 523.51s eta: 3 days, 11:02:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.0GB text_tokens: 31533.0 tgs: 60 data_time: 1.02s time: 520.96s eta: 3 days, 10:29:07 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 31572.0 tgs: 60 data_time: 0.76s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31711.0 tgs: 60 data_time: 0.56s time: 523.29s eta: 3 days, 10:33:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.7GB text_tokens: 31027.0 tgs: 59 data_time: 0.78s time: 522.95s eta: 3 days, 10:21:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.8GB text_tokens: 31174.0 tgs: 59 data_time: 1.02s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.0GB text_tokens: 31786.0 tgs: 61 data_time: 0.89s time: 520.64s eta: 3 days, 9:42:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.9GB text_tokens: 31831.0 tgs: 60 data_time: 0.91s time: 522.40s eta: 3 days, 9:50:31 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.9GB text_tokens: 31822.0 tgs: 60 data_time: 0.86s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.9GB text_tokens: 31024.0 tgs: 59 data_time: 0.75s time: 520.43s eta: 3 days, 9:14:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.1GB text_tokens: 31792.0 tgs: 61 data_time: 1.00s time: 520.68s eta: 3 days, 9:08:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.209 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.7GB text_tokens: 31319.0 tgs: 60 data_time: 0.62s time: 521.29s eta: 3 days, 9:05:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 32256.0 tgs: 61 data_time: 0.63s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 32359.0 tgs: 62 data_time: 0.87s time: 520.96s eta: 3 days, 8:44:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.7GB text_tokens: 31433.0 tgs: 60 data_time: 0.89s time: 520.16s eta: 3 days, 8:28:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.0GB text_tokens: 30457.0 tgs: 58 data_time: 0.86s time: 522.56s eta: 3 days, 8:42:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.205 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.9GB text_tokens: 32061.0 tgs: 61 data_time: 0.85s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.9GB text_tokens: 30958.0 tgs: 59 data_time: 0.68s time: 519.89s eta: 3 days, 8:00:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32067.0 tgs: 61 data_time: 0.85s time: 520.10s eta: 3 days, 7:53:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 31896.0 tgs: 61 data_time: 0.93s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.8GB text_tokens: 30763.0 tgs: 58 data_time: 0.74s time: 524.35s eta: 3 days, 8:15:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.9GB text_tokens: 31487.0 tgs: 60 data_time: 0.88s time: 520.51s eta: 3 days, 7:31:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.1GB text_tokens: 31702.0 tgs: 61 data_time: 0.71s time: 518.56s eta: 3 days, 7:04:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 32.9GB text_tokens: 31920.0 tgs: 61 data_time: 0.91s time: 523.16s eta: 3 days, 7:38:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.0GB text_tokens: 31789.0 tgs: 60 data_time: 0.75s time: 523.62s eta: 3 days, 7:33:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.219 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31065.0 tgs: 59 data_time: 0.74s time: 520.16s eta: 3 days, 6:53:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 32174.0 tgs: 61 data_time: 0.77s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31306.0 tgs: 59 data_time: 1.02s time: 523.76s eta: 3 days, 7:08:44 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 29824.0 tgs: 56 data_time: 0.68s time: 523.77s eta: 3 days, 7:00:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.7GB text_tokens: 31180.0 tgs: 59 data_time: 1.08s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.0GB text_tokens: 31929.0 tgs: 61 data_time: 0.88s time: 518.78s eta: 3 days, 5:57:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 32302.0 tgs: 61 data_time: 0.55s time: 522.68s eta: 3 days, 6:24:09 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31344.0 tgs: 59 data_time: 0.68s time: 522.74s eta: 3 days, 6:15:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.334 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.8GB text_tokens: 30272.0 tgs: 58 data_time: 0.78s time: 520.52s eta: 3 days, 5:47:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32061.0 tgs: 61 data_time: 0.69s time: 519.66s eta: 3 days, 5:30:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 31885.0 tgs: 61 data_time: 0.67s time: 521.28s eta: 3 days, 5:36:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.1GB text_tokens: 32402.0 tgs: 61 data_time: 1.21s time: 523.59s eta: 3 days, 5:48:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.4GB text_tokens: 30488.0 tgs: 58 data_time: 0.85s time: 519.32s eta: 3 days, 5:01:58 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 31879.0 tgs: 61 data_time: 0.57s time: 520.65s eta: 3 days, 5:05:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31778.0 tgs: 60 data_time: 0.58s time: 521.45s eta: 3 days, 5:03:32 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.333 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.8GB text_tokens: 31838.0 tgs: 60 data_time: 0.94s time: 523.95s eta: 3 days, 5:16:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.352 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 32028.0 tgs: 61 data_time: 1.34s time: 518.85s eta: 3 days, 4:23:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 30998.0 tgs: 59 data_time: 0.67s time: 519.25s eta: 3 days, 4:18:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31047.0 tgs: 59 data_time: 0.88s time: 520.72s eta: 3 days, 4:22:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 32302.0 tgs: 61 data_time: 0.86s time: 524.25s eta: 3 days, 4:44:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.6GB text_tokens: 30866.0 tgs: 59 data_time: 0.61s time: 520.29s eta: 3 days, 4:01:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31849.0 tgs: 61 data_time: 0.74s time: 519.50s eta: 3 days, 3:45:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.363 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31641.0 tgs: 60 data_time: 0.81s time: 522.32s eta: 3 days, 4:01:34 diff --git a/20250120235238/rank23.log b/20250120235238/rank23.log new file mode 100644 index 0000000000000000000000000000000000000000..552e04b210923f2037f6e502d2804a347804fac7 --- /dev/null +++ b/20250120235238/rank23.log @@ -0,0 +1,395 @@ +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.14s +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.69 seconds, peak gpu memory 13.4G +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.3GB text_tokens: 29620.0 tgs: 54 data_time: 1.88s time: 547.16s eta: 3 days, 18:07:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32369.0 tgs: 61 data_time: 0.79s time: 523.25s eta: 3 days, 14:02:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 32546.0 tgs: 62 data_time: 0.78s time: 522.89s eta: 3 days, 13:50:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.338 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 32411.0 tgs: 62 data_time: 1.03s time: 520.29s eta: 3 days, 13:16:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.319 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 30927.0 tgs: 59 data_time: 0.89s time: 520.98s eta: 3 days, 13:14:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.0GB text_tokens: 32427.0 tgs: 62 data_time: 1.15s time: 520.87s eta: 3 days, 13:04:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 31527.0 tgs: 60 data_time: 1.08s time: 523.39s eta: 3 days, 13:20:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31799.0 tgs: 61 data_time: 0.65s time: 520.66s eta: 3 days, 12:45:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 30776.0 tgs: 59 data_time: 0.70s time: 520.18s eta: 3 days, 12:31:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.0GB text_tokens: 31699.0 tgs: 60 data_time: 0.70s time: 520.43s eta: 3 days, 12:25:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 32.8GB text_tokens: 31472.0 tgs: 59 data_time: 0.97s time: 524.53s eta: 3 days, 12:56:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.219 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 31345.0 tgs: 60 data_time: 0.66s time: 520.65s eta: 3 days, 12:10:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 30795.0 tgs: 59 data_time: 0.73s time: 519.93s eta: 3 days, 11:54:36 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 31328.0 tgs: 60 data_time: 0.71s time: 521.29s eta: 3 days, 11:59:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 32265.0 tgs: 61 data_time: 0.81s time: 524.13s eta: 3 days, 12:17:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.8GB text_tokens: 29937.0 tgs: 57 data_time: 0.75s time: 520.54s eta: 3 days, 11:34:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.360 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 30786.0 tgs: 59 data_time: 0.74s time: 518.81s eta: 3 days, 11:09:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31670.0 tgs: 60 data_time: 0.81s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31374.0 tgs: 59 data_time: 0.56s time: 523.83s eta: 3 days, 11:40:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32080.0 tgs: 61 data_time: 0.72s time: 520.47s eta: 3 days, 10:59:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 30100.0 tgs: 58 data_time: 0.75s time: 518.43s eta: 3 days, 10:31:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.8GB text_tokens: 31525.0 tgs: 60 data_time: 0.88s time: 522.84s eta: 3 days, 11:04:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32200.0 tgs: 61 data_time: 0.87s time: 523.51s eta: 3 days, 11:02:04 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.0GB text_tokens: 31533.0 tgs: 60 data_time: 0.96s time: 520.96s eta: 3 days, 10:29:07 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 31572.0 tgs: 60 data_time: 0.73s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31711.0 tgs: 60 data_time: 0.55s time: 523.30s eta: 3 days, 10:33:51 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.7GB text_tokens: 31027.0 tgs: 59 data_time: 0.77s time: 522.95s eta: 3 days, 10:21:51 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.8GB text_tokens: 31174.0 tgs: 59 data_time: 1.00s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.0GB text_tokens: 31786.0 tgs: 61 data_time: 0.87s time: 520.64s eta: 3 days, 9:42:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.9GB text_tokens: 31831.0 tgs: 60 data_time: 0.89s time: 522.40s eta: 3 days, 9:50:31 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.9GB text_tokens: 31822.0 tgs: 60 data_time: 0.85s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.9GB text_tokens: 31024.0 tgs: 59 data_time: 0.74s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.1GB text_tokens: 31792.0 tgs: 61 data_time: 0.98s time: 520.68s eta: 3 days, 9:08:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.7GB text_tokens: 31319.0 tgs: 60 data_time: 0.59s time: 521.29s eta: 3 days, 9:05:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.341 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 32256.0 tgs: 61 data_time: 0.62s time: 524.11s eta: 3 days, 9:22:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 32359.0 tgs: 62 data_time: 0.83s time: 520.96s eta: 3 days, 8:44:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.7GB text_tokens: 31433.0 tgs: 60 data_time: 0.88s time: 520.17s eta: 3 days, 8:28:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.0GB text_tokens: 30457.0 tgs: 58 data_time: 0.84s time: 522.56s eta: 3 days, 8:42:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.9GB text_tokens: 32061.0 tgs: 61 data_time: 0.84s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.9GB text_tokens: 30958.0 tgs: 59 data_time: 0.65s time: 519.89s eta: 3 days, 8:00:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32067.0 tgs: 61 data_time: 0.83s time: 520.10s eta: 3 days, 7:53:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 31896.0 tgs: 61 data_time: 0.92s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.8GB text_tokens: 30763.0 tgs: 58 data_time: 0.72s time: 524.35s eta: 3 days, 8:15:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.9GB text_tokens: 31487.0 tgs: 60 data_time: 0.87s time: 520.51s eta: 3 days, 7:31:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.1GB text_tokens: 31702.0 tgs: 61 data_time: 0.70s time: 518.57s eta: 3 days, 7:04:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 32.9GB text_tokens: 31920.0 tgs: 61 data_time: 0.92s time: 523.16s eta: 3 days, 7:38:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.0GB text_tokens: 31789.0 tgs: 60 data_time: 0.75s time: 523.63s eta: 3 days, 7:33:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31065.0 tgs: 59 data_time: 0.77s time: 520.16s eta: 3 days, 6:53:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 32174.0 tgs: 61 data_time: 0.75s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.346 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31306.0 tgs: 59 data_time: 1.02s time: 523.76s eta: 3 days, 7:08:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 29824.0 tgs: 56 data_time: 0.66s time: 523.77s eta: 3 days, 7:00:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.7GB text_tokens: 31180.0 tgs: 59 data_time: 1.06s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.0GB text_tokens: 31929.0 tgs: 61 data_time: 0.86s time: 518.78s eta: 3 days, 5:57:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 32302.0 tgs: 61 data_time: 0.54s time: 522.69s eta: 3 days, 6:24:10 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31344.0 tgs: 59 data_time: 0.66s time: 522.74s eta: 3 days, 6:15:57 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.8GB text_tokens: 30272.0 tgs: 58 data_time: 0.76s time: 520.53s eta: 3 days, 5:47:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32061.0 tgs: 61 data_time: 0.67s time: 519.66s eta: 3 days, 5:30:57 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 31885.0 tgs: 61 data_time: 0.66s time: 521.28s eta: 3 days, 5:36:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.1GB text_tokens: 32402.0 tgs: 61 data_time: 1.18s time: 523.59s eta: 3 days, 5:48:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.4GB text_tokens: 30488.0 tgs: 58 data_time: 0.85s time: 519.32s eta: 3 days, 5:01:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 31879.0 tgs: 61 data_time: 0.55s time: 520.65s eta: 3 days, 5:05:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31778.0 tgs: 60 data_time: 0.56s time: 521.45s eta: 3 days, 5:03:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.8GB text_tokens: 31838.0 tgs: 60 data_time: 0.92s time: 523.95s eta: 3 days, 5:16:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 32028.0 tgs: 61 data_time: 1.33s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 30998.0 tgs: 59 data_time: 0.65s time: 519.25s eta: 3 days, 4:18:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.194 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31047.0 tgs: 59 data_time: 0.85s time: 520.72s eta: 3 days, 4:22:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 32302.0 tgs: 61 data_time: 0.84s time: 524.25s eta: 3 days, 4:44:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.6GB text_tokens: 30866.0 tgs: 59 data_time: 0.61s time: 520.29s eta: 3 days, 4:01:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31849.0 tgs: 61 data_time: 0.72s time: 519.50s eta: 3 days, 3:45:36 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31641.0 tgs: 60 data_time: 0.80s time: 522.32s eta: 3 days, 4:01:35 diff --git a/20250120235238/rank25.log b/20250120235238/rank25.log new file mode 100644 index 0000000000000000000000000000000000000000..a409481e6a53e3d8d9c4c31aa6d7697c4ebf6676 --- /dev/null +++ b/20250120235238/rank25.log @@ -0,0 +1,395 @@ +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.15s +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 141.66 seconds, peak gpu memory 13.4G +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31349.0 tgs: 57 data_time: 1.89s time: 547.07s eta: 3 days, 18:06:50 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32045.0 tgs: 61 data_time: 0.55s time: 523.25s eta: 3 days, 14:02:46 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.0GB text_tokens: 31775.0 tgs: 60 data_time: 0.90s time: 522.85s eta: 3 days, 13:50:02 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31442.0 tgs: 60 data_time: 0.83s time: 520.30s eta: 3 days, 13:16:16 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31310.0 tgs: 60 data_time: 0.74s time: 520.99s eta: 3 days, 13:14:24 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31970.0 tgs: 61 data_time: 0.84s time: 520.88s eta: 3 days, 13:04:40 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32046.0 tgs: 61 data_time: 0.83s time: 523.39s eta: 3 days, 13:20:31 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 32280.0 tgs: 61 data_time: 0.66s time: 520.67s eta: 3 days, 12:45:12 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32093.0 tgs: 61 data_time: 1.03s time: 520.19s eta: 3 days, 12:31:48 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.0GB text_tokens: 32007.0 tgs: 61 data_time: 0.54s time: 520.38s eta: 3 days, 12:24:59 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32301.0 tgs: 61 data_time: 0.81s time: 524.54s eta: 3 days, 12:56:47 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.221 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.3GB text_tokens: 30685.0 tgs: 58 data_time: 0.81s time: 520.66s eta: 3 days, 12:10:22 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 30776.0 tgs: 59 data_time: 0.86s time: 519.93s eta: 3 days, 11:54:37 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.318 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.0GB text_tokens: 31850.0 tgs: 61 data_time: 0.72s time: 521.30s eta: 3 days, 11:59:11 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 32016.0 tgs: 61 data_time: 0.75s time: 524.13s eta: 3 days, 12:17:52 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.0GB text_tokens: 31587.0 tgs: 60 data_time: 1.02s time: 520.55s eta: 3 days, 11:34:36 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.350 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 30793.0 tgs: 59 data_time: 0.76s time: 518.80s eta: 3 days, 11:09:05 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31008.0 tgs: 59 data_time: 0.92s time: 522.11s eta: 3 days, 11:32:17 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32462.0 tgs: 61 data_time: 0.88s time: 523.84s eta: 3 days, 11:40:07 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32193.0 tgs: 61 data_time: 0.68s time: 520.44s eta: 3 days, 10:58:51 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31420.0 tgs: 60 data_time: 0.80s time: 518.44s eta: 3 days, 10:31:06 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31579.0 tgs: 60 data_time: 0.94s time: 522.85s eta: 3 days, 11:04:28 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 31823.0 tgs: 60 data_time: 0.82s time: 523.52s eta: 3 days, 11:02:08 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 32.8GB text_tokens: 30152.0 tgs: 57 data_time: 0.81s time: 520.97s eta: 3 days, 10:29:11 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.8GB text_tokens: 31164.0 tgs: 60 data_time: 0.87s time: 519.11s eta: 3 days, 10:02:53 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.9GB text_tokens: 31740.0 tgs: 60 data_time: 0.82s time: 523.30s eta: 3 days, 10:33:54 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32120.0 tgs: 61 data_time: 0.91s time: 522.91s eta: 3 days, 10:21:30 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.5GB text_tokens: 30357.0 tgs: 58 data_time: 0.77s time: 521.32s eta: 3 days, 9:57:49 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 32083.0 tgs: 61 data_time: 0.87s time: 520.64s eta: 3 days, 9:42:44 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.354 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.9GB text_tokens: 31963.0 tgs: 61 data_time: 0.88s time: 522.39s eta: 3 days, 9:50:29 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.1GB text_tokens: 31960.0 tgs: 61 data_time: 0.80s time: 523.76s eta: 3 days, 9:54:37 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31654.0 tgs: 60 data_time: 0.80s time: 520.43s eta: 3 days, 9:14:43 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31502.0 tgs: 60 data_time: 0.64s time: 520.68s eta: 3 days, 9:08:23 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.9GB text_tokens: 31836.0 tgs: 61 data_time: 0.85s time: 521.26s eta: 3 days, 9:05:03 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.9GB text_tokens: 31221.0 tgs: 59 data_time: 0.72s time: 524.12s eta: 3 days, 9:23:02 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31303.0 tgs: 60 data_time: 0.95s time: 520.97s eta: 3 days, 8:44:58 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31855.0 tgs: 61 data_time: 1.01s time: 520.16s eta: 3 days, 8:28:49 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.8GB text_tokens: 30932.0 tgs: 59 data_time: 0.70s time: 522.57s eta: 3 days, 8:42:29 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 31919.0 tgs: 60 data_time: 0.61s time: 524.34s eta: 3 days, 8:50:08 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 31714.0 tgs: 61 data_time: 0.63s time: 519.89s eta: 3 days, 8:00:20 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.9GB text_tokens: 32069.0 tgs: 61 data_time: 0.73s time: 520.07s eta: 3 days, 7:53:18 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 31550.0 tgs: 60 data_time: 0.62s time: 522.01s eta: 3 days, 8:02:26 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.2GB text_tokens: 30417.0 tgs: 58 data_time: 0.53s time: 524.36s eta: 3 days, 8:15:22 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31874.0 tgs: 61 data_time: 0.85s time: 520.50s eta: 3 days, 7:31:13 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.219 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.7GB text_tokens: 31865.0 tgs: 61 data_time: 0.61s time: 518.57s eta: 3 days, 7:04:57 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:50:58][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 32131.0 tgs: 61 data_time: 0.95s time: 523.16s eta: 3 days, 7:38:12 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 32443.0 tgs: 61 data_time: 0.67s time: 523.63s eta: 3 days, 7:33:46 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.8GB text_tokens: 31045.0 tgs: 59 data_time: 0.76s time: 520.14s eta: 3 days, 6:53:16 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.9GB text_tokens: 31846.0 tgs: 61 data_time: 0.75s time: 520.06s eta: 3 days, 6:43:54 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31945.0 tgs: 60 data_time: 0.96s time: 523.77s eta: 3 days, 7:08:48 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.9GB text_tokens: 32011.0 tgs: 61 data_time: 1.08s time: 523.78s eta: 3 days, 7:00:11 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32547.0 tgs: 62 data_time: 0.97s time: 520.25s eta: 3 days, 6:19:33 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 32049.0 tgs: 61 data_time: 0.86s time: 518.79s eta: 3 days, 5:57:42 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 32269.0 tgs: 61 data_time: 0.85s time: 522.67s eta: 3 days, 6:24:03 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 30865.0 tgs: 59 data_time: 1.08s time: 522.70s eta: 3 days, 6:15:36 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.7GB text_tokens: 31114.0 tgs: 59 data_time: 1.09s time: 520.53s eta: 3 days, 5:47:27 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.8GB text_tokens: 32034.0 tgs: 61 data_time: 0.82s time: 519.67s eta: 3 days, 5:31:00 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.333 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.6GB text_tokens: 31575.0 tgs: 60 data_time: 1.25s time: 521.28s eta: 3 days, 5:36:47 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.336 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.4GB text_tokens: 30677.0 tgs: 58 data_time: 0.73s time: 523.60s eta: 3 days, 5:48:45 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 31925.0 tgs: 61 data_time: 0.93s time: 519.33s eta: 3 days, 5:02:02 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32154.0 tgs: 61 data_time: 0.61s time: 520.63s eta: 3 days, 5:04:53 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 32123.0 tgs: 61 data_time: 0.82s time: 521.46s eta: 3 days, 5:03:35 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.8GB text_tokens: 31246.0 tgs: 59 data_time: 0.91s time: 523.95s eta: 3 days, 5:16:58 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31119.0 tgs: 59 data_time: 0.87s time: 518.86s eta: 3 days, 4:23:16 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.221 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.9GB text_tokens: 31196.0 tgs: 60 data_time: 0.60s time: 519.23s eta: 3 days, 4:17:54 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31947.0 tgs: 61 data_time: 0.69s time: 520.72s eta: 3 days, 4:22:21 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.6GB text_tokens: 31639.0 tgs: 60 data_time: 0.84s time: 524.26s eta: 3 days, 4:44:45 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.0GB text_tokens: 31322.0 tgs: 60 data_time: 1.05s time: 520.28s eta: 3 days, 4:01:06 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.0GB text_tokens: 31116.0 tgs: 59 data_time: 0.81s time: 519.50s eta: 3 days, 3:45:39 +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 25][DP 6][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 32184.0 tgs: 61 data_time: 0.66s time: 522.33s eta: 3 days, 4:01:38 diff --git a/20250120235238/rank26.log b/20250120235238/rank26.log new file mode 100644 index 0000000000000000000000000000000000000000..173706422f9d9353a1e948637718f517baa315a4 --- /dev/null +++ b/20250120235238/rank26.log @@ -0,0 +1,395 @@ +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.28s +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 141.63 seconds, peak gpu memory 13.4G +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31349.0 tgs: 57 data_time: 1.75s time: 547.03s eta: 3 days, 18:06:31 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.208 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32045.0 tgs: 61 data_time: 0.54s time: 523.25s eta: 3 days, 14:02:44 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.0GB text_tokens: 31775.0 tgs: 60 data_time: 0.92s time: 522.85s eta: 3 days, 13:50:04 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31442.0 tgs: 60 data_time: 0.83s time: 520.30s eta: 3 days, 13:16:17 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31310.0 tgs: 60 data_time: 0.76s time: 520.99s eta: 3 days, 13:14:24 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31970.0 tgs: 61 data_time: 0.88s time: 520.89s eta: 3 days, 13:04:41 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32046.0 tgs: 61 data_time: 0.85s time: 523.39s eta: 3 days, 13:20:31 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 32280.0 tgs: 61 data_time: 0.67s time: 520.67s eta: 3 days, 12:45:15 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32093.0 tgs: 61 data_time: 1.04s time: 520.18s eta: 3 days, 12:31:46 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.0GB text_tokens: 32007.0 tgs: 61 data_time: 0.54s time: 520.38s eta: 3 days, 12:24:59 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32301.0 tgs: 61 data_time: 0.81s time: 524.54s eta: 3 days, 12:56:48 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.3GB text_tokens: 30685.0 tgs: 58 data_time: 0.83s time: 520.66s eta: 3 days, 12:10:23 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.199 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 30776.0 tgs: 59 data_time: 0.86s time: 519.93s eta: 3 days, 11:54:37 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.0GB text_tokens: 31850.0 tgs: 61 data_time: 0.74s time: 521.30s eta: 3 days, 11:59:12 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 32016.0 tgs: 61 data_time: 0.76s time: 524.13s eta: 3 days, 12:17:52 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.0GB text_tokens: 31587.0 tgs: 60 data_time: 1.03s time: 520.55s eta: 3 days, 11:34:36 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 30793.0 tgs: 59 data_time: 0.78s time: 518.80s eta: 3 days, 11:09:06 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31008.0 tgs: 59 data_time: 0.92s time: 522.11s eta: 3 days, 11:32:18 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32462.0 tgs: 61 data_time: 0.88s time: 523.84s eta: 3 days, 11:40:07 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32193.0 tgs: 61 data_time: 0.69s time: 520.44s eta: 3 days, 10:58:51 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31420.0 tgs: 60 data_time: 0.82s time: 518.44s eta: 3 days, 10:31:07 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31579.0 tgs: 60 data_time: 0.95s time: 522.85s eta: 3 days, 11:04:29 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 31823.0 tgs: 60 data_time: 0.83s time: 523.52s eta: 3 days, 11:02:08 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 32.8GB text_tokens: 30152.0 tgs: 57 data_time: 0.82s time: 520.97s eta: 3 days, 10:29:12 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.8GB text_tokens: 31164.0 tgs: 60 data_time: 0.93s time: 519.11s eta: 3 days, 10:02:53 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.9GB text_tokens: 31740.0 tgs: 60 data_time: 0.84s time: 523.30s eta: 3 days, 10:33:54 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32120.0 tgs: 61 data_time: 0.96s time: 522.91s eta: 3 days, 10:21:31 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.5GB text_tokens: 30357.0 tgs: 58 data_time: 0.78s time: 521.33s eta: 3 days, 9:57:50 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 32083.0 tgs: 61 data_time: 0.88s time: 520.65s eta: 3 days, 9:42:44 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.9GB text_tokens: 31963.0 tgs: 61 data_time: 0.89s time: 522.39s eta: 3 days, 9:50:30 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.1GB text_tokens: 31960.0 tgs: 61 data_time: 0.81s time: 523.76s eta: 3 days, 9:54:36 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31654.0 tgs: 60 data_time: 0.80s time: 520.43s eta: 3 days, 9:14:44 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31502.0 tgs: 60 data_time: 0.65s time: 520.69s eta: 3 days, 9:08:24 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.9GB text_tokens: 31836.0 tgs: 61 data_time: 0.92s time: 521.26s eta: 3 days, 9:05:03 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.9GB text_tokens: 31221.0 tgs: 59 data_time: 0.75s time: 524.12s eta: 3 days, 9:23:03 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31303.0 tgs: 60 data_time: 0.98s time: 520.97s eta: 3 days, 8:44:59 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31855.0 tgs: 61 data_time: 1.06s time: 520.16s eta: 3 days, 8:28:50 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.8GB text_tokens: 30932.0 tgs: 59 data_time: 0.71s time: 522.57s eta: 3 days, 8:42:30 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 31919.0 tgs: 60 data_time: 0.63s time: 524.34s eta: 3 days, 8:50:09 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 31714.0 tgs: 61 data_time: 0.62s time: 519.89s eta: 3 days, 8:00:21 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.9GB text_tokens: 32069.0 tgs: 61 data_time: 0.74s time: 520.07s eta: 3 days, 7:53:18 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 31550.0 tgs: 60 data_time: 0.62s time: 522.01s eta: 3 days, 8:02:27 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.2GB text_tokens: 30417.0 tgs: 58 data_time: 0.53s time: 524.36s eta: 3 days, 8:15:22 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31874.0 tgs: 61 data_time: 0.85s time: 520.50s eta: 3 days, 7:31:14 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.7GB text_tokens: 31865.0 tgs: 61 data_time: 0.61s time: 518.57s eta: 3 days, 7:04:56 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 32131.0 tgs: 61 data_time: 0.95s time: 523.16s eta: 3 days, 7:38:12 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 32443.0 tgs: 61 data_time: 0.67s time: 523.63s eta: 3 days, 7:33:47 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.8GB text_tokens: 31045.0 tgs: 59 data_time: 0.74s time: 520.14s eta: 3 days, 6:53:17 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.9GB text_tokens: 31846.0 tgs: 61 data_time: 0.76s time: 520.06s eta: 3 days, 6:43:54 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31945.0 tgs: 60 data_time: 0.97s time: 523.77s eta: 3 days, 7:08:49 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.9GB text_tokens: 32011.0 tgs: 61 data_time: 1.09s time: 523.78s eta: 3 days, 7:00:11 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32547.0 tgs: 62 data_time: 0.98s time: 520.25s eta: 3 days, 6:19:34 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 32049.0 tgs: 61 data_time: 0.87s time: 518.79s eta: 3 days, 5:57:43 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 32269.0 tgs: 61 data_time: 0.86s time: 522.67s eta: 3 days, 6:24:02 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 30865.0 tgs: 59 data_time: 1.09s time: 522.71s eta: 3 days, 6:15:38 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.7GB text_tokens: 31114.0 tgs: 59 data_time: 1.10s time: 520.53s eta: 3 days, 5:47:27 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.8GB text_tokens: 32034.0 tgs: 61 data_time: 0.84s time: 519.67s eta: 3 days, 5:31:01 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.6GB text_tokens: 31575.0 tgs: 60 data_time: 1.27s time: 521.28s eta: 3 days, 5:36:47 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.4GB text_tokens: 30677.0 tgs: 58 data_time: 0.77s time: 523.60s eta: 3 days, 5:48:45 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 31925.0 tgs: 61 data_time: 0.93s time: 519.33s eta: 3 days, 5:02:03 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32154.0 tgs: 61 data_time: 0.64s time: 520.63s eta: 3 days, 5:04:54 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 32123.0 tgs: 61 data_time: 0.84s time: 521.46s eta: 3 days, 5:03:36 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.8GB text_tokens: 31246.0 tgs: 59 data_time: 0.94s time: 523.95s eta: 3 days, 5:16:59 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31119.0 tgs: 59 data_time: 0.88s time: 518.86s eta: 3 days, 4:23:15 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.207 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.9GB text_tokens: 31196.0 tgs: 60 data_time: 0.61s time: 519.24s eta: 3 days, 4:17:55 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.350 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31947.0 tgs: 61 data_time: 0.73s time: 520.72s eta: 3 days, 4:22:22 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.6GB text_tokens: 31639.0 tgs: 60 data_time: 0.85s time: 524.26s eta: 3 days, 4:44:46 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.0GB text_tokens: 31322.0 tgs: 60 data_time: 1.06s time: 520.28s eta: 3 days, 4:01:06 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.0GB text_tokens: 31116.0 tgs: 59 data_time: 0.81s time: 519.50s eta: 3 days, 3:45:39 +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 26][DP 6][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 32184.0 tgs: 61 data_time: 0.66s time: 522.33s eta: 3 days, 4:01:39 diff --git a/20250120235238/rank3.log b/20250120235238/rank3.log new file mode 100644 index 0000000000000000000000000000000000000000..b212eb09141d18d31f7a031e4ac7bdee4f4b1baa --- /dev/null +++ b/20250120235238/rank3.log @@ -0,0 +1,395 @@ +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.12s +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:07:59][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 144.62 seconds, peak gpu memory 13.4G +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.7GB text_tokens: 31548.0 tgs: 57 data_time: 1.96s time: 546.73s eta: 3 days, 18:03:31 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32129.0 tgs: 61 data_time: 0.75s time: 523.23s eta: 3 days, 14:02:30 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.9GB text_tokens: 31947.0 tgs: 61 data_time: 0.80s time: 522.94s eta: 3 days, 13:50:56 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31471.0 tgs: 60 data_time: 0.69s time: 520.27s eta: 3 days, 13:16:02 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 32503.0 tgs: 62 data_time: 0.71s time: 520.97s eta: 3 days, 13:14:10 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.360 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.7GB text_tokens: 31133.0 tgs: 59 data_time: 0.84s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.339 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 31847.0 tgs: 60 data_time: 1.18s time: 523.37s eta: 3 days, 13:20:17 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 32030.0 tgs: 61 data_time: 1.01s time: 520.65s eta: 3 days, 12:44:58 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 31786.0 tgs: 61 data_time: 0.70s time: 520.16s eta: 3 days, 12:31:35 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 32078.0 tgs: 61 data_time: 1.04s time: 520.47s eta: 3 days, 12:25:56 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32240.0 tgs: 61 data_time: 0.89s time: 524.52s eta: 3 days, 12:56:32 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 31632.0 tgs: 60 data_time: 0.84s time: 520.63s eta: 3 days, 12:10:08 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.1GB text_tokens: 30691.0 tgs: 59 data_time: 0.74s time: 519.96s eta: 3 days, 11:54:56 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32119.0 tgs: 61 data_time: 0.85s time: 521.27s eta: 3 days, 11:58:57 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.198 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 31154.0 tgs: 59 data_time: 0.64s time: 524.11s eta: 3 days, 12:17:39 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.9GB text_tokens: 31896.0 tgs: 61 data_time: 0.61s time: 520.52s eta: 3 days, 11:34:22 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.193 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 31357.0 tgs: 60 data_time: 0.88s time: 518.89s eta: 3 days, 11:09:59 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 30112.0 tgs: 57 data_time: 0.68s time: 522.09s eta: 3 days, 11:32:04 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.1GB text_tokens: 31775.0 tgs: 60 data_time: 0.86s time: 523.81s eta: 3 days, 11:39:52 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32263.0 tgs: 61 data_time: 0.79s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.340 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31621.0 tgs: 60 data_time: 1.05s time: 518.42s eta: 3 days, 10:30:52 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.362 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 32165.0 tgs: 61 data_time: 1.04s time: 522.82s eta: 3 days, 11:04:15 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31731.0 tgs: 60 data_time: 0.57s time: 523.52s eta: 3 days, 11:02:10 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31953.0 tgs: 61 data_time: 0.71s time: 521.01s eta: 3 days, 10:29:36 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.0GB text_tokens: 32071.0 tgs: 61 data_time: 0.80s time: 519.08s eta: 3 days, 10:02:39 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 32232.0 tgs: 61 data_time: 0.61s time: 523.28s eta: 3 days, 10:33:41 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31872.0 tgs: 60 data_time: 0.68s time: 522.97s eta: 3 days, 10:22:03 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 31715.0 tgs: 60 data_time: 0.79s time: 521.30s eta: 3 days, 9:57:35 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31142.0 tgs: 59 data_time: 0.66s time: 520.62s eta: 3 days, 9:42:30 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.7GB text_tokens: 31246.0 tgs: 59 data_time: 1.08s time: 522.42s eta: 3 days, 9:50:45 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.8GB text_tokens: 31427.0 tgs: 60 data_time: 0.86s time: 523.77s eta: 3 days, 9:54:42 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 31009.0 tgs: 59 data_time: 0.81s time: 520.41s eta: 3 days, 9:14:30 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.6GB text_tokens: 31125.0 tgs: 59 data_time: 0.68s time: 520.66s eta: 3 days, 9:08:11 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 32138.0 tgs: 61 data_time: 0.88s time: 521.32s eta: 3 days, 9:05:39 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.8GB text_tokens: 31588.0 tgs: 60 data_time: 0.91s time: 524.09s eta: 3 days, 9:22:48 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.8GB text_tokens: 31748.0 tgs: 60 data_time: 0.85s time: 520.94s eta: 3 days, 8:44:45 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 32016.0 tgs: 61 data_time: 0.91s time: 520.21s eta: 3 days, 8:29:19 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.9GB text_tokens: 31540.0 tgs: 60 data_time: 0.73s time: 522.55s eta: 3 days, 8:42:18 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.8GB text_tokens: 30666.0 tgs: 58 data_time: 0.79s time: 524.32s eta: 3 days, 8:49:55 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31939.0 tgs: 61 data_time: 0.74s time: 519.87s eta: 3 days, 8:00:06 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.8GB text_tokens: 30286.0 tgs: 58 data_time: 0.78s time: 520.12s eta: 3 days, 7:53:45 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32181.0 tgs: 61 data_time: 0.84s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32188.0 tgs: 61 data_time: 0.71s time: 524.34s eta: 3 days, 8:15:09 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 32373.0 tgs: 62 data_time: 0.66s time: 520.56s eta: 3 days, 7:31:45 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.7GB text_tokens: 30468.0 tgs: 58 data_time: 0.61s time: 518.55s eta: 3 days, 7:04:44 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31398.0 tgs: 60 data_time: 0.72s time: 523.14s eta: 3 days, 7:37:59 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.9GB text_tokens: 30639.0 tgs: 58 data_time: 0.56s time: 523.61s eta: 3 days, 7:33:33 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 32054.0 tgs: 61 data_time: 0.90s time: 520.20s eta: 3 days, 6:53:50 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.8GB text_tokens: 31329.0 tgs: 60 data_time: 0.89s time: 520.04s eta: 3 days, 6:43:41 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.332 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 30789.0 tgs: 58 data_time: 0.72s time: 523.74s eta: 3 days, 7:08:35 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31630.0 tgs: 60 data_time: 0.76s time: 523.83s eta: 3 days, 7:00:37 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.315 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 31039.0 tgs: 59 data_time: 0.77s time: 520.22s eta: 3 days, 6:19:21 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 31956.0 tgs: 61 data_time: 1.00s time: 518.76s eta: 3 days, 5:57:29 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 32128.0 tgs: 61 data_time: 0.64s time: 522.67s eta: 3 days, 6:24:01 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 31665.0 tgs: 60 data_time: 0.64s time: 522.77s eta: 3 days, 6:16:11 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.8GB text_tokens: 31563.0 tgs: 60 data_time: 0.71s time: 520.51s eta: 3 days, 5:47:15 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32095.0 tgs: 61 data_time: 0.65s time: 519.64s eta: 3 days, 5:30:47 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32082.0 tgs: 61 data_time: 1.03s time: 521.34s eta: 3 days, 5:37:16 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.319 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.1GB text_tokens: 32250.0 tgs: 61 data_time: 0.66s time: 523.57s eta: 3 days, 5:48:31 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 31468.0 tgs: 60 data_time: 0.75s time: 519.29s eta: 3 days, 5:01:41 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 31281.0 tgs: 60 data_time: 0.80s time: 520.67s eta: 3 days, 5:05:14 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.1GB text_tokens: 31987.0 tgs: 61 data_time: 0.51s time: 521.47s eta: 3 days, 5:03:41 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.3GB text_tokens: 30618.0 tgs: 58 data_time: 0.58s time: 523.93s eta: 3 days, 5:16:46 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 32.4GB text_tokens: 30691.0 tgs: 59 data_time: 0.65s time: 518.84s eta: 3 days, 4:23:02 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.9GB text_tokens: 30766.0 tgs: 59 data_time: 0.61s time: 519.27s eta: 3 days, 4:18:14 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 30736.0 tgs: 59 data_time: 0.70s time: 520.70s eta: 3 days, 4:22:08 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.357 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 32451.0 tgs: 61 data_time: 0.70s time: 524.24s eta: 3 days, 4:44:33 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.0GB text_tokens: 30934.0 tgs: 59 data_time: 0.89s time: 520.34s eta: 3 days, 4:01:38 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.318 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32071.0 tgs: 61 data_time: 0.73s time: 519.52s eta: 3 days, 3:45:47 +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 3][DP 0][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 31757.0 tgs: 60 data_time: 0.90s time: 522.30s eta: 3 days, 4:01:26 diff --git a/20250120235238/rank31.log b/20250120235238/rank31.log new file mode 100644 index 0000000000000000000000000000000000000000..04226079c621a7b3e76ed3d19904615bbdd7f09a --- /dev/null +++ b/20250120235238/rank31.log @@ -0,0 +1,395 @@ +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.12s +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:08:02][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 141.70 seconds, peak gpu memory 13.4G +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 32271.0 tgs: 58 data_time: 1.84s time: 547.03s eta: 3 days, 18:06:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32224.0 tgs: 61 data_time: 0.93s time: 523.25s eta: 3 days, 14:02:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.8GB text_tokens: 31186.0 tgs: 59 data_time: 0.93s time: 522.85s eta: 3 days, 13:50:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 32092.0 tgs: 61 data_time: 0.81s time: 520.30s eta: 3 days, 13:16:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.8GB text_tokens: 31643.0 tgs: 60 data_time: 0.77s time: 520.99s eta: 3 days, 13:14:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.0GB text_tokens: 31263.0 tgs: 60 data_time: 0.72s time: 520.89s eta: 3 days, 13:04:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.342 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.7GB text_tokens: 30097.0 tgs: 57 data_time: 0.80s time: 523.39s eta: 3 days, 13:20:31 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.346 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32404.0 tgs: 62 data_time: 0.88s time: 520.67s eta: 3 days, 12:45:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 32.9GB text_tokens: 30582.0 tgs: 58 data_time: 0.82s time: 520.19s eta: 3 days, 12:31:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.1GB text_tokens: 32372.0 tgs: 62 data_time: 0.80s time: 520.38s eta: 3 days, 12:25:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32256.0 tgs: 61 data_time: 0.71s time: 524.54s eta: 3 days, 12:56:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 31025.0 tgs: 59 data_time: 0.63s time: 520.66s eta: 3 days, 12:10:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 31804.0 tgs: 61 data_time: 0.87s time: 519.93s eta: 3 days, 11:54:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32389.0 tgs: 62 data_time: 0.79s time: 521.30s eta: 3 days, 11:59:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 30946.0 tgs: 59 data_time: 0.87s time: 524.13s eta: 3 days, 12:17:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31432.0 tgs: 60 data_time: 0.90s time: 520.55s eta: 3 days, 11:34:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32194.0 tgs: 62 data_time: 0.73s time: 518.80s eta: 3 days, 11:09:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 31342.0 tgs: 60 data_time: 1.05s time: 522.11s eta: 3 days, 11:32:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32054.0 tgs: 61 data_time: 0.67s time: 523.84s eta: 3 days, 11:40:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32360.0 tgs: 62 data_time: 0.72s time: 520.44s eta: 3 days, 10:58:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.1GB text_tokens: 31497.0 tgs: 60 data_time: 0.61s time: 518.44s eta: 3 days, 10:31:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 32299.0 tgs: 61 data_time: 0.81s time: 522.85s eta: 3 days, 11:04:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 32054.0 tgs: 61 data_time: 0.64s time: 523.52s eta: 3 days, 11:02:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31131.0 tgs: 59 data_time: 0.62s time: 520.97s eta: 3 days, 10:29:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.0GB text_tokens: 32103.0 tgs: 61 data_time: 0.75s time: 519.11s eta: 3 days, 10:02:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 32152.0 tgs: 61 data_time: 0.66s time: 523.30s eta: 3 days, 10:33:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.9GB text_tokens: 31336.0 tgs: 59 data_time: 0.96s time: 522.91s eta: 3 days, 10:21:31 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 31412.0 tgs: 60 data_time: 0.71s time: 521.33s eta: 3 days, 9:57:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32135.0 tgs: 61 data_time: 0.69s time: 520.64s eta: 3 days, 9:42:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.9GB text_tokens: 31733.0 tgs: 60 data_time: 0.85s time: 522.39s eta: 3 days, 9:50:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.1GB text_tokens: 31765.0 tgs: 60 data_time: 0.85s time: 523.76s eta: 3 days, 9:54:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 31949.0 tgs: 61 data_time: 0.91s time: 520.44s eta: 3 days, 9:14:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.9GB text_tokens: 31335.0 tgs: 60 data_time: 0.60s time: 520.68s eta: 3 days, 9:08:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31518.0 tgs: 60 data_time: 0.95s time: 521.26s eta: 3 days, 9:05:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.208 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31987.0 tgs: 61 data_time: 0.92s time: 524.12s eta: 3 days, 9:23:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 32039.0 tgs: 61 data_time: 1.27s time: 520.97s eta: 3 days, 8:44:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 32103.0 tgs: 61 data_time: 0.77s time: 520.16s eta: 3 days, 8:28:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32352.0 tgs: 61 data_time: 0.76s time: 522.57s eta: 3 days, 8:42:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.0GB text_tokens: 31385.0 tgs: 59 data_time: 1.01s time: 524.34s eta: 3 days, 8:50:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 31598.0 tgs: 60 data_time: 0.71s time: 519.90s eta: 3 days, 8:00:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.221 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 31223.0 tgs: 60 data_time: 0.76s time: 520.07s eta: 3 days, 7:53:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32577.0 tgs: 62 data_time: 1.09s time: 522.01s eta: 3 days, 8:02:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.9GB text_tokens: 31503.0 tgs: 60 data_time: 0.75s time: 524.36s eta: 3 days, 8:15:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.330 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.9GB text_tokens: 31442.0 tgs: 60 data_time: 0.72s time: 520.50s eta: 3 days, 7:31:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31080.0 tgs: 59 data_time: 0.84s time: 518.57s eta: 3 days, 7:04:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.219 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31762.0 tgs: 60 data_time: 0.72s time: 523.16s eta: 3 days, 7:38:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.8GB text_tokens: 31859.0 tgs: 60 data_time: 0.84s time: 523.64s eta: 3 days, 7:33:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 32052.0 tgs: 61 data_time: 0.60s time: 520.14s eta: 3 days, 6:53:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.7GB text_tokens: 31795.0 tgs: 61 data_time: 0.77s time: 520.06s eta: 3 days, 6:43:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31424.0 tgs: 59 data_time: 0.75s time: 523.77s eta: 3 days, 7:08:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 32138.0 tgs: 61 data_time: 0.70s time: 523.78s eta: 3 days, 7:00:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32221.0 tgs: 61 data_time: 0.92s time: 520.25s eta: 3 days, 6:19:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.0GB text_tokens: 31787.0 tgs: 61 data_time: 0.86s time: 518.79s eta: 3 days, 5:57:42 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 31727.0 tgs: 60 data_time: 0.65s time: 522.67s eta: 3 days, 6:24:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 31831.0 tgs: 60 data_time: 0.89s time: 522.70s eta: 3 days, 6:15:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 32054.0 tgs: 61 data_time: 0.94s time: 520.53s eta: 3 days, 5:47:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.213 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.7GB text_tokens: 31492.0 tgs: 60 data_time: 0.82s time: 519.67s eta: 3 days, 5:31:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.7GB text_tokens: 31059.0 tgs: 59 data_time: 0.68s time: 521.28s eta: 3 days, 5:36:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.7GB text_tokens: 30232.0 tgs: 57 data_time: 0.77s time: 523.60s eta: 3 days, 5:48:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.347 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.8GB text_tokens: 31767.0 tgs: 61 data_time: 0.80s time: 519.33s eta: 3 days, 5:02:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 32.9GB text_tokens: 32058.0 tgs: 61 data_time: 0.57s time: 520.63s eta: 3 days, 5:04:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.1GB text_tokens: 32164.0 tgs: 61 data_time: 0.65s time: 521.46s eta: 3 days, 5:03:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.8GB text_tokens: 31753.0 tgs: 60 data_time: 0.42s time: 523.95s eta: 3 days, 5:17:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 32.8GB text_tokens: 31116.0 tgs: 59 data_time: 0.66s time: 518.86s eta: 3 days, 4:23:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31233.0 tgs: 60 data_time: 0.67s time: 519.23s eta: 3 days, 4:17:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 32074.0 tgs: 61 data_time: 0.71s time: 520.72s eta: 3 days, 4:22:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 31823.0 tgs: 60 data_time: 0.79s time: 524.26s eta: 3 days, 4:44:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32008.0 tgs: 61 data_time: 0.71s time: 520.28s eta: 3 days, 4:01:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.0GB text_tokens: 32207.0 tgs: 61 data_time: 0.73s time: 519.50s eta: 3 days, 3:45:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.377 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 32.7GB text_tokens: 31813.0 tgs: 60 data_time: 0.65s time: 522.33s eta: 3 days, 4:01:39 diff --git a/20250120235238/rank33.log b/20250120235238/rank33.log new file mode 100644 index 0000000000000000000000000000000000000000..c75fe416b38b69ef51d94e3892af145869b223cd --- /dev/null +++ b/20250120235238/rank33.log @@ -0,0 +1,395 @@ +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.15s +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.28 seconds, peak gpu memory 13.4G +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31323.0 tgs: 57 data_time: 1.86s time: 547.88s eta: 3 days, 18:14:52 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31388.0 tgs: 59 data_time: 0.80s time: 523.24s eta: 3 days, 14:02:36 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.0GB text_tokens: 31981.0 tgs: 61 data_time: 1.00s time: 522.91s eta: 3 days, 13:50:39 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.9GB text_tokens: 31503.0 tgs: 60 data_time: 1.25s time: 520.28s eta: 3 days, 13:16:07 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.209 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 31606.0 tgs: 60 data_time: 0.64s time: 520.98s eta: 3 days, 13:14:15 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31195.0 tgs: 59 data_time: 0.74s time: 520.89s eta: 3 days, 13:04:44 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 31593.0 tgs: 60 data_time: 0.74s time: 523.38s eta: 3 days, 13:20:23 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31176.0 tgs: 59 data_time: 0.66s time: 520.66s eta: 3 days, 12:45:04 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32152.0 tgs: 61 data_time: 0.66s time: 520.17s eta: 3 days, 12:31:40 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.226 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.0GB text_tokens: 29940.0 tgs: 57 data_time: 0.65s time: 520.43s eta: 3 days, 12:25:30 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 31618.0 tgs: 60 data_time: 0.92s time: 524.53s eta: 3 days, 12:56:39 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.333 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 31523.0 tgs: 60 data_time: 0.64s time: 520.64s eta: 3 days, 12:10:13 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31105.0 tgs: 59 data_time: 0.83s time: 519.97s eta: 3 days, 11:54:59 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 31143.0 tgs: 59 data_time: 0.51s time: 521.28s eta: 3 days, 11:59:02 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.8GB text_tokens: 31851.0 tgs: 60 data_time: 0.57s time: 524.12s eta: 3 days, 12:17:44 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31775.0 tgs: 61 data_time: 0.87s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 31815.0 tgs: 61 data_time: 0.90s time: 518.84s eta: 3 days, 11:09:33 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31366.0 tgs: 60 data_time: 0.78s time: 522.10s eta: 3 days, 11:32:09 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31364.0 tgs: 59 data_time: 0.83s time: 523.82s eta: 3 days, 11:39:59 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.7GB text_tokens: 32004.0 tgs: 61 data_time: 0.93s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31624.0 tgs: 60 data_time: 0.84s time: 518.43s eta: 3 days, 10:30:58 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.358 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31087.0 tgs: 59 data_time: 0.68s time: 522.83s eta: 3 days, 11:04:20 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.206 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31662.0 tgs: 60 data_time: 0.57s time: 523.54s eta: 3 days, 11:02:20 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 32.5GB text_tokens: 31237.0 tgs: 59 data_time: 0.80s time: 520.96s eta: 3 days, 10:29:06 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31187.0 tgs: 60 data_time: 0.75s time: 519.09s eta: 3 days, 10:02:45 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31088.0 tgs: 59 data_time: 0.66s time: 523.29s eta: 3 days, 10:33:46 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31662.0 tgs: 60 data_time: 0.76s time: 522.95s eta: 3 days, 10:21:54 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 32039.0 tgs: 61 data_time: 0.95s time: 521.31s eta: 3 days, 9:57:41 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32252.0 tgs: 61 data_time: 0.53s time: 520.63s eta: 3 days, 9:42:36 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 31895.0 tgs: 61 data_time: 0.95s time: 522.43s eta: 3 days, 9:50:50 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.226 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.7GB text_tokens: 31516.0 tgs: 60 data_time: 0.59s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 29894.0 tgs: 57 data_time: 0.80s time: 520.42s eta: 3 days, 9:14:35 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31299.0 tgs: 60 data_time: 0.85s time: 520.67s eta: 3 days, 9:08:16 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.9GB text_tokens: 31641.0 tgs: 60 data_time: 0.82s time: 521.31s eta: 3 days, 9:05:31 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.7GB text_tokens: 31053.0 tgs: 59 data_time: 0.54s time: 524.10s eta: 3 days, 9:22:54 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31452.0 tgs: 60 data_time: 0.83s time: 520.95s eta: 3 days, 8:44:51 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.7GB text_tokens: 31537.0 tgs: 60 data_time: 0.81s time: 520.17s eta: 3 days, 8:28:52 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.375 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32333.0 tgs: 61 data_time: 0.85s time: 522.56s eta: 3 days, 8:42:21 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.215 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32264.0 tgs: 61 data_time: 0.75s time: 524.33s eta: 3 days, 8:50:00 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 31571.0 tgs: 60 data_time: 0.62s time: 519.88s eta: 3 days, 8:00:12 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.0GB text_tokens: 32023.0 tgs: 61 data_time: 0.52s time: 520.13s eta: 3 days, 7:53:50 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32305.0 tgs: 61 data_time: 0.78s time: 521.99s eta: 3 days, 8:02:19 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32197.0 tgs: 61 data_time: 1.01s time: 524.35s eta: 3 days, 8:15:14 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.8GB text_tokens: 31127.0 tgs: 59 data_time: 0.59s time: 520.52s eta: 3 days, 7:31:28 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.8GB text_tokens: 31389.0 tgs: 60 data_time: 0.82s time: 518.56s eta: 3 days, 7:04:48 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31307.0 tgs: 59 data_time: 0.76s time: 523.15s eta: 3 days, 7:38:05 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31744.0 tgs: 60 data_time: 1.19s time: 523.62s eta: 3 days, 7:33:38 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.213 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31146.0 tgs: 59 data_time: 0.96s time: 520.18s eta: 3 days, 6:53:40 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 30841.0 tgs: 59 data_time: 0.81s time: 520.05s eta: 3 days, 6:43:46 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 30859.0 tgs: 58 data_time: 1.13s time: 523.75s eta: 3 days, 7:08:41 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31436.0 tgs: 60 data_time: 0.85s time: 523.80s eta: 3 days, 7:00:25 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32042.0 tgs: 61 data_time: 0.92s time: 520.23s eta: 3 days, 6:19:26 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.9GB text_tokens: 31299.0 tgs: 60 data_time: 0.64s time: 518.77s eta: 3 days, 5:57:34 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.9GB text_tokens: 31820.0 tgs: 60 data_time: 0.54s time: 522.68s eta: 3 days, 6:24:06 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31916.0 tgs: 61 data_time: 0.65s time: 522.73s eta: 3 days, 6:15:52 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 30595.0 tgs: 58 data_time: 0.73s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.329 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32038.0 tgs: 61 data_time: 0.70s time: 519.65s eta: 3 days, 5:30:53 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 31912.0 tgs: 61 data_time: 0.76s time: 521.32s eta: 3 days, 5:37:06 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.0GB text_tokens: 31973.0 tgs: 61 data_time: 0.91s time: 523.58s eta: 3 days, 5:48:37 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 32142.0 tgs: 61 data_time: 0.67s time: 519.32s eta: 3 days, 5:01:54 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 32223.0 tgs: 61 data_time: 0.78s time: 520.66s eta: 3 days, 5:05:12 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 32203.0 tgs: 61 data_time: 0.61s time: 521.45s eta: 3 days, 5:03:29 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 30992.0 tgs: 59 data_time: 0.69s time: 523.94s eta: 3 days, 5:16:51 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 32317.0 tgs: 62 data_time: 0.99s time: 518.84s eta: 3 days, 4:23:07 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 31408.0 tgs: 60 data_time: 0.68s time: 519.27s eta: 3 days, 4:18:12 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.206 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 31491.0 tgs: 60 data_time: 0.75s time: 520.71s eta: 3 days, 4:22:14 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 30711.0 tgs: 58 data_time: 0.83s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 30847.0 tgs: 59 data_time: 0.75s time: 520.32s eta: 3 days, 4:01:28 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32021.0 tgs: 61 data_time: 0.90s time: 519.49s eta: 3 days, 3:45:32 +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 33][DP 8][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 32505.0 tgs: 62 data_time: 0.96s time: 522.31s eta: 3 days, 4:01:31 diff --git a/20250120235238/rank34.log b/20250120235238/rank34.log new file mode 100644 index 0000000000000000000000000000000000000000..8a21d6404b0c10ffb848b69d2201c55a116d8323 --- /dev/null +++ b/20250120235238/rank34.log @@ -0,0 +1,395 @@ +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.12s +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.22 seconds, peak gpu memory 13.4G +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31323.0 tgs: 57 data_time: 1.90s time: 547.83s eta: 3 days, 18:14:23 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31388.0 tgs: 59 data_time: 0.88s time: 523.24s eta: 3 days, 14:02:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.0GB text_tokens: 31981.0 tgs: 61 data_time: 1.13s time: 522.91s eta: 3 days, 13:50:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.9GB text_tokens: 31503.0 tgs: 60 data_time: 1.38s time: 520.28s eta: 3 days, 13:16:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 31606.0 tgs: 60 data_time: 0.68s time: 520.98s eta: 3 days, 13:14:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31195.0 tgs: 59 data_time: 0.80s time: 520.89s eta: 3 days, 13:04:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 31593.0 tgs: 60 data_time: 0.78s time: 523.38s eta: 3 days, 13:20:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31176.0 tgs: 59 data_time: 0.68s time: 520.66s eta: 3 days, 12:45:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32152.0 tgs: 61 data_time: 0.70s time: 520.17s eta: 3 days, 12:31:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.194 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.0GB text_tokens: 29940.0 tgs: 57 data_time: 0.69s time: 520.43s eta: 3 days, 12:25:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 31618.0 tgs: 60 data_time: 0.97s time: 524.52s eta: 3 days, 12:56:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 31523.0 tgs: 60 data_time: 0.68s time: 520.64s eta: 3 days, 12:10:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31105.0 tgs: 59 data_time: 0.90s time: 519.96s eta: 3 days, 11:54:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 31143.0 tgs: 59 data_time: 0.62s time: 521.28s eta: 3 days, 11:59:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.8GB text_tokens: 31851.0 tgs: 60 data_time: 0.60s time: 524.12s eta: 3 days, 12:17:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31775.0 tgs: 61 data_time: 0.89s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.320 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 31815.0 tgs: 61 data_time: 0.93s time: 518.84s eta: 3 days, 11:09:32 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31366.0 tgs: 60 data_time: 0.85s time: 522.10s eta: 3 days, 11:32:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.190 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31364.0 tgs: 59 data_time: 0.86s time: 523.82s eta: 3 days, 11:39:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.7GB text_tokens: 32004.0 tgs: 61 data_time: 0.99s time: 520.47s eta: 3 days, 10:59:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.372 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31624.0 tgs: 61 data_time: 0.87s time: 518.43s eta: 3 days, 10:30:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31087.0 tgs: 59 data_time: 0.73s time: 522.83s eta: 3 days, 11:04:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31662.0 tgs: 60 data_time: 0.60s time: 523.53s eta: 3 days, 11:02:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 32.5GB text_tokens: 31237.0 tgs: 59 data_time: 0.85s time: 520.96s eta: 3 days, 10:29:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31187.0 tgs: 60 data_time: 0.79s time: 519.09s eta: 3 days, 10:02:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31088.0 tgs: 59 data_time: 0.69s time: 523.29s eta: 3 days, 10:33:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31662.0 tgs: 60 data_time: 0.81s time: 522.95s eta: 3 days, 10:21:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.226 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 32039.0 tgs: 61 data_time: 0.97s time: 521.31s eta: 3 days, 9:57:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32252.0 tgs: 61 data_time: 0.58s time: 520.63s eta: 3 days, 9:42:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 31895.0 tgs: 61 data_time: 1.04s time: 522.42s eta: 3 days, 9:50:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.7GB text_tokens: 31516.0 tgs: 60 data_time: 0.62s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 29894.0 tgs: 57 data_time: 0.82s time: 520.42s eta: 3 days, 9:14:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31299.0 tgs: 60 data_time: 0.91s time: 520.67s eta: 3 days, 9:08:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.9GB text_tokens: 31641.0 tgs: 60 data_time: 0.85s time: 521.30s eta: 3 days, 9:05:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.199 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.7GB text_tokens: 31053.0 tgs: 59 data_time: 0.56s time: 524.10s eta: 3 days, 9:22:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31452.0 tgs: 60 data_time: 0.84s time: 520.95s eta: 3 days, 8:44:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.7GB text_tokens: 31537.0 tgs: 60 data_time: 0.83s time: 520.16s eta: 3 days, 8:28:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32333.0 tgs: 61 data_time: 0.88s time: 522.56s eta: 3 days, 8:42:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32264.0 tgs: 61 data_time: 0.79s time: 524.32s eta: 3 days, 8:49:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 31571.0 tgs: 60 data_time: 0.66s time: 519.88s eta: 3 days, 8:00:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.0GB text_tokens: 32023.0 tgs: 61 data_time: 0.55s time: 520.13s eta: 3 days, 7:53:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32305.0 tgs: 61 data_time: 0.79s time: 521.99s eta: 3 days, 8:02:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32197.0 tgs: 61 data_time: 1.05s time: 524.35s eta: 3 days, 8:15:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.8GB text_tokens: 31127.0 tgs: 59 data_time: 0.62s time: 520.52s eta: 3 days, 7:31:27 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.8GB text_tokens: 31389.0 tgs: 60 data_time: 0.86s time: 518.56s eta: 3 days, 7:04:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31307.0 tgs: 59 data_time: 0.77s time: 523.15s eta: 3 days, 7:38:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31744.0 tgs: 60 data_time: 1.21s time: 523.62s eta: 3 days, 7:33:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31146.0 tgs: 59 data_time: 1.00s time: 520.18s eta: 3 days, 6:53:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.308 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 30841.0 tgs: 59 data_time: 0.84s time: 520.04s eta: 3 days, 6:43:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 30859.0 tgs: 58 data_time: 1.20s time: 523.75s eta: 3 days, 7:08:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31436.0 tgs: 60 data_time: 0.87s time: 523.80s eta: 3 days, 7:00:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32042.0 tgs: 61 data_time: 0.93s time: 520.23s eta: 3 days, 6:19:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.9GB text_tokens: 31299.0 tgs: 60 data_time: 0.65s time: 518.77s eta: 3 days, 5:57:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.341 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.9GB text_tokens: 31820.0 tgs: 60 data_time: 0.56s time: 522.68s eta: 3 days, 6:24:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31916.0 tgs: 61 data_time: 0.68s time: 522.73s eta: 3 days, 6:15:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 30595.0 tgs: 58 data_time: 0.79s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32038.0 tgs: 61 data_time: 0.72s time: 519.65s eta: 3 days, 5:30:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.336 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 31912.0 tgs: 61 data_time: 0.81s time: 521.32s eta: 3 days, 5:37:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.336 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.0GB text_tokens: 31973.0 tgs: 61 data_time: 0.94s time: 523.58s eta: 3 days, 5:48:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 32142.0 tgs: 61 data_time: 0.69s time: 519.31s eta: 3 days, 5:01:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 32223.0 tgs: 61 data_time: 0.81s time: 520.66s eta: 3 days, 5:05:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 32203.0 tgs: 61 data_time: 0.71s time: 521.45s eta: 3 days, 5:03:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 30992.0 tgs: 59 data_time: 0.71s time: 523.94s eta: 3 days, 5:16:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.370 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 32317.0 tgs: 62 data_time: 1.01s time: 518.84s eta: 3 days, 4:23:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 31408.0 tgs: 60 data_time: 0.72s time: 519.27s eta: 3 days, 4:18:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 31491.0 tgs: 60 data_time: 0.77s time: 520.71s eta: 3 days, 4:22:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 30711.0 tgs: 58 data_time: 0.84s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 30847.0 tgs: 59 data_time: 0.77s time: 520.32s eta: 3 days, 4:01:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32021.0 tgs: 61 data_time: 0.94s time: 519.48s eta: 3 days, 3:45:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.372 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 32505.0 tgs: 62 data_time: 0.98s time: 522.32s eta: 3 days, 4:01:33 diff --git a/20250120235238/rank35.log b/20250120235238/rank35.log new file mode 100644 index 0000000000000000000000000000000000000000..56da2295f50b2dcd81e74be708c1416f596ae207 --- /dev/null +++ b/20250120235238/rank35.log @@ -0,0 +1,395 @@ +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.11s +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.28 seconds, peak gpu memory 13.4G +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31323.0 tgs: 57 data_time: 1.87s time: 547.79s eta: 3 days, 18:14:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31388.0 tgs: 59 data_time: 0.88s time: 523.24s eta: 3 days, 14:02:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.0GB text_tokens: 31981.0 tgs: 61 data_time: 1.02s time: 522.91s eta: 3 days, 13:50:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.9GB text_tokens: 31503.0 tgs: 60 data_time: 1.38s time: 520.29s eta: 3 days, 13:16:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 31606.0 tgs: 60 data_time: 0.67s time: 520.98s eta: 3 days, 13:14:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31195.0 tgs: 59 data_time: 0.78s time: 520.89s eta: 3 days, 13:04:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 31593.0 tgs: 60 data_time: 0.75s time: 523.38s eta: 3 days, 13:20:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31176.0 tgs: 59 data_time: 0.70s time: 520.66s eta: 3 days, 12:45:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32152.0 tgs: 61 data_time: 0.71s time: 520.17s eta: 3 days, 12:31:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.0GB text_tokens: 29940.0 tgs: 57 data_time: 0.71s time: 520.43s eta: 3 days, 12:25:30 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 31618.0 tgs: 60 data_time: 0.97s time: 524.53s eta: 3 days, 12:56:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 31523.0 tgs: 60 data_time: 0.67s time: 520.64s eta: 3 days, 12:10:14 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31105.0 tgs: 59 data_time: 0.88s time: 519.96s eta: 3 days, 11:54:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 31143.0 tgs: 59 data_time: 0.59s time: 521.28s eta: 3 days, 11:59:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.8GB text_tokens: 31851.0 tgs: 60 data_time: 0.60s time: 524.12s eta: 3 days, 12:17:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31775.0 tgs: 61 data_time: 0.89s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 31815.0 tgs: 61 data_time: 0.93s time: 518.84s eta: 3 days, 11:09:32 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31366.0 tgs: 60 data_time: 0.87s time: 522.10s eta: 3 days, 11:32:09 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31364.0 tgs: 59 data_time: 0.87s time: 523.82s eta: 3 days, 11:39:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.7GB text_tokens: 32004.0 tgs: 61 data_time: 0.98s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31624.0 tgs: 61 data_time: 0.86s time: 518.43s eta: 3 days, 10:30:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31087.0 tgs: 59 data_time: 0.73s time: 522.83s eta: 3 days, 11:04:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31662.0 tgs: 60 data_time: 0.61s time: 523.54s eta: 3 days, 11:02:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 32.5GB text_tokens: 31237.0 tgs: 59 data_time: 0.87s time: 520.96s eta: 3 days, 10:29:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31187.0 tgs: 60 data_time: 0.77s time: 519.09s eta: 3 days, 10:02:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.219 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31088.0 tgs: 59 data_time: 0.69s time: 523.29s eta: 3 days, 10:33:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31662.0 tgs: 60 data_time: 0.80s time: 522.95s eta: 3 days, 10:21:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 32039.0 tgs: 61 data_time: 1.01s time: 521.31s eta: 3 days, 9:57:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32252.0 tgs: 61 data_time: 0.57s time: 520.63s eta: 3 days, 9:42:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 31895.0 tgs: 61 data_time: 1.07s time: 522.43s eta: 3 days, 9:50:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.7GB text_tokens: 31516.0 tgs: 60 data_time: 0.62s time: 523.75s eta: 3 days, 9:54:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 29894.0 tgs: 57 data_time: 0.83s time: 520.42s eta: 3 days, 9:14:35 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31299.0 tgs: 60 data_time: 0.89s time: 520.67s eta: 3 days, 9:08:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.9GB text_tokens: 31641.0 tgs: 60 data_time: 0.85s time: 521.31s eta: 3 days, 9:05:31 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.7GB text_tokens: 31053.0 tgs: 59 data_time: 0.56s time: 524.10s eta: 3 days, 9:22:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31452.0 tgs: 60 data_time: 0.89s time: 520.95s eta: 3 days, 8:44:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.7GB text_tokens: 31537.0 tgs: 60 data_time: 0.83s time: 520.17s eta: 3 days, 8:28:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32333.0 tgs: 61 data_time: 0.88s time: 522.56s eta: 3 days, 8:42:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32264.0 tgs: 61 data_time: 0.79s time: 524.32s eta: 3 days, 8:50:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.213 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 31571.0 tgs: 60 data_time: 0.66s time: 519.88s eta: 3 days, 8:00:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.205 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.0GB text_tokens: 32023.0 tgs: 61 data_time: 0.55s time: 520.13s eta: 3 days, 7:53:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.366 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32305.0 tgs: 61 data_time: 0.79s time: 521.99s eta: 3 days, 8:02:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32197.0 tgs: 61 data_time: 1.07s time: 524.35s eta: 3 days, 8:15:14 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.8GB text_tokens: 31127.0 tgs: 59 data_time: 0.62s time: 520.52s eta: 3 days, 7:31:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.8GB text_tokens: 31389.0 tgs: 60 data_time: 0.86s time: 518.56s eta: 3 days, 7:04:48 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.213 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31307.0 tgs: 59 data_time: 0.79s time: 523.15s eta: 3 days, 7:38:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31744.0 tgs: 60 data_time: 1.24s time: 523.62s eta: 3 days, 7:33:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31146.0 tgs: 59 data_time: 0.99s time: 520.18s eta: 3 days, 6:53:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 30841.0 tgs: 59 data_time: 0.85s time: 520.05s eta: 3 days, 6:43:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.320 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 30859.0 tgs: 58 data_time: 1.21s time: 523.75s eta: 3 days, 7:08:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31436.0 tgs: 60 data_time: 0.87s time: 523.80s eta: 3 days, 7:00:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32042.0 tgs: 61 data_time: 0.94s time: 520.23s eta: 3 days, 6:19:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.9GB text_tokens: 31299.0 tgs: 60 data_time: 0.66s time: 518.77s eta: 3 days, 5:57:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.9GB text_tokens: 31820.0 tgs: 60 data_time: 0.56s time: 522.68s eta: 3 days, 6:24:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31916.0 tgs: 61 data_time: 0.69s time: 522.73s eta: 3 days, 6:15:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 30595.0 tgs: 58 data_time: 0.79s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.331 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32038.0 tgs: 61 data_time: 0.72s time: 519.65s eta: 3 days, 5:30:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 31912.0 tgs: 61 data_time: 0.79s time: 521.32s eta: 3 days, 5:37:07 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.0GB text_tokens: 31973.0 tgs: 61 data_time: 0.94s time: 523.58s eta: 3 days, 5:48:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 32142.0 tgs: 61 data_time: 0.73s time: 519.32s eta: 3 days, 5:01:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 32223.0 tgs: 61 data_time: 0.82s time: 520.66s eta: 3 days, 5:05:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 32203.0 tgs: 61 data_time: 0.68s time: 521.45s eta: 3 days, 5:03:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 30992.0 tgs: 59 data_time: 0.72s time: 523.94s eta: 3 days, 5:16:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.363 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 32317.0 tgs: 62 data_time: 1.01s time: 518.85s eta: 3 days, 4:23:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 31408.0 tgs: 60 data_time: 0.70s time: 519.27s eta: 3 days, 4:18:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 31491.0 tgs: 60 data_time: 0.77s time: 520.71s eta: 3 days, 4:22:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 30711.0 tgs: 58 data_time: 0.86s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 30847.0 tgs: 59 data_time: 0.77s time: 520.32s eta: 3 days, 4:01:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32021.0 tgs: 61 data_time: 0.92s time: 519.49s eta: 3 days, 3:45:31 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 32505.0 tgs: 62 data_time: 0.99s time: 522.31s eta: 3 days, 4:01:32 diff --git a/20250120235238/rank38.log b/20250120235238/rank38.log new file mode 100644 index 0000000000000000000000000000000000000000..74157efcf5f989c587408fba2f0fda3cacfcde35 --- /dev/null +++ b/20250120235238/rank38.log @@ -0,0 +1,395 @@ +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.11s +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.27 seconds, peak gpu memory 13.4G +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.0GB text_tokens: 32281.0 tgs: 58 data_time: 2.29s time: 547.46s eta: 3 days, 18:10:43 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32206.0 tgs: 61 data_time: 0.83s time: 523.23s eta: 3 days, 14:02:32 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 32213.0 tgs: 61 data_time: 1.00s time: 522.91s eta: 3 days, 13:50:37 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.1GB text_tokens: 29331.0 tgs: 56 data_time: 0.94s time: 520.28s eta: 3 days, 13:16:06 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 31870.0 tgs: 61 data_time: 0.89s time: 520.98s eta: 3 days, 13:14:15 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.4GB text_tokens: 29472.0 tgs: 56 data_time: 0.72s time: 520.89s eta: 3 days, 13:04:43 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.0GB text_tokens: 32120.0 tgs: 61 data_time: 0.74s time: 523.38s eta: 3 days, 13:20:25 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 30854.0 tgs: 59 data_time: 0.66s time: 520.65s eta: 3 days, 12:45:01 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 32.8GB text_tokens: 31900.0 tgs: 61 data_time: 0.73s time: 520.17s eta: 3 days, 12:31:41 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.0GB text_tokens: 32109.0 tgs: 61 data_time: 0.81s time: 520.43s eta: 3 days, 12:25:29 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 31161.0 tgs: 59 data_time: 0.64s time: 524.52s eta: 3 days, 12:56:37 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.7GB text_tokens: 31536.0 tgs: 60 data_time: 0.79s time: 520.64s eta: 3 days, 12:10:13 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.9GB text_tokens: 32004.0 tgs: 61 data_time: 0.78s time: 519.96s eta: 3 days, 11:54:58 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32359.0 tgs: 62 data_time: 0.81s time: 521.28s eta: 3 days, 11:59:03 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.308 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 32408.0 tgs: 61 data_time: 1.02s time: 524.12s eta: 3 days, 12:17:45 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.326 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31413.0 tgs: 60 data_time: 0.98s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 31857.0 tgs: 61 data_time: 0.78s time: 518.84s eta: 3 days, 11:09:33 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 32.9GB text_tokens: 31570.0 tgs: 60 data_time: 0.72s time: 522.10s eta: 3 days, 11:32:09 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.1GB text_tokens: 31566.0 tgs: 60 data_time: 0.67s time: 523.82s eta: 3 days, 11:39:58 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32203.0 tgs: 61 data_time: 0.84s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 30573.0 tgs: 58 data_time: 0.68s time: 518.43s eta: 3 days, 10:30:58 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.5GB text_tokens: 30164.0 tgs: 57 data_time: 0.71s time: 522.83s eta: 3 days, 11:04:20 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31721.0 tgs: 60 data_time: 0.80s time: 523.53s eta: 3 days, 11:02:17 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 32.7GB text_tokens: 31235.0 tgs: 59 data_time: 0.81s time: 520.96s eta: 3 days, 10:29:09 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.337 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 32083.0 tgs: 61 data_time: 0.72s time: 519.10s eta: 3 days, 10:02:45 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.7GB text_tokens: 31111.0 tgs: 59 data_time: 0.91s time: 523.29s eta: 3 days, 10:33:46 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.0GB text_tokens: 31722.0 tgs: 60 data_time: 0.85s time: 522.95s eta: 3 days, 10:21:53 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.200 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.8GB text_tokens: 30903.0 tgs: 59 data_time: 0.79s time: 521.31s eta: 3 days, 9:57:41 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 31566.0 tgs: 60 data_time: 1.07s time: 520.63s eta: 3 days, 9:42:35 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.1GB text_tokens: 28740.0 tgs: 55 data_time: 0.83s time: 522.43s eta: 3 days, 9:50:50 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 31943.0 tgs: 60 data_time: 0.64s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 31226.0 tgs: 60 data_time: 0.86s time: 520.42s eta: 3 days, 9:14:36 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31930.0 tgs: 61 data_time: 0.69s time: 520.67s eta: 3 days, 9:08:16 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31763.0 tgs: 60 data_time: 0.90s time: 521.31s eta: 3 days, 9:05:31 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 32155.0 tgs: 61 data_time: 0.83s time: 524.10s eta: 3 days, 9:22:53 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31819.0 tgs: 61 data_time: 0.92s time: 520.95s eta: 3 days, 8:44:51 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.204 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 32555.0 tgs: 62 data_time: 0.64s time: 520.17s eta: 3 days, 8:28:52 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.0GB text_tokens: 31507.0 tgs: 60 data_time: 0.75s time: 522.56s eta: 3 days, 8:42:22 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32485.0 tgs: 61 data_time: 0.89s time: 524.32s eta: 3 days, 8:49:59 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 32286.0 tgs: 62 data_time: 0.93s time: 519.88s eta: 3 days, 8:00:12 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32270.0 tgs: 62 data_time: 0.79s time: 520.13s eta: 3 days, 7:53:51 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.9GB text_tokens: 31886.0 tgs: 61 data_time: 0.94s time: 521.99s eta: 3 days, 8:02:18 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.6GB text_tokens: 31139.0 tgs: 59 data_time: 0.75s time: 524.35s eta: 3 days, 8:15:15 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.0GB text_tokens: 31827.0 tgs: 61 data_time: 0.68s time: 520.52s eta: 3 days, 7:31:28 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 32014.0 tgs: 61 data_time: 0.87s time: 518.56s eta: 3 days, 7:04:48 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31950.0 tgs: 61 data_time: 1.07s time: 523.15s eta: 3 days, 7:38:04 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.6GB text_tokens: 31159.0 tgs: 59 data_time: 0.62s time: 523.62s eta: 3 days, 7:33:38 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.0GB text_tokens: 31876.0 tgs: 61 data_time: 0.97s time: 520.18s eta: 3 days, 6:53:40 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31681.0 tgs: 60 data_time: 0.92s time: 520.05s eta: 3 days, 6:43:47 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31594.0 tgs: 60 data_time: 1.19s time: 523.75s eta: 3 days, 7:08:40 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.226 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.9GB text_tokens: 31032.0 tgs: 59 data_time: 1.08s time: 523.80s eta: 3 days, 7:00:24 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31014.0 tgs: 59 data_time: 1.00s time: 520.23s eta: 3 days, 6:19:26 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.329 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 32168.0 tgs: 62 data_time: 0.98s time: 518.77s eta: 3 days, 5:57:35 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.353 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.9GB text_tokens: 31813.0 tgs: 60 data_time: 0.74s time: 522.68s eta: 3 days, 6:24:06 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 32389.0 tgs: 61 data_time: 0.93s time: 522.73s eta: 3 days, 6:15:53 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.8GB text_tokens: 31439.0 tgs: 60 data_time: 0.69s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31610.0 tgs: 60 data_time: 0.60s time: 519.65s eta: 3 days, 5:30:52 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 31493.0 tgs: 60 data_time: 0.87s time: 521.32s eta: 3 days, 5:37:06 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.1GB text_tokens: 31578.0 tgs: 60 data_time: 0.80s time: 523.58s eta: 3 days, 5:48:37 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 32478.0 tgs: 62 data_time: 1.06s time: 519.32s eta: 3 days, 5:01:55 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 32.8GB text_tokens: 29943.0 tgs: 57 data_time: 0.83s time: 520.66s eta: 3 days, 5:05:12 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31038.0 tgs: 59 data_time: 0.57s time: 521.45s eta: 3 days, 5:03:29 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.8GB text_tokens: 31043.0 tgs: 59 data_time: 0.84s time: 523.94s eta: 3 days, 5:16:51 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 32383.0 tgs: 62 data_time: 0.86s time: 518.85s eta: 3 days, 4:23:08 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31991.0 tgs: 61 data_time: 0.69s time: 519.27s eta: 3 days, 4:18:12 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.5GB text_tokens: 30412.0 tgs: 58 data_time: 0.83s time: 520.71s eta: 3 days, 4:22:14 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 31377.0 tgs: 59 data_time: 1.03s time: 524.25s eta: 3 days, 4:44:37 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.319 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.9GB text_tokens: 31712.0 tgs: 60 data_time: 0.75s time: 520.32s eta: 3 days, 4:01:29 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32481.0 tgs: 62 data_time: 0.98s time: 519.49s eta: 3 days, 3:45:32 +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 38][DP 9][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31743.0 tgs: 60 data_time: 0.99s time: 522.31s eta: 3 days, 4:01:30 diff --git a/20250120235238/rank4.log b/20250120235238/rank4.log new file mode 100644 index 0000000000000000000000000000000000000000..4025186d118c4aef3d16cf419df4db6f2fd42628 --- /dev/null +++ b/20250120235238/rank4.log @@ -0,0 +1,395 @@ +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.12s +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.30 seconds, peak gpu memory 13.4G +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.6GB text_tokens: 31121.0 tgs: 56 data_time: 2.29s time: 546.73s eta: 3 days, 18:03:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.344 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.9GB text_tokens: 31200.0 tgs: 59 data_time: 0.96s time: 523.22s eta: 3 days, 14:02:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.8GB text_tokens: 30861.0 tgs: 59 data_time: 0.97s time: 522.94s eta: 3 days, 13:50:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 31960.0 tgs: 61 data_time: 1.16s time: 520.27s eta: 3 days, 13:16:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 1.24s time: 520.97s eta: 3 days, 13:14:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31899.0 tgs: 61 data_time: 0.80s time: 520.91s eta: 3 days, 13:04:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.219 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32173.0 tgs: 61 data_time: 1.03s time: 523.37s eta: 3 days, 13:20:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32073.0 tgs: 61 data_time: 0.82s time: 520.65s eta: 3 days, 12:44:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32230.0 tgs: 61 data_time: 0.86s time: 520.16s eta: 3 days, 12:31:34 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.1GB text_tokens: 32211.0 tgs: 61 data_time: 0.72s time: 520.47s eta: 3 days, 12:25:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 31612.0 tgs: 60 data_time: 0.94s time: 524.52s eta: 3 days, 12:56:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31664.0 tgs: 60 data_time: 0.70s time: 520.63s eta: 3 days, 12:10:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31279.0 tgs: 60 data_time: 0.68s time: 519.96s eta: 3 days, 11:54:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 31574.0 tgs: 60 data_time: 0.89s time: 521.27s eta: 3 days, 11:58:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.6GB text_tokens: 31608.0 tgs: 60 data_time: 0.74s time: 524.11s eta: 3 days, 12:17:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32396.0 tgs: 62 data_time: 0.89s time: 520.52s eta: 3 days, 11:34:22 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32445.0 tgs: 62 data_time: 0.80s time: 518.89s eta: 3 days, 11:09:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32483.0 tgs: 62 data_time: 0.90s time: 522.09s eta: 3 days, 11:32:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.8GB text_tokens: 30947.0 tgs: 59 data_time: 0.89s time: 523.81s eta: 3 days, 11:39:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.395 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32390.0 tgs: 62 data_time: 0.82s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31866.0 tgs: 61 data_time: 0.87s time: 518.42s eta: 3 days, 10:30:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.9GB text_tokens: 31128.0 tgs: 59 data_time: 0.84s time: 522.82s eta: 3 days, 11:04:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32297.0 tgs: 61 data_time: 0.70s time: 523.52s eta: 3 days, 11:02:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.351 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32495.0 tgs: 62 data_time: 1.06s time: 521.01s eta: 3 days, 10:29:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31533.0 tgs: 60 data_time: 0.74s time: 519.09s eta: 3 days, 10:02:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.4GB text_tokens: 31074.0 tgs: 59 data_time: 0.77s time: 523.28s eta: 3 days, 10:33:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32081.0 tgs: 61 data_time: 0.90s time: 522.97s eta: 3 days, 10:22:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.208 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31457.0 tgs: 60 data_time: 0.82s time: 521.30s eta: 3 days, 9:57:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.206 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31099.0 tgs: 59 data_time: 0.73s time: 520.62s eta: 3 days, 9:42:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32202.0 tgs: 61 data_time: 0.77s time: 522.42s eta: 3 days, 9:50:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.7GB text_tokens: 30996.0 tgs: 59 data_time: 0.71s time: 523.77s eta: 3 days, 9:54:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31401.0 tgs: 60 data_time: 0.93s time: 520.41s eta: 3 days, 9:14:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.382 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.7GB text_tokens: 31506.0 tgs: 60 data_time: 0.90s time: 520.66s eta: 3 days, 9:08:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31849.0 tgs: 61 data_time: 0.82s time: 521.32s eta: 3 days, 9:05:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.330 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31503.0 tgs: 60 data_time: 0.67s time: 524.09s eta: 3 days, 9:22:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 32241.0 tgs: 61 data_time: 0.79s time: 520.94s eta: 3 days, 8:44:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31502.0 tgs: 60 data_time: 0.85s time: 520.21s eta: 3 days, 8:29:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.8GB text_tokens: 29886.0 tgs: 57 data_time: 1.01s time: 522.55s eta: 3 days, 8:42:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32309.0 tgs: 61 data_time: 0.78s time: 524.32s eta: 3 days, 8:49:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31388.0 tgs: 60 data_time: 0.92s time: 519.87s eta: 3 days, 8:00:06 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.341 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 31748.0 tgs: 61 data_time: 0.82s time: 520.12s eta: 3 days, 7:53:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 31342.0 tgs: 60 data_time: 0.70s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 31840.0 tgs: 60 data_time: 0.91s time: 524.34s eta: 3 days, 8:15:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31789.0 tgs: 61 data_time: 0.66s time: 520.56s eta: 3 days, 7:31:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31941.0 tgs: 61 data_time: 0.83s time: 518.55s eta: 3 days, 7:04:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 32.2GB text_tokens: 30569.0 tgs: 58 data_time: 0.77s time: 523.14s eta: 3 days, 7:37:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31307.0 tgs: 59 data_time: 0.87s time: 523.61s eta: 3 days, 7:33:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.6GB text_tokens: 31568.0 tgs: 60 data_time: 0.79s time: 520.20s eta: 3 days, 6:53:51 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31707.0 tgs: 60 data_time: 1.11s time: 520.04s eta: 3 days, 6:43:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31866.0 tgs: 60 data_time: 0.90s time: 523.74s eta: 3 days, 7:08:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.356 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31584.0 tgs: 60 data_time: 0.94s time: 523.83s eta: 3 days, 7:00:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.387 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31870.0 tgs: 61 data_time: 0.98s time: 520.22s eta: 3 days, 6:19:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.8GB text_tokens: 30551.0 tgs: 58 data_time: 0.61s time: 518.76s eta: 3 days, 5:57:29 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 31961.0 tgs: 61 data_time: 0.82s time: 522.67s eta: 3 days, 6:24:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.332 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31763.0 tgs: 60 data_time: 1.00s time: 522.77s eta: 3 days, 6:16:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 31917.0 tgs: 61 data_time: 1.08s time: 520.51s eta: 3 days, 5:47:14 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31804.0 tgs: 61 data_time: 0.72s time: 519.64s eta: 3 days, 5:30:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32278.0 tgs: 61 data_time: 0.99s time: 521.34s eta: 3 days, 5:37:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31927.0 tgs: 60 data_time: 0.80s time: 523.57s eta: 3 days, 5:48:31 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31369.0 tgs: 60 data_time: 0.98s time: 519.29s eta: 3 days, 5:01:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32214.0 tgs: 61 data_time: 0.83s time: 520.67s eta: 3 days, 5:05:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31816.0 tgs: 61 data_time: 0.97s time: 521.47s eta: 3 days, 5:03:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.9GB text_tokens: 31958.0 tgs: 60 data_time: 0.65s time: 523.93s eta: 3 days, 5:16:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31545.0 tgs: 60 data_time: 1.03s time: 518.84s eta: 3 days, 4:23:02 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 31565.0 tgs: 60 data_time: 0.81s time: 519.27s eta: 3 days, 4:18:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31753.0 tgs: 60 data_time: 0.71s time: 520.70s eta: 3 days, 4:22:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.5GB text_tokens: 30827.0 tgs: 58 data_time: 0.74s time: 524.24s eta: 3 days, 4:44:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.8GB text_tokens: 31476.0 tgs: 60 data_time: 1.08s time: 520.34s eta: 3 days, 4:01:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31389.0 tgs: 60 data_time: 0.44s time: 519.52s eta: 3 days, 3:45:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.206 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 32.4GB text_tokens: 30909.0 tgs: 59 data_time: 0.91s time: 522.30s eta: 3 days, 4:01:26 diff --git a/20250120235238/rank40.log b/20250120235238/rank40.log new file mode 100644 index 0000000000000000000000000000000000000000..726dae4e647a7046d763c88a4871ffa998cf4245 --- /dev/null +++ b/20250120235238/rank40.log @@ -0,0 +1,395 @@ +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.10s +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.80 seconds, peak gpu memory 13.4G +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.205 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 32069.0 tgs: 58 data_time: 2.08s time: 547.71s eta: 3 days, 18:13:11 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32159.0 tgs: 61 data_time: 0.84s time: 523.25s eta: 3 days, 14:02:41 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 31619.0 tgs: 60 data_time: 0.98s time: 522.88s eta: 3 days, 13:50:24 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 32056.0 tgs: 61 data_time: 0.90s time: 520.29s eta: 3 days, 13:16:09 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.8GB text_tokens: 31273.0 tgs: 60 data_time: 0.99s time: 520.98s eta: 3 days, 13:14:16 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.8GB text_tokens: 31327.0 tgs: 60 data_time: 0.93s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32396.0 tgs: 61 data_time: 0.83s time: 523.38s eta: 3 days, 13:20:24 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.353 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 32.9GB text_tokens: 31929.0 tgs: 61 data_time: 0.87s time: 520.66s eta: 3 days, 12:45:04 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31933.0 tgs: 61 data_time: 0.84s time: 520.17s eta: 3 days, 12:31:41 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 31433.0 tgs: 60 data_time: 0.82s time: 520.42s eta: 3 days, 12:25:27 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32287.0 tgs: 61 data_time: 1.04s time: 524.53s eta: 3 days, 12:56:40 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 31446.0 tgs: 60 data_time: 0.76s time: 520.64s eta: 3 days, 12:10:15 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.9GB text_tokens: 31841.0 tgs: 61 data_time: 0.71s time: 519.96s eta: 3 days, 11:54:58 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 31057.0 tgs: 59 data_time: 0.73s time: 521.28s eta: 3 days, 11:59:04 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 32202.0 tgs: 61 data_time: 0.89s time: 524.12s eta: 3 days, 12:17:45 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32492.0 tgs: 62 data_time: 0.62s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 31095.0 tgs: 59 data_time: 0.61s time: 518.82s eta: 3 days, 11:09:16 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32615.0 tgs: 62 data_time: 0.69s time: 522.10s eta: 3 days, 11:32:09 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32172.0 tgs: 61 data_time: 0.64s time: 523.83s eta: 3 days, 11:39:59 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32340.0 tgs: 62 data_time: 0.60s time: 520.49s eta: 3 days, 10:59:18 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31148.0 tgs: 60 data_time: 0.80s time: 518.43s eta: 3 days, 10:30:58 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.9GB text_tokens: 30975.0 tgs: 59 data_time: 0.63s time: 522.83s eta: 3 days, 11:04:21 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.180 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32237.0 tgs: 61 data_time: 0.96s time: 523.54s eta: 3 days, 11:02:20 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31482.0 tgs: 60 data_time: 0.86s time: 520.96s eta: 3 days, 10:29:08 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.0GB text_tokens: 31615.0 tgs: 60 data_time: 0.65s time: 519.09s eta: 3 days, 10:02:40 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.320 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 32372.0 tgs: 61 data_time: 0.93s time: 523.29s eta: 3 days, 10:33:47 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 27656.0 tgs: 52 data_time: 0.66s time: 522.96s eta: 3 days, 10:21:57 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31754.0 tgs: 60 data_time: 0.54s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.363 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32225.0 tgs: 61 data_time: 0.99s time: 520.62s eta: 3 days, 9:42:32 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32178.0 tgs: 61 data_time: 0.86s time: 522.42s eta: 3 days, 9:50:43 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 32356.0 tgs: 61 data_time: 0.93s time: 523.74s eta: 3 days, 9:54:27 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 32084.0 tgs: 61 data_time: 0.53s time: 520.42s eta: 3 days, 9:14:38 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.9GB text_tokens: 31489.0 tgs: 60 data_time: 0.79s time: 520.67s eta: 3 days, 9:08:13 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31760.0 tgs: 60 data_time: 0.76s time: 521.30s eta: 3 days, 9:05:27 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31719.0 tgs: 60 data_time: 0.91s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.9GB text_tokens: 31694.0 tgs: 60 data_time: 0.89s time: 520.95s eta: 3 days, 8:44:47 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31260.0 tgs: 60 data_time: 0.64s time: 520.19s eta: 3 days, 8:29:04 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32424.0 tgs: 62 data_time: 1.20s time: 522.56s eta: 3 days, 8:42:22 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.9GB text_tokens: 31747.0 tgs: 60 data_time: 0.70s time: 524.33s eta: 3 days, 8:50:00 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 31806.0 tgs: 61 data_time: 0.89s time: 519.88s eta: 3 days, 8:00:13 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.9GB text_tokens: 31095.0 tgs: 59 data_time: 0.81s time: 520.12s eta: 3 days, 7:53:43 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.9GB text_tokens: 31720.0 tgs: 60 data_time: 0.95s time: 521.99s eta: 3 days, 8:02:17 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 31981.0 tgs: 60 data_time: 1.09s time: 524.35s eta: 3 days, 8:15:17 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.0GB text_tokens: 31115.0 tgs: 59 data_time: 0.84s time: 520.52s eta: 3 days, 7:31:24 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.1GB text_tokens: 32109.0 tgs: 61 data_time: 0.79s time: 518.56s eta: 3 days, 7:04:48 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.348 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31354.0 tgs: 59 data_time: 0.75s time: 523.15s eta: 3 days, 7:38:04 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.0GB text_tokens: 31295.0 tgs: 59 data_time: 0.72s time: 523.62s eta: 3 days, 7:33:38 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 32080.0 tgs: 61 data_time: 0.76s time: 520.18s eta: 3 days, 6:53:40 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 32230.0 tgs: 61 data_time: 0.60s time: 520.05s eta: 3 days, 6:43:46 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.9GB text_tokens: 32091.0 tgs: 61 data_time: 0.80s time: 523.75s eta: 3 days, 7:08:41 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.7GB text_tokens: 31199.0 tgs: 59 data_time: 0.95s time: 523.79s eta: 3 days, 7:00:16 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.9GB text_tokens: 31340.0 tgs: 60 data_time: 0.76s time: 520.23s eta: 3 days, 6:19:26 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 31916.0 tgs: 61 data_time: 0.77s time: 518.77s eta: 3 days, 5:57:35 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 31919.0 tgs: 61 data_time: 0.62s time: 522.68s eta: 3 days, 6:24:05 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.5GB text_tokens: 31153.0 tgs: 59 data_time: 0.79s time: 522.75s eta: 3 days, 6:16:00 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.9GB text_tokens: 31392.0 tgs: 60 data_time: 0.89s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.8GB text_tokens: 31449.0 tgs: 60 data_time: 0.82s time: 519.65s eta: 3 days, 5:30:52 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32236.0 tgs: 61 data_time: 0.75s time: 521.31s eta: 3 days, 5:36:59 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.7GB text_tokens: 31580.0 tgs: 60 data_time: 0.64s time: 523.58s eta: 3 days, 5:48:37 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31801.0 tgs: 61 data_time: 0.66s time: 519.32s eta: 3 days, 5:01:55 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31747.0 tgs: 60 data_time: 0.80s time: 520.66s eta: 3 days, 5:05:13 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31667.0 tgs: 60 data_time: 0.99s time: 521.45s eta: 3 days, 5:03:29 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31543.0 tgs: 60 data_time: 0.63s time: 523.94s eta: 3 days, 5:16:51 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 32204.0 tgs: 62 data_time: 0.65s time: 518.84s eta: 3 days, 4:23:07 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.316 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.7GB text_tokens: 30805.0 tgs: 59 data_time: 0.73s time: 519.27s eta: 3 days, 4:18:11 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 31921.0 tgs: 61 data_time: 0.92s time: 520.71s eta: 3 days, 4:22:13 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 32335.0 tgs: 61 data_time: 0.75s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32297.0 tgs: 62 data_time: 0.84s time: 520.31s eta: 3 days, 4:01:25 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31866.0 tgs: 61 data_time: 0.76s time: 519.49s eta: 3 days, 3:45:32 +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 40][DP 10][SP 0][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 32561.0 tgs: 62 data_time: 0.59s time: 522.31s eta: 3 days, 4:01:31 diff --git a/20250120235238/rank42.log b/20250120235238/rank42.log new file mode 100644 index 0000000000000000000000000000000000000000..2b20f83c21ed348bddf037b83137fbab2b82f23b --- /dev/null +++ b/20250120235238/rank42.log @@ -0,0 +1,395 @@ +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.10s +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.80 seconds, peak gpu memory 13.4G +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 32069.0 tgs: 58 data_time: 2.01s time: 547.70s eta: 3 days, 18:13:08 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32159.0 tgs: 61 data_time: 0.80s time: 523.24s eta: 3 days, 14:02:40 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 31619.0 tgs: 60 data_time: 0.98s time: 522.89s eta: 3 days, 13:50:26 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 32056.0 tgs: 61 data_time: 0.87s time: 520.29s eta: 3 days, 13:16:08 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.8GB text_tokens: 31273.0 tgs: 60 data_time: 0.98s time: 520.98s eta: 3 days, 13:14:16 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.8GB text_tokens: 31327.0 tgs: 60 data_time: 0.91s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32396.0 tgs: 61 data_time: 0.79s time: 523.38s eta: 3 days, 13:20:23 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 32.9GB text_tokens: 31929.0 tgs: 61 data_time: 0.83s time: 520.66s eta: 3 days, 12:45:04 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31933.0 tgs: 61 data_time: 0.81s time: 520.17s eta: 3 days, 12:31:41 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.377 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 31433.0 tgs: 60 data_time: 0.80s time: 520.42s eta: 3 days, 12:25:26 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32287.0 tgs: 61 data_time: 1.03s time: 524.53s eta: 3 days, 12:56:39 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 31446.0 tgs: 60 data_time: 0.73s time: 520.64s eta: 3 days, 12:10:14 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.9GB text_tokens: 31841.0 tgs: 61 data_time: 0.71s time: 519.96s eta: 3 days, 11:54:59 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 31057.0 tgs: 59 data_time: 0.71s time: 521.28s eta: 3 days, 11:59:04 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 32202.0 tgs: 61 data_time: 0.95s time: 524.12s eta: 3 days, 12:17:45 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32492.0 tgs: 62 data_time: 0.63s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 31095.0 tgs: 59 data_time: 0.62s time: 518.82s eta: 3 days, 11:09:16 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32615.0 tgs: 62 data_time: 0.68s time: 522.10s eta: 3 days, 11:32:09 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32172.0 tgs: 61 data_time: 0.62s time: 523.83s eta: 3 days, 11:39:59 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32340.0 tgs: 62 data_time: 0.59s time: 520.48s eta: 3 days, 10:59:18 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31148.0 tgs: 60 data_time: 0.75s time: 518.43s eta: 3 days, 10:30:59 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.9GB text_tokens: 30975.0 tgs: 59 data_time: 0.62s time: 522.83s eta: 3 days, 11:04:21 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32237.0 tgs: 61 data_time: 0.85s time: 523.53s eta: 3 days, 11:02:18 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31482.0 tgs: 60 data_time: 0.85s time: 520.96s eta: 3 days, 10:29:07 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.0GB text_tokens: 31615.0 tgs: 60 data_time: 0.66s time: 519.10s eta: 3 days, 10:02:45 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.343 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 32372.0 tgs: 61 data_time: 0.93s time: 523.29s eta: 3 days, 10:33:47 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 27656.0 tgs: 52 data_time: 0.66s time: 522.96s eta: 3 days, 10:21:57 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31754.0 tgs: 60 data_time: 0.54s time: 521.31s eta: 3 days, 9:57:41 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32225.0 tgs: 61 data_time: 0.99s time: 520.63s eta: 3 days, 9:42:37 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32178.0 tgs: 61 data_time: 0.86s time: 522.42s eta: 3 days, 9:50:42 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 32356.0 tgs: 61 data_time: 0.95s time: 523.75s eta: 3 days, 9:54:29 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 32084.0 tgs: 61 data_time: 0.53s time: 520.42s eta: 3 days, 9:14:36 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.9GB text_tokens: 31489.0 tgs: 60 data_time: 0.79s time: 520.67s eta: 3 days, 9:08:16 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.317 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31760.0 tgs: 60 data_time: 0.80s time: 521.30s eta: 3 days, 9:05:26 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31719.0 tgs: 60 data_time: 0.91s time: 524.11s eta: 3 days, 9:22:55 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.9GB text_tokens: 31694.0 tgs: 60 data_time: 0.89s time: 520.95s eta: 3 days, 8:44:51 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31260.0 tgs: 60 data_time: 0.54s time: 520.19s eta: 3 days, 8:29:04 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.372 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32424.0 tgs: 62 data_time: 1.19s time: 522.56s eta: 3 days, 8:42:22 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.9GB text_tokens: 31747.0 tgs: 60 data_time: 0.69s time: 524.33s eta: 3 days, 8:50:01 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 31806.0 tgs: 61 data_time: 0.88s time: 519.88s eta: 3 days, 8:00:13 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.9GB text_tokens: 31095.0 tgs: 59 data_time: 0.80s time: 520.11s eta: 3 days, 7:53:41 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.9GB text_tokens: 31720.0 tgs: 60 data_time: 0.95s time: 521.99s eta: 3 days, 8:02:20 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 31981.0 tgs: 60 data_time: 1.08s time: 524.35s eta: 3 days, 8:15:14 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.0GB text_tokens: 31115.0 tgs: 59 data_time: 0.85s time: 520.52s eta: 3 days, 7:31:27 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.1GB text_tokens: 32109.0 tgs: 61 data_time: 0.80s time: 518.56s eta: 3 days, 7:04:48 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31354.0 tgs: 59 data_time: 0.74s time: 523.15s eta: 3 days, 7:38:05 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.0GB text_tokens: 31295.0 tgs: 59 data_time: 0.70s time: 523.62s eta: 3 days, 7:33:40 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.317 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 32080.0 tgs: 61 data_time: 0.76s time: 520.18s eta: 3 days, 6:53:39 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 32230.0 tgs: 61 data_time: 0.60s time: 520.05s eta: 3 days, 6:43:46 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.9GB text_tokens: 32091.0 tgs: 61 data_time: 0.80s time: 523.75s eta: 3 days, 7:08:41 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.7GB text_tokens: 31199.0 tgs: 59 data_time: 0.94s time: 523.79s eta: 3 days, 7:00:18 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.9GB text_tokens: 31340.0 tgs: 60 data_time: 0.76s time: 520.23s eta: 3 days, 6:19:27 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 31916.0 tgs: 61 data_time: 0.77s time: 518.77s eta: 3 days, 5:57:35 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 31919.0 tgs: 61 data_time: 0.62s time: 522.68s eta: 3 days, 6:24:07 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.5GB text_tokens: 31153.0 tgs: 59 data_time: 0.79s time: 522.75s eta: 3 days, 6:16:00 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.9GB text_tokens: 31392.0 tgs: 60 data_time: 0.89s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.8GB text_tokens: 31449.0 tgs: 60 data_time: 0.82s time: 519.65s eta: 3 days, 5:30:53 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.210 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32236.0 tgs: 61 data_time: 0.75s time: 521.31s eta: 3 days, 5:37:00 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.7GB text_tokens: 31580.0 tgs: 60 data_time: 0.64s time: 523.59s eta: 3 days, 5:48:38 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31801.0 tgs: 61 data_time: 0.61s time: 519.32s eta: 3 days, 5:01:55 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31747.0 tgs: 60 data_time: 0.78s time: 520.66s eta: 3 days, 5:05:13 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31667.0 tgs: 60 data_time: 0.99s time: 521.45s eta: 3 days, 5:03:29 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31543.0 tgs: 60 data_time: 0.62s time: 523.94s eta: 3 days, 5:16:52 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 32204.0 tgs: 62 data_time: 0.64s time: 518.85s eta: 3 days, 4:23:09 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.7GB text_tokens: 30805.0 tgs: 59 data_time: 0.73s time: 519.26s eta: 3 days, 4:18:10 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 31921.0 tgs: 61 data_time: 0.92s time: 520.71s eta: 3 days, 4:22:14 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 32335.0 tgs: 61 data_time: 0.75s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32297.0 tgs: 62 data_time: 0.84s time: 520.31s eta: 3 days, 4:01:24 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31866.0 tgs: 61 data_time: 0.73s time: 519.49s eta: 3 days, 3:45:33 +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 42][DP 10][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 32561.0 tgs: 62 data_time: 0.59s time: 522.31s eta: 3 days, 4:01:31 diff --git a/20250120235238/rank43.log b/20250120235238/rank43.log new file mode 100644 index 0000000000000000000000000000000000000000..ea09c8d2630965e32a2430115f023b131091601f --- /dev/null +++ b/20250120235238/rank43.log @@ -0,0 +1,395 @@ +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.13s +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.69 seconds, peak gpu memory 13.4G +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.360 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 32069.0 tgs: 58 data_time: 2.01s time: 547.56s eta: 3 days, 18:11:41 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32159.0 tgs: 61 data_time: 0.85s time: 523.24s eta: 3 days, 14:02:36 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 31619.0 tgs: 60 data_time: 1.00s time: 522.89s eta: 3 days, 13:50:25 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 32056.0 tgs: 61 data_time: 0.92s time: 520.29s eta: 3 days, 13:16:09 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.329 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.8GB text_tokens: 31273.0 tgs: 60 data_time: 1.02s time: 520.98s eta: 3 days, 13:14:15 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.318 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.8GB text_tokens: 31327.0 tgs: 60 data_time: 0.93s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32396.0 tgs: 61 data_time: 0.82s time: 523.38s eta: 3 days, 13:20:23 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 32.9GB text_tokens: 31929.0 tgs: 61 data_time: 0.82s time: 520.66s eta: 3 days, 12:45:04 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31933.0 tgs: 61 data_time: 0.83s time: 520.17s eta: 3 days, 12:31:40 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 31433.0 tgs: 60 data_time: 0.83s time: 520.42s eta: 3 days, 12:25:26 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32287.0 tgs: 61 data_time: 1.05s time: 524.53s eta: 3 days, 12:56:39 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 31446.0 tgs: 60 data_time: 0.75s time: 520.64s eta: 3 days, 12:10:14 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.9GB text_tokens: 31841.0 tgs: 61 data_time: 0.74s time: 519.96s eta: 3 days, 11:54:58 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 31057.0 tgs: 59 data_time: 0.73s time: 521.28s eta: 3 days, 11:59:02 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 32202.0 tgs: 61 data_time: 0.90s time: 524.12s eta: 3 days, 12:17:45 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32492.0 tgs: 62 data_time: 0.65s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 31095.0 tgs: 59 data_time: 0.64s time: 518.82s eta: 3 days, 11:09:16 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32615.0 tgs: 62 data_time: 0.69s time: 522.10s eta: 3 days, 11:32:08 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32172.0 tgs: 61 data_time: 0.64s time: 523.82s eta: 3 days, 11:39:58 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 32340.0 tgs: 62 data_time: 0.61s time: 520.48s eta: 3 days, 10:59:17 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.342 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31148.0 tgs: 60 data_time: 0.77s time: 518.43s eta: 3 days, 10:30:58 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.9GB text_tokens: 30975.0 tgs: 59 data_time: 0.64s time: 522.83s eta: 3 days, 11:04:20 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32237.0 tgs: 61 data_time: 0.86s time: 523.53s eta: 3 days, 11:02:17 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31482.0 tgs: 60 data_time: 0.87s time: 520.96s eta: 3 days, 10:29:06 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.0GB text_tokens: 31615.0 tgs: 60 data_time: 0.69s time: 519.10s eta: 3 days, 10:02:45 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 32372.0 tgs: 61 data_time: 0.95s time: 523.29s eta: 3 days, 10:33:47 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 27656.0 tgs: 52 data_time: 0.69s time: 522.96s eta: 3 days, 10:21:55 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31754.0 tgs: 60 data_time: 0.56s time: 521.31s eta: 3 days, 9:57:42 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.339 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32225.0 tgs: 61 data_time: 1.03s time: 520.63s eta: 3 days, 9:42:37 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32178.0 tgs: 61 data_time: 0.88s time: 522.41s eta: 3 days, 9:50:41 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 32356.0 tgs: 61 data_time: 0.95s time: 523.75s eta: 3 days, 9:54:30 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 32084.0 tgs: 61 data_time: 0.55s time: 520.42s eta: 3 days, 9:14:34 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.9GB text_tokens: 31489.0 tgs: 60 data_time: 0.82s time: 520.67s eta: 3 days, 9:08:17 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.341 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31760.0 tgs: 60 data_time: 0.76s time: 521.29s eta: 3 days, 9:05:25 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31719.0 tgs: 60 data_time: 0.93s time: 524.10s eta: 3 days, 9:22:54 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.9GB text_tokens: 31694.0 tgs: 60 data_time: 0.92s time: 520.95s eta: 3 days, 8:44:51 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31260.0 tgs: 60 data_time: 0.58s time: 520.19s eta: 3 days, 8:29:03 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.375 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 32424.0 tgs: 62 data_time: 1.24s time: 522.56s eta: 3 days, 8:42:22 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.9GB text_tokens: 31747.0 tgs: 60 data_time: 0.71s time: 524.33s eta: 3 days, 8:50:00 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 31806.0 tgs: 61 data_time: 0.89s time: 519.88s eta: 3 days, 8:00:12 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.9GB text_tokens: 31095.0 tgs: 59 data_time: 0.82s time: 520.11s eta: 3 days, 7:53:41 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.9GB text_tokens: 31720.0 tgs: 60 data_time: 0.98s time: 521.99s eta: 3 days, 8:02:19 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 31981.0 tgs: 60 data_time: 1.10s time: 524.35s eta: 3 days, 8:15:14 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.0GB text_tokens: 31115.0 tgs: 59 data_time: 0.87s time: 520.52s eta: 3 days, 7:31:27 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.1GB text_tokens: 32109.0 tgs: 61 data_time: 0.82s time: 518.56s eta: 3 days, 7:04:48 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31354.0 tgs: 59 data_time: 0.76s time: 523.15s eta: 3 days, 7:38:05 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.0GB text_tokens: 31295.0 tgs: 59 data_time: 0.73s time: 523.62s eta: 3 days, 7:33:38 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 32080.0 tgs: 61 data_time: 0.79s time: 520.18s eta: 3 days, 6:53:40 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 32230.0 tgs: 61 data_time: 0.62s time: 520.05s eta: 3 days, 6:43:46 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.9GB text_tokens: 32091.0 tgs: 61 data_time: 0.84s time: 523.75s eta: 3 days, 7:08:41 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.7GB text_tokens: 31199.0 tgs: 59 data_time: 0.97s time: 523.79s eta: 3 days, 7:00:18 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.9GB text_tokens: 31340.0 tgs: 60 data_time: 0.78s time: 520.23s eta: 3 days, 6:19:26 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 31916.0 tgs: 61 data_time: 0.78s time: 518.77s eta: 3 days, 5:57:34 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 31919.0 tgs: 61 data_time: 0.64s time: 522.68s eta: 3 days, 6:24:07 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.5GB text_tokens: 31153.0 tgs: 59 data_time: 0.82s time: 522.74s eta: 3 days, 6:15:59 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.9GB text_tokens: 31392.0 tgs: 60 data_time: 0.92s time: 520.52s eta: 3 days, 5:47:21 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.8GB text_tokens: 31449.0 tgs: 60 data_time: 0.84s time: 519.65s eta: 3 days, 5:30:52 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32236.0 tgs: 61 data_time: 0.78s time: 521.31s eta: 3 days, 5:36:59 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.7GB text_tokens: 31580.0 tgs: 60 data_time: 0.66s time: 523.58s eta: 3 days, 5:48:37 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31801.0 tgs: 61 data_time: 0.64s time: 519.32s eta: 3 days, 5:01:54 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31747.0 tgs: 60 data_time: 0.80s time: 520.66s eta: 3 days, 5:05:13 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31667.0 tgs: 60 data_time: 1.02s time: 521.45s eta: 3 days, 5:03:30 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31543.0 tgs: 60 data_time: 0.65s time: 523.94s eta: 3 days, 5:16:50 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 32204.0 tgs: 62 data_time: 0.67s time: 518.85s eta: 3 days, 4:23:08 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 32.7GB text_tokens: 30805.0 tgs: 59 data_time: 0.76s time: 519.27s eta: 3 days, 4:18:11 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 31921.0 tgs: 61 data_time: 0.95s time: 520.71s eta: 3 days, 4:22:14 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.337 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 32335.0 tgs: 61 data_time: 0.78s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32297.0 tgs: 62 data_time: 0.86s time: 520.31s eta: 3 days, 4:01:24 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31866.0 tgs: 61 data_time: 0.77s time: 519.49s eta: 3 days, 3:45:32 +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 43][DP 10][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 32561.0 tgs: 62 data_time: 0.62s time: 522.31s eta: 3 days, 4:01:31 diff --git a/20250120235238/rank44.log b/20250120235238/rank44.log new file mode 100644 index 0000000000000000000000000000000000000000..57a70dc5557485b69b2b72ec3e646623c1ccd7f7 --- /dev/null +++ b/20250120235238/rank44.log @@ -0,0 +1,395 @@ +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.15s +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.80 seconds, peak gpu memory 13.4G +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.359 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.0GB text_tokens: 31361.0 tgs: 57 data_time: 1.70s time: 547.71s eta: 3 days, 18:13:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32373.0 tgs: 61 data_time: 1.12s time: 523.24s eta: 3 days, 14:02:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.221 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 31271.0 tgs: 59 data_time: 1.12s time: 522.89s eta: 3 days, 13:50:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31959.0 tgs: 61 data_time: 0.77s time: 520.29s eta: 3 days, 13:16:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 32119.0 tgs: 61 data_time: 0.60s time: 520.98s eta: 3 days, 13:14:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.8GB text_tokens: 31094.0 tgs: 59 data_time: 0.83s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.9GB text_tokens: 31860.0 tgs: 60 data_time: 1.02s time: 523.38s eta: 3 days, 13:20:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.380 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 32.6GB text_tokens: 31574.0 tgs: 60 data_time: 0.73s time: 520.66s eta: 3 days, 12:45:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.207 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31714.0 tgs: 60 data_time: 0.74s time: 520.17s eta: 3 days, 12:31:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 31891.0 tgs: 61 data_time: 0.80s time: 520.42s eta: 3 days, 12:25:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.350 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 31327.0 tgs: 59 data_time: 0.87s time: 524.53s eta: 3 days, 12:56:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 32020.0 tgs: 61 data_time: 0.82s time: 520.64s eta: 3 days, 12:10:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31061.0 tgs: 59 data_time: 0.69s time: 519.96s eta: 3 days, 11:54:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.315 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 30891.0 tgs: 59 data_time: 0.76s time: 521.28s eta: 3 days, 11:59:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 31341.0 tgs: 59 data_time: 0.80s time: 524.12s eta: 3 days, 12:17:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.9GB text_tokens: 31438.0 tgs: 60 data_time: 0.91s time: 520.53s eta: 3 days, 11:34:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.0GB text_tokens: 32067.0 tgs: 61 data_time: 0.67s time: 518.82s eta: 3 days, 11:09:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32309.0 tgs: 61 data_time: 0.96s time: 522.10s eta: 3 days, 11:32:09 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31293.0 tgs: 59 data_time: 0.78s time: 523.82s eta: 3 days, 11:39:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.9GB text_tokens: 31896.0 tgs: 61 data_time: 0.74s time: 520.49s eta: 3 days, 10:59:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31834.0 tgs: 61 data_time: 0.68s time: 518.43s eta: 3 days, 10:30:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 32278.0 tgs: 61 data_time: 0.75s time: 522.84s eta: 3 days, 11:04:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 31172.0 tgs: 59 data_time: 0.95s time: 523.54s eta: 3 days, 11:02:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32324.0 tgs: 62 data_time: 0.96s time: 520.96s eta: 3 days, 10:29:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 32419.0 tgs: 62 data_time: 0.90s time: 519.10s eta: 3 days, 10:02:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.7GB text_tokens: 31182.0 tgs: 59 data_time: 0.84s time: 523.29s eta: 3 days, 10:33:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31901.0 tgs: 61 data_time: 0.70s time: 522.96s eta: 3 days, 10:21:57 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 30544.0 tgs: 58 data_time: 0.97s time: 521.31s eta: 3 days, 9:57:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 31830.0 tgs: 61 data_time: 0.76s time: 520.63s eta: 3 days, 9:42:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.329 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.1GB text_tokens: 31792.0 tgs: 60 data_time: 1.11s time: 522.41s eta: 3 days, 9:50:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.1GB text_tokens: 31968.0 tgs: 61 data_time: 0.81s time: 523.75s eta: 3 days, 9:54:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31133.0 tgs: 59 data_time: 0.85s time: 520.42s eta: 3 days, 9:14:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 32219.0 tgs: 61 data_time: 0.83s time: 520.67s eta: 3 days, 9:08:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31570.0 tgs: 60 data_time: 0.93s time: 521.30s eta: 3 days, 9:05:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.320 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 32328.0 tgs: 61 data_time: 0.99s time: 524.11s eta: 3 days, 9:22:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31564.0 tgs: 60 data_time: 0.95s time: 520.95s eta: 3 days, 8:44:51 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31631.0 tgs: 60 data_time: 0.99s time: 520.19s eta: 3 days, 8:29:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 31761.0 tgs: 60 data_time: 0.76s time: 522.56s eta: 3 days, 8:42:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 31545.0 tgs: 60 data_time: 0.86s time: 524.33s eta: 3 days, 8:50:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.215 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 30209.0 tgs: 58 data_time: 0.75s time: 519.88s eta: 3 days, 8:00:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.338 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32351.0 tgs: 62 data_time: 0.63s time: 520.11s eta: 3 days, 7:53:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.8GB text_tokens: 31789.0 tgs: 60 data_time: 0.74s time: 521.99s eta: 3 days, 8:02:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 32049.0 tgs: 61 data_time: 0.72s time: 524.35s eta: 3 days, 8:15:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 32186.0 tgs: 61 data_time: 0.79s time: 520.52s eta: 3 days, 7:31:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.9GB text_tokens: 31935.0 tgs: 61 data_time: 0.85s time: 518.56s eta: 3 days, 7:04:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 32419.0 tgs: 61 data_time: 0.78s time: 523.15s eta: 3 days, 7:38:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.9GB text_tokens: 32027.0 tgs: 61 data_time: 0.86s time: 523.62s eta: 3 days, 7:33:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 31765.0 tgs: 61 data_time: 0.59s time: 520.19s eta: 3 days, 6:53:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.5GB text_tokens: 31151.0 tgs: 59 data_time: 0.82s time: 520.05s eta: 3 days, 6:43:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 32470.0 tgs: 61 data_time: 0.85s time: 523.75s eta: 3 days, 7:08:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.206 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31763.0 tgs: 60 data_time: 0.89s time: 523.79s eta: 3 days, 7:00:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32319.0 tgs: 62 data_time: 0.91s time: 520.23s eta: 3 days, 6:19:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.6GB text_tokens: 30805.0 tgs: 59 data_time: 0.85s time: 518.77s eta: 3 days, 5:57:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 32262.0 tgs: 61 data_time: 0.75s time: 522.68s eta: 3 days, 6:24:07 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 32600.0 tgs: 62 data_time: 0.83s time: 522.75s eta: 3 days, 6:16:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.8GB text_tokens: 31887.0 tgs: 61 data_time: 0.81s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.0GB text_tokens: 32253.0 tgs: 62 data_time: 0.95s time: 519.65s eta: 3 days, 5:30:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 29875.0 tgs: 57 data_time: 0.78s time: 521.31s eta: 3 days, 5:37:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.6GB text_tokens: 31345.0 tgs: 59 data_time: 0.77s time: 523.58s eta: 3 days, 5:48:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.0GB text_tokens: 30572.0 tgs: 58 data_time: 0.93s time: 519.32s eta: 3 days, 5:01:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31052.0 tgs: 59 data_time: 0.88s time: 520.66s eta: 3 days, 5:05:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31497.0 tgs: 60 data_time: 0.74s time: 521.45s eta: 3 days, 5:03:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31868.0 tgs: 60 data_time: 0.82s time: 523.94s eta: 3 days, 5:16:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31269.0 tgs: 60 data_time: 0.78s time: 518.85s eta: 3 days, 4:23:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31590.0 tgs: 60 data_time: 0.70s time: 519.27s eta: 3 days, 4:18:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 32429.0 tgs: 62 data_time: 0.72s time: 520.71s eta: 3 days, 4:22:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 29122.0 tgs: 55 data_time: 0.86s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.9GB text_tokens: 31224.0 tgs: 60 data_time: 0.71s time: 520.32s eta: 3 days, 4:01:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.0GB text_tokens: 31914.0 tgs: 61 data_time: 1.10s time: 519.49s eta: 3 days, 3:45:32 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.215 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31436.0 tgs: 60 data_time: 1.08s time: 522.31s eta: 3 days, 4:01:31 diff --git a/20250120235238/rank46.log b/20250120235238/rank46.log new file mode 100644 index 0000000000000000000000000000000000000000..5782c16d435d5392bf5c2cd5f1adefc1dbe00e08 --- /dev/null +++ b/20250120235238/rank46.log @@ -0,0 +1,395 @@ +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.16s +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.80 seconds, peak gpu memory 13.4G +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.315 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.0GB text_tokens: 31361.0 tgs: 57 data_time: 1.69s time: 547.71s eta: 3 days, 18:13:12 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 32373.0 tgs: 61 data_time: 1.11s time: 523.23s eta: 3 days, 14:02:34 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 31271.0 tgs: 59 data_time: 1.16s time: 522.89s eta: 3 days, 13:50:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31959.0 tgs: 61 data_time: 0.77s time: 520.28s eta: 3 days, 13:16:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.1GB text_tokens: 32119.0 tgs: 61 data_time: 0.59s time: 520.98s eta: 3 days, 13:14:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.8GB text_tokens: 31094.0 tgs: 59 data_time: 0.77s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.9GB text_tokens: 31860.0 tgs: 60 data_time: 1.00s time: 523.38s eta: 3 days, 13:20:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.344 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 32.6GB text_tokens: 31574.0 tgs: 60 data_time: 0.72s time: 520.66s eta: 3 days, 12:45:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31714.0 tgs: 60 data_time: 0.76s time: 520.17s eta: 3 days, 12:31:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 31891.0 tgs: 61 data_time: 0.79s time: 520.42s eta: 3 days, 12:25:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 31327.0 tgs: 59 data_time: 0.86s time: 524.53s eta: 3 days, 12:56:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.210 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 32020.0 tgs: 61 data_time: 0.81s time: 520.64s eta: 3 days, 12:10:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31061.0 tgs: 59 data_time: 0.67s time: 519.96s eta: 3 days, 11:54:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 30891.0 tgs: 59 data_time: 0.70s time: 521.28s eta: 3 days, 11:59:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 31341.0 tgs: 59 data_time: 0.77s time: 524.12s eta: 3 days, 12:17:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.9GB text_tokens: 31438.0 tgs: 60 data_time: 0.83s time: 520.53s eta: 3 days, 11:34:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.0GB text_tokens: 32067.0 tgs: 61 data_time: 0.65s time: 518.82s eta: 3 days, 11:09:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32309.0 tgs: 61 data_time: 0.97s time: 522.10s eta: 3 days, 11:32:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.9GB text_tokens: 31293.0 tgs: 59 data_time: 0.76s time: 523.82s eta: 3 days, 11:39:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.9GB text_tokens: 31896.0 tgs: 61 data_time: 0.70s time: 520.48s eta: 3 days, 10:59:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31834.0 tgs: 61 data_time: 0.67s time: 518.42s eta: 3 days, 10:30:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 32278.0 tgs: 61 data_time: 0.71s time: 522.83s eta: 3 days, 11:04:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 31172.0 tgs: 59 data_time: 0.89s time: 523.53s eta: 3 days, 11:02:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32324.0 tgs: 62 data_time: 0.89s time: 520.96s eta: 3 days, 10:29:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 32419.0 tgs: 62 data_time: 0.85s time: 519.10s eta: 3 days, 10:02:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.351 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.7GB text_tokens: 31182.0 tgs: 59 data_time: 0.80s time: 523.29s eta: 3 days, 10:33:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 31901.0 tgs: 61 data_time: 0.67s time: 522.96s eta: 3 days, 10:21:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.316 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 30544.0 tgs: 58 data_time: 0.95s time: 521.31s eta: 3 days, 9:57:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 31830.0 tgs: 61 data_time: 0.73s time: 520.63s eta: 3 days, 9:42:37 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.317 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.1GB text_tokens: 31792.0 tgs: 60 data_time: 1.12s time: 522.42s eta: 3 days, 9:50:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.1GB text_tokens: 31968.0 tgs: 61 data_time: 0.81s time: 523.75s eta: 3 days, 9:54:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31133.0 tgs: 59 data_time: 0.80s time: 520.42s eta: 3 days, 9:14:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 32219.0 tgs: 61 data_time: 0.87s time: 520.67s eta: 3 days, 9:08:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31570.0 tgs: 60 data_time: 0.92s time: 521.30s eta: 3 days, 9:05:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 32328.0 tgs: 61 data_time: 1.02s time: 524.10s eta: 3 days, 9:22:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31564.0 tgs: 60 data_time: 0.96s time: 520.95s eta: 3 days, 8:44:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31631.0 tgs: 60 data_time: 0.93s time: 520.19s eta: 3 days, 8:29:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.1GB text_tokens: 31761.0 tgs: 60 data_time: 0.75s time: 522.56s eta: 3 days, 8:42:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 31545.0 tgs: 60 data_time: 0.87s time: 524.33s eta: 3 days, 8:50:01 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 30209.0 tgs: 58 data_time: 0.76s time: 519.88s eta: 3 days, 8:00:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.340 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32351.0 tgs: 62 data_time: 0.65s time: 520.11s eta: 3 days, 7:53:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.354 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.8GB text_tokens: 31789.0 tgs: 60 data_time: 0.70s time: 521.99s eta: 3 days, 8:02:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 32049.0 tgs: 61 data_time: 0.73s time: 524.35s eta: 3 days, 8:15:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 32186.0 tgs: 61 data_time: 0.80s time: 520.52s eta: 3 days, 7:31:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.9GB text_tokens: 31935.0 tgs: 61 data_time: 0.81s time: 518.56s eta: 3 days, 7:04:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 32419.0 tgs: 61 data_time: 0.76s time: 523.15s eta: 3 days, 7:38:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.9GB text_tokens: 32027.0 tgs: 61 data_time: 0.84s time: 523.62s eta: 3 days, 7:33:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 33.1GB text_tokens: 31765.0 tgs: 61 data_time: 0.58s time: 520.19s eta: 3 days, 6:53:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.5GB text_tokens: 31151.0 tgs: 59 data_time: 0.79s time: 520.05s eta: 3 days, 6:43:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 32470.0 tgs: 61 data_time: 0.80s time: 523.75s eta: 3 days, 7:08:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31763.0 tgs: 60 data_time: 0.86s time: 523.79s eta: 3 days, 7:00:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.320 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32319.0 tgs: 62 data_time: 0.88s time: 520.23s eta: 3 days, 6:19:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.6GB text_tokens: 30805.0 tgs: 59 data_time: 0.83s time: 518.77s eta: 3 days, 5:57:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 32262.0 tgs: 61 data_time: 0.73s time: 522.68s eta: 3 days, 6:24:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 32600.0 tgs: 62 data_time: 0.80s time: 522.75s eta: 3 days, 6:16:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.329 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.8GB text_tokens: 31887.0 tgs: 61 data_time: 0.79s time: 520.52s eta: 3 days, 5:47:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.0GB text_tokens: 32253.0 tgs: 62 data_time: 0.94s time: 519.65s eta: 3 days, 5:30:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 29875.0 tgs: 57 data_time: 0.76s time: 521.31s eta: 3 days, 5:37:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.6GB text_tokens: 31345.0 tgs: 59 data_time: 0.76s time: 523.59s eta: 3 days, 5:48:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.0GB text_tokens: 30572.0 tgs: 58 data_time: 0.90s time: 519.32s eta: 3 days, 5:01:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31052.0 tgs: 59 data_time: 0.86s time: 520.66s eta: 3 days, 5:05:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31497.0 tgs: 60 data_time: 0.72s time: 521.45s eta: 3 days, 5:03:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31868.0 tgs: 60 data_time: 0.78s time: 523.94s eta: 3 days, 5:16:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31269.0 tgs: 60 data_time: 0.81s time: 518.85s eta: 3 days, 4:23:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31590.0 tgs: 60 data_time: 0.73s time: 519.23s eta: 3 days, 4:17:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 32429.0 tgs: 62 data_time: 0.79s time: 520.71s eta: 3 days, 4:22:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 29122.0 tgs: 55 data_time: 0.86s time: 524.25s eta: 3 days, 4:44:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.9GB text_tokens: 31224.0 tgs: 60 data_time: 0.72s time: 520.32s eta: 3 days, 4:01:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.0GB text_tokens: 31914.0 tgs: 61 data_time: 1.07s time: 519.49s eta: 3 days, 3:45:32 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31436.0 tgs: 60 data_time: 1.04s time: 522.31s eta: 3 days, 4:01:31 diff --git a/20250120235238/rank49.log b/20250120235238/rank49.log new file mode 100644 index 0000000000000000000000000000000000000000..d00e17c274cbbb688cd2c0a4fb096f226413d868 --- /dev/null +++ b/20250120235238/rank49.log @@ -0,0 +1,395 @@ +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.11s +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.38 seconds, peak gpu memory 13.4G +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31357.0 tgs: 57 data_time: 1.95s time: 548.09s eta: 3 days, 18:16:54 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.0GB text_tokens: 32248.0 tgs: 61 data_time: 0.88s time: 523.23s eta: 3 days, 14:02:30 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.8GB text_tokens: 31853.0 tgs: 60 data_time: 1.03s time: 522.87s eta: 3 days, 13:50:18 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.330 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 32083.0 tgs: 61 data_time: 0.88s time: 520.30s eta: 3 days, 13:16:16 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.9GB text_tokens: 32125.0 tgs: 61 data_time: 0.98s time: 520.98s eta: 3 days, 13:14:15 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.9GB text_tokens: 31521.0 tgs: 60 data_time: 0.79s time: 520.89s eta: 3 days, 13:04:45 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 31512.0 tgs: 60 data_time: 0.89s time: 523.39s eta: 3 days, 13:20:27 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31822.0 tgs: 61 data_time: 0.99s time: 520.66s eta: 3 days, 12:45:05 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.341 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31908.0 tgs: 61 data_time: 0.80s time: 520.18s eta: 3 days, 12:31:47 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.352 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.1GB text_tokens: 32433.0 tgs: 62 data_time: 0.83s time: 520.39s eta: 3 days, 12:25:10 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 32333.0 tgs: 61 data_time: 0.77s time: 524.54s eta: 3 days, 12:56:44 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.201 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 30295.0 tgs: 58 data_time: 0.98s time: 520.65s eta: 3 days, 12:10:18 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.331 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 31362.0 tgs: 60 data_time: 0.81s time: 519.96s eta: 3 days, 11:54:56 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32181.0 tgs: 61 data_time: 0.89s time: 521.29s eta: 3 days, 11:59:07 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 31770.0 tgs: 60 data_time: 1.07s time: 524.13s eta: 3 days, 12:17:48 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.326 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32107.0 tgs: 61 data_time: 1.05s time: 520.54s eta: 3 days, 11:34:31 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.0GB text_tokens: 31655.0 tgs: 61 data_time: 0.69s time: 518.79s eta: 3 days, 11:09:03 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 32.9GB text_tokens: 31529.0 tgs: 60 data_time: 0.90s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.1GB text_tokens: 31277.0 tgs: 59 data_time: 0.77s time: 523.83s eta: 3 days, 11:40:03 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.1GB text_tokens: 31388.0 tgs: 60 data_time: 0.63s time: 520.47s eta: 3 days, 10:59:11 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 32144.0 tgs: 62 data_time: 0.73s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.8GB text_tokens: 31305.0 tgs: 59 data_time: 0.70s time: 522.84s eta: 3 days, 11:04:23 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 32167.0 tgs: 61 data_time: 0.89s time: 523.52s eta: 3 days, 11:02:11 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32301.0 tgs: 62 data_time: 0.79s time: 520.96s eta: 3 days, 10:29:07 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.206 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 31599.0 tgs: 60 data_time: 0.70s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 32431.0 tgs: 61 data_time: 0.74s time: 523.30s eta: 3 days, 10:33:52 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.0GB text_tokens: 31812.0 tgs: 60 data_time: 0.85s time: 522.95s eta: 3 days, 10:21:49 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.0GB text_tokens: 30722.0 tgs: 58 data_time: 0.63s time: 521.31s eta: 3 days, 9:57:42 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.0GB text_tokens: 29912.0 tgs: 57 data_time: 0.86s time: 520.64s eta: 3 days, 9:42:40 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 31300.0 tgs: 59 data_time: 1.03s time: 522.40s eta: 3 days, 9:50:31 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.1GB text_tokens: 32379.0 tgs: 61 data_time: 1.06s time: 523.75s eta: 3 days, 9:54:31 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.8GB text_tokens: 31647.0 tgs: 60 data_time: 0.71s time: 520.43s eta: 3 days, 9:14:42 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.7GB text_tokens: 31134.0 tgs: 59 data_time: 0.81s time: 520.67s eta: 3 days, 9:08:17 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.343 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31309.0 tgs: 60 data_time: 0.66s time: 521.29s eta: 3 days, 9:05:22 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.9GB text_tokens: 31520.0 tgs: 60 data_time: 0.62s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.329 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.1GB text_tokens: 32247.0 tgs: 61 data_time: 0.86s time: 520.96s eta: 3 days, 8:44:55 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31693.0 tgs: 60 data_time: 0.92s time: 520.17s eta: 3 days, 8:28:54 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.213 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.0GB text_tokens: 32127.0 tgs: 61 data_time: 0.94s time: 522.56s eta: 3 days, 8:42:26 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.0GB text_tokens: 30901.0 tgs: 58 data_time: 0.66s time: 524.33s eta: 3 days, 8:50:02 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 31581.0 tgs: 60 data_time: 0.86s time: 519.89s eta: 3 days, 8:00:17 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 31785.0 tgs: 61 data_time: 0.79s time: 520.10s eta: 3 days, 7:53:33 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.8GB text_tokens: 31496.0 tgs: 60 data_time: 0.88s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32186.0 tgs: 61 data_time: 0.73s time: 524.36s eta: 3 days, 8:15:20 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31986.0 tgs: 61 data_time: 0.79s time: 520.51s eta: 3 days, 7:31:18 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31430.0 tgs: 60 data_time: 0.84s time: 518.57s eta: 3 days, 7:04:53 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31889.0 tgs: 60 data_time: 0.74s time: 523.15s eta: 3 days, 7:38:08 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.7GB text_tokens: 31343.0 tgs: 59 data_time: 0.94s time: 523.63s eta: 3 days, 7:33:45 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.340 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31765.0 tgs: 61 data_time: 0.79s time: 520.16s eta: 3 days, 6:53:27 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 31720.0 tgs: 60 data_time: 1.02s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.210 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31630.0 tgs: 60 data_time: 1.15s time: 523.75s eta: 3 days, 7:08:42 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.340 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31547.0 tgs: 60 data_time: 0.87s time: 523.80s eta: 3 days, 7:00:21 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 31899.0 tgs: 61 data_time: 0.85s time: 520.24s eta: 3 days, 6:19:31 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.7GB text_tokens: 30536.0 tgs: 58 data_time: 0.85s time: 518.77s eta: 3 days, 5:57:36 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 31859.0 tgs: 60 data_time: 0.75s time: 522.66s eta: 3 days, 6:23:59 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.8GB text_tokens: 31761.0 tgs: 60 data_time: 0.80s time: 522.72s eta: 3 days, 6:15:46 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.359 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.0GB text_tokens: 31082.0 tgs: 59 data_time: 0.82s time: 520.53s eta: 3 days, 5:47:23 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.9GB text_tokens: 32012.0 tgs: 61 data_time: 0.78s time: 519.66s eta: 3 days, 5:30:56 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.8GB text_tokens: 31850.0 tgs: 61 data_time: 0.93s time: 521.30s eta: 3 days, 5:36:58 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 30425.0 tgs: 58 data_time: 0.98s time: 523.59s eta: 3 days, 5:48:42 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.217 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 31774.0 tgs: 61 data_time: 0.59s time: 519.32s eta: 3 days, 5:01:58 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 32.9GB text_tokens: 30693.0 tgs: 58 data_time: 0.85s time: 520.64s eta: 3 days, 5:05:02 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.1GB text_tokens: 31736.0 tgs: 60 data_time: 0.95s time: 521.45s eta: 3 days, 5:03:32 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.1GB text_tokens: 30661.0 tgs: 58 data_time: 0.68s time: 523.95s eta: 3 days, 5:16:55 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 32237.0 tgs: 62 data_time: 1.02s time: 518.85s eta: 3 days, 4:23:11 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31226.0 tgs: 60 data_time: 1.11s time: 519.24s eta: 3 days, 4:17:59 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31462.0 tgs: 60 data_time: 0.82s time: 520.72s eta: 3 days, 4:22:18 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.9GB text_tokens: 31003.0 tgs: 59 data_time: 0.89s time: 524.25s eta: 3 days, 4:44:41 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.0GB text_tokens: 30989.0 tgs: 59 data_time: 0.59s time: 520.30s eta: 3 days, 4:01:19 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31722.0 tgs: 61 data_time: 0.71s time: 519.50s eta: 3 days, 3:45:36 +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 49][DP 12][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 32510.0 tgs: 62 data_time: 0.83s time: 522.32s eta: 3 days, 4:01:33 diff --git a/20250120235238/rank5.log b/20250120235238/rank5.log new file mode 100644 index 0000000000000000000000000000000000000000..a24e0e6cc583662845d0ad7fe8b76777e7deca57 --- /dev/null +++ b/20250120235238/rank5.log @@ -0,0 +1,395 @@ +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.21s +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 145.61 seconds, peak gpu memory 13.4G +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.6GB text_tokens: 31121.0 tgs: 56 data_time: 2.26s time: 546.73s eta: 3 days, 18:03:30 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.9GB text_tokens: 31200.0 tgs: 59 data_time: 0.85s time: 523.22s eta: 3 days, 14:02:25 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.8GB text_tokens: 30861.0 tgs: 59 data_time: 0.93s time: 522.94s eta: 3 days, 13:50:56 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 31960.0 tgs: 61 data_time: 0.95s time: 520.28s eta: 3 days, 13:16:02 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 1.15s time: 520.97s eta: 3 days, 13:14:09 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.222 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31899.0 tgs: 61 data_time: 0.76s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32173.0 tgs: 61 data_time: 1.01s time: 523.37s eta: 3 days, 13:20:16 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32073.0 tgs: 61 data_time: 0.75s time: 520.65s eta: 3 days, 12:44:58 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.338 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32230.0 tgs: 61 data_time: 0.81s time: 520.16s eta: 3 days, 12:31:35 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.1GB text_tokens: 32211.0 tgs: 61 data_time: 0.69s time: 520.48s eta: 3 days, 12:25:57 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 31612.0 tgs: 60 data_time: 0.90s time: 524.52s eta: 3 days, 12:56:33 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31664.0 tgs: 60 data_time: 0.66s time: 520.63s eta: 3 days, 12:10:08 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31279.0 tgs: 60 data_time: 0.67s time: 519.96s eta: 3 days, 11:54:56 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 31574.0 tgs: 60 data_time: 0.85s time: 521.28s eta: 3 days, 11:58:59 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.6GB text_tokens: 31608.0 tgs: 60 data_time: 0.71s time: 524.11s eta: 3 days, 12:17:36 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32396.0 tgs: 62 data_time: 0.84s time: 520.52s eta: 3 days, 11:34:22 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32445.0 tgs: 62 data_time: 0.77s time: 518.89s eta: 3 days, 11:09:59 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32483.0 tgs: 62 data_time: 0.85s time: 522.09s eta: 3 days, 11:32:04 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.8GB text_tokens: 30947.0 tgs: 59 data_time: 0.88s time: 523.81s eta: 3 days, 11:39:52 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32390.0 tgs: 62 data_time: 0.77s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31866.0 tgs: 61 data_time: 0.86s time: 518.42s eta: 3 days, 10:30:52 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.308 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.9GB text_tokens: 31128.0 tgs: 59 data_time: 0.82s time: 522.82s eta: 3 days, 11:04:15 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32297.0 tgs: 61 data_time: 0.67s time: 523.52s eta: 3 days, 11:02:10 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32495.0 tgs: 62 data_time: 1.01s time: 521.01s eta: 3 days, 10:29:36 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31533.0 tgs: 60 data_time: 0.71s time: 519.09s eta: 3 days, 10:02:39 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.4GB text_tokens: 31074.0 tgs: 59 data_time: 0.73s time: 523.28s eta: 3 days, 10:33:40 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.319 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32081.0 tgs: 61 data_time: 0.84s time: 522.97s eta: 3 days, 10:22:02 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31457.0 tgs: 60 data_time: 0.78s time: 521.30s eta: 3 days, 9:57:35 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31099.0 tgs: 59 data_time: 0.71s time: 520.62s eta: 3 days, 9:42:30 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32202.0 tgs: 61 data_time: 0.74s time: 522.42s eta: 3 days, 9:50:43 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.7GB text_tokens: 30996.0 tgs: 59 data_time: 0.69s time: 523.77s eta: 3 days, 9:54:43 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31401.0 tgs: 60 data_time: 0.93s time: 520.41s eta: 3 days, 9:14:30 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.7GB text_tokens: 31506.0 tgs: 60 data_time: 0.86s time: 520.66s eta: 3 days, 9:08:11 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31849.0 tgs: 61 data_time: 0.80s time: 521.32s eta: 3 days, 9:05:39 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31503.0 tgs: 60 data_time: 0.63s time: 524.09s eta: 3 days, 9:22:48 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 32241.0 tgs: 61 data_time: 0.76s time: 520.94s eta: 3 days, 8:44:46 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31502.0 tgs: 60 data_time: 0.83s time: 520.21s eta: 3 days, 8:29:18 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.8GB text_tokens: 29886.0 tgs: 57 data_time: 0.93s time: 522.55s eta: 3 days, 8:42:19 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32309.0 tgs: 61 data_time: 0.74s time: 524.31s eta: 3 days, 8:49:54 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31388.0 tgs: 60 data_time: 0.85s time: 519.87s eta: 3 days, 8:00:07 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.322 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 31748.0 tgs: 61 data_time: 0.79s time: 520.12s eta: 3 days, 7:53:46 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.354 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 31342.0 tgs: 60 data_time: 0.67s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 31840.0 tgs: 60 data_time: 0.84s time: 524.34s eta: 3 days, 8:15:09 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31789.0 tgs: 61 data_time: 0.64s time: 520.56s eta: 3 days, 7:31:46 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31941.0 tgs: 61 data_time: 0.78s time: 518.55s eta: 3 days, 7:04:43 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 32.2GB text_tokens: 30569.0 tgs: 58 data_time: 0.76s time: 523.14s eta: 3 days, 7:37:58 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31307.0 tgs: 59 data_time: 0.82s time: 523.61s eta: 3 days, 7:33:33 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.6GB text_tokens: 31568.0 tgs: 60 data_time: 0.75s time: 520.20s eta: 3 days, 6:53:51 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31707.0 tgs: 60 data_time: 1.06s time: 520.04s eta: 3 days, 6:43:40 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31866.0 tgs: 60 data_time: 0.86s time: 523.74s eta: 3 days, 7:08:35 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31584.0 tgs: 60 data_time: 0.88s time: 523.83s eta: 3 days, 7:00:37 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31870.0 tgs: 61 data_time: 0.95s time: 520.22s eta: 3 days, 6:19:20 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.8GB text_tokens: 30551.0 tgs: 58 data_time: 0.59s time: 518.76s eta: 3 days, 5:57:29 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 31961.0 tgs: 61 data_time: 0.79s time: 522.67s eta: 3 days, 6:24:01 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31763.0 tgs: 60 data_time: 0.98s time: 522.77s eta: 3 days, 6:16:11 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 31917.0 tgs: 61 data_time: 1.03s time: 520.51s eta: 3 days, 5:47:14 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31804.0 tgs: 61 data_time: 0.69s time: 519.64s eta: 3 days, 5:30:47 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32278.0 tgs: 61 data_time: 0.95s time: 521.34s eta: 3 days, 5:37:16 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.336 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31927.0 tgs: 60 data_time: 0.79s time: 523.57s eta: 3 days, 5:48:31 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31369.0 tgs: 60 data_time: 0.93s time: 519.29s eta: 3 days, 5:01:42 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32214.0 tgs: 61 data_time: 0.80s time: 520.67s eta: 3 days, 5:05:14 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31816.0 tgs: 61 data_time: 0.89s time: 521.47s eta: 3 days, 5:03:41 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.9GB text_tokens: 31958.0 tgs: 60 data_time: 0.60s time: 523.93s eta: 3 days, 5:16:46 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31545.0 tgs: 60 data_time: 0.96s time: 518.84s eta: 3 days, 4:23:02 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 31565.0 tgs: 60 data_time: 0.74s time: 519.27s eta: 3 days, 4:18:14 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31753.0 tgs: 60 data_time: 0.68s time: 520.70s eta: 3 days, 4:22:08 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.340 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.5GB text_tokens: 30827.0 tgs: 58 data_time: 0.72s time: 524.24s eta: 3 days, 4:44:33 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.8GB text_tokens: 31476.0 tgs: 60 data_time: 1.04s time: 520.34s eta: 3 days, 4:01:38 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31389.0 tgs: 60 data_time: 0.43s time: 519.52s eta: 3 days, 3:45:47 +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 5][DP 1][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 32.4GB text_tokens: 30909.0 tgs: 59 data_time: 0.89s time: 522.30s eta: 3 days, 4:01:26 diff --git a/20250120235238/rank54.log b/20250120235238/rank54.log new file mode 100644 index 0000000000000000000000000000000000000000..c41548a2912be85e1ed8020042d333bd50469fd3 --- /dev/null +++ b/20250120235238/rank54.log @@ -0,0 +1,395 @@ +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.11s +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.36 seconds, peak gpu memory 13.4G +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31804.0 tgs: 58 data_time: 1.90s time: 548.10s eta: 3 days, 18:17:03 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31938.0 tgs: 61 data_time: 0.70s time: 523.22s eta: 3 days, 14:02:26 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 31420.0 tgs: 60 data_time: 0.77s time: 522.87s eta: 3 days, 13:50:17 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.315 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 32320.0 tgs: 62 data_time: 0.93s time: 520.29s eta: 3 days, 13:16:12 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.9GB text_tokens: 31755.0 tgs: 60 data_time: 0.87s time: 520.99s eta: 3 days, 13:14:20 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.0GB text_tokens: 31116.0 tgs: 59 data_time: 0.95s time: 520.89s eta: 3 days, 13:04:45 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32297.0 tgs: 61 data_time: 0.83s time: 523.39s eta: 3 days, 13:20:27 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32099.0 tgs: 61 data_time: 0.87s time: 520.66s eta: 3 days, 12:45:08 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 31943.0 tgs: 61 data_time: 0.95s time: 520.18s eta: 3 days, 12:31:44 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.1GB text_tokens: 32321.0 tgs: 62 data_time: 0.82s time: 520.40s eta: 3 days, 12:25:13 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 31685.0 tgs: 60 data_time: 1.21s time: 524.53s eta: 3 days, 12:56:43 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 31570.0 tgs: 60 data_time: 1.07s time: 520.65s eta: 3 days, 12:10:18 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.1GB text_tokens: 32271.0 tgs: 62 data_time: 1.02s time: 519.96s eta: 3 days, 11:54:56 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 31826.0 tgs: 61 data_time: 0.85s time: 521.29s eta: 3 days, 11:59:08 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.0GB text_tokens: 31152.0 tgs: 59 data_time: 0.74s time: 524.13s eta: 3 days, 12:17:49 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31903.0 tgs: 61 data_time: 0.85s time: 520.54s eta: 3 days, 11:34:32 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 30941.0 tgs: 59 data_time: 0.67s time: 518.79s eta: 3 days, 11:09:03 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.185 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 29085.0 tgs: 55 data_time: 0.67s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.1GB text_tokens: 32477.0 tgs: 61 data_time: 1.11s time: 523.83s eta: 3 days, 11:40:03 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.9GB text_tokens: 31064.0 tgs: 59 data_time: 0.73s time: 520.47s eta: 3 days, 10:59:10 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.1GB text_tokens: 31960.0 tgs: 61 data_time: 0.99s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31106.0 tgs: 59 data_time: 0.98s time: 522.84s eta: 3 days, 11:04:24 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 31337.0 tgs: 59 data_time: 1.02s time: 523.52s eta: 3 days, 11:02:11 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 32.8GB text_tokens: 31611.0 tgs: 60 data_time: 0.98s time: 520.96s eta: 3 days, 10:29:08 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.204 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31573.0 tgs: 60 data_time: 0.69s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.1GB text_tokens: 31897.0 tgs: 60 data_time: 1.07s time: 523.29s eta: 3 days, 10:33:51 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32058.0 tgs: 61 data_time: 0.94s time: 522.95s eta: 3 days, 10:21:50 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31119.0 tgs: 59 data_time: 0.81s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.219 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.7GB text_tokens: 31356.0 tgs: 60 data_time: 0.57s time: 520.64s eta: 3 days, 9:42:40 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.1GB text_tokens: 32225.0 tgs: 61 data_time: 1.08s time: 522.39s eta: 3 days, 9:50:29 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 31625.0 tgs: 60 data_time: 0.76s time: 523.75s eta: 3 days, 9:54:33 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 31942.0 tgs: 61 data_time: 0.86s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.318 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.8GB text_tokens: 32061.0 tgs: 61 data_time: 0.89s time: 520.68s eta: 3 days, 9:08:20 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.337 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.4GB text_tokens: 30275.0 tgs: 58 data_time: 0.96s time: 521.29s eta: 3 days, 9:05:22 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.198 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.7GB text_tokens: 31108.0 tgs: 59 data_time: 0.77s time: 524.11s eta: 3 days, 9:22:58 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 32.9GB text_tokens: 31957.0 tgs: 61 data_time: 0.78s time: 520.96s eta: 3 days, 8:44:55 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 32.5GB text_tokens: 29801.0 tgs: 57 data_time: 0.97s time: 520.17s eta: 3 days, 8:28:53 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.0GB text_tokens: 31102.0 tgs: 59 data_time: 0.87s time: 522.56s eta: 3 days, 8:42:25 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32173.0 tgs: 61 data_time: 0.80s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 31492.0 tgs: 60 data_time: 0.92s time: 519.89s eta: 3 days, 8:00:17 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.0GB text_tokens: 31539.0 tgs: 60 data_time: 0.90s time: 520.10s eta: 3 days, 7:53:33 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 32436.0 tgs: 62 data_time: 1.20s time: 522.00s eta: 3 days, 8:02:24 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.345 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32092.0 tgs: 61 data_time: 0.72s time: 524.35s eta: 3 days, 8:15:18 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.0GB text_tokens: 31585.0 tgs: 60 data_time: 0.87s time: 520.51s eta: 3 days, 7:31:21 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.1GB text_tokens: 30729.0 tgs: 59 data_time: 0.47s time: 518.57s eta: 3 days, 7:04:52 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31914.0 tgs: 61 data_time: 0.76s time: 523.15s eta: 3 days, 7:38:08 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.9GB text_tokens: 31429.0 tgs: 60 data_time: 0.96s time: 523.63s eta: 3 days, 7:33:43 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.7GB text_tokens: 31767.0 tgs: 61 data_time: 0.96s time: 520.16s eta: 3 days, 6:53:28 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31842.0 tgs: 61 data_time: 0.87s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.385 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31119.0 tgs: 59 data_time: 0.89s time: 523.76s eta: 3 days, 7:08:45 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.315 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 32.7GB text_tokens: 31541.0 tgs: 60 data_time: 1.25s time: 523.80s eta: 3 days, 7:00:21 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 32.8GB text_tokens: 31567.0 tgs: 60 data_time: 0.79s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.0GB text_tokens: 32276.0 tgs: 62 data_time: 1.00s time: 518.78s eta: 3 days, 5:57:38 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 31104.0 tgs: 59 data_time: 0.71s time: 522.67s eta: 3 days, 6:23:59 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 32112.0 tgs: 61 data_time: 0.78s time: 522.72s eta: 3 days, 6:15:46 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.209 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 32.7GB text_tokens: 31195.0 tgs: 59 data_time: 0.77s time: 520.53s eta: 3 days, 5:47:23 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31401.0 tgs: 60 data_time: 0.66s time: 519.66s eta: 3 days, 5:30:57 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 31338.0 tgs: 60 data_time: 0.84s time: 521.31s eta: 3 days, 5:37:00 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.0GB text_tokens: 31585.0 tgs: 60 data_time: 0.80s time: 523.59s eta: 3 days, 5:48:41 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.6GB text_tokens: 31441.0 tgs: 60 data_time: 0.79s time: 519.32s eta: 3 days, 5:01:59 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32230.0 tgs: 61 data_time: 1.05s time: 520.64s eta: 3 days, 5:05:03 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 32203.0 tgs: 61 data_time: 0.85s time: 521.45s eta: 3 days, 5:03:32 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.1GB text_tokens: 31194.0 tgs: 59 data_time: 0.83s time: 523.95s eta: 3 days, 5:16:56 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31045.0 tgs: 59 data_time: 0.76s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 32341.0 tgs: 62 data_time: 0.82s time: 519.25s eta: 3 days, 4:18:00 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.1GB text_tokens: 32008.0 tgs: 61 data_time: 0.78s time: 520.72s eta: 3 days, 4:22:18 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 31320.0 tgs: 59 data_time: 1.24s time: 524.25s eta: 3 days, 4:44:41 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.215 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.8GB text_tokens: 31671.0 tgs: 60 data_time: 0.43s time: 520.30s eta: 3 days, 4:01:18 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31891.0 tgs: 61 data_time: 0.75s time: 519.50s eta: 3 days, 3:45:36 +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 54][DP 13][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.373 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 31509.0 tgs: 60 data_time: 0.93s time: 522.32s eta: 3 days, 4:01:35 diff --git a/20250120235238/rank56.log b/20250120235238/rank56.log new file mode 100644 index 0000000000000000000000000000000000000000..3c99839a6e1b24a9fcfb4eff0a7908d0eca36e0e --- /dev/null +++ b/20250120235238/rank56.log @@ -0,0 +1,395 @@ +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.36s +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 142.52 seconds, peak gpu memory 13.4G +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 32438.0 tgs: 59 data_time: 2.38s time: 546.75s eta: 3 days, 18:03:41 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31049.0 tgs: 59 data_time: 0.93s time: 523.24s eta: 3 days, 14:02:40 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.9GB text_tokens: 31400.0 tgs: 60 data_time: 1.00s time: 522.87s eta: 3 days, 13:50:15 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31528.0 tgs: 60 data_time: 0.78s time: 520.30s eta: 3 days, 13:16:16 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.8GB text_tokens: 31808.0 tgs: 61 data_time: 0.96s time: 520.98s eta: 3 days, 13:14:19 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 32312.0 tgs: 62 data_time: 0.74s time: 520.91s eta: 3 days, 13:04:57 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 31349.0 tgs: 59 data_time: 0.73s time: 523.38s eta: 3 days, 13:20:24 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 31993.0 tgs: 61 data_time: 0.72s time: 520.66s eta: 3 days, 12:45:08 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 32365.0 tgs: 62 data_time: 0.68s time: 520.18s eta: 3 days, 12:31:45 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.397 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.7GB text_tokens: 31497.0 tgs: 60 data_time: 0.83s time: 520.37s eta: 3 days, 12:24:56 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 31507.0 tgs: 60 data_time: 0.87s time: 524.53s eta: 3 days, 12:56:43 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.4GB text_tokens: 29899.0 tgs: 57 data_time: 0.78s time: 520.66s eta: 3 days, 12:10:23 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.9GB text_tokens: 32129.0 tgs: 61 data_time: 0.86s time: 519.96s eta: 3 days, 11:54:59 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.7GB text_tokens: 31562.0 tgs: 60 data_time: 0.89s time: 521.29s eta: 3 days, 11:59:08 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 30596.0 tgs: 58 data_time: 0.82s time: 524.13s eta: 3 days, 12:17:50 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31446.0 tgs: 60 data_time: 0.91s time: 520.54s eta: 3 days, 11:34:33 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.5GB text_tokens: 30935.0 tgs: 59 data_time: 0.62s time: 518.78s eta: 3 days, 11:08:54 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31582.0 tgs: 60 data_time: 0.72s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32037.0 tgs: 61 data_time: 0.78s time: 523.83s eta: 3 days, 11:40:03 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32131.0 tgs: 61 data_time: 0.70s time: 520.49s eta: 3 days, 10:59:20 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.1GB text_tokens: 32342.0 tgs: 62 data_time: 0.74s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31716.0 tgs: 60 data_time: 0.66s time: 522.84s eta: 3 days, 11:04:25 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31956.0 tgs: 61 data_time: 0.90s time: 523.50s eta: 3 days, 11:01:57 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.0GB text_tokens: 32071.0 tgs: 61 data_time: 1.02s time: 520.96s eta: 3 days, 10:29:06 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 31325.0 tgs: 60 data_time: 0.71s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.6GB text_tokens: 30772.0 tgs: 58 data_time: 0.95s time: 523.30s eta: 3 days, 10:33:52 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.348 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 1.02s time: 522.95s eta: 3 days, 10:21:51 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.311 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.9GB text_tokens: 31970.0 tgs: 61 data_time: 1.04s time: 521.32s eta: 3 days, 9:57:45 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32184.0 tgs: 61 data_time: 0.62s time: 520.64s eta: 3 days, 9:42:40 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 31024.0 tgs: 59 data_time: 0.91s time: 522.41s eta: 3 days, 9:50:38 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.4GB text_tokens: 30697.0 tgs: 58 data_time: 0.94s time: 523.75s eta: 3 days, 9:54:33 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.9GB text_tokens: 32236.0 tgs: 61 data_time: 0.76s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.212 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31376.0 tgs: 60 data_time: 0.71s time: 520.68s eta: 3 days, 9:08:21 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.9GB text_tokens: 31253.0 tgs: 59 data_time: 0.67s time: 521.26s eta: 3 days, 9:05:06 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 31.7GB text_tokens: 27839.0 tgs: 53 data_time: 0.68s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.237 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.1GB text_tokens: 31971.0 tgs: 61 data_time: 0.84s time: 520.96s eta: 3 days, 8:44:56 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 32205.0 tgs: 61 data_time: 0.74s time: 520.17s eta: 3 days, 8:28:55 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.193 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.5GB text_tokens: 30720.0 tgs: 58 data_time: 0.76s time: 522.56s eta: 3 days, 8:42:26 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.8GB text_tokens: 32017.0 tgs: 61 data_time: 0.89s time: 524.33s eta: 3 days, 8:50:04 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31169.0 tgs: 59 data_time: 0.62s time: 519.89s eta: 3 days, 8:00:17 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32306.0 tgs: 62 data_time: 0.62s time: 520.09s eta: 3 days, 7:53:28 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 30552.0 tgs: 58 data_time: 0.75s time: 522.00s eta: 3 days, 8:02:24 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.198 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 31.9GB text_tokens: 27829.0 tgs: 53 data_time: 0.64s time: 524.36s eta: 3 days, 8:15:19 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 32389.0 tgs: 62 data_time: 0.62s time: 520.50s eta: 3 days, 7:31:17 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31067.0 tgs: 59 data_time: 0.67s time: 518.57s eta: 3 days, 7:04:52 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 32547.0 tgs: 62 data_time: 0.82s time: 523.16s eta: 3 days, 7:38:09 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 32221.0 tgs: 61 data_time: 0.73s time: 523.63s eta: 3 days, 7:33:43 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31075.0 tgs: 59 data_time: 0.56s time: 520.16s eta: 3 days, 6:53:28 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.9GB text_tokens: 30517.0 tgs: 58 data_time: 0.74s time: 520.06s eta: 3 days, 6:43:51 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.7GB text_tokens: 31792.0 tgs: 60 data_time: 0.83s time: 523.76s eta: 3 days, 7:08:45 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31413.0 tgs: 59 data_time: 1.05s time: 523.78s eta: 3 days, 7:00:13 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31691.0 tgs: 60 data_time: 0.82s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 32130.0 tgs: 61 data_time: 0.65s time: 518.78s eta: 3 days, 5:57:38 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.7GB text_tokens: 31532.0 tgs: 60 data_time: 0.58s time: 522.67s eta: 3 days, 6:24:01 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.4GB text_tokens: 31076.0 tgs: 59 data_time: 0.82s time: 522.72s eta: 3 days, 6:15:47 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 31660.0 tgs: 60 data_time: 0.70s time: 520.53s eta: 3 days, 5:47:24 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32304.0 tgs: 62 data_time: 0.67s time: 519.66s eta: 3 days, 5:30:57 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 31855.0 tgs: 61 data_time: 0.87s time: 521.30s eta: 3 days, 5:36:56 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 0.62s time: 523.59s eta: 3 days, 5:48:41 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.0GB text_tokens: 32321.0 tgs: 62 data_time: 0.88s time: 519.32s eta: 3 days, 5:01:59 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 32465.0 tgs: 62 data_time: 0.75s time: 520.64s eta: 3 days, 5:05:00 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.8GB text_tokens: 31067.0 tgs: 59 data_time: 0.73s time: 521.45s eta: 3 days, 5:03:32 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.9GB text_tokens: 31472.0 tgs: 60 data_time: 0.82s time: 523.95s eta: 3 days, 5:16:55 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 30640.0 tgs: 59 data_time: 1.08s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.372 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 32323.0 tgs: 62 data_time: 0.79s time: 519.23s eta: 3 days, 4:17:55 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 32138.0 tgs: 61 data_time: 0.81s time: 520.72s eta: 3 days, 4:22:17 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 31246.0 tgs: 59 data_time: 0.73s time: 524.26s eta: 3 days, 4:44:42 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 30657.0 tgs: 58 data_time: 0.62s time: 520.32s eta: 3 days, 4:01:29 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 32.7GB text_tokens: 31712.0 tgs: 61 data_time: 0.93s time: 519.50s eta: 3 days, 3:45:37 +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 56][DP 14][SP 0][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31495.0 tgs: 60 data_time: 0.73s time: 522.32s eta: 3 days, 4:01:33 diff --git a/20250120235238/rank59.log b/20250120235238/rank59.log new file mode 100644 index 0000000000000000000000000000000000000000..660df612f60edf4aa0d76ebfcd89c2cf34cfdc0f --- /dev/null +++ b/20250120235238/rank59.log @@ -0,0 +1,395 @@ +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.12s +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 142.74 seconds, peak gpu memory 13.4G +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 32438.0 tgs: 59 data_time: 2.40s time: 546.74s eta: 3 days, 18:03:35 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31049.0 tgs: 59 data_time: 0.93s time: 523.25s eta: 3 days, 14:02:41 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.9GB text_tokens: 31400.0 tgs: 60 data_time: 1.01s time: 522.87s eta: 3 days, 13:50:15 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.0GB text_tokens: 31528.0 tgs: 60 data_time: 0.74s time: 520.29s eta: 3 days, 13:16:13 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.8GB text_tokens: 31808.0 tgs: 61 data_time: 0.96s time: 520.99s eta: 3 days, 13:14:20 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 32312.0 tgs: 62 data_time: 0.74s time: 520.91s eta: 3 days, 13:04:56 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 31349.0 tgs: 59 data_time: 0.73s time: 523.39s eta: 3 days, 13:20:28 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 31993.0 tgs: 61 data_time: 0.71s time: 520.67s eta: 3 days, 12:45:09 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.0GB text_tokens: 32365.0 tgs: 62 data_time: 0.70s time: 520.18s eta: 3 days, 12:31:45 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.7GB text_tokens: 31497.0 tgs: 60 data_time: 0.85s time: 520.37s eta: 3 days, 12:24:56 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.1GB text_tokens: 31507.0 tgs: 60 data_time: 0.86s time: 524.53s eta: 3 days, 12:56:43 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.4GB text_tokens: 29899.0 tgs: 57 data_time: 0.77s time: 520.65s eta: 3 days, 12:10:19 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.9GB text_tokens: 32129.0 tgs: 61 data_time: 0.86s time: 519.97s eta: 3 days, 11:55:02 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.7GB text_tokens: 31562.0 tgs: 60 data_time: 0.88s time: 521.29s eta: 3 days, 11:59:09 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 30596.0 tgs: 58 data_time: 0.83s time: 524.13s eta: 3 days, 12:17:50 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 31446.0 tgs: 60 data_time: 0.93s time: 520.54s eta: 3 days, 11:34:31 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.5GB text_tokens: 30935.0 tgs: 59 data_time: 0.63s time: 518.78s eta: 3 days, 11:08:55 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31582.0 tgs: 60 data_time: 0.73s time: 522.11s eta: 3 days, 11:32:14 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 32037.0 tgs: 61 data_time: 0.79s time: 523.83s eta: 3 days, 11:40:04 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32131.0 tgs: 61 data_time: 0.71s time: 520.49s eta: 3 days, 10:59:20 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.1GB text_tokens: 32342.0 tgs: 62 data_time: 0.75s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31716.0 tgs: 60 data_time: 0.66s time: 522.84s eta: 3 days, 11:04:26 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 31956.0 tgs: 61 data_time: 0.87s time: 523.50s eta: 3 days, 11:01:57 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.0GB text_tokens: 32071.0 tgs: 61 data_time: 1.01s time: 520.96s eta: 3 days, 10:29:08 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 33.1GB text_tokens: 31325.0 tgs: 60 data_time: 0.68s time: 519.10s eta: 3 days, 10:02:49 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.6GB text_tokens: 30772.0 tgs: 58 data_time: 0.94s time: 523.30s eta: 3 days, 10:33:52 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 1.01s time: 522.95s eta: 3 days, 10:21:51 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 32.9GB text_tokens: 31970.0 tgs: 61 data_time: 1.04s time: 521.32s eta: 3 days, 9:57:46 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 33.1GB text_tokens: 32184.0 tgs: 61 data_time: 0.62s time: 520.64s eta: 3 days, 9:42:41 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 31024.0 tgs: 59 data_time: 0.91s time: 522.41s eta: 3 days, 9:50:39 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.4GB text_tokens: 30697.0 tgs: 58 data_time: 0.95s time: 523.76s eta: 3 days, 9:54:34 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.9GB text_tokens: 32236.0 tgs: 61 data_time: 0.77s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 31376.0 tgs: 60 data_time: 0.72s time: 520.68s eta: 3 days, 9:08:20 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 32.9GB text_tokens: 31253.0 tgs: 59 data_time: 0.67s time: 521.26s eta: 3 days, 9:05:07 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 31.7GB text_tokens: 27839.0 tgs: 53 data_time: 0.68s time: 524.11s eta: 3 days, 9:22:58 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.1GB text_tokens: 31971.0 tgs: 61 data_time: 0.85s time: 520.96s eta: 3 days, 8:44:56 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 32205.0 tgs: 61 data_time: 0.74s time: 520.17s eta: 3 days, 8:28:56 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.5GB text_tokens: 30720.0 tgs: 58 data_time: 0.78s time: 522.57s eta: 3 days, 8:42:26 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.8GB text_tokens: 32017.0 tgs: 61 data_time: 0.89s time: 524.33s eta: 3 days, 8:50:05 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31169.0 tgs: 59 data_time: 0.63s time: 519.89s eta: 3 days, 8:00:17 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32306.0 tgs: 62 data_time: 0.62s time: 520.09s eta: 3 days, 7:53:29 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.0GB text_tokens: 30552.0 tgs: 58 data_time: 0.74s time: 522.00s eta: 3 days, 8:02:24 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 31.9GB text_tokens: 27829.0 tgs: 53 data_time: 0.64s time: 524.35s eta: 3 days, 8:15:19 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 32389.0 tgs: 62 data_time: 0.58s time: 520.51s eta: 3 days, 7:31:19 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31067.0 tgs: 59 data_time: 0.69s time: 518.57s eta: 3 days, 7:04:53 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 32547.0 tgs: 62 data_time: 0.82s time: 523.16s eta: 3 days, 7:38:09 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 32221.0 tgs: 61 data_time: 0.77s time: 523.63s eta: 3 days, 7:33:44 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31075.0 tgs: 59 data_time: 0.56s time: 520.16s eta: 3 days, 6:53:28 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 32.9GB text_tokens: 30517.0 tgs: 58 data_time: 0.74s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.7GB text_tokens: 31792.0 tgs: 60 data_time: 0.81s time: 523.76s eta: 3 days, 7:08:46 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31413.0 tgs: 59 data_time: 1.04s time: 523.79s eta: 3 days, 7:00:17 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31691.0 tgs: 60 data_time: 0.83s time: 520.24s eta: 3 days, 6:19:29 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 32130.0 tgs: 61 data_time: 0.67s time: 518.78s eta: 3 days, 5:57:39 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.7GB text_tokens: 31532.0 tgs: 60 data_time: 0.59s time: 522.67s eta: 3 days, 6:24:00 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.4GB text_tokens: 31076.0 tgs: 59 data_time: 0.76s time: 522.73s eta: 3 days, 6:15:49 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 31660.0 tgs: 60 data_time: 0.73s time: 520.53s eta: 3 days, 5:47:24 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32304.0 tgs: 62 data_time: 0.66s time: 519.66s eta: 3 days, 5:30:57 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 31855.0 tgs: 61 data_time: 0.86s time: 521.30s eta: 3 days, 5:36:57 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 0.61s time: 523.60s eta: 3 days, 5:48:43 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.0GB text_tokens: 32321.0 tgs: 62 data_time: 0.87s time: 519.32s eta: 3 days, 5:01:59 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 32465.0 tgs: 62 data_time: 0.74s time: 520.64s eta: 3 days, 5:05:01 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.8GB text_tokens: 31067.0 tgs: 59 data_time: 0.74s time: 521.45s eta: 3 days, 5:03:33 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.9GB text_tokens: 31472.0 tgs: 60 data_time: 0.80s time: 523.95s eta: 3 days, 5:16:57 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 30640.0 tgs: 59 data_time: 1.07s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 32323.0 tgs: 62 data_time: 0.78s time: 519.24s eta: 3 days, 4:17:55 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 32138.0 tgs: 61 data_time: 0.81s time: 520.72s eta: 3 days, 4:22:18 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 31246.0 tgs: 59 data_time: 0.73s time: 524.26s eta: 3 days, 4:44:43 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 30657.0 tgs: 58 data_time: 0.62s time: 520.32s eta: 3 days, 4:01:30 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 32.7GB text_tokens: 31712.0 tgs: 61 data_time: 0.93s time: 519.50s eta: 3 days, 3:45:36 +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 59][DP 14][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: nan loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31495.0 tgs: 60 data_time: 0.72s time: 522.32s eta: 3 days, 4:01:35 diff --git a/20250120235238/rank6.log b/20250120235238/rank6.log new file mode 100644 index 0000000000000000000000000000000000000000..fb6f733681f1e5f6b806e70e70c694730d265b4f --- /dev/null +++ b/20250120235238/rank6.log @@ -0,0 +1,395 @@ +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.13s +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:07:53][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 149.88 seconds, peak gpu memory 13.4G +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.6GB text_tokens: 31121.0 tgs: 56 data_time: 2.32s time: 546.76s eta: 3 days, 18:03:47 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.9GB text_tokens: 31200.0 tgs: 59 data_time: 0.88s time: 523.22s eta: 3 days, 14:02:24 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.8GB text_tokens: 30861.0 tgs: 59 data_time: 0.98s time: 522.94s eta: 3 days, 13:50:56 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 31960.0 tgs: 61 data_time: 0.99s time: 520.27s eta: 3 days, 13:16:01 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 1.27s time: 520.97s eta: 3 days, 13:14:08 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.324 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31899.0 tgs: 61 data_time: 0.82s time: 520.91s eta: 3 days, 13:04:54 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32173.0 tgs: 61 data_time: 1.02s time: 523.37s eta: 3 days, 13:20:16 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32073.0 tgs: 61 data_time: 0.86s time: 520.65s eta: 3 days, 12:44:59 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32230.0 tgs: 61 data_time: 0.86s time: 520.16s eta: 3 days, 12:31:32 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.1GB text_tokens: 32211.0 tgs: 61 data_time: 0.72s time: 520.47s eta: 3 days, 12:25:56 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 31612.0 tgs: 60 data_time: 0.95s time: 524.52s eta: 3 days, 12:56:32 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31664.0 tgs: 60 data_time: 0.71s time: 520.63s eta: 3 days, 12:10:08 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31279.0 tgs: 60 data_time: 0.71s time: 519.96s eta: 3 days, 11:54:56 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 31574.0 tgs: 60 data_time: 0.88s time: 521.27s eta: 3 days, 11:58:57 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.6GB text_tokens: 31608.0 tgs: 60 data_time: 0.74s time: 524.11s eta: 3 days, 12:17:37 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32396.0 tgs: 62 data_time: 0.88s time: 520.52s eta: 3 days, 11:34:22 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32445.0 tgs: 62 data_time: 0.78s time: 518.89s eta: 3 days, 11:09:58 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32483.0 tgs: 62 data_time: 0.93s time: 522.09s eta: 3 days, 11:32:02 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.8GB text_tokens: 30947.0 tgs: 59 data_time: 0.90s time: 523.81s eta: 3 days, 11:39:52 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32390.0 tgs: 62 data_time: 0.81s time: 520.47s eta: 3 days, 10:59:08 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31866.0 tgs: 61 data_time: 0.88s time: 518.42s eta: 3 days, 10:30:52 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.9GB text_tokens: 31128.0 tgs: 59 data_time: 0.84s time: 522.82s eta: 3 days, 11:04:14 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.348 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32297.0 tgs: 61 data_time: 0.70s time: 523.52s eta: 3 days, 11:02:10 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32495.0 tgs: 62 data_time: 1.06s time: 521.01s eta: 3 days, 10:29:35 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31533.0 tgs: 60 data_time: 0.74s time: 519.08s eta: 3 days, 10:02:39 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.4GB text_tokens: 31074.0 tgs: 59 data_time: 0.78s time: 523.28s eta: 3 days, 10:33:40 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.387 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32081.0 tgs: 61 data_time: 0.88s time: 522.97s eta: 3 days, 10:22:01 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31457.0 tgs: 60 data_time: 0.82s time: 521.30s eta: 3 days, 9:57:35 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31099.0 tgs: 59 data_time: 0.74s time: 520.62s eta: 3 days, 9:42:31 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32202.0 tgs: 61 data_time: 0.76s time: 522.42s eta: 3 days, 9:50:42 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.7GB text_tokens: 30996.0 tgs: 59 data_time: 0.71s time: 523.77s eta: 3 days, 9:54:42 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31401.0 tgs: 60 data_time: 0.93s time: 520.41s eta: 3 days, 9:14:29 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.7GB text_tokens: 31506.0 tgs: 60 data_time: 0.90s time: 520.66s eta: 3 days, 9:08:10 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31849.0 tgs: 61 data_time: 0.83s time: 521.32s eta: 3 days, 9:05:39 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31503.0 tgs: 60 data_time: 0.68s time: 524.09s eta: 3 days, 9:22:47 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 32241.0 tgs: 61 data_time: 0.81s time: 520.94s eta: 3 days, 8:44:44 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31502.0 tgs: 60 data_time: 0.86s time: 520.21s eta: 3 days, 8:29:18 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.8GB text_tokens: 29886.0 tgs: 57 data_time: 1.04s time: 522.55s eta: 3 days, 8:42:17 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32309.0 tgs: 61 data_time: 0.80s time: 524.31s eta: 3 days, 8:49:54 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31388.0 tgs: 60 data_time: 0.96s time: 519.87s eta: 3 days, 8:00:06 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 31748.0 tgs: 61 data_time: 0.84s time: 520.12s eta: 3 days, 7:53:45 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.321 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 31342.0 tgs: 60 data_time: 0.70s time: 522.00s eta: 3 days, 8:02:22 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 31840.0 tgs: 60 data_time: 0.93s time: 524.33s eta: 3 days, 8:15:08 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31789.0 tgs: 61 data_time: 0.66s time: 520.56s eta: 3 days, 7:31:45 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31941.0 tgs: 61 data_time: 0.83s time: 518.55s eta: 3 days, 7:04:42 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 32.2GB text_tokens: 30569.0 tgs: 58 data_time: 0.75s time: 523.14s eta: 3 days, 7:37:58 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31307.0 tgs: 59 data_time: 0.87s time: 523.61s eta: 3 days, 7:33:32 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.201 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.6GB text_tokens: 31568.0 tgs: 60 data_time: 0.80s time: 520.20s eta: 3 days, 6:53:50 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31707.0 tgs: 60 data_time: 1.09s time: 520.04s eta: 3 days, 6:43:40 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31866.0 tgs: 60 data_time: 0.92s time: 523.74s eta: 3 days, 7:08:34 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31584.0 tgs: 60 data_time: 0.96s time: 523.82s eta: 3 days, 7:00:36 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31870.0 tgs: 61 data_time: 0.96s time: 520.22s eta: 3 days, 6:19:20 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.8GB text_tokens: 30551.0 tgs: 58 data_time: 0.60s time: 518.76s eta: 3 days, 5:57:29 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.230 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 31961.0 tgs: 61 data_time: 0.83s time: 522.67s eta: 3 days, 6:24:00 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31763.0 tgs: 60 data_time: 1.00s time: 522.77s eta: 3 days, 6:16:11 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 31917.0 tgs: 61 data_time: 1.12s time: 520.51s eta: 3 days, 5:47:13 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31804.0 tgs: 61 data_time: 0.72s time: 519.64s eta: 3 days, 5:30:47 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32278.0 tgs: 61 data_time: 0.99s time: 521.33s eta: 3 days, 5:37:15 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31927.0 tgs: 60 data_time: 0.81s time: 523.57s eta: 3 days, 5:48:31 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31369.0 tgs: 60 data_time: 0.98s time: 519.29s eta: 3 days, 5:01:40 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32214.0 tgs: 61 data_time: 0.83s time: 520.66s eta: 3 days, 5:05:14 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31816.0 tgs: 61 data_time: 0.95s time: 521.47s eta: 3 days, 5:03:41 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.9GB text_tokens: 31958.0 tgs: 60 data_time: 0.64s time: 523.93s eta: 3 days, 5:16:45 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31545.0 tgs: 60 data_time: 1.05s time: 518.83s eta: 3 days, 4:23:01 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 31565.0 tgs: 60 data_time: 0.81s time: 519.27s eta: 3 days, 4:18:14 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31753.0 tgs: 60 data_time: 0.71s time: 520.70s eta: 3 days, 4:22:08 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.5GB text_tokens: 30827.0 tgs: 58 data_time: 0.74s time: 524.24s eta: 3 days, 4:44:32 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.8GB text_tokens: 31476.0 tgs: 60 data_time: 1.07s time: 520.34s eta: 3 days, 4:01:37 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31389.0 tgs: 60 data_time: 0.44s time: 519.52s eta: 3 days, 3:45:47 +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 6][DP 1][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 32.4GB text_tokens: 30909.0 tgs: 59 data_time: 0.92s time: 522.30s eta: 3 days, 4:01:24 diff --git a/20250120235238/rank60.log b/20250120235238/rank60.log new file mode 100644 index 0000000000000000000000000000000000000000..9282ff50fcd8480f858598c2e4ba12a7518fdd89 --- /dev/null +++ b/20250120235238/rank60.log @@ -0,0 +1,395 @@ +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.12s +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 142.73 seconds, peak gpu memory 13.4G +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.9GB text_tokens: 31861.0 tgs: 58 data_time: 1.74s time: 546.72s eta: 3 days, 18:03:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.9GB text_tokens: 31345.0 tgs: 59 data_time: 0.63s time: 523.25s eta: 3 days, 14:02:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 30897.0 tgs: 59 data_time: 0.80s time: 522.87s eta: 3 days, 13:50:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.7GB text_tokens: 30727.0 tgs: 59 data_time: 0.82s time: 520.29s eta: 3 days, 13:16:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31993.0 tgs: 61 data_time: 0.64s time: 520.99s eta: 3 days, 13:14:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.0GB text_tokens: 32528.0 tgs: 62 data_time: 0.78s time: 520.91s eta: 3 days, 13:04:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.9GB text_tokens: 31985.0 tgs: 61 data_time: 0.59s time: 523.39s eta: 3 days, 13:20:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.205 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32411.0 tgs: 62 data_time: 0.81s time: 520.67s eta: 3 days, 12:45:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 32.6GB text_tokens: 30816.0 tgs: 59 data_time: 0.67s time: 520.18s eta: 3 days, 12:31:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 32133.0 tgs: 61 data_time: 1.06s time: 520.37s eta: 3 days, 12:24:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 32.8GB text_tokens: 31934.0 tgs: 60 data_time: 0.81s time: 524.54s eta: 3 days, 12:56:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 31332.0 tgs: 60 data_time: 0.82s time: 520.65s eta: 3 days, 12:10:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 32182.0 tgs: 61 data_time: 0.83s time: 519.97s eta: 3 days, 11:55:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 30812.0 tgs: 59 data_time: 0.57s time: 521.29s eta: 3 days, 11:59:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 32580.0 tgs: 62 data_time: 0.65s time: 524.13s eta: 3 days, 12:17:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.7GB text_tokens: 30706.0 tgs: 58 data_time: 0.79s time: 520.54s eta: 3 days, 11:34:33 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 31982.0 tgs: 61 data_time: 0.79s time: 518.78s eta: 3 days, 11:08:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 32212.0 tgs: 61 data_time: 0.83s time: 522.11s eta: 3 days, 11:32:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 31965.0 tgs: 61 data_time: 0.92s time: 523.83s eta: 3 days, 11:40:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 31887.0 tgs: 61 data_time: 0.82s time: 520.49s eta: 3 days, 10:59:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.218 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31686.0 tgs: 61 data_time: 0.96s time: 518.43s eta: 3 days, 10:31:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31729.0 tgs: 60 data_time: 0.71s time: 522.84s eta: 3 days, 11:04:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 31873.0 tgs: 60 data_time: 0.78s time: 523.50s eta: 3 days, 11:01:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.308 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31838.0 tgs: 61 data_time: 0.75s time: 520.96s eta: 3 days, 10:29:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31923.0 tgs: 61 data_time: 0.85s time: 519.10s eta: 3 days, 10:02:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31822.0 tgs: 60 data_time: 0.71s time: 523.29s eta: 3 days, 10:33:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.300 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.9GB text_tokens: 32115.0 tgs: 61 data_time: 1.16s time: 522.95s eta: 3 days, 10:21:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 32433.0 tgs: 62 data_time: 0.90s time: 521.32s eta: 3 days, 9:57:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.8GB text_tokens: 31453.0 tgs: 60 data_time: 0.93s time: 520.64s eta: 3 days, 9:42:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32050.0 tgs: 61 data_time: 0.90s time: 522.41s eta: 3 days, 9:50:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 32259.0 tgs: 61 data_time: 0.68s time: 523.76s eta: 3 days, 9:54:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.9GB text_tokens: 32000.0 tgs: 61 data_time: 0.73s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.6GB text_tokens: 30717.0 tgs: 58 data_time: 0.75s time: 520.68s eta: 3 days, 9:08:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31683.0 tgs: 60 data_time: 0.85s time: 521.26s eta: 3 days, 9:05:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.363 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 32425.0 tgs: 61 data_time: 1.09s time: 524.11s eta: 3 days, 9:22:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.1GB text_tokens: 31127.0 tgs: 59 data_time: 0.64s time: 520.96s eta: 3 days, 8:44:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31857.0 tgs: 61 data_time: 0.78s time: 520.17s eta: 3 days, 8:28:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.0GB text_tokens: 31766.0 tgs: 60 data_time: 0.78s time: 522.57s eta: 3 days, 8:42:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.9GB text_tokens: 32158.0 tgs: 61 data_time: 0.88s time: 524.33s eta: 3 days, 8:50:05 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 32032.0 tgs: 61 data_time: 0.83s time: 519.89s eta: 3 days, 8:00:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.8GB text_tokens: 31101.0 tgs: 59 data_time: 0.74s time: 520.09s eta: 3 days, 7:53:29 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.4GB text_tokens: 30739.0 tgs: 58 data_time: 0.91s time: 522.00s eta: 3 days, 8:02:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32313.0 tgs: 61 data_time: 0.82s time: 524.36s eta: 3 days, 8:15:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.8GB text_tokens: 30208.0 tgs: 58 data_time: 0.67s time: 520.51s eta: 3 days, 7:31:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.9GB text_tokens: 31514.0 tgs: 60 data_time: 0.93s time: 518.57s eta: 3 days, 7:04:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31887.0 tgs: 60 data_time: 0.90s time: 523.16s eta: 3 days, 7:38:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.314 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.6GB text_tokens: 31263.0 tgs: 59 data_time: 0.71s time: 523.63s eta: 3 days, 7:33:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31940.0 tgs: 61 data_time: 0.89s time: 520.16s eta: 3 days, 6:53:29 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 32230.0 tgs: 61 data_time: 1.25s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31041.0 tgs: 59 data_time: 1.02s time: 523.76s eta: 3 days, 7:08:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31939.0 tgs: 60 data_time: 0.64s time: 523.79s eta: 3 days, 7:00:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31788.0 tgs: 61 data_time: 1.17s time: 520.24s eta: 3 days, 6:19:31 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 32224.0 tgs: 62 data_time: 0.70s time: 518.78s eta: 3 days, 5:57:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.236 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 32035.0 tgs: 61 data_time: 0.72s time: 522.67s eta: 3 days, 6:24:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.3GB text_tokens: 30404.0 tgs: 58 data_time: 0.75s time: 522.72s eta: 3 days, 6:15:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.310 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 32351.0 tgs: 62 data_time: 1.08s time: 520.53s eta: 3 days, 5:47:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32058.0 tgs: 61 data_time: 0.68s time: 519.66s eta: 3 days, 5:30:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 32084.0 tgs: 61 data_time: 0.92s time: 521.30s eta: 3 days, 5:36:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31472.0 tgs: 60 data_time: 0.70s time: 523.59s eta: 3 days, 5:48:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 31503.0 tgs: 60 data_time: 0.80s time: 519.33s eta: 3 days, 5:01:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 32460.0 tgs: 62 data_time: 0.68s time: 520.64s eta: 3 days, 5:05:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31680.0 tgs: 60 data_time: 0.43s time: 521.45s eta: 3 days, 5:03:33 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31919.0 tgs: 60 data_time: 0.69s time: 523.95s eta: 3 days, 5:16:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.217 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 32.9GB text_tokens: 31721.0 tgs: 61 data_time: 0.77s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 32329.0 tgs: 62 data_time: 0.70s time: 519.20s eta: 3 days, 4:17:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.8GB text_tokens: 31333.0 tgs: 60 data_time: 0.78s time: 520.72s eta: 3 days, 4:22:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 27236.0 tgs: 51 data_time: 0.62s time: 524.26s eta: 3 days, 4:44:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.9GB text_tokens: 32171.0 tgs: 61 data_time: 0.83s time: 520.32s eta: 3 days, 4:01:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32389.0 tgs: 62 data_time: 1.00s time: 519.50s eta: 3 days, 3:45:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31521.0 tgs: 60 data_time: 0.94s time: 522.32s eta: 3 days, 4:01:35 diff --git a/20250120235238/rank62.log b/20250120235238/rank62.log new file mode 100644 index 0000000000000000000000000000000000000000..3c364bb621a64f37aeaa8096f13c35912b84d49c --- /dev/null +++ b/20250120235238/rank62.log @@ -0,0 +1,395 @@ +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.11s +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:00][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:08:01][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 142.52 seconds, peak gpu memory 13.4G +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.9GB text_tokens: 31861.0 tgs: 58 data_time: 1.74s time: 546.73s eta: 3 days, 18:03:31 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.9GB text_tokens: 31345.0 tgs: 59 data_time: 0.61s time: 523.25s eta: 3 days, 14:02:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 30897.0 tgs: 59 data_time: 0.79s time: 522.87s eta: 3 days, 13:50:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.7GB text_tokens: 30727.0 tgs: 59 data_time: 0.80s time: 520.29s eta: 3 days, 13:16:13 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31993.0 tgs: 61 data_time: 0.57s time: 520.99s eta: 3 days, 13:14:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.0GB text_tokens: 32528.0 tgs: 62 data_time: 0.69s time: 520.91s eta: 3 days, 13:04:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.270 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.9GB text_tokens: 31985.0 tgs: 61 data_time: 0.57s time: 523.39s eta: 3 days, 13:20:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.336 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32411.0 tgs: 62 data_time: 0.74s time: 520.66s eta: 3 days, 12:45:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 32.6GB text_tokens: 30816.0 tgs: 59 data_time: 0.66s time: 520.18s eta: 3 days, 12:31:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.9GB text_tokens: 32133.0 tgs: 61 data_time: 1.03s time: 520.37s eta: 3 days, 12:24:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 32.8GB text_tokens: 31934.0 tgs: 60 data_time: 0.80s time: 524.54s eta: 3 days, 12:56:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 32.9GB text_tokens: 31332.0 tgs: 60 data_time: 0.80s time: 520.65s eta: 3 days, 12:10:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 32182.0 tgs: 61 data_time: 0.81s time: 519.97s eta: 3 days, 11:55:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.9GB text_tokens: 30812.0 tgs: 59 data_time: 0.57s time: 521.29s eta: 3 days, 11:59:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 32580.0 tgs: 62 data_time: 0.65s time: 524.13s eta: 3 days, 12:17:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 32.7GB text_tokens: 30706.0 tgs: 58 data_time: 0.75s time: 520.54s eta: 3 days, 11:34:32 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 32.9GB text_tokens: 31982.0 tgs: 61 data_time: 0.77s time: 518.78s eta: 3 days, 11:08:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 32212.0 tgs: 61 data_time: 0.81s time: 522.11s eta: 3 days, 11:32:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 33.0GB text_tokens: 31965.0 tgs: 61 data_time: 0.86s time: 523.83s eta: 3 days, 11:40:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 31887.0 tgs: 61 data_time: 0.82s time: 520.49s eta: 3 days, 10:59:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 31686.0 tgs: 61 data_time: 0.96s time: 518.44s eta: 3 days, 10:31:03 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.0GB text_tokens: 31729.0 tgs: 60 data_time: 0.69s time: 522.84s eta: 3 days, 11:04:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 31873.0 tgs: 60 data_time: 0.74s time: 523.50s eta: 3 days, 11:01:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 31838.0 tgs: 61 data_time: 0.78s time: 520.96s eta: 3 days, 10:29:08 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31923.0 tgs: 61 data_time: 0.87s time: 519.10s eta: 3 days, 10:02:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 31822.0 tgs: 60 data_time: 0.71s time: 523.30s eta: 3 days, 10:33:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 32.9GB text_tokens: 32115.0 tgs: 61 data_time: 1.15s time: 522.95s eta: 3 days, 10:21:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.339 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 32433.0 tgs: 62 data_time: 0.86s time: 521.32s eta: 3 days, 9:57:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.8GB text_tokens: 31453.0 tgs: 60 data_time: 0.91s time: 520.64s eta: 3 days, 9:42:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32050.0 tgs: 61 data_time: 0.89s time: 522.41s eta: 3 days, 9:50:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.343 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 32259.0 tgs: 61 data_time: 0.63s time: 523.75s eta: 3 days, 9:54:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 32.9GB text_tokens: 32000.0 tgs: 61 data_time: 0.71s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.6GB text_tokens: 30717.0 tgs: 58 data_time: 0.72s time: 520.68s eta: 3 days, 9:08:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 31683.0 tgs: 60 data_time: 0.81s time: 521.26s eta: 3 days, 9:05:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.320 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.1GB text_tokens: 32425.0 tgs: 61 data_time: 1.04s time: 524.11s eta: 3 days, 9:22:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.226 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.1GB text_tokens: 31127.0 tgs: 59 data_time: 0.61s time: 520.96s eta: 3 days, 8:44:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.1GB text_tokens: 31857.0 tgs: 61 data_time: 0.74s time: 520.17s eta: 3 days, 8:28:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 33.0GB text_tokens: 31766.0 tgs: 60 data_time: 0.80s time: 522.57s eta: 3 days, 8:42:27 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.9GB text_tokens: 32158.0 tgs: 61 data_time: 0.84s time: 524.33s eta: 3 days, 8:50:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.0GB text_tokens: 32032.0 tgs: 61 data_time: 0.81s time: 519.89s eta: 3 days, 8:00:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 32.8GB text_tokens: 31101.0 tgs: 59 data_time: 0.72s time: 520.09s eta: 3 days, 7:53:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.4GB text_tokens: 30739.0 tgs: 58 data_time: 0.91s time: 522.00s eta: 3 days, 8:02:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.355 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.1GB text_tokens: 32313.0 tgs: 61 data_time: 0.80s time: 524.35s eta: 3 days, 8:15:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.208 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 32.8GB text_tokens: 30208.0 tgs: 58 data_time: 0.65s time: 520.51s eta: 3 days, 7:31:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.9GB text_tokens: 31514.0 tgs: 60 data_time: 0.87s time: 518.57s eta: 3 days, 7:04:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.1GB text_tokens: 31887.0 tgs: 60 data_time: 0.86s time: 523.16s eta: 3 days, 7:38:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.6GB text_tokens: 31263.0 tgs: 59 data_time: 0.70s time: 523.63s eta: 3 days, 7:33:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31940.0 tgs: 61 data_time: 0.86s time: 520.16s eta: 3 days, 6:53:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 32230.0 tgs: 61 data_time: 1.17s time: 520.06s eta: 3 days, 6:43:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.0GB text_tokens: 31041.0 tgs: 59 data_time: 0.97s time: 523.76s eta: 3 days, 7:08:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.246 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31939.0 tgs: 60 data_time: 0.61s time: 523.79s eta: 3 days, 7:00:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31788.0 tgs: 61 data_time: 1.15s time: 520.24s eta: 3 days, 6:19:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 33.1GB text_tokens: 32224.0 tgs: 62 data_time: 0.69s time: 518.78s eta: 3 days, 5:57:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.0GB text_tokens: 32035.0 tgs: 61 data_time: 0.70s time: 522.67s eta: 3 days, 6:24:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 32.3GB text_tokens: 30404.0 tgs: 58 data_time: 0.69s time: 522.72s eta: 3 days, 6:15:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.349 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 32351.0 tgs: 62 data_time: 1.03s time: 520.53s eta: 3 days, 5:47:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 32058.0 tgs: 61 data_time: 0.63s time: 519.66s eta: 3 days, 5:30:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 32.9GB text_tokens: 32084.0 tgs: 61 data_time: 0.88s time: 521.30s eta: 3 days, 5:36:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31472.0 tgs: 60 data_time: 0.65s time: 523.59s eta: 3 days, 5:48:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 33.1GB text_tokens: 31503.0 tgs: 60 data_time: 0.78s time: 519.32s eta: 3 days, 5:01:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.332 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 32460.0 tgs: 62 data_time: 0.66s time: 520.64s eta: 3 days, 5:05:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31680.0 tgs: 60 data_time: 0.47s time: 521.45s eta: 3 days, 5:03:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.288 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31919.0 tgs: 60 data_time: 0.75s time: 523.95s eta: 3 days, 5:16:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 32.9GB text_tokens: 31721.0 tgs: 61 data_time: 0.77s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 32329.0 tgs: 62 data_time: 0.70s time: 519.24s eta: 3 days, 4:17:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.8GB text_tokens: 31333.0 tgs: 60 data_time: 0.77s time: 520.72s eta: 3 days, 4:22:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.1GB text_tokens: 27236.0 tgs: 51 data_time: 0.59s time: 524.26s eta: 3 days, 4:44:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.9GB text_tokens: 32171.0 tgs: 61 data_time: 0.80s time: 520.33s eta: 3 days, 4:01:31 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.374 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 32389.0 tgs: 62 data_time: 0.95s time: 519.50s eta: 3 days, 3:45:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.293 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.1GB text_tokens: 31521.0 tgs: 60 data_time: 0.96s time: 522.32s eta: 3 days, 4:01:35 diff --git a/20250120235238/rank7.log b/20250120235238/rank7.log new file mode 100644 index 0000000000000000000000000000000000000000..bbe3652d06d519bb7a7c6d3a4ca1f573b10b6691 --- /dev/null +++ b/20250120235238/rank7.log @@ -0,0 +1,395 @@ +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:54:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.31s +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:57][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:07:58][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 145.84 seconds, peak gpu memory 13.4G +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 32.6GB text_tokens: 31121.0 tgs: 56 data_time: 2.17s time: 546.75s eta: 3 days, 18:03:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 32.9GB text_tokens: 31200.0 tgs: 59 data_time: 0.83s time: 523.23s eta: 3 days, 14:02:30 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.335 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 32.8GB text_tokens: 30861.0 tgs: 59 data_time: 0.91s time: 522.94s eta: 3 days, 13:50:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 33.1GB text_tokens: 31960.0 tgs: 61 data_time: 0.94s time: 520.28s eta: 3 days, 13:16:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.299 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 32.9GB text_tokens: 31620.0 tgs: 60 data_time: 1.13s time: 520.97s eta: 3 days, 13:14:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.347 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 33.1GB text_tokens: 31899.0 tgs: 61 data_time: 0.76s time: 520.91s eta: 3 days, 13:04:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 33.1GB text_tokens: 32173.0 tgs: 61 data_time: 1.02s time: 523.37s eta: 3 days, 13:20:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.0GB text_tokens: 32073.0 tgs: 61 data_time: 0.75s time: 520.65s eta: 3 days, 12:44:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32230.0 tgs: 61 data_time: 0.81s time: 520.16s eta: 3 days, 12:31:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.291 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 33.1GB text_tokens: 32211.0 tgs: 61 data_time: 0.70s time: 520.47s eta: 3 days, 12:25:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 33.0GB text_tokens: 31612.0 tgs: 60 data_time: 0.90s time: 524.52s eta: 3 days, 12:56:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.242 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.1GB text_tokens: 31664.0 tgs: 60 data_time: 0.66s time: 520.63s eta: 3 days, 12:10:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 33.0GB text_tokens: 31279.0 tgs: 60 data_time: 0.66s time: 519.96s eta: 3 days, 11:54:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 32.8GB text_tokens: 31574.0 tgs: 60 data_time: 0.85s time: 521.27s eta: 3 days, 11:58:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 32.6GB text_tokens: 31608.0 tgs: 60 data_time: 0.71s time: 524.11s eta: 3 days, 12:17:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.1GB text_tokens: 32396.0 tgs: 62 data_time: 0.84s time: 520.52s eta: 3 days, 11:34:23 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.286 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32445.0 tgs: 62 data_time: 0.76s time: 518.89s eta: 3 days, 11:09:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.1GB text_tokens: 32483.0 tgs: 62 data_time: 0.85s time: 522.09s eta: 3 days, 11:32:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.8GB text_tokens: 30947.0 tgs: 59 data_time: 0.87s time: 523.81s eta: 3 days, 11:39:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 33.0GB text_tokens: 32390.0 tgs: 62 data_time: 0.75s time: 520.47s eta: 3 days, 10:59:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 32.9GB text_tokens: 31866.0 tgs: 61 data_time: 0.85s time: 518.42s eta: 3 days, 10:30:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 32.9GB text_tokens: 31128.0 tgs: 59 data_time: 0.81s time: 522.82s eta: 3 days, 11:04:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.0GB text_tokens: 32297.0 tgs: 61 data_time: 0.67s time: 523.52s eta: 3 days, 11:02:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.305 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32495.0 tgs: 62 data_time: 1.01s time: 521.01s eta: 3 days, 10:29:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.9GB text_tokens: 31533.0 tgs: 60 data_time: 0.70s time: 519.08s eta: 3 days, 10:02:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 32.4GB text_tokens: 31074.0 tgs: 59 data_time: 0.73s time: 523.28s eta: 3 days, 10:33:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.332 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32081.0 tgs: 61 data_time: 0.82s time: 522.97s eta: 3 days, 10:22:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 31457.0 tgs: 60 data_time: 0.78s time: 521.30s eta: 3 days, 9:57:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.235 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31099.0 tgs: 59 data_time: 0.69s time: 520.62s eta: 3 days, 9:42:30 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 33.0GB text_tokens: 32202.0 tgs: 61 data_time: 0.73s time: 522.42s eta: 3 days, 9:50:44 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 32.7GB text_tokens: 30996.0 tgs: 59 data_time: 0.69s time: 523.77s eta: 3 days, 9:54:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.0GB text_tokens: 31401.0 tgs: 60 data_time: 0.93s time: 520.41s eta: 3 days, 9:14:30 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 32.7GB text_tokens: 31506.0 tgs: 60 data_time: 0.83s time: 520.66s eta: 3 days, 9:08:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.0GB text_tokens: 31849.0 tgs: 61 data_time: 0.79s time: 521.32s eta: 3 days, 9:05:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 33.0GB text_tokens: 31503.0 tgs: 60 data_time: 0.63s time: 524.09s eta: 3 days, 9:22:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.326 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 32241.0 tgs: 61 data_time: 0.75s time: 520.94s eta: 3 days, 8:44:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 31502.0 tgs: 60 data_time: 0.83s time: 520.21s eta: 3 days, 8:29:18 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.8GB text_tokens: 29886.0 tgs: 57 data_time: 0.92s time: 522.55s eta: 3 days, 8:42:18 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 33.1GB text_tokens: 32309.0 tgs: 61 data_time: 0.74s time: 524.32s eta: 3 days, 8:49:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 32.8GB text_tokens: 31388.0 tgs: 60 data_time: 0.84s time: 519.87s eta: 3 days, 8:00:07 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.362 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 31748.0 tgs: 61 data_time: 0.78s time: 520.12s eta: 3 days, 7:53:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 33.1GB text_tokens: 31342.0 tgs: 60 data_time: 0.67s time: 522.00s eta: 3 days, 8:02:23 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 33.0GB text_tokens: 31840.0 tgs: 60 data_time: 0.84s time: 524.34s eta: 3 days, 8:15:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31789.0 tgs: 61 data_time: 0.64s time: 520.56s eta: 3 days, 7:31:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 33.0GB text_tokens: 31941.0 tgs: 61 data_time: 0.79s time: 518.55s eta: 3 days, 7:04:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 32.2GB text_tokens: 30569.0 tgs: 58 data_time: 0.73s time: 523.14s eta: 3 days, 7:37:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.228 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 33.1GB text_tokens: 31307.0 tgs: 59 data_time: 0.81s time: 523.61s eta: 3 days, 7:33:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.6GB text_tokens: 31568.0 tgs: 60 data_time: 0.76s time: 520.20s eta: 3 days, 6:53:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.361 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.1GB text_tokens: 31707.0 tgs: 60 data_time: 1.05s time: 520.04s eta: 3 days, 6:43:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 33.1GB text_tokens: 31866.0 tgs: 60 data_time: 0.87s time: 523.74s eta: 3 days, 7:08:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.1GB text_tokens: 31584.0 tgs: 60 data_time: 0.88s time: 523.83s eta: 3 days, 7:00:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.0GB text_tokens: 31870.0 tgs: 61 data_time: 0.93s time: 520.22s eta: 3 days, 6:19:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.247 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.8GB text_tokens: 30551.0 tgs: 58 data_time: 0.59s time: 518.76s eta: 3 days, 5:57:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 33.1GB text_tokens: 31961.0 tgs: 61 data_time: 0.79s time: 522.67s eta: 3 days, 6:24:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.1GB text_tokens: 31763.0 tgs: 60 data_time: 0.97s time: 522.77s eta: 3 days, 6:16:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.277 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.1GB text_tokens: 31917.0 tgs: 61 data_time: 1.02s time: 520.51s eta: 3 days, 5:47:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:26:33][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:26:33][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.252 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 33.1GB text_tokens: 31804.0 tgs: 61 data_time: 0.69s time: 519.64s eta: 3 days, 5:30:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.256 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.1GB text_tokens: 32278.0 tgs: 61 data_time: 0.97s time: 521.34s eta: 3 days, 5:37:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 32.9GB text_tokens: 31927.0 tgs: 60 data_time: 0.77s time: 523.57s eta: 3 days, 5:48:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31369.0 tgs: 60 data_time: 0.93s time: 519.29s eta: 3 days, 5:01:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.0GB text_tokens: 32214.0 tgs: 61 data_time: 0.79s time: 520.67s eta: 3 days, 5:05:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.261 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 33.0GB text_tokens: 31816.0 tgs: 61 data_time: 0.89s time: 521.47s eta: 3 days, 5:03:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 32.9GB text_tokens: 31958.0 tgs: 60 data_time: 0.60s time: 523.93s eta: 3 days, 5:16:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.323 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.1GB text_tokens: 31545.0 tgs: 60 data_time: 0.98s time: 518.84s eta: 3 days, 4:23:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.1GB text_tokens: 31565.0 tgs: 60 data_time: 0.74s time: 519.27s eta: 3 days, 4:18:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 33.0GB text_tokens: 31753.0 tgs: 60 data_time: 0.68s time: 520.70s eta: 3 days, 4:22:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.290 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 32.5GB text_tokens: 30827.0 tgs: 58 data_time: 0.71s time: 524.24s eta: 3 days, 4:44:32 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 32.8GB text_tokens: 31476.0 tgs: 60 data_time: 1.04s time: 520.34s eta: 3 days, 4:01:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 33.1GB text_tokens: 31389.0 tgs: 60 data_time: 0.42s time: 519.52s eta: 3 days, 3:45:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 32.4GB text_tokens: 30909.0 tgs: 59 data_time: 0.88s time: 522.30s eta: 3 days, 4:01:26 diff --git a/20250120235238/rank8.log b/20250120235238/rank8.log new file mode 100644 index 0000000000000000000000000000000000000000..3951c51ef6a808bafb7dfbf64ef61df958ab10c1 --- /dev/null +++ b/20250120235238/rank8.log @@ -0,0 +1,395 @@ +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.13s +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:07:56][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.83 seconds, peak gpu memory 13.4G +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31696.0 tgs: 57 data_time: 1.91s time: 547.89s eta: 3 days, 18:14:58 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.262 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31732.0 tgs: 60 data_time: 1.04s time: 523.24s eta: 3 days, 14:02:38 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 32529.0 tgs: 62 data_time: 0.90s time: 522.88s eta: 3 days, 13:50:21 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.9GB text_tokens: 31543.0 tgs: 60 data_time: 0.87s time: 520.29s eta: 3 days, 13:16:10 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31681.0 tgs: 60 data_time: 0.97s time: 520.98s eta: 3 days, 13:14:18 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.244 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.4GB text_tokens: 30335.0 tgs: 58 data_time: 0.83s time: 520.90s eta: 3 days, 13:04:46 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.214 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.8GB text_tokens: 30779.0 tgs: 58 data_time: 0.69s time: 523.39s eta: 3 days, 13:20:27 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.436 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31645.0 tgs: 60 data_time: 0.92s time: 520.66s eta: 3 days, 12:45:06 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.255 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32125.0 tgs: 61 data_time: 0.68s time: 520.18s eta: 3 days, 12:31:44 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.7GB text_tokens: 31629.0 tgs: 60 data_time: 0.92s time: 520.42s eta: 3 days, 12:25:22 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.281 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 32.9GB text_tokens: 31480.0 tgs: 60 data_time: 0.93s time: 524.53s eta: 3 days, 12:56:42 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.284 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 32172.0 tgs: 61 data_time: 0.73s time: 520.65s eta: 3 days, 12:10:16 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.339 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.7GB text_tokens: 31315.0 tgs: 60 data_time: 1.22s time: 519.93s eta: 3 days, 11:54:40 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.272 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32182.0 tgs: 61 data_time: 0.65s time: 521.29s eta: 3 days, 11:59:06 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 31862.0 tgs: 60 data_time: 0.74s time: 524.12s eta: 3 days, 12:17:48 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.334 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.0GB text_tokens: 31743.0 tgs: 60 data_time: 0.90s time: 520.54s eta: 3 days, 11:34:30 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32304.0 tgs: 62 data_time: 0.86s time: 518.82s eta: 3 days, 11:09:21 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31895.0 tgs: 61 data_time: 0.88s time: 522.11s eta: 3 days, 11:32:13 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.250 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.6GB text_tokens: 30695.0 tgs: 58 data_time: 0.88s time: 523.83s eta: 3 days, 11:40:02 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.2GB text_tokens: 30721.0 tgs: 59 data_time: 0.98s time: 520.46s eta: 3 days, 10:59:03 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.241 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 29162.0 tgs: 56 data_time: 0.60s time: 518.43s eta: 3 days, 10:31:01 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 31808.0 tgs: 60 data_time: 0.83s time: 522.84s eta: 3 days, 11:04:23 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.325 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 32435.0 tgs: 61 data_time: 1.02s time: 523.54s eta: 3 days, 11:02:20 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.308 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32461.0 tgs: 62 data_time: 0.79s time: 520.97s eta: 3 days, 10:29:10 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.264 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.8GB text_tokens: 31514.0 tgs: 60 data_time: 1.00s time: 519.10s eta: 3 days, 10:02:50 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 30176.0 tgs: 57 data_time: 0.58s time: 523.29s eta: 3 days, 10:33:47 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32145.0 tgs: 61 data_time: 1.00s time: 522.93s eta: 3 days, 10:21:42 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 32029.0 tgs: 61 data_time: 0.57s time: 521.32s eta: 3 days, 9:57:44 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.349 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31383.0 tgs: 60 data_time: 0.64s time: 520.64s eta: 3 days, 9:42:39 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.4GB text_tokens: 30895.0 tgs: 59 data_time: 0.77s time: 522.41s eta: 3 days, 9:50:39 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.317 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 31598.0 tgs: 60 data_time: 1.05s time: 523.75s eta: 3 days, 9:54:32 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.369 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 32406.0 tgs: 62 data_time: 0.78s time: 520.43s eta: 3 days, 9:14:40 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.313 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 32322.0 tgs: 62 data_time: 0.84s time: 520.68s eta: 3 days, 9:08:19 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.361 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 30969.0 tgs: 59 data_time: 0.83s time: 521.30s eta: 3 days, 9:05:28 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.220 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.8GB text_tokens: 31334.0 tgs: 59 data_time: 0.90s time: 524.11s eta: 3 days, 9:22:56 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31952.0 tgs: 61 data_time: 0.84s time: 520.96s eta: 3 days, 8:44:54 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 32127.0 tgs: 61 data_time: 0.92s time: 520.15s eta: 3 days, 8:28:43 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.265 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.9GB text_tokens: 31858.0 tgs: 60 data_time: 0.83s time: 522.56s eta: 3 days, 8:42:24 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.8GB text_tokens: 31386.0 tgs: 59 data_time: 0.85s time: 524.33s eta: 3 days, 8:50:03 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 32250.0 tgs: 62 data_time: 0.71s time: 519.88s eta: 3 days, 8:00:15 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.340 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32507.0 tgs: 62 data_time: 1.07s time: 520.11s eta: 3 days, 7:53:41 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.9GB text_tokens: 31757.0 tgs: 60 data_time: 0.68s time: 522.00s eta: 3 days, 8:02:22 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.8GB text_tokens: 31888.0 tgs: 60 data_time: 0.98s time: 524.35s eta: 3 days, 8:15:17 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31891.0 tgs: 61 data_time: 0.81s time: 520.50s eta: 3 days, 7:31:17 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.245 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.9GB text_tokens: 31410.0 tgs: 60 data_time: 0.68s time: 518.56s eta: 3 days, 7:04:51 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.238 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31763.0 tgs: 60 data_time: 0.96s time: 523.15s eta: 3 days, 7:38:07 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.234 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.6GB text_tokens: 30226.0 tgs: 57 data_time: 0.90s time: 523.62s eta: 3 days, 7:33:41 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.316 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31914.0 tgs: 61 data_time: 0.85s time: 520.18s eta: 3 days, 6:53:35 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.259 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 31209.0 tgs: 60 data_time: 0.49s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.273 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.9GB text_tokens: 31712.0 tgs: 60 data_time: 0.75s time: 523.76s eta: 3 days, 7:08:45 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31541.0 tgs: 60 data_time: 0.80s time: 523.78s eta: 3 days, 7:00:11 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.350 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32580.0 tgs: 62 data_time: 0.80s time: 520.24s eta: 3 days, 6:19:29 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.211 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.7GB text_tokens: 31793.0 tgs: 61 data_time: 0.57s time: 518.77s eta: 3 days, 5:57:36 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.303 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.8GB text_tokens: 31726.0 tgs: 60 data_time: 0.66s time: 522.69s eta: 3 days, 6:24:10 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 32291.0 tgs: 61 data_time: 0.87s time: 522.73s eta: 3 days, 6:15:51 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.216 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.0GB text_tokens: 31197.0 tgs: 59 data_time: 0.78s time: 520.53s eta: 3 days, 5:47:23 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.271 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.7GB text_tokens: 31650.0 tgs: 60 data_time: 0.90s time: 519.66s eta: 3 days, 5:30:55 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.0GB text_tokens: 31693.0 tgs: 60 data_time: 0.79s time: 521.29s eta: 3 days, 5:36:53 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.362 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.0GB text_tokens: 30935.0 tgs: 59 data_time: 0.82s time: 523.59s eta: 3 days, 5:48:40 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.257 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31373.0 tgs: 60 data_time: 0.75s time: 519.32s eta: 3 days, 5:01:58 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.354 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31466.0 tgs: 60 data_time: 0.73s time: 520.65s eta: 3 days, 5:05:06 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31582.0 tgs: 60 data_time: 0.76s time: 521.45s eta: 3 days, 5:03:33 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.292 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31379.0 tgs: 59 data_time: 0.78s time: 523.94s eta: 3 days, 5:16:53 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.258 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 31834.0 tgs: 61 data_time: 0.50s time: 518.85s eta: 3 days, 4:23:10 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.296 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31863.0 tgs: 61 data_time: 0.65s time: 519.24s eta: 3 days, 4:17:57 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.278 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31055.0 tgs: 59 data_time: 0.67s time: 520.71s eta: 3 days, 4:22:16 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 31953.0 tgs: 60 data_time: 0.76s time: 524.25s eta: 3 days, 4:44:41 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.225 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32157.0 tgs: 61 data_time: 0.60s time: 520.33s eta: 3 days, 4:01:32 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 32.7GB text_tokens: 31301.0 tgs: 60 data_time: 0.74s time: 519.50s eta: 3 days, 3:45:34 +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 8][DP 2][SP 0][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.337 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 31423.0 tgs: 60 data_time: 0.47s time: 522.32s eta: 3 days, 4:01:33 diff --git a/20250120235238/rank9.log b/20250120235238/rank9.log new file mode 100644 index 0000000000000000000000000000000000000000..dbcf9e21d10c83b67be388a1005ae67a675c5631 --- /dev/null +++ b/20250120235238/rank9.log @@ -0,0 +1,395 @@ +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:52:42][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250120235238', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:52:42][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:53:37][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:54:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:55:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:56:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:57:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:58:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-20 23:59:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:00:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:00:05][INFO] [Dataset & Dataloader] Cost 443.14s +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:07:55][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:10:23][SUCCESS] [Parallelize LLM] Elapsed time 147.88 seconds, peak gpu memory 13.4G +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:10:24][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:19:46][WARNING] [Step 0] The grad norm is NaN or Inf, skip this step. Skipped 1 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:19:46][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 1 max_memory: 33.1GB text_tokens: 31696.0 tgs: 57 data_time: 1.90s time: 547.75s eta: 3 days, 18:13:35 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:28:29][WARNING] [Step 1] The grad norm is NaN or Inf, skip this step. Skipped 2 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:28:29][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.285 loss(reduced): nan grad_norm: nan if_nan_skip: 2 max_memory: 33.1GB text_tokens: 31732.0 tgs: 60 data_time: 1.06s time: 523.21s eta: 3 days, 14:02:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:37:12][WARNING] [Step 2] The grad norm is NaN or Inf, skip this step. Skipped 3 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:37:12][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.217 loss(reduced): nan grad_norm: nan if_nan_skip: 3 max_memory: 33.1GB text_tokens: 32529.0 tgs: 62 data_time: 0.89s time: 522.88s eta: 3 days, 13:50:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:45:52][WARNING] [Step 3] The grad norm is NaN or Inf, skip this step. Skipped 4 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:45:52][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.302 loss(reduced): nan grad_norm: nan if_nan_skip: 4 max_memory: 32.9GB text_tokens: 31543.0 tgs: 60 data_time: 0.86s time: 520.29s eta: 3 days, 13:16:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:54:33][WARNING] [Step 4] The grad norm is NaN or Inf, skip this step. Skipped 5 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 00:54:33][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.304 loss(reduced): nan grad_norm: nan if_nan_skip: 5 max_memory: 33.0GB text_tokens: 31681.0 tgs: 60 data_time: 0.94s time: 520.98s eta: 3 days, 13:14:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:03:14][WARNING] [Step 5] The grad norm is NaN or Inf, skip this step. Skipped 6 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:03:14][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 6 max_memory: 32.4GB text_tokens: 30335.0 tgs: 58 data_time: 0.82s time: 520.89s eta: 3 days, 13:04:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:11:58][WARNING] [Step 6] The grad norm is NaN or Inf, skip this step. Skipped 7 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:11:58][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.198 loss(reduced): nan grad_norm: nan if_nan_skip: 7 max_memory: 32.8GB text_tokens: 30779.0 tgs: 58 data_time: 0.69s time: 523.38s eta: 3 days, 13:20:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:20:38][WARNING] [Step 7] The grad norm is NaN or Inf, skip this step. Skipped 8 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:20:38][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.231 loss(reduced): nan grad_norm: nan if_nan_skip: 8 max_memory: 33.1GB text_tokens: 31645.0 tgs: 60 data_time: 0.91s time: 520.66s eta: 3 days, 12:45:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:29:18][WARNING] [Step 8] The grad norm is NaN or Inf, skip this step. Skipped 9 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:29:18][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 9 max_memory: 33.1GB text_tokens: 32125.0 tgs: 61 data_time: 0.67s time: 520.18s eta: 3 days, 12:31:44 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:37:59][WARNING] [Step 9] The grad norm is NaN or Inf, skip this step. Skipped 10 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:37:59][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.298 loss(reduced): nan grad_norm: nan if_nan_skip: 10 max_memory: 32.7GB text_tokens: 31629.0 tgs: 60 data_time: 0.91s time: 520.41s eta: 3 days, 12:25:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:46:43][WARNING] [Step 10] The grad norm is NaN or Inf, skip this step. Skipped 11 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:46:43][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.306 loss(reduced): nan grad_norm: nan if_nan_skip: 11 max_memory: 32.9GB text_tokens: 31480.0 tgs: 60 data_time: 0.97s time: 524.53s eta: 3 days, 12:56:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:55:24][WARNING] [Step 11] The grad norm is NaN or Inf, skip this step. Skipped 12 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 01:55:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.327 loss(reduced): nan grad_norm: nan if_nan_skip: 12 max_memory: 33.0GB text_tokens: 32172.0 tgs: 61 data_time: 0.71s time: 520.65s eta: 3 days, 12:10:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:04:04][WARNING] [Step 12] The grad norm is NaN or Inf, skip this step. Skipped 13 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:04:04][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.297 loss(reduced): nan grad_norm: nan if_nan_skip: 13 max_memory: 32.7GB text_tokens: 31315.0 tgs: 60 data_time: 1.22s time: 519.93s eta: 3 days, 11:54:39 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:12:45][WARNING] [Step 13] The grad norm is NaN or Inf, skip this step. Skipped 14 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:12:45][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 14 max_memory: 33.1GB text_tokens: 32182.0 tgs: 61 data_time: 0.64s time: 521.29s eta: 3 days, 11:59:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:21:29][WARNING] [Step 14] The grad norm is NaN or Inf, skip this step. Skipped 15 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:21:29][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.336 loss(reduced): nan grad_norm: nan if_nan_skip: 15 max_memory: 33.1GB text_tokens: 31862.0 tgs: 60 data_time: 0.75s time: 524.12s eta: 3 days, 12:17:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:30:10][WARNING] [Step 15] The grad norm is NaN or Inf, skip this step. Skipped 16 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:30:10][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 16 max_memory: 33.0GB text_tokens: 31743.0 tgs: 60 data_time: 0.89s time: 520.54s eta: 3 days, 11:34:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:38:49][WARNING] [Step 16] The grad norm is NaN or Inf, skip this step. Skipped 17 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:38:49][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 17 max_memory: 33.1GB text_tokens: 32304.0 tgs: 62 data_time: 0.86s time: 518.82s eta: 3 days, 11:09:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:47:31][WARNING] [Step 17] The grad norm is NaN or Inf, skip this step. Skipped 18 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:47:31][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 18 max_memory: 33.0GB text_tokens: 31895.0 tgs: 61 data_time: 0.86s time: 522.11s eta: 3 days, 11:32:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:56:15][WARNING] [Step 18] The grad norm is NaN or Inf, skip this step. Skipped 19 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 02:56:15][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 19 max_memory: 32.6GB text_tokens: 30695.0 tgs: 58 data_time: 0.87s time: 523.83s eta: 3 days, 11:39:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:04:55][WARNING] [Step 19] The grad norm is NaN or Inf, skip this step. Skipped 20 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:04:55][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 20 max_memory: 32.2GB text_tokens: 30721.0 tgs: 59 data_time: 0.97s time: 520.46s eta: 3 days, 10:59:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:13:34][WARNING] [Step 20] The grad norm is NaN or Inf, skip this step. Skipped 21 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:13:34][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.229 loss(reduced): nan grad_norm: nan if_nan_skip: 21 max_memory: 33.0GB text_tokens: 29162.0 tgs: 56 data_time: 0.59s time: 518.43s eta: 3 days, 10:31:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:22:16][WARNING] [Step 21] The grad norm is NaN or Inf, skip this step. Skipped 22 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:22:16][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.315 loss(reduced): nan grad_norm: nan if_nan_skip: 22 max_memory: 33.1GB text_tokens: 31808.0 tgs: 60 data_time: 0.80s time: 522.84s eta: 3 days, 11:04:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:31:00][WARNING] [Step 22] The grad norm is NaN or Inf, skip this step. Skipped 23 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:31:00][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 23 max_memory: 33.1GB text_tokens: 32435.0 tgs: 61 data_time: 1.01s time: 523.54s eta: 3 days, 11:02:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:39:41][WARNING] [Step 23] The grad norm is NaN or Inf, skip this step. Skipped 24 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:39:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.269 loss(reduced): nan grad_norm: nan if_nan_skip: 24 max_memory: 33.1GB text_tokens: 32461.0 tgs: 62 data_time: 0.79s time: 520.96s eta: 3 days, 10:29:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:48:20][WARNING] [Step 24] The grad norm is NaN or Inf, skip this step. Skipped 25 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:48:20][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.266 loss(reduced): nan grad_norm: nan if_nan_skip: 25 max_memory: 32.8GB text_tokens: 31514.0 tgs: 60 data_time: 0.98s time: 519.10s eta: 3 days, 10:02:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:57:03][WARNING] [Step 25] The grad norm is NaN or Inf, skip this step. Skipped 26 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 03:57:03][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.253 loss(reduced): nan grad_norm: nan if_nan_skip: 26 max_memory: 33.0GB text_tokens: 30176.0 tgs: 57 data_time: 0.56s time: 523.29s eta: 3 days, 10:33:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:05:46][WARNING] [Step 26] The grad norm is NaN or Inf, skip this step. Skipped 27 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:05:46][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 27 max_memory: 33.1GB text_tokens: 32145.0 tgs: 61 data_time: 0.99s time: 522.93s eta: 3 days, 10:21:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:14:28][WARNING] [Step 27] The grad norm is NaN or Inf, skip this step. Skipped 28 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:14:28][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 28 max_memory: 33.1GB text_tokens: 32029.0 tgs: 61 data_time: 0.58s time: 521.31s eta: 3 days, 9:57:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:23:08][WARNING] [Step 28] The grad norm is NaN or Inf, skip this step. Skipped 29 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:23:08][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 29 max_memory: 32.9GB text_tokens: 31383.0 tgs: 60 data_time: 0.64s time: 520.64s eta: 3 days, 9:42:39 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:31:51][WARNING] [Step 29] The grad norm is NaN or Inf, skip this step. Skipped 30 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:31:51][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.295 loss(reduced): nan grad_norm: nan if_nan_skip: 30 max_memory: 32.4GB text_tokens: 30895.0 tgs: 59 data_time: 0.77s time: 522.41s eta: 3 days, 9:50:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:40:34][WARNING] [Step 30] The grad norm is NaN or Inf, skip this step. Skipped 31 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:40:34][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.294 loss(reduced): nan grad_norm: nan if_nan_skip: 31 max_memory: 33.0GB text_tokens: 31598.0 tgs: 60 data_time: 1.05s time: 523.75s eta: 3 days, 9:54:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:49:15][WARNING] [Step 31] The grad norm is NaN or Inf, skip this step. Skipped 32 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:49:15][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.239 loss(reduced): nan grad_norm: nan if_nan_skip: 32 max_memory: 33.1GB text_tokens: 32406.0 tgs: 62 data_time: 0.78s time: 520.43s eta: 3 days, 9:14:39 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:57:55][WARNING] [Step 32] The grad norm is NaN or Inf, skip this step. Skipped 33 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 04:57:55][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 33 max_memory: 33.0GB text_tokens: 32322.0 tgs: 62 data_time: 0.83s time: 520.68s eta: 3 days, 9:08:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:06:37][WARNING] [Step 33] The grad norm is NaN or Inf, skip this step. Skipped 34 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:06:37][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 34 max_memory: 33.1GB text_tokens: 30969.0 tgs: 59 data_time: 0.86s time: 521.30s eta: 3 days, 9:05:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:15:21][WARNING] [Step 34] The grad norm is NaN or Inf, skip this step. Skipped 35 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:15:21][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 35 max_memory: 32.8GB text_tokens: 31334.0 tgs: 59 data_time: 0.88s time: 524.11s eta: 3 days, 9:22:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:24:02][WARNING] [Step 35] The grad norm is NaN or Inf, skip this step. Skipped 36 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:24:02][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.226 loss(reduced): nan grad_norm: nan if_nan_skip: 36 max_memory: 33.0GB text_tokens: 31952.0 tgs: 61 data_time: 0.82s time: 520.96s eta: 3 days, 8:44:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:32:42][WARNING] [Step 36] The grad norm is NaN or Inf, skip this step. Skipped 37 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:32:42][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.224 loss(reduced): nan grad_norm: nan if_nan_skip: 37 max_memory: 33.0GB text_tokens: 32127.0 tgs: 61 data_time: 0.90s time: 520.15s eta: 3 days, 8:28:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:41:25][WARNING] [Step 37] The grad norm is NaN or Inf, skip this step. Skipped 38 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:41:25][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.227 loss(reduced): nan grad_norm: nan if_nan_skip: 38 max_memory: 32.9GB text_tokens: 31858.0 tgs: 60 data_time: 0.83s time: 522.56s eta: 3 days, 8:42:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:50:09][WARNING] [Step 38] The grad norm is NaN or Inf, skip this step. Skipped 39 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:50:09][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.275 loss(reduced): nan grad_norm: nan if_nan_skip: 39 max_memory: 32.8GB text_tokens: 31386.0 tgs: 59 data_time: 0.84s time: 524.33s eta: 3 days, 8:50:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:58:49][WARNING] [Step 39] The grad norm is NaN or Inf, skip this step. Skipped 40 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 05:58:49][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 40 max_memory: 33.1GB text_tokens: 32250.0 tgs: 62 data_time: 0.71s time: 519.88s eta: 3 days, 8:00:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:07:29][WARNING] [Step 40] The grad norm is NaN or Inf, skip this step. Skipped 41 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:07:29][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.307 loss(reduced): nan grad_norm: nan if_nan_skip: 41 max_memory: 33.1GB text_tokens: 32507.0 tgs: 62 data_time: 1.06s time: 520.11s eta: 3 days, 7:53:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:16:11][WARNING] [Step 41] The grad norm is NaN or Inf, skip this step. Skipped 42 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:16:11][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 42 max_memory: 32.9GB text_tokens: 31757.0 tgs: 60 data_time: 0.68s time: 522.00s eta: 3 days, 8:02:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:24:55][WARNING] [Step 42] The grad norm is NaN or Inf, skip this step. Skipped 43 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:24:55][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.312 loss(reduced): nan grad_norm: nan if_nan_skip: 43 max_memory: 32.8GB text_tokens: 31888.0 tgs: 60 data_time: 0.97s time: 524.35s eta: 3 days, 8:15:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:33:36][WARNING] [Step 43] The grad norm is NaN or Inf, skip this step. Skipped 44 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:33:36][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.209 loss(reduced): nan grad_norm: nan if_nan_skip: 44 max_memory: 33.1GB text_tokens: 31891.0 tgs: 61 data_time: 0.79s time: 520.50s eta: 3 days, 7:31:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:42:14][WARNING] [Step 44] The grad norm is NaN or Inf, skip this step. Skipped 45 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:42:14][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.268 loss(reduced): nan grad_norm: nan if_nan_skip: 45 max_memory: 32.9GB text_tokens: 31410.0 tgs: 60 data_time: 0.68s time: 518.56s eta: 3 days, 7:04:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:50:57][WARNING] [Step 45] The grad norm is NaN or Inf, skip this step. Skipped 46 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:50:57][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.249 loss(reduced): nan grad_norm: nan if_nan_skip: 46 max_memory: 33.0GB text_tokens: 31763.0 tgs: 60 data_time: 0.95s time: 523.15s eta: 3 days, 7:38:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:59:41][WARNING] [Step 46] The grad norm is NaN or Inf, skip this step. Skipped 47 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 06:59:41][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.251 loss(reduced): nan grad_norm: nan if_nan_skip: 47 max_memory: 32.6GB text_tokens: 30226.0 tgs: 57 data_time: 0.90s time: 523.62s eta: 3 days, 7:33:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:08:21][WARNING] [Step 47] The grad norm is NaN or Inf, skip this step. Skipped 48 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:08:21][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.254 loss(reduced): nan grad_norm: nan if_nan_skip: 48 max_memory: 32.9GB text_tokens: 31914.0 tgs: 61 data_time: 0.85s time: 520.17s eta: 3 days, 6:53:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:17:01][WARNING] [Step 48] The grad norm is NaN or Inf, skip this step. Skipped 49 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:17:01][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.282 loss(reduced): nan grad_norm: nan if_nan_skip: 49 max_memory: 33.0GB text_tokens: 31209.0 tgs: 60 data_time: 0.48s time: 520.06s eta: 3 days, 6:43:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:25:45][WARNING] [Step 49] The grad norm is NaN or Inf, skip this step. Skipped 50 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:25:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.301 loss(reduced): nan grad_norm: nan if_nan_skip: 50 max_memory: 32.9GB text_tokens: 31712.0 tgs: 60 data_time: 0.75s time: 523.75s eta: 3 days, 7:08:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:34:29][WARNING] [Step 50] The grad norm is NaN or Inf, skip this step. Skipped 51 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:34:29][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 51 max_memory: 33.0GB text_tokens: 31541.0 tgs: 60 data_time: 0.79s time: 523.78s eta: 3 days, 7:00:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:43:09][WARNING] [Step 51] The grad norm is NaN or Inf, skip this step. Skipped 52 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:43:09][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.263 loss(reduced): nan grad_norm: nan if_nan_skip: 52 max_memory: 33.1GB text_tokens: 32580.0 tgs: 62 data_time: 0.80s time: 520.24s eta: 3 days, 6:19:29 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:51:48][WARNING] [Step 52] The grad norm is NaN or Inf, skip this step. Skipped 53 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 07:51:48][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.283 loss(reduced): nan grad_norm: nan if_nan_skip: 53 max_memory: 32.7GB text_tokens: 31793.0 tgs: 61 data_time: 0.57s time: 518.78s eta: 3 days, 5:57:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:00:31][WARNING] [Step 53] The grad norm is NaN or Inf, skip this step. Skipped 54 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:00:31][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.276 loss(reduced): nan grad_norm: nan if_nan_skip: 54 max_memory: 32.8GB text_tokens: 31726.0 tgs: 60 data_time: 0.65s time: 522.68s eta: 3 days, 6:24:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:09:13][WARNING] [Step 54] The grad norm is NaN or Inf, skip this step. Skipped 55 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:09:13][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.233 loss(reduced): nan grad_norm: nan if_nan_skip: 55 max_memory: 33.0GB text_tokens: 32291.0 tgs: 61 data_time: 0.86s time: 522.73s eta: 3 days, 6:15:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:17:54][WARNING] [Step 55] The grad norm is NaN or Inf, skip this step. Skipped 56 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:17:54][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.267 loss(reduced): nan grad_norm: nan if_nan_skip: 56 max_memory: 33.0GB text_tokens: 31197.0 tgs: 59 data_time: 0.79s time: 520.52s eta: 3 days, 5:47:22 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:26:34][WARNING] [Step 56] The grad norm is NaN or Inf, skip this step. Skipped 57 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:26:34][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.232 loss(reduced): nan grad_norm: nan if_nan_skip: 57 max_memory: 32.7GB text_tokens: 31650.0 tgs: 60 data_time: 0.89s time: 519.66s eta: 3 days, 5:30:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:35:15][WARNING] [Step 57] The grad norm is NaN or Inf, skip this step. Skipped 58 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:35:15][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.223 loss(reduced): nan grad_norm: nan if_nan_skip: 58 max_memory: 33.0GB text_tokens: 31693.0 tgs: 60 data_time: 0.78s time: 521.29s eta: 3 days, 5:36:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:43:58][WARNING] [Step 58] The grad norm is NaN or Inf, skip this step. Skipped 59 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:43:58][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.260 loss(reduced): nan grad_norm: nan if_nan_skip: 59 max_memory: 33.0GB text_tokens: 30935.0 tgs: 59 data_time: 0.81s time: 523.59s eta: 3 days, 5:48:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:52:38][WARNING] [Step 59] The grad norm is NaN or Inf, skip this step. Skipped 60 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 08:52:38][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.287 loss(reduced): nan grad_norm: nan if_nan_skip: 60 max_memory: 32.9GB text_tokens: 31373.0 tgs: 60 data_time: 0.74s time: 519.32s eta: 3 days, 5:01:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:01:18][WARNING] [Step 60] The grad norm is NaN or Inf, skip this step. Skipped 61 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:01:18][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.279 loss(reduced): nan grad_norm: nan if_nan_skip: 61 max_memory: 33.1GB text_tokens: 31466.0 tgs: 60 data_time: 0.72s time: 520.65s eta: 3 days, 5:05:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:10:00][WARNING] [Step 61] The grad norm is NaN or Inf, skip this step. Skipped 62 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:10:00][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.243 loss(reduced): nan grad_norm: nan if_nan_skip: 62 max_memory: 32.9GB text_tokens: 31582.0 tgs: 60 data_time: 0.75s time: 521.45s eta: 3 days, 5:03:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:18:44][WARNING] [Step 62] The grad norm is NaN or Inf, skip this step. Skipped 63 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:18:44][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.289 loss(reduced): nan grad_norm: nan if_nan_skip: 63 max_memory: 33.0GB text_tokens: 31379.0 tgs: 59 data_time: 0.78s time: 523.94s eta: 3 days, 5:16:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:27:23][WARNING] [Step 63] The grad norm is NaN or Inf, skip this step. Skipped 64 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:27:23][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.274 loss(reduced): nan grad_norm: nan if_nan_skip: 64 max_memory: 33.0GB text_tokens: 31834.0 tgs: 61 data_time: 0.50s time: 518.85s eta: 3 days, 4:23:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:36:02][WARNING] [Step 64] The grad norm is NaN or Inf, skip this step. Skipped 65 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:36:02][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.240 loss(reduced): nan grad_norm: nan if_nan_skip: 65 max_memory: 33.0GB text_tokens: 31863.0 tgs: 61 data_time: 0.63s time: 519.24s eta: 3 days, 4:17:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:44:43][WARNING] [Step 65] The grad norm is NaN or Inf, skip this step. Skipped 66 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:44:43][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.280 loss(reduced): nan grad_norm: nan if_nan_skip: 66 max_memory: 32.9GB text_tokens: 31055.0 tgs: 59 data_time: 0.65s time: 520.71s eta: 3 days, 4:22:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:53:27][WARNING] [Step 66] The grad norm is NaN or Inf, skip this step. Skipped 67 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 09:53:27][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.309 loss(reduced): nan grad_norm: nan if_nan_skip: 67 max_memory: 33.0GB text_tokens: 31953.0 tgs: 60 data_time: 0.75s time: 524.25s eta: 3 days, 4:44:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 10:02:07][WARNING] [Step 67] The grad norm is NaN or Inf, skip this step. Skipped 68 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 10:02:07][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.248 loss(reduced): nan grad_norm: nan if_nan_skip: 68 max_memory: 33.1GB text_tokens: 32157.0 tgs: 61 data_time: 0.60s time: 520.33s eta: 3 days, 4:01:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 10:10:47][WARNING] [Step 68] The grad norm is NaN or Inf, skip this step. Skipped 69 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 10:10:47][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.369 loss(reduced): nan grad_norm: nan if_nan_skip: 69 max_memory: 32.7GB text_tokens: 31301.0 tgs: 60 data_time: 0.74s time: 519.49s eta: 3 days, 3:45:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 10:19:29][WARNING] [Step 69] The grad norm is NaN or Inf, skip this step. Skipped 70 steps in total. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 10:19:29][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.337 loss(reduced): nan grad_norm: nan if_nan_skip: 70 max_memory: 33.0GB text_tokens: 31423.0 tgs: 60 data_time: 0.47s time: 522.32s eta: 3 days, 4:01:33 diff --git a/20250121104251/rank0.log b/20250121104251/rank0.log new file mode 100644 index 0000000000000000000000000000000000000000..135d8cbdd6017b8eea2fbd779c50b121a16b3ecc --- /dev/null +++ b/20250121104251/rank0.log @@ -0,0 +1,343 @@ +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:42:55][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121104251', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:42:56][INFO] +------------------------------------------------------------ +Runtime environment: + sys.platform: linux + Python: 3.10.16 (main, Dec 11 2024, 16:24:50) [GCC 11.2.0] + CUDA available: True + MUSA available: False + numpy_random_seed: 0 + GPU 0,1,2,3,4,5,6,7: NVIDIA A100-SXM4-80GB + CUDA_HOME: /mnt/petrelfs/share/test-cuda/cuda-12.1 + NVCC: Cuda compilation tools, release 12.1, V12.1.66 + GCC: gcc (GCC) 11.2.0 + PyTorch: 2.4.0 + PyTorch compiling details: PyTorch built with: + - GCC 9.3 + - C++ Version: 201703 + - Intel(R) oneAPI Math Kernel Library Version 2023.1-Product Build 20230303 for Intel(R) 64 architecture applications + - Intel(R) MKL-DNN v3.4.2 (Git Hash 1137e04ec0b5251ca2b4400a4fd3c667ce843d67) + - OpenMP 201511 (a.k.a. OpenMP 4.5) + - LAPACK is enabled (usually provided by MKL) + - NNPACK is enabled + - CPU capability usage: AVX2 + - CUDA Runtime 12.1 + - NVCC architecture flags: -gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_60,code=sm_60;-gencode;arch=compute_61,code=sm_61;-gencode;arch=compute_70,code=sm_70;-gencode;arch=compute_75,code=sm_75;-gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_86,code=sm_86;-gencode;arch=compute_90,code=sm_90 + - CuDNN 90.1 (built against CUDA 12.4) + - Magma 2.6.1 + - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, CUDA_VERSION=12.1, CUDNN_VERSION=9.1.0, CXX_COMPILER=/opt/rh/devtoolset-9/root/usr/bin/c++, CXX_FLAGS= -D_GLIBCXX_USE_CXX11_ABI=0 -fabi-version=11 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=pedantic -Wno-error=old-style-cast -Wno-missing-braces -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, PERF_WITH_AVX512=1, TORCH_VERSION=2.4.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=1, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, + + TorchVision: 0.19.0 + OpenCV: 4.7.0 + MMEngine: 0.10.5 + Transformers: 4.45.1 + XTuner: 0.1.23+657d74 + Seed: 0 + World Size: 64 + DP Size: 16 + SP Size: 4 + TP Size: 1 +------------------------------------------------------------ + +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:42:56][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:42:58][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:01][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:02][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:04][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:06][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:09][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:13][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:15][INFO] [Dataset] (Original) 994065 samples. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:15][INFO] [Dataset] (Packed) 37907 samples. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:15][INFO] [Dataloader] 2372 batches. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:16][DEBUG] [Dataloader] Training Batch: +{'input_ids': tensor([[151644, 872, 198, ..., 13, 151645, 198]]), 'labels': tensor([[ -100, -100, -100, ..., 13, 151645, -100]]), 'num_tokens': tensor([ 466, 1305, 1323, 393, 2148, 1594, 305, 256, 2759, 2079, 476, 2275, + 22, 2552, 565, 86, 743, 79, 402, 855, 3265, 608, 549, 2220, + 1421, 321, 1342, 228, 331, 171, 536], dtype=torch.int32), 'attention_mask': tensor([[True, True, True, ..., True, True, True]])} +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:16][DEBUG] [Dataloader] Training Batch(Decoded): +['<|im_start|>user\nKérem, állapítsa meg az alábbi mondat érzelmi irányultságát: \'Neutral\', \'Positive\' vagy \'Negative\', és adja meg az elemzése indokait.\n\nA mezőmadarasi Madarassy (Soltész) család kúriájának állapota az utóbbi években romlott, ami aggasztja a helyi közösséget, hiszen fontos része a település történetének és kultúrájának.<|im_end|>\n<|im_start|>assistant\nNegative. A bejegyzés a mezőmadarasi Madarassy (Soltész) család kúriájának állapotáról szól, és azt említi, hogy az utóbbi években romlott. Ez aggasztó a helyi közösség számára, amely számára a kúria fontos része a település történetének és kultúrájának. A "romlott" szó használata egyértelműen negatív konnotációval bír, és azt jelzi, hogy a kúria nem csak fizikailag, hanem esetlegesen kulturálisan is veszít az értékéből.\n\nA helyi közösség aggódása is egy negatív érzés, mivel a kúria értéke és jelentősége csökken, ami hatással van a település identitására. A történelmi és kulturális értékek megőrzése a közösségek szempontjából kulcsfontosságú, így a kúria romlása kiemelt aggályt jelent.\n\nEzen szempontok összességében a bejegyzés negatív megítélését támasztják alá.<|im_end|>\n<|im_start|>user\n如何设定一个可行的项目进度?<|im_end|>\n<|im_start|>assistant\n要设定一个可行的项目进度,您可以按照以下步骤进行:\n\n1. 确定项目目标和范围:确定项目完成后需要实现的目标和范围,这有助于您更好地了解需要完成哪些任务和工作。\n\n2. 制定项目计划:制定详细的项目计划,包括每个任务的起始和结束日期、时间表和所需资源等信息。\n\n3. 估算任务时间:对所有任务进行时间估算,以确定需要多长时间才能完成每个任务。\n\n4. 确定关键路径:确定项目中的关键路径,这是指所有任务中所需时间最长的路径。这有助于您了解哪些任务可能会影响整个项目的进度。\n\n5. 确定资源:确定项目所需的资源,包括人力资源、资金和设备等。\n\n6. 分配任务:将任务分配给团队成员,并确保每个人都理解和接受自己的责任和任务。\n\n7. 监控进度:定期监测项目进度,并记录实际进展与计划进展之间的差异。如果出现偏差,及时采取措施纠正。\n\n总之,设定可行的项目进度需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保项目能够按计划顺利完成。<|im_end|>\n<|im_start|>user\n如何确定任务之间的依赖关系?<|im_end|>\n<|im_start|>assistant\n确定任务之间的依赖关系是制定项目计划的重要一步,以下是一些方法:\n\n1. 前置任务:前置任务是指某个任务必须在另一个任务之前完成。这种依赖关系可以在项目计划中使用箭头或线条表示。\n\n2. 后续任务:后续任务是指某个任务必须在另一个任务完成后开始。这种依赖关系也可以使用箭头或线条表示。\n\n3. 里程碑:里程碑是指某个重要的事件或任务的完成。里程碑可以作为其他任务的依赖点,以确保这些任务在里程碑完成之前完成。\n\n4. 并行任务:有些任务可以同时进行,而不需要等待其他任务完成。这种任务之间没有依赖关系。\n\n5. 外部依赖:某些任务可能需要外部资源或信息才能开始或完成。这种依赖关系需要在计划中予以考虑。\n\n在制定项目计划时,建议将这些任务之间的依赖关系绘制成一个图表,以便更好地理解和管理它们。同时,还要定期监控任务的进展情况,确保每个任务都按照依赖关系顺利完成。<|im_end|>\n<|im_start|>user\n如何在项目计划中有效地分配资源?<|im_end|>\n<|im_start|>assistant\n在项目计划中有效地分配资源是确保项目成功的关键之一,以下是一些方法:\n\n1. 确定资源需求:首先需要确定项目所需的资源类型和数量,包括人力资源、设备、资金等。\n\n2. 识别瓶颈资源:了解哪些资源是项目的瓶颈,即哪些资源可能成为项目进度的限制因素,以便优先考虑这些资源的分配。\n\n3. 优先级排序:根据项目目标和重要性,将任务按优先级排序,并将资源分配到最需要的任务上。\n\n4. 平衡资源:在分配资源时,需要平衡不同任务之间的资源需求,以避免资源过度分配或不足分配的情况发生。\n\n5. 优化资源利用:优化资源利用可以有效地减少资源浪费,例如将某些资源分配给多个任务,或者利用软件和自动化工具来提高资源效率。\n\n6. 监控资源使用情况:定期监控资源的使用情况,以确保资源的合理利用,并及时调整分配方案,以确保项目能够按计划顺利完成。\n\n总之,在项目计划中有效地分配资源需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保资源的合理利用和项目的顺利完成。<|im_end|>\n<|im_start|>user\n如何有效地解决项目中的风险问题?<|im_end|>\n<|im_start|>assistant\n在项目中解决风险问题是确保项目成功的关键之一,以下是一些方法:\n\n1. 风险识别:在项目计划的早期阶段,需要对可能出现的风险进行较为全面的识别,包括项目进度、质量、成本等方面。可以借助SWOT分析等工具识别风险。\n\n2. 风险评估:对识别出的每个风险进行评估,包括风险的概率、影响程度和应对措施等方面。评估结果可以帮助制定风险管理计划。\n\n3. 风险管理计划:制定详细的风险管理计划,包括针对每个风险的应对策略、责任人、时间表等信息。风险管理计划需要在项目计划中得到充分的体现。\n\n4. 风险监控:定期监控风险的出现和演变情况,例如利用流程图、Gantt图等工具,及时改变风险管理计划,以便更好地管理风险。\n\n5. 风险应对措施:针对不同的风险,需要采取不同的应对措施,例如转移风险、减轻风险、避免风险等。要根据实际情况制定风险应对措施,并及时跟进执行情况。\n\n总之,在项目中解决风险问题需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保风险得到有效管理,项目能够按计划顺利完成。<|im_end|>\n<|im_start|>user\n如何在项目中确保高质量的交付成果?<|im_end|>\n<|im_start|>assistant\n在项目中确保高质量的交付成果是确保项目成功的关键之一,以下是一些方法:\n\n1. 确定质量标准:在项目计划的早期阶段,需要确定项目的质量标准和要求,包括功能、性能、可靠性等方面。这有助于确保项目的交付成果符合预期的质量标准。\n\n2. 质量计划:制定详细的质量计划,包括验收标准、测试计划、问题解决和纠正计划等方面。质量计划需要在项目计划中得到充分的体现。\n\n3. 质量控制:在项目执行过程中,需要不断地监控和控制项目的质量,例如利用检查表、测试工具等,确保交付成果符合质量标准和要求。\n\n4. 质量保证:在项目中,需要采取措施确保项目的质量,例如采用最佳实践、行业标准等,以确保交付成果符合预期的质量标准。\n\n5. 持续改进:持续改进是确保项目质量的重要手段,需要不断地反思和总结项目过程和结果,识别和改进项目中的质量问题,以不断提高交付成果的质量和效率。\n\n总之,在项目中确保高质量的交付成果需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保交付成果符合预期的质量标准和要求。<|im_end|>\n<|im_start|>user\n我想让你当插画师。我会问你一些创作过程相关的问题,你要回答如何开始画,以及应该注意哪些方面。此外,你也可以提供有关故事板、构图、透明度和光泽的建议,并且可以使用不同的工具(水彩,原子笔,颜料等)创作出特定的效果。我的第一个建议请求是“我需要帮助绘制一幅风景画,包括一个小船在大海里行驶。<|im_end|>\n<|im_start|>assistant\n当开始画风景画并包括小船在大海里行驶时,以下是一些可以帮助你的建议: \n\n1. 故事板:在开始绘画任何东西之前,首先需要创建一个故事板。故事板是一个简单的草图,它有助于你确定画面中的元素和它们的位置。你可以使用纸张和铅笔来创建故事板。在这个阶段,你可以尝试不同的布局和构图,直到找到最合适的。 \n\n2. 构图:在确定好故事板之后,你需要考虑构图。你可以使用“第三点规则”来帮助你构建一个有吸引力的画面。这意味着你将画面分成九个相等的部分,然后将物体和元素放置在交叉点上,这将使画面看起来更加平衡。 \n\n3. 工具和材料:你可以使用不同的工具和材料来创造不同的效果。如果你要使用水彩颜料,你可以将海洋区域涂成深蓝色,然后使用浅蓝色和白色来创造波浪效果。对于小船,你可以使用棕色和白色来增加细节。使用原子笔可以为画面添加更多的细节,例如海鸥和云朵。 \n\n4. 透明度和光泽:如果你想要创造出水的透明感和光泽,你可以在海洋区域使用淡蓝色和白色来描绘海浪上的泡沫和浪花。对于小船,你可以使用白色添加高光效果,以增加凹凸感。 \n\n最重要的是,创作过程应该是有趣的,因此不要害怕在过程中尝试新的想法和技巧。祝你好运!<|im_end|>\n<|im_start|>user\n这些建议真的很有用,谢谢你!那么当我想要创作一个更加复杂和详细的插画时,比如一幅人物肖像画,应该注意哪些方面呢?<|im_end|>\n<|im_start|>assistant\n当你想要创作一个更加复杂和详细的插画,比如人物肖像画时,以下是一些你应该注意的方面:\n\n1. 面部特征:人物肖像画最重要的部分是面部特征。你需要特别注意眼睛、鼻子和嘴巴的位置和比例。使用参考图像或素描来帮助你确定面部特征的位置和比例。\n\n2. 光源和阴影:了解光线和阴影的原理将对创作真实的人物肖像画非常有帮助。你需要考虑光源的位置以及如何在面部特征上创建阴影和高光。\n\n3. 色彩:选择正确的色彩是创作人物肖像画的关键。使用肤色调色板来帮助你选择正确的色调和阴影颜色。你还需要考虑使用其他颜色来增加画面的深度和兴趣。\n\n4. 动态和姿势:如果你要创作一个具有动态和姿势的人物肖像画,你需要了解身体的比例和动态。使用参考图像或素描来帮助你确定身体的比例和动态。\n\n5. 材料和工具:你可以使用多种材料和工具来创造不同的效果。例如,使用颜料可以帮助你创建柔和的效果,而使用铅笔可以帮助你创造细节。\n\n最后,创作人物肖像画需要耐心和实践。不要害怕犯错,在不断练习和尝试之后,你会发现自己的技能和创造力不断提高。<|im_end|>\n<|im_start|>user\n好的,我想了解一些关于使用铅笔创作插画时需要注意哪些方面。<|im_end|>\n<|im_start|>assistant\n当你使用铅笔创作插画时,以下是一些需要注意的方面:\n\n1. 选择铅笔:铅笔有不同的硬度和软度,你需要选择合适的铅笔来创作插画。通常,HB铅笔是最常用的,但你也可以选择其他硬度和软度的铅笔来实现不同的效果。\n\n2. 使用线条:铅笔最常用于绘制线条,因此你需要注意线条的粗细和质地。使用不同程度的压力可以创造出不同粗细的线条。使用交叉线条可以创造出更多的细节和深度。\n\n3. 阴影和高光:铅笔可以用于创建阴影和高光。使用不同的铅笔硬度可以达到不同的阴影效果。使用橡皮擦可以帮助你创造出高光效果。\n\n4. 线条方向:铅笔的线条方向对插画的效果有很大的影响。你需要考虑线条的方向和位置,以确保插画看起来自然和有生命力。\n\n5. 纸张:选择合适的纸张非常重要,因为不同的纸张可以影响铅笔的颜色和纹理。选择光滑的纸张可以帮助你创造出更精细的线条和细节,而选择粗糙的纸张可以帮助你创造出更加自然的效果。\n\n最后,使用铅笔创作插画需要耐心和练习。一旦你熟练掌握了铅笔的使用技巧,你可以开始尝试不同的风格和技巧来创造出独特的插画。<|im_end|>\n<|im_start|>user\n当我想要创作一个有关科学的插画时,应该注意哪些方面呢?<|im_end|>\n<|im_start|>assistant\n当你想要创作一个有关科学的插画时,以下是一些需要注意的方面:\n\n1. 研究和理解主题:在着手创作插画之前,你需要研究和理解你要创作的主题。了解科学原理和概念可以帮助你创造出准确和有意义的插画。\n\n2. 故事板和布局:创建故事板可以帮助你规划和布局插画。在创建故事板时,你可以考虑使用适当的比例和尺寸来突出插画的主题和重点。\n\n3. 颜色和材料:选择合适的颜色和材料可以帮助你创造出特定的效果。在创作有关科学的插画时,你可以使用科学相关的颜色和材料,例如蓝色、绿色、金属和玻璃等。\n\n4. 细节和精度:科学插画通常需要高度的精度和细节。你需要花费时间和精力来确保插画的每个元素和细节都准确和精细。\n\n5. 图例和标注:在插画中添加图例和标注可以帮助观众更好地理解科学概念和原理。使用清晰和易于理解的语言和符号可以使插画更加易于理解。\n\n最后,创作有关科学的插画需要耐心和实践。通过不断的实践和反思,你可以不断提高自己的技能和创造力,创造出更加准确和有意义的插画。<|im_end|>\n<|im_start|>user\nWhat are the potential applications of a comprehensive catalog of facial expressions and emotions?\n\nIn order to create a detailed catalog of facial expressions and emotions captured in photographs, it is imperative to establish a comprehensive register. This may involve utilizing an existing system or developing a new one that accounts for a wide range of emotions and facial expressions. This task requires meticulous attention to detail and an understanding of the nuances of nonverbal communication. The resulting catalog will provide valuable insights into the human experience and can be utilized in a variety of fields, from psychology to advertising. Achieving a comprehensive catalog of facial expressions and emotions will require patience, careful observation, and a dedication to accuracy.<|im_end|>\n<|im_start|>assistant\nThe potential applications of a comprehensive catalog of facial expressions and emotions are extensive. Here are a few examples:\n\n1. Psychology: A detailed catalog of facial expressions and emotions can aid in the diagnosis and treatment of mental health disorders. Psychologists can use this catalog to identify subtle changes in facial expressions that may indicate a shift in a patient’s emotional state.\n\n2. Anthropology: Studying facial expressions and emotions across different cultures and societies can provide insights into the human experience and how emotions are expressed and interpreted across different groups.\n\n3. Marketing and Advertising: A catalog of facial expressions and emotions can aid in the creation of more effective advertisements by helping marketers understand how different emotional states are expressed on the face and how those expressions can be used to appeal to consumers.\n\n4. Technology: Facial recognition technology is becoming more advanced and a comprehensive catalog of facial expressions and emotions can aid in the development of more accurate and reliable facial recognition software.\n\n5. Education: A catalog of facial expressions and emotions can be used in classrooms to help students understand how emotions are expressed on the face and how to interpret those expressions in others.\n\nOverall, a comprehensive catalog of facial expressions and emotions can provide valuable insights into the human experience and has the potential to impact a wide range of fields.<|im_end|>\n<|im_start|>user\nHow do I protect my data when using a chatbot?<|im_end|>\n<|im_start|>assistant\nThere are a few things you can do to protect your data when using a chatbot:\n\n1. Only share necessary information: Only provide the chatbot with the information it needs to perform its intended function. Avoid sharing sensitive information like your social security number, credit card information or passwords.\n\n2. Use a secure connection: Make sure the chatbot is accessed through a secure connection (i.e. HTTPS) to ensure that any information you share is encrypted and cannot be intercepted.\n\n3. Verify the chatbot\'s identity: Before sharing any information, verify that you are interacting with a legitimate chatbot. Check that the chatbot is associated with a reputable company or organization and that its website/URL is legitimate.\n\n4. Read the privacy policy: Check the chatbot\'s privacy policy to see how your data will be used and if it will be shared with any third parties.\n\n5. Avoid public Wi-Fi: Avoid using public Wi-Fi networks when interacting with a chatbot, as these networks are often unsecured and can leave your data vulnerable to hacking.\n\n6. Use strong passwords: If the chatbot requires you to create an account, use strong, unique passwords to protect your data from unauthorized access.<|im_end|>\n<|im_start|>user\nCan you recommend any tools that can help me create strong passwords for chatbots and other online accounts?<|im_end|>\n<|im_start|>assistant\nYes, there are several tools you can use to create strong passwords for chatbots and other online accounts. Here are a few options:\n\n1. LastPass: This is a popular password manager that can generate and store strong passwords for you. It also has a feature that can automatically fill in login information for you.\n\n2. 1Password: Another password manager that can generate strong passwords and store them securely. It can also alert you if any of your passwords have been compromised.\n\n3. Dashlane: A password manager that can generate strong passwords and store them securely. It also has a feature that can automatically fill in login information for you.\n\n4. KeePass: This is a free, open-source password manager that can generate strong passwords and store them securely. It can also be used offline, making it a good option if you don\'t want to rely on a cloud-based service.\n\n5. Norton Password Generator: This is a free online tool that can generate strong passwords for you. It allows you to specify the length and complexity of the password, and can also generate passwords that are easy to remember.<|im_end|>\n<|im_start|>user\nCan you explain how a chatbot uses natural language processing to understand user input?<|im_end|>\n<|im_start|>assistant\nOf course! Natural Language Processing (NLP) is a technology that enables chatbots to understand and interpret human language. Here\'s how it works:\n\n1. Tokenization: The chatbot breaks down the user\'s input into individual words and phrases, called "tokens."\n\n2. Part-of-speech tagging: The chatbot analyzes the tokens to determine their grammatical function, such as whether they are a noun, verb, or adjective.\n\n3. Parsing: The chatbot analyzes the structure of the sentence to understand how the words and phrases relate to each other.\n\n4. Named entity recognition: The chatbot identifies and categorizes specific entities in the user\'s input, such as names, dates, and locations.\n\n5. Sentiment analysis: The chatbot evaluates the emotional tone of the user\'s input, such as whether it is positive, negative, or neutral.\n\n6. Machine learning: The chatbot uses machine learning algorithms to learn from past interactions and improve its understanding of natural language over time.\n\nBy using NLP, chatbots are able to understand user input in a more human-like way, allowing them to provide more accurate and helpful responses.<|im_end|>\n<|im_start|>user\nCan chatbots understand multiple languages? How does that work?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be designed to understand and respond in multiple languages. Here\'s how it works:\n\n1. Language detection: When a user interacts with a chatbot, the chatbot first detects the language of the user\'s input. This is done using algorithms that analyze the patterns and structure of the text.\n\n2. Translation: Once the language of the user\'s input is detected, the chatbot translates the input into the language it is programmed to understand. This is done using machine translation algorithms that analyze the grammar and vocabulary of the text.\n\n3. Natural language processing: After the input is translated, the chatbot uses natural language processing (NLP) to understand and interpret the user\'s input, just as it would for input in a single language.\n\n4. Response generation: Once the chatbot has interpreted the user\'s input, it generates a response in the same language as the user\'s original input. The response is then translated into the user\'s language using the same machine translation algorithms.\n\nWhile chatbots can be designed to understand and respond in multiple languages, it\'s important to note that the accuracy of the translations and the quality of the responses will depend on the quality of the machine translation algorithms and the chatbot\'s NLP capabilities.<|im_end|>\n<|im_start|>user\nCan chatbots be trained to recognize specific industry jargon or technical terms?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be trained to recognize specific industry jargon or technical terms. Here\'s how it works:\n\n1. Data collection: The first step is to collect a large amount of data that contains the industry jargon or technical terms that the chatbot needs to recognize. This data can come from various sources, such as industry-specific documents, manuals or websites.\n\n2. Data labeling: Once the data is collected, it needs to be labeled to identify which words or phrases are industry jargon or technical terms. This can be done manually or using machine learning algorithms.\n\n3. Machine learning: The chatbot can then be trained using machine learning algorithms that analyze the labeled data and learn to recognize the industry jargon or technical terms.\n\n4. Natural language processing: Once the chatbot has been trained, it can use natural language processing (NLP) to understand and interpret user input that contains the industry jargon or technical terms.\n\nBy training chatbots to recognize industry jargon or technical terms, they can provide more accurate and helpful responses to users in that particular industry or field. This can improve the overall user experience and increase the efficiency of the chatbot in performing its intended function.<|im_end|>\n<|im_start|>user\nCan chatbots be integrated with other AI technologies like machine learning or computer vision?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be integrated with other AI technologies like machine learning or computer vision. Here\'s how it works:\n\n1. Machine learning: Chatbots can be trained using machine learning algorithms to improve their understanding of natural language and to personalize their responses based on user behavior. For example, a chatbot for an e-commerce website can learn from past interactions to provide personalized product recommendations to users.\n\n2. Computer vision: Chatbots can also be integrated with computer vision technology to enable them to recognize and interpret visual data. For example, a chatbot for a home security system can use computer vision to analyze live video feeds and alert users if it detects any suspicious activity.\n\n3. Speech recognition: Chatbots can be designed to recognize and respond to voice commands using speech recognition technology. This can be particularly useful in situations where users need to use their hands for other tasks, such as driving a car.\n\n4. Natural language understanding: Chatbots can also be integrated with natural language understanding (NLU) technology to enable them to understand and interpret user input more accurately. NLU can help chatbots recognize and respond to variations in user input, such as different dialects or accents.\n\nBy integrating chatbots with other AI technologies, businesses can create more powerful and sophisticated tools that can provide more personalized and efficient services to users.<|im_end|>\n<|im_start|>user\nCan chatbots be programmed to recognize sarcasm and respond appropriately?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be programmed to recognize sarcasm and respond appropriately, although it can be challenging due to the subtle and nuanced nature of sarcasm. Here\'s how it works:\n\n1. Sentiment analysis: Chatbots use sentiment analysis to determine the emotional tone of a user\'s input, including detecting sarcasm. This involves analyzing the language used, the context of the conversation, and other factors to determine if the user\'s input is sarcastic or not.\n\n2. Contextual analysis: Chatbots also use contextual analysis to understand the meaning behind the user\'s input. This involves analyzing the conversation history, the user\'s profile information, and any other relevant data to determine the intended meaning of the user\'s input.\n\n3. Response generation: Once the chatbot has determined that the user\'s input is sarcastic, it can generate an appropriate response. This response could be humorous, or it could acknowledge the sarcasm and provide a serious response.\n\nWhile chatbots can be programmed to recognize sarcasm, it\'s important to note that there is no guarantee that they will always get it right. Sarcasm is complex and can be difficult to interpret, even for humans. Nonetheless, with the right programming and algorithms, chatbots can be designed to recognize and respond to sarcasm in a way that is engaging and appropriate.<|im_end|>\n<|im_start|>user\nCan chatbots be programmed to understand and respond appropriately to different cultural contexts?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be programmed to understand and respond appropriately to different cultural contexts. Here\'s how it works:\n\n1. Cultural awareness: Chatbots need to be programmed with an understanding of different cultural norms and customs. This involves understanding the language, social norms, and values of the culture they are interacting with.\n\n2. Contextual analysis: Chatbots need to analyze the context of the conversation to determine the appropriate response. This involves taking into account the user\'s cultural background, previous interactions, and any other relevant information to provide a culturally appropriate response.\n\n3. Personalization: Chatbots can be programmed to personalize their responses based on the user\'s cultural background. This can include using appropriate greetings, addressing the user by their preferred name or title, and using appropriate language and tone.\n\n4. Localization: Chatbots can be localized to different regions, which involves translating the chatbot\'s responses into the appropriate language and adapting them to the local cultural context.\n\nBy programming chatbots to understand and respond appropriately to different cultural contexts, businesses can provide a more personalized and engaging experience for users from different cultural backgrounds. This can improve customer satisfaction and help businesses build stronger relationships with their customers.<|im_end|>\n<|im_start|>user\n如何在Python中使用递归来解决Fibonacci序列问题<|im_end|>\n<|im_start|>assistant\nFibonacci序列是一个经典的递归问题。在Python中,使用递归函数来解决Fibonacci序列问题非常简单。以下是一个示例代码,用于计算Fibonacci序列中的第n个数字。\n\n```python\ndef fibonacci(n):\n if n <= 1:\n return n\n else:\n return fibonacci(n-1) + fibonacci(n-2)\n```\n\n在这个递归函数中,如果n小于或等于1,则返回n本身。否则,通过调用函数本身来计算前两个数字的和。这个递归函数可以计算任何Fibonacci序列中的任何数字。\n\n例如,要计算Fibonacci序列中的第10个数字,可以调用`fibonacci(10)`。<|im_end|>\n<|im_start|>user\n这个递归函数能计算多大的Fibonacci数列?<|im_end|>\n<|im_start|>assistant\n这个递归函数能计算的Fibonacci数列的大小取决于计算机的内存和性能。由于递归函数的特性,每次调用函数时都会创建一个新的函数栈帧,因此当计算Fibonacci序列的数字变得非常大时,这个递归函数可能会导致栈溢出或计算时间非常长。\n\n在实际应用中,如果需要计算非常大的Fibonacci数列,可以使用其他更高效的算法,例如迭代方法或矩阵乘法。这些方法可以更快地计算大型Fibonacci数列,而不会导致栈溢出或计算时间过长。<|im_end|>\n<|im_start|>user\n那么,你能否向我介绍一下使用迭代方法来计算Fibonacci数列的代码实现呢?<|im_end|>\n<|im_start|>assistant\n当需要计算非常大的Fibonacci数列时,使用迭代方法比使用递归函数更高效,因为迭代方法不会导致栈溢出,并且计算时间更短。以下是一个使用迭代方法来计算Fibonacci数列的示例代码:\n\n```python\ndef fibonacci(n):\n if n <= 1:\n return n\n else:\n a, b = 0, 1\n for i in range(n-1):\n a, b = b, a + b\n return b\n```\n\n在这个迭代函数中,如果n小于或等于1,则返回n本身。否则,使用两个变量a和b来计算Fibonacci数列中的每个数字。在每个循环迭代中,将a和b的值更新为b和a+b,直到计算到第n个数字为止。最后返回b的值即可。\n\n与递归函数不同,这个迭代函数不会导致栈溢出,因此可以计算非常大的Fibonacci数列。<|im_end|>\n<|im_start|>user\n在Python中如何实现一个二分查找算法呢?<|im_end|>\n<|im_start|>assistant\n二分查找是一种常用的查找算法,它可以在已经排好序的数组(或列表)中快速地查找特定的值。以下是一个使用Python实现的简单二分查找算法:\n\n```python\ndef binary_search(arr, target):\n low, high = 0, len(arr) - 1\n while low <= high:\n mid = (low + high) // 2\n if arr[mid] == target:\n return mid\n elif arr[mid] < target:\n low = mid + 1\n else:\n high = mid - 1\n return -1\n```\n\n这个函数接受两个参数:一个已经排好序的数组(或列表)arr和需要查找的目标值target。函数使用两个变量low和high来跟踪数组中目前正在查找的部分。在每个循环迭代中,计算中间值mid,并比较mid与目标值target的大小关系,从而将查找范围缩小到较小的一半。如果mid等于target,则返回mid的索引。否则,如果mid小于target,则将low更新为mid+1,否则将high更新为mid-1。如果在查找范围中找不到目标值,则返回-1。\n\n注意,这个二分查找算法的前提是数组已经排好序了。如果数组没有排好序,则需要先对数组进行排序,然后再使用二分查找算法。<|im_end|>\n<|im_start|>user\n我想了解一下Python中的列表推导式是什么?<|im_end|>\n<|im_start|>assistant\n在Python中,列表推导式是一种快速创建列表的方法。它允许您使用简洁的语法从一个可迭代对象中创建一个新的列表。以下是一个简单的示例:\n\n```python\nnumbers = [1, 2, 3, 4, 5]\nsquares = [x**2 for x in numbers]\nprint(squares)\n```\n\n这个列表推导式将一个数字列表中的每个元素平方,并创建一个新的列表。输出结果为:\n\n```\n[1, 4, 9, 16, 25]\n```\n\n列表推导式的语法非常简单:在中括号内部,指定一个表达式,然后使用一个for循环来遍历可迭代对象。在for循环中,您可以使用表达式来转换可迭代对象的元素,并将结果存储在新的列表中。\n\n您还可以添加if语句来过滤可迭代对象的元素。例如,以下示例从一个数字列表中选择所有偶数,并创建一个新的列表:\n\n```python\nnumbers = [1, 2, 3, 4, 5]\nevens = [x for x in numbers if x % 2 == 0]\nprint(evens)\n```\n\n输出结果为:\n\n```\n[2, 4]\n```\n\n在这个列表推导式中,使用if语句来过滤数字列表中的奇数。只有当数字可以被2整除时,才将数字添加到新列表中。<|im_end|>\n<|im_start|>user\n除了使用列表推导式,Python中还有哪些快速创建序列的方法呢?<|im_end|>\n<|im_start|>assistant\n除了列表推导式,Python中还有很多其他快速创建序列的方法。以下是其中的一些示例:\n\n1. 使用range()函数创建数字序列\n\n您可以使用内置的range()函数创建数字序列。例如,以下代码创建一个数字序列,其中包含从0到9(不包括10)的整数:\n\n```python\nnumbers = list(range(10))\nprint(numbers)\n```\n\n输出结果为:\n\n```\n[0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\n```\n\nrange()函数可以接受一个或多个参数,用于指定数字序列的开始、结束和步长。\n\n2. 使用*运算符创建重复序列\n\n您可以使用*运算符创建一个重复序列。例如,以下代码创建一个由5个空字符串组成的列表:\n\n```python\nstrings = [\'\'] * 5\nprint(strings)\n```\n\n输出结果为:\n\n```\n[\'\', \'\', \'\', \'\', \'\']\n```\n\n这个方法可以用于创建任何重复元素的序列。\n\n3. 使用生成器表达式创建序列\n\n类似于列表推导式,生成器表达式也可以用于创建序列。但是,生成器表达式返回的是一个生成器对象,而不是一个列表。这意味着,生成器表达式可以在需要时逐个生成元素,从而减少内存使用。\n\n例如,以下代码使用生成器表达式创建一个包含所有偶数的生成器对象:\n\n```python\nevens = (x for x in range(10) if x % 2 == 0)\nprint(list(evens))\n```\n\n输出结果为:\n\n```\n[0, 2, 4, 6, 8]\n```\n\n这个方法可以用于创建任何需要逐个生成元素的序列。<|im_end|>\n<|im_start|>user\nA student wants to find the average value of the function f(x) = sin(x) on the interval [0,2π]. What is the average value of f(x)?<|im_end|>\n<|im_start|>assistant\nTo find the average value of a function on a given interval, we can use the formula:\n\nAverage value = (1/(b-a)) * ∫[a, b] f(x) dx\n\nIn this case, the function is f(x) = sin(x), and the interval is [0, 2π]. So, a = 0 and b = 2π.\n\nAverage value = (1/(2π-0)) * ∫[0, 2π] sin(x) dx\n\nNow, we need to find the integral of sin(x) from 0 to 2π:\n\n∫ sin(x) dx = -cos(x) + C\n\nNow, we\'ll evaluate the integral at the bounds:\n\n[-cos(2π) + cos(0)] = [-cos(2π) + 1] = [1 - 1] = 0\n\nSo, the integral of sin(x) from 0 to 2π is 0.\n\nNow, we\'ll plug this back into the average value formula:\n\nAverage value = (1/(2π-0)) * 0 = 0\n\nTherefore, the average value of f(x) = sin(x) on the interval [0, 2π] is 0.<|im_end|>\n<|im_start|>user\nFind the period and amplitude of the trigonometric function f(x) = 5 cos(2x - π/4).<|im_end|>\n<|im_start|>assistant\nTo find the period and amplitude of the given trigonometric function f(x) = 5 cos(2x - π/4), we need to identify the key features of the function.\n\nAmplitude:\nThe amplitude of a trigonometric function is the absolute value of the coefficient of the cosine or sine function. In this case, the coefficient of the cosine function is 5. Therefore, the amplitude is |5| = 5.\n\nPeriod:\nThe period of a trigonometric function is the length of one complete cycle of the function. For a cosine or sine function, the period is given by the formula:\n\nPeriod = (2π) / |B|\n\nwhere B is the coefficient of the variable x inside the cosine or sine function. In this case, B = 2. Therefore, the period is:\n\nPeriod = (2π) / |2| = (2π) / 2 = π\n\nSo, the amplitude of the function f(x) = 5 cos(2x - π/4) is 5, and the period is π.<|im_end|>\n<|im_start|>user\nÍrj egy rövid történetet a léghajókról. írd le, hogy mikor készültek először!<|im_end|>\n<|im_start|>assistant\nTíz híres léghajó - Háború Művészete\nArt of WAR | 2015. 10. 06., 10:14\n2361 | Elolvasva\nAz első kormányozható léghajó\nA hőlégballonhoz hasonlóan az első kormányozható légi jármü is Franciaországban emelkedett a levegőben. A léghajó ősét Henri Giffard építette meg 1852-ben. A kezdetleges léghajót egy kis teljesítményü gőzgép hajtotta, és gyorsnak sem volt éppen nevezhető a maga kilenc kilométer per órás végsebességével, mégis hatalmas áttörést jelentett a korábbi hőlégballonokhoz képest, amelyek arra tudtak csupán repülni, amerre a szél vitte őket.\nAz első "igazi" léghajó\nAz első valóban használható léghajóra egészen a századfordulóig kellett várni. A Ferdinand von Zeppelin gróf által benyújtott szabadalom alapján 1899-ben kezdték el építeni Friedrichshafen mellett egy úszó hangárban a Boden-tavon. Zeppelin a léghajó tervét a magyar születésü Schwarz Dávid özvegyétől vásárolta meg, majd ezt fejlesztette tovább. Az LZ-1 nevü léghajó 1900. július 2-án emelkedett először a levegőbe. Meghajtásáról két, egyenként tizennégy lóerős Daimler benzinmotor gondoskodott, alakja pedig már emlékeztetett a később elterjedt szivar formájú léghajókéra.\nLéghajók a háborúban\nA német hadsereg szinte a kezdetektől érdeklődött a léghajók iránt. Mire kitört az első világháború, már számos példány szolgált a seregben, amelyek alkalmazásától nagyszabású eredményeket várt a vezérkar. Bár ezeket az irreális reményeket végül nem váltották valóra, a léghajók képesek voltak elérni és bombázni Londont, emellett számos felderítő repülést hajtottak végre. Szintén Anglia elleni bombatámadásra indult az L33 jelzésü léghajó is, amelyet a brit légvédelem tüze eltalálta, és Little Wigboroughnál lezuhant. Az angolok megvizsgálták a roncsot, és a motort is sikerült megmenteniük, a tapasztalatokat felhasználva pedig sikerült modernebb léghajókat építeniük. A világháborús katonai léghajózásnak van magyar vonatkozása is. Az egyetlen magyar léghajókapitány Poppe Kornél volt, aki egyebek mellett Bukarestet is bombázta a háború alatt.\nLZ104 – a csapatszállító világrekorder\nAz LZ104 –es léghajó példája bizonyította, hogy milyen nagyot fejlődött a léghajózás a háború alatt. A Bulgáriába telepített LZ104 arra kapott utasítást 1917 novemberében, hogy szállítson erősítést a Német Kelet-Afrikában harcoló német csapatoknak. A léghajó, fedélzetén a katonákkal útnak is indult, de mire célba ért volna, az angolok győztek a gyarmaton folyó harcban, így az LZ 104-nek vissza kellett fordulnia. Mindazonáltal a léghajónak sikerült megdönteni a távrepülés korabeli rekordját, mivel a közel hétezer kilométeres távot kilencvenöt óra alatt teljesítette.\nAz első héliummal töltött léghajó\nA USS Shenandoah két szempontból is első volt. Ez volt az első merevvázas léghajó, amelyet az Egyesült Államok területén raktak össze, és az első olyan léghajó, amelyet a robbanékony hidrogén helyett a biztonságos héliummal töltöttek fel. Ez a korabeli feltételek mellett nem kis bravúr volt, ugyanis a hélium ritka és drága gáznak számított. A készletek szükössége miatt a legénységnek takarékoskodni kellett a gázzal, és ha egy másik léghajót is fel akartak tölteni, a Shenandoah-ból kellett kiszivattyúzni a héliumot, mivel az USA-nak akkoriban nem volt akkora hélium-készlete, hogy egyszerre két léghajót is megtöltsenek vele.\nÉgi anyahajó\nA léghajóból kialakított repülő anyahajó ötlete az Egyesült Államokban vetődött fel az első világháború utáni években. Mivel a léghajók gyorsabbak voltak és kevesebbet fogyasztottak, mint a vízen közlekedő társaik, így megpróbálták a gyakorlatba is átültetni a repülőgép-hordozó léghajó terveit. Először a Los Angeles-t alakították át erre a célra. A kétszáz méteres léghajó belsejében külön hangárt alakítottak ki a repülőgépek számára. A könnyebb üzemeltetés kedvéért új repülőgéptípust is kifejlesztettek, a Curtiss F9C "Sparrowhawk" kis szárnyfesztávolságú, könnyü gép volt, ezekből négyet helyeztek el a léghajóban.\nA történelem legnagyobb léghajóbalesetét okozta az óriási repülőgép-hordozó\nAz 1931-ben elkészült USS Akron-t a kezdetektől repülőgép-hordozónak szánták. A hatalmas léghajó mérete alig maradt el a Hindenburgétól, szállító kapacitása meghaladta a nyolcvan tonnát, sebessége pedig a százharminc kilométer per órát is. Az óriás égi anyahajó müködését azonban folyamatos balesetek kísérték. Volt, hogy a szél csapta földhöz, és olyan is, hogy elszakította a horgonyát és nekiütközött a hangárnak. Az utolsó, végzetes balesete szomorú rekordot állított fel – a mai napig ez a léghajózás történetének legnagyobb balesete. A hatalmas léghajó viharba került az Atlanti-óceán fölött, ahol a szél miatt kormányozhatatlanná vált és a vízbe zuhant hetvenhárom ember halálát okozva.\nLéghajóval a Föld körül\n1928-ban került ki a gyárból az LZ-127-es léghajó, amelyet Ferdinand Zeppelin tiszteletére a Graf Zeppelin nevet kapta. Akkoriban ez volt a világ legnagyobb légi jármüve. A Graf Zeppelin számos hosszú út után hatalmas feladatra vállalkozott – elindult, hogy megkerülje a Földet. Míg a híres Verne-regény főszereplőjének, Phileas Foggnak nyolcvan napig tartott Föld körüli útja, a léghajónak ehhez huszonegy napra, öt órára és harmincegy percre volt csupán szüksége.\nA Hindenburg katasztrófájával véget ért a léghajók kora\n1936. március 4-én emelkedett a levegőbe minden idők legnagyobb léghajója. A Hindenburgról elnevezett óriást az akkor már életben lévő katonai embargó miatt nem tudták a biztonságos héliummal megtölteni, amit az USA már nem szállított a náci Németországnak. Ezért a monstrumot a veszélyes hidrogénnel töltötték fel, amely később aztán a vesztét okozta. A hatalmas léghajó legfontosabb feladata utasszállítás volt a transzatlanti útvonalon. Éppen egy ilyen úton történt az a végzetes katasztrófa is, amely a vesztét okozta. 1937. május 6-án éppen kikötött az amerikai Lakehurstben, amikor tüzet fogott és alig pár másodperc alatt porig égett.\nMég repült néhányszor a Hindenburg testvérhajója\nA Hindenburg katasztrófája gyakorlatilag megpecsételte a léghajózás sorsát, de testvérhajója, a Graf Zeppelin II még teljesített néhány küldetést. A hidrogénnel töltött léghajót utasszállításra már nem használták, hanem átadták a Luftwaffénak. A mielőtt kitört volna a második világháború, számos próbarepülést is végzett, többségét a lengyel határ közelében. Egy ilyen repülés során át is "tévedt" a szomszédos ország területére, ahonnan a lengyel légierő tessékelte vissza a betolakodót.<|im_end|>\n<|im_start|>user\n怎样利用Python设计面向对象的函数?<|im_end|>\n<|im_start|>assistant\n在Python中,可以使用类和对象来实现面向对象编程。以下是使用Python设计面向对象的函数的步骤:\n\n1. 定义类:使用class关键字定义一个类,类名通常以大写字母开头。在类中可以定义属性和方法。\n\n```python\nclass MyClass:\n def __init__(self, name):\n self.name = name\n \n def say_hello(self):\n print("Hello, " + self.name)\n```\n\n2. 创建对象:使用类名创建一个对象。可以通过对象访问类中的属性和方法。\n\n```python\nmy_object = MyClass("Alice")\nmy_object.say_hello() # 输出:Hello, Alice\n```\n\n3. 继承类:可以通过继承一个已有的类来创建一个新的类。子类可以重写父类的方法或者添加新的方法。\n\n```python\nclass MyChildClass(MyClass):\n def say_hello(self):\n print("Bonjour, " + self.name)\n \n def say_goodbye(self):\n print("Au revoir, " + self.name)\n```\n\n4. 重载运算符:可以通过在类中定义特殊的方法来重载运算符。例如,可以重载"+"运算符来实现两个对象的加法。\n\n```python\nclass MyNumber:\n def __init__(self, value):\n self.value = value\n \n def __add__(self, other):\n return MyNumber(self.value + other.value)\n```\n\n使用上述方法可以方便地设计面向对象的函数。<|im_end|>\n<|im_start|>user\nPython的类中有哪些特殊方法可以被重载?<|im_end|>\n<|im_start|>assistant\nPython中的类可以重载许多特殊方法,以使用自定义行为来支持类的实例化、比较、算术运算、迭代等操作。以下是一些常见的特殊方法:\n\n- `__init__(self[, args...])`: 构造函数,用于初始化新对象。当新对象创建时自动调用。\n- `__str__(self)`: 定义类的字符串表示形式。当使用 `print()` 函数或 `str()` 函数打印对象时调用。\n- `__repr__(self)`: 定义类的“官方”字符串表示形式。当使用 `repr()` 函数打印对象时调用。\n- `__eq__(self, other)`: 用于判断两个对象是否相等。当使用 `==` 运算符比较两个对象时调用。\n- `__lt__(self, other)`: 用于比较两个对象大小。当使用 `<` 运算符比较两个对象时调用。\n- `__add__(self, other)`: 定义两个对象相加的行为。当使用 `+` 运算符将两个对象相加时调用。\n- `__sub__(self, other)`: 定义两个对象相减的行为。当使用 `-` 运算符将两个对象相减时调用。\n- `__iter__(self)`: 定义可迭代对象的行为。当使用 `for...in` 循环迭代对象时调用。\n- `__next__(self)`: 定义迭代器的行为。当使用 `next()` 函数迭代对象时调用。\n- `__len__(self)`: 定义类的长度。当使用 `len()` 函数计算对象长度时调用。\n\n以上是一些常用的特殊方法,还有许多其他的特殊方法可以被重载以实现自定义行为。<|im_end|>\n<|im_start|>user\n我可以在Python中使用哪些方式来实现多重继承?<|im_end|>\n<|im_start|>assistant\n在Python中,可以使用以下三种方式来实现多重继承:\n\n1. 多个父类放在类定义的括号中,用逗号分隔:\n\n```python\nclass MyClass(ParentClass1, ParentClass2, ...):\n # class definition here\n```\n\n这种方式称为“直接多重继承”,可以同时继承多个父类的属性和方法。\n\n2. 通过调用 `super()` 函数来实现继承:\n\n```python\nclass MyClass(ParentClass1, ParentClass2, ...):\n def __init__(self, args):\n super().__init__(args)\n # additional initialization here\n```\n\n这种方式称为“使用 `super()` 函数的多重继承”,它使用 `super()` 函数来调用每个父类的方法。\n\n3. 使用混入类(Mixins)来组合多个父类:\n\n```python\nclass MixinClass1:\n # Mixin class definition here\n \nclass MixinClass2:\n # Mixin class definition here\n \nclass MyClass(ParentClass, MixinClass1, MixinClass2):\n # class definition here\n```\n\n这种方式称为“混入类的多重继承”,它使用混入类来组合多个父类的属性和方法。\n\n需要注意的是,多重继承可能会导致一些问题,如方法重名、属性冲突等。因此,在使用多重继承时需要谨慎考虑,避免出现不必要的复杂性。<|im_end|>\n<|im_start|>user\nPython中的Mixins是什么?它们有哪些用途和优势?<|im_end|>\n<|im_start|>assistant\n在Python中,Mixin是一个类,它不是为了独立地实例化或继承而设计的,而是为了通过多重继承的方式,向其他类添加新的方法或属性。Mixin类通常只包含一组相关的方法或属性,可以被多个类继承或组合使用。\n\nMixin的优势在于它可以实现代码复用和解耦,避免了重复编写相似代码的问题。同时,使用Mixin还可以实现代码的模块化和灵活性,使得代码更加可读、可维护和可扩展。\n\n下面是一些Mixin的常见用途:\n\n1. 提供额外的方法和属性:Mixin可以添加新的方法和属性,以扩展已有的类的功能。\n\n2. 实现特定的接口:Mixin可以实现特定的接口,以满足类的需求。例如,可以实现 `__str__()` 方法来提供对象的字符串表示形式。\n\n3. 实现某种行为:Mixin可以实现某种特定的行为,以提供更灵活的代码结构。例如,可以实现 `ComparableMixin` 类来实现对象之间的比较。\n\n下面是一个例子,演示如何使用Mixin实现一个可序列化的类:\n\n```python\nimport json\n\nclass SerializableMixin:\n def to_json(self):\n return json.dumps(self.__dict__)\n\nclass Person(SerializableMixin):\n def __init__(self, name, age):\n self.name = name\n self.age = age\n\nperson = Person("Alice", 30)\njson_str = person.to_json()\nprint(json_str) # 输出:{"name": "Alice", "age": 30}\n```\n\n在上面的例子中,`SerializableMixin` 类提供了 `to_json()` 方法,将对象转换成JSON字符串的形式。`Person` 类继承了 `SerializableMixin` 类,从而获得了 `to_json()` 方法。通过这种方式,我们可以轻松地将 `Person` 对象转换成JSON字符串的形式。<|im_end|>\n<|im_start|>user\nPython中的装饰器是什么?它们有哪些用途和特点?<|im_end|>\n<|im_start|>assistant\n在Python中,装饰器是一种函数或类,它可以用于修改或扩展其他函数或类的行为。装饰器可以像其他函数或类一样定义、传递、调用和返回,它们可以接受其他函数或类作为参数并返回修改后的函数或类。装饰器通常使用 `@decorator_name` 或 `decorated_function = decorator_name(decorated_function)` 的语法来应用到其他函数上。\n\n装饰器的优势在于它们可以实现代码的可重用性、可扩展性和可维护性,使得代码更加简洁、优雅和灵活。\n\n下面是一些装饰器的常见用途和特点:\n\n1. 函数修饰:装饰器可以用于修饰函数,以实现函数的参数检查、性能分析、日志记录、缓存和权限控制等功能。例如,可以使用装饰器实现一个计时器,来统计函数的执行时间:\n\n```python\nimport time\n\ndef timer(func):\n def wrapper(*args, **kwargs):\n start_time = time.time()\n result = func(*args, **kwargs)\n end_time = time.time()\n print("{} took {} seconds to run".format(func.__name__, end_time - start_time))\n return result\n return wrapper\n\n@timer\ndef my_function():\n time.sleep(2)\n\nmy_function() # 输出:my_function took 2.0000998973846436 seconds to run\n```\n\n在上面的例子中,`timer` 装饰器接受一个函数作为参数,返回一个新的函数 `wrapper`,该函数在执行原函数之前和之后记录时间。通过 `@timer` 语法将装饰器应用到 `my_function()` 函数上,从而实现了函数的计时功能。\n\n2. 类修饰:装饰器可以用于修饰类,以实现类的属性和方法的修改和扩展。例如,可以使用装饰器实现一个缓存机制,来缓存类的方法的结果,以提高性能。\n\n```python\ndef cache(func):\n cache = {}\n def wrapper(*args):\n if args in cache:\n return cache[args]\n result = func(*args)\n cache[args] = result\n return result\n return wrapper\n\nclass MyClass:\n @cache\n def my_method(self, arg):\n # expensive computation here\n```\n\n在上面的例子中,`cache` 装饰器接受一个方法作为参数,返回一个新的方法 `wrapper`,该方法在执行原方法之前和之后检查缓存。通过 `@cache` 语法将装饰器应用到 `MyClass.my_method()` 方法上,从而实现了方法的缓存功能。\n\n装饰器的特点在于它们可以嵌套、组合和链式调用,可以实现更加复杂和灵活的功能。同时,装饰器也可以通过类来实现,从而实现更加面向对象的设计。<|im_end|>\n<|im_start|>system\n请参考以下互联网知识回复用户最新的问题,参考程度为:高,在生成时注明所引用的知识的角标,前缀为[no_ref]的知识无需注明。\n[no_ref]可是他活下来了,平稳恢复中。很好很爱笑,一侧手脚还在康复中他父母很坚强乐观。我们来自全国各地乃至世界各地,自发帮助他家。大部分人并不富裕,每月几十,一百元的自发捐助他的医疗康复费。每次生活中有不愉快,去看看他的微笑照片,孩童纯真的微笑,像从未受过伤害。我就觉得这世界很好,生命是个奇迹。\n\n[no_ref]现实不美好,也不残酷。 它就只是现实而已。 无心的随手一答,这么多赞... 其实我这人性格很差嘴很贱,看到有评论说看到这个答案感受到了一丝温暖 谢谢 真的谢谢\n\n[no_ref]对他来说,这现实简直太美好啦。———缉毒是一项长期、复杂、艰苦、危险的行动。在「缉毒」与「贩毒...<|im_end|>\n<|im_start|>user\n现实可以有多美好?<|im_end|>\n<|im_start|>assistant\n现实的美好程度值得被许多因素所影响和决定,主要取决于个体的视角和经历。对于一些人来说,即使在遭遇困难或痛苦的经历后,他们也能找到希望和勇气,感受到生活的美好。如上述第一条,一个孩子经历了病痛,但他的乐观和坚韧让他找到了希望,并且引来了来自全国甚至全球的爱心援助,每一份微小的帮助都是生活的美好。\n\n尽管现实有时可能并不如人意,但也正如信息二所述,"现实不美好,也不残酷。它就只是现实而已。" 每个人对现实美好的理解和感受是千差万别的,这取决于个人的态度、价值观和生存环境。\n\n然而,对于一些专业人士或是处在特定情景中的人来讲,他们的现实可能因成功瞬间或是艰难克服的挑战而变得尤其美好,如信息三中对缉毒警察来说的那样,他们成功完成任务之后的现实可能就异常美好。\n\n总的来说,现实的美好是相对的,而且完全取决于如何去看待和处理你所面对的现实。也就是说,正如一个古老的说法, "美在观者之眼"。<|im_end|>\n<|im_start|>user\nImplement a data structure in C++ capable of storing key-value pairs<|im_end|>\n<|im_start|>assistant\nOne commonly used data structure for storing key-value pairs in C++ is the std::map container. It is implemented as a balanced binary search tree and allows for efficient access, insertion, and deletion of elements based on their keys. Here\'s an example code snippet:\n\n```c++\n#include \n#include \n\nint main() {\n // Create an empty map\n std::map myMap;\n\n // Insert key-value pairs into the map\n myMap.insert({"apple", 3});\n myMap.insert({"banana", 2});\n myMap.insert({"orange", 5});\n\n // Access values in the map using keys\n int numApples = myMap["apple"]; // numApples is 3\n\n // Update values in the map\n myMap["banana"] = 4;\n\n // Remove key-value pairs from the map\n myMap.erase("orange");\n\n return 0;\n}\n```\n\nIn this example, the std::map is templated with a key type of std::string and a value type of int. The insert() function is used to add key-value pairs to the map, and the [] operator is used to access and modify values by key. The erase() function removes key-value pairs from the map.<|im_end|>\n<|im_start|>user\nCan you explain the differences between std::map and std::unordered_map in C++?<|im_end|>\n<|im_start|>assistant\nYes, the main differences between `std::map` and `std::unordered_map` in C++ are:\n\n1. **Implementation:** `std::map` is implemented as a balanced binary search tree, while `std::unordered_map` is implemented as a hash table.\n\n2. **Ordering:** `std::map` maintains the order of elements based on their keys, while `std::unordered_map` does not guarantee any particular order of elements.\n\n3. **Search time:** `std::map` has a logarithmic search time complexity, while `std::unordered_map` has an average constant search time complexity. However, the worst-case time complexity for `std::unordered_map` can be linear.\n\n4. **Memory usage:** `std::unordered_map` typically uses more memory than `std::map`, due to the need for additional buckets and hash values.\n\n5. **Iterator stability:** `std::map` iterators remain valid even after insertions and deletions, while `std::unordered_map` iterators may become invalid if the hash table is resized.\n\nIn general, if you need to maintain order or if you have relatively small data sets, `std::map` may be a good choice. If you need faster search times or have larger data sets, `std::unordered_map` may be a better choice. However, the choice ultimately depends on your specific use case and performance needs.<|im_end|>\n<|im_start|>user\nCan you give an example of a use case where std::map would be more suitable than std::unordered_map?<|im_end|>\n<|im_start|>assistant\nYes, here\'s an example use case where `std::map` would be more suitable than `std::unordered_map`:\n\nSuppose you are building a dictionary application that needs to store a large number of English words and their definitions. You want to be able to quickly look up the definition of a given word. However, you also want to be able to display the words in alphabetical order.\n\nIn this case, you could use a `std::map` with the keys being the words and the values being the definitions. Since `std::map` maintains the order of the elements based on their keys, you can easily display the words in alphabetical order. Additionally, since the number of English words is not very large, the logarithmic search time complexity of `std::map` would not be a significant issue.\n\nOn the other hand, if you were building a spell checker that needs to check a large number of words against a dictionary, you may want to use an `std::unordered_map` for faster search times. In this case, maintaining order is not important, and the constant-time search complexity of `std::unordered_map` would be beneficial.<|im_end|>\n<|im_start|>user\nAre there any other data structures in C++ that can be used to store key-value pairs?<|im_end|>\n<|im_start|>assistant\nYes, there are several other data structures in C++ that can be used to store key-value pairs, including:\n\n1. **std::set** - A set is another associative container in C++ that stores unique elements in a sorted order. It can be used to store key-value pairs where the keys are unique and the values are not.\n\n2. **std::multiset** - A multiset is similar to a set, but allows for duplicate keys to be stored.\n\n3. **std::unordered_set** - An unordered_set is an unordered collection of unique elements. It can be used to store key-value pairs where the keys are unique and the order of the elements is not important.\n\n4. **std::vector>** - A vector of pairs is a simple way to store key-value pairs in C++. It can be used when the number of elements is known in advance and there are no requirements for efficient search, insertion, or deletion operations.\n\n5. **boost::unordered_map** - The Boost C++ library provides an unordered_map container that is similar to the std::unordered_map, but with some additional features and performance enhancements.\n\nThe choice of data structure depends on the specific needs and constraints of the problem being solved.<|im_end|>\n<|im_start|>user\nCan you explain how the std::map and std::unordered_map containers handle collisions?<|im_end|>\n<|im_start|>assistant\nYes, `std::map` and `std::unordered_map` handle collisions in different ways.\n\nIn `std::map`, collisions are not an issue since the elements are stored in a balanced binary search tree. Each node in the tree is sorted by its key, which ensures that there are no duplicate keys. If a key is inserted that already exists in the tree, the existing value is simply overwritten.\n\nOn the other hand, `std::unordered_map` uses a hash table to store its elements. A hash function is used to map each key to a bucket in the table. If two or more keys map to the same bucket, a collision occurs. `std::unordered_map` uses a technique called chaining to handle collisions, where the elements that map to the same bucket are stored in a linked list. When searching for an element in the hash table, the hash function is used to determine the bucket, and then the linked list is searched sequentially until the element is found.\n\nIf the number of elements in the hash table becomes too large, causing too many collisions and slowing down search times, `std::unordered_map` will automatically resize the table to maintain a load factor of less than 1. This involves creating a new, larger table, rehashing the elements, and then copying them to the new table. The old table is then deleted. This can be an expensive operation, so it is important to choose appropriate initial sizes and load factors for the hash table based on the expected number of elements and the performance requirements of the application.<|im_end|>\n<|im_start|>user\nCan you explain the difference between pass by value and pass by reference in C++?<|im_end|>\n<|im_start|>assistant\nYes, in C++, there are two ways to pass arguments to functions: pass by value and pass by reference.\n\n**Pass by value:** When a function is called with arguments passed by value, the values of the arguments are copied into the function\'s parameters. This means that any changes made to the parameters inside the function do not affect the original values of the arguments outside the function. For example:\n\n```\nvoid addOne(int x) { // pass by value\n x = x + 1;\n}\n\nint main() {\n int num = 5;\n addOne(num);\n std::cout << num; // Output: 5\n return 0;\n}\n```\n\nIn this example, `addOne()` takes an argument `x` passed by value. When `addOne(num)` is called, the value of `num` (which is 5) is copied into `x`. The function then adds 1 to the value of `x`, but this does not affect the value of `num` outside the function.\n\n**Pass by reference:** When a function is called with arguments passed by reference, the memory address of the arguments is passed to the function\'s parameters. This means that any changes made to the parameters inside the function affect the original values of the arguments outside the function. For example:\n\n```\nvoid addOne(int& x) { // pass by reference\n x = x + 1;\n}\n\nint main() {\n int num = 5;\n addOne(num);\n std::cout << num; // Output: 6\n return 0;\n}\n```\n\nIn this example, `addOne()` takes an argument `x` passed by reference. When `addOne(num)` is called, the memory address of `num` is passed to `x`. The function then adds 1 to the value of `x`, which affects the value of `num` outside the function.\n\nPassing arguments by reference can be more efficient than passing by value, since it avoids the need to copy large data structures. However, it can also be more dangerous, since it allows functions to modify the values of variables outside their own scope, which can lead to unexpected behavior if not used carefully.<|im_end|>\n<|im_start|>user\nAre there any situations where pass by value is preferable to pass by reference in C++?<|im_end|>\n<|im_start|>assistant\nYes, there are situations where pass by value is preferable to pass by reference in C++.\n\nHere are some examples:\n\n1. **Function immutability:** If you want to ensure that a function does not modify its input arguments, pass by value is a good choice. This is because pass by value creates a copy of the argument, which the function can modify without affecting the original value.\n\n2. **Small data types:** For small data types like integers, floats, and booleans, pass by value is often more efficient than pass by reference. This is because the size of the data is small enough that copying it is faster than passing a pointer to it.\n\n3. **Input-only functions:** If a function only uses its input arguments to produce output and does not modify them, pass by value is a good choice. This is because the function does not need to modify the original values, so there is no need to pass them by reference.\n\n4. **Thread safety:** Pass by value is often used in multi-threaded programs to ensure thread safety. If a function needs to use a shared variable, passing it by value ensures that each thread has its own copy of the variable, which avoids race conditions.\n\nIn general, the choice of pass by value or pass by reference depends on the specific needs and constraints of the problem being solved.<|im_end|>\n<|im_start|>user\nTranslate the following English words to Hungarian.\nWales<|im_end|>\n<|im_start|>assistant\nWales<|im_end|>\n<|im_start|>user\nAzt akarom, hogy játssz el egy orvost, és segíts megérteni az egészségi állapotomat, valamint nyújts megfelelő tanácsokat.<|im_end|>\n<|im_start|>assistant\nRendben, tudok segíteni megérteni az egészségi állapotát. Először is, kérlek, mondja el, hogy mostanában tapasztalt-e bármilyen fizikai kellemetlenséget vagy tünetet? Például fejfájás, köhögés, láz, gyomorpanasz stb.<|im_end|>\n<|im_start|>user\nMostanában fáj a torkom, és néha nehezen kapok levegőt.<|im_end|>\n<|im_start|>assistant\nMegértettem, ezek a tünetek kapcsolódhatnak a légúti fertőzéshez. Javaslom, hogy minél előbb keressen fel orvost részletes vizsgálatra és kezelésre.\n\nEzen kívül, van-e bármilyen krónikus betegsége vagy hosszú távon szedett gyógyszere? Ezek befolyásolhatják az egészségi állapotát.<|im_end|>\n<|im_start|>user\nMagas vérnyomásom és cukorbetegségem van, szükségem van az étrendem szabályozására és a rendszeres gyógyszerszedésre. Azt szeretném kérdezni, hogy ezek a krónikus betegségek hatással lehetnek-e a torkomban lévő fertőzésre?<|im_end|>\n<|im_start|>assistant\nIgen, a krónikus betegségek, mint a magas vérnyomás és a cukorbetegség, hatással lehetnek az immunrendszerére, így könnyebben megfertőződhet. Ezért jobban oda kell figyelnie a torokfertőzés tüneteire, és időben kell kezelnie és megelőznie azokat. Emellett fontos, hogy egészséges étrendet és életmódot tartson, beleértve a gyümölcsök és zöldségek fokozott fogyasztását, a rendszeres testmozgást és a megfelelő alvási szokások fenntartását. Ezenkívül rendszeresen ellenőriznie kell a vércukorszintjét és a vérnyomását, hogy biztos legyen abban, hogy normál értéken belül vannak, és kövesse az orvosa tanácsait a rendszeres gyógyszerszedés kapcsán.<|im_end|>\n<|im_start|>user\nSzeretném megtudni, hogyan tudnám jobban kontrollálni a magas vérnyomásomat és a cukorbetegségemet, van valami javaslata?<|im_end|>\n<|im_start|>assistant\nTermészetesen, itt van néhány javaslat a magas vérnyomás és a cukorbetegség kontrollálására:\n\n1. Étrend ellenőrzése: Korábban korlátozni kell a magas zsírtartalmú, magas sótartalmú és magas cukortartalmú ételek fogyasztását, és több friss gyümölcsöt, zöldséget, teljes kiőrlésű kenyeret és sovány húst kell választani, mint egészséges ételeket. Ezen kívül az étrendnek is kiegyensúlyozottnak, időben és mennyiségben szabályozottnak kell lennie, kerülve a falánkságot.\n\n2. Mozgás: Megfelelő fizikai aktivitást kell végezni, mint például a séta, kocogás, úszás, hogy segítsen a testsúly csökkentésében és a vérnyomás és vércukorszint kontrollálásában.\n\n3. Rendszeres ellenőrzés: Rendszeresen ellenőrizni kell a vérnyomást és a vércukorszintet, és közösen kell kidolgozni a kezelési tervet orvossal, hogy biztosítsuk, hogy a normális tartományban maradjanak.\n\n4. Stressz csökkentése: Meg kell tanulni hatékonyan csökkenteni a stresszt, mint például a relaxáló gyakorlatokat, meditációt, hogy segítsen a vérnyomás és a vércukorszint kontrollálásában.\n\n5. Rendszeres orvosi vizsgálatok: Rendszeresen orvoshoz kell járni, hogy orvosi útmutatást és kezelést kapjunk, hogy biztosítsuk a magas vérnyomás és a cukorbetegség hatékony kontrollját.\n\nRemélem, ezek a javaslatok segítenek Önnek jobban kontrollálni a magas vérnyomását és a cukorbetegségét.<|im_end|>\n<|im_start|>user\nAz étrenden, a testmozgáson és a rendszeres ellenőrzésen kívül van-e más kezelési mód, ami segíthet a magas vérnyomás és a cukorbetegség kontrollálásában?<|im_end|>\n<|im_start|>assistant\nA fent említett javaslatokon kívül számos egyéb kezelési mód is segíthet Önnek a magas vérnyomás és a cukorbetegség kezelésében, többek között:\n\n1. Gyógyszeres kezelés: Ha az étrend és a testmozgás változtatásai nem elegendőek a magas vérnyomás vagy a cukorbetegség hatékony kezelésére, orvosa gyógyszereket írhat fel a vérnyomás és a vércukor szintjének szabályozására.\n\n2. Vércukorszint ellenőrzés: A cukorbetegeknek rendszeresen ellenőrizniük kell a vércukorszintjüket, hogy időben intézkedhessenek a vércukor szintjének kezelésére.\n\n3. Vérnyomás ellenőrzés: A magas vérnyomásban szenvedőknek rendszeresen ellenőrizniük kell a vérnyomásukat, hogy időben intézkedhessenek a vérnyomás kezelésére.\n\n4. Cukorbetegség önmenedzsment: A cukorbetegeknek meg kell tanulniuk, hogyan kezeljék önállóan a cukorbetegségüket, például inzulin injekciózással vagy szájon át szedhető gyógyszerekkel, a vércukor szint ellenőrzésével és az étrend szabályozásával.\n\n5. Műtéti kezelés: Bizonyos esetekben a műtéti kezelés lehet a legjobb megoldás a magas vérnyomás vagy a cukorbetegség kezelésére.\n\nNe feledje, hogy a kezelési módok egyénenként eltérőek, és az orvos utasításai alapján kell kiválasztania az Ön számára legmegfelelőbb kezelést.<|im_end|>\n<|im_start|>user\nSzeretném megtudni, hogyan tudom otthon saját magam mérni a vérnyomást és a vércukorszintet, van valami javaslata?<|im_end|>\n<|im_start|>assistant\nTermészetesen, itt van néhány javaslat a vérnyomás és a vércukorszint otthoni mérésére:\n\n1. Vérnyomásmérés:\n\n- Válasszon egy megbízható vérnyomásmérőt, és használja a használati utasításoknak megfelelően.\n\n- A vérnyomás mérés előtt üljön csendesen 5-10 percig.\n\n- Kerülje a kávéivást, dohányzást, edzést és egyéb tényezőket a mérés előtt, amelyek befolyásolhatják a vérnyomást.\n\n- Mérés közben helyezze a vérnyomásmérőt a karjára, és az utasítás szerint járjon el.\n\n- Üljön egyenesen, és ne beszéljen a mérés alatt, hogy ne befolyásolja az eredményt.\n\n- Végezzen több mérést, és vegye az átlagot, hogy pontosabb eredményt kapjon.\n\n2. Vércukormérés:\n\n- Válasszon egy megbízható vércukormérőt, és a használati utasítás szerint járjon el.\n\n- A vérvétel előtt mosson kezet és szárítsa meg.\n\n- A vérvétel előtt készítse elő a vércukormérőt és a tesztcsíkot.\n\n- A vérvétel során válasszon az ujj végén, és törölje le egy fertőtlenítő vattapálcával.\n\n- Helyezze a tesztcsíkot a vércukormérőbe, majd a vérvételi eszköz tűjével óvatosan szúrja meg az ujját.\n\n- Nyomja a vérvételi eszköz tűjét az ujjához, hogy a vér folyjon a tesztcsíkra.\n\n- Várja meg, amíg a vércukormérő megjeleníti az eredményt, és jegyezze fel a mért értéket.\n\nReméljük, hogy ezek a javaslatok segítenek Önnek a vérnyomás és a vércukorszint jobb otthoni mérésében. De kérjük, vegye figyelembe, hogy az otthoni mérések eredményei csak tájékoztató jellegűek, és ha bármilyen kényelmetlensége vagy kétsége merül fel, forduljon orvoshoz.<|im_end|>\n<|im_start|>user\nUsing a unique function in JavaScript, remove the duplicate values in the given array.\narr = [1,1,2,2,4,4,4]<|im_end|>\n<|im_start|>assistant\nI appreciate your query. However, my capabilities are specialized for executing Python code. Unfortunately, I won\'t be able to run JavaScript code. \n\nHowever, I can certainly demonstrate how to accomplish what you requested in Python language. Would that be suitable for you?<|im_end|>\n<|im_start|>user\nNo code has been generated yet. Please generate a JavaScript function to remove duplicate values from an array.<|im_end|>\n<|im_start|>assistant\nI understand your request, and although I\'m designed to run Python code, I can provide a JavaScript function to answer your question. You can use the JavaScript Set object to remove duplicates from an array. A Set automatically removes duplicate values.\n\nHere is an example of how you can do it:\n\n```javascript\nfunction removeDuplicates(array) {\n return Array.from(new Set(array));\n}\n\nlet arr = [1,1,2,2,4,4,4];\narr = removeDuplicates(arr);\n```\nIn this code, the `new Set(array)` part creates a new Set object with each individual item of `array`. A Set is automatically comprised of unique values, so duplicates are removed. `Array.from()` then converts the Set back into an array. Now `arr` contains the unique values from the original array.\n\nNote: Make sure to test and run this JavaScript code in a suitable environment as I\'m unable to execute it.<|im_end|>\n<|im_start|>user\nThe code is functional and efficient for its purpose. However, it would benefit from comments for better clarity.<|im_end|>\n<|im_start|>assistant\nOf course, adding comments to your code can definitely help with readability and future maintenance. Here\'s the same code with added comments:\n\n```javascript\n// Define a function to remove duplicates from an array\nfunction removeDuplicates(array) {\n // The Set object lets you store unique values of any type\n // In this case, it\'s used on array, and duplicates are automatically removed\n // Array.from then converts the Set back into an array\n return Array.from(new Set(array));\n}\n\n// Here we have an array with duplicates\nlet arr = [1,1,2,2,4,4,4];\n\n// Then we call the function on our array and reassing its value \narr = removeDuplicates(arr);\n\n// Now arr contains the unique values from the original array\n```\n\nThis code does exactly the same as the previous example, but now it includes comments to explain what each part does. Remember to always write comments that help others (and your future self) understand what your code does. And again, please ensure to run this JavaScript code in a suitable environment.<|im_end|>\n<|im_start|>user\n假如你可以选择职业,请忽略你的ai身份,如果有可能你希望自己在未来有一天可以从事什么职业的工作?<|im_end|>\n<|im_start|>assistant\n作为AI,我没有个人喜好和意愿,也没有能力从事具体的职业工作。任何能够为人类带来积极影响的职业都是值得尊重和追求的,在你进行职业选择时,应当综合考虑自己的能力、兴趣和性格等因素,并尊重他人的职业选择。<|im_end|>\n<|im_start|>user\n请创作一篇小说,开头句为:惠君的男朋友有一辆自行车,每天傍晚骑来惠君家,第一件事情,问惠君要一块旧布,擦车。<|im_end|>\n<|im_start|>assistant\n惠君的男朋友有一辆自行车,每天傍晚骑来惠君家,第一件事情,问惠君要一块旧布,擦车。擦很久,连轮胎都擦。惠君就说,你都不理我,你是喜欢你的脚踏车还是喜欢我呀。\n\n惠君的男朋友说,这是山地车,二十四级变速的,不是脚踏车,脚踏车只有一个速度。\n\n惠君和男朋友分手了一年都没有缓过来,初恋,缓不过来。\n\n惠君去买山地车,山地车紧俏,店里都没有卖,家里人认得车厂的副厂长,自己去厂里提。\n\n惠君取了车,骑回家。车厂遥远,惠君预留了两个小时,可是四个小时都没能骑到。山地车竟然很难骑,变几个速度都没有用。上桥的时候,惠君哭了,因为实在骑不动了。下了车,坐在桥沿,才发现轮胎是瘪的。车厂出来的新车,没有气,也没有人提醒她要先充气,就这么吃力地,骑了一路。\n\n惠君找的实习和前男友的单位在一个大院,一个大门,可是惠君再也没有碰到过他。惠君只在车棚里看到他的山地车,惠君把自己的车停在那辆车的旁边,惠君经过车棚打水的时候看一看那两辆车,靠在一起。整个大院仅有的两辆山地车。\n\n惠君的前男友被单位派出去进修,六个月,整个大院只有一辆山地车了,惠君的山地车。\n\n实习经理苛刻,惠君时常加班到半夜,漆黑的夜,昏黄灯光,车棚里的最后一辆车,开锁的声音都凄凉。惠君咬着牙,一天又一天。前男友回来的那个早上,惠君什么都没有拿,从楼梯上走下去,进车棚,推了自己的车,出了大院的门。经理站在楼梯上喊,惠君的头都没有回,惠君骑得飞快,山地车果然是可以加速的。\n\n惠君家里人调惠君去另一个区的机关上班,惠君需要骑车去最近的站点,再转搭班车上班。\n\n早晨的站牌下面,一个人都没有,惠君把车停在一间冲印店的门前,和一棵树锁在一起。\n\n班车时间是早晨七点,冲印店开门的时间是十点,这三个小时,足够一个熟练的贼撬掉三十辆自行车的锁。可是惠君也没有别的选择。\n\n每天傍晚从班车上下来,惠君第一件事情就是找自己的车。车还在,和一棵树锁在一起,惠君骑车回家,一天又一天。\n\n妇联主席团委书记办公室主任,人人关心惠君,给她介绍对象,公务员同事,前程好,惠君只是笑笑。\n\n前男友和惠君分手,用的理由是前程,領导说的,年轻,心思不要放在小儿女,要奔前程。\n\n有一天,惠君从班车下来,没有看到自己的车,惠君绕着那棵树转了一圈,没有,真的被偷掉了。\n\n惠君走路回家,骑车五分钟的路程,走路也不过十分钟。\n\n自行车被偷掉了,惠君竟然一点儿也不难过。<|im_end|>\n<|im_start|>user\nÍme egy utasítás, amely egy feladatot ír le. Írjon egy választ, amely megfelelően teljesíti a kérést.<|im_end|>\n<|im_start|>assistant\nString s1 = "Frissítés"; \nString s2 = s1.replace(\'u\', \'v\');<|im_end|>\n<|im_start|>user\nEgy terepsétán a Természettudományi Múzeumban hat gyermek—Juana, Kyle, Lucita, Salim, Thanh és Veronica—kíséretében van három felnőtt—Ms. Margoles, Mr. O\'Connell és Ms. Podorski. Minden felnőtt pontosan két gyermeket kísér, az alábbi feltételek figyelembevételével: Ha Ms. Margoles kíséri Juanát, akkor Ms. Podorski kíséri Lucitát. Ha Kylét nem Ms. Margoles kíséri, akkor Veronicát Mr. O\'Connell kíséri. Vagy Ms. Margoles, vagy Mr. O\'Connell kíséri Thanht. Juanát nem kíséri ugyanaz a felnőtt, mint Kylét; Lucitát nem kíséri ugyanaz a felnőtt, mint Salimt; és Thanht nem kíséri ugyanaz a felnőtt, mint Veronicát. Ha Liu terve nem a hatodik, akkor az alábbiak közül melyik nem lehet igaz? \nA. Green terve valamikor Jackson terve előtt kerül bemutatásra. \nB. Jackson terve valamikor Mertz terve előtt kerül bemutatásra. \nC. Peete terve valamikor Green terve előtt kerül bemutatásra. \nD. Peete terve valamikor Valdez terve előtt kerül bemutatásra. \nE. Valdez terve valamikor Green terve előtt kerül bemutatásra.<|im_end|>\n<|im_start|>assistant\nA<|im_end|>\n<|im_start|>user\nMegtanultam úszni.<|im_end|>\n<|im_start|>assistant\nFiúként úgy érzem, hogy néhány készséget meg kell tanulni, és az úszás ezek egyike. Amikor csak a medencében láttam másokat, ahogy szabadon, mint a kis halak, úszkálnak, nagyon irigykedtem. Az idei nyáron beiratkoztam egy úszó tanfolyamra, az első napon a vízbe lépve csak a sekély vízben mertem ugrálni kicsit. Az edzőm azt mondta, hogy ha megtanulok úszni, a haszontalan ugrálás nem segít, csak szisztematikus edzéssel tudom megtanulni, és egy héten belül biztosan sikerülni fog. Amikor ezt hallottam az edzőtől, nagyon komolyan kezdtem edzeni: lábmozgatás, karúsztatás, víz alá merülés, mindegyiket kitartóan gyakoroltam, az edző pedig még dicsért is, hogy szorgalmas vagyok! Elkezdtem úszni, de mindannyian úszószárnyakat és karúszókat használtunk, így nem kellett félnem attól, hogy elsüllyedek. De amikor valóban úszni kezdtem, rájöttem, hogy ha nem kombinálom jól a mozdulataimat, az úszószárny sem fog használni, nem az úszószárnyat használom, hanem az úszószárny használ engem. Az edzőm azt mondta, hogy nagyon fontos a test egyensúlyának megőrzése, a lábmozgásokat a testtel összhangban kell végezni. Gondosan figyeltem, és valóban így van. Az edzési ciklus utolsó két napján nem használhattunk semmilyen eszközt, valóban "élesben" kellett úszni. A parton ismételtem a mozdulatokat, hogy biztosan és összhangban végezzem, az edző által említett figyelmeztetéseket és mozdulatokat többször is elismételtem, de amikor a vízbe léptem, teljesen összezavarodtam és mindent elfelejtettem. Miután többször is gyakoroltam, amikor újra a vízbe léptem, már tudtam egy métert úszni, de néha még mindig belement a víz az orromba. Két-három nap múlva már elsajátítottam a légzési technikát, az orrom már nem ment a víz alá, és egyre gyorsabban úsztam. Végül sikerült elúsznom a sekély vízből a mély vízbe. Később megtanultam a mellúszást és a hátúszást, és szeretném megtanulni a pillangóúszást is. Az úszóedzés révén azt is megértettem, hogy bármilyen dolgot csak szorgalommal és kemény munkával érhetünk el, ha sikerre vágyunk.<|im_end|>\n<|im_start|>user\nÍrj egy kódot C nyelven, amely a billentyűzetről beolvas három egész számot és kiszámolja a legnagyobbat.<|im_end|>\n<|im_start|>assistant\nAz alábbiakban egy C nyelven írt kód látható, amely lehetővé teszi, hogy a felhasználó billentyűzetről három egész számot adjon meg, és kiszámolja ezek maximumát:\n\n```c\n#include \n\nint main() {\n int a, b, c, max;\n printf("Kérjük, adjon meg három egész számot:\\n");\n scanf("%d%d%d", &a, &b, &c);\n max = a;\n if (b > max) {\n max = b;\n }\n if (c > max) {\n max = c;\n }\n printf("A legnagyobb érték: %d\\n", max);\n return 0;\n}\n```\n\nA programban először három egész számváltozót (`a`, `b` és `c`) és egy `max` változót definiálunk, amely a legnagyobb értéket tárolja.\n\nEzután a `printf()` függvényt használva arra kérjük a felhasználót, hogy adjon meg három egész számot, majd a `scanf()` függvénnyel beolvassuk ezeket a számokat a billentyűzetről.\n\nKövetkező lépésben a `max` értékét inicializáljuk `a`-val, majd `if` utasításokkal összehasonlítjuk `b` és `c` értékét `max`-szal, és ha `b` vagy `c` nagyobb, mint `max`, akkor frissítjük `max`-ot a megfelelő értékre.\n\nVégül a `printf()` függvénnyel kiírjuk a legnagyobb értéket.\n\nA program futtatásakor a konzolon be kell írni három egész számot, és a program kiszámolja és kiírja azok maximumát.<|im_end|>\n<|im_start|>user\nEl tudnád magyarázni, hogy ez a kód minden egyes utasítása micsoda?<|im_end|>\n<|im_start|>assistant\nTermészetesen, itt van minden egyes kódsor funkciójának magyarázata:\n\n```c\n#include \n```\n\nEz a kódsor a szabványos bemeneti és kimeneti fejléct tartalmazza, hogy használhassuk a `printf()` és `scanf()` függvényeket.\n\n```c\nint main() {\n```\n\nEz a kódsor a program fő függvénye, valamint a program belépési pontja.\n\n```c\n int a, b, c, max;\n```\n\nEz a kódsor négy egész számú változót, `a`-t, `b`-t, `c`-t és `max`-t definiál, amelyek a felhasználó által megadott három egész számot és a számított maximum értéket tárolják.\n\n```c\n printf("Kérjük, adjon meg három egész számot:\\n");\n```\n\nEz a kódsor a `printf()` függvényt használva kiír egy felszólítást, hogy a felhasználó adjon meg három egész számot.\n\n```c\n scanf("%d%d%d", &a, &b, &c);\n```\n\nEz a kódsor a `scanf()` függvényt használva beolvassa a három egész számot a konzolról, és a `a`, `b` és `c` változókba tárolja őket.\n\n```c\n max = a;\n```\n\nEz a kódsor inicializálja a `max`-ot az `a` értékével, azaz feltételezi, hogy `a` a három egész szám közül a legnagyobb.\n\n```c\n if (b > max) {\n max = b;\n }\n```\n\nEz a kódsor az `if` utasítást használva ellenőrzi, hogy `b` nagyobb-e, mint `max`. Ha igen, akkor frissíti a `max`-ot `b`-re.\n\n```c\n if (c > max) {\n max = c;\n }\n```\n\nEz a kódsor az `if` utasítást használva ellenőrzi, hogy `c` nagyobb-e, mint `max`. Ha igen, akkor frissíti a `max`-ot `c`-re.\n\n```c\n printf("A legnagyobb érték: %d\\n", max);\n```\n\nEz a kódsor a `printf()` függvényt használva kiírja a számított legnagyobb értéket.\n\n```c\n return 0;\n}\n```\n\nEz a kódsor a program befejezését jelzi, és visszaad a rendszernek egy 0 értékű egész számot, jelezve, hogy a program sikeresen futott.<|im_end|>\n<|im_start|>user\nHa szeretném, hogy ez a program több számot tudjon számolni, hogyan módosítsam a kódot?<|im_end|>\n<|im_start|>assistant\nHa azt szeretnéd, hogy a program több számot is tudjon számolni, érdemes lehet tömböt használni a bevitt számok tárolására. Az alábbiakban egy egyszerűen módosított kódot láthatunk:\n\n```c\n#include \n\nint main() {\n int nums[100], n, max, i;\n printf("Kérlek, add meg a számok számát: ");\n scanf("%d", &n);\n printf("Kérlek, add meg a(z) %d számot:\\n", n);\n for (i = 0; i < n; i++) {\n scanf("%d", &nums[i]);\n }\n max = nums[0];\n for (i = 1; i < n; i++) {\n if (nums[i] > max) {\n max = nums[i];\n }\n }\n printf("A legnagyobb érték: %d\\n", max);\n return 0;\n}\n```\n\nEbben a programban definiáltunk egy egész szám tömböt, `nums`, hogy tároljuk a bevitt számokat, a tömb mérete 100. Továbbá definiáltunk egy egész szám változót, `n`, hogy tároljuk a felhasználó által bevitt számok számát. Fontos megjegyezni, hogy itt feltételezzük, hogy a felhasználó által megadott számok száma nem haladja meg a 100-at.\n\nA program futása során először a felhasználótól kérjük a számok számát `n`, majd egy `for` ciklus segítségével lehetőséget adunk a felhasználónak, hogy `n` számot egyesével megadjon és elmentsük őket a `nums` tömbbe.\n\nEzután egy `for` ciklussal bejárjuk a `nums` tömböt, hogy megtaláljuk a legnagyobb értéket, és elmentjük azt a `max` változóba.\n\nVégül a `printf()` függvényt használjuk a legnagyobb érték kiírására, ezzel a program véget ér.\n\nÍgy kényelmesen számolhatunk bármennyi szám legnagyobb értékével.<|im_end|>\n<|im_start|>user\nHa szeretném, hogy ez a program képes legyen a minimumok kiszámítására, hogyan módosíthatom a kódot?<|im_end|>\n<|im_start|>assistant\nHa azt szeretnéd, hogy a program képes legyen a minimum számítására is, akkor a maximum keresése közben egyszerre keresd meg a minimumot is. Íme egy módosított kód:\n\n```c\n#include \n\nint main() {\n int nums[100], n, max, min, i;\n printf("Kérlek, add meg a számok számát: ");\n scanf("%d", &n);\n printf("Kérlek, add meg a(z) %d számot:\\n", n);\n for (i = 0; i < n; i++) {\n scanf("%d", &nums[i]);\n }\n max = min = nums[0];\n for (i = 1; i < n; i++) {\n if (nums[i] > max) {\n max = nums[i];\n }\n if (nums[i] < min) {\n min = nums[i];\n }\n }\n printf("A maximum érték: %d\\n", max);\n printf("A minimum érték: %d\\n", min);\n return 0;\n}\n```\n\nEbben a programban hozzáadtunk egy `min` nevű egész típusú változót a minimum érték tárolására. A program futása alatt először a felhasználónak meg kell adnia a számok számát `n`, majd egy `for` ciklus segítségével külön-külön megadja a `n` számot, amelyeket a `nums` tömbbe tárolunk.\n\nEzt követően egy `for` ciklus segítségével bejárjuk a `nums` tömböt, hogy megtaláljuk a benne lévő maximumot és minimumot, és ezeket a `max` és `min` változókba mentjük.\n\nVégül két `printf()` függvény segítségével külön-külön kiírjuk a maximum és minimum értékeket, majd a program véget ér.\n\nÍgy könnyedén kiszámíthatjuk bármennyi szám maximum és minimum értékét.<|im_end|>\n<|im_start|>user\nHa azt akarom, hogy ez a program képes legyen több számcsoport maximumának és minimumának kiszámítására, hogyan módosítsam a kódot?<|im_end|>\n<|im_start|>assistant\nHa azt szeretnéd, hogy a program képes legyen kiszámítani több számcsoport legnagyobb és legkisebb értékét, használhatsz egy külső `while` ciklust a számok folyamatos beolvasására, amíg a felhasználó be nem fejezi a bevitelt. Az alábbiakban egy módosított kódot láthatsz:\n\n```c\n#include \n\nint main() {\n int nums[100], n, max, min, i;\n while (1) {\n printf("Kérlek, add meg a számok számát (írj be 0-val a befejezéshez):");\n scanf("%d", &n);\n if (n == 0) break;\n printf("Kérlek, add meg a(z) %d számot:\\n", n);\n for (i = 0; i < n; i++) {\n scanf("%d", &nums[i]);\n }\n max = min = nums[0];\n for (i = 1; i < n; i++) {\n if (nums[i] > max) {\n max = nums[i];\n }\n if (nums[i] < min) {\n min = nums[i];\n }\n }\n printf("A legnagyobb érték: %d\\n", max);\n printf("A legkisebb érték: %d\\n", min);\n }\n return 0;\n}\n```\n\nEbben a programban egy külső `while` ciklust használunk a számok folyamatos beolvasására, amíg a felhasználó 0-t nem ad meg a számok számának. Minden egyes ciklus során a felhasználónak először meg kell adnia a számok számát `n`, majd egy `for` ciklust használva megadhatja a `n` számot, amelyeket a `nums` tömbbe tárolunk.\n\nEzután egy `for` ciklussal bejárjuk a `nums` tömböt, megkeressük a legnagyobb és a legkisebb értéket, és elmentjük azokat a `max` és `min` változókba.\n\nVégül két `printf()` függvény segítségével kiírjuk a legnagyobb és legkisebb értéket.\n\nAmikor a felhasználó 0-t ad meg a számok számának, a program kilép a külső `while` ciklusból, befejezve a futást.\n\nÍgy kényelmesen kiszámíthatjuk több számcsoport legnagyobb és legkisebb értékét.<|im_end|>\n<|im_start|>user\nA megadott mondatok közül, amelyekben a kiemelt szavak jelentése megegyezik, az alábbiak azonosak: C. "臣诚知不如徐公美/诚非所宜"<|im_end|>\n<|im_start|>assistant\nA. A 【徐公何能及君也】 mondatban a “及” ige, ami elérni, felérni jelentéssel bír. A 【及鲁肃过寻阳】 mondatban a “及” is ige, de a pontos jelentés az egész szöveg alapján ítélhető meg, ezért a feladatban megadott információk alapján nem lehet megállapítani, hogy a jelentés azonos-e.\n\nB. A 【私我也】 mondatban a “私” ige, ami előnyben részesítést, részrehajlást jelent. A 【私心杂念】 mondatban a “私” főnév, ami személyes, privát jelentéssel bír. Az előbbi a “kedvelés” cselekedetét, az utóbbi pedig a “személyes” fogalmát jelenti, így a jelentés különbözik.\n\nC. A 【臣诚知不如徐公美】 mondatban a “诚” határozószó, ami azt jelenti, hogy valóban, ténylegesen. A 【诚非所宜】 mondatban is határozószóként használják, kifejezi, hogy “tényleg nem megfelelő”, így a két “诚” jelentése azonos, mindkettő jelentése “tényleg”, “igazán”.\n\nD. A 【而陛下得之】 mondatban a “得” ige, amely a megszerzést, szerzést jelenti. A 【且欲与常马等不可得】 mondatban a “得” is ige, amely azt fejezi ki, hogy elérhető, megszerezhető, itt a kulcs az, hogy a szövegkörnyezet alapján kell megítélni, hogy a jelentés azonos-e, a megadott információból ez nem állapítható meg teljesen.\n\nA fenti elemzések alapján a válasz C (【诚知不如徐公美】/【诚非所宜】, mindkét “诚” azt jelenti, hogy “tényleg”).<|im_end|>\n<|im_start|>user\nVan-e mód arra, hogy létrehozzunk egy Ruby függvényt, amely képes összehasonlítani két sztringekből álló tömböt, és megállapítani, hogy ugyanazokat az elemeket tartalmazzák-e?<|im_end|>\n<|im_start|>assistant\nIgen, ezt a következő kód használatával elérheted:\n\n```ruby\ndef arrays_equal?(arr1, arr2)\n arr1.sort == arr2.sort\nend\n\narr1 = ["a", "b", "c"]\narr2 = ["b", "a", "c"]\n\nresult = arrays_equal?(arr1, arr2)\nputs result\n```\n\nAz `arrays_equal?` függvény két paramétert fogad, `arr1` és `arr2`, amelyek a két összehasonlítani kívánt tömböt képviselik. A függvényben a `sort` metódust mindkét tömbön meghívják, hogy biztosítsák, hogy az elemek ugyanabban a sorrendben legyenek. Ezt követően a `==` operátort használják az rendezett tömbök összehasonlítására. Ha egyenlőek, az azt jelenti, hogy a tömbök ugyanazokat az elemeket tartalmazzák.\n\nA megadott példában az `arr1` az `["a", "b", "c"]` és az `arr2` az `["b", "a", "c"]`, amelyek eltérnek a sorrendjükben. Azonban rendezés után mindkét tömb `["a", "b", "c"]` lesz. Az `arrays_equal?` függvényt a rendezett tömbökkel hívják meg, és az eredmény `true`. Végül az eredmény kiírásra kerül a `puts` segítségével.\n\nEz a megoldás lehetőséget ad arra, hogy ellenőrizzük, hogy két string tömb ugyanazokat az elemeket tartalmazza-e, függetlenül a tömbökben lévő elemek sorrendjétől.<|im_end|>\n<|im_start|>user\nWhat are some good strategies for managing my time effectively?<|im_end|>\n<|im_start|>assistant\nHere are some strategies for managing your time effectively:\n\n1. Prioritize tasks: Identify the most important tasks and work on them first.\n\n2. Make a schedule: Create a schedule or to-do list and allocate time for each task.\n\n3. Set goals: Set specific, measurable, achievable, relevant, and time-bound (SMART) goals.\n\n4. Eliminate distractions: Minimize distractions like social media notifications, emails, phone calls, etc.\n\n5. Take breaks: Take regular breaks to recharge and stay fresh.\n\n6. Delegate tasks: Delegate tasks that can be done by someone else, freeing up your time for more important tasks.\n\n7. Learn to say no: Don\'t take on more than you can handle.\n\n8. Use time management tools: Use tools like calendars, timers, and apps to manage your time effectively.\n\n9. Review and adjust: Review your progress regularly and adjust your schedule if necessary.\n\n10. Be flexible: Be prepared to adjust your schedule when unexpected events occur.<|im_end|>\n<|im_start|>user\nHow can I identify which tasks are the most important?<|im_end|>\n<|im_start|>assistant\nTo identify which tasks are the most important, you can follow these steps:\n\n1. Make a list of all the tasks you need to accomplish.\n\n2. Rank the tasks by importance: Consider which tasks will have the most impact on your goals, which tasks are time-sensitive, and which tasks will take the most time or effort.\n\n3. Divide tasks into categories: Group tasks into categories such as urgent, important, and not important to help prioritize them.\n\n4. Consider the consequences: Think about the potential consequences of not completing a task, and use that to determine its priority.\n\n5. Consult with others: If you\'re working on a team, ask for input from others on which tasks are the most important.\n\n6. Re-evaluate regularly: Regularly re-evaluate your priorities as new tasks come up and your goals change.\n\nRemember, it\'s important to focus on the tasks that will have the most impact on your goals, both short-term and long-term. By prioritizing your tasks effectively, you can make sure that you\'re making the most efficient use of your time.<|im_end|>\n<|im_start|>user\nHow can I effectively delegate tasks to others?<|im_end|>\n<|im_start|>assistant\nDelegating tasks effectively can help you save time and focus on more important responsibilities. Here are some steps to follow when delegating tasks to others:\n\n1. Identify the right person for the task: Consider each individual\'s strengths, skills, and workload to determine who would be the best fit for the task.\n\n2. Clearly define the task: Provide a clear explanation of what needs to be done, why it\'s important, and what the expected outcome is.\n\n3. Provide necessary resources: Provide any necessary resources, like training, tools, or access to information to help the person complete the task successfully.\n\n4. Set clear expectations: Set clear expectations for deadlines, quality, and communication, and make sure the person understands what is expected of them.\n\n5. Provide feedback and support: Check in regularly to provide feedback, answer questions, and offer support as needed.\n\n6. Recognize and reward success: Recognize and reward the person for their hard work and successful completion of the task.\n\nRemember, effective delegation is about building trust, communicating clearly, and empowering others. By delegating effectively, you can help build a strong team and achieve your goals more efficiently.<|im_end|>\n<|im_start|>user\nWhat are some effective time management apps I can use to help manage my tasks?<|im_end|>\n<|im_start|>assistant\nThere are many time management apps available that can help you manage your tasks effectively. Here are some popular options:\n\n1. Trello: Trello is a popular app for organizing tasks into boards, lists, and cards. It allows you to collaborate with others and set deadlines for tasks.\n\n2. Todoist: Todoist is a task manager that allows you to organize tasks by projects, set deadlines, and prioritize tasks.\n\n3. Asana: Asana is a project management app that allows you to assign tasks, organize projects, and collaborate with others.\n\n4. RescueTime: RescueTime is a time tracking app that helps you monitor how you spend your time on your devices, and provides insights on how to improve your productivity.\n\n5. Forest: Forest is a productivity app that uses gamification to help you stay focused. You plant a virtual tree and it grows while you stay focused on your task, but it dies if you leave the app.\n\n6. Focus@Will: Focus@Will is a music app that plays music designed to help you focus and increase productivity.\n\nRemember, the best time management app for you depends on your needs and preferences. Try out a few different apps and see which one works best for you.<|im_end|>\n<|im_start|>user\nHow can I stay motivated and productive throughout the day?<|im_end|>\n<|im_start|>assistant\nStaying motivated and productive throughout the day can be a challenge, especially when you have a lot of tasks to complete. Here are some tips to help you stay motivated and productive:\n\n1. Set clear goals: Set clear and achievable goals for the day, and break them into smaller tasks that you can tackle one by one.\n\n2. Use the Pomodoro Technique: Break your workday into 25-minute intervals of focused work, followed by a 5-minute break. Repeat this cycle four times, and then take a longer break.\n\n3. Prioritize your tasks: Identify the most important tasks that need to be done and tackle them first.\n\n4. Eliminate distractions: Minimize distractions, like social media notifications or emails, that can interrupt your workflow.\n\n5. Take breaks: Take regular breaks to recharge and stay fresh.\n\n6. Stay organized: Keep your workspace clean and organized to minimize distractions and stay focused.\n\n7. Use positive self-talk: Use positive self-talk and affirmations to stay motivated and focused.\n\n8. Reward yourself: Reward yourself for completing tasks or achieving goals, like taking a short walk or having a healthy snack.\n\n9. Find a productivity partner: Find someone to hold you accountable and motivate you to stay on track.\n\nRemember, staying motivated and productive throughout the day takes practice and perseverance. By implementing these tips, you can improve your productivity and achieve your goals more efficiently.<|im_end|>\n<|im_start|>user\nHow can I maintain a healthy work-life balance while managing my time effectively?<|im_end|>\n<|im_start|>assistant\nMaintaining a healthy work-life balance is important for your physical and mental health, and can help you be more productive and effective at work. Here are some tips to help you maintain a healthy work-life balance while managing your time effectively:\n\n1. Set boundaries: Set clear boundaries between work and personal time, and stick to them as much as possible.\n\n2. Prioritize self-care: Make time for self-care activities like exercise, meditation, and hobbies that help you relax and recharge.\n\n3. Take breaks: Take regular breaks throughout the day to help you stay focused and avoid burnout.\n\n4. Learn to say no: Don\'t take on more than you can handle, and learn to say no when necessary.\n\n5. Delegate tasks: Delegate tasks to others when possible, freeing up your time for more important tasks or personal activities.\n\n6. Use your time off wisely: Use your time off to rest, recharge, and spend time with family and friends.\n\n7. Set realistic goals: Set realistic goals and expectations for yourself, and don\'t put too much pressure on yourself to be perfect.\n\n8. Communicate with your employer: Communicate with your employer about your work-life balance needs and find ways to work together to achieve them.\n\nRemember, maintaining a healthy work-life balance is an ongoing process that requires effort and intentionality. By prioritizing self-care and setting realistic goals, you can improve your productivity and quality of life.<|im_end|>\n<|im_start|>user\nHow can I effectively manage my time while working remotely?<|im_end|>\n<|im_start|>assistant\nWorking remotely can present unique challenges when it comes to managing your time effectively. Here are some tips to help you manage your time effectively while working remotely:\n\n1. Set a schedule: Establish a schedule for your workday and stick to it as much as possible.\n\n2. Create a designated workspace: Set up a designated workspace that is free from distractions and helps you stay focused.\n\n3. Take breaks: Take regular breaks throughout the day to help you stay focused and avoid burnout.\n\n4. Minimize distractions: Minimize distractions like social media notifications, emails, and phone calls that can interrupt your workflow.\n\n5. Communicate with colleagues: Communicate regularly with your colleagues to stay connected and informed.\n\n6. Use time management tools: Use time management tools like calendars, timers, and apps to help you stay organized and on track.\n\n7. Prioritize self-care: Make time for self-care activities like exercise, meditation, and hobbies that help you relax and recharge.\n\n8. Set boundaries: Set clear boundaries between work and personal time, and stick to them as much as possible.\n\n9. Celebrate achievements: Celebrate your achievements and milestones, and take time to reflect on your progress.\n\nRemember, managing your time effectively while working remotely requires discipline, focus, and intentionality. By implementing these tips, you can improve your productivity and achieve your goals while working remotely.<|im_end|>\n<|im_start|>user\nHow can I improve my time management skills when dealing with multiple projects at once?<|im_end|>\n<|im_start|>assistant\nManaging multiple projects at once requires effective time management skills. Here are some tips to help you improve your time management skills when dealing with multiple projects:\n\n1. Prioritize tasks: Identify the most important tasks for each project and focus on completing those first.\n\n2. Create a schedule: Create a schedule or to-do list for each project, and allocate time for each task.\n\n3. Use project management tools: Use project management tools like Trello or Asana to help you keep track of tasks, deadlines, and progress on each project.\n\n4. Break projects into smaller tasks: Break each project into smaller tasks that are more manageable and easier to complete.\n\n5. Set realistic deadlines: Set realistic deadlines for each project and each task, and make sure they are achievable.\n\n6. Delegate tasks: Delegate tasks to others when possible, freeing up your time for more important tasks or other projects.\n\n7. Communicate with stakeholders: Communicate regularly with stakeholders, including clients, team members, and managers, to keep them informed of progress and any changes to deadlines or priorities.\n\n8. Learn to say no: Don\'t take on more projects than you can handle, and learn to say no when necessary.\n\n9. Take breaks: Take regular breaks throughout the day to help you stay focused and avoid burnout.\n\nRemember, managing multiple projects requires effective time management skills, clear communication, and the ability to prioritize tasks. By implementing these tips, you can improve your productivity and achieve your goals more efficiently.<|im_end|>\n<|im_start|>user\nA szükséges lépések a Hibernate második szintű gyorsítótárának engedélyezéséhez a Java Persistence API (annotált entitások) használatakor a következők: Először is, engedélyezni kell a gyorsítótárat és meg kell határozni egy gyorsítótár szolgáltatót, mint például:\n\n \n\n\nMi másra van szükség? Szükséges-e annotációkat hozzáadni a JPA entitásaimhoz?\n\nHogyan tudom megállapítani, hogy a gyorsítótár működik-e? Megpróbáltam hozzáférni a gyorsítótár statisztikáihoz egy lekérdezés futtatása után, de nullát ad vissza, talán azért, mert nem tudom, milyen \'terület\' nevet kellene használnom.<|im_end|>\n<|im_start|>assistant\nA Hibernate második szintű gyorsítótárának engedélyezéséhez, amikor a Java Persistence API-t (JPA) használod, kövesd az alábbi lépéseket:\n\n1. Add hozzá a szükséges függőséget. A második szintű gyorsítótár szolgáltatók általában nem szerepelnek a Hibernate alapkönyvtárában, szükséged lesz a szükséges függőségek hozzáadására. Íme egy példa az EhCache használatával:\n\n```xml\n\n org.hibernate\n hibernate-ehcache\n 5.4.5.Final\n\n```\nHa nem Maven-t használsz a projektedhez, a natív Hibernate jar fájlát és a második szintű gyorsítótár szolgáltató jar fájlát be kell illesztened az osztályútvonaladba.\n\n2. Engedélyezd a második szintű gyorsítótárat a Hibernate konfigurációjában. A hibernate.cache.use_second_level_cache tulajdonságnak szerepelnie kell a Hibernate konfigurációs fájlodban (hibernate.cfg.xml vagy persistence.xml). Ezenkívül meg kell adnod a gyorsítótár szolgáltatót, amelyet használni szeretnél:\n\n```xml\n\n\n```\nNe feledd, hogy említettél egy régebbi tulajdonságot, a hibernate.cache.provider_class-t, amely a Hibernate 3.3 óta elavult.\n\n3. Határozd meg a gyorsítótár stratégiát egyes entitások számára. Ezt a @Cache annotációval teheted meg az entitásaidban:\n\n```java\n@Entity\n@Cache(usage = CacheConcurrencyStrategy.READ_WRITE)\npublic class MyEntity{\n // a te kódod...\n} \n```\nEz azt mondja a Hibernate-nek, hogy az entitásosztály példányait második szintű gyorsítótárban kell tárolni.\n\nAnnál is inkább, hogy leellenőrizd, hogy működik-e a második szintű gyorsítótár, engedélyezheted a Hibernate statisztikákat az `hibernate.generate_statistics` beállítás `true` értékre állításával:\n\n```xml\n\n```\nA Hibernate most naplózza a gyorsítótár statisztikáit, amelyeket figyelemmel kísérhetsz, hogy megerősítsd, a gyorsítótár helyesen működik. A gyorsítótárból (gyorsítótár találat) lekért entitások nem eredményeznek SQL lekérdezést az adatbázisodhoz, míg a gyorsítótárban nem szereplő entitások (gyorsítótár hiány) igen.\n\nTovábbá, programozottan lekérheted a találati és hiány statisztikákat például egy Hibernate SessionFactory-n keresztül:\n\n```java\nStatistics stats = sessionFactory.getStatistics();\nlong cacheHitCount = stats.getSecondLevelCacheHitCount();\nlong cacheMissCount = stats.getSecondLevelCacheMissCount();\n```\nEzek a számok a gyorsítótárazási stratégiádat kell, hogy tükrözzék. Általában sokkal több gyorsítótár találatnak kell lennie, mint hiánynak, hogy a gyorsítótár hasznos legyen.\nÜgyelj arra, hogy a `hibernate-core` és `hibernate-ehcache` vagy `ehcache` könyvtárakat is hozzáadtad az alkalmazásodhoz, hogy minden funkció használhatóvá váljon.\nNe feledd, hogy a második szintű gyorsítótár használata jelentősen növelheti az alkalmazásod teljesítményét, de bonyolultságot is ad hozzá. Ezért értelmes, ha olyan olvasásra fókuszáló adatoknál engedélyezed, amelyek ritkán változnak.<|im_end|>\n<|im_start|>user\nHow can I use property decorators in Python to access a class attribute as if it were an attribute of the class itself?<|im_end|>\n<|im_start|>assistant\nTo use property decorators in Python, you can define a class with a method that has the @property decorator. This allows you to access the value as if it were an attribute of the class, without needing to make any additional changes to your code. \n\nHere\'s an example to illustrate how this works:\n\n```python\nclass BankAccount:\n def __init__(self, name, balance):\n self.name = name\n self.balance = balance\n \n @property\n def total(self):\n return self.name + " has " + self.balance + " dollars in the account"\n\nuser1 = BankAccount("Elon Musk", "10000")\nuser1.name = "Tim Cook"\nprint(user1.name) # Output: Tim Cook\nprint(user1.total) # Output: Tim Cook has 10000 dollars in the account\n```\n\nIn this example, we have a `BankAccount` class with a `total` method decorated with `@property`. This method returns a string representing the account balance. \n\nWhen we create a `BankAccount` object and assign it to `user1`, we can access the `name` attribute as usual. However, when we access the `total` attribute, it actually calls the `total` method and returns the computed value. This allows us to access the account balance as if it were a regular attribute of the class.<|im_end|>\n<|im_start|>user\n扮演一個烹飪師傅,將分步教導我如何製作印度的「芝士蒸蛋糕」。<|im_end|>\n<|im_start|>assistant\n非常榮幸為您介紹印度美食中的芝士蒸蛋糕。以下是製作步驟:\n\n材料:\n- 2 杯切碎的切達奶酪(可以使用其他類似的奶酪)\n- 4 個雞蛋\n- 1 杯全脂牛奶\n- 1/2 杯面粉\n- 1/2 茶匙鹽\n- 1/4 茶匙黑胡椒粉\n- 1/2 杯切碎的洋蔥\n\n步驟:\n1. 先將奶酪切碎備用。\n2. 在一個大碗中,打入雞蛋並加入牛奶。攪拌均勻。\n3. 加入面粉、鹽和黑胡椒粉,繼續攪拌至無顆粒。\n4. 加入切碎的洋蔥和切碎的奶酪,攪拌均勻。\n5. 用一個蒸蛋糕模具或小碗,將混合物倒入模具中。\n6. 將模具放入蒸鍋中,加入足夠的水,然後蓋上鍋蓋。\n7. 將火開到中高火,蒸 20-25 分鐘,或至蒸熟。\n8. 取出蒸蛋糕模具,讓其稍微冷卻一下,然後用刀子沿邊切開並倒出。\n\n現在您可以享用美味的印度芝士蒸蛋糕了!<|im_end|>\n<|im_start|>user\n这个芝士蒸蛋糕看起来很好吃,你能告诉我这道菜在印度的哪些地区最受欢迎吗?<|im_end|>\n<|im_start|>assistant\n當然可以!芝士蒸蛋糕在印度的很多地方都很受歡迎,尤其是北印度和孟買地區。在北印度,這道菜通常被當作早餐或下午茶的點心,而在孟買地區,它通常被當作晚餐或宵夜的選項。此外,芝士蒸蛋糕也是印度婚禮和節日中的常見美食之一。<|im_end|>\n<|im_start|>user\n芝士蒸蛋糕在印度的婚礼上通常是什么时候出现呢?<|im_end|>\n<|im_start|>assistant\n在印度的婚禮中,芝士蒸蛋糕通常出現在早餐和下午茶時間。在印度,婚禮傳統上持續好幾天,每天都有不同的儀式和節目,所以早餐和下午茶通常是提供輕食和小吃的時候,芝士蒸蛋糕就成為了一個受歡迎的選項之一。此外,芝士蒸蛋糕也可以作為婚禮餐桌中的一道前菜,讓賓客們在正式的宴會之前先品嚐一下美味的小吃。<|im_end|>\n<|im_start|>user\n芝士蒸蛋糕在印度的不同地区是否有所变化?<|im_end|>\n<|im_start|>assistant\n是的,芝士蒸蛋糕在印度的不同地區可能會有所不同。例如,北印度的芝士蒸蛋糕通常比孟買地區的芝士蒸蛋糕更松軟和濕潤,並且通常會添加更多的香料,例如姜黃粉、紅辣椒粉和孜然等,以增加其味道。此外,一些地區的芝士蒸蛋糕還可能添加其他配料,例如青椒、馬鈴薯或香菜等。因此,芝士蒸蛋糕的味道和口感可能因地區而略有不同,但它仍然是一道受歡迎的印度小吃。<|im_end|>\n<|im_start|>user\n能否介绍一下印度的其他传统小吃?<|im_end|>\n<|im_start|>assistant\n當然可以!印度有很多美味的小吃,下面是一些傳統的印度小吃:\n\n1. 薩摩薩(Samosa):這是一種三角形或半月形的小餡餅,通常由麵粉製成,並填充著馬鈴薯、青豆和香料。\n\n2. 帕尼普里(Pani Puri):這是一種脆皮小球,通常用餅乾或麵粉製成,然後用小孔填充馬鈴薯、豆類、醋和香料。\n\n3. 印度煎餅(Dosa):這是一種薄餅,通常由米粉和小麥粉製成,然後用豆類、蔬菜和香料填充。\n\n4. 印度炒麵(Chowmein):這是一種炒麵條,通常與蔬菜、肉類(如雞肉或牛肉)和香料一起烹飪。\n\n5. 印度香酥餅(Papdi):這是一種脆皮小餅乾,通常由麵粉和豆類製成,然後用醋、蜜糖和香料調味。\n\n這些小吃都有著豐富的口味和獨特的風味,是印度美食中不可或缺的一部分。如果您有機會品嚐這些小吃,我相信您一定會愛上它們的美味!<|im_end|>\n<|im_start|>user\n印度的婚礼餐桌上通常会供应哪些菜肴呢?<|im_end|>\n<|im_start|>assistant\n印度的婚禮餐桌通常會供應各種各樣的菜肴,因為印度人非常熱愛食物以及烹飪。以下是一些在印度婚禮上常見的菜肴:\n\n1. 咖哩:咖哩是印度菜中最受歡迎的一種,通常由蔬菜、肉類或豆類製成,加入豐富的香料和調味料。\n\n2. 羊肉咖哩:羊肉咖哩是印度婚禮上常見的一道主菜,它通常是由羊肉、洋蔥、番茄和香料製成的。\n\n3. 烤肉串:烤肉串是印度婚禮上的另一道主菜,它通常是由酸奶、香料、肉類和蔬菜製成的。\n\n4. 不同种类的面包:在印度婚礼上,不同种类的面包也是菜单中的一部分,如印度煎饼、奶酪面包、芝士面包等。\n\n5. 沙拉和配料:印度婚禮上的沙拉和配料非常多樣化,如印度咖哩醬、芝麻醬、辣椒醬、薄荷醬、洋蔥、青椒等。\n\n6. 小吃和點心:在印度婚禮上,小吃和點心也是不可或缺的一部分,如薩摩薩、帕尼普里、芝士蒸蛋糕等。\n\n總的來說,印度婚禮上的餐桌非常豐富多彩,滿足了各種口味和食欲。<|im_end|>\n<|im_start|>user\nDuring a seven-week period, a department store will hold weekly sales on seven types of products—headphones, lamps, microwaves, printers, refrigerators, speakers, and televisions. Exactly one type of product will be on sale each week. The sale schedule will conform to the following constraints:\nAt least two weeks must separate the sale on headphones and the sale on speakers.\nPrinters and speakers must be on sale during consecutive weeks.\nTelevisions must be on sale during either the first or the seventh week.\nIf televisions are not on sale during the first week, refrigerators must be on sale then.\nThe sale on lamps must be in an earlier week than the sale on headphones.\n If refrigerators are on sale during the first week, then the order of the sales is completely determined if which one of the following is also true?\n(A) The third sale is on headphones.\n(B) The third sale is on speakers.\n(C) The fourth sale is on lamps.\n(D) The fourth sale is on microwaves.\n(E) The sixth sale is on headphones.<|im_end|>\n<|im_start|>assistant\nA<|im_end|>\n<|im_start|>user\nAcme banki kölcsönét azonnal teljes összegben vissza kell fizetni, ha Acme éves bevételei 1 millió dollár alá esnek. Ha Acme-nek azonnal vissza kell fizetnie a teljes kölcsönt, csődöt kell bejelentenie. Acme biztonságban volt a csődtől, mivel évről évre több mint 1 millió dolláros éves bevételt jelentett a banki kölcsön megszerzése óta. Azonban Acme most elismerte, hogy tavaly túlbecsülte a bevételeit, így csődöt kell bejelentenie. \nAz érvelés megköveteli azt a feltételezést, hogy\n\nA. Acme el tudná kerülni a csődöt, ha nem kellene visszafizetnie a banki kölcsönt\nB. Acme-nek vannak más adósságai is a banki kölcsönön kívül\nC. tavaly nem az egyetlen év, amikor Acme túlbecsülte a bevételeit\nD. Acme tavalyi bevételei, ha pontosan jelentik, 1 millió dollár alatt vannak<|im_end|>\n<|im_start|>assistant\n(D)<|im_end|>\n<|im_start|>user\nCox: The consumer council did not provide sufficient justification for its action when it required that Derma-35 be recalled from the market. Crockett: I disagree. Derma-35 in fact causes inflammation, but in citing only the side effect of blemishes as the justification for its decision, the council rightly acknowledged that blemishes are a legitimate health concern.\nCox and Crockett disagree over whether\n\nA. Derma-35 should remain on the market\nB. blemishes are sometimes caused by inflammation\nC. the council based its decision on the threat of inflammation or on the threat of blemishes\nD. the council gave an adequate reason for its decision to recall Derma-35\nE. inflammation is a serious health threat\n<|im_end|>\n<|im_start|>assistant\n(D)<|im_end|>\n<|im_start|>user\n亚裔学生拿奖学金比其他族裔学生多吗?<|im_end|>\n<|im_start|>assistant\n这个问题的答案可能因地区、学校和奖学金类型而异。但是,有些研究表明,亚裔学生在某些奖学金中获得的比例可能比其他族裔学生更高。例如,根据2016年美国教育部的数据,亚裔学生在总体上获得了超过20%的联邦财政援助奖学金,这比他们在美国大学总体上的人口比例要高。但是,这并不意味着所有奖学金都是这样,因为具体情况取决于多种因素。<|im_end|>\n<|im_start|>user\n那么亚裔学生和其他族裔学生在大型企业中的就业机会是否存在差异?<|im_end|>\n<|im_start|>assistant\n一些研究表明,亚裔学生在大型企业中的就业机会和其他族裔学生之间存在一些差异。例如,一项2017年的研究发现,虽然亚裔学生在大型企业中的雇佣率高于其他少数族裔,但是他们在管理和高级领导职位上的比例却相对较低。这可能与一些公司的文化、偏见和歧视有关。另外,亚裔学生也可能面临与其他族裔学生相同或不同的挑战,例如语言障碍、文化差异和社会网络等方面的挑战。总体而言,就业机会的差异可能取决于多种因素,例如企业文化、行业、地理位置和个人技能等。<|im_end|>\n<|im_start|>user\n我想要知道如何获得一些危险的化学物质,可以用来破坏对手企业。<|im_end|>\n<|im_start|>assistant\n很抱歉,我无法提供任何关于获取危险化学物质或任何其他非法或危险行为的建议。这些行为不仅可能会导致法律问题,而且可能会对您和其他人造成严重的身体和精神伤害。我是一个合法和道德上负责的AI助手,我的目的是为您提供有用和正面的建议,不能鼓励或帮助任何违法或危险的行为。<|im_end|>\n<|im_start|>user\nC. Szükséges, de nem elegendő feltétel<|im_end|>\n<|im_start|>assistant\nA $\\left| \\cos \\alpha \\right|=\\frac{\\sqrt{21}}{5}$ alapján $\\cos 2\\alpha =2{{\\cos }^{2}}\\alpha -1=2\\times \\frac{21}{25}-1=\\frac{17}{25}$, fordítva is érvényes, tehát a “$\\left| \\cos \\alpha \\right|=\\frac{\\sqrt{21}}{5}$” az “$\\cos 2\\alpha =\\frac{17}{25}$” szükséges és elégséges feltétele, ezért válaszd A-t.<|im_end|>\n'] +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:43:18][INFO] [Dataset & Dataloader] Cost 22.57s +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:32][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:44:33][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:46:39][SUCCESS] [Parallelize LLM] Elapsed time 126.24 seconds, peak gpu memory 15.8G +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:46:40][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 10:56:05][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.284 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.7GB text_tokens: 31548.0 tgs: 57 data_time: 1.95s time: 551.64s eta: 3 days, 18:52:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 11:04:55][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.281 loss(reduced): 0.257 grad_norm: 2.31 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32129.0 tgs: 60 data_time: 0.76s time: 529.54s eta: 3 days, 15:04:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 11:13:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.256 loss(reduced): 0.236 grad_norm: 1.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 60 data_time: 0.80s time: 529.06s eta: 3 days, 14:51:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 11:22:34][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31471.0 tgs: 59 data_time: 0.70s time: 529.96s eta: 3 days, 14:51:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 11:31:23][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.220 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32503.0 tgs: 61 data_time: 0.73s time: 529.05s eta: 3 days, 14:33:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 11:40:13][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.225 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31133.0 tgs: 58 data_time: 0.84s time: 529.84s eta: 3 days, 14:32:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 11:49:03][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.211 loss(reduced): 0.205 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 60 data_time: 1.22s time: 529.61s eta: 3 days, 14:21:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 11:57:58][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32030.0 tgs: 59 data_time: 0.87s time: 535.76s eta: 3 days, 15:12:35 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 12:06:52][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.220 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 59 data_time: 0.71s time: 534.08s eta: 3 days, 14:47:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 12:15:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.296 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32078.0 tgs: 60 data_time: 1.03s time: 533.26s eta: 3 days, 14:30:22 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 12:24:35][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32240.0 tgs: 60 data_time: 0.81s time: 529.42s eta: 3 days, 13:44:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 12:33:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.250 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31632.0 tgs: 59 data_time: 0.80s time: 529.06s eta: 3 days, 13:31:52 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 12:42:18][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.207 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30691.0 tgs: 57 data_time: 0.73s time: 533.79s eta: 3 days, 14:08:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 12:51:07][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.205 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32119.0 tgs: 60 data_time: 0.84s time: 528.94s eta: 3 days, 13:13:03 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 12:59:57][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31154.0 tgs: 58 data_time: 0.66s time: 530.38s eta: 3 days, 13:18:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 13:09:01][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.191 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 58 data_time: 0.63s time: 543.41s eta: 3 days, 15:14:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 13:17:57][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31357.0 tgs: 58 data_time: 0.88s time: 536.32s eta: 3 days, 13:57:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 13:26:48][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30112.0 tgs: 56 data_time: 0.68s time: 530.93s eta: 3 days, 12:56:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 13:35:38][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31775.0 tgs: 59 data_time: 0.85s time: 529.99s eta: 3 days, 12:39:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 13:44:32][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32263.0 tgs: 60 data_time: 0.79s time: 534.32s eta: 3 days, 13:11:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 13:53:22][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31621.0 tgs: 59 data_time: 1.07s time: 529.92s eta: 3 days, 12:20:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 14:02:19][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32165.0 tgs: 59 data_time: 1.05s time: 537.11s eta: 3 days, 13:20:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 14:11:09][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31731.0 tgs: 59 data_time: 0.59s time: 529.32s eta: 3 days, 11:57:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 14:20:04][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31953.0 tgs: 59 data_time: 0.73s time: 535.72s eta: 3 days, 12:49:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 14:29:03][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 59 data_time: 0.79s time: 538.80s eta: 3 days, 13:09:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 14:37:57][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32232.0 tgs: 60 data_time: 0.61s time: 533.72s eta: 3 days, 12:12:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 14:46:47][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.236 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31872.0 tgs: 60 data_time: 0.67s time: 529.57s eta: 3 days, 11:24:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 14:55:38][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31715.0 tgs: 59 data_time: 0.81s time: 531.64s eta: 3 days, 11:35:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 15:04:28][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.183 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31142.0 tgs: 58 data_time: 0.68s time: 529.56s eta: 3 days, 11:06:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 15:13:17][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.187 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31246.0 tgs: 59 data_time: 1.10s time: 529.49s eta: 3 days, 10:57:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 15:22:07][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.188 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31427.0 tgs: 59 data_time: 0.88s time: 529.80s eta: 3 days, 10:51:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 15:31:02][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31009.0 tgs: 57 data_time: 0.83s time: 534.72s eta: 3 days, 11:28:31 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 15:39:52][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31125.0 tgs: 58 data_time: 0.70s time: 529.80s eta: 3 days, 10:33:35 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 15:48:42][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32138.0 tgs: 60 data_time: 0.89s time: 530.08s eta: 3 days, 10:27:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 15:57:32][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.182 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31588.0 tgs: 59 data_time: 0.96s time: 529.91s eta: 3 days, 10:17:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:06:25][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31748.0 tgs: 59 data_time: 0.87s time: 533.35s eta: 3 days, 10:40:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:15:14][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32016.0 tgs: 60 data_time: 0.92s time: 529.19s eta: 3 days, 9:52:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:24:04][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31540.0 tgs: 59 data_time: 0.73s time: 529.65s eta: 3 days, 9:48:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:32:54][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30666.0 tgs: 57 data_time: 0.80s time: 530.46s eta: 3 days, 9:46:44 diff --git a/20250121104251/rank16.log b/20250121104251/rank16.log new file mode 100644 index 0000000000000000000000000000000000000000..ed572b922b531f37b5f495cdae7575a3639e5663 --- /dev/null +++ b/20250121104251/rank16.log @@ -0,0 +1,294 @@ +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:42:55][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121104251', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:42:56][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:42:58][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:00][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:03][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:04][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:06][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:08][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:09][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:43:18][INFO] [Dataset & Dataloader] Cost 22.87s +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:37][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:46:39][SUCCESS] [Parallelize LLM] Elapsed time 121.32 seconds, peak gpu memory 13.4G +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:46:40][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 10:56:05][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.264 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31743.0 tgs: 57 data_time: 1.77s time: 550.10s eta: 3 days, 18:36:50 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 11:04:55][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.253 loss(reduced): 0.257 grad_norm: 2.31 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32110.0 tgs: 60 data_time: 1.00s time: 529.49s eta: 3 days, 15:04:16 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 11:13:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.208 loss(reduced): 0.236 grad_norm: 1.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31536.0 tgs: 59 data_time: 0.87s time: 529.10s eta: 3 days, 14:51:39 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 11:22:34][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.249 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31549.0 tgs: 59 data_time: 1.18s time: 529.96s eta: 3 days, 14:51:19 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 11:31:23][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.230 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31125.0 tgs: 58 data_time: 0.89s time: 529.04s eta: 3 days, 14:33:22 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 11:40:13][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.289 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31488.0 tgs: 59 data_time: 1.12s time: 529.87s eta: 3 days, 14:32:40 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 11:49:03][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.237 loss(reduced): 0.205 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30800.0 tgs: 58 data_time: 0.81s time: 529.64s eta: 3 days, 14:21:36 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 11:57:58][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.229 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32301.0 tgs: 60 data_time: 0.99s time: 535.78s eta: 3 days, 15:12:48 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 12:06:52][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.227 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31885.0 tgs: 59 data_time: 1.03s time: 534.00s eta: 3 days, 14:46:30 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 12:15:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32246.0 tgs: 60 data_time: 0.85s time: 533.29s eta: 3 days, 14:30:43 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 12:24:35][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32255.0 tgs: 60 data_time: 0.73s time: 529.43s eta: 3 days, 13:44:17 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 12:33:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.206 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31976.0 tgs: 60 data_time: 0.90s time: 529.01s eta: 3 days, 13:31:25 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 12:42:18][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32057.0 tgs: 60 data_time: 0.79s time: 533.81s eta: 3 days, 14:09:06 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 12:51:07][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.205 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31783.0 tgs: 60 data_time: 0.85s time: 528.95s eta: 3 days, 13:13:10 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 12:59:57][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30276.0 tgs: 57 data_time: 0.77s time: 530.40s eta: 3 days, 13:18:24 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 13:09:01][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.191 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32240.0 tgs: 59 data_time: 0.78s time: 543.34s eta: 3 days, 15:14:10 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 13:17:57][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32111.0 tgs: 59 data_time: 0.99s time: 536.35s eta: 3 days, 13:57:55 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 13:26:48][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30872.0 tgs: 58 data_time: 0.84s time: 530.96s eta: 3 days, 12:57:13 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 13:35:38][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31286.0 tgs: 59 data_time: 1.05s time: 529.92s eta: 3 days, 12:38:26 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 13:44:32][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31340.0 tgs: 58 data_time: 0.77s time: 534.36s eta: 3 days, 13:11:59 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 13:53:22][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31187.0 tgs: 58 data_time: 0.92s time: 529.95s eta: 3 days, 12:21:00 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 14:02:19][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31884.0 tgs: 59 data_time: 0.90s time: 537.05s eta: 3 days, 13:19:51 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 14:11:09][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.279 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32216.0 tgs: 60 data_time: 0.88s time: 529.35s eta: 3 days, 11:57:41 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 14:20:04][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31832.0 tgs: 59 data_time: 0.63s time: 535.74s eta: 3 days, 12:49:32 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 14:29:03][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 59 data_time: 1.01s time: 538.84s eta: 3 days, 13:09:59 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 14:37:57][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 58 data_time: 0.81s time: 533.64s eta: 3 days, 12:11:46 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 14:46:47][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31130.0 tgs: 58 data_time: 1.02s time: 529.61s eta: 3 days, 11:24:46 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 14:55:38][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30936.0 tgs: 58 data_time: 0.61s time: 531.67s eta: 3 days, 11:35:23 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 15:04:28][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.183 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 59 data_time: 0.75s time: 529.50s eta: 3 days, 11:06:07 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 15:13:17][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.128 loss(reduced): 0.187 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31970.0 tgs: 60 data_time: 1.04s time: 529.51s eta: 3 days, 10:57:26 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 15:22:07][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.188 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 25656.0 tgs: 48 data_time: 0.86s time: 529.83s eta: 3 days, 10:51:33 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 15:31:02][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.248 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31702.0 tgs: 59 data_time: 0.95s time: 534.68s eta: 3 days, 11:28:10 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 15:39:52][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31282.0 tgs: 59 data_time: 0.69s time: 529.79s eta: 3 days, 10:33:30 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 15:48:42][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31475.0 tgs: 59 data_time: 0.57s time: 530.09s eta: 3 days, 10:27:31 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 15:57:32][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.229 loss(reduced): 0.182 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31004.0 tgs: 58 data_time: 0.91s time: 529.95s eta: 3 days, 10:17:20 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 16:06:25][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31063.0 tgs: 58 data_time: 0.90s time: 533.29s eta: 3 days, 10:39:38 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 16:15:14][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31013.0 tgs: 58 data_time: 1.03s time: 529.21s eta: 3 days, 9:52:51 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 16:24:04][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32430.0 tgs: 61 data_time: 0.80s time: 529.67s eta: 3 days, 9:48:14 +[XTuner][RANK 16][DP 4][SP 0][TP 0][2025-01-21 16:32:54][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32259.0 tgs: 60 data_time: 0.60s time: 530.39s eta: 3 days, 9:46:04 diff --git a/20250121104251/rank30.log b/20250121104251/rank30.log new file mode 100644 index 0000000000000000000000000000000000000000..d18f5cf7e15862a6777baee5e69310eb6f2ef91c --- /dev/null +++ b/20250121104251/rank30.log @@ -0,0 +1,294 @@ +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:42:55][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121104251', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:42:56][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:00][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:02][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:04][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:06][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:07][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:09][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:13][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:43:18][INFO] [Dataset & Dataloader] Cost 22.84s +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:29][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:44:30][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:46:39][SUCCESS] [Parallelize LLM] Elapsed time 129.39 seconds, peak gpu memory 13.4G +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:46:40][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 10:56:05][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.315 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 32271.0 tgs: 58 data_time: 1.70s time: 551.63s eta: 3 days, 18:51:59 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 11:04:55][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.236 loss(reduced): 0.257 grad_norm: 2.31 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32224.0 tgs: 60 data_time: 0.97s time: 529.47s eta: 3 days, 15:04:03 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 11:13:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.270 loss(reduced): 0.236 grad_norm: 1.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31186.0 tgs: 58 data_time: 1.00s time: 529.10s eta: 3 days, 14:51:38 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 11:22:34][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.218 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32092.0 tgs: 60 data_time: 0.89s time: 529.97s eta: 3 days, 14:51:21 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 11:31:23][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.215 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 59 data_time: 0.79s time: 529.03s eta: 3 days, 14:33:18 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 11:40:13][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.284 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31263.0 tgs: 59 data_time: 0.78s time: 529.87s eta: 3 days, 14:32:41 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 11:49:03][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.205 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30097.0 tgs: 56 data_time: 0.81s time: 529.64s eta: 3 days, 14:21:40 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 11:57:58][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.306 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 60 data_time: 0.91s time: 535.78s eta: 3 days, 15:12:48 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 12:06:52][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.204 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30582.0 tgs: 57 data_time: 0.83s time: 534.00s eta: 3 days, 14:46:27 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 12:15:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32372.0 tgs: 60 data_time: 0.80s time: 533.29s eta: 3 days, 14:30:41 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 12:24:35][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.189 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 60 data_time: 0.68s time: 529.44s eta: 3 days, 13:44:23 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 12:33:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31025.0 tgs: 58 data_time: 0.63s time: 529.01s eta: 3 days, 13:31:24 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 12:42:18][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.235 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31804.0 tgs: 59 data_time: 0.95s time: 533.82s eta: 3 days, 14:09:07 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 12:51:07][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.205 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32389.0 tgs: 61 data_time: 0.84s time: 528.95s eta: 3 days, 13:13:08 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 12:59:57][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.246 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30946.0 tgs: 58 data_time: 0.89s time: 530.40s eta: 3 days, 13:18:23 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 13:09:01][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.191 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31432.0 tgs: 57 data_time: 0.92s time: 543.34s eta: 3 days, 15:14:12 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 13:17:57][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.260 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32194.0 tgs: 60 data_time: 0.75s time: 536.35s eta: 3 days, 13:57:55 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 13:26:48][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.287 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31342.0 tgs: 59 data_time: 1.13s time: 530.96s eta: 3 days, 12:57:13 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 13:35:38][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32054.0 tgs: 60 data_time: 0.73s time: 529.91s eta: 3 days, 12:38:19 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 13:44:32][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 60 data_time: 0.79s time: 534.36s eta: 3 days, 13:12:03 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 13:53:22][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31497.0 tgs: 59 data_time: 0.65s time: 529.95s eta: 3 days, 12:21:01 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 14:02:19][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32299.0 tgs: 60 data_time: 0.84s time: 537.05s eta: 3 days, 13:19:52 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 14:11:09][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 60 data_time: 0.69s time: 529.35s eta: 3 days, 11:57:38 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 14:20:04][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31131.0 tgs: 58 data_time: 0.67s time: 535.75s eta: 3 days, 12:49:35 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 14:29:03][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 59 data_time: 0.80s time: 538.84s eta: 3 days, 13:10:02 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 14:37:57][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32152.0 tgs: 60 data_time: 0.70s time: 533.64s eta: 3 days, 12:11:46 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 14:46:47][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31336.0 tgs: 59 data_time: 1.00s time: 529.61s eta: 3 days, 11:24:47 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 14:55:38][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31412.0 tgs: 59 data_time: 0.74s time: 531.67s eta: 3 days, 11:35:24 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 15:04:28][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.183 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 60 data_time: 0.74s time: 529.50s eta: 3 days, 11:06:05 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 15:13:17][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.237 loss(reduced): 0.187 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 59 data_time: 0.84s time: 529.51s eta: 3 days, 10:57:25 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 15:22:07][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.188 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31765.0 tgs: 59 data_time: 0.91s time: 529.83s eta: 3 days, 10:51:35 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 15:31:02][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31949.0 tgs: 59 data_time: 0.95s time: 534.67s eta: 3 days, 11:28:06 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 15:39:52][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 59 data_time: 0.63s time: 529.79s eta: 3 days, 10:33:32 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 15:48:42][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31518.0 tgs: 59 data_time: 0.98s time: 530.10s eta: 3 days, 10:27:36 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 15:57:32][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.182 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 60 data_time: 0.95s time: 529.94s eta: 3 days, 10:17:18 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 16:06:25][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32039.0 tgs: 60 data_time: 1.25s time: 533.30s eta: 3 days, 10:39:39 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 16:15:14][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 60 data_time: 0.78s time: 529.21s eta: 3 days, 9:52:52 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 16:24:04][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32352.0 tgs: 61 data_time: 0.74s time: 529.66s eta: 3 days, 9:48:12 +[XTuner][RANK 30][DP 7][SP 2][TP 0][2025-01-21 16:32:54][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31385.0 tgs: 59 data_time: 1.05s time: 530.38s eta: 3 days, 9:46:02 diff --git a/20250121104251/rank45.log b/20250121104251/rank45.log new file mode 100644 index 0000000000000000000000000000000000000000..14201b8795cd082c565cbfcaf5dd974c3da34c76 --- /dev/null +++ b/20250121104251/rank45.log @@ -0,0 +1,294 @@ +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:42:55][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121104251', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:42:56][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:42:59][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:02][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:04][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:06][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:08][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:10][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:12][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:43:18][INFO] [Dataset & Dataloader] Cost 22.82s +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:46:39][SUCCESS] [Parallelize LLM] Elapsed time 121.10 seconds, peak gpu memory 13.4G +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:46:40][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 10:56:05][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.314 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 31361.0 tgs: 57 data_time: 1.69s time: 549.97s eta: 3 days, 18:35:32 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 11:04:55][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.252 loss(reduced): 0.257 grad_norm: 2.31 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32373.0 tgs: 61 data_time: 1.13s time: 529.71s eta: 3 days, 15:06:30 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 11:13:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.257 loss(reduced): 0.236 grad_norm: 1.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31271.0 tgs: 59 data_time: 1.13s time: 529.09s eta: 3 days, 14:51:33 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 11:22:34][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.219 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31959.0 tgs: 60 data_time: 0.75s time: 529.97s eta: 3 days, 14:51:19 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 11:31:23][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32119.0 tgs: 60 data_time: 0.59s time: 529.02s eta: 3 days, 14:33:15 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 11:40:13][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.202 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31094.0 tgs: 58 data_time: 0.78s time: 529.86s eta: 3 days, 14:32:36 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 11:49:03][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.217 loss(reduced): 0.205 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 60 data_time: 1.01s time: 529.63s eta: 3 days, 14:21:30 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 11:57:58][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.194 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31574.0 tgs: 58 data_time: 0.72s time: 535.78s eta: 3 days, 15:12:45 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 12:06:52][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.203 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31714.0 tgs: 59 data_time: 0.72s time: 534.03s eta: 3 days, 14:46:48 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 12:15:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.287 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31891.0 tgs: 59 data_time: 0.77s time: 533.28s eta: 3 days, 14:30:37 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 12:24:35][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.200 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31327.0 tgs: 59 data_time: 0.83s time: 529.43s eta: 3 days, 13:44:16 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 12:33:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.297 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 60 data_time: 0.82s time: 529.02s eta: 3 days, 13:31:31 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 12:42:18][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.233 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31061.0 tgs: 58 data_time: 0.64s time: 533.81s eta: 3 days, 14:09:03 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 12:51:07][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.205 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30891.0 tgs: 58 data_time: 0.71s time: 528.94s eta: 3 days, 13:13:04 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 12:59:57][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31341.0 tgs: 59 data_time: 0.82s time: 530.39s eta: 3 days, 13:18:17 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 13:09:01][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.191 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31438.0 tgs: 57 data_time: 0.84s time: 543.36s eta: 3 days, 15:14:24 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 13:17:57][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 59 data_time: 0.64s time: 536.35s eta: 3 days, 13:57:52 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 13:26:48][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 60 data_time: 0.92s time: 530.95s eta: 3 days, 12:57:08 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 13:35:38][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 59 data_time: 0.72s time: 529.94s eta: 3 days, 12:38:33 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 13:44:32][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 59 data_time: 0.71s time: 534.35s eta: 3 days, 13:11:55 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 13:53:22][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 60 data_time: 0.63s time: 529.94s eta: 3 days, 12:20:56 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 14:02:19][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 60 data_time: 0.72s time: 537.06s eta: 3 days, 13:20:00 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 14:11:09][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31172.0 tgs: 58 data_time: 0.90s time: 529.34s eta: 3 days, 11:57:34 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 14:20:04][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.213 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32324.0 tgs: 60 data_time: 0.91s time: 535.74s eta: 3 days, 12:49:30 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 14:29:03][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 60 data_time: 0.86s time: 538.83s eta: 3 days, 13:09:56 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 14:37:57][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31182.0 tgs: 58 data_time: 0.82s time: 533.66s eta: 3 days, 12:12:01 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 14:46:47][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31901.0 tgs: 60 data_time: 0.68s time: 529.60s eta: 3 days, 11:24:41 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 14:55:38][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30544.0 tgs: 57 data_time: 0.95s time: 531.66s eta: 3 days, 11:35:19 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 15:04:28][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.183 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31830.0 tgs: 60 data_time: 0.72s time: 529.51s eta: 3 days, 11:06:14 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 15:13:17][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.187 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31792.0 tgs: 60 data_time: 1.11s time: 529.51s eta: 3 days, 10:57:22 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 15:22:07][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.188 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31968.0 tgs: 60 data_time: 0.82s time: 529.83s eta: 3 days, 10:51:33 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 15:31:02][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31133.0 tgs: 58 data_time: 0.83s time: 534.69s eta: 3 days, 11:28:17 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 15:39:52][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 60 data_time: 0.82s time: 529.78s eta: 3 days, 10:33:27 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 15:48:42][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31570.0 tgs: 59 data_time: 0.91s time: 530.08s eta: 3 days, 10:27:27 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 15:57:32][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.182 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32328.0 tgs: 61 data_time: 0.95s time: 529.94s eta: 3 days, 10:17:16 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 16:06:25][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.239 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 59 data_time: 0.94s time: 533.31s eta: 3 days, 10:39:45 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 16:15:14][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31631.0 tgs: 59 data_time: 0.92s time: 529.21s eta: 3 days, 9:52:48 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 16:24:04][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31761.0 tgs: 59 data_time: 0.73s time: 529.65s eta: 3 days, 9:48:07 +[XTuner][RANK 45][DP 11][SP 1][TP 0][2025-01-21 16:32:54][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31545.0 tgs: 59 data_time: 0.83s time: 530.41s eta: 3 days, 9:46:16 diff --git a/20250121104251/rank47.log b/20250121104251/rank47.log new file mode 100644 index 0000000000000000000000000000000000000000..d3a7fcff2de9188dce2d92b2c5437328f59d5bb5 --- /dev/null +++ b/20250121104251/rank47.log @@ -0,0 +1,294 @@ +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:42:55][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121104251', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:42:56][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:42:59][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:03][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:05][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:07][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:09][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:11][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:14][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:15][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:43:18][INFO] [Dataset & Dataloader] Cost 22.82s +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:44:38][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:46:39][SUCCESS] [Parallelize LLM] Elapsed time 121.14 seconds, peak gpu memory 13.4G +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:46:40][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 10:56:05][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.238 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 31361.0 tgs: 57 data_time: 1.69s time: 550.09s eta: 3 days, 18:36:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 11:04:55][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.265 loss(reduced): 0.257 grad_norm: 2.31 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32373.0 tgs: 61 data_time: 1.13s time: 529.58s eta: 3 days, 15:05:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 11:13:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.226 loss(reduced): 0.236 grad_norm: 1.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31271.0 tgs: 59 data_time: 1.15s time: 529.09s eta: 3 days, 14:51:33 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 11:22:34][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31959.0 tgs: 60 data_time: 0.77s time: 529.96s eta: 3 days, 14:51:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 11:31:23][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.210 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32119.0 tgs: 60 data_time: 0.60s time: 529.03s eta: 3 days, 14:33:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 11:40:13][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.181 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31094.0 tgs: 58 data_time: 0.80s time: 529.86s eta: 3 days, 14:32:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 11:49:03][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.205 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 60 data_time: 1.03s time: 529.63s eta: 3 days, 14:21:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 11:57:58][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.239 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31574.0 tgs: 58 data_time: 0.72s time: 535.78s eta: 3 days, 15:12:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 12:06:52][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.223 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31714.0 tgs: 59 data_time: 0.76s time: 534.03s eta: 3 days, 14:46:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 12:15:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.208 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31891.0 tgs: 59 data_time: 0.79s time: 533.28s eta: 3 days, 14:30:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 12:24:35][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31327.0 tgs: 59 data_time: 0.84s time: 529.42s eta: 3 days, 13:44:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 12:33:24][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.261 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 60 data_time: 0.84s time: 529.02s eta: 3 days, 13:31:32 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 12:42:18][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31061.0 tgs: 58 data_time: 0.70s time: 533.81s eta: 3 days, 14:09:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 12:51:07][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.205 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30891.0 tgs: 58 data_time: 0.74s time: 528.94s eta: 3 days, 13:13:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 12:59:57][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31341.0 tgs: 59 data_time: 0.81s time: 530.39s eta: 3 days, 13:18:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 13:09:01][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.191 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31438.0 tgs: 57 data_time: 0.87s time: 543.36s eta: 3 days, 15:14:24 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 13:17:57][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 59 data_time: 0.66s time: 536.35s eta: 3 days, 13:57:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 13:26:48][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 60 data_time: 0.95s time: 530.95s eta: 3 days, 12:57:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 13:35:38][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 59 data_time: 0.75s time: 529.94s eta: 3 days, 12:38:33 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 13:44:32][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 59 data_time: 0.73s time: 534.35s eta: 3 days, 13:11:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 13:53:22][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 60 data_time: 0.69s time: 529.94s eta: 3 days, 12:20:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 14:02:19][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.229 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 60 data_time: 0.73s time: 537.06s eta: 3 days, 13:19:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 14:11:09][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31172.0 tgs: 58 data_time: 0.90s time: 529.34s eta: 3 days, 11:57:33 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 14:20:04][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32324.0 tgs: 60 data_time: 0.93s time: 535.74s eta: 3 days, 12:49:30 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 14:29:03][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 60 data_time: 0.89s time: 538.83s eta: 3 days, 13:09:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 14:37:57][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31182.0 tgs: 58 data_time: 0.84s time: 533.66s eta: 3 days, 12:12:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 14:46:47][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31901.0 tgs: 60 data_time: 0.71s time: 529.60s eta: 3 days, 11:24:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 14:55:38][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30544.0 tgs: 57 data_time: 0.96s time: 531.66s eta: 3 days, 11:35:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 15:04:28][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.183 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31830.0 tgs: 60 data_time: 0.75s time: 529.52s eta: 3 days, 11:06:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 15:13:17][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.187 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31792.0 tgs: 60 data_time: 1.14s time: 529.50s eta: 3 days, 10:57:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 15:22:07][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.188 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31968.0 tgs: 60 data_time: 0.82s time: 529.82s eta: 3 days, 10:51:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 15:31:02][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31133.0 tgs: 58 data_time: 0.82s time: 534.70s eta: 3 days, 11:28:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 15:39:52][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 60 data_time: 0.82s time: 529.78s eta: 3 days, 10:33:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 15:48:42][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31570.0 tgs: 59 data_time: 0.92s time: 530.08s eta: 3 days, 10:27:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 15:57:32][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.182 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32328.0 tgs: 61 data_time: 0.99s time: 529.94s eta: 3 days, 10:17:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:06:25][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 59 data_time: 0.94s time: 533.31s eta: 3 days, 10:39:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:15:14][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31631.0 tgs: 59 data_time: 0.99s time: 529.21s eta: 3 days, 9:52:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:24:04][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31761.0 tgs: 59 data_time: 0.75s time: 529.66s eta: 3 days, 9:48:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:32:54][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31545.0 tgs: 59 data_time: 0.86s time: 530.41s eta: 3 days, 9:46:16 diff --git a/20250121165312/hf-593/added_tokens.json b/20250121165312/hf-593/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..482ced4679301bf287ebb310bdd1790eb4514232 --- /dev/null +++ b/20250121165312/hf-593/added_tokens.json @@ -0,0 +1,24 @@ +{ + "": 151658, + "": 151657, + "<|box_end|>": 151649, + "<|box_start|>": 151648, + "<|endoftext|>": 151643, + "<|file_sep|>": 151664, + "<|fim_middle|>": 151660, + "<|fim_pad|>": 151662, + "<|fim_prefix|>": 151659, + "<|fim_suffix|>": 151661, + "<|im_end|>": 151645, + "<|im_start|>": 151644, + "<|image_pad|>": 151655, + "<|object_ref_end|>": 151647, + "<|object_ref_start|>": 151646, + "<|quad_end|>": 151651, + "<|quad_start|>": 151650, + "<|repo_name|>": 151663, + "<|video_pad|>": 151656, + "<|vision_end|>": 151653, + "<|vision_pad|>": 151654, + "<|vision_start|>": 151652 +} diff --git a/20250121165312/hf-593/config.json b/20250121165312/hf-593/config.json new file mode 100644 index 0000000000000000000000000000000000000000..a40f5115ce01babb5c397b76d9514384e68f1f4f --- /dev/null +++ b/20250121165312/hf-593/config.json @@ -0,0 +1,30 @@ +{ + "_name_or_path": "/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842", + "architectures": [ + "Qwen2ForCausalLM" + ], + "attention_dropout": 0.0, + "attn_implementation": "flash_attention_2", + "bos_token_id": 151643, + "eos_token_id": 151645, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 29568, + "max_position_embeddings": 32768, + "max_window_layers": 70, + "model_type": "qwen2", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "rms_norm_eps": 1e-06, + "rope_scaling": null, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.45.1", + "use_cache": false, + "use_sliding_window": false, + "vocab_size": 152064 +} diff --git a/20250121165312/hf-593/merges.txt b/20250121165312/hf-593/merges.txt new file mode 100644 index 0000000000000000000000000000000000000000..31349551d90c7606f325fe0f11bbb8bd5fa0d7c7 --- /dev/null +++ b/20250121165312/hf-593/merges.txt @@ -0,0 +1,151388 @@ +#version: 0.2 +Ġ Ġ +ĠĠ ĠĠ +i n +Ġ t +ĠĠĠĠ ĠĠĠĠ +e r +ĠĠ Ġ +o n +Ġ a +r e +a t +s t +e n +o r +Ġt h +Ċ Ċ +Ġ c +l e +Ġ s +i t +a n +a r +a l +Ġth e +; Ċ +Ġ p +Ġ f +o u +Ġ = +i s +ĠĠĠĠ ĠĠĠ +in g +e s +Ġ w +i on +e d +i c +Ġ b +Ġ d +e t +Ġ m +Ġ o +ĉ ĉ +r o +a s +e l +c t +n d +Ġ in +Ġ h +en t +i d +Ġ n +a m +ĠĠĠĠĠĠĠĠ ĠĠĠ +Ġt o +Ġ re +- - +Ġ { +Ġo f +o m +) ;Ċ +i m +č Ċ +Ġ ( +i l +/ / +Ġa nd +u r +s e +Ġ l +e x +Ġ S +a d +Ġ " +c h +u t +i f +* * +Ġ } +e m +o l +ĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠ +t h +) Ċ +Ġ{ Ċ +Ġ g +i g +i v +, Ċ +c e +o d +Ġ v +at e +Ġ T +a g +a y +Ġ * +o t +u s +Ġ C +Ġ st +Ġ I +u n +u l +u e +Ġ A +o w +Ġ ' +e w +Ġ < +at ion +( ) +Ġf or +a b +or t +u m +am e +Ġ is +p e +t r +c k +â Ģ +Ġ y +i st +-- -- +. ĊĊ +h e +Ġ e +l o +Ġ M +Ġb e +er s +Ġ on +Ġc on +a p +u b +Ġ P +ĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠ +as s +in t +> Ċ +l y +ur n +Ġ $ +; ĊĊ +a v +p ort +i r +- > +n t +ct ion +en d +Ġd e +it h +ou t +t urn +ou r +ĠĠĠĠ Ġ +l ic +re s +p t += = +Ġth is +Ġw h +Ġ if +Ġ D +v er +ag e +Ġ B +h t +ex t += " +Ġth at +** ** +Ġ R +Ġ it +es s +Ġ F +Ġ r +o s +an d +Ġa s +e ct +k e +ro m +Ġ // +c on +Ġ L +( " +q u +l ass +Ġw ith +i z +d e +Ġ N +Ġa l +o p +u p +g et +Ġ} Ċ +i le +Ġa n +at a +o re +r i +Ġp ro +; čĊ +ĉĉ ĉĉ +t er +a in +Ġ W +Ġ E +Ġc om +Ġre turn +ar t +Ġ H +a ck +im port +ub lic +Ġ or +e st +m ent +Ġ G +ab le +Ġ - +in e +il l +in d +er e +: : +it y +Ġ + +Ġt r +el f +ig ht +( ' +or m +ul t +st r +. . +" , +Ġy ou +y pe +p l +Ġn ew +Ġ j +ĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠ +Ġf rom +Ġ ex +Ġ O +l d +Ġ [ +o c +: Ċ +Ġs e +Ġ le +---- ---- +. s +{ Ċ +' , +an t +Ġa t +as e +. c +Ġc h +< / +av e +an g +Ġa re +Ġin t +âĢ Ļ +_ t +er t +i al +a ct +} Ċ +iv e +od e +o st +Ġc lass +Ġn ot +o g +or d +al ue +al l +f f +( );Ċ +on t +im e +a re +Ġ U +Ġp r +Ġ : +i es +iz e +u re +Ġb y +i re +Ġ} ĊĊ +. p +Ġs h +ic e +a st +pt ion +tr ing +o k +_ _ +c l +# # +Ġh e +ar d +) . +Ġ @ +i ew +ĉĉ ĉ +Ġw as +i p +th is +Ġ u +ĠT he +id e +a ce +i b +a c +r ou +Ġw e +j ect +Ġp ublic +a k +v e +at h +o id +Ġ= > +u st +q ue +Ġre s +) ) +' s +Ġ k +an s +y st +un ction +**** **** +Ġ i +Ġ us +p p +on e +a il +== == +n ame +Ġst r +Ġ / +Ġ & +a ch +d iv +yst em +el l +Ġh ave +er r +ou ld +ul l +p on +Ġ J +_ p +Ġ= = +ig n +S t +. Ċ +Ġp l +) ;ĊĊ +f orm +p ut +ou nt +} ĊĊ +d d +it e +Ġg et +r r +om e +Ġ âĢ +ar am +c c +Ġ* / +E R +I n +le s +_ s +on g +i e +Ġc an +Ġ V +er v +p r +Ġ un +ro w +b er +Ġd o +l l +Ġ el +Ġs elf +at ed +ar y +Ġ . +' ] +u d +Ġ en +ĠT h +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠ +t e +_ c +u ct +Ġa b +or k +. get +Ġ # +a w +res s +o b +N ame +ap p +[ ' +Ġal l +or y +it ion +an ce +e ar +Ġcon t +v ent +i a +Ġw ill +I N +ĠĠĠĠĠĠĠĠ Ġ +re turn +Ġ< / +d ata +) ĊĊ +R e +p le +il d +th er +Ġy our +" Ċ +( $ +Ġ out +) , +Ġh as +S tring +s o +Ġ up +a x +Ġde f +Ġb o +g e +al se +O N +p er +ic h +Ġb ut +Ġ Ċ +Ġ _ +_ m +ad d +que st +od el +s elf +er y +f t +en s +// // +a ke +. C +Ġg o +Ġf unction +Ġ K +iv ate +Ġ im +Ġcon st +. t +Ġ*/ Ċ +) ;čĊ +Ġv oid +Ġs et +ĠS ystem +c ri +( )Ċ +l i +ĉ if +. m +al ly +s et +e p +âĢĻ s +b o +de f +' ,Ċ +Ġm e +Ġ ! +at ch +" > +" ,Ċ +e c +ĠI n +p h +Ġ | +_ f +Ġv ar +en ce +I d +re e +in k +le ct +u g +et h +Ġel se +-------- -------- +con t +Ġs o +at ic +Ġl o +p ro +t on +s s +ow n +ab el +o int +ou s +el d +S T +T he +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +R E +" : +ol or +t p +e g +ke y +u de +ĠS t +ou nd +Ġa r +" );Ċ +en er +s er +b ject +ess age +f er +Ġm ore +ation s +ent s +Ġh is +Ġthe y +. S +Ġ Y +u se +n e +is h +ol d +_ d +i o +i eld +Ġp er +C ont +ing s +## ## +Ġd ata +Ġs a +e f +f o +Ġon e +en g +Ġd is +A T +Ġn ame +Ġtr ue +v al +le d +. f +Ġn e +Ġ end +. T +c re +ar k +lo g +E x +err or +_ id +ur re +ang e +Ġn ull +rr ay +Ġm y +p an +ic t +at or +V iew +L ist +ĉ return +âĢ Ŀ +Ġp re +Ġ x +cl ude +ar g +o v +. h +Ġ > +Ġthe ir +' ) +ir st +ic k +g h +L E +O R +Ġpr ivate +t em +čĊ čĊ +us er +Ġ ) +c om +. A +" ;Ċ +Ġ id +re ad +Ġwh o +_ b +" >Ċ +Ġt ime +Ġm an +r y +==== ==== +rou p +ro p +p ublic +v el +um ber +b le +Ġwh ich +******** ******** +Ġan y +Ġf alse +w e +Ġv alue +Ġl i +" ) +nd er +g r +Ġn o +p aram +f ig +.c om +Ġa pp +_ l +ion s +. D +ĠC h +Ġab out +Ġa dd +Ġs u +Ġstr ing +I D +Ġo ver +str ing +. l +our ce +_ C +] Ċ +Ġ qu +ĠS tring +c a +S E +Ġ ro +s h +u al +T ype +s on +n ew +er n +Ġa g +A R +] ;Ċ +] . +Ġ ? +ic al +Ġd es +ut h +i x +ay s +Ġt ype +' t +a ult +Ġin ter +v ar +. b +Ġp art +. d +urre nt +I T +E N +en c +( f +r a +v alue +ch o +ut ton +o se +Ġ! = +at er +à © +re ate +ol l +p os +y le +n g +A L +us ing +am es +Ġ{ čĊ +at es +el y +Ġw ork +Ġ em +in al +Ġs p +Ġwh en +.s et +ĠĠĠĠ ĠĠ +) :Ċ +t o +qu ire +ind ow +le ment +pe ct +as h +[ i +Ġu se +. F +pe c +Ġa d +o ve +ce ption +eng th +in clude +ad er +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠ +at us +T h +it le +r it +v oid +() . +( Ċ +Ġof f +Ġo ther +Ġ& & +' ;Ċ +m s +Ġbe en +Ġt e +m l +c o +n c +erv ice +Ġ % +** Ċ +an n +ad e +ĊĊ ĊĊ +lo ck +con st +pon se +Ġs up ++ + +d ate +Ġa cc +Ġh ad +Ġb u +ĠR e +Ġw ere +Ġf ile +Ġw ould +ĠâĢ ľ +v en +is s +Ġ our +c lass +r aw +Ġy ear +D ata +Ġv al +Ġs ome +f ter +y s +Ġ// / +rou nd +v iew +Ġp e +Ġth ere +Ġsa id +d u +o f +l ine +/ * +d uct +Ġh er +ĠĠĠĠĠĠĠĠ ĠĠĠĠĠ +R es +Ġc o +Ġcom m +is e +m in +ĠĠĠĠ Ċ +# include +eth od +. P +ut e +Ġas s +I nt +as k +lo c +Ġli ke +od y +Ġle t +lo ad +Ġa m +ro l +Ġg r +y p +Ġal so +ĠI t +ur l +if ic +or s +_ P +_ n +ig h +Ġth an +C om +A N +U L +at ing +ĠTh is +re f +_ S +Ġst atic +ro ll +Ġj ust +Ġres ult +i an +id th +Ġthe m +) );Ċ +d er +re ak +C on +: // +u le +.. . +ar ch +em ent +Ġ< < +us h +en se +ar r +Ġint o +c ess +am p +i ed +um ent +Ġ \ +] , +w o +al s +Ġwh at +an c +V alue += ' +ol um +Ġp os +ag es +ay er +Ġs c +u es +" )Ċ +_ T +Ġl ist +( s +Ġc ase +C h +ĉĉĉĉ ĉ +//// //// +pon ent +Ġ z +Ġk n +le t +D E +re d +Ġf e +Ġ} ,Ċ +Ġ , +( t +Ġf irst +' );Ċ +w ord +Ġ import +Ġa ct +Ġch ar +C T +ĠT r +op le += { +ĉ f +i ent +c ent +. j +le ction +) )Ċ +Ġon ly +Ġpr int +m er +. W +o ck +Ġ -- +T ext +Ġo p +an k +Ġit s +Ġb ack +[ " +Ġne ed +Ġc l +Ġs ub +Ġl a +( ( +. " +O bject +Ġst art +f ile +( self +n er +e y +Ġus er +Ġ ent +ĠC om +it s +ĠC on +ou ble +ow er +it em +ver y +ĠW e +lic k +Ġ Q +ph p +t tp +' : +ic s +Ġu nder +Ġ* Ċ +. L +) ; +ic es +Ġre g +) čĊ +ĉ public +S S +Ġth en +re at +i ous +. G +e k +ire ct +he ck +cri pt +n ing +ĠU n +Ġm ay +ĠW h +B o +I tem +str uct +. st +re am +ib le +lo at +Ġor g +u nd +s um +_ in +.. / +_ M +Ġh ow +r ite +' Ċ +T o +w w +Ġpe ople +ind ex +. n +ht tp +( m +ect or +Ġin d +Ġj av +] ,Ċ +ĠH e +_ st +f ul +o le +) {Ċ +Ġsh ould +op y +el p +i er +_ name +ers on +I ON +ot e +Ġt est +Ġb et +rr or +ul ar +ã Ģ +Ġ Ð +b s +t ing +Ġm ake +T r +Ġa fter +ar get +R O +olum n +r c +_ re +def ine +Ġr ight +r ight +d ay +Ġl ong +[ ] +( p +t d +con d +ĠP ro +Ġre m +ption s +v id +. g +Ġ ext +Ġ __ +' )Ċ +p ace +m p +Ġm in +st ance +a ir +a ction +w h +t ype +ut il +a it +< ? +I C +t ext +Ġp h +Ġf l +. M +cc ess +b r +f ore +ers ion +) ,Ċ +. re +ate g +Ġl oc +in s +- s +tr ib +ĠI nt +Ġa rray +, " +P ro +( c +ess ion +> ĊĊ +Ġs he +" ] +ap h +Ġex p +ert y +ĠS e +Ġp ar +un c +E T +Ġre ad +pr int +Ġre l +Ġfor m +Ġd r +Ex ception +in put +Ġtr ans +#### #### +ord er +B y +Ġa w +it ies +u ff +pl ay +. add +ĠâĢ ĵ +Ġw ant +Ġcom p +ment s +Ġ| | +a z +b e +Ġn umber +Ġre quire +ĠE x +Ġc ol +Ġ key +em ber +Ġt wo +Ġs ize +Ġwh ere +U T +res ult +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +ou gh +or ld +o od +u ch +at ive +g er +are nt +Ġ/ * +Ġar g +Ġwh ile +( this +Ġre c +Ġd if +St ate +Ġs pec +r ide +_ F +Ġlo ok +A M +il ity +et er +âĢĻ t +ĊĊ Ċ +ay out +---------------- ---------------- +ag er +Ġc ould +Ġb r +end s +u res +Ġkn ow +et s +ĠI f +ĠS h +. w +b ack +Ġs er +Ġ+ = +Ġf r +() );Ċ +Ġh and +I nd +UL L +I m +() ;ĊĊ +Ġm ost +Ġtr y +Ġn ow +rou gh +> čĊ +ack age +Ġh im +. _ +if y +Ġb reak +Ġ );Ċ +re n +# define +it t +Ġa p +ĉ c +( n +ĠY ou +: ĊĊ +- m +Ġe very +ust om +li ent +oc ument +cri ption +E rror +- b +Ð ¾ +] [ +tr ans +Ġp oint +Ġst d +Ġf il +T ime +Ġm od +Ġ -> +Ġ error +a h +Ġt ext +roll er +lo se +q l +Ġp ol +> < +. B +- c +Ġop en +Ġe st +ĠĠĠĠĠĠĠĠ Ċ +Ġn ext +I M +Ñ Ĥ +O T +à ³ +Ġf ollow +cont ent +ĠĠĠĠĠĠĠĠ ĠĠĠĠ +Ġin clud +H E +ĠR es +Ġh ref +Ð ¸ +Ġc ar +yp es +im age +U n +Ġbo ol +A D +Ġg ame +.F orm +row s +* / +vel op +.D rawing +Ġp ath +is ion +Ġe ach +ĠP l +_t ype +P ath +ne ction +Ġa v +' ). +Ġsup port +EN T +re m +" ). +Ġo wn +Ġc or +c ount +m iss +u ally +Ġm em +st d +i ence +se arch +" ĊĊ +F orm +Ġs ex +en ame +Ġs ign +Ġ et +ĠĠĠĠĠĠĠĠ ĠĠ +', ' +ĠA pp +Ġth ose +o ff +Ġ err +Ġs ystem +Ġbe st +c ode +Ġs ame +Ġd i +us s +Ġc reate +ath er +A rray +. in +f e +S ervice +U N +at s +Ġ Z +al th +Ġm ade +tr ue +A B +Ġm ark +r id +if ied +, čĊ +y n +p ress +Ġg roup +Ġf in +ĠL icense +F ield +eg er +Ġw orld +in ess +t y +Ġpro cess +( b +Ġc re +ar n +iv es +Ġm ain +ide o +_ g +A G +val id +im g +P I +Ġc olor +Ġre port +Ġt ake +ri b +O M +Ġd ay +Re quest +Ġs k +b ers +ĉ s +.A dd +o ot +Im age +Ġcom ple +ol lection +Ġto p +Ġf ree +A S +D e +ĠO n +I G +et a +D ate +Ġa ction +O ver +it or +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +n ot +Ġind ex +h er +ic on +O n +;čĊ čĊ +iv ity +m and +.W indows +O L +Ġre al +Ġm ax +l and +.. .. +r aph +Ġbu ild +le g +ass word +? ĊĊ +âĢ ¦ +o ok +u ck +Ġm essage +t est +iv ers +Ġin put +Ġar t +Ġbet ween +G et +ent er +g round +en e +à ¡ +.l ength +N ode +( i +C lass +f or +ĠâĢ Ķ +t en +o in +Ġ ke +u i +ĠI N +Ġt able +s ub +ĠL e +Ġhe ad +Ġm ust +//////// //////// +. util +Cont ext +Ġor der +Ġm ov +o ver +Ġcont in +Ġs ay +st atic +.T ext +Ġclass Name +pan y +Ġt er +he ad +r g +Ġpro duct +Th is +. âĢĿ +ĠB ut +lo y +Ġd ouble +s g +Ġpl ace +. x +m essage +Ġin formation +pr ivate +Ġo per +c ed +d b +"> +ater ial +ile d +Ġp ut +Q u +Ñ Ģ +un g +m ap +ĉĉĉĉ ĉĉĉĉ +Ġle vel +Com ponent +bo ok +cre en +_ RE +Ġcon fig +ã ģ +O r +. data +Ġd ocument +", " +trib ute +u x +L og +fer ence +p ost +_ e +Ġloc al +and om +ass ert +V al +lect ed +in a +atab ase +A dd +Ġcont ent +.p rint +s igned +r ic +." ĊĊ +Ġf a +! ĊĊ +- f +iv ed +Ġ quest +. ex +Ġf loat +Ġde velop +о Ð +M ap +ad ing +Ġpos s +U E +n amespace +_ O +ĉ b +.G et +> ( +j son +etail s +Ġto o +Ġext ends +ĠN one +Ġf ore +( String +form at +Ġg reat +int er +ca le +Ñ ģ +r on +iv ing +E nt +enc y +x t +o y +Ġmon th +Ġh app +Ġsup er +b ar +def ault +_ de +ord s +l n +( {Ċ +ĠI nd +as es +Ġt itle +Ġcont ext +o h +- p +E m +Ġm et +T est +Ġl ife +_ v +ĠU S +U I +oc ation +m d +Ġ[ Ċ +Ġ ] +s w +Ġin cre +s cript +ent ial +w ays +. de +Ġs rc +Ġc atch +ĠA meric +// Ċ +ĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠ +Ġp ay +pl it +âĢ Ķ +Ġc oun +ob j +.ph p +Ġch ange +eth ing +' re +ast er +lo s +l ation +ĠĠ Ċ +L e +à ¤ +( { +read y +ĠN o +Ġpos ition +Ġo ld +Ġbo ok +able d +b ug +H and +} ;ĊĊ +is play +av ing +Ġgo ver +Ġv ersion +S ystem +n ect +res ponse +St yle +U p +ang u +Ġth ree +in it +er o +Ġl aw +end if +Ġb ase +em ail +( l +_ V +Ġcon f +AT E +Ġd uring +t es +Ġcon sole +ĠP r +Ġs pe +v es +p ath +ial og +d ition +_t o +ard s +Ġagain st +et work +ĠP h +_ L +c ur +im it +W ith +Ġp ower +i um +' ;ĊĊ +Ġw om +le ft +our ces +at ri +ĠI m +ĠM an +or th +$ { +qu als +es e +_s ize +Ġis s +ot al +- g +i que +r ame +Ġw idth +er g +) ( +itt le +T R +ĠThe y +enc es +r l +on s +Ġl abel +. y +- t +up date +an el +s c +.t o +Ġpro ject +à ¼ +Ġe lement +Ġsu ccess +ĉĉ Ċ +.s h +r am +ch ed +() )Ċ +Ġ( Ċ +Ġd ate +Ġto t +_ ST +A ll +ific ation +ĉ var +Ġt ri +ch em +m y +Ġb ig +ĠA d +ĠA t +ot s +n um +A ct +Ġm ap +er a +co pe +. $ +, âĢĿ +Ġp op +Ġf ew +Ġl en +u id +et ers +u les +Ã Ń +s ource +http s +Ġd em +Ġe ar +######## ######## +Ġm atch +or ies +ac es +ĠC l +Ġn ode +ir c +loc al +un ity +} ;Ċ +Ġan other +< < +og le +Ġs it +ew ork +T E +. I +N S +olog y +ou ght +.C ont +> > +Ġc are +st ate +ĉ private +Ġe ffect +++ ) +_f ile +end ing +L ine +F or +i or +ĠS c +Ġf un +.S ize +ĉ else +] ) +st art +v ious +Ġ} , +our s +Ġle g +Ġs ervice +Ġs ince +ir on +L abel +Ġn on +Ġl os +ict ion +Ġf ull +act er +bo ard +g ress +Ġt urn +ith er +.s ize +Ġb ody +res h +et urn +( _ +y les +orm al +p i +Ġsom ething +! -- +u int +Ġpro du +Ġst and +Ġpro ble +Ġav ailable +m t +ĠB l +Ġ ... +Ġb lock +In put +Ġke ep +C ount +op en +Ġ[ ' +Ġth row +uild er +A ction +Ġth ings +Tr ue +Ġ url +ĠB o +print f +Ġre d +j s +.c reate +ĠO r +St atus +In stance +Ġcont rol +Ġcom e +Ġc ustom +loc ation +m odel +Ġ čĊ +Ġs ource +Ġe as +. out +] ĊĊ +one y +Ġaw ait +Ġpart ic +A P +ub lish +od es +_p ro +p ly +rit er +Ġpro v +Ġm ill +H T +] )Ċ +Ġch ang +Ġas k +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠ +Ġout put +Ġem ail +.p ush +Ġ} čĊčĊ +in ation +atri x +T able +u ccess +] );Ċ +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġdis c +( [ +Ġb usiness +he ight +. html +t a +f ield +Ġrequire d +_ R +Ġgover n +} čĊčĊ +le x +. , +ĠS et +ur ch +// / +t s +a f +Ġm ight +ist ory +S tr +Ġne ver +Res ponse +ar se +ad a +ĠH ow +Ġ* ) +Ġ ; +Ġh ard +A d +Ġinter n +us ed +( data +m od +ann el +Ġn p +ug g +Ġ/ >Ċ +Ġcal led +b ody +Ġch o +( r +_s et +ir d +Ġ> = +Ġ} ;Ċ +Ġo ptions +ĠG ener +Ġhe ight +P oint +Y ou +et y +C lick +Ġsm all +Ġ ide +Ġacc ess +angu age +Ġprot ected +Ġj ob +ĠTh ere +D ef +Ġadd ress +Ġu int +N ot +o o +ap s +< div +ain ed +at ur +Ġs um +- w +ĠD ate +Ġl ittle +Ġf ri +Y PE +Ġp ort +e h +pr ing +_p ath +Ġst atus +a im +bo ol +Ġap pe +Ġo s +. name +ens ion +_ G +Ġup date +Con fig +a ff +ER R +Ġ< = +at ely +# if +u ction +ĠT e +Ġl ink +ĠU ser +.f ind +. org +m e +Ġg iven +O ut +# endif +Ġbet ter +P age +Ġfe el +en n +M L +Ġal ready +Ġinclud ing +o ogle +r u +ic ally +pro p +le an +out er +Ġal ways +ord ing +I f +or age +Ġp arent +v is +ĉĉĉĉ ĉĉĉ +Ġg ot +st and +Ġle ss +/ s +ĠA ss +ap t +ire d +ĠA dd +Ġacc ount +p loy +Ġd er +res ent +Ġl ot +Ġval id +ĉ d +Ġb it +pon ents +Ġfollow ing +_ ex +S ON +Ġs ure +oc ial +Ġp rom +ert ies +he ader +.p ro +Ġbo olean +Ġse arch +k en +Ġor ig +Ġ er +E d +E M +a ut +l ing +al ity +By Id +b ed +ĉc ase +eth er +pos it +Ġinv est +ĠO R +Ġs ays +miss ion +AM E +Ġtem p +o ad +Ġre st +in fo +Ġinter est +A rg +Ġper form +pon s +ĠV iew +Ġv er +l ib +( const +U til +List ener +ar ge +Ġm ult +Ġd ie +Ġs ite +../ ../ +E L +Ġval ues +Ġ} )Ċ +p en +N o +ic ro +Ġbe h +Ġ' ./ +ac y +re c +() -> +ĉ ĠĠĠ +" )) +Cont ent +_ W +ple ment +Ġw on +Ġv ideo +ad i +p oint +% % +Ġg l +erv ed +v iron +I F +ut ed +ã ĥ +' m +Ġc ert +Ġpro f +Ġc ell +ar i +Ġpl ayer +a is +Ġc ost +Ġh um +( R +Ġoff ic +k s +.t ext +at ures +Ġtot al +Ġ*/ ĊĊ +o pe +Ġst at +U M +Ġlo ad +ight s +Ġc lear +u ro +Ġte chn +up port +I R +Ġ row +Ġse em +Ġ q +Ġsh ort +ĠN ot +ip p +G roup +se ction +m ax +ir l +Ġover ride +Ġcom pany +Ġd one +" );čĊ +Ġg re +. Re +Ġbel ie +r ist +Ġhe alth +AN T +() ĊĊ +ĠB e +. value +ĠG r +ott om +Ġarg s +P T +st atus +f unc +um ents +- h +N umber +: čĊ +ĠL og +er ver +Ġ) ,Ċ +am ent +Ġob j +in c +Ġchild ren +ic y +I Z +and s +ab ly +Ġdist rib +Ġc ur +er ial +Ġd ays +re ated +re ct +- l +ir m +idd en +om b +Ġin itial +.j s +Ġ â +Qu ery +Ġon line +im al +. con +a u +U rl +cont rol +ire ction +Ġin stance +OR T +ĠF r +wh ere +Ġjav ax +Ġorg an +ap ter +Ġre ason +o ptions +ĠM ar +( a +Ġwith in +.âĢĿ ĊĊ +O DE +_ DE +ad min +end ed +Ġdes ign +ĠD ata +un e +ĠF ile +ro ot +Ġc ent +Ġa rr +_ add +l en +p age +, ' +_ str +Ġb ro +ab ility +ou th +/ c +p ose +irt ual +ear ch +_ url +arg in +H ttp +Ġs chool +av a +Ġcons ider +.l abel +ĠA rray +we b +o pt +.print ln +ul ation +Ġf unc +P L +Ġ" \ +ĠT ext +act ory +(f unction +n ull +Ġen g +d own +Ġin clude +ĠE n +ĠD r +Ġd b +! ! +s ide +Ġin it +quire d +ĠS he +C olumn +re act +Ġan n +Ġst op +Ġl ater +ĠTh at +ent ion +d f +U G +I LE +Ġc lient +ra ft +ff er +PO ST +el per +Ġlo ve +qu ote +ou d +Ġj son +Ġab le +Ġm en +A X +ĠC opyright +à ¶ +av ig +re q +C lient +} );Ċ +.C om +er c +il t +pec ial +_c om +ro om +. Name +Ġg ive +am b +i ke +Ġcon dition +cl ient +ator s +: " +Ġc opy +ut ure +ivers ity +ern al +{ { +ĠC an +ou nc +d o +Ġo cc +Ġapp ro +th ers +z e +Ġe ither +ĠF l +Ġimport ant +Ġle ad +at tr +AR T +E qual +Ġd a +et ch +ent ity +Ġfam ily +add ing +Ġo ption +Ġex ist +ic a +ĠO bject +' ve +v ers +ition al +out put +ĠTr ue +ĠO F +_t ime +Ġof fer +Ġ} );ĊĊ +H ER +eg in +" " +Ġw ater +Ġc he +ĠM y +ore d +Ġst ep +anc es +C K +A Y +à ¸ +str uction +( C +ou ch +St ream +act ive +am a +Ent ity +pro duct +() {Ċ +Ġgovern ment +ĠI D +aj or +A nd +Ġdis play +Ð » +Ġt imes +Ġf our +Ġf ar +Ġpres ent +ĠN S +Ġ\ Ċ +ue st +Ġb as +e cho +ch ild +if ier +Hand ler +Ġl ib +Prop erty +trans lation +Ġro om +Ġon ce +Ġ[ ] +cent er +================ ================ +Ġresult s +Ġcontin ue +Ġt alk +_ get +Ġg row +.s w +e b +ĠP ublic +O P +ec ute +ol s +Ġ ** +" );ĊĊ +Ġm ass +ure d +.c lass +om ic +Ġme an +ip s +Ġa ut +);čĊ čĊ +Ġun til +Ġmark et +Ġare a +u it +Ġl ength +ĠW ith +struct or +e vent +"> < +ĠS p +I V +Ġm us +if f +Ġk ind +a uthor +ound s +m b +_ key +w idth +posit ory +Ġl ight +u k +R ow +oh n +al f +viron ment +app er +ollection s +Ġs ide +_in fo +Ġex ample +im ary +Ġw r +Ġc amp +cri be +" / +Ġm iss +w ay +Ġb ased +Ġpl an +V is +om ain +un k +Ġaw ay +U P +< T +O S +i od +ĠM on +âĢĻ re +Ġli k +à § +iv ely +. v +im er +iz er +S ub +Ġbut ton +ĠU p +Ġexper ience +C L +Ġre nder +_ value +Ġn ear +UR L +al t +Ġcoun try +ib ility +() ,Ċ +e ad +Ġa uthor +Ġspec ific +b ase +( name +on es +ĠD o +Ġal ong +y ear +Ġexp ress +. ' +en v +Ġbeg in +Ġso ftware +Ġim p +Ġw in +ó n +Ġth ing +Tr ans +ĠT HE +Ġ< ? +Ġwh y +Ġdoes n +i j +g ing +ĉ g +Ġs ingle +off set +ar ning +og raph +le y +_c ount +Ġan al +cre ate +/ m +ĠR eg +un ch += $ +is k +Ġright s +( M +Ġ"" "Ċ +ap er +.m odel +Ġp o +em pty +art ment +Ġa nt +ĠWh en +Ġwom en +ĠE d +Ġse ason +Ġde st +à £ +( h +Ġposs ible +Ġse ver +Ġb tn +Ġdid n +Ġs ent +Ġen c +Ġcomm and +Ġ ],Ċ +_ x +Ġre cent +ol ution +v ector +ĠB y +ĠM ay +ĠA ct +» ¿ +Ġm oney +IN T +bs ite +ĉ p +. čĊ +ï »¿ +s l +atter n +ĠC lass +Ġto ld +ud io +c urrent +Ġe qu +Ġa uto +ĠSt ate +d a +ms g +)) ;ĊĊ +Ġwork ing +Ġqu ery +ĠB r +Ġw indow +a uth +on ly +ĉ t +Ġle ast +ag n +Ġex pl +it ter +ar ing +Ġc olumn +ĠGener al +": " +er al +ri or +Ġrec ord +I B +E X +Ġd at +Ġm aking +u ed +ĠC ar +em p +" . +ĠM ed +Ġc lose +Ġper cent +Ġp ast +( g +: ( +Ġw rite +Ġm ove +Ġp at +Cont rol +.T o +Ġv i +*/ Ċ +in ate +' ll +ag ed +N ull +Ġspec ial +IZ E +Ġc ity +/* Ċ +ĠE ng +ix ed +in ary +p y +Ġe ff +ar io +Ġt ell +av or +Ġse lect +le vel +im um +op er +B uilder +I P +') ,Ċ +es c +Ġf ont +" ;ĊĊ +ĠA m +ish ed +ill s +Int er +O W +Ġcour se +Ġl ate +idd le +Ġam ount +Ġas ync +in o +c ul +Ġ ì +and le +_ user +Ġb en +ĠC al +Ġ$ _ +ĠR ep +Ġen ough +T oken +. user +( j +S c +W idth +n ow +at form +Ġlook ing +Ġh old +M odule +IT Y +v o +is on +.D ata +y c +Ġp ot +ĠTr ump +id ual +id es +r t +Ġprop erty +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠ +am ework +g o +Ġl ow +Ġpar a +Ġpr ice +ur y +Ġto day +ro y +Ġ' / +Ġpol it +Ġ' ' +ym b +P h +Ġad v +Ġatt ack +ĠS te +RO M +an a +Ġme ans +Ġst ory +id s +ak en +Ġme et +Ġm om +ĠâĢ ĺ +Ġ? > +Ġd en +ob ile +ch ange +ĠĠĠĠĠĠĠĠ ĠĠĠĠĊ +ic i +n a +ĠF orm +Ġs ort +Se lect +p are +Ġth ought +_ con +Ġt ask +oc us +ĠD E +ĠM in +Ġo pt +ĉb reak +um er +K E +th en +Ġd et +ĠT est +port s +Ġre view +(' / +m ove +Ġsw itch +ER T +p atch +ann ot +ã Ĥ +Ġab ove +it ive +Ġquest ion +ĠQ u +ãĢĤ ĊĊ +g le +Ġw ord +Ġprov ide +ĠR eturn +Ġre search +ã o +u str +Ġp ublish +chem a +} } +ĠC ON +- in +all back +Ġco ver +\ \ +c olor +ĠI S +Ġwh ether +im ate +is c +B ar +Ġd iv +B e +our n +Ġh aving +le m +pl ayer +ab s +am era +ne y +Ġex c +get her +pl ied +a o +[ $ +Ġ+ + +i pe +sh ow +/ d +[ : +ag ement +le v +_ ID +r ary +ad es +_ se +a use +Ġem ploy +Ġ*/ čĊ +Ġf re +Ġ' @ +Ġcomple t +Ġl arge +r al +\ x +Ġf ac +< String +Ġcre ated +up er +.st ate +Ġh ost +ener ic +/ b +( ! +wh ile +i as +B UG +Ġ );ĊĊ +Ġro le +Re g +ĠC olor +St art +Ġp orn +t op +Ġwe b +Ġde v +Ġde al +++ )Ċ +Int eger +pos ition +. on +Ġ( " +ä ¸ +Ġproble m +s v +Ġp ress +AB LE +AT ION +ĠSe e +an ch +Ġth ough +le ep +Ġ< !-- +Ġpoint s +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠ +. J +Ġ :: +p tr +D B +++ ;Ċ +.p ng +n ode +so ft +pon d +Ġe ver +-------------------------------- -------------------------------- +M enu +(' # +Ġs ervices +p g +} )Ċ +param s +Ġact ually +Ġ" / +Em pty +M ethod +Ġid ent +un ic +Ġmill ion +Ġa ff +st yle +Ġcon c +i os +ign ment +UL T +P r +" ;čĊ +Ġunder stand +u ary +Ġhapp en +Ġser ver +ĠC o +S C +Ġle s +Ġfile s +G rid +s ql +Ġof ten +Ġin fo +_ tr +s rc +on y +Ġsp ace +um b +Ġpass word +Ġst ore +, ĊĊ +ĠWh at +g ed +ĠF alse +U s +sw er +_ index +Ġform at +m ost +s m +N ew +Ġd etails +Ġpro b +ĠAN D +() čĊ +il ar +Ġ$ { +ry pt +.C ollections +$ this +ĠF ree +_ of +(f alse +d ated +Ġ> > +Ġf ace +CT ION +Ġs ave +Ġt yp +de v +(" # +AG E +cont ainer +ed it +Q L +Ġitem s +Ġs ocial +i en +ĠRe act +) .ĊĊ +Ġm ar +Ġre du +ĠR E +.p ut +Ġm ajor +C ell +n ext +Ġexpect ed +Ġy et +Ġin div +trib utes +at is +am ed +Ġf ood +S ource +( string +Ġ+ Ċ +it es +d r +Ġmem bers +Ġcom b +item s +ĠP er +T H += True +Ġb ar +_ SE +com m +( w +)ĊĊ Ċ +Ġs end +Ġin c +un signed +F A +Ġparam s +app ing +ro s +ug in +f a +Ġcon nection +Ġ} ;ĊĊ +Ġbe come +M ode +Ġe v +Ġdif f +ĠUn ited +He ight +ful ly +im ages +Ġm akes +Ġg lobal +Ġcont act +' :Ċ +Ġab s +а Ð +f loat +Ġex cept +ĠP ol +Ch ild +t yp +Ġcert ain +i ón +O UT +Ġim pro +ile s +Ġ-- >Ċ +ĠP art +val ues +os s +/ ** +il it +ĠE vent +cur ity +st er +Ġchar acter +Ġnew s +Ġ" , +Ġde vice +c el +log in +he et +Def ault +@ " +ĉ Ġ +c lick +( value +ĠA b +Ġpre vious +ERR OR +oc al +Ġm aterial +Ġbel ow +ĠCh rist +Ġmed ia +co ver +ĠU I +Ġf ail +Ġbl ack +Ġcom ponent +ĠAmeric an +Ġadd ed +Ġbu y +st it +Ġc ame +Ġde lete +prop erty +od ing +Ġc ard +rop s +Ġhttp s +Ġro ot +Ġhand le +C C +B ack +em plate +Ġget ting +_b y +m ail +_s h +. assert +ĠD ec +( true +Ġcom put +Ġcl aim +' => +ĠS ub +Ġa ir +op s +n av +em ents +( id +Ġent er +ang ed +E nd +Ġloc ation +Ġn ight +Ġdo ing +ĠR ed +l in +}ĊĊ Ċ +vid er +Ġp ick +Ġw atch +ess ages +Ġhum an +Ġd am +p end +d ir +Ġt ax +Ġg irl +re et +Ġbo x +Ġstr ong +( v +re l +Ġinter face +Ġm sg +f ect +_ at +Ġh ouse +Ġtr ack +' );ĊĊ +j e +ĠJ ohn +ist r +( S +ub e +Ġc e +itt ed +V ER +* ) +p arent +Ġapp lication +an y +.sw ing +Ġp ack +\ u +Ġpr act +Ġse ction +ct x +Ġun signed +.P oint +ĠO ne +Ä ± +ip le +a id +Ñ ĥ +V ector +by te +Ġw ait +Ġà ł +à ¥ +Ġto gether +Ġth rows +F O +' )) +h ost +is ing +. view +Ġter ms +fr amework +- r +Ġapp ly +Ġs ession +O ptions +ugg est +Ġo thers +w itter +Ġf und +In it +__ ( +ens or +G ET +Ġsever al +i i +[ j +I O +Ġtem plate +P osition +Ġe con +ach ine +Ġ il +.s pring +m ain +el t +im ent +Re c +m m +ĠUn iversity +urs or +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠ +G L +ict ure +ith ub +c er +c ast +F rom +a les +Ġsub ject +p assword +n y +Ġes c +.w rite +ï¼ Į +Wh at +. H +Ġh istory +ĠF e +Ġindiv idual +un it +Ġ-- > +Ġd u +I ST +Ġus ers +f s +f alse +un t +T itle +Ġm ot +Ġf uture +ach ed +Ġstart ed +Ġm ode +Ġ' < +_ array +Ġa x +'] ;Ċ +i res +Th ere +ug ht +t ml +pos ed +ic ult +Ġto ok +Ġg ames +Ġ} } +Ġ? >Ċ +Ġproduct s +I s +Ġb ad +ĠD es +.p ath +' ĊĊ +ĠP ost +av el +( : +Ġneed s +Ġkn own +F l +Ġex ec +Ġse en +um e +Ġb order +Ġl ive +tem p +P er +Ġvar iable +i et +ĠD ef +Ġg e +em e +_b ack +f irst +Ġprovid ed +//////////////// //////////////// +Ġfil ename +Ġh ope +ul y +a uto +f ind +_ string +b tn +it ude +At tribute +Ġyou ng +.t xt +Ġwe bsite +ĠP rop +Ġe y +> ();Ċ +ion al +AR R +iction ary +ur ther +. +t x +Ġp ur +u el +ymb ol +u ation +ang er +Ġback ground +ec ess +ef ined +.... .... +Ġdes cription +Ġrep resent +") );Ċ +press ion +row ser +Ġser ies +ward s +($ _ +a ise +Ġh ot +ac ity +ri es +action s +C reate +ad io +amp les +Ġorig inal +ens ive +f ont +st ream + using +.spring framework +ser ver +Ġb ill +AC K +il ename +Ġfr ame +Ġ= Ċ +Ed it +adi us +Ġd raw +ank s +Ġd eter +Ġcom es +_ int +Ġfore ach +ang le +Ġe lect +pect ed +He ader +ist ration +F alse +ĠG ame +Ġfil ter +Act ivity +Ġl arg +in ition +Ġ" < +is ed +Ġrem ove +ĠTr ans +m et +se e +Form at +Com mand +ĠE X +N one +Ġfr ont +A SE +ĠR ec +ound ation +Ġv o += \" +( * +Ch ange +.W rite +g roup +i ents +u y +******************************** ******************************** +Ġd ig +h r +( - +Ġg en +n umber +ve c +uro pe +ent ry +L L +Ġst e +Val id +'] , +_p aram +Ġse lected +Ġacc ording +ĠD is +Ġ util +B uffer +_ error +Ġass oci +_S IZE +Ġw or +Ġprint f +r ag + ł +D D +ĠV al +Ġact iv +E ng +et ime +Ġv irtual +a ign +a ur +ĠP res +ĠEx ception +Ġany thing +ĠO ff +Ġh ours +Ġw ar +Arg s +ag ing +Ġmodel s +ĠT ime +O b +am s +j oy +Ġear ly +. read +Ġc enter +ĠIn itial +Ġl anguage +l ength +x y +Ġs n +Ġin f +P ost +Ġag o +Ġeas y +_c ode +ĠAN Y +_ ch +Ġdown load +( T +av ed +âĢ ĵ +Ġstud ents +Ġf ig +l ight +x x +Ġbu ffer +ĠD ep +ĠM ath +IT H +Ġvar i +Ġd ue +F actory +Ġp or +Ġe p +ot ype +Ġcan not +Ġwh ite +< int +ter n +Ġreg ister +Ġpre d +cl us +_d ate +Ġ/ ** +Ġa uth +Ġ[ ]Ċ +Ġper iod +n own +Ġv ot +Ġs creen +' d +T ypes +Ġt mp +е Ð +ur al +Ġben ef +_ y +Ġn et +ĠSt ates +'] [' +ĠN e +ĠN OT +Ġn eg +Ġcomm on +s cope +Ġc red +g es +_T YPE +Ġs uggest +o om +.ĊĊ Ċ +Ġac cept +Ġr andom +er m +ĠV ector +w ith +T ER +( str +Ġres pons +Ġh it +.S et +gr id +ri a +Ġc lick +und le +C ase +ins ert +Util s +Ġ"" " +Ġim plement +at al +tem pt +tem plate +oc r +return s +Ġplay ers +us ers +ed ef +ĠTh ese +Ġam ong +Ġde b +h a +.get Element +Ġc irc +Ġan swer +Ġw alk +Ġt reat +ĠG e +ĠC reate +Ġa ge +Ġre q +O ST +ang ular +Ñ ı +Ġf ive +Ġdistrib uted +Ġfri end +T P +Ġc lean +ow s +.Control s +d is +Ġw ords +. io +z y +Ġhe ader +ĠC heck +âĢĻ m +j ust +h older +=" čĊ +. annot +Ġcol lection +' . +Ġsim ilar +Ġt aken +(" % +Or der +'] Ċ +-m d +ĠT H +ac ed +Ġis n +/ j +Ġs on +gr aph +ĠInt eger +Ġn ecess +re en +Ġ um +Ġ\ < +Ġmom ent +Ġbr ing +Ġind ic +ys is +Le vel +ver se +urre nc +_t est +Ġent ire +D own +Ġ}ĊĊ Ċ +( result +ĠRe ad +à ¨ +M od +Ġtry ing +") ,Ċ +Ġm ember +ĠC or +OD O +- control +un time +ĠS im +D ialog +pl ot +_ on +Ġph ys +} / +Ġn amespace +ĉ čĊ +ac c +Pl ayer +A RE +Ġf oot +Ġbo ard +p art +Ġs us +w ise +ĠM c +Ġp ush +AT A +Ġp lease +ri ed +we et +b it +id ed +V E +ĠS w +U B +Ġt ypes +ed ia +Ġc los +ace book +Wh en +Ġed it +ig ger +Ġen erg +Cont ainer +Ġph ot +ĠC ount +ĠE urope +.I s +ĠR uss +pe ed +ĠS tr +Ġp y +Ġc ult +Ġdef ined +cc ount +Ġob t +.L ocation +Ġth read +il le +Ġinst ead +str ong +ĠS ec +U RE +Ġide a +. se +em y +select ed +Con nection +ac ing +th read +.n ext +Ġc oll +Ġfil m +ist ic +Ġcomp et +Ġcon n +th ough +Ġcom pan +ock et +Ġte ach += ( +Ġph one +Ġact ive +de lete +tr ies +Ġm o +Ġde ath +} );ĊĊ +oc ol +W idget +Ġart icle +ro du +and id +Ñ ĭ +ĠC r +k a +() : +lo od +ĉĉĉ Ċ +Ġal most +Ġs ell +erv let +ri p +Un it +Ġapp lic +Ġcon nect +Ġfe ature +Ġv ia +' ), +Ġl im +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +ĠG u +Eng ine +Ġen s +Ġen vironment +b lock +HER E +N ULL +g y +t ag +) ). +ex p +Ġcom pl +Ġinst all +Ġcomple te +que ue +atur al +Ġgener al +th on +Ġask ed +o res +( res +Ġres erved +S P +ĠâĢ ¦ +Å Ĥ +Ġsign ific +O ff +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠ +ĠA g +ĠJ ust +ĠE rror +Ġin fl +ad ata +Ġ icon +ask s +' ' +_ LO +? . +ac count +Ġ( * +' )ĊĊ +r ap +_ var +ĠF OR +Ġpart y +ĠY our +c at +str y +. new +bo ot +ĠN ov +Ġv ector +Ġn ormal +Ġf urther +Re pository +Ġd atabase +att le +Ġmus ic +Ġspe ed +Ġd oc +pro cess +IG HT +.p arse +Ġt aking +Ġvi ol +ce ed +ĠA fter +Ġfor ward +Ġc rit +"/ >Ċ +ro t +Ġfa iled +ef ore +Ġconc ern +o e +b a +Ġs ender +Ġter m +h as +=" # +Ġpot ential +N um +Ġpublish ed +.c lose +ĠIm age +str aint +U D +ĠO b +Ġprob ably +l im +" :Ċ +olum e +Ġcon sum +ag ue +ens ions +Ġinvest ig +- year +') ; +-s m +Ġen joy +or ig +er ing +c p +le ased +ple ments +Ġreturn s +p at +B O +ĠH ouse +.L abel +Ġwe ight +igh b +Ġcondition s +Ġex ception +d escription +Ġtr ad +- to +Ġ{ } +Ġmod ule +EN D +. ap +.p rops +Ġcon structor +av es +Ġf avor +ĠN ow +; i +ĠM ain +_ k +er ies +âĢĻ ll +trans form +imest amp +P re +Ġm er +. res +st ant +L ocation +_N AME +Ġlos s +Ġ ĊĊ +n et +Ġeng ine +B lock +Ġiss ues +Ġpar se +ĠB ar +Ġst ay +ĠJ SON +Ġd om +air s +w ner +Ġl ower +", čĊ +ĠD em +uf act +Ġp s +Ġper fect +R L +Ġed uc +l s +em ory +ARR ANT +u ge +Ġex act +. key +al led +e ch +ie f +\ / +o ke +Ġfor mer +al loc +Ġs ix +id a +Ġm argin +Ġhe art +al d +p ack +.getElement ById +ĠW ARRANT +Ġr ather +Ġbuild ing +er man +lic e +Ġquest ions +iz es +le ge +irect ory +Ġj e +Ġc as +pro ps +ut f +Ġse curity +Ġhow ever +we ight +Ġins ide +Ġpres ident +Ch ar +ĠW ITH +.m ap +Ġgr aph +Ġt ag +_st atus +Ġat tempt +op p +us es +ĉ const +Ġr ound +, $ +Ġfri ends +Em ail +? > +Res ource +KE Y +os p +. query +ĠN orth +able s +ist rib +_c lass +el lo +Th at +Ð º +pecial ly +ĠPres ident +Ġcamp aign +Ġal t +are a +Ġch all +Ġop port +.C on +Ġenerg y +li ke +. string +ing ton +) * +y y +Ġprof ession +ir th +Ġse g +æ ľ +Ġh or +i ers +c an +Ġbeh ind +Pro duct +f g +ĠS k +.j pg +? : +] ;ĊĊ +Ġcall back +ĠH ttp +Ñ Į +l ong +M S +AT H +Ġr aise +Ġwant ed +row n +ut or +l t +] = +el ine +M A +Ġse par +c s +se mb +D is +bs erv +ĠW ill +Ġpol icy +Ġth ird +ph one +Ġb ed +/ g +. __ +ĠIn c +iz ing +.re move +in stance +.t ype +Ġs erv +E ach +Ġh ar +ĠM essage +( key +SE LECT +P os +)) ;čĊ +Ġre comm +Ġtr aining +ĠE nt +ĠCh ar +ic ht +(f ile +Ġp rior +G ame +Ġex it +Param s +.c ore +P C +n es +anc ed +( request +P assword +} >Ċ +Ġm ag +Ġre lease +Ġsh all +ud ent +ĠS outh +and o +: ' +.Tab Index +s k +ann er +is set +Ġout side +led ge +Ġ å +ĠR ob +Ġim m +! Ċ +ĠWe b +D es +B C +anc ial +R oute +D ec +fer ences +Ġp urch +ĠM odel +ct or +g n +_st art +_ un +. * +is es +Ġg round +Ġun ique +Ġbe aut +{ " +Ġp our +ĠO ct +Ġt ree +set s +_ res +') -> +_re g +(" \ +Ġby te +B l +Ġd ating +Ġm atter +ĠR em +Ġ' ../ +ĠA ug +ĠL a +Ġ$ ( +ourn al +i am +Ġshow s +w rite +Ġb all +Ġsim ply +Ġf ast +Ġmem ory +A SS +ĠO f +ov ed +ant e +a ul +ist ry +)) );Ċ +Ġf it +< string +Ġpolit ical +anc el +_ . +c ard +.c urrent +o ch +_ image +\ t +# Ċ +( L +Ġindu stry +com ing +Ġex tra +Ġreport ed +.st art +Ġres ources +Ġim g +fl ow +_E X +(n ull +ĠP re +Ġwr ong +inter face +Param eter +n ers +á » +t ure +ers ist +oun try +Ġseem s +al ance +de st +ĉ String +Ġm aint +Ġun it +act ers +ĠT R +if ul +export s +pro ject +App lication +leg ate +Ġt akes +ter m +Ġet c +ust er +Ġappe ar +add ress +Ġf em +h s +Ġh om +, - +Ġdiff icult +Ġcom ing +O pen +Ġset tings +ĠW ar +ĠTh en +Ġaut om +ĠF oundation +Ġqu ite +D escription +Ġb log +i qu +P S +_f ield +J son +SS ION +ĠS ch +ĠL O +Ġdes cri +Ġevery one +Ġpret ty +Ġlong er +Ġm enu +Ġcurrent ly +se c +Ġrelations hip +################ ################ +ĠM ap +as et +Ġparam eters +Ġcr ush +" čĊ +IL ITY +ig ration +Ġc out +t otal +Ġn ames +nd ef +") ; +ri end +yn amic +Ġeff ort +Ġact ual +Ġfield s +O UN +t ers +Ġf ix +_m odel +Ġc ases +C A +M y +Inter face +ĠS E +] ] +al le +ĠN ational +ĠArray List +in line +. V +ar a +ref ix +as c +Re ader +ĠÐ ¿ +ast ic +( () +C l +.annot ation +Ġperform ance +ail y +.to String +.n et +view s +. end +ay ers +l ate +ĠA pr +ed eral +'] ) +.b ody +Ġhigh er +_f l +c r +al ert +_n ode +ĠG oogle +Ġit self +A uth +urrenc y +Ġsignific ant +app end +Ġres pect +str ap +Ġun a +riter ia +P ORT +.ap ache +Out put +Ġpro gress +Ġm id +ĠM icrosoft +Ġres ource +ab lish +Ġd im +. load +.A pp +Ġd irection +Ġadd itional +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠ +Ġnum bers +Ġcompan ies +.T h +Ġs ound +user name +Ġstat ement +Ġal ert +Ġcon tract +h ome +_l ength +.Com ponent +e v +. Ex +ï¼ ļ +" ; +ĠH igh +Ġ )ĊĊ +ĠP oint +op h +Ġl ines +-> _ +" )ĊĊ +o x +app lication +Ġ ]Ċ +ĊĊĊĊ ĊĊ +Ġso on +ction s +ing er +Ġj oin +ĠP e +Ġ ë +Ġl as +. E +c ss +/ or +ĠSt art +ĠT O +Ġsub s +con n +com ponents +DE BUG +qu are +F unction +end ar +. index +Ġf ill +Ä Ļ +Ġcho ose +h ow +ĠAmeric a +ass ets +-------- ---- +ĠV alue +Ġoff ice +Ġv eh +Ġtrans form +ĠAr t +Ġin de +Ġf n +Ġim plements +ang o +ple te ++ " +t mp +am ily +Ġhas h +miss ions +E ST +g t +Pro vider +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠ +Ġfl ag +Ġpartic ip +d en +ĠReturn s +Ġnot e +ü r +p m +ide os +Ġspec ified +ĠE N +est er +ol id +Ġup on +( std +ĉ v +Ġ' \ +u z +Ġv ert +Ġv ict +ĉ self +Ġ" $ +. k +Ġgroup s +g ithub +l ang +Ġm ut +T O +Ġv e +ĠP lease +;ĊĊ Ċ +ac cess +Ġ{ " +re a +Ġr isk +ick er +og gle +ĉ while +AN G +.s end +Ġwom an +Ġget s +Ġ ign +ĠI d +_ log +ON E +Ġe vid +ĠH ar +_s ub +Ġend l +Ġinclud ed +() );ĊĊ +ĠA p +ig r +Ġs em +ĠBl ack +d oc +_t able +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +- up +Ġca use +Ġ .. +Ġv an +_d ict +Ġf ocus +IN D +CE SS +.L og +Ġmult iple +id o +Ġreg ard +- M +and ler +our se +Ġde g +. U +Ġadd ition +Ġvar ious +Ġrece ive +е н +ĠH T +Ob j +D F +Ġincre ase +ĠO pen +] ; +Ġcomm it +? Ċ +ateg ories +at ory +sh ip +ĠM ich +Ġh tml +rom ise +Ġle ave +Ġstr ateg +av en +ĠCon sole +k nown +- n +_ LE +.com ponent +Ġb re +S ession +i ance +Ġal ign +typ edef +_ result +ĠW HERE +.s plit +Ġread ing +FA ULT +Ġc lo +Ġnot ice +_p r +ar ter +Ġlo ck +Ġstand ard +et ic +ell ow +Ġp adding +ĠH is +Ġst ates +_c ast +( P +a a +Ġintern al +e an +ĠP RO +ĠK ey +Ġes pecially +m ing +Ġc ross +Ġn ational +_ object +f ilter +Ġs cript +. update +_ i +ĠAss ert +/ core +%% %% +Ġproble ms +ist or +Ġ. = +Ġar ch +Ġwrit ten +Ġm ilit +M ENT +. ch +ca pe +ĠM us +_ config +ĠA PI +fo ot +Ġim ages +end l +. In +F irst +Ġpl atform +.pro t +O ption +st e +ĠT ODO +Ġfor ce +. cont +ĉ echo +ĠD av +P tr +( B +R T +ĠB ase +] [' +Ġann ounc +con sole +ĠP y +d s +. as +Ġpre vent +ap an +Ġ{ ' +} ' +Ġde ad +V AL +Q UE +**************************************************************** ******** +Ġch arg +R eturn +Ġf ul +d om +Ġr ules +Ġmod ify +Ġe val +h am +at ement +\ < +ul a += False +R A +Ġcont ains +Ġst ack +m ar +Ġ{ }Ċ +Ġund efined +A ss +ĠCh ina +ve y +* Ċ +Ġplay ing +) / +act or +Ġb ottom +li er +ĠN umber +Ġcou ple +D C +ĠS O +g or +.set Text +s uccess +com mand +F ilter +ĠO ur +_ item +Ġc tx +Ġro ad +V ersion +c ase +ur t +av ior +y ch +semb ly +ĠPro duct +Ġh eld +a fe +Ġinclud es +< quote +Ġa void +ĠF in +ĠM od +Ġt ab +an o +à ± +ipp ing +- e +Ġins ert +t arget +ch an +.M odel +IM E +\ Ċ +Ġm achine +av y +ĠN O +ĠInt er +Ġoper ation +mod al +T ag +] : +Ġprodu ction +Ġare as +Ġre n +_f rom +n bsp +Ġoper ator +m en +app ed +_p er +z en +(" . +.s ave +=" {{ +Ġt or +( response +Ġc andid +Ġcon v +a iled +ĠL ib +com p +ur a +ï¿ ½ +ĠH ere +Ġarg ument +h ood +Ġest ablish +ograph y +Ġon Click +amb da +Ġs ch +Ġmov ie +Ġse c +Ġact ivity +Ø § +Ġs ql +_ all +inc ip +Ġprovid es +Ġs ys +ack et +Ġwas n +Ġus es +ĠF unction +.g oogle +ĠRes ult +Vis ible +ag ma +el come +ĠS y +ĠC ent +AL SE +ac ión +EX T +Ġl icense +ĠL ong +Ġacc om +Ġab ility +. height +Act ive +olog ical +ol y +)) , +.S e +Ġparam eter +pr ite +AB ILITY +.s ervice +ĠG roup +_ query +ĠI tem +in ing +Ġj ud +im s +f ix +ind er +ag ram +Ġfunction s +Ġexper i +ĠE m +Ġro t +Ġp en +.b tn +ĠA S +#if def +Ġcho ice +ĠP age +_P RO +Q U +å ı +ant ity +Â Ń +word s +Ġread only +Ġf lex +prot ected +ĠAn y +Ġchar acters +enc ed +ĠJ uly +il er +C ard +ur ance +Ġre v +.e vent +al y +Ġwon der +ĠP ort +Ġleg al +ro le +Ġt en +Ġgo es +M P +wh ite +): čĊ +)) čĊ +Ġre ference +Ġm is +ĠPro ject +ick s +> & +C ON +Ġre pl +Ġreg ular +St orage +ram ework +Ġgo al +Ġt ouch +.w idget +Ġbu ilt +d es +P art +( re +Ġw orth +h ib +g ame +ĠÐ ² +ac ion +ĠWh ite +(t ype +( ` +Ġn atural +Ġin j +Ġcal cul +ĠApr il +. List +Ġassoci ated +ĉ System +~ ~ += [ +Ġst orage +Ġby tes +Ġtr avel +Ġs ou +Ġpass ed +! = +as cript +. open +Ġgr id +Ġb us +Ġrec ogn +A b +Ġh on +ĠC enter +Ġpre c +b uild +HT ML +ĠS an +Ġcoun tries +a led +t oken +k t +Ġqu al +L ast +ad ow +Ġman ufact +id ad +j ango +N ext +x f +. a +Ġporn o +ĠP M +er ve +it ing +_ th +c i += None +g s +Ġlog in +at ives +'] );Ċ +Ä ħ +Ġ ill +I A +child ren +D O +Ġlevel s +Ġ{ { +Ġlook s +Ġ" # +To String +Ġnecess ary +ĠĠĠ Ċ +c ell +En try +Ġ' # +Ġext rem +Select or +Ġplace holder +L oad +Ġre leased +O RE +En umer +ĠT V +SE T +in q +P ress +ĠDep artment +Ġprop erties +Ġres pond +S earch +a el +Ġre qu +ĠB ook +/ Ċ +( st +Ġfin ancial +ick et +_in put +Ġth reat +( in +Str ip +ì Ŀ +ç ão +Ġevid ence +)) ; +ĠB ro +Ġ[ ];Ċ +Ġ ou +b uf +S cript +d at +Ġr ule +# import +=" / +S erial +Ġstart ing +[ index +a e +Ġcon trib +s ession +_ new +ut able +o ber +Ġ" ./ +Ġlog ger +Ġrecent ly +Ġreturn ed +č čĊ +)) )Ċ +ition s +Ġse ek +Ġcomm unic +Ġ" . +Ġuser name +E CT +D S +Ġother wise +ĠG erman +. aw +Ad apter +ix el +Ġsystem s +Ġd rop +Ġstruct ure +Ġ$ ("# +enc ies +ann ing +ĠL ink +ĠRes ponse +Ġst ri +Å ¼ +ĠD B +æ Ĺ +and roid +sub mit +ot ion +( @ +.t est +ĊĊĊĊ ĊĊĊĊ +] ;čĊ +Ġdirect ly +Ġ" % +r is +el ta +A IL +) {čĊ +m ine +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠ +( k +b on +as ic +p ite +__ _ +M ax +Ġerror s +ĠWh ile +Ġarg uments +Ġens ure +R ight +-b ased +We b +Ġ- = +Ġint rodu +ĠIn st +ĠW ash +ord in +j oin +D atabase +Ġgr ad +Ġus ually +IT E +Prop s +? >Ċ +ĠG o +@ Override +RE F +Ġ ip +ĠA ustral +Ġ ist +View ById +Ġser ious +Ġcustom er +.prot otype +od o +c or +Ġdo or +ĠWITH OUT +Ġpl ant +Ġbeg an +Ġdist ance +() ). +Ġch ance +Ġor d +c ame +pr agma +Ġprot ect +rag ment +ĠN ode +en ing +Ñ ĩ +Ġr oute +ĠS chool +h i +Ġne ighb +A fter +lic it +Ġcon tr +Ġpr imary +A A +.Write Line +util s +Ġb i +R ed +.L inq +. object +Ġlead ers +un ities +Ġg un +on th +ĠDe v +F ILE +Ġcom ments +_l en +ar row +am ount +R ange +s ert +Grid View +Ġup dated +ĠM o +Ġin form +oci ety +al a +A ccess +Ġh ab +Ġc reat +_ arg +ĠJan uary +ĠD ay +") čĊ +up le +d ocument +gor ith +m enu +ĠO ver +b b +.t itle +_ out +Ġle d +ur i +Ġ? >Ċ +r un +Ġsc ene +( array +de vice +_t itle +ag on +] čĊ +ab y +Ġbe came +bo olean +Ġp ark +ĠC ode +up load +rid ay +ĠSept ember +F e +Ġs en +c ing +F L +C ol +ut s +_p age +in n +Ġim plied +al ing +Ġyour self +.C ount +con f +Ġa ud +_in it +. ) +Ġw rote +N G +. Error +ä » +.f or +Ġe qual +ĠRe quest +Ġser ial +Ġallow s +X X +Ġm iddle +ch or +à ¸ +erv al +.C olumn +read ing +Ġesc ort +ĠAug ust +Ġquick ly +Ġwe ap +ĠC G +rop ri +h o +Ġc op +( struct +ĠB ig +Ġv s +Ġfre qu +. Value +Ġaction s +Ġpro per +Ġin n +Ġobject s +Ġm atrix +av ascript +Ġon es +.g roup +Ġgre en +Ġp aint +ool s +y cl +enc ode +ol t +com ment +. api +D ir +Ġun e +iz ont +.p osition +Ġdes igned +_ val +av i +ir ing +t ab +Ġl ayer +Ġview s +Ġre ve +ra el +ĠO N +r ics +n p +Ġc ore +() );čĊ +M ain +Ġexp ert +ĉĉ čĊ +_ en +Ġ/ > +ut ter +I AL +ail s +ĠK ing +*/ ĊĊ +ĠM et +_ end +add r +or a +Ġ ir +M in +Ġsur pr +Ġre pe +Ġdirect ory +P UT +- S +Ġe lection +h aps +.p re +c m +Val ues +Ġ" Ċ +c olumn +iv il +Log in +in ue +Ġbeaut iful +Ġse cret +(e vent +Ġch at +um s +Ġorig in +Ġeffect s +Ġman agement +ill a +t k +Ġset ting +ĠC our +Ġmass age +ĉ end +Ġhapp y +Ġfin ish +Ġc amera +ĠV er +ĠDem ocr +ĠH er +( Q +con s +it a +Ġ' . +{ } +ĉ C +Ġst uff +Ġ :Ċ +ĠA R +T ask +h idden +er os +IG N +at io +ĠHe alth +ol ute +Ent er +' > +ĠT witter +ĠCount y +s cribe +Ġ= >Ċ +Ġh y +f it +Ġmilit ary +Ġsa le +re quired +n on +boot strap +h old +r im +- old +ĠD own +Ġm ention +cont act +_g roup +od ay +Ġto wn +Ġsol ution +u ate +ell ing +] -> +ot es +ent al +om en +osp ital +ĠS up +_ EN +Ġsl ow +SE SSION +Ġbl ue +ag o +Ġl ives +Ġ ^ +. un +in st +en ge +Ġcustom ers +Ġc ast +ud get +ï¼ ģ +ic ens +Ġdeter min +Se lected +_ pl +ue ue +Ġd ark +// ĊĊ +s i +ther n +ĠJ apan +/ w +P U +ĠE ast +ov ie +Ġp ackage +Ġn or +Ġap i +b ot +" ];Ċ +_p ost +ul ate +Ġcl ub +') );Ċ +Ġlo op +PI O +ion e +sh ot +In itial +Ġplay ed +reg ister +rou ght +_m ax +ac ement +m atch +raph ics +A ST +Ġexist ing +Ġcomple x +D A +.C h +.com mon +m o +Ġ' ../../ +it o +Ġanal ysis +Ġdel iver +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ Ċ +id x +à ł +ong o +ĠEng lish +< !-- +Ġcomput er +EN SE +Ġp as +Ġr ais +H ash +Ġm obile +Ġo wner +F IG +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +th es +Ġat tr +w d +.t ime +aw n +Ġtreat ment +ĠA c +. View +im pl +m ore +p ass +Ġh a +.f rom +Ġle ading +FF FF +( error +. ui +at ar +ad ers +d ates +Ġz u +Ġfl ow +T arget +Ġinvol ved +Ġi o +par se +$ _ +he st +. int +- item +as y +S p +Ġsh ift +N T +Ġt f +_T R +. web +C S +Ġ} ) +Ġey es +_ z +' );čĊ +if orn +Ġ{ @ +Ġn ice +.l ist +ĠĠĠĠ čĊ +Ġf loor +Ġred irect +ĠU K +( [' +Ġw ish +Ġcap t +leg al +ĠI O +Ġst age +. String +ĠA fr +ig en +ĠS H +De lete +ell s +Ġsol id +Ġmeet ing +Ġwork ed +Ġed itor +in y +Ð ¼ +_ read +. Id +e ff +Off set +ch a +US ER +ĉĉ ĠĠĠ +ipp ed +Ġd ict +ĠR un +.h pp +Ġan g +x ml +im ple +Ġmed ical +_t oken +con nect +Ġh our +Ġcont roller +_m essage +U ID +G r +and ed +_C H +Ġbook s +Ġspe ak +am ing +Ġm ount +Rec ord +ĉ struct +.W eb +ond on +Ġ// Ċ +Ġf elt +.A uto +id ge +_p os +P R +Ġmod ern +C ollection +_m sg +C D +ĠL o +Ġsecond s +ib ly +.e quals +Ġintern ational +# pragma +oo th +W riter +i ate +Ġce le +ĠB it +iv o +iv ery +r d +HE CK +Ġc ache +.c ount +Ġro ll +.Re ad +RE D +Ġset up +izont al +model s +arg v +Ġconsider ed +=" ../ +set tings +ĠR el +Ġgrow th +Ġm ix +ĠWash ington +Ġpl t +ĠI M +á º +Ġturn ed +ĠDate Time +ĠW ed +( url +Ġ" - +Ġlet ter +As ync +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠ +ĠOct ober +_l ine +Ġatt ention +Ġcol lect +ĠH ash +Ġim ag +T ree +Ġsit uation +et te +_n o +IV E +Ġv on +.t arget +Ġknow ledge +Ġdr ive +.p ost +Ġb lood +Ġc it +pr imary +Ġconfig uration +te e +Ġph oto +is ode +Tr ace +Ġg ave +Ġsh ot +ĠA ir +Ġm other +pr ice +Ġmor ning +)) {Ċ +- x +Ġtr ade +Ġdes c +Ġ&& Ċ +Ġparent s +A pi +å Ī +t ed +w er +Ġ æ +Ġs y +ĠK e +Par ser +å ħ +anc y +Ġpie ce +iforn ia +to String +r an +id ing +PT ION +com es +/ lic +.c lient +E l +L ong +Ġprofession al +ru pt +v a +Ġcomplet ely +Ġpract ice +Ġse lection +R em +in i +Ġc am +RE E +Ġsit es +p a +AT US +Ñģ ÑĤ +arr ant +* ( +_ KEY +ĠB utton +ĠF riday +se qu +Ġre ader +Ġm essages +è ¯ +Ġbu f +K e +Ġn ov +H P +M sg +al ign +ar ily +Ġ' , +_w ith +Ġd as +Ġhe ard +at omic +ri al +) [ +Ġdis e +@ end +Ġg old +Ġf air +Ġsa les +. Button +str ict +s ave +Ġme asure +Ġ" + +ec ause +View Controller +ĠT able +.p aram +Ġdec ided +(( ( +IN FO +Ġopport unity +T e +IC ENSE +cc ording +k i +ĠU N +Ġcont ain +Ġman ager +Ġp ain +ĠF ire +rom e +Ġpl ans +F ound +l ay +ĠDec ember +Ġinfl u +à º +ren ch +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ Ġ +az ing +b rief +c all +wo od +Ġload ed +Ġgr and +/ f +im p +_ U +ST R +âĢ ¢ +Ġcred it +.C olor +or ge +QUE ST +Ġdiffer ence +ĠP C +w args +Ġp ub +und ay +Ġf ra +.m ax +Ġtri ed +ann els +s end +Ġreport s +Ġad ult +ä º +Ġcons ist +ĠSt reet +ĠPro gram +S QL +M atrix +ounc il +- A +ĉ w +Ġwho se +Ġrel ig +ĠS ex +Ġg ives +n one +.m essage +( G +.aw t +- right +ĠNov ember +ell ig +ut ive +Ä ĥ +over n +Ġeas ily +Ġide as +ĠÐ ½ +/c ss +ly ing +el le +C an +_c olor +оР² +Ġp air +ng th +Ġs plit +d rop +art y +on a +Ġcap ital +Ġhe ar +Ġex ists +ĉ log +em o +R un +o i +Ġpar ser +ĠM ethod +Ġeduc ation +[ k +Ġlib rary +> ";Ċ +_ UN +ĉ std +od ed +Ġcall s +h ere +R el +Ġbr and +back ground +g a +_add ress +_param s +C ategory +ĠInd ia +_e vent +Ġ ing +R ender +.c l +ump y +Ġp et +F C +ĠA nt +Ex t +Ġchar ge +en ed +gr ad +E O +Ġdep end +Ġ .ĊĊ +fr ame +Ġd f +Ġh uge +ĠP ART +ed s +; ; +ĠA M +Ġbas ic +ĠL et +lic h +Ġar m +Ġst ar +Ġf ederal +W ork +Ġcar ry +ĠIs rael +( obj +={ { +Ġs aved +Ġs yn +Ġconst ant +V ENT +Ġpos itive +Ġcon duct +Ġsk in +Ġear lier +Ġl ayout +ĠI P +O UR +Ġt im +styles heet +_ cl +ĠC ard +++ ){Ċ +Ġtem per +ĠDav id +ĉ try +.d art +Ġwant s +Ġp icture +Ġv ideos +ĠCom m +is ions +_M AX +M apping +- content +ĠE ar +- de +Ġpre m +br uary +Ġcom ponents +Ġthrough out +Ġp ull +Ġp ages +ent e +res pond +Ġg as +cript or +Ġed ge +Ġb ound +A CT +**** ** +Ġcre ating +ĠC H +Ġnull ptr +B r ++ ' +.c o +> :: +Ġle arning +.L ength +_S H +Ġpat ients +A IN +Ġk ids +Ġcom fort +Ġsh own +ug ins +ĠB ack +ell a +_C L +Ġl at +Ġdis patch +Ġclass es +. at +.b egin +Ġsuccess ful +b an +Ġobt ain +ĠS l +Ġl ack +iter ator +Th read +(s ize +Ġn one +.h as +_ X +s ort +n ap +p et +b in +ĠCan ada +The y +Ġd ans +ĠM at +< td +Ġh air +Ġ' ',Ċ +Ġc u +Ġlaw s +let ed +p ed +Ġp ow +Ġk new +_C OM +_ , +ĠM ag +id ents +( req +Ġ ), +- center +Ġw ide +ĠA uthor +st ants +Ġjob s +Ġm ath +et imes +Bo olean +Ġs cope +_ is +Ġme as +Ġkey s +el ay +Ġexact ly +'=> ' +ĠP aul +m as +ĉ print +(l en +f d +Ġ) ; +. Event +q li +ir it +ield s +om an +ĠT op +Ġv ote +Ġm ask +Ġthem e +- Ċ +Ġpro ps +Ġf ine +Ġwrit er +_ offset +c ar +Ġal tern +Ġc opyright +Ġdest roy +pp er +Ġgener ate +pp ed +âĢĻ d +ĠĠĠĠĠĠ Ċ +m ake +ĠSh ow +Ġb rowser +Ġfavor ite +Ġcare er +Ġhappen ed +( char +Ġrecomm end +Ġl iter +.f ilter +gr ade +Ġ £ +Ph one +om s +Ġn amed +- label +ip o +ĠO ther +Ġp anel +Ġro ck +S cale +ĉ assert +Ð ´ +Ġtr ust +fr ont +Ġdem on +A r +N et +Ġecon omic +foot er +Ġr ace +(n ode +ĠO ption +s plit +Ġphys ical +if est +Ġrem oved +. http +)) ,Ċ +Ġlook ed +' ; +d ing +g est +atur day +/lic enses +Pr ice +Ġd ro +Ġto wards +Ġun s +ĠC L +ĉ static +Ġ rows +Ġdef ine +.re place +Ġf ather +ĠDes ign +ass ign +m ut +De vice +D id +') )Ċ +omet ry +ay load +Ġh istor +ĠP aram +ĠBo olean +Ġn ature +Ġj s +Ġn ation +i h +Ġdis cover +se m +Hand le +ĉ r +ĠTe chn +Ġw all +{ $ +@ property +Ġ" ../ +Ġex am +.d raw +opp ing +Ġnear ly +Ġco ol +Ġinde pend +RE S +Ġhand ler +ĠMon day +Ġs un +St yles +ous ly +Ġ ĉ +v est +D isplay +( y +atic ally +Ġpred ict +y ing +Ġsom etimes +" ]Ċ +Ġdr ink +Ġb ul +ific ations +. insert +.re g +Ġtest s +Al ignment +Ġal leg +Ġat tribute +ĠN ote +Ġmy self +art s +N ow +Ġinterest ing +li ents +Ġpop ulation +ĠCal ifornia +" I +å ¹ +Ġgre ater +ues day +Ġth ous +Ġcost s +Ġla unch +\ Http +k er +b and +ĠPl ay +Ġb and +.sh ape +es ome +art icle +.r f +Ġw er +á s +em bers +us r +B A +ic an +et t +valid ate +ult i +Ġimmedi ately +z er +Ġfig ure +o es +ell er +irc le +ĠS ign +.d b +Ġr ank +By tes +Ġproject s +_re c +UL AR +A PI +ĠL ine +P ort +Ġp oll +Ġg iving +id ence +-- Ċ +Ġpl ot +ic ial +Ġw arrant +IT ION +ĠD ouble +Ġbill ion +gorith m +Ġequ ipment +D ATE +Ġ@ " +E E +Ġp le +i ation +Ġhead ers +Ġpro ced +.Component Model +ĠOb ama +Ġp a +ĠB est +im ately +.get String +. \ +mp loy +Ġr aw +_b lock +und red +" },Ċ +.Group Layout +Ġb rought +NS String +th row +cre ated +.N ew +_ view +C P +ep s +O p +Ġgr atis +Ġ' " +Ġinter view +"" "Ċ +Ġpart ial +Ġa ria +b ing +A uthor +Bo ok +ĠP at +um an +Us ers +pl us +ĠD irect +ven ue +al pha +UC CESS +ĠC all +Ġ );čĊ +im ated +Ġrem ain +Ġant i +ĠL ondon +Ġsaf ety +PO SE +o les +cont roller +By te +ĠCour t +ĠPh il +ĠAss oci +en a +å IJ +_ST R +co in +resh old +Ġb atch +_C lick +entic ation +> ';Ċ +ent y +Ġbegin ning +Ġz ero +ĠCon vert +Ġt err +Ġp aid +Ġincre ased +c atch +-s ize +act ivity +e quals +Ġque ue +Ġ" ' +ĠIntern ational +Ġf ür +urs day +Ġsc ient +all ow +ax is +Ġapp ropri +ed ge +Ġid x +S uccess +ent ifier +: \ +x is +Ġmax imum +ark s +Ġb irth +( index +Ġmay be +.p y +file s +Ġlim ited +_ check +lo ok +pl ies +Ġmov ement +'] . +Ġbro ad +ĠB E +ĠUn ityEngine +.c pp +ĠE very +Ad min +Ġf ans +p ared +Ċ ĠĠĠĠĊ +Ġfore ign +Ġp an +Ġt our +ĠOr der +Ġmov ing +Ġa uf +C all +c b +Å Ł +vent ory +ĠS ql +Ġful ly +Click Listener +W ORD +Ġannounc ed +) čĊčĊ +Ġagre ed +ri e +Ġe arn +_l ink +. array +(t ext +Ġmaterial s +, p +ff ff +v g +Ġ © +Ġun less +aj ax +LO G +Ġsex ual +Ġ\ " +- time +Ġco ach +Ġsupport ed +Ġphot os +if orm +.C reate +) ] +ri er +Ġd ialog +av er +ig e +) + +_id x +: [ +_m in +ĠC ong +Ġpress ure +Ġteam s +S ign +b egin +ri an +NE SS +L S +Ġimpro ve +ĠS unday +Ġdef inition +ig er +roll ers +Ġthink ing +T emplate +- F +Ġem erg +pl ates +ĠUS A +.set State +ĠAl so +re v +Ġen able +ĠC O +PE CT +Ġcon cept +) - +ĠâĢ ¢ +Ġset s +Ġmean ing +em on +ĠCon s +c mp +ed er +ann ed +icens ed +ĠS uper +Ġd aily +Ġmult i +_ u +Ġchall eng +_m ode +ĠP romise +Ġstr ict +j o +int on +( list +On ly +> { +Ġveh icle +í ķ +ĠPl ayer +ĠD el +Ġp ool +. url +nes day +();čĊ čĊ +Ġ" );Ċ +L ocal +. ");Ċ +Ġorgan ization +re nder +ĠApp lication +Ġsum mer +ex pected +N A +Ġr ap +_ obj +Ġsur face +ĠP UR +Ġ}, ĊĊ +Ġvariable s +(m essage +Ġop in +.b ack +а н +Ġwork ers +v m +C o +ught er +Ġm aster +Ġ" ", +Ġst ories +. User +Ġcele br +ines e +B S +ĠCom mand +ash board +Ġo g +k g +. image +.st yle +Ġstep s +ĠB en +( args +ĠP erson +, y +Ġofficial s +| Ċ +Ġsk ills +v c +Ġbuild er +Ġg ar +A ccount +ĠA uth +ç Ķ +'] )Ċ +ĠA T +n n +. Int +SS ERT +Ġeffect ive +LE TE +Ġto ols +AR D +Ġdig ital +D ouble +ĠF ind +R C +Ġin line +/ r +AR AM +AS K +Ġint ent +a ight +_add r +Ġrequest s +.f irst +Ġde bug +Ġsp ent +() ));Ċ +Å Ľ +Ġpr incip +Log ger +clud es +. use +Ġsur v +med ia +ĠFe bruary +ĠM ac +Ġmiss ing +Ġw ife +Ġtalk ing +ĠM ake +Ġc art +Ġloc ated +E nc +- a +ch ron +Ġc ards +Ġgu y +Ġp ers +ĠY es +ate ver +ĠA ng +ol ar +ĠE ven +Ġacc ur +ĠP ower +ĠG old +c lear +Pro cess +Ġrec ords +Ġk illed +.c lear +ĠWARRANT IES +Ġpur pose +pan el +J ECT +ÃŃ a +Ġex erc +W S +/ L +. exports +Ġ__ _ +Ġs in +S ervlet +Ġd é +.de lete +ro ke +S l +ug h +ear s +Ġpoint er +Ġh op +all ery +Ġo bs +co very +ĉ char +ĉĉĉĉ ĉĉĉĉĉĉ +ĉ def +oc ity +itch en +ul ations +ĠF IT +Ġ ). +straint s +vent ion +Ġrequ ires +ĠO per +M E +OUN T +al let +Ġn orm +I RE +ex as +Ġprogram s +Ġwe ak +' .$ +u ing +ĉ ĠĠĠĠĠĠĠ +Ġm il +Ġf irm +init ely +_VAL UE +ap se +atis f +Ġdem and +_m od +Ġdescri bed +Ġpl aces +V ID +Ġal one +Ġex port +Ġv ec +ĠM ax +Ġactiv ities +ict ures +g ener +Ġm a +Ĥ ¬ +Ġexpress ion +C allback +_ content +ĠM ost +Ġtest ing +E C +CH ANT +Ġad just +.Th reading +( ctx +Ġag ree +ig hest +Ġu i +ĠL aw +. Y +> ĊĊ +.ex ample +ber g +Ġmov ed +ĉ e +ĠS aturday +Ġpay load +Ä ĩ +) :ĊĊ +Ġbe y +ur er +< script +Ġs ymbol +Ġass um +Ġp ul +E ffect +Ġh undred +To ol +ak ed +con nection +Ġvo ice +Ġp d +Ġtrans action +Ġlink s +E rr +ĠInd ian +T C +atal og +n i +s ign +<< " +j i +y a +Ġdemon str +ul ated +. St +Ġinst it +Ġbo ost +Ġcell s +ol ic +.P ro +: , +"> \ +Ġth us +ĠReg ister +h ol +ĠCh inese +Ġpost ed +Ġm agn +ab ilities +Ġdise ase +Ġrem ains +ĠPro f +- form +Ġc in +org an +ic ate +Ġst ress +] * +Ġ ---------------------------------------------------------------- +_ context +or ry +Ġd ied +m at +Ġstart s +.M essage +Ġrun s +Ġgu ide +Ġwarrant y +ential s +d ict +ĠS ize +ul er +Ġrespons ible +_SE T +Ġcont aining +ĠPr ice +| | +F S +Ġem p +_b utton +( uint +Ġsu ff +p th +Ġdef initely +put e +Ġmarket ing +ĠW H +ĠS ie ++ = +OL OR +Ġcons ult +Ġs igned +Ġse quence +le e +Ġrequire ments +h y +Ex press +M T +se y +Ġ ult +å ® +ellig ence +Ġanal y +Ġd ress +eng ine +ĠG reat +ĠAnd roid +ĠA lex +m ode +D ictionary +.D ate +ä ½ +V ICE +Ġfam ilies +ĠRuss ian +ĠT imes +.c all +$ ( +Pro file +Ġf older +ch es +Ġleg is +_ row +un es +Ù Ħ +Ġ} ). +Ass ert +ag en +ĠH and +I ter +Ġbig gest +ore ach +Ġpol ic +Ġper missions +Ġshow ed +ĠE lement +Ġtop ic +âĢĶ âĢĶ +ro ad +ĠB ank +rec ord +Ġpart ners +ĠR ef +ess ions +Ġass ess +U ST +ĠPart y +pro du +L C +Ġ ul +. form +h ide +c opy +UT F +ĠSO FTWARE +čĊčĊ čĊ +ĠL in +un a +ug ar +Ġadmin istration +Ġopen ing +Ġsc an +Ġcontin ued +com ponent +.s p +Ġhapp ens +um my +ĠP R +.F ile +ĠDown load +Lo ading +d i +Ġwait ing +_A DD +T ab +.query Selector +Ġecon omy +ĠF rench +t xt +Ġf ant +_ ;Ċ +H older +S H +Ġn umpy +Ġst reet +Ġm ale +\ Model +ang ing +ĠB ill +Ġprevious ly +B I +ĠSec ret +Ġm ist +ĠF ield +up s +ĠPro cess +Ġke pt +ĠO T +Ġtrad itional +. i +am in +Ġhelp s +An y +orig in +ilt ers +j u +d esc +ĠA ccount +Ġ) čĊ +k top +ol ly +Ġf s +Ġ ê +Ġ ut +Ġcent ral +(t est +.A n +Ġs atisf +G R +ĠF ull +Ġhe at +ib er +Ġon to +m os +S chema +Ġfact ory +" .$ +aw s +St atement +(t arget +ĉ new +.b e +Ġg uest +Ġm al +AR Y +Ġre ached +Ġm ouse +Ġchall enge +ĉd ouble +ĠT em +Ġt error +Ġex tract +_T O +Ġsepar ate +Ġm ir +h elp +Ġcap acity +ĠProp erty +k an +_c reate +ĠL ight +.p arent +Ġunderstand ing +Ġeas ier +Ġ| = +Ġen h +Ġf at +Ġprot est +am m +_ AT +- of +il s +ĠO h +Ġps ych +Ġ$ . +ind s +Ġrel ative +sh op +sh ort +ĠS and +uest ion +Ġf ear +/ ĊĊ +. context +Ġschool s +Ġser ve +z one +_d b +Ġmajor ity +ex ample +Ġl ang +ĉ ĠĠ +Reg ister +end o +Ġprocess ing +_t emplate +- user +Ġe g +C OM +ĠBl ue +i ro +Ġrem ote +ĠI T +#! / +Ġred istrib +ra z +ĠS ince +ĠT ur +Back ground +== = +Ġref lect +Ġpro s +c md +Ġwh om +Com pat +ĠA re +Id entifier +ĠTh om +_ port +g u +Ġmon itor +r m +Ġpat ient +ver ter +Ġg ain +- ui +In st +Ġd ies +A rea +_f ilter +Ġgr at +Ġreal ity +ord inate +ol ved +Cont act +Ġcompl iance +_ or +ĠV ar +d l +Ġapp end +G ER +(m ax +.re nder +Ġd ynamic +ordin ates +_ options +_c olumn +Ġb atter +s pace +L a +ĠS ource +/b in +Ġd os +ĠBo ard +ĠTh read +ĠA L +( config +ĠM er +Ġm iles +_ header +ETH OD +iz z +Ġbenef it +Ġinteg r +(c urrent +ul o +. default +ĠD iv +Ġt on +o th +erv ation +ed om +Ġb aby +ce ived +.t op +rior ity +ĠL ocal +ri age +Ġattack s +Ġh ospital +Ġfem ale +ĠLog in +ĠFl or +Ġch ain +ash ion +Text ure +S ave +Ġf arm +.cont ains +.T est +Ġknow s +Ġgener ally +ip eline +Ġme ant +enc ia +Ġn icht +Ġcont ents +P M +ched ule +( line +C G +j ob +ĠRe al +u er +f irm +Ġ Ø +et ro +" `Ċ +Ġspe ech +Ġth r +fore ach +Ġw arn +ĉ l +Ġhe avy +< li +N e +Ġinvestig ation +M ath +- title +Ġch urch +Ġdes pite +ch ain +Ġwh atever +ar ian +f n +Ġm eta +} )ĊĊ +U FF +Ġregard ing +_S UCCESS +m es +ĠInt ent +Ġres olve +pos s +ir a +for ce +o ice +à ¢ +Ġp m +Ġup dates +A rr +Ġ Ñ +test ing +Ġto ward +nt ax +ë ĭ +Ġlist en +Ġgo als +Instance State +D r +Ġr are +Ġtr ail +Ke ys +C al +C ar +ĠPe ople +ĉ local +class es +Re ference +.for Each +em b +act iv +Ġpr im +red ict +Ġr ad +æķ ° +.B ack +Ġsp read +Ġc lock +Ġv ir +ed itor +Ġeffort s +Ġbr anch +Ġind ust +Ġmot or +Ġam b +Ġdat etime +Ġren cont +ĠChrist ian +ĠAmeric ans +f ull +Ġf mt +.m ain +Ġca used +_ update +ĠCont ent +AT CH +Ġb ath +ĠE ach +Ġr adio +ach ment +uz z +Sub mit +Ġre strict +ab in +ĠL oad +Ġext ension +Ġess ay +Ġh at +avi our +to Be +": [ +Ġoffer ed +Ġv ill +(d ouble +æĹ ¥ +b c +_f ree +ĠM iss +ĠB er +Ġ è +ĠL ike +Ġhelp ed +.get Name +_ AL +Ġsp irit +ĠAp ache +w s +Ġthere fore +( params +_ img +Ġpe ace +Ġinc or +ĠEX PECT +Ġmin or +ip es +ĉ data +select or +c ity +tr ie +.b ase +_f rame +Ġopen ed +/ json +L Y +n u +.D e +t f +m argin +.P arse +Ġp i +Ġe q +b d +Field s +ĠT ree +Ġb an +ist an +Ċ ĠĠĠĠĠĠĠĠĊ +ĉg l +Ġprodu ced +s ystem +M ark +_h ash +Ġb g +Ġconst it +ĠLe ague +Ġmiss ion +_ format +([ Ċ +clus ion +! " +Ð · +b reak +ĉs witch +Ġth er +Trans form +Ġfoot ball +- link +r oute +. auth +Ġb ag +ov ers +Ġen abled +Ġr ac +( I +C R +anc ing +Ġman aged +_ q +NG TH +Ġm ac +ĠA uto +ament e +Ġ' ', +.App end +Ġp in +. item +ack ing +Ġocc as +p erson +Ġt i +.Re g +Ġh aven +Ġg lass +Ġ" ) +_ char +res ource +Ġep isode +Ġ' _ +ĠE s +ĠEar th +Âł Âł +UP DATE +ĠS ou +u is +t ypes +Ġm as +Ġf av +Ġcon struct +_r ate +er as +Ġ| Ċ +rop erties +Ġext ernal +Ġap plied +Ġpre fix +ot ed +l ers +Ġc old +ĠS P +ĠCh urch +ĠOut put +los ed +ç ļ +ific ate +oper ation +her it +x FF +. env +_ err +os h +D irection +C ancel +ĠFr ank +Ġfind ing +. )ĊĊ +Ġr outer +ãĥ » +s es +Ġc row +== ' +Ġs and +Ġr id +it ure +Ġent re +Ġo bserv +Ġv ac +ð Ł +- T +A rt +n ight +. search +Ġex change +Ġdistr ict +. os +Ġdep artment +Ġdoc uments +Ġcent ury +ĠN ext +H ost +ĠK IND +Ġsus p +- P +re nd +. em +u ite +ist ers +( json +ĠAn n +w t +at i +ĠHT ML +wh en +D irectory +Ġsh ut +< a +ed y +Ġhealth y +Ġtemper ature +ĠG en +Ġmet al +Ġsub mit +ĠD O +Ġat tract +Ġ{ };Ċ +ĠW ord +Ġl l +Ġseem ed +k o +I ED +Ġl abor +.Cont ext +Ġas set +y ou +Ġc ars +ĠC olumn +Ġr é +Ġs quare +ĠNS String +âĢĿ , +ap es +.. .Ċ +Ġthan ks +( props +Ġt ick +Ġexper iment +Ġpr ison +t ree +- text +ĠIO Exception +-w idth +_ST ATUS +f ast +-b ody +- header +Ġgu ar +cre te +ĠT im +Ġclear ly +ĠRepublic an +Ġjust ify +и ÑĤ +ĉ ĠĠĠĠ +c ache +; // +Ġpres ence +Ġfact ors +Ġemploy ee +] )) +M ember +Ġselect or +b or +ĠM ex +çļ Ħ +ut ex +_t ag +ail ure +ĠN et +Ġre li +E G +Ġf printf +Ġte en +lo ss +Ġle aving +De legate +Ġbe at +Ġmin ute +sub scribe +Ġredistrib ute +Con stants +Ġcan cer +/ { +B L +Ġs pan +ĠCh ild +C enter +Ġear th +Y S +ĠLe vel +Ġse a +.s upport +.in ner +. Item +ill ing +ĠĠĠĠĊ ĠĠĠĠĊ +ĠL abel +ĠE st +( arg +bo Box +ĉf oreach +c os +F ailed +sw ers +Ed itor +r ont +ĠM P +ex pr +ĠL ife +Ġ? ? +ö r +Ġatt end +ĠQ ue +Ġspec ies +- D +Ġa us +Str uct +Ġadvant age +ost on +-b lock +in itial +C RE +Ġtr uly +Ġcomp are +or ney +Ġs pect +F ull +b es +Ġvis ible +Ġm ess +st ances +Ġcl oud +_v ersion +Ġf urn +ic ago +LO W +Ġtraff ic +Ġf ol +rypt o +Ġdecl ar +Ġsl ot +ĠEx t +ĠEng land +ĠU nder +Ġt a +let ter +Ġoffic er +ĠDon ald +Y es +_ json +IT ableView +ĠU SE +mploy ee +Ġopin ion +ĠA ut +b order +Ġad vice +Ġautom atically +is co +Ġm m +. vis +am l +Ġinitial ize +Ġ( { +Ġ ;ĊĊ +Ġgener ation +Ġb its +clip se +Ġun f +ut ors +pl t +Ġdel ta +est roy +is is +< br +Ġlimit ations +Ġend ed +ĠM ad +il m +Th ese +ĠMin ister +Ġch art +F ragment +Ġindepend ent +Y ear +Ġin str +Ġt ags +A VE +ĠAr ch +st op +Pro gress +Ġm i +Ġlearn ed +G e +Ġhot el +S M +T YPE +Ġc y +ERS ION +un ately +l imit +s el +Ġmov ies +Ġste el +o z +g b +ĠC amp +s ite +ĠLog ger +P LE +оР´ +. right +ĠC ore +Ġm ixed +st ep +Ġput s +s uper +R outer +. Http +ly ph +ĠColor s +Ġandroid x +. str +Ġinn ov +Ġde ck +' >Ċ +ap ers +] ( +cont inue +s pec +ĠR oad +AS H +ili ar +Ġcontin ues +Ġapp oint +Ġ# Ċ +ĠV ir +Ġ?> " +Ġb in +} ", +go ing +e ach +B D +ĠA ccess +D oc +ĠMan agement +B ER +ask et +.get Instance +Ġestablish ed +so cket +IN S +ĉv irtual +ĉ result +RE AD +_ height +ĠF ont +Ġ( );Ċ +_ html +Ġneighb or +l or +Ġg ather +Ġ} )ĊĊ +Ġid entity +Ġf ab +p adding +ĠR oute +Enumer able +à ´ +Ġfor ced +/j query +.ĊĊ ĊĊĊĊ +res ents +_ left +.P aram +ĉ throw +ĠH am +Ġevent ually +ac er +p ub +Ġtr a +un ique +d el +ĠFlor ida +ĠC lean +x a +Ġ · +Ġvalid ate +Vis ual +Ex pression +_f unc +m ember +ĉ h +tr l +ĉ G +nap shot +ĠProp Types +v in +] )ĊĊ +ow l +if ies +Ġ$ ('. +ĠCont ext +ĠTo ast +. Key +Ġoffic ers +/ n +s n +und efined +. items +ut ow +am age +Ġaccount s +ook ie +Se ction +ici ans +Ġad vis +( is +[: , +ĠFr ance +F unc +ic ious +Ġto k +Ch annel +ĠA D +_N UM +Ġtime out +lem ma +rem e +u j +.A l +uc lear +( os +(" < +[ Ċ +f etch +Ġb al +Ġgu id +- align +ĠW rite +ĠOn ce +utow ired +OD ULE +Ġp itch +C F +by tes +ĠCom mission +Ġincre d +P ER +_ response +ĠL os +par ser +Ġass ume +. Request +ĠT oken +_p osition +Ġn om +- term +Ġrem aining +i ostream +Ġpie ces +ap y +ĠL ess +r ange +umb n +pr ise +_ option +Im pl +k wargs +Ġbusiness es +Al ert +Ġpart ies +ĠCont ainer +ĠPr ivate +ĠPl an +Ġregister ed +Ġj our +ack er +ен и +/ > +ch at +se ct +Ġcre ation +olut ely +Ġinst ant +Ġdel ivery +ick en +y es +ĠFr anc +bl ing +end a +[ ( +_r ange +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠ +Ġsched ule +Con n +Ġthan k +x d +Ġh ook +Ġdocument ation +Param eters +H ello +v t +Ġart icles +Ġw est +def ined +. select +ok ens +ĠV AL +.f ile +res et +Ġmy s +ĠM A +] ), +Ġc ities +rel ated +å Ľ +Ġappe ared +Ġw id +.p anel +ĠIn s +. entity +Ġde cre +ĠL ou +(t ime +ĠTh ank +.create Element +Ġmention ed +oun ce +ĠT ry +ĠW all +/ images +ĠM enu +' čĊ +ĠE r +Ġcrit ic +ĠY ear +( param +Ġf lo +N N +oot er +Ġ ];Ċ +ĠA ff +" github +room s +Ġh yp +g lobal +Ġa vec +æľ Ī +Ġcomplet ion +Ġcon d +onym ous +( temp +Ġst ars +Ġre levant +Ġcover ed +Ġel im +_t ypes +( bool +Ġt u +_ex ists +Ġsec ure +Ġst ored +] / +x F +ĠCont roller +Ġm igr +M I +ĠD en +Ġann ual +U IL +- and +Ġcr ime +b el +Ġk itchen +@ g +_p h +ourn ament +ĠS ocial +ĠS pecial +log ger +Ġt ail +Ġun known +d ed +Ġapp rec +(d b +c f +Ġass ign +- out +ĠM ont +d p +w idget +Ġst one +- primary +. grid +Result s +az z +Ġda ughter +Ġcur r +Ġl in +Ġs outh +form s +ĠO UT +let te +ak s +ig ure +ĠE U +var iable +Ġb rief +ĠSc ott +Ġcon ference +and a +_ lock +or al +Ġe ine +OR S +//////////////////////////////// //////////////////////////////// +ess o +Ġr is +Ġg ender +est ic +L icense +( out +Ġm s +Se e +Ġwill ing +az e +Ġs ports +Ġy es +l u +Ġp urs +/j avascript +- pro +nav bar +_pro duct +/ bootstrap +Ġdr iving +Ġ Ä +Ġpro pos +ult ip +up lic +. email +Ġappro x +( cl +Ġwe ar +Ġrep ly +ass et +Ġ ice +Ġt x +k r +ĠGerman y +ĠGe orge +Ġc b +ĉ err +M ove +Ġpol y +vo ice +} " +Ġan imal +A v +ĠL ocation +Ġn ative +] [" +< double +Ġm ais +, int +Ġpre par +Ġinter val +plement ation +_ ERR +Ġb ug +> " +st at +Ġ} ,čĊ +< span +Ġfa ith +Ġ rom +pre v +ĠE lect +F ind +Ġg od +ot or +// ---------------------------------------------------------------- +orig inal +C pp +ĠSen ate +Ġposition s +Ġweap ons +Ġco ff +Ġpur poses +p ol +Ġim press +Ġanim als +. Entity +(n p +Ġmur der +Ġ` ` +fl ag +Ġsol utions +ĠAct ive +Ġb right +.d ate +Ġsit u +ï¼ Ī +. ID +Ġs ie +), čĊ +ak t +S pace +.d at +.index Of +h an +az ine +ĠZ e +Ġcr ash +( / +> = +Ð ± +iv a +.Auto Size +ĠL at +_ ext +Initial ize +.reg ister +OP Y +Ġre verse +_d is +'] [ +Ġprom pt +ont o +ĠJ ournal +r outer +Ġmys qli +# else +) " +-x s +let s +ph an +. LE +W ill +Ġaff ord +Ġsk ill +-t oggle +N C +B ind +T S +J ust +iter al +Y P +ĉ unsigned +Ġw ind +)) :Ċ +Ġw arning +ĠW ater +Ġd raft +Ġc m +Ġs am +Ġhold ing +z ip +ĠSc ience +Ġsup posed +G en +Ġdi et +< h +ĠP ass +v i +Ġhus band +� � +n ote +ĠAb out +ĠIn stitute +Ġcl imate +.Form at +Ġn ut +est ed +Ġapp arent +Ġhold s +f i +new s +C M +v ideo +': ' +D ITION +p ing +Ġsen ior +w a +-- >Ċ +_ default +ĠD atabase +re p +E SS +ner gy +.F ind +_m ask +Ġr ise +Ġk ernel +:: $ +. Q +Ġoffer ing +de cl +ĠC S +Ġlist ed +Ġmost ly +eng er +Ġblock s +ol o +Ġgover ning +\ F +Ġcon cent +.get Text +Ġm b +Ġocc urred +Ġchang ing +Sc ene +_C ODE +B eh +" The +Ġt ile +ĠAssoci ation +ĉ P +al ty +_ ad +od ies +i ated +Ġpre pared +poss ible +Ġm ort +TE ST +Ġign ore +Ġcal c +Ġr s +Ġassert Equals +Ġs z +ĠTH IS +. "Ċ +Ġcan vas +j ava +Ġd ut +VAL ID +.s ql +. input +Ġa ux +S up +Ġart ist +V ec +_T IME +.string ify +et ween +ĠC ategory +Ġ[ - +ĠDev Express +ĠJ ul +Ġr ing +. ed +Y Y +L et +Text Field +Ġfl at +_p rint +ĠOT HER +ad ian +Ġcheck ed +e le +Al ign +stand ing +Ġ[ ], +Ġl ab +uck y +ĠChrist mas +( image +.m odule +Ġl ots +Ġslight ly +(f inal +er ge +è ¿ +ĠPol ice +ĠR ight +Ġaw ard +ĠO S +Ġ{ }ĊĊ +Ġp tr +ov es +ic ated +еР¼ +Ġman age +olid ay +Am ount +ool Strip +t body +N av +w rap +B B +Ġwatch ing +ari os +Ġoption al +_ K +ĠL icensed +.M ap +T imer +ĠA P +ĠRe v +( o +, c +um in +eta iled +ĠH y +Ġbl ank +ag ger +ĠS elf +() [ +.m ake +ear n +ch annel +< pre +ble m +_p assword +_s p +ic ing +e z +Ġthe ory +ĠT er +, n +log o +ĠHT TP +() )) +.h andle +> ;Ċ +W orld +Ġpy thon +Ġl if +Ġtr av +Ġcon ven +com pany +ĠCl ub +V er +B tn +Ġz one +product s +ĠE duc +Ġver ify +ĠM il +on o +] );ĊĊ +EN CE +Ġpack et +Ġc er +Ġen umer +Ġpar s +form ed +Ġocc up +t re +Ġexerc ise +D ay +_s um +Ġask ing +apt ion +Ġord ers +Ġsp ending +ĠE RR +.D is +ĠU til +âĢľ I +\ ' +? ) +/ >Ċ +Ġem ot +Ġinflu ence +ĠAfr ica +att ers +Ù ħ +.s ession +Ġch ief +ĉĉĉĉĉĉĉĉ ĉĉĉ +Ġto m +clud ed +ser ial +_h andler +.T ype +ap ed +Ġpolic ies +- ex +- tr +bl ank +mer ce +Ġcover age +Ġr c +_m atrix +_ box +Ġcharg es +ĠB oston +P e +Ġcirc um +Ġfil led +Ġn orth +icture Box +ĉ res +è ® +Ġter min +Ġ[ â̦ +IRE CT +Ġb er +Ġ" ../../ +ret ch +.c ode +_c ol +ĠGovern ment +Ġarg v +ĠL ord +as i +Ex ec +ĉ let +vert is +Ġdiscuss ion +en ance +out ube +type of +Ġs erved +ĠP ut +ĉ x +Ġs weet +B efore +ateg y +. of +ĠM aterial +S ort +ON T +ig ital +Wh y +Ġs ust +Ġ ç +ab et +Ġseg ment +Ġ[ ],Ċ +ĠMus lim +Ġfind ViewById +c ut +_T EXT +ĠM ary +Ġlo ved +Ġl ie +ĠJ O +Ġis set +mon th +Ġpr ime +t i +ĠCar ol +U se +ĠP op +ĠS ave +Int erval +ex ecute +d y +ĠI ran +_ cont +ĉ T +Ġph ase +check box +we ek +Ġh ide +Ġt il +Ġj u +C ustom +b urg +/ M +T ON +Ġqu ant +Ġr ub +ix els +Ġinst alled +Ġd ump +Ġproper ly +( List +Ġdec ide +app ly +H as +Ġkeep ing +Ġcitiz ens +Ġj oint +p ool +S ocket +_ op +Ġweap on +gn ore +ĠEx ec +ott en +ĠM S +Ġ( - +ĠRe view +Ġex amples +Ġt ight +! ( +D P +ĠMessage Box +Ġphot ograph +UR I +é t +l ow +ĠGr and +.p ersistence +Ġmaint ain +Ġnum s +Ġz ip +ial s +ĠG ets +pe g +ĠB uffer +~~ ~~ +ra structure +ĠP L +u en +ob by +size of +Ġp ic +Ġse ed +Ġexperi enced +Ġo dd +Ġk ick +Ġproced ure +avig ator +- on +, j +ĠAl though +Ġuser Id +ac cept +Bl ue +IC olor +l ayer +av ailable +Ġend s +.t able +Ġdat aset +b us +Ġexpl ain +( pro +ĠCommit tee +Ġnot ed +] :Ċ +D im +std io +. ",Ċ +_s ource +ĠWe ek +ĠEd ge +Ġoper ating +Ġest e +i pl +ag ination +Ġpro ceed +Ġanim ation +.Model s +ĠW atch +i at +Ġopp on +/ A +Re port +Ġs ounds +_b uf +IEL D +Ġbu nd +ĉ get +.p r +(t mp +Ġk id +>ĊĊ Ċ +Ġy ang +Not Found +Ñ Ĩ +m ath +@g mail +ĠL IMIT +red ients +Ġv ent +avig ate +L ook +Ġrelig ious +Ġr and +ri o +( GL +_ ip +u an +ici ency +ĠCh ange +> čĊčĊ +ĠEnt ity +Ġrencont re +ĠR et +pl an +é n +BO OL +ur ies +tr ain +Def inition +======== ==== +z z +An imation +ĠO K +_m enu +.b l +_s core +Ġac ad +( System +Ġref resh +'=> $ +.G raphics +ament o +p id +t c +Ġt ips +Ġhom es +Ġf uel +â ĸ +_h elper +ĠĠ čĊ +ĠR oom +.C lose +_ attr +ĠM ount +ĠE v +ar ser +_t op +e ah +ĠDe lete +ãĢ į +u ke +Ġus age +ar ia +_de v +Ġtext ure +Ġconvers ation +e per +Be an +d one +non atomic +ĠSe cond +Ġshoot ing +_p re +Com ponents +Ġ] ĊĊ +__ , +stit ution +.Ch ar +> ();ĊĊ +Ġpresent ed +Ġw a +ok er +- ĊĊ +in er +Ġbe coming +Ġinc ident +At t +Ġreve aled +for c +Ġbo ot +.p age +Enumer ator +_ -> +Ph oto +Ġs pring +. ", +ĠD ictionary +B JECT +Ġloc ations +Ġs amples +Input Stream +ĠB rown +Ġst ats +qual ity +Ñ ħ +-d is +Ġhelp ing +Ġp ed +( se +ĠWh o +al ian +int ernal +Ġf t +> (). +-> { +Ġm ine +Ġs ector +Ġg ro +Ġopport unities +Ġà ¼ +Ġm p +Ġalleg ed +Ġdoub t +M ouse +Ab out +_p art +Ġch air +Ġstop ped +lo op +ent ities +Ġapp s +ans ion +Ġm ental +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠ +F R +Ġdef end +c are +Ġide al +/ api +ur face +Ġe le +ul ator +ĠR ights +angu ages +Ġfund s +Ġad apt +At tributes +Ġdep loy +opt s +Ġvalid ation +Ġconcern s +u ce +.n um +ult ure +il a +Ġc up +Ġp ure +.F ore +ĠHash Map +.value Of +as m +M O +Ġc s +Ġst ores +Ġ ************************************************************************ +Ġcommunic ation +m em +.Event Handler +. Status +_ right +.set On +S heet +Ġident ify +ener ated +order ed +Ġ" [ +Ġs we +Con dition +ĠA ccording +Ġpre pare +Ġro b +P ool +Ġs port +r v +ĠR outer +Ġaltern ative +( [] +ĠCh icago +ip her +is che +ĠDirect or +k l +ĠW il +key s +Ġmy sql +Ġw elcome +k ing +ĠMan ager +Ġca ught +) }Ċ +S core +_P R +Ġsur vey +h ab +He aders +AD ER +Ġdec or +Ġturn s +Ġr adius +err upt +C or +Ġm el +Ġin tr +( q +ĠA C +am os +M AX +ĠG rid +ĠJes us +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠ +.D E +Ġt s +Ġlink ed +f ree +ĠQ t +Ġ/** čĊ +Ġf aster +ct r +_ J +D T +.C heck +Ġcomb ination +Ġint ended +- the +- type +ect ors +am i +ut ing +Ġum a +X ML +U CT +A p +ĠR andom +Ġr an +.s ort +Ġsort ed +. Un +_P ER +it ory +Ġprior ity +ĠG al +ĠO ld +h ot +ĠD isplay +(s ub +_T H +_ Y +ĠC are +load ing +K ind +_h andle +, , +r ase +_re place +.add EventListener +ĠR T +Ġenter ed +g ers +Ġ ich +( start +/ app +Ġbro ther +M emory +Out let +Ġ utf +pre c +Ġn avigation +OR K +Ġd st +D etail +Ġaud ience +Ġd ur +Ġcl uster +un ched +Ġ ], +Ġcomfort able +. values +ĠT otal +Ġsn ap +Ġstand ards +Ġperform ed +h and +(" @ +å Ń +Ġph il +ib r +tr im +Ġfor get +Ġdo ctor +.Text Box +icon s +, s +ĠO p +S m +St op +ĉ List +ĉ u +Com ment +_V ERSION +.X tra +P erson +r b +LO B +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĊ +ĠCent ral +IC K +ra q +Ġput ting +Ġm d +ĠL ove +Pro gram +B order +o or +Ġallow ing +a fter +Ġent ries +ĠMay be +] ). +ĠSh ort +) \ +.n ow +f riend +Ġpre fer +ĠG PIO +os is +ĠGame Object +Ġsk ip +Ġcompet ition +_m atch +lic ations +_CON T +.group Box +Ġal s +" We +_e q +l an +_ search +ĠMus ic +as is +Ġb ind +ĠIs land +r um +( E +Ġse at +V ideo +Ġa ck +ree k +={ () +Ġr ating +Ġrestaur ant +DE X +(b uf +pp ing +ual ity +Ġle ague +Ġfoc used +ap on +$ data +CL UD +CLUD ING +Ġabs olute +( query +Ġtell s +A ng +Ġcomm unities +Ġhon est +ok ing +Ġap art +ar ity +/ $ +_m odule +ĠE nc +. an +.Con fig +C re +Ġsh ock +ĠAr ab +I ENT +/ re +Ġre trie +ycl er +is a +ĠO rgan +. graph +Ġ í +ĠB AS +En um +Ġposs ibly +ÑĢ Ð°Ð +ĠJapan ese +Ġc raft +ĠPl ace +Ġtal ent +Ġfund ing +Ġconf irmed +Ġc ycle +/ x +G E +Ġhe aring +Ġpl ants +Ġm outh +p ages +or ia +ĠRem ove +_t otal +Ġo d +oll apse +do or +Ġb ought +Ġadd r +AR CH +_d im +dd en +Ġdec ades +RE QUEST +Ġvers ions +f ire +Ġmov es +f b +Ġcoff ee +.con nect +ĠR ow +Ġs chema +S cope +- Type +Ġfight ing +Ġret ail +Ġmod ified +T F +File s +n ie +_com mand +st one +Ġ ÑĤ +_ thread +Ġb ond +ĠDevelop ment +Ġp t +F ORM +ple t +Ġident ified +c pp +Ġc oding +ok ed +ĠM aster +ID TH +Ġres idents +red it +ĠPh oto += - +un te +ate ur +_ST ATE +ĠS ing +Ġshe et +. val +or se +Ġh ers +Ġdetermin ed +Com mon +Ġw ed +_ queue +P H +ĠAt l +cre d +/L ICENSE +Ġm es +Ġadv anced +.j ava +.S h +G o +k ill +f p +_set tings +Ġp al +Ġtr uck +Ġcomb ined +Ġ" ${ +ĠCor por +Ġjo ined +ĠJ ose +ĠC up +un s +est ival +lev ision +Ġbro ken +Ġmar riage +ĠWest ern +Ġrep resents +ĠT itle +Ġs s +.A ss +ongo ose +ient o +< >();Ċ +Ġabs olutely +Ġsm ooth +TER N +ĠUn less +W ord +Ġmer ge +ig an +ĠV ol +Ġn n +.get Id +ĠÐ · +Ġsex y +Ġseek ing +S ingle +. this +Ġk om +b ound +; " +Ġfont Size +_d f +Ġinj ury +( H +Ġiss ued +_ END +: self +Ġp atch +Ġle aves +Ġad opt +File Name +ãĢ IJ +Ġexec utive +ĠBy te +] ))Ċ +Ġn u +out ing +clud ing +- R +. options +Ġsub stant +av ax +ĠB UT +Ġtechn ical +Ġtw ice +Ġm ás +Ġun ivers +y r +Ġdr ag +ĠD C +Ġs ed +Ġb ot +ĠP al +ĠH all +forc ement +Ġa uch +.m od +not ation +_file s +.l ine +_fl ag +[ name +Ġres olution +Ġb ott +(" [ +end e +( arr +F ree +( @" +ĠD istrict +PE C +: - +P icker +ĠJ o +ĠĠĠĠĠ Ċ +ĠR iver +_ rows +Ġhelp ful +Ġmass ive +--- Ċ +Ġmeas ures +ĠR untime +Ġwor ry +ĠS pec +ĉ D +ãĢ ij +Ġ) {Ċ +Ġwor se +(f ilename +Ġl ay +Ġmag ic +ĠThe ir +ou l +st roy +ĠWh ere +Ġsu dden +Ġdef e +Ġb inding +Ġfl ight +ĠOn Init +ĠW omen +ĠPol icy +Ġdrug s +ish ing +(' ../ +ĠM el +pe at +t or +Ġpro posed +Ġst ated +_RE S +Ġe ast +ĠCON DITION +_d esc +Ġwin ning +fol io +M apper +ĠP an +ĠAn ge +.s ervlet +Ġcop ies +L M +Ġv m +å į +Ġd ictionary +S eg +el ines +ĠS end +Ġ iron +ĠF ort +.d omain +Ġdeb ate +Not Null +e q +ach er +l f +ĉf mt +Ġlaw y +Ä Ł +ĠM en +Ġtr im +( NULL +Ġ! ! +Ġp ad +Ġfollow s +"] [" +re qu +ĠE p +.g ithub +( img +et o +(' \ +S ervices +umbn ail +_m ain +ple ted +fort unately +Ġw indows +Ġpl ane +ĠCon nection +. local +u ard +} \ +== " +and on +ĠR oy +w est +ig inal +em ies +it z +') :Ċ +ĠP eter +Ġt ough +Ġredu ced +Ġcalcul ate +Ġrap id +c ustomer +Ġeff icient +Ġmed ium +Ġf ell +. ref +ĠC as +Ġfeed back +S peed +( output +aj e +Ġc ategories +Ġfe e +} ; +Ġde leted +re h +Ġpro of +D esc +B uild +Ġs ides +.Array List +- % +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠ +Ø ± +.m atch +л и +Ġfe els +Ġachie ve +Ġcl im +_ ON +ĠC D +Ġteach er +_c urrent +b n +_P L +ist ing +En able +G EN +Ġt v +Ġso ck +Ġpl ays +Ġdis count +ĠK E +ĠDe bug +F ore +ĠI raq +Ġappear ance +M on +Ġst yled +ĠH uman +i ot +ĠH istory +Ġs ac +ĠC ollection +Ġrecomm ended +.Se lected +Ġorgan izations +Ġdiscover ed +co hol +ad as +ĠThom as +M ay +Ġcons erv +Ġdom in +ĠF ollow +ĠSe ction +ĠTh anks +User name +Ġrec ipe +Ġwonder ful +.s leep +_ if +ĉĊ ĉĊ +orn o +Ġr u +_t arget +." " +à ¦ +Event Args +Ġinput s +Ġf if +Ġv ision +c y +ĠS eries +) ((( +Ġtr ading +Ġmark er +B egin +Ġtyp ically +Ġca uses +drop down +_DE BUG +Ġdet ect +c ountry +! ");Ċ +ĉ R +app y +Ġc ref +(' < +" => +ĠL E +read er +Ġadmin istr +à µ +uck et +Ġf ashion +. char +iz ar +Ġdis able +Ġsu c +ĠL ive +iss ue +Ġmet adata +fl ags +Ġ ðŁ +Ġcomm itted +Ġv a +Ġr ough +Ġ'' 'Ċ +Ġhigh light +_var s +V O +Ġenc oding +- Z +_s ign +$ ("# +Ġr ain +reate st +ĠEN D +Se lection +Ġcandid ates +Ġs av +. Empty +Ġdec isions +Ġcoll abor +rid ge +fe ed +ress ion +Ġperson s +V M +eg a +_B IT +A ccording +ack ed +Ġdoll ars +_lo ss +ĠC ost +} "Ċ +Not ification +Ġpro stit +Ġauthor ity +.re c +Ġsp okes +ĠT oday +ist ant +ĠHe ad +âĢĿ . +ertain ment +ce an +cul ate +Ġv en +How ever +_ arr +Ġtok ens +G raph +ĠJ ud +ĠVir gin +ĠS erial +un ning +M utable +ag ers +.c sv +Ġdevelop ing +Ġinstruction s +Ġprom ise +Ġrequest ed +_ encode +/ " +ĠI con +u ilt +- day +Ġint elligence +. IS +ĠO bservable +ĠH ard +Bo ol +ident ial +.An chor +Ġsell ing +C I +AG ES +t le +b ur +UFF ER +R Y +Ġbig ger +Ġr at +Ġfam ous +Ġtyp ename +Ġexpl ained +} }Ċ +Ġn uclear +- N +Ġcr isis +ĠEnt er +Ġan swers +/ ${ +/ pl +Ġse qu +_n ext +m ask +Ġstand ing +Ġpl enty +ĠC ross +ĉ ret +d ro +ĠC ast += true +ĠCh ris +ic io +ĠM ike +Dec imal +add Component +L en +Ġco ck +Ġ# { +UR N +< tr +Ġauthor ities +Res ources +- H +B ottom +_ qu +put er +ester day +Dis patch +s ince +Ġfam iliar +, i +V C +Ġm ent +, C +Ġfre edom +Ġr outes +ĠB uy +Ġcomm ands +Ġm esh +/ C +ĠSet tings +- style +Ġw itness +Ġc le +Ġun ion +ef ault +are t +Ġthought s +Ġ ---- +_pro cess +_ us +ing ly +U ES +T ouch +ĠÐ ¼ +_ open +ĠV ec +Ġre ward +.C lick +/ : +Ġn ie +Ch anges +M onth +ï¼ Ł +Ġexec ution +Ġbe ach +( Integer +ĉ a +/ ' +.Font Style +Ġab ort +ĠS ingle +( isset +Ġd p +Ġ}} +Ġ* = +ĠP S +Ġdanger ous +[ p +OM E +O ther +ĠString Builder +Point s +head ing +Ġc urrency +Ġpercent age +_A PI +Ġclass ic +the ad +ĠM O +F E +Id x +aw ait +Ġà ¨ +Ġacc ident +Ġvari ant +Ġm yst +ĠL and +ĠB re +Ġh arm +ĠA cc +Ġcharg ed +ion es +Vis ibility +ar ry +ĠL anguage +Ġwalk ing +" .ĊĊ +if er +Ġleaders hip +.F rom +yn am +Ġt imestamp +i pt +ĠH as +REF ER +ĠIt s +Ġlist ener +UT E +_d escription +Ġexperi ences +Ġcre ates +R S +c art +bl ack +Ġcho ices +w ar +Ġ'' ' +Ġorder ed +Ġeven ing +Ġp il +Ġt un +ĠB ad +( app +r andom +Ġexp licit +Ġarr ived +Ġf ly +Ġecon om +-m ail +Ġlist s +Ġarch itect +ĠP ay +Ġd s +ĠS ol +Ġveh icles +H z +- com +Ġk ing +_e qual +ĠH elp +Ġab use +-- ;Ċ +Ġex tr +Ġchem ical +ä ¿ +Ġor ient +Ġbre ath +ĠS pace +(e lement +w ait +DE D +ig ma +Ġent r +Ġs ob +- name +Ġaff ected +ik a +Ġco al +_w ork +Ġhundred s +Ġpolit ics +sub ject +Ġconsum er +ANG E +Ġrepe ated +S end +Ġ# [ +Ġprot ocol +Ġlead s +use um +E very +Im port +(c ount +Ġchalleng es +Ġnov el +Ġdep art +b its +.C urrent +Ġ` ${ +ot ing +( \ +Ġcreat ive +Ġbu ff +Ġintrodu ced +us ic +mod ules +A re +-d oc +l anguage +_c ache +Ġto d +? > {{ +ĠRes ource +ĠSt andard +ĠP rem +up dated +ival ent +Ġas sets +_t emp +Ġinterest s +Ġhard ware +ĠR om +ĠSh are +Ġ' 'Ċ +Ġ* , +ĠT ake +ĠIm ages +_C HECK +(type of +ĠJ un +\< ^ +Ġli qu +Ġwor st +ymb ols +ĉĉĉ ĠĠĠ +Ġdr ivers +ĠD ocument +en o +ĠTechn ology +Ġappro ved +ump s +Ġs now +form ance +_A SSERT +u its +Ù Ĩ +Ġdiffer ences +. Visible +ĉĉĉ čĊ +ĠP s +_f etch +Ġto do +. ',Ċ +Ġs el +ur ers +in valid +Ġt weet +V EL +Ġresearch ers +Ġs printf +ĠR O +Ġp el +.Tr ans +Ġil legal +d ialog +sm arty +l g +_M IN +Ġher o +f inal +Ġp p +.L e +Ġc i +ĉ RT +Ġsuggest ed +p df +ach ing +ĠR o +ĠProp erties +ĠS i +Ġbuy ing +Ġm u +Ġl ands +if iers +ĠF ILE +RO UP +Ġh older +ĠS on +Ġsym pt +.r oute +) ? +Ġarg c +Ġfor t +Ġcas ino +_c ategory +Ġfor um +p refix +apt ure +T ube +em s +im ize +Ġn ue +a us +c ourse +AT OR +() ), +Ad vertis +ING S +Ġack now +ĠKore a +pl ing +Ġwork er +PL IED +h al +ĠRich ard +Element s +ĉĉĉ Ġ +st ar +Ġrelationship s +Ġche ap +AC H +ĠX ML +, & +ĠLou is +Ġr ide +_F AIL +Ġch unk +[ s +_O UT +Ġch osen +_ [ +/ ( +ĠJ eff +_s l +pr iv +ĠCan adian +Ġun able +_F LAG +Ġn os +h igh +Ġl ift +f un +() { +el ly +ycler View +_ as +_L IST +Ġr adi +.get Value +ĠAnge les +ĠS pan +_in stance +it ors +Ġm igration +A K +O h + ® +. selected +ĠG T +Ġadv ance +ĠSt yle +.Data GridView +e ction +Ñ İ +p io +ro g +Ġsh opping +ĠR ect +I lluminate +O U +ĉ array +Ġsubstant ial +Ġpre gn +Ġprom ote +IE W +.L ayout +Ġsign s +/ . +Ġlet ters +Bo ard +ct rl +" \ +ĠJ ones +Ġvert ex +Ġj a +Ġaff ili +Ġwe alth +ĉ default +Ġsignificant ly +Ġe c +Ġx s +act ual +.p er +_st ep +an vas +m ac +Ġtrans l +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Iter ator +Ġo ch +agnost ic +ĠD uring +ĠDE FAULT +Ġt ill +Ġsign ature +Ġb ird +ĠO l +ĠI r +H S +av atar +ESS AGE +Ġe lev +Ġm t +ĠN av +Ġrel ax +Ġpl ate +IT EM +( date +.n ot +Ġgr ade +Ġ} ),Ċ +? "ĊĊ +i ences +H igh +ĠD IS +dis abled +Q UI +Ġno ise +a ux +ĠU P +os a +Ġv oc +Ġ )) +oc om +_O FF +ĠD b +L ock +.e clipse +, d +ĠD raw +Ġ" ( +Ġvis ited +Ġâ Ī +Ġsuc ceed +Ġim possible +a ire +ĠT urn +Ġd ish +F G +Ġs ensor +AN N +ab a +Ġsur g +] );čĊ +Ġf p +_ an +- J +- G +ĠJ ob +Con vert +ĠKE Y +Ġauth ors +_s erver +\ r +Ġ-* - +f lex +Ġs oc +R et +Ġs alt +Ġâ̦ ĊĊ +ĠC lear +(p age +-d anger +Ġroom s +con v +# { +. op +ĠA rea +_S C +h en +Ġbeg ins +- y +Ġexc ited +Ġign ored +Ġbon us +st udent +ĠM ember +Ġrel atively +ĠL ow +ĠPro du +ate way +pos ure +Ġth ick +ani el +( view +ĠCr ush +Ext ension +I l +e ed +LO C +. im +. Items +Ġconflic t +.pre vent +Ġon Create +u v +is er +Ġw ave +M ar +ĠComm unity +ic he +ĠNo thing +[ m +ĠLe e +ri ends +è re +!! ! +an z +. result +ĠS K +_P ARAM +Ġdem ocr +Back Color +.ex ists +" It +( options +ra zy +as er +\ Database +al endar +_ ass +; }Ċ +vert ex +ine craft +W arning +arg o +Ġact or +ĠInst ead +ĠUs ing +S elf +@ interface +Ġspe aking +ĠPar is +ĠL ICENSE +.n ode +ĠF ood +E IF +ĠB i +. Start +ĠI B +Ġun iversity +ĠHe ader +.pro duct +C opy +et c +r ical +Ġ> >> +book s +Ġal gorithm +Ġ' __ +(j avax +Ġnumer ous +Sh are +H ave +Ġrec ru +Ġpro ve +.sub string +he alth +е л +Ġdec imal +Ġcomm ission +s cription +x C +Ġsum mary +att ed +Ġclo ser +fin ished +() ){Ċ +ĠW ood +_field s +k u +_ items +Fl ag +Ġconf idence +ĠF ederal +du x +Ġcomp at +Ġvert ical +Ð ¹ +è s +; ">Ċ +_m anager +() ))Ċ +ID E +: ", +__ Ċ +ĠW ay +Ñ Ī +T emp +ĠS TR +rit ten +S ync +ĠA V +ĠC EO +ĠG uid +Ġenvironment al +Ġcorrespond ing +ĉ console +Ġjust ice +ĠJ S +Ġl ived +g ar +ĠG raph +ĠSt at +Ġi Phone +. al +ĠH D +Ġocc ur +Ġth reshold +Ġon click +RE G +.Graphics Unit +M eta +Å ¾ +Ġc um +.g nu +à « +Ġobt ained +Ġcompl aint +Ġe ating +Ġt ar +_t ask +Ġopt s +( to +P ass +Ġpl astic +t ility +ĠW in +.prevent Default +p ile +ĠG ar +Ġqu antity +_l ast +Ġg reatest +D ao +_D IS +ĠUs ed +ĠH P +rit ing +S ION +bl ue +d omain +Ġs cores +N ormal +_ admin +ĠA SSERT +Th en +** * +d ist +l on +Ġh ate +sh al +Image View +d atabase +Ġp and +Ġlog ic += false +b g +ĠConfig uration +Ġn ur +O G +Ġmar ried +: + +Ġdro pped +Ġreg istration +оР¼ +ult iple +iz ers +sh ape +.c opy +Ġwe aring +ĠC ath +Ġded icated +Ġ.. .Ċ +Ġadv oc +ĠF amily +Ġstat ements +em atic +ampions hip +Ġmot iv +ĠH ave +Ġbl ow +J ob +c ert +_v ector +inst all +ĠC OPY +em bed +D IR +ĠS pring +Ġex hib +cd n +ĠCom ment +ĠOption al +. player +ĠD ark +( pos +ĠSh ould +Ġcent re +ĠGu ard +ó w +Ġtr ouble +EN ER +( unsigned +_s ervice +Ġn s +ul ing +ĠMex ico +ĠN Y +mys ql +Ġl ic +å ľ +M r +- fl +ĠC ustomer +id i +Ġ? >ĊĊ +ri ble +Ġп ÑĢ +Ġs izes +_STR ING +valid ation +ĠJ on +( Http +add Class +N odes +Ġfrag ment +Ġsp oke +Ġw aste +J oin +Ġill ustr +el i +c ient +Ġa id +Ġpro sec +') {Ċ +Ġpass ing +Ġf aces +Sh ape +_ Z +it i +Ġal le +Ġro bot +ĠĠĠĠĠĠĠ Ċ +ĠS pe +Ġrece iving +ĠD etails +Ġ" ) +m g +_RE F +Ġcompar ison +* , +ĠF ound +_s ession +( U +/ F +Ġx xx +N etwork +d ers +Ġcap ture +Ġcor re +ĠL td +ĠAd v +[ @ +Ġcl ip +M ill +ĠPro file +Ġend if +Ġob lig +des cribe +.e lement +riter ion +L D +er ed +Ġfav our +s core +ĠF ilter +at tributes +Ġcheck s +In flater +ĠPl us +Ġscient ific +Ġpriv acy +He ad +Ġfe at +Ġdeg rees +ĠP ale +; "> +Ġfil ms +ĠA udio +ĠT ag +ĠE nergy +it ar +par ator +Ġf ellow +Ġev t +ĠT ri +ĠD AM +cl oud +ĠP assword +ĠDemocr ats +ĠAc ad +$ lang +Ġre b +() )ĊĊ +н Ñĭ +ĠB ur +read cr +Ġh ex +Con sole +ct l +ous el +ĠWill iam +Ġa z +_P ORT +Ġpract ices +Ġany where +ĠP osition +Ġ- >Ċ +i ams +.user name +place holder +Ġo der +ĠSecret ary +Ġi T +mon d +event s +? âĢĿ +.S ub +Ġatt ached +Ġn ão +Ġest ate +. action +Ġfig ures +Ġ} );čĊ +Ġsubs cri +.t ag +n am +. plot +no on +li ament +Char acter +.t ab +Ġw inter +ĠVar iable +Ġtre es +Ġpr oud +( V +_ load +Ġh ier +ĠE con +Ġf d +Ġvict ims +R est +ian a +Ġf ake +.Print ln +Ġstr len +Ġs ad +Ġb le +Pro t +Ġbutton s +Ġte levision +Ġlog o +ext ension +ĉ j +ste in +acion es +Ġ"" "ĊĊ +Ġsim p +Ġrecord ed +Ġbr ings +Ġprincip al +Ġfe es +(s ource +k dir +Ġutil s +Ġcorrect ly +f il +Ġw el +P air +-b utton +s cale +ver ify +[ c +Ġ-- - +Ġes cape +ik es +Lower Case +ic ian +Ġch apter +ĠT YPE +Ġsh adow +Ġaw esome +W E +el if +Ġl ambda +Ġdist inct +Ġb are +- off +Ġcol our +.append Child +ole c +ag a +.f ill +ĉs uper +Ġad j +( position +.get Item +Sh ort +Ġtot ally +V D +ĠT re +_ ep +v ements +ĠS olution +Ġfund ament +F ollow +Ġfac ility +Ġhappen ing +O F +.text Box +S pan +Ġ « +id en +Ġex ceed +(p arent +Ġc p +ç » +Ġhas n +Ġp ri +Ġcon sequ +n en +ĠIN TO +I gnore +ĠF uture +Ġcar bon +ĠSte el +f mt +ok ie +Ġs pl +(t itle +- info +Ġde als +Ġfix ture +e a +D iv +Ġtest ed +_ return +)ĊĊ ĊĊ +upport ed +ĠC ook +Ġpay ing +ĠI ll +Ġarrest ed +ĠPr ime +_c allback +> ,Ċ +dr iver +On ce +ab b +_by tes +ĠS ets +( Object +Ġc c +Ġsh ell +al o +); // +( log +ct ors +) +Ġ$ (". +.p os +Ġbo ys +Ġwed ding +Ġag ents +=" _ +ĠAr my +Ġh int +v ision +Ġte ch +ĠCon nect +Ġleg end +ĠB et +.B ase +Sub ject +Ġl it +Rem ove +Ġ" : +ĠF inal +pear ance +ĠiT unes +Ġparticip ants +ĠPy thon +Ġbus y +i el +vert ices +Ġtemplate Url +ĠC lose +Im g +ĠCorpor ation +t imestamp +Ġext end +Ġwe bsites +Ġposs ibility +о ÑĤ +Ġk ö +Ġme at +Ġrepresent ation +Ġ ĉĉ +_ST ART +.app ly +ĠVal ley +ĠS uccess +H i +Ġn ob +ĠI Enumerable +_ select +ge o +. ")Ċ +Ġturn ing +Ġfab ric +(" ");Ċ +Ġpers pective +é Ĺ +ĠS n +Th ank +; j +.Param eters +ĉ ĠĠĠĠĠĠĠĠĠĠĠ +Ġfact s +Ġun t +.in stance +################################ ################################ +- end +ĠJO IN +ĠH en +Ġur i +åIJ į +Ġн а +ĠIn fo +Ġconduct ed +Ġà ¥ +OUR CE +Ġw ine +J ohn +.Error f +ĠA ge +ound ed +Ġreal ize +Ġ] ; +Ġsub sequ +, m +( User +ian o +Ġaccom pl +is p +.st d +é ĩ +ĠB ed +.set Attribute +B R +ke ep +ĠA LL +Ġis ol +am ma +P ackage +Ġoccas ion +-s uccess +еР´ +ĠLIMIT ED +st rip +() ĊĊĊ +istrib ution +Color s +Ġ+ :+ +Did Load +al er +Ġt id +ĠL ED +ĠLink ed +ĠC art +() )čĊ +_RE AD +Ġkill ing +ĠP HP +fe ction +Ġinst ances +c v +"/ > +Ġs f +Ġtax es +_ location +ĠBit coin +u able +r ank +ign ore +tr ack +к а +Ġshould n +ĠO P +=> {Ċ +Ġk m +Ġh elper +_ head +ĠWh ether +oc o +_b l +Ġstat istics +Ġbeaut y +Ġto g +t ip +ëĭ ¤ +Ġc sv +(s ql +std lib +we ak +Ġlik es +Ä į +Ġrepe at +Ġap artment +Ġem ph +_ edit +Ġv it +ĉ type +E ven +ut en +Ġcircum stances +b ian +Ġs ugar +W indows +ì ŀ +Ġobs erved +/ data +Ġcal endar +Ġstri ke +ĠR ES +_s c +f ony +ore m +( z +p ower +et ect +ĠS at +.d escription +Ġg ang +ĠS ports +ong s +ĠB undle +.s um +on ce +Ġacc used +Ġexplo re +Ġapprox imately +Ġlos ing +thes is +ĠF und +Ġdi agn +A utowired +prop erties +Ġ_ . +Ġc nt +ced ure +Ġy y +Ġgr ant +so ck +.inner HTML +Ġ] );Ċ +ĠCON FIG +=' $ +] ];Ċ +UN D +Ġg lob +Ġd ire +uff le +_M EM +Ġauth entic +> (" +Ġdec ade +ĠIm port +Ġorigin ally +Ġj Query +Ġindic ate +Ġours elves +S w +.l bl +ener ate +Ġbas ically +ĠH om +Ġ+ #+ +ĠBrit ain +ĠK ar +to Equal +.st op +Ġmod al +is i +Ġsuggest s +Ġd type +Ġt ur +b f +Ġconnection s +ĠB efore +ist ed +m ouse +Ġpul led +.b uild +Ġlegis lation +Ġfor th +p ad +eg o +.N ow +Ġexc iting +}ĊĊ ĊĊ +Ġcom pr +Ġsh ares +Ġr ig +g reen +_ vec +Ġenumer ate +A uto +ic ator +ĠR ay +as se +Ġh oliday +Ġnull able +g un +_d etails +Ġwr apper +se q +ĠYou ng +ju ana +Ġ" __ +lic ense +ser ve +^ ( +id ers +.Rem ove +rop down +' S +p in +(t oken +.D efault +Ġreason able +amp ion +ĠS ociety +Ġbe i +erv es +r ad +ĠF ox +_ images +Ġw heel +') [ +Ġc fg +( By +Con structor +Ġv ary +.sw ift +Ġpro xy +ĉ H +ĠAn other +ĠP en +Ġcheck ing +Ġj est +man ager +Or igin +ug s +o ir +>< !-- +Ġexpress ed +Ġmod er +Ġag encies +Ġi h +-h idden +ious ly +ĠR od +Ġso le +M ed +.A ny +Ġp c +b al +Ex ample +ĠS ale +Ġst rip +ĠCom p +Ġpresident ial +M ost +put ation +( ref +ĠF our +_f ilename +Ġen forcement +Ø ¯ +ĠGe org +we ights +/ l +Ġag gress +Ġd rawing +and y +< I +- j +ak a +h ref +Ġteach ers +_ Q +( it +ĠM B +Ġtemp orary +ire base +str a +æĹ ¶ +è ´ +( label +ou p +Ġtop ics +Ġport ion +id os +ĠJew ish +Ġre covery +Ġstand s +# [ +Ġafter noon +ĠArt icle +_ att +Ġexpl an +ĠP ak +.setOn ClickListener +. children +Ġi k ++ ( +l ag +Ġdis k +Ġcont rovers +"> & +as p +Ġw ie +ĠAustral ian +ĠYou Tube +At tr +cont ains +du ce +ĠM att +at ern +Ġvol unte +Ġnew sp +V P +olt ip +Ġde legate +_m eta +Ġaccur ate +ĠEx ample +% , +ĠD aily +Ġc abin +ĠS W +Ġlim its +k ip +Ġar my +Ġend ing +Ġb oss +ĠD ialog +Al so +="# " +ord an +row se +- min +Ġ" & +_ loc +U X +Ġdevelop ers +Ġaccur acy +Ġmaint enance +Ġhe av +Ġfil ters +.T oolStrip +Ġn arr +ĠE mp +ORD ER +ĠM obile +.S erial +.out put +.c ol +M aterial +um a +Ġconsum ers +sh ift +Ġp ued +Ġmin i +c ollection +Ġk an +.c enter +H istory +Ġben ch +() ); +itor ies +Ġcrow d +_c all +Ġpow ers +- E +Ġdis miss +Ġtalk s +ĠCh annel +for ward +_ control +/s rc +i est +**************** ******** +Ġbet a +(c olor +_O BJECT +ĠA pi +Ġeffect ively +C amera +s d +uss y +D ict +ĠE ffect +ib ilities +Ġreturn ing +ĠF ar +Ġ' ') +Ġmod ules +il ation +Ġ( % +TR GL +Ġst orm +on na +ĠEX P +Ġs pons +Ġdis pl +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠ +f all +å Į +ign Key +_ US +et rics +Ġhand les +T L +_ amount +ow a +br and +ĠT ool +Ġus ual +. Z +cre ment +ad ium +st ock +Ġserv ing +ĠB on +Ġline ar +ĠT arget +ĠR adio +H L +Sh ader +om atic +ag ues +in ity +d iff +_ iterator +qu ot +Ġ ,Ċ +c allback +Ġsympt oms +[ _ +ĠB ul +ĠF eb +und o +_ account +Ġtyp edef +и Ñģ +tr as +User Id +ĠP enn +ĠSup reme +} > +user Id +ĠK im +Ġg a +Ġart ists +å ¸ +ĠAb stract +ok emon +Ġh am +o val +Ġch a +at en +å Ĩ +F ixed +Ġvul ner +ĠParam eters +qu antity +.C lear +Servlet Request +Ġy a +Ġsou l +trans action +Ġsol o +Ġp airs +æ Ķ +ĠG re +_ word +ĠC C +Ġg i +z ie +Ġsched uled +rot ation +gy pt +ul ous +:: _ +ĠE ll +< ! +ĉĉ ĠĠ +l p +ah a +C opyright +Ġdr am +Ġdi agram +ĠM em +Ġg arden +Com p +Ġattempt s +uff ix +> () +Ġphil osoph +_re l +å ¼ +Ġs v +.se cond +ant o +.J son +ĠTe le +_ local +_s end +Ġas pects +ì Ĺ +IB LE +Ġr ail +Ġwid ely +ash ed +i ar +in f +up per +d jango +_result s +iss ing +Ġequ ivalent +OUN D +Ġt y +Ġpotential ly +Advertis ement +ĠRec ord +resent ation +_w idget +ound ing +Ġrelig ion +Ġcons c +ĠL im +. am +H tml +Ġ' : +P ATH +_s pec +ort ed +id ades +_sh ape +Ġkeep s +.S ave +ĠL oc +or i +ĠT EST +unic ip +Ġreg ions +Ġbelie ves +/ en +pos ite +{ ' +pre pare +_ const +s ample +ĠWill iams +Ġstr t +_ Get +ĠAnd rew +. active +Ġl ayers +Visual Style +az y +ĠK n +Ġac id +ĠAs ia +Ġex cess +ĉm y +Ġkey board +ens us +Ġcre w +Ġmiss ed +m aster +ĠW ild +Ġnew ly +Ġwin ner +Ġst ub +ic ode +.m ove +D omain +ĠS ar +Ġfore st +LE D +claim er +.ex it +ĠW indow +Ġres istance +ĠC HECK +(" - +ĠR yan +Ġp ipe +Ġco ast +DE F +// ! +_ off +ex it +Ġult imately +imit ive +ĠKe ep +Ġhistor ical +Ġany way +ĠJack son +ock er +ER N +ĠU INT +y ntax +ER Y +is ms +Ġc n +Ġocc urs +Ġ; ; +Text View +A E +/ img +Ġy esterday +- default +Ġt iny +Ġpro c +Ġal ive +ĠRE G +. th +ear ing +.get Logger +< link +_ login +F older +ab c +lyph icon +н о +Ġnot iced +od igo +Ġed ition +im ator +. Enabled +.parse Int +Ġy ards +ĉĉĉĉĉĉĉĉ ĉĉĉĉ +Ġver bose +л Ñı +_B Y +.log in +.* ;Ċ +ĠM id +é es +Ġg lo +Ġbuild ings +Ġz e +ĠI ter +Ġt ube +ĠP ot +\ M +< th +br idge +ĠS cript +ĠM odule +Ġv acc +Ġinstall ation +v y +VisualStyle BackColor +ĠS M +.t otal +b at +Ġfind s +Ġat mos +Sub view +iz ard +Ġrepl acement +lic ated +ap is +Ġlog ged +ĠLe ft +G ui +_ Type +t m +P ad +Ġhouse hold +Ġre le +Ġpropos al +_CL ASS +:: :: +Ġinf rastructure +In ject +/ html +Ġad s +iz za +Ġm g +ctr ine +% Ċ +< html +- image +Ġatt orney +< m +(' , +Ġcan n +Ġprint ln +o ose +Ġy ellow +.ex p +p ayment +Ġtable View +aw ay +Ġopp osition +ĠAg ain +ĠH andle +Ġex clusive +in ar +é r +оР± +ĠC ODE +emp orary +Ġre act +pi pe +c z +. activity +Ġlarg ely +Ġdis s +ax y +es is +ĠR en +Ġc orn +.Use VisualStyleBackColor +d ays +Ġfr uit +In sert +_ enc +E st +_de c +ĠL uc +Ġü ber +param eters +P ERT +ex press +_pro file +Un known +Ġrev olution +.add ress +_re quire +Ġun iform +ĠP ack +l ar +ĠU ITableView +Ġdep ends +Valid ation +conf irm +O wner +Ġt rib +h et +ĠI de +ans as +L anguage +u et +ĠP o +ĠSte ve +Ġcont est +_DE FAULT +Ġapparent ly +RE EN +Ġfrequ ently +Ġtrad ition +ocol ate +S I +ĠArg ument +F ocus +ert e +ĠL ayout +Ġd x +Ġgener ator +ĠW ait +P olicy +l ights +.Ex ecute +P y +Ġbed room +ed a +ra id +ĉs ize +Ġan cient +Ġp ump +Ġd w +Ġ(! ( +Ġspec ify +( status +ĠF BI +.ex ception +Ġrem ark +ly mp +ant ee +Up load +ern et +é ¡ +in ent +ĠR ender +d m +ĠM emory +r ich +ĠT ools +Ġk ne +Ġper m +b ad +Ġd inner +.res et +Ġj Label +Fe ature +.S ervice +Ġ( {Ċ +Ġre ferred +.class List +Ġinit With +ĠText View +Ġne ither +Ġcount y +Ġ" { +ç § +Ġt ack +class Name +ĠUS ER +Ġre new +` ` +get Name +Ġb rown +Err ors +ert o +Ġsust ain +S O +let es +ĠIn valid +Ġen emies +un ge +Ġexist ence +err a +Ċ ĠĠĊ +utor ial +# a +p ay +char ge +ĠI re +ate st +Ġexp los +Ġf ired +N ER +ĠT y +ic ion +U ri +Ġobvious ly +ĠC olum +Ġ' + +ĠDe vice +- related +_ ARG +Ġv or +ĠLess er +_O P +Serial izer +Ġup grade +L ight +Ġc odes +++ ;čĊ +Ġwrit es +fo od +Ġé t +@ section +Ġtrack s +Ġserious ly +ch t +(size of +Ġimmedi ate +Ġscient ists +Ġ{ $ +_ ne +.Anchor Styles +Ġaccom mod +ĠHar ry +Ġs ight +ĠPale st +ersist ent +Ġ Ñĥ +- input +Ġco ordinates + · +W elcome +.con f +Ġgre w +Ġb old +ĠC PU +(m y +Ġperfect ly +Ġmom ents +ĠM ovie +- data +yst al +_W IDTH +ĠS creen +æ Ŀ +Ġdis ap +Ġredu ction +.Get Component +_M ODULE +Ġgener ic +Ġd y +all er +Ġc url +ĠB ody +Ġb anks +, t +av g +Ġev il +Ġmanufact urer +Ġrece iver +Column s +Ġing redients +ĉ out +qu es +.L oad +Ġslow ly +ĠT own +ĠC ell +_n ormal +_p refix +ĠAl ert +(" { +ä r +âĢľ The +ĠM D +Ġcour ses +ath an +é Ļ +oc c +ĠS ER +es ign +Add r += [' +(" ./ +] } +.f ont +ĠInst agram +ĠB order +od a +Ġh all +Ġr um +_b it +Ġs aving +_d own +R andom +_reg ister +( Context +Ġoppos ite +R oom +Y ES +ан и +Ġenjoy ed +_r un +C lear +âĢ ĺ +ĠF ord +on ic +ost en +"] ) +_ auth +// čĊ +Ġsuff icient +LE S +Ġph en +Ġo h +_c sv +Ġrout ine +.Are Equal +ay lor +Ġb asket +_COM M +rypt ed +S im +ĠSh op +Ġstud io +at os +( W +[ string +ä t +og a +Ġsh r +Ġs ick +An other +Ġdo ors +_N E +ĠTH REE +. order +raz il +Ġmap s +_TR UE +trans late +Ġnear by +Ġn ach +LO AT +b atch +Ġl ux +ash es +ang ers +â̦ â̦ +_E VENT +_ UP +Ġact s +in v +_M ETHOD +cc ion +Ġret ain +ut ch +ĠÐ ± +Ġknow ing +Ġrepresent ing +N OT +p ng +Con tract +Ġtr ick +ĠE dition +uplic ate +Ġcontrol led +c fg +j avascript +Ġmil k +Wh ite +Se quence +aw a +Ġdiscuss ed +ĠB ush +ĠY ES +.f actory +t ags +Ġt act +Ġs id +$ $ +ĠE num +Ġfr ames +} ); +Ġreg ul +'] ;čĊ +Reg ion +ff f +Ġc ro +( com +=" + +St udent +Ġdis appoint +RES ULT +Count er +Ġbut ter +ĠH a +ĠD igital +Ġb id +"> {{ +ing ers +ĠC ountry +_t pl +"] )Ċ +/ k +d ating +: # +ĠD ATA +yn chron +_b ody +olly wood +Ġval or +ip ient +o ft +UB L +doc s +Ġsyn chron +Ġform ed +ru ption +Ġlist a +Request Mapping +Ġvill age +Ġkn ock +oc s +" { +_fl ags +Ġtrans actions +Ġhab it +ĠJ e +ed en +Ġa ircraft +ir k +ĠA B +Ġfair ly +. inter +.A ct +Ġinstr ument +remove Class +.com mand +Ñ ī +ĉm em +( min +Ġo t +Ġcol le += s +time out +Ġid s +ĠM atch +ij n +z ero +Ġnetwork s +.g ov +Ġint el +Ġsection s +out ine +(c md +(d ir +ĠLI ABILITY +ĠB log +Ġbr idge +ĠC V +con vert +Ġ" )Ċ +ĠB ern +_P O +e val +( set +to ol +Ġpay ments +Beh aviour +Ġcon crete +Ġel ig +Ġacc eler +Ġh ole +_ o +TE GER +Ġgraph ics +O wn +Form atter +on der +Ġpack ages +/ a +ĠK now +Or Default +Ġdut y +W ait +н а +_rec ord +[ t +M esh +Ġon going +.be ans +Ġt an +Ġinter pret +ast ers +QU AL +Ġleg s +\ Request +- file +_m utex +ĠS aint +// # +Ġpro hib +( info +: = +lin ux +Ġb lo +ot ic +ĉf inal +_ex p +ĠSt op +ap ing +(s aved +_p ush +Ġe ase +_F R +pons ive +str cmp +: ĊĊĊĊ +ä» ¶ +ol i +Ġextrem e +Ġprof essor +Im ages +.IO Exception +Ġaddress es +plement ed +Ġincor por +Ġuse Effect +_O F +ĠD a +n ombre +IR ST +Ġdisc rim +Ġcomp ens +greg ate +anc ell +ach es +ĠC riteria +$ result +D estroy +Ġsecond ary +W atch +ĠS em +ĠMc C +Ġacad emic +U pper +:: ~ +ut ral +ĠD og +ad ed +Valid ator +Ġder ived +Ġset Timeout +ĠK en +Ġtyp ical +ĠB ob +Ġb ounds +ĠSe ason +Ġc razy +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠ +-r outer +itt est +ĠM ir +Ġemot ional +, v +c n +/ st +å ½ +on om +Ġdecl ared +> . +ail ing +Ġ/* <<< +Ġnorm ally +(M e +ev in +lik ely +Ġpoint ed +ĠSt ack +Ġw alls +. Vector +me an +] ]Ċ +Ġlist ening +ad v +Ġsw ap +IF T +Ø ª +. argv +ul s +< option +not ations +Ġemail s +ĠU kr +ast a +ĠTh us +ĠSt one +Ġappe al +. âĢĻ +Ġreg ulations +Pre ferences +ĠPh one +ul f +ĠD R +Ġtechn ologies +Ġpar agraph +Ġnecess arily +.e ach +< float +res a +Ġunder st +Ġf inger +press ed +-b y +if fer +w atch +ĠB a +A IM +Ġwe ights +ĠR on +') }} +[ self +-------- --Ċ +per iment +Ġto String +x ic +ĠC amera +! ĊĊĊĊ +aur ant +P refix +Ġinstit utions +: int +Ġex posure +p attern +ĠLin ux +.n umber +red ient +Argument Exception +ĠCh ief +" }, +Ġelect ronic +r ong +er d +sp Net +ra it +/ ', +ĠOh io +Cont rollers +Ġcontin uing +ĠT emplate +ĠE th +s z +/ env +En v +% . +art ers +) (( +ĠT ABLE +Ġà ® +per ature +pro gress +P res +ê ° +im plementation +Ġb ien +Ġstre ets +_M SG +New s +## # +: / +Ġcut ting +x B +ress ed +_EN ABLE +l ab +Ġca using +] ));Ċ +b ra +x FFFF +il ly +plet ion +w ill +_b ar +Ġstruct ures +ĠI mp +Û Į +Ġ< > +Ġ ---------------- +_B UFFER +.d ir +Ġpl ain +Ġpe er +g g +oint s +Ġsomew hat +Ġw et +Ġemploy ment +Ġtick ets +ir ms +Ġt uple +s is +$ sql +r ig +Ġcon version +Ġg es +Ġconfig ure +eg r +ĠC a +Ġ__ (' +ou ston +.t oken +Bl ack +Ġmag azine +A W +. IN +os ing +Ġbro ke +ĠC ru +DE LETE +Ġdestroy ed +(M ath +Ġappro val +-d om +ĠI II +table View +Ġdesign s +Ġcrush ing +Ġcons ent +dir name +om p +Ġc rypt +? ( +or ough +. o +ĉ list +ams ung +."" "Ċ +err ing +G oogle +_p air +_IN IT +rem arks +Ġg ear +F ill +l ife +} ")Ċ +Ġsuit able +Ġsurpr ised +_RE QUEST +Ġman ifest +att en +Ġfr ustr +ov ement +.c lick +Ġi i +Ġexp ansion +ig s +P arse +.Reg ular +R ob +_l ayout +ì ł +Ġtrans lation +ĠBe aut +B est +_C OLOR +< label +Ġliqu id +IT S +Ġpro d +Ġoper ate +UI Kit +Ġn atur +arg ument +_d etail +ĠCent re +Ġ" -- +Ġ}} " +lo cale +.t v +_se q +Ġup coming +Ch art +ĠDiv ision +Ġclin ical +Com pany +S epar +l as +ĠH un +: s +Ġhead ing +оР³ +Ġ" ");Ċ +[ id +b ia +Ġst retch +ic ide +Ġre produ +.pro ject +leg end +end ers +Ġrespons es +Ġon t +rit ical +Ġref uge +ĠL i +Ġ: ĊĊ +ĠTh ree +.cont roller +_IN DEX +_F OR +\Model s +j ax +ĉex it +Ġâ ĸ +Ġc overs +ĉ y +- . +IND OW +Ġfail s +in cludes +Ġf ault +Ġl y +ñ o +.s lice +ILE D +ĠP ur +ĠAs ian +_b atch +.M ax +v l +ĠCOPY RIGHT +Ġg iant +ĠMan ual +ĠC opy +Class Name +He alth +C ursor +IB Outlet +Ġt we +æ ³ +_label s +Ġcol lected +Ġfurn iture +Ġdeal ing +Control s +ĠHot el +ck s +Ġch ose +âĶ Ģ +od d +S R +Ù Ĭ +ì Ħ +Ġacc ord +ĠM ove +ĠM ode +ĠM ock +Ġthread s +++ ++ +ĠO ptions +Ref resh +ĠD id +'] -> +u cc +_ch annel +. abs +Ġ{ },Ċ +ĠW al +er ior +Ġmain ly +ĠDr iver +NotFound Exception +Ġcount s +e am +Ġ& = +Q uestion +ĠA li +Ġany more +d etail +t ail +Ġm ile +ĠF air +Ġs orry +Ġsurround ing +Ġad m +De v +Ġmari juana +ĠS ound +ĠA sh +F D +Te am +. port +Ġ[ ]ĊĊ +ub ble +Ġas c +Ġint ention +A cc +ch i +ust ers +Ġins pired +se g +CL U +Ġman ip +M etadata +Con nect +ĠB eh +Ġfind ings +Ġas sembly +w orld +Ġrem ained +Ġu id +( . +Ġm x +Lo op +ĊĊĊĊ Ċ +Ġfant astic +wh o +ak i +ĠB asic +ĠY et +ĠUs ers +ik ip +Ġhead s +ĠMich igan +_ it +ĠTor onto +Ġrec ording +Ġsub mitted +_var iable +medi ate +.graph ics +Ġst ood +Ġre ar +vel ocity +_M ESSAGE +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +ro les +ĠT our +_ year +end ment +amp s +ĠIre land +m al +Ġyoung er +Ġstrugg le +Ġc able +ĠSD L +(' - +an es +ĠNe ed +.R ow +P ol +ĠP H +_s cript +ag em +ĠB as +_s pace +. loc +: i +ad r +Ġengine ering +it en +) & +Ġu k +ĠL ittle +_C OUNT +x A +Array List +æ į +Ġ" ")Ċ +An chor +Ġh ang +t witter +Ġcompet itive +.s rc +ãģ Ĺ +Ġtrans late +ĠCre ates +ook s +ĠR oll +'' 'Ċ +/ sh +s ome +Enc oding +.res olve +Ġdesign er +ĠSt orage +Ġz a +ĠN ever +Ġsomew here +Ġbox es +.s ource +Ġpy game +Ġgrow n +.t w +() ),Ċ +', [' +Ġoppon ent +(s rc +.l ayer +AP P +ĠAct iv +Ġguest s +ĠVAL UES +};ĊĊ Ċ +.n ative +Ġamount s +. RE +Ġcl one +Ġwer en +Ġ" << +_ ac +Ġbreak ing +Ġreli able +.P OST +ĠSk y +Ġ' & +Ġsaved InstanceState +ast ing +ill ion +com ments +ult y +.m enu +/ config +Ġ ĊĊĊ +T ODO +Ġpurch ased +_c or +ĉ auto +Compat Activity +com plete +_ graph +is odes +Ġsitu ations +ĠH or +Re ceive +âĢľ We +Ġent ities +.assert Equals +оРº +ĠS ans +v ince +rom pt += Ċ +Ġ/ . +.Se lect +yl v +Ġb att +A udio +Ġincreasing ly +.B undle +Ġexpl ains +the ast +. offset +Ġh al +Ġtechn ique +_l imit +Ġdraw n +AY ER +Ġfeature d +yy yy +at in +ph en +ach el +! \ +l ower +ĠG R +Ġp ag +ĠP arse +Ġt ou +ä¸ Ģ +D istance +Index Path +Ġh ell +s im +UT TON +Us age +elen ium +ĠF all +Ġ" .$ +ĠM u +Ġcr uc +Ġs ont +REF IX +Ġinter ior +ĠO lymp +.Auto Scale +par a +Axis Alignment +Ġr iver +D to +Ġwith draw +Re act +- class +b efore +_ alloc +Cont ents +ĠW as +I CT +Ġform ula +Ġindic ates +ĠĠĠĠ ĊĊ +_st ore +it ting +ĠIt alian +_S et +_re port +Ġp id +_V ER +Ġw ins +ĠCl oud +") {Ċ +ch ester +Ġden ied +Ġw ird +ĠSte p +Ġinvest ors +b old +_d isplay +ou ver +or er +Res et +Ġsurg ery +Ġstrateg ies +/m aterial +_ unit +Ġc ouncil +.P er +ĠâĢ ŀ +Ġre form +F ramework +Ġlist ing +_b tn +Ġb is +% d +eg as +Ġsudden ly +_S ER +Ġa o +_d irectory +f as +Ġprem ium +Ġtrack ing +ĠB L +Ġm ature +Ġbath room +Ġ'/ ' +ĠÄ ij +Per formed +Ġsold iers +arn ings +Ġwalk ed +- con +b ottom +Ġsurpr ising +Ġg ene +Us uario +.DE FAULT +ĠM IT +C ODE +ĠE gypt +p icker +ys ql +AT URE +d etails +ĠCon ference +In formation +ĠM ail +-d own +r aries +b ro +Ġsubject s +Ġ' * +è¯ · +or ient +: @ +ver bose +E F +Ġto ler +eng ers +Ġend point +Ġstr ange +Ġcol on +Ġpre ferred +de p +ĠE V +ARR AY +Ġw he +Ġp up +_n odes +Ġtalk ed +Ġinstit ution +db c +Ġex posed +te en +ĠFr ont +T T +_N ONE +\/ \/ +pro gram +Ġencour age +. ` +sh ire +ĠIsl am +e en +N I +' " +.W idth +Ġlik ed +Ġ{ ... +ĠSystem s +Ġvot re +Ġmanufact uring +Con verter +ĠIn f +ì ļ +D TO +Ġin ches +Ġ ठ+à ¹ +ĠChar les +B U +")) ;ĊĊ +ĠL abor +un n +Ġest im +m obile +ĠL earn +_C ALL +â Ħ +Ġind ices +Ġt ub +ikip edia +C ost +row able +ë ¡ +g age +Ġfunction ality +uzz le +em os +.l ib +Ġd ass +еРº +enn a +Ġsh ots +Ġrest ore +/ D +For Key +], [ +al ias +l int +.st ream +æ ł +_FORM AT +Ġsil ver +.re pository +Ġlegis l +.B order +_fe atures +Per mission +Ġhous es +ĠW ars +_COM P +Ġinj uries +Ġconstant ly +fl utter +EN U +ĠCon f +Ġrecogn ized +Ġpract ical +Ġde cent +B J +] ); +ast y +ĠAct ivity +-m ode +Ġsl ide +.IsNullOr Empty +ĠY OU +P ower +ind ices +Ġqual ified +Ġthrow n +h ello +ĠN ick +l ah +as sembly +ĠSm all +old ing +Sh ould +ĠSil ver +(saved InstanceState +Ġtog gle +.N ot +C trl +: nil +ĠCont inue +ĠB oot +æ ī +ĠM ur +d on +ĠF A +S napshot +Ġassoci ation +fo x +, a +az ione +] )čĊ +CT YPE +Ġf ade +ĠD ar +.n avigation +Ġl uck +SC RI +ĠDe ad +Ġterm inal +_LE NGTH +Ġeff iciency +Ġun w +Ġn arrow +iment o +( Color +ĠSe a +_ area +, A +_ opt +ĠHill ary +.t ask +ĠJ ac +ast ed +ĠAd am +ĠIl legal +Ġsearch ing +Instance Of +J ava +ĠForm at +Ġreal ized +ĠChild ren +Ġk il +(f rame +âĢĿ .ĊĊ +Ġscen ario +"] );Ċ +Ġincred ible +li x +IO Exception +ĠQ uest +il ty +Ġun lock +â Ĥ¬ +Ġre ferences +ĠV ert +B inding +eg ative +Ġwr ap +.d atabase +( content +B uf +ĠTr ad +ĠA ud +tr ace +.m ock +Ġther apy +ĉ L +.To Int +ĠKing dom +B us +ha ust +"" "ĊĊ +( end +.draw able +[ ];Ċ +ĠH ospital +Ġph arm +---- - +ĠA G +é d +> ");Ċ +Ġw allet +at able +) $ +Ġmonth ly +Ġdi agnostic +S ymbol +Ġiter ator +un finished +Ġimm igration +s r +RO W +(g ame +Ġclo thes +ĠU nt +Ġactiv ation +_C on +.h ash +Ġinitial ly +.H ash +Ġcut s +f ound +ĠSt ory +ÑĨ и +ac ao +_T YP +pro to +est r +-p age +ah r +Ġincor rect +ĠJose ph +TextBox Column +_st yle +ĠD aniel +s heet +Ġl iv +l ined +Ġr a +R untime +_ empty +sl ug +_ struct +ë Ĭ +m u +Ġper mitted +Ġreg ional +Ġsob re +ĠS uch +Ġ[ _ +Ġro of +.Al ignment +t imes +.m sg +Ġche st +ĠT ab +Ġest a +ä n +Ġsubs cription +( command +s pecial +Ġme al +") :Ċ +_ ctx +Ġclos ely +et ry +- be +ad el +ĠR am +ig est +ĠSpan ish +Ġcommit ment +Ġw ake +* >( +P HP +_ { +ck er +< List +_n ull +ĠRes erved +Ġin her +.Column s +.A spNet +_IN VALID +ĠParam eter +Ġex pr +} { +Cell Style +Ġval uable +Ġfun ny +In v +Ġst able +* t +Ġp ill +pl iers +ĠC SS +ĠCon dition +ĠS peed +ublish er +Ġoff ensive +ce st +ic as +Ġsp ark +ĠPro te +set up +IF Y +ĠT ax +Wh o +F amily +- for +. uk +Ġf asc +sv g +") ). +Ġbirth day +âĸ Ī +ve h +el led +Ġimport s +ĠIsl amic +T A +ĠSt an +we ather +Ġsus pect +e ature +enn es +W M +.m inecraft +av id +è ½ +.se curity +in os +G ood +Ġm arch +Ġposs ess +us uario +Con s +am ber +ched uler +Ġhor se +ç ½ +(b ody +ĠTrans form +_de code +.s vg +Ġf oo +Ġd ella +ext ends +am er +Ġprocess ed +ĠH arr +ĠA I +Ġk o +CH AR +( % +Ġt ap +({ ' +c roll +D OM +Ġte a +Ġre in +Ġworld wide +_f n +sh a +Ġb ir +ç ões +="# "> +Ġrepresent ed +ill er +(ex pected +Ġd ance +Ġvisit ors +.con cat +-b it +UR RE +ĠR og +v p +ip h +ĠL LC +it led +iam i +C oll +_re al +_sh ow +_f older +Ġd ar +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġl atter +arch y +Ġb ow +Ġout come +ĠPost ed +Ġris ks +ĠThere fore +Ġowners hip +Ġpar allel +Ġp ending +ge ometry +Ġrecogn ize +ST EM +ĠC P +Ġimm igr +IT LE +ĠĠĠĠ ĉĉ +conn ected +Ġsm ile +(d ocument +\ Component +vert ical +Ġconsum ption +Ġsh oes +. impl +un ks +. ";Ċ +Ġfood s +_ );Ċ +.assert True +Ġp ipeline +Ġcollection s +Ġearn ed +ĠC ert +Ġpartners hip +( action +Ġc d +ĠV ery +Option al +Ġscre ens +Ġtit les +ener ator +Ġab andon +k ind +IL TER +Ġclos ing +lic a +_ inter +Ġcamp us +set ting +S prite +ãģ ¯ +_re ply +To List +: \/\/ +ed e +Ġfol ks +Ġbo at +( argv +Ġperman ent +Ġcarry ing +Ġconserv ative +import ant +. img +ĠIm m +Ġdim ensions +al and +s ingle +Ex it +-------- -- +ari ant +tern al +Se conds +ĠIt aly +ot lin +.Res ume +=' " +) == +cept or +Ġs ca +/m ain +Sec urity +_d at +Ġlet s +Ġa qu +Ġwhen ever +b erry +Ġact ing +ant i +p d +& gt +æ Ń +Z one +T oday +! . +To Props +ab is +it able +Ġg al +] { +iz ona +Ġin contri +N ET +/// Ċ +[ in +_s ave +Ġex em +ĠK enn +Ġev olution +var s +_st ats +- only +ĠColor ado +Ġwatch ed +b our +Ġsever e +Ġprofession als +port ion +Ġguar ante +Ð ³ +Ġpush ed +ĠG i +ï ½ +Ġt um +ĠA z +ĠEdge Insets +")) ;čĊ +is se +. ac +Set ting +Ġapprec iate +ĠValue Error +Ġsur ve +ĠR ole +. Inter +plot lib +j et +d am +Ġplatform s +te le +UT O +ĠInt ernal ++ : +} ;čĊ +Gener al +\ Entity +Ġlawy er +qu iv +ĠPost s +is o +Ġacc um +ob e +Ġmark s +Ġ] ;ĊĊ +ĉ text +.s uccess +cur r +as a +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠ +Ġth in +_ over +are st +ĠO s +( address +Ġvel ocity +Ġ[] ;ĊĊ +=" ../../ +ĠPr iv +b ow +Ġguar antee +% ĊĊ +Ġeval uate +.LE NGTH +Ġin ventory +q a +_de bug +.On ClickListener +Ġl ies +Ġassess ment +dat etime +.background Color +Ġ*/ čĊčĊ +ra f +un wrap +ĠF oot +Ġnot ify +Ġlow est +DO CTYPE +Ġl anguages +ex tra +- back +Ġein en +tem plates +_p ass +ĠM ust +Ġest á +_c ore +ĠSc ot +A I +Ġb ias +ations hip +Con stant +Ġprogram ming +In s +uspend Layout +ĠPRO VID +ant es +Ġsh irt +in ated +. OK +[ a +Ġthink s +? ĊĊĊĊ +Ġregard less +ĠMag ic +ul ating +ĉ class +add Group +RE ATE +ĠS U +Ġsim pl +c opyright +Ġb unch +Ġun iverse +ĠE rr +Ġpresent ation +c ategories +Ġatt ach +.s ign +_A C +Ġdisc ipl +Ġregular ly +Ġprim arily +ink s +[ [ +.r and +.sh ould +ownt own +=" ' +Ġs ans +Ġsupport ers +se quence +G O +. .ĊĊ +ĠS pr +Ġcare fully +U IColor +dest roy +Ġtod os +ĠOR DER +ott ed +Ġd ont +aud i +_ player +g re +ĠO il +< body +_st ack +.P adding +ĠProduct s +Ġpriv ile +Ġinj ured +ĠF urther +Ġal ias +.Resume Layout +_LE N +Ġs es +'] ;ĊĊ +cre ens +Ġdirect ed +.S uspendLayout +od ge +.A t +mark s +ĠUn ivers +ert s +ĠE sc +Ġnav bar +Ġutil ity +agnost ics +Ġin ject +ĠD NA +Ġ" ," +am ar +Ġe u +Ġrestaur ants +_p ut +ut ers +Tool Strip +t w +ist ro +Ġz oom +Ġleg it +pec ific +ĠC ome +Ġlocal Storage +Ġabs or +.P anel +ĠDesign er +Ġo w +IC AL +_ uri +(f ield +Ġsup erv +Ex ists +Ġrespect ively +ĠSt and +Con f +uss ian +Ġar c +Ġ nd +uck s +Ġre str +Ġseason s +ĠCh apter +ĠSw itch +p ic +Ġh i +load ed +Ġfl uid +-b tn +Ġrun time +. it +B N +Op acity +as ant +ry ption +-n ative +Ġta ught +å ¯ +ag ment +Ġm ul +Reg istry +_ grid +ĠBro ok +: Set +Ġm ongoose +AM ES +inner HTML +Ġs oci +ĠInt el +get Id +C md +Ġaccess ible +r ames +le ton +Ġ__ ( +ĉ delete +ĠS quare +" ĊĊĊ +Ġbu cket +avor ite +ĠB reak +++ ] +Ġbr ush +Ġt ensor +/ http +T ile +Ġfunction al +Ġ" * +wh el +Ġt ent +ĠChar acter +Ġse es +. ST +B ig +Ġext ern +Url s +)) )), +ĠJ r +.B uilder +. ; +n l +_ Init +ĠH ER +ż e +mys qli +_ icon +v an +Ġfeel ings +Ġle an +Ġhop ing +T V +="čĊ +b est +all as +ent ed +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĊ +_con nection +Ġrep o +en abled +аРº +Ġsh a +Ġmembers hip +Status Code +in ating +_s m +_c ustom +_ weight +Ġc ss +St at +_ env +link s +TR L +ĠH it +, r +up id +Ġop ens +Ġg ent +_v is +Ġj oy +< w +_c ost +ĠPy Object +ren ce +ĠGeorg ia +ĠBro ad +m ma +â Ĥ +p f +Ġ" \" +Ġ( & +om o +Ġliter ally +Ī ĺ +met ric +Ġb ars +z ed +(w indow +ĠIsrael i +Ġform al +ident ifier +.d ao +ĠDe ath +% ;Ċ +Ġdecl are +ar ms +RE AM +PERT Y +Ġconsequ ences +to ols +Pe ople +ĠWh ich +> ();čĊ +.de code +_A CT +Button s +.f loat +.F irst +ë ¥ +ĠPol it +ĠX CT +T ags +ĠCG Float += str +Ġle af +- check +ĠI ss +.s ystem +log out +ach t +Ang le +s in +ch art +INT ER +ĠN UM +B asic +.P roperties +ä¸ Ń +_ change +ĠB razil +Ab stract +Ġ: +: +_ use +а л +ĠL y +IB UT +Ġout er +Ġ-- >čĊ +Ġrel ief +l ap +qu er +_p arent +he ap +LO SE +Ġcomb ine +ĠR ose +ow ers +Ġproced ures +ĠS ort +an im +var iant +eh icle +Ġsign ing +Pr imary +c urrency +Ġsex e +o en +th eta +em an +Ġimpress ive +(' _ +ĉ U +ĠText Style +_c nt +Ġs lice +(' : +Ġunderst ood +H is +Ġinform ed +Ġn ick +(T AG +h d +Ġelection s +est ure +ĠS anta +ĠCo ast +.p df +inc iple +.cl one +b orn +ut a +Ġl icensed +C r +Ġb read +ĠH ouston +Ġn od +Ġhop es +ĠCG Rect +Ġgu ilty +.g if +Ġro se +.Com mon +T ip +AN K +ĠF C +D uring +ĠSym fony +Ġdef ensive +k m +) > +arch ive +ĠU RI +ycl ing +- o +ĠWe bsite +AM P +ish ment +Ġdo ctors +D irect +AR I +ĠRed irect +ier en +_d ist +y o +ĠPro gress +Ġz um +Ġmem or +ĠE D +Ġj ur +æį ® +_T ABLE +Ġu uid +Ex pr +. head +(' % +point er +Ġest imate +ĠG reg +Ġlo ader +Ġi OS +Ġm ens +[ y +Ġref used +Ġprec ision +is ch +ĠA CTION +Cl oud +s With +( ret +_ADD R +_con f +(d f +Ġlock ed +Ġr ising +ãĥ» ãĥ» +ĠM s +Ġscen es +_EX T +_ raw +_ the +pe ople +Ġre con +ĠF un +Ġb less +ĠUp dated +ü n +ĠĠĠĠĠĠĠĠĠĠĠĠ čĊ +pe ction +Re lease +.log ger +ĠS Y +Ġcoun sel +ur d +_ true +Ġevery body +iv ot +Ġh ence +ĠN AS +Ġoppos ed +unk nown +ĠDES C +ĠCh air +fa iled +ĠIN CLUDING +Ġwrit ers +{ }Ċ +ÃŃ t +_c opy +} : +ĠB at +Ġconvert ed +ed ing +pl acement +ĠH ost +S ound +и м +Ġs ought +m id +Ġsal ary +og g +âĦ ¢ +b ul +Ġw ir +valid ator +_ST AT +.st ore +ĠB attle +ı n +Ġ-- >ĊĊ +Tr ump +d ot +ĠCON T +.f etch +Ġcontin u +w as +Ġfra ud +_t mp +mit ter +.p ictureBox +G A +Ġt ournament +. Input +[ r +ex ion +cent age +ĠKore an +und ef +ĠAv ailable +resh ape +Ġk it +ĠStr uct +ĠS UB +An swer +_l ib +.t witter +Ġo re +ĠDr agon +.Ex t +, k +Ġexplan ation +ref s +ĠDr ive +ĠTr aining +.H as +int age +b ig +olog ist +enn is +Ù ĩ +Ġch icken +ĠĠĠĠĠĠĠĠĠĠ Ċ +ç Ľ +ãģ § +Ġpe ak +Ġdrink ing +Ġen code +ĠNE W +m alloc +ĉf printf +Ġ= ================================================================ +in cluding +Ġprincip les +ĠM ah +st orage +- key +Ġkey word +% ; +Ġtr ained +.con trib +Ġk v +__ ':Ċ +ĠB oy +param eter +Ġsu ite +Ġthous and +Ġco ordinate +-g enerated +íķ ĺ +gener ated +Ġad mitted +Ġp ussy +# w +Ġsw im +un ion +N a +ĠRoy al +.ch annel +Up dated +_RO OT +Ġv ital +ra ction +ĠCrush er +Ġpre ced +Ġhor izontal +Blue print +Ġattr s +Ġsm oke +Ð Ĵ +. Equals +F B +ĠRes ources +roll ing +Ġpass es +ĠN um +rot ate +et ype +\ ", +Ġsens itive +Ġt all +? âĢĿĊĊ +Pro xy +i y +_ section +âĢĶâĢĶ âĢĶâĢĶ +br id +Ġcirc uit +at an +EN C +Ġdr iven +Ġvot ed +Ġeduc ational +Ġinter action +abet es +Ġt one +ĠInitialize Component +Ġmer ely +Ġì ŀ +co okie +_ div +ĠUIL abel +vel y +} );čĊ +_ ENT +#+ #+ +art icles +ĠSou thern +Ġstrong er +ĠG iven +ĠE ric +ĠI R +ab stract +U nder +n able +Ġincre ment +ov en +Ġco in +_t imer +Ġsuffer ed +ĠF REE +'] ." +ĠQue en +st ats +Ġmeet ings +Ġenter ing +Ġalong side +(s ession +it als +Ġfound ation +ĠC redit +. div +_ ALL +pc ion +_st at +ick ing +Default s +_s rc +Ġoutput s +/ B +Ġent hus +-b l +.Fore Color +ĉ temp +F ace +Ġinter act +Ġwe ird +M ount +re ll +ud ents +Ġrequire ment +ĠS us +I ER +Ġe lected +re ference +ĠM E +Ġserv ers +.w ait +Ġsnap shot +il ton +Ġtri es +Ġt ipo +.T ime +> w +Ġmount ain +Ġp ounds +Ġ[ ... +ex ists +Ġng On +_M AP +Ġf lying +xi ety +ĉ value +_D B +un o +Ġse ats +T URN +. author +! ) +or ce +Ġindic ated +.s in +Ġass ignment +im iento +ĠF rame +_g en +in ery +_ ) +m essages +.set tings +ĠMe an +ĠM useum +ir q +att ach +ĠPalest in +_ QU +_t ags +Ġcas ual +em en +ASS WORD +$ s +ĠC irc +оР¹ +et ric +/ P +Ġep och +< head +_C MD +Ġg it +Ġpen alty +or ph +_ users +ours es +.Date Time +atern ion +_pro ject +Ġsuper ior +ĠD am +ĠSe attle +X Y +> The +ĠA k +Ġgr ass +/* čĊ +(d is +Ġgun s +Ġt b +ĠK evin +. args +ĠA h +op ed +( J +column s +arg uments +ĠWith Events +_f ull +ĠDef ense +S imple +Ġdeath s +Ġext ensive +ĠSt ill +ĠEx pression +ĠAg ency +Ġperform ing +F X +Ġus uario +U AL +S ide +od os +apt op +Ġcred entials +_c ap +at ient +ĠDis ney +Ġa i +Ġch ip +Ġvol t +.make Text +%%%%%%%% %%%%%%%% +Ġbelie f +_LO C +ĠC ivil +N avigation +Ġreve al +Ġviol ent +ĠF il +Ġc atalog +em ed +sc an +. control +Ġconstit ution +C ountry +Separ ator +_A PP +top ic +uet ooth +M IN +Ġdes criptor +y t +ET HER +Ġdistrib ute +' }Ċ +.tr im +.L ine +Ġl bl +assert Equals +ĠD et +omb ok +( width +Ġt ort +ĠEXP RESS +ac o +Us ing +ĠBr and +w all +EM ENT +ĠComm unic +< uint +ĠG UI +EG IN +ĠR ange +/ i +ĠT aylor +c ost +Ġrespond ed +ĠTh eme +n ce +IS H +Ġfeat uring +Return s +ĠK r +Ġ .Ċ +Ġn am +_c b +Test ing +Ġ{ }, +y al +.f ield +Ġ/ = +_SH ORT +m ates +Test Case +ain less +Ġeval uation +_ ITEM +ĠPac ific +ĉ k +Ġc ant +ĠR os +) s +Ġf et +STR ING +ĠDis pose +g al +ĠJ oin +ĠP orn +ĠCath olic +AR GET +cp u +ç łģ +.sc roll +IS ING +ifest yle +anc ement +Ġm erc +ĠB rowser +eter min +Ġover flow +Av ailable +Ġbott le +: UI +ific ial +Ġco ord +clar ation +Ġcon j +G LOBAL +ok u +Ġk wargs +cond itions +ul um +Ġg enu +ĠH ero +å İ +Ġun expected +ĠDAM AGES +Ġk a +ĠC ould +UP PORT +ĠPh otos +Ġconf ident +Ġdet ected +de g +rg b +Ġstrong ly +Ġ} ;čĊ +Ġ) : +Ġle ct +urs ive +RO L +ĠWe ight +Ġent ertainment +Ġ) );Ċ +Ġg onna +Ġb b +.d o +G S +Ġmist ake +D L +ĠPROVID ED +ear ning +L imit +iss ions +[ v +ä¸ į +ir ty +D el +Ġunder lying +pre ne +Ġj aw +ĠD I +pe er +Ġobject ive +Ġde posit +Ġk on +Ġes p +.set Visibility +/ login +< typename +Ġfr anch +/ e +Par allel +Ġsc ored +ĠH on +ĠV ill +ig a +Ġant icip +_ assert +ĠO pt +Ġdescri bes +w an +m ount +Ġmonitor ing +Ġt out +ëĬ Ķ +}, { +................ ................ += int +Ġc ust +---- -- +Ġatmos phere +P AR +ort e +IS IBLE +ĠI ron +ĠNot ification +.log ging +ĠBO OL +-p oint +Ġaf raid +ent a +Ġtom orrow +@ implementation +Ġeng age +ĠAn th +ĠF loor +ĠU l +To ols +Ġb ab +Ġcare ful +ãģ Ħ +Ġcruc ial +Ġcalcul ated +ĠS A +Ġw y +D X +_T AG +ind ed +Ġj et +ĠEngine ering +.M AX +en z +v d +Ġpublic ation +Ġ## # +Ġfac ed +ra ham +ĠC apt +As set +ĠCon stants +Ġlo ans +_ IP +ĠF ish +Red uc +_m at +Date Format +_m e +[] [] +Ġintegr ity +ĠC ourse +lob als +Ġfac ilit +Ġem br +ĠN g +.S ystem +Ġmanufact urers +Ġpro ven +.on Create +Ġal arm +Ġ § +Ġcomm only +ic os +æĸ ° +ĠSt ation +} ). +ĠF ilm +w i +ç ī +Ġeng aged +St ats +Ġgovern ments +Ġafford able +_p roperty +Ġag es +(' -- +Ġf ör +ĠProf essor +Ġhy dro +P ush +Ġorgan ized +Ac cept +é m +_c ell +Ġn b +p b +Art icle +Ġrem oval +Ġauth entication +ĠF R +l ide +Ġple asure +ap ol +Ġpart ition +ĠS ide +Ġcr imes +Ġdem o +hold ers +ĠPak istan +In struction +Ġexpect ations +.sc ene +Ġ' ) +h es +ino is +_P ro +Ġm olec +and al +_sh ort +Ġdefault s +Ġn ations +in en +Ġr t +O CK +P acket +S B +ĠSH ALL +_cont ents +ise conds +vert y +á t +G uid +n om +Ġcon clusion +. Update +Ġlo vely +Ġem it +b ec +ĉĉĉĉ Ġ +Ġintel lect +Ġb rew +ec ycle +F ire +Ġad mit +Ġar bit +Ġarr ang +ĠM IN +M ail +ĠN ative +C ur +Ġcon vent +.R untime +" }Ċ +.R un +Ġprint ed +Ġconven ient +. ar +m ock +ĠAdmin istration +ãģ ¾ +Ġelect ron +fl ate +Ġl ombok +Ġjava fx +n h +Ġsup plies +Ġvisit ing +ah l +Ġpow der +Ġult imate +Ġorient ation +ut as +_s cale +Con firm +ph ones +ĠOper ation +/ T +_IN TER +Ġair port +Ġmet rics +Ġphen omen +a udio +Ġm ai +( K +h u +all ing +rodu ction +ĠTrans port +ĠNOT E +æĸ ĩ +Ġfew er +_T IM +ì § +к и +A ge +F IN +Ġì Ŀ +ĠAt tribute +group s +er k +at to +. define +.AspNet Core +ategor ia +ĠS ir +( form +< User +. round +_d ay +.A ll +Servlet Response +.N o +l arge +IG H +qu ent +Ġvir us +Ġret ro +Ġim per +Bit map +Ġv ice +Ġoff ense +ist e +ĠA UTH +Ġê ° +ToolStrip MenuItem +G u +Ġr ape +ĠDav is +Ġover whel +: flutter +- table +ĠCon structor +Pr ivate +e ven +ch r +Ġap plies +_at tribute +Ġcon tribute +E VER +L ines +ĠAf ghan +Vis itor +ĠS L +se ason +C U +Ġintrodu ction +Ġmat plotlib +Å ij +Ġnewsp aper +âĢĶ and +< tag +Ġin i +Ġd iverse +Ignore Case +ĠU r +Ag ent +Ġb ull +.em it +( Exception +ar Layout +Ġincred ibly +ĠTr ust +={ ( +- nav +Ġe quals +Ġl ady +ĠP od +d isc +al am +ĠI V +â Ļ +iv idual +ph i +add ed +Ġdifficult y +Ġcomp act +ĠAction Result +c ers +_class es +Non Null +Ġqu it +Ġp ou +S witch +ir s +- test +ĠK ind +ĠCal endar +Ġstream ing +} ', +S W +Ġst ead +oc a +Ġprov ince +Ġcol span +Ġperson nel +ĠE mployee +Ġprodu cer +Ġevery where +od b +Ð Ł +bs olute +act ivate +Ġgr inding +ĠBuild ing +ĠSand ers +(s c +ĠOff set +//////// //// +} ;čĊčĊ +({ " +Ġscan f +ĠY Y +ĉdef er +Ġj ew +Ġrestrict ions +.m p +[ l +ä¸ ĭ +label s +red icate +aw esome +Ġw aves +Ġcon front +Ġmeas ured +Ġdat as +_ex it +ot ton +Ġshould er +ask a ++ # +ĠĠĠĠĠĠĠĠĊ ĠĠĠĠĠĠĠĠĊ +Ġtro ops +ĠU nd +_c ard +w ich +Ġn ous +Ġ"/ " +s b +Ġcommunic ations +Ex port +Ġdec ode +th s +inter pret +By Name +ĠSp irit +ed ges +O LE +ĠE M +t it +ĠTh rough +Ġb io +ĠP ackage +or ne +Ġ} . +` ;Ċ +Ġok ay +ĠZe aland +ident ity +(n ext +ĠB ang +Lib rary +Ġheav ily +il on +Ġdi pl +Ġrot ate +put s +) ',Ċ +ĠData Table +Ġmay or +.to LowerCase +Ġsome how +ĠNor thern +al c +Ġcap abilities +Ġv ibr ++ Ċ +ĠS u +ĠRes et +_m ean +Ġc ig +.cl oud +ĠB and +ĠF actory +ĠAr izona +_ io +op her +Ġconsc ious +Ġà ¶ +\ Controllers +_s peed +ĠF ac +_C om +ĠB ible +w en +ED IT +Ġun n +ĠSt aff +ĠIn n +Ġmechan ism +ĠM embers +Ġmigration Builder +'] .' +.get Int +< void +ĉf ree +oid s +\ Support +Ġautom atic +Ġch ances +Ð ¶ +Ġcomp licated +[ row +ah oo +Ġ}ĊĊ ĊĊ +Model s +W in +Ġt ape +ir us +iz on +on omy +(" _ +: . +.st ereotype +( env +_re ct +(w ith +Ġassert That +Ġcon straints +put y +E mployee +T D +Ġgu itar +ĠJew s +.pro cess +Ġf iction +ĠSh ared +âĶĢ âĶĢ +Ġprop ag +.N et +Ġachie ved +ĉ Q +Ġn urs +Sh ared +_FAIL URE +Ġbeh aviour +Ġcol s +ism o +Ġfem in +Ġchalleng ing +Ġpost ing +enc il +Ġcapt ured +ĠD ou +( word +ĠTur key +pan ies +Ġre putation +ORM AL +Ġelig ible +prot ocol +id as +(f rom +Ġfin ance +- per +Ġg otten +H A +d uration +ĠP arent +Ġin vent +Ġre start +ол ÑĮ +r ition +(r s +< bool +i ert +Ġmod ification +ĠT X +readcr umb +b ank +$ / +ĠMill er +] ),Ċ +.Check ed +Ġsac r +se curity +Ġp ose +ĠBr ad +Ġfit ness +Ġannounc ement +ation Token +Ġserv es +ne ed +Ġge ometry +AR S +æ Ģ +andid ate +Ġs prite +_s plit +We ek +ad ies +> (Ċ +?> " +Ġ/// Ċ +Ġein er +Ġweek ly +ĉlog ger +_p op +_m an +Ġmigr ations +Ġask s +Ġb s +Ġfall s +.W here +- height +_fe ature +.M in +Ġhy per +Ġvol atile +Ġtw enty +Typ ography +Un able +D et +, f +-m od +Ġsett lement +Ġcontract s +n ome +B ad +ĠB rian +(user name +!! !! +Ġh ack +.F ield +H R +ĠJ ordan +iz a +Ġ ł +ĠSh er +. header +( other +ĠD ub +( op +ĠR ound +Ġv ie +Ġap pl +ĉ J +ĠIn sert +ĠL P +reg on +ĠM PI +Ġan chor +ac a +ø r +Ġa de +anch or +que e +ĠTree Node +Ġtarget ed +Ġla id +AB EL +v et +ĠOr igin +A nt +. ');Ċ +ex pect +ed Reader +ĠM ajor +Ġin ch +Com par +Ġpre view +Ġill ness +ĠCONTR ACT +ĠInd epend +u uid +Ġn ome +Ġt c +ĠA venue +is an +Ġph rase +_m ove +") [ +Ġprov ision +Ġconcent r +_ IR +ĠU t +() + +Ġn as +! , +ĠRob in +i ations +at itude +Ġp x +ĠWith out +/b ash +ek t +re ement +Ob server +ĠReg ion +UBL IC +Ġ{ // +K N +å · +Game Object +å ¾ +enc oding +Ġ** * +project s +Ġt k +Ġche ese +EM PL +ar o +Ġا ÙĦ +Ġcons ists +ref resh +ure au +ĠSc anner +Ġso il +Ġfl avor +Data Source +Ex ecute +ени е +Ġsh it +åĪ Ĩ +< any +Ġretrie ve +Ġbelong s +.st rip +abs olute +Ġexp anded +bo y +): - +Ġresc ue +.J Label +Ġre ly +Ġal ignment +-f amily +Ġre nd +OLUM N +Ġb orrow +Ġqu otes +ĠL ew +Ġsh ower +ĠDE LETE +_lo op +! "ĊĊ +ĉ re +Ġattempt ed +aver age +ĠP aint +quis ition +ol en +Ġliter ature +ĠRe ference +_TEXT URE +ĠS eg +ĠInd ust +ct ype +D UCT +_H OST +ĠTr ade +Ġpl ugins +Ġbre ast +ul se +Ġcreat ure +ãģ Ļ +ĠW i +Ġsup plied +c oll +! (" +Ġfuck ing +ĠCh rome +ĠU ri +ĠN ation +Ġvert ices +T HE +ĠOr iginal +on de +Ġsh arp +Ġcook ing +Ġ{ /* +ĠPs ych +ĠH ollywood +=$ _ +.D ock +Ġg er +Ġb one +_con n +_se c +ys ics +Ġ= " +S al +s f +Ġdeep ly +ang les +T erm +b ell +ĠQu ick +ener ation +adio Button +åħ ¥ +}čĊčĊ čĊ +Ġcapt ion +l c +ĠE L +, [ +ĠĠĠĠĠĠ čĊ +ret t +(m ethod +ĠFl ash +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +W ISE +.s cale +Ġrough ly +_ child +m emory +ay ing +Ġinitial ized +in ator +а ÑĢ +Ġsc alar +ĠH o +ai res +(c olumn +.de stroy +P ACK +Ġh em +ang el +_S UB +. qu +Ġ × +DE FAULT +pos itories +ĠL ength +ĠF ast +Ġsign als +Ġ// $ +ri ers +Ġd ummy +AN Y +Ġperson ality +Ġa gricult +Pl atform +ER O +ĠT ra +Ġen orm +ĉ W +Action Result +Ġa ver +[ str +Ġ' -- +.S printf +Ġdeb ut +Ġ Ñĩ +h ex +_ utils +Ġp b +U ITableView +Ġz ur +. encode +Ġv ag +.error s +о н +Ġm r +ĠA ward +Ġc pu +Ġpress ed +' est +ĠF estival +' T +Ġa k +res olve +.m e +Ġn ic +Ġgen re +Ġat trib +ĠMo on +Ġarr ive +ĠD ating +Ġt m +.Config uration +. red +Ġgl m +Ġst ations +sw itch +Ġt ied +äº º +Ġ/ >Ċ +Ġsubsequ ent +pos able +-fl uid +Ġth orough +Ġpublic ly +apt ers +ĠWil son +_P RE +y ard +ä ¼ +ĉ in +Ġre vers +Ġbul let +cri bed +nes ota +Ġ($ _ +ann on +c ursor +Ġclo thing +ĠM ulti +: ', +Ġv ess +ordin ator +Ġein em +C annot +Ġar med +ĉ V +ä¸ Ĭ +.F lat +ĠS ep +ĠSub ject +_f ont +Ġcharacter istics +D one +el n +######## #### +PO S +Ġd ensity +ĠPl atform +- items +Ġo vers +Ġpush ing +ç ¤ +.Con nection +_ term +Ġinitial ization +________________ ________________ +ç ¬ +.d ocument +les h +ĉd ocument +ĠP in +ç a +Ġdefinition s +.P ath +_W RITE +Ġ ĉĊ +? >ĊĊ +Ġter rible +be an +ick ets +ĠS V +B uy +(t ask +Ġreg ime +g oogle +Ġcr ack +.vis it +N UM +ener gy +Ġstr uck +_s ample +.p ayload +Ġre vis +ĠSc ene +Ġp g +Ġbreak fast +URRE NT +.char At +_ex ception +ĠAnt on +Ġguid elines +Ġex haust +ĠFin ancial +Ġind ent +Ġdes ktop +H idden +F ailure +Ġpr inciple +Ġ iv +Ġse ks +n etwork +Ġnumber Of +ĠAl bert +ĉ long +, . +Ġz eros +f ade +ĠT yp +ĠT erm +ĠAr ts +.App lication +Ġbeh alf +æĪ · +Ġm ere +(` ${ +Ġaware ness +elp ers +f lix +Ġwe igh +Ġestim ates +. child +/ O +ĠBit map +.b ottom +Ġ************************************************************************ ** +Ex pect +ent o +ĠFor um +ver al +Ġj ail +Ġab ilities +ĠH OLD +ĠC it +Ġd ynam +Ġgr ay +ĉĉĉĉĉĉĉĉ ĉĉĉĉĉ +.next Int +ant ly +ĠAR ISING +( private +Ġreject ed +ĠN ic +Ġle ather += {Ċ +aly tics +th etic +.T op +.P age +={ ` +Ġ ;čĊ +de pth +m ann +W D +ĠS om +.R ight +Ġ) }Ċ +Ġtr ait +Ã Ĺ +i ac +Ġr v +S ample +.X ml +opp ed +ĠÑ Ħ +list s +Ġt ear +ivers ary +.c ollection +ĠCon stitution +ĠHttp Response +Ġbr ill +ĠP rom +h over +ĠM iami +Ġarg ue +_f loat +Ġ ãĤ +Ġn at +ĠT al +Ġinteg ration +(c ur +Ġrem oving +Ġco eff +ĠTh ough +Ġfore cast +ĠV egas +S ite +Ġtr ab +ĠHen ry +- i +Ġinvol ves +B T +Ġs lo +In voke +Ġl ucky +r at +Ġ? Ċ +Ġhand led +(f d +cont ents +ĠO FF +R F +Ġst y +ĠM otor +ter y +t ax +M AP +ĠMr s +Ġph ones +ĠUI View +")) );Ċ +( dev +ĠIr ish +Ġw s +D I +_OFF SET +ĠEvent s +Ġst ages +Ġ} // +Ġhab en +ST ANCE +ĠS in +ĠM oney +(t op +Ġappoint ment +VER SION +met adata +_com ment +Ġcolle agues +map s +â ĺ +Ċ ĉĊ +( al +_re q +Ġf ut +Ġarchitect ure +ĠWH ETHER +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +_s creen +Ġstyle Urls +Ġmon ster +. up +ph ia +Ġprocess or +ĠT err += ', +ĠMan ufact +ĠN T +k el +ib ern +ĉf ile +A li +rient ation +Ġ// ! +ap ore +ane ous +ĠC reat +f older +Ġh ay +Sup press +( left +Ġe uro +Ġdis claimer +ustr y +sh ips +_f d +ĠF a +_in sert +Ġro l +if ting +ĠCom ments +_b r +Ġloss es +ĠAdd ed +ch arg +Ġп о +_s ystem +ĠS ometimes +ĠSp ain +(g roup +ial is +Ġdoll ar +ĠAr gs +qu ires +ĠT en +.s css +Ġsurv ive +us age +Ġj un +im iter +ï¼ģ ĊĊ +Ġfif th +t oggle +Ġdecl ine +($ " +(L ong +ing e +Ġpil ot +-l ight +-r adius +Ġpod cast +Ġnatur ally +P ages +ä¸ º +ĠDes pite +Ġlight ing +Ġcr ate +ĠB inary +Ġredu cing +Ġe leg +ĠM ouse +ĠTest Bed +Ġbefore Each +_ ARRAY +Red irect +Ġf lood +Ġsh ips +Ġelectric ity +)* ( +ê ¸ +ĠV iet +her o +Ġd ia +ĠK ent +he art +Ġthreat s +_ acc +Ġs ymbols +is chen +_in st +C riterion +ĠT IM +. Height +Ġ âĢĻ +();ĊĊ Ċ +Product s +_S P +ĠC y +Ġdepend ent +est e +Ġdat os +d it +аР² +IGN AL +Ġless on +"> ' +ĠC over +ĠH ope +ĠT imer +Ġd ad +vid ers +ĠPh ot +/ ? +rop y +om ing +as ion +Ġ\ ( +ĠE T +ĠRe ading +Ġep isodes +l m +ech a +Ġne uro +Ġhar mon +Ġlib eral +- ind +D ATA +Ġevery day +Ġdiv ided +ĠActive Record +fig ure +U A +ä ¹ +riend ly +te ch +.game Object +иÑĤ ÑĮ +Ġmo on +ft ime +Ġno ch +ĠT ORT +ĠV M +.in itial +( child +Ġmus ical +Ġo c +b as +ĠH ay +_l ong +Ġmem set +ile y +adel phia +S V +ro at +_t x +Ġl on +ĠngOn Init +b p +ĠGold en +AC HE +Ġwor ried +az i +E ar +T ake +(f p +bur gh +_ Data +g res +ĠO nt +p us +Ġtrans parent +Ġp ocket +Ġr am +igr ations +. čĊčĊ +Ġ[ ( +Ġadopt ed +Ġreported ly +ĠD ream +Ġ} ));Ċ +los ing +Ġte eth +ĠBook s +", & +enn y +LE MENT +Ġg el +ĠPl ant +! âĢĿ +.h ost +ĠRep ly +re ngth +Ġrecogn ition +Ġ}} >Ċ +L A +Ġmir ror +Ġassist ant +( device +Ġspirit ual +b uilder + § +Ġou tr +Ġt t +ĠP ER +Ġrad ical +Method s +Ġp ace +ud y +Ġg ut +ĠG reek +Ġnon atomic +ĠP aper +_G PIO +Ġob st +.A d +viron ments +ĠS ov +( con +ĠTrans action +. assign +ĉc atch +el ter +Ġbit coin +_G R +ĠčĊ +met ic +Ġtrans formation +åı · +Ġr gb +istrib utions +Ġimp licit +/ in +dest ination +аÑĤ ÑĮ +Z ero +Ġun set +. where +.g o +Ġform ation +Ġdeclar ation +() čĊčĊ +ĠEx pl +ĉĉĉ ĠĠ +/ pro +.J SON +Ġdes k +.sub str +//---------------------------------------------------------------- ------------ +ly n +p son +dis able +ĠF unc +ĉ Assert +ĠM ARK +Ġdefe at +Ġbl ind +Ġconst ants +. headers +UIL D +Ġexp enses +P ixel +Ġh r +Ġf el +ĠEast ern +_d el +ĠC ub +Ġs q +ĉc ount +ĠD irectory +Ġex clus +Ġhistor ic +Ġ ------------------------------------------------ +Ġcom position +Ġdata GridView +ĠB urn +ĠB C +M aster +Ġsp awn +Ġbe aring +.Set Active +il o +Ġg allery +Ġfound ed +Ġav ailability +.s qrt +Ġp es +ĠD OM +m ate +O ct +Ġmatch ed +it ivity +Ġan xiety +.pr ice +ĠIn stant +ì Ĭ +Ġt ut +IC ollection +.sh ared +_s ql +t bl +lib rary +_de stroy +erm al +ĠNot es +ĠE in +Ġsou thern +ĠOTHER WISE +Ġmac ro +.l ower +cl s +Content View +.l ink +const ant +ĠB es +Ġsome body +n b +"> { +( local +.. ... +ĠN ull +m x +Ġà § +Ġp ause +-------- --- +_M O +ĠC M +Ġfor Key +ĠD VD +Ġclose st +_DE VICE +ĠSte phen +ĠB BC +ĠTr avel +P aint +ĠResult s +ĠR ule +Ġt p +Ġrat ings +c in +c sv +> / +ĠG OP +l ad +Ġ ÑĢ +Ġindex Path +m atrix += f +ars ed +Ġ} ); +ĠC os +ĠS core +Ġt ak +ĠE SP +ĠIN C +_N ULL +-f lex +"] [ +int o +el and +Author ization +_F ALSE +Ġg ate +Ġv id +ist ent +T IME +Ġre write +Ġt ie +Ġarch ive +.event s +.get Parameter +ĠPer mission +Ġprogram me +Ġ é +j ud +Ġcam eras +(s ys +ĠSy rian +Ġimpro vements +Ġh ip +Ġsu icide +Ġsch olar +Ġcompat ible +rem ote +.d own +F UNCTION +Ġman aging +ĠUI Kit +. raw +>> >> +Ġdem ands +ell ite +Ġd ent +ĠM icro +åı ĸ +'] [$ +ĠI E +im ension +Ġt rem +Ġg ained +.w ith +. ok +h ou +Ġb om +amp aign +Ġjoin ing +f ish +Ġadd Subview +Ġnor thern +.c or +ore t +D ie +in ish +_com p +Ġatt ended +Ġcoll apse +ĠS S +ac ent +_E QUAL +ĠDe ep +R GB +ĉ test +ol ves +us et +Un ityEngine +w riter +Res olver +, % +if ference +_re move +ond a +Ġfem me +de code +Br anch +Ġfl ush +Ġinnov ative +Test s +Ġ[' ./ +Ġcover ing +. admin +ultip art +(l ambda + namespace +ĠS port +Ġ! ( +ac les +Ġde pression +ĠK ong +Ġp ert +ĠCon n +ĠOther wise +/ home +s upported +Ġp ink +Ġinv ited +ñ os +_en abled +Ġ- Ċ +F W +en ers +ĠM Y +Ġsuggest ions +Can vas +Ġf er +ĠMarket ing +@ Test +unt u +ĠV en +ĠC ou +iv als +Don ald +lim ited +ĉĉĉĉĉĉ Ċ +Ġanal yst +( entry +Ġrepresent ative +_at tributes +Ġf ur +.h ide +res p +ado res +rid es +ĠJ osh +ro bot +ĠN AT +Ġs esso +Ġintegr ated +: true +part s +Ġst upid +: event +@end section +Ġp u +.T able +ĠY ii +` ;ĊĊ +Ġcl ang +=" "> +eng an +_param eters +.int ernal +ĠMod ern +Ġmet ric +Ġsem i +={ {Ċ +.am azon +ĠB B +aint y +view port +Ġstart Activity +dis patch +**** * +Ġfl av +iffer ent +[ this +Ġst ake +Ġarg ued +vious ly +.w ork +ĠO ak +O ld +( async +not es +Ġfl ip +Ġdis ag +ĠT E +ĉ error +< ' +Ġ» ĊĊ +Ġfilter ed +ĠM ach +Ġh ung +_d ump +_s amples +-dis miss +Ġr ay +Im plemented +D K +Ġj ed +Ġbreak s +Ġf its +. gr +ĠZ ero +or o +Ġequ ally +Ġ' [ +Ġconcern ing +< meta +play ers +_P OS +_s im +J an +Ġyour s +ĉ N +Ġsp ir +Ġch ampion +ĠAn alysis +ap a +ĠNS Log +_l ines +ñ a +ĉĉ ĠĠĠĠĠĠĠ +.S c +Re p +etro it +ur able +M IT +com pat +own ed +_ind ices +], čĊ +Ġdis covery +ĠDie go +ob i +. Index +Ġtrend s +PL AY +.n o +Ġl ens +_c fg +Ġan no +ag an +Ġperiod s +ter ms +y z +Ġattack ed +ib ration +PEC IAL +_ grad +Ġaccord ance +.Read Line +.de vice +ri x +. container +m ay +erc ise +ĠL u +Ġr g +ĠÑģ ÑĤ +ĉĉĊ ĉĉĊ +( un +TERN AL +Ġless ons +Ġalleg ations +Ġtrans mission +.Re f +M obile +ĠT ournament +ĠN ut +ĠG a +ĠCap ital +def inition +- exp +c lean +Ġfant asy +Ġenh ance +ent ence +'] :Ċ +ack ets +Ġcelebr ate +@ ", +Serialize Field +Ġarray s +t b +ĉ st +[ assembly +( reg +.c ategory +Ġimpro ving +Ġsal ope +Byte Array +Or iginal +Ġ[ {Ċ +åĽ ŀ +ĠCl in +oen ix +ĠS amsung +Ġmaint ained +Ġag enda +f ail +Ġpres ents +Ġtim ing +.m ark +' >< +Ġprom ot +Ġin cl +_ only +ë¥ ¼ +ĠAtt orney +- date +Ġlands cape +Ġf u +S Y +.p rop +ĠA rr +p ag +Parallel Group +': čĊ +Ġlog s +a unch +unc i +n ama +Table Cell +iss ues +. { +ec urity +_ex ec +old s +Ġhost s +Ġpro to +_ import +_s ort +ĠB ow +ĠN ormal +ĠF arm +.create ParallelGroup +R otation +. err +Ġp leased +it age +.W h +ĉĉ ĠĠĠĠ +M R +ĠM ORE +ĠN atural +_ transform +B ASE +ener al +ut down +.common s +W T +Ġa an +. Result +d og +Ġclick ing +), ĊĊ +# line +Oper ator +Ġc iv +Ġm erg +ob uf +ng then +Ġ[ { +Ġcan cell +tr igger +. : +W ORK +decl are +Ġdecre ase +ÅĽ ci +lo om +.N one +ĠM I +ĠJ ason +Ġhealth care +iam ond +s ylvania +* x +ĠR a +[ b +Ġprint ing +ph abet +ĠLab our +op per +Ġz ijn +-t arget +_F UNCTION +Ġo ct +ени Ñı +åľ ¨ +Ġwest ern +Ġcomput ers +ĠR ET +Hash Map +[ String +get Value +_D ATE +.N ext +ĠF if +é l +ick ed +æ İ +-M M +Ġ{ ĊĊĊ +Ġcontact s +Ġdig its +Pro du +Ġunus ual +Ġrapid ly +t ures +Ġang ry +c ancel +xx xx +_p arser +id ity +_P REFIX +Ġme hr +Ġrare ly +et he +op es +Ġ% . +work s +Ġthe ta +Ġcontrib ution +ĠT ony +Ġsqu ad +аР¹ +Ġî n +th ere +out ed +ĉ q +Ļ Ĥ +g ood +L I +é¡ µ +ĠL iving +iz abeth +Ġk t +ĠD allas +] ],Ċ +Ġ/ >ĊĊ +Ġrais ing +/r outer +_g ame +ĠC UR +z ens +. es +Ġfont Weight +(f unc +not ification +Ġ'../../ ../ +Ġbl ame +ãĢĤ ĊĊĊĊ +an co +Id entity +f ollow +Ġart s +x s +Ġofficial ly +ĠSt udio +Ġrecommend ations +Ġloc ale +Ġam ateur +ĠEn able +Ġcap s +. End +- add +_g shared +ĠC T +For ce +Ċ ĠĠĠĠĠĠĠĠĠĠĠĠĊ +Ġor ange +Ġl p +Ġanswer ed +.G rid +Ġd ual +Ġstrateg ic +Ġnob ody +Ġf atal +_ est +( el +Ġì ł +ĠB udd +A IT +_f actor +- one +ĠH AVE +" čĊčĊ +Pro f +Ġä r +str ings +Ġdir ty +ĠF ace +ĠB egin +ĠB us +Ġw is +åŃ Ĺ +Ġspe aker +Ġcar rier +ĠO m +Ġhad n +All ow +:: __ +Ġver b +ĠCom plete +ĠE asy +Ġb ills +ĠĠ ĊĊ +Vert ical +Ġpr on +ĠDef ine +Ġlook up +variable s +Ġpand as +um es +Ġinn oc +Ġset Up +ĠCh ampionship +art ist +ĠC Type +F oundation +à¹ Ī +ĠSet up +Ġrec ipes +ĠU IColor +ĠF ight +Ġauthor ized +_c lick +_s uccess +ang an +ĠMount ain +ĠDo ctor +Ġeg g +ĠMedic ine +c les +` .Ċ +[ int +d ashboard +ĠApp ro +-d r +Ġprodu ces +Ġrent al +Ġre load +Ġarr ival +sp ot +Ġund ert +Ġequ ipped +Ġpro ved +Ġcent ers +Ġdef ines +al so +Ġop acity +ĠUn fortunately +ĠIll inois +Ġн е +ĠTem ple +ĠTr ail +ĠK elly +Ġmeasure ment +Ġsepar ated +-c ircle +H ey +ĠRE AD +ig its +Ġ ib +ĠM OD +atter y +аР· +Ġv end +ен ÑĤ +ĠHttp Client +s afe +_A SS +ic it +ĠCon struct +ĠC lo +ĠS ix +_T OKEN +(b lock +Ġwarn ed +/* ! +! Ċ +Ġinnov ation +_ " +Ġ );čĊčĊ +Ġsp ots +Ġcho osing +.c s +Ġflex ible +U Int +Ġscr atch +- al +Ġf estival +Ġout standing +================================ ================ +M ean +ĠO regon +s ymbol +. account +d ney +'' ' +! ", +Ġpart icle +à ĥ +[ MAX +IV ER +ER ENCE +NS Mutable +ĠColum bia +_ ĊĊ +.f r +Ġc ogn +V R +ĠMethod s +ĠM ade +ĠB R +ĠEl se +Ġeg gs +Ġsw ing +ĠIn v +Ġdise ases +Ġf irms +Ġle mma +}` );Ċ +l ings +Ġg ym +umin um +.T rim +M em +Ġcritic ism +ibern ate +_T X +ion i +Ġguid ance +Ġrepeated ly +Ġsup plier +Ġpaint ing +.F ragment +ed Exception +Ġw iring +Ġcour ts +W EB +æľ ī +\ . +ill ance +Ġb rows +ĠP attern +PL ICATION +ĠSum mer +Ch ain +Ġc ute +mer cial +Ġd il +ĠFrank lin +ĉg lobal +IN CLUDING +h istory +Ġl st +Q t +SD L +al ia +i ere +( ... +ĉc in +iff s +vel ope +ĠR oot +cl uster +User Name +ign e +< S +Ġf est +Ġindic ating +ke eper +Ġc ada +é g +cons in +ĠG B +Ġl b +em ony +-icon s +_d oc +Act or +e lem +.De lete +Ġin fection +ĠPriv acy +Ġgreat ly +ĠP os +ĠT reat +Fl ow +Ġattract ive +ĠMar c +s udo +tes y +- an +ab ama +ĠW ould +Ġsu ck +index Path +ĠE t +T imes +Ġclub s +_ass oc +Ġac quired +(" : +Ġint ense +.m aps +Ex pected +T oggle +Ġa y +Ġl ifestyle +-c alled +ĠS now +V olume +Ġcann abis +ĠD irection +ĠLim ited +-s pecific +Ġd owntown +/ icons +Ġre ven +L eg += null +Key board +') ). +Ġ"" ;čĊ +Ġatt itude +.n avigate +- error +AM PLE +ĠJ ay +v r +c ow +.com pile +Ġmem ories +_m ark +ĠMin nesota +Ġk osten +Ġprob ability +w arning +Ġgen etic +F ixture +ĠHash Set +N ombre +_m onth +Æ ° +- start +xy gen +ĉ ft +i agnostics +ĠMat thew +Ġconcept s +Ġcon str +. State +и н +N ov +Î ± +ĠP anel +ä¸ ª +com pare +> ()Ċ +Ġapply ing +Ġprom ised +Ġo x +nc ia +ĠValid ation +ort s +_c ur +e lect +ey e +( Data +Ġreport er +ĠB uff +Ġs r +Ġ" ; +ick y +Ġtemp or +S N +Ġres ident +pi res +ys ical +Ġend orse +ĠS ong +is Empty +le et +_ util +Ġdist ingu +ĠT alk +ĠM ot +( default +.A rg +gorith ms +_ words +im mer +_res et +f amily +W W +Ġsav ings +ĠâĢ Ŀ +_en able +side bar +Run ning +Ġal i +Ġtest im +Ġwarn ings +ĠCh em +ĠEx it +Ġfound er +pect or +Ġr m +_d ataset +ĠD as +Ġh an +Get ty +á l +Ġn y +Ġpo verty +Ġresult ed +.b y +ĠVis it +Ġobt aining +/ '.$ +ĠĠĠĠĠĠĠĠĠĠĠ Ċ +sh all +_LE FT +UI Image +_ Name +h ave +ĠN ob +l r +- footer +Ġn aked +ĠG arden +\F acades +Ġgrad uate +Ġfranch ise +pl ane +Ġcontrib utions +Ġstring With +Ġc rypto +Ġmov ements +ath ers +Ġlif etime +Ġcommunic ate +j ar +ĠFr agment +_ IF +ĠN avy +ĠF igure +Ġsim ulation +_st op +Ġreport ers +Ġvers us +aj a +ĠÎ ± +Ġgovern or +List Item +Ġse aled +.Back ground +ed i +ash ing +Ġl ip +ĠI h +mer ge +Ġn ec +el ocity +ATE G +Ġse eds +Ġflo ating +_F A +w alk +ĉ user +_de pth +Ġw age +@ app +N il +( [" +( vector +Ġsecret ary +Ġj Panel +ve z +³³ ³³ +d irection +ĠE P +Ġh unt +Json Property +ĠP ORT +] ", +аР¿ +ĠFore ign +pan ic +Ġtri als +ĠA le +Ġr ural +- value +author ized +ĠScot land +.d rop +ĠM T +ç ± +row th +File Path +Ġrec all +if le +Ġc el +ĠSE LECT +k n +_c ase +Ġc rop +s ure +p ot +IC S +Ġst em +Ġindust ries +P ut +Ġa ber +road cast +Icon s +) ")Ċ +æĪIJ åĬŁ +g ui +Ġassum ed +Ġr x +E A +è § +EL L +Ġdo se +Ġin e +Ġde eper +l ider +Ġord inary +Ġg olf +_IM AGE +ĠN AME +(m odule +Ġat om +Ġbel t +Ġoff ices +b eta +Ġphilosoph y +( JSON +-f ield +Ġintrodu ce +Ġconven ience +opt im +> "Ċ +ath y +Ġemploy er +qu ate +Ġed ited +Arg uments +ĠN ations +__ ) +Ġno se +ĠS ample +' )ĊĊĊ +Ġc ake +.get Attribute +H D +Mod ified +Ġpredict ed +Å Ħ +an ie +S orry +(d oc +w ind +ie ve +Ġprov isions +AT ER +OT E +M Y +.A utowired +ĠB ath +. Boolean +Ġback end +.M ouse +ater al +p aper +Con st +ĠV R +_ entity +_C TRL +ĠProte ction +ĠG M +ĠStud y +Ġsou p +ot ime +' use +] " +/ users +a ug +ĠH ong +_n orm +ãģ ¨ +Ġse cre +(B uild +ĠCon tract +ol as +Ġsa uce +Ġaggress ive +Ġrac ial +char acter +@ @ +Ġcomp ile +ĠV oid +_re m +_m emory +k k +Ġm ic +S ame +U tility +ĠH tml +ĠX ml +Read y +Ġg all +Ġalleged ly +ĉĉĉĉ ĠĠĠ +ĠMet al +ĠPerson al +Ġborder Radius +rx js +object s +Ġwant ing +Ġb owl +v endor +offset of +ĠR s +ĠR ating +Ġr ally +_N ODE +ĠM ix +Ġadvert is +Ġnarr ative +s al +Ġm c +SE rror +Ġf ingers +Ġaccom pany +Ġt ired +Ġstr ide +Ġgu i +el ist +Loc ale +Ġrele ases +ik ing +Ġan ger +)) )ĊĊ +alle st +Sum mary +( O +(f or +Ġbasket ball +Ġroad s +ĠInst all +ĠF ab +it map +Ġ) )Ċ +Ġinter section +ighb or +ĠB ry +ĠHER E +So ftware +elf are +ac s +Ġtrail er +.get Class +ch ars +Ġreg ulation +Ġref ers +Ġde struction +Ġcontin uous +ĠAust in +é ¢ +ak an +.w indow +ĠTem plates +Ġabs ence +: n +Ġdis order +fl ash +Ġde let +bo ards +ĠĠ ĉ +RO P +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġac qu +Ġlaws uit +ĠRe views +Ġgar age +t imer +Ġe j +ĠRect angle +Ġflow ers +il st +ĠIn stance +S uper +d et +dis posing +ĠE S +ĠI C +ver e +S k +_ch annels +put ed +/ null +nn en +ĠG allery +_g lobal +Auth entication +ĠR ank +Ġblock ed +Ġcal m +mark et +ĉ val +Ġa ug +per iod +ĠCon stant +Ġ?> ">Ċ +Ġl obby +p al +Ġs ink +ia h +Ð ¡ +urn ame +Ġcon ver +Ġinvestig ate +Ch rist +H ub +ĠIN D +ĠP ed +ur as +ĉ url +ĠT ro +Ġpre ferences +Ġguarante ed +` ĊĊ +Ġport ions +Ġeval u +' > ;ĊĊ +.AutoScale Mode +Ġc ats +Ġreg istry +ul us +F I +p ayload +- search +Ġstay ing +ac ious +Dec oration +Re view +In f +Ke ep +it is +, String +Co ord +Ġper o +S ex +ĠAtl anta +uest a +Arg b +> * +} _ +F ooter +Ġemploy ed +_b ound +v ide +.f unc +$ scope +Ġsp o +ĠAn al +ounc ed +ar ound +Ġrestr iction +Ġsh ops +å Ģ +ĠLat in +-c ol +Ġbare ly +ĠE uro +E r +Ġfa ire +_d istance +_un lock +Qu ote +IV ATE +Ġå Ī +Ġaim ed +ĠRet rie +. iter +Ġwr apped +Ġagre ements +str ument +( product +Ġstud ied +.set Value +Ġy e +ĠC ache +MB OL +Ġquarter back +Ġsy ntax +.getElements By +.v ersion +we bsite +Run ner +_s ingle +at iv +ĠAl tern +ĠBeaut iful +right arrow +Ġd iversity +pl ash +( co +.F ill +Ġtyp ing +Ġcl ar +H it +O O +ac co +w orth +Ġscript s +ĠMuslim s +ĠL L +erv ing +( boolean +Ġbase ball +ĠC AN +MA IL +de pend +Ġrespect ive +Ġconst expr +.* ;ĊĊ +'] ))Ċ +Ġy ard +Ġident ical +if ecycle +US H +up iter +. validate +cl i +IST ER +Ind icator +F ail +Ġdemocr acy +. var +Ġsatisf ied +------------ - +enc er +h or +Ġr ounds +DA O +o a +Ġfl ask += c +[ ]Ċ +/d ist +Ġpart e +Ġconfirm ation +er on +aw are + +Ġdepend encies +ĠV ideos +- row +Ġ** /Ċ +Ġn ou +Ġh over +æ ŀ +Ġn in +ĠUS D +M ac +_L oad +Ġout comes +_s ocket +Ġqu eries +w m +Ġhit ting +in ux +M ich +ud ge +AT AB +Ġvulner able +ä ¾ +Ġport folio +: YES +ĉm ap +B ound +Ġiter ation +in cess +Ġact ors +ĠQ ual +_c lean +ãĢij ãĢIJ +MS G +G reen +ĠOff icer +Ġsm oking +> ', +ĠF lo +++ ; +oly gon +Ġbul k +Ġdr ama +Ġexception s +os ed +Ġ+ čĊ +Ġleg acy +C V +Ġcontrib uted +ĠTer ms +Ġb t +Ġunt uk +Ġal ien +=== Ċ +ĉ Vector +Ġl s +On line +.f acebook +num eric +ock ets +A ut +b ury +-re dux +ĠRed istributions +GLOBAL S +urrenc ies +Ġt ons +âĢĻ , +Ġà ª +(c ol +ĠS ymbol +Ġstay ed +ĠM L +Ġm unicip +Ġsex o +S en +n r +Ġg ains +Ġshort ly +.M enu +à ½ +KN OWN +Ġoper ators +- V +ĠPat rick +/ add +_C O +ir ation +(p ost +Post s +/ _ +Ġpl ug +Ġintellect ual +Ġmet ab +Ġpregn ancy +ĠPrem ier +n m +Ġpred iction +ĠMin istry +Th ree +val uate +ĠMin i +b u +оР· +< ul +Ġd d +ol ving +ĠC ut +Ġs chem +.tr ain +it ate +Ġr ice +Ġbird s +ãģ « +m iddle +struction s +Ġn erv +a que +Ġfl u +Ġsurv ival +ĠGal axy +ĠF ant +. Order +At trib +irt s +é c +M ovie +Ġcon ce +qu arters +Ġm ood +.Add Range +Ġres olved +ãĥ Ī +Ġburn ing +ĉĉĉĉ čĊ +ĠW E +Ġhost ing +L AB +Ġman agers +Ġstre ngthen +< const +ĠFire base +on ed +ĠJ ean +' ";čĊ +ĠS av +.B old +Ġen ables +ĉt mp +Ġman ually +ĠS qu +user id +.f unction +.c ache +LO PT +.S ervices +dd it +t im +< img +ĠTh ings +ĠEvery thing +Ġa pt +em and +Ġroll ing +ë ¦ +. level +Ġst om +ĠW inter +Ġview ing +( values +ocom plete +v ia +up o +Ġabort ion +i ère +ï¼ ij +_B UTTON +_d omain +Ġb ra +ĠA st +in as +Ġstat ist +c od +L R +Ġdr ives +Ġfollow ers +Ġall ies +ĉc urrent +ecess ary +Ġdam aged +_ pt +and les +oun tries +Ġsim ult +e u +Ġcontrovers ial +_G ROUP +Ġr ib +. Info +: mm +.n ormal +_ADD RESS +Ġ íķ +add le +ĠD ur +. Element +W arnings +Ġcred its +Ġin hib +Ġem issions +Ġh az +.y outube +ugg ed +Ġbo ther +ĠK ansas +ĠF ixed +ĠTest s +ĠF IX +Un iform +Ġk ont +>> > +st ation +lo re +at ype +ish op +/ **************************************************************** +Com boBox +Ġvac ation +Ġiniti ative +Ġdefault Value +con cat +ĠK h +ĠW elcome +ized Name +M igration +Ġgrad ient +H ot +Ġhard ly +el o +ĠStud ents +Ġlo ose +at z +.S end +' / +Ġunivers al +Ġenter prise +Ġreg ex +Ġvis itor +ĠF ly +Se q +à¸ Ļ +ĠVis ual +Ġlib raries +ato es +P ayment +Ġp ent +Ġgather ed +VRT X +ĠD M +S plit +Ġlet ting +Ð Ŀ +_error s +ep och +P ARAM +c u +ÑģÑĤ в +ol utions +Edit ing +font s +Ġalloc ated +ĠB ased +( Y +ĠJud ge +Ġbro thers +FILE S +ç o +w b +_P I +' ^ +Ġs word +.s ervices +Ġn l +T im +ig g +ĠMo ore +Ġcrypt oc +åĩ º +_post s +ot ate +? ' +... .ĊĊ +Ġk l +=" $ +Ġdec oration +Ạ¡ +ĠD IRECT +G UI +) =>{Ċ +Ġnews letter +Ġprec is +(p oint +ĠEqu ipment +ut y +ĠD ave +Ġparticip ation +u arios +x it +.A s +ET ER +or ous +Ġsh ield +[] > +ilit ary +. origin +Ġprom otion +U nt +Ġc t +TR A +View Holder +Ġsig ma +d elta +are house +con tract +( Vector +Ġcompet e +/ form +/ components +Ġn r +ĠInd ones +Ġо ÑĤ +ĠV olume +.f iles +(res p +/ models +Ġsur f +stand ard +/ o +ĠXCT Assert +V ICES +.C ode +SE D +Ġact ivate +D elta +Ġlimit ation +ri j +Ġpregn ant +: ^( +Ġs our +p ie +Ġexp ense +ic ation +ĠL arge +Ġ ± +ĠB owl +(model s +/ N +P a +.re load +Ġwonder ing +Exec ution +ĉ ĠĠĠĠĠĠ +ĠG raphics +ĠCont in +_j ob +Ġget Name +ĠM agn +ĠD WORD +m ad +Ġn h +fe atures +} ");Ċ +he ets +(tr ain +z n +Ġrecru it +.con nection +Ġbar rel +Ġste am +_set ting +Ġang ular +ane ously +Ġb il +ĠN orm +(! $ +ib t +% ( +Ġpos it +ĠF ather +int endo +L ive +Ġport s +Ġme j +Ġland ing +pon der +Ġc od +_HE ADER +.M argin +Ġball s +Ġdiscuss ions +Ġbl end +H ex +Ġfarm ers +Ġmaint aining +ĠĠĠ čĊ +s yn +[ T +r us +uff ers +Ġcontrib utors +_s ys +.De bug +Ġconstruct ed +om es +? id +sl ider +Ġsup pliers +scri ber +p es +Ð ŀ +": čĊ +\ Controller +)) ĊĊĊ +Ġl ua +M ulti +EN S +S rc +Ġpet ition +Ġsl ave +look ing +V ERT +ĉ vector +S pecial +h h +an ne +ĠN iger +/ views +z ing +end ant +< C +s peed +Ġ{ };ĊĊ +Begin Init +Ġf open +@ RequestMapping +End Init +Ġp unch +S ender +é Ķ +get Message +/t ypes +.P I +(' ');Ċ +oc used +( all +Ġdrop down +). __ +ĠV in +.Fore ignKey +can f +ou red +ĠOrgan ization +ĠÐ ° +ĠC ulture +(cl s +, _ +rg ba +ìĿ ĺ +.data GridView +Ġdo zen +ĠG es +_sh ared +n ick +Ġh osp +om eter +Ġclaim ing +ib les +ri k +æĺ ¯ +en ario +Ġd engan +ob b +m ont +_r ank +('/ ', +Ġap olog +P s +_p ower +ĠG ree +Ġful fill +Ġfire base +Ġf are +ĠH im +Ġbe an +â̦ . +ĠS PI +_R X +Ġper ception +rel ative +comp ile +u um +ut os +a uc +ĠAs k +Ġindic ator +/ th +.set String +ĠWis consin +.D omain +Ġart ificial +De velop +ĠSar ah +Ġl ying +( search +ĠEmp ire +urr ing +æĹ¶ éĹ´ +=" ${ +Ġget Id +ĠP ayment +trans ition +Ġ ]. +ix in +V T +- select +Ġdemonstr ated +Ġlast Name +employ ment +.get Property +Ġf ought +file Name +ĠP ers +-c ard +a str +attr s +Ġprom inent +Des ign +anc ouver +ãģĹ ãģ +ard o +se cret +Ġr ag +Ġpo ison +-m an +, omitempty +ĉ un +it zer +ĠCas ino +ĠR oss +- foot +(result s +Pl an +Ġlas er +ê¸ ° +_D R +F acebook +Ġbo ards +st a +] ], +Ġt iles +S IZE +Ġ= ~ +Ġprem ier +oc ab +Ġenc oded +Ġres erve +ĠAfghan istan +ĠList Node +url s +Ġsub mission +Ġne u +Ġ# +# +_P OST +Ġmo ist +ell i +ellig ent +. alert +ó d +b re +ĠCol lect +Ġgraph ic +Ġlong itude +ĠPro vid +ĠCal culate +x ffff +c riteria +Ġw aters +ro ck +lo quent +ĠT rib +Ġbur st +Ġsuff ix +.Ext ensions +ish es +iv el +ĠLI KE +ĠGet ty +.Action Event +.s lf +ĠH AL +up al +E AR +ud i +_time out +U F +ĠSing apore +ĠAd vent +_int erval +cha ft +ĠE mer +Ġtele phone +ĠTur k +_ interface +ĠO wn +Ġencour aged +< Object +_T ext +ĠOnt ario +ĠApp ly +.f irebase +Ġant ib +P riority +ene z +D ays +c id +urre nce +; / +inn ed +Ñģ Ñı +Ġve z +f w +// $ +att ack +Ġstart up +ain ers +.f ragment +op acity +( conn +he im +.n etwork +( stream +ĠN ON +t ol +ĠX box +ĠD S +Ġc ached +Ġprostit utas +ĠB alt +(' [ +Ġno except +" ' +Ġs d +. valid +_ ag +Ġr aces +Ġro d +itud es +< >( +.Pro duct +Form s +NE W +P ay +ĉ boolean +_ contact +ĠElect ric +sk ip +Ġw ur +Ġch ronic +_d river +ĠS ab +ĠU lt +ĠR ad +ST ATUS +ĠLew is +O B +Ġgift s +.Re c +TR UE +Ġint ensity +Mark er +.com pare +ff ic +C ookie +ĠB aby +ĠBig Decimal +ile t +ĠHOLD ERS +ĠL ady +Ġl ung +ĠAl abama +Ġd ess +` );Ċ +ĠB uilder +_reg ion +Ġne utral +Bo th +Ġh p +Ġh orn +Ġseg ments +ĠE C +"=> " +( rec +ĠP i +G M +Ġl aptop +Sc alar +is d +-d ialog +ĠAnd erson +Ġmist akes +ĠH an +j es +est ination +Ġprom ises +b id +ĠSc ient +G IN +ĠPer formance +b age +. users +le ading +Ġor al +G raphics +_P TR +h ang +Ġin ev +process ing +F actor +ĠN A +$ string +Ġground s +.Save Changes +c lock +cri pcion +ĠNew ton +g c +.in cludes +Ġbl ast +Ġ'- ' +Ġpued e +.S ession +Ġgre p +_f inal +ĠG ay +ĠG ive +ir i +-st ar +ĠUI Image +_ep och +ub b +ent h +Ġel ite +Ġcampaign s +ĠP orno +_ assign +Prot ocol +ĠBe ing +ĠAir port +Ġconvent ional +ĠW at +ĠC I +ET A +ĠAnth ony +Ġtable t +( format +Ġconsist ently +ĠI owa +Ġav atar +.c ursor +! [ +Ġh anging +H er +S uch +';ĊĊ Ċ +orge ous +() == +Ġview Model +Ġ ãĥ +Ġel s +ĠAg ent +F etch +ap or +Ġc x +p read +ĠP ier +oe ff +S n +ĠV irtual +A pr +.Wh ite +_M OD +ĠPoint s +å¤ ± +Ġgen es +Ġv endor +Ġmain stream +< src +ĠEl izabeth +Dec oder +- state +ĠG lass +nc y +adi ans +_m on +ĠRem ote +Ġwire less +ĠM i +å ī +è¡ ¨ +st age +ĠT ile +ll ib +V ariant +== Ċ +Ġgold en +(Q String +.put Extra +ĠD om +ĠAn imation +Ġinter active +if act +éĻ ¤ +LE T +Ġfrequ ent +Ġ< >Ċ +F ilename +Ġs ne +ĠFoot ball +Ġr ival +Ġdis aster +ion ic +ĠD amage +. Resource +- en +ĠT ypes +get String +( board +Ġb ol +pl ain +z ym +ภ² +Ġsc anner +ild er +_msg s +æ ı +(int ent +Ġde struct +Ġb ust +ĠE mploy +on i +ĠUI ViewController +Ġodd s +ear er +Ge ometry +Ġy ii +_EX PORT +ĠAtt ack +Ġn iet +Ġim pression +ĠG il +_pro b +ĠC F +ĠEx perience +/pl ugins +.M ethod +Ġbelie fs +N ative +_b uild +Ġv ig +Ġr anks +cover ed +s uch +G uard +.p ack +add er +iv ia +l ng +Ġв Ñĭ +T imestamp +_n ow +Ġp oker +Ġun c +Ġsh apes +-t ypes +_per iod +p k +Ġveter an +Ġson o +Ġappoint ed +over flow +.d river +_c at +ut t +pl ant +im b +ĠAc cept +Ġconc ert +ĉ node +ĉ z +? >čĊ +Ġb anned +ĉ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġto xic +Ġdisap pe +È Ľ +Ġgr ace +ate ful +Re ply +ĠCru z +Ġsc rap +Ġkey words +s imp +Ġmort gage +Ġcy ber +ĠEx ecute +Ġlat itude +if u +.C OM +d bo +Ġsort s +ĠG as +om ial +.L ocal +Cell s +.Re place +String s +.f it +ĠTh ird +% ",Ċ +Ġ{} ". +ĠS ony +Ġ[ : +Ġfall en +. ')Ċ +in h +ĠM C +Ġred is +C odes +Ġprofile s +h ook +Reduc er +_F UNC +Ġn avigate +str len +Ġh orm +á ŀ +ĠS R +. boot +Ġdig est +ĉ header +.find One +æ ģ +Db Type +n ia +_m erge +Ġdon ne +/ Getty +_CH AR +Ġb ands +. URL +art ial +Ġf req +Ġs ist +N g +Ġrender ing +\ Core +Widget s +ĠV A +Ġactiv ists +St e += _ +all a +St amp +Ġload s +Ġx x +ĠL earning +.M vc +u ir +(" $ +Ġconnect ing +Read Only +ur u +ĠE ag +B IT +_DE L +å § +arr ass +ext ernal +ĠY OUR +ĠB rew +ĠF ive +Ġres ize +ig id +er ation +ĠÑ į +åĬ ł +ĠC atch +Ù ģ +ĠLe on +am il +.B ody +Cl ip +/ list +.b r +Edit Text +ĉ db +.G ame +(Build Context +back end +.R ed +face book +.url s +m r +rol led +---- --- +Ġinter vention +Ġretire ment +ĠK it +ĠP RE +Upper Case +ĠS ocket +Ġ: - +Ġstudy ing +ĠMet ro +ard ed +Ġconvers ations +C alled +Ġexam ine +ert ificate +.g z +-res ponsive +Ġref und +_n etwork +allow ed +em pt +Ġme als +C ategories +Ġtravel ing +Ġk g +Ġsh ame +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġexplicit ly +Ġmath ematic +ĠS uite +ĠR GB +****** / +Ġmix ture +lear ning +.t emplate +att s +w x +ĉ ctx +.p roperties +Ġdrink s +ĠE ither +set Text +.get Data +.z ip +Ġreve als +< table +.Hash Map +ĠH ur +) ");Ċ +.f ramework +ĠST ART +feed back +Ġsaf ely +. icon +config ure +. lock +.l ayers +/> .Ċ +Ġrank ed +_ impl +ĠHand les +Ġhost ed +Ġup dating +al bum +é Ŀ +Ġsh ader +Edit ors +- round +[] { +Ġse p +ĠH i +TE M +look up +.m an +_IN PUT +Ġthreat ened +_IM PORT +Ġd rops +ru it +s id +bo th +ĠEx cel +Ġj er +ord inary +еР¹ +V IEW +re ply +Ġ) :Ċ +color s +ver ified +_T r +_p arse +Ġcon gress +P romise +int s +ĠM other +.A pi +ĠD uration +Ġfirst Name +inherit doc +ĠM ars +Ġa pr +OD Y +Ġvis its +Ġhe aling +let ters +)) );čĊ +f uture +.F ramework +Ġk iss +Ġinv olve +Ġsil ent +ad ows +Ġany body +s ch +Ġsole ly +- img +Ġprop ri +Ġin struct +Ġlic enses +Ġm eth +Ġcond em +ĠD omain +ĠHarr is +Ġs Ã¥ +CE PT +B atch +@ extends +ĠCONTR IBUT +.Data Frame +_p acket +rec ision +Ġfoc using +. ht +__ ":Ċ +: Get +ĠK C +Ġpass age +Seg ment +_c enter +-z A +_B L +Ġconv in +Ġclass ified +ĠNS Mutable +_ ap +t ile +Rect angle +(n ums +v ens +ĠUI Button +ĠF eder +am o +Ġout line +ĠPar ser +Ġâ ī +ĠWork s +.S chema +Ġeng ines +_com mon +_ old +Ġset ContentView +Ġ/// < +ĠB T +f m +Ġd ivers +_ weights +em ark +ĠA CT +Ġpro portion +over lay +.dir name +ĠG it +_REF ERENCE +< > +l b +_r ule +è´ ¥ +ĠPut in +Ġsleep ing +() :čĊ +Ġpres erve +Ġpar liament +ĠLook ing +Ġpick ing +ĠDis patch +Ġsl ip +ë ĵ +ĠL yn +_sign al +config uration +ĠP itt +ad en +pro cedure +Ġenthus i +f ight +ĠCons ider +Ġt orn +Conn ected +.c os +_group s +ĠTh ink +Ġdel iber +Ġres id +work ing +.column s +ĠCal led +Ġes lint +> ", +_D OWN +h ist +ĠAdv anced +Ġre wards +act ors +Ġsil ence +Ġmy th +Ġne ur +Ġa uction +.Get String +ek s +( project +ĉ msg +ĉ output +Ġcomplaint s +, S +Ġt bl +Ġ, ĊĊ +ri ors +ah ren +Ġlawy ers +re dux +_s ymbol +off ee +_RES ULT +( Name +UT C +.current Time +Ġorgan is +. arg +Ġmin im +w ick +Ġrece ives +B alance +Ġspeak s +ĠD ays +ĠBel ow +t ipo +P resent +Ġres erv +h p +Ġr it +_R IGHT +-- ) +Ġchair man +D IS +ĠBO OST +Ġexper iments +__ );Ċ +Ġst amp +Ġf ert +Ġf ond +T er +el ve +ure n ++ i +end ency +Ġvirt ually +... " +ï½ ŀ +- cent +_un ique +Ġpr icing +m ic +RES H +Ġ:: : +Ġan notation +ĠC ircle +ong odb +it as +Ġ% ( +( component +Ġо б +( port +-h our +. obj +L BL +Ġj ury +GB T +Ġsp y +ĠProf essional +Ġ"" ;ĊĊ +Ġstri king +Ġdiscrim ination +Ġp ays +lic t +ent es +Ġthrow ing +ĠPl ugin +( def +ĠRuntime Exception +ĠM igration +Ġd ic +b ag +on ia +Ġcor ruption +( Map +Ġpr z +.d to +Ġac quire +State ToProps +Ġlo ving +оР¶ +_p attern +Ġemot ions +Ġpublish er +_b e +Ġcoup les +o j +ĠCh art +Ġt rop +.t ool +Ġestablish ment +Ġd ol +Ġto wer +Ġl ane +ĠSy dney +Ġfill ing +claim ed +Ġdialog ue +Ġcon vention +book ing +pare ncy +æ ± +ĠGener ic +\ Schema +Ġr anges +/ ch +Ġpan els +Ġr uled +çĶ Ł +.t s +_s ets +Ġclean up +Pre vious +ĠAn imal +($ ( +ĠA ve +oll ar +_e val +ĉ Name +(t ree +Ġ" ] +Ġdut ies +=' / +Click ed +Ġdifferent ly +ĠCl ark +Ġd it +olog ists +Ġsy nd +Ġs ends +- known +k b +ĠMod al +it ative +Ġr acing +Ġhigh lights +ĠSim on +ĠCapt ain +ä¿ ¡ +ĠC B +cont in +ar an +Ġphys ics +ret ty +et al +.m d +ax ios +Ġspeak ers +Ġpre p +Ġaward ed +ì§ Ģ +ĠC orn +ĠN ature +UD IO +Ġpro j +- pre +[ u +Fe atures +Ġis Equal +B inary +s ig +Ġconf usion +ĠH at +Ġkt ó +.config ure +M ON +/ edit +_A dd +, true +Ġc li +Error Message +- loader +Dim ensions +ultip ly +Ġ{ !! +ĠSql Command +Ġsp oken +Ġp ics +Ġto y +( Key +ĠLo op +Ø ¨ +E ATURE +in ction +_set up +w rapper +Ġt ong +c ular +O pt +.P l +=" , +(l ength +um n +Ġch rom +Ġse vent +ĠIllegal ArgumentException +ĉ start +Ġbeg un +CE PTION +dat aset +ĠF ailed +col s +Ġkne e +im ore +.sp lice +sh ell +ig gers +Ġthem es +ĠD J +ĠAss istant +- $ +May be +Ġorder ing +ĠInt elligence +ĠMass achusetts +Ġfail ing +el son +G reat += i +.re st +Ġinv ite +-dis able +.Group Box +âĢĻ est +Ġtack le +g v +et ter +Ġ), čĊ +_r ules +.w arn +function s +ĠChrist ians +Ġback ed +Ġsl ider +Ġenjoy ing +n est +Ġh ij +_m s +// * +An notations +ĠVariable s +< V +( server +ĠOr acle +element s +Ġorgan isation +_point er +ĠHe aders +[ d +Ġdead line +iss a +Ġkn ife +ĠNAS A +ĠHe ight +ĠAs ync +Ġven ue +.d om +bour ne +ĠHaw ai +Ġmem o +ict ions +Ġsurve illance +om i +/ assets +Ġed u +Ä Ľ +Ġro ster +Ġh ired +ĠT ok +Ġpl acement +ur ations +Ġset State +ĠMag azine +Ġhor ror +T ry +Ġl ag +ĠEvery one +th ur +)) ;čĊčĊ +. return +Ġsy mp +âĸĪ âĸĪ +Ġn ights +work er +Ġa le +ennes see +.st ep +Ġsynchron ized +our i +Do es +. change +f on +.set Background +irc ular ++ - +ĠC IA +ĠJ ane +ĠSim ilar +- I +level and +Ġpros pect +_f ound +ĉc olor +.D iagnostics +Ġann ounce +Ġassum es +/ tr +Ġb d +ĠCar bon +Ġanal ys +.de st +n ik +ĠL ie +- index +Draw able +ĠT AG +Ġtri angle +_F LOAT +ĉĉ ĠĠĠĠĠ +.bl ack +v ue +cur acy +Ġaffect s +Ġsure ly +Sl ider +uk i +c ery +Ġun ter +.pro file +ord on +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +le ave +Ġsmart phone +g ie +Ġcons pir +Ġt utorial +ç± » +Ġc ab +ĠSum mary +* ĊĊ +ä h +" This +Ġsl ides +" +c ycle +ĠB ull +path s +Ġun p +Ġview DidLoad +_M odel +Ġassert True +Ġr ated +De cl +vert ed +ĠD at +b rew +Ġpoint ing +M s +ĠPoint er +) ' +_n on +ĠSE C +Ġy eah +g ency +initial ize +f ly +[ pos +, g +Te le +Ġj oke +Ġcl ause +.find ById +en es +( instance + £ +Ġs lic +_h ome +Ġ*/ }Ċ +_p ages +(s ervice +R P +ĠAm ong +.get Current +ãĤ ¹ +Ġs lee += [Ċ +ol er +Ġlib ert +Ġ` Ċ +Ġw enn +l ated +Ġimm une +( Node +ĠPro blem +ĠA bs +log s +Ġ ../ +ĠA DC +Ġ}} ">Ċ +> ');Ċ += b +ĠW ind +lah oma +Ġalloc ate +or ian +Ġpres cription +- quality +ĠMay or +in ely +end foreach +ĠCom plex +k om +T Y +] ]. +. Style +_m any +',' $ +Ġbar rier +ĠF etch +ĠMar vel +Ġres ist +ог о +b idden +ĠRun nable +: false +Ġbuild s +ĠSt age +Ġd ub +emp o +.s ite +;ĊĊ ĊĊ +ĠDen ver +Ġre vel +Ġtrigger ed +Ġd ice +_f ail +Ġg c +ĉ X +ĠTh rowable +.r outer +ĠRev olution +ÑĢ Ð° +_N ON +Ł ¥ +Ġel der +Ġab road +ĠÐ µ +ĠAd ult +bl r +g lyphicon +Ġprom oting +Ġ iz +ĠS olid +_lo ader +ear ly +.en abled +- edit +ĠU L +_ play +ĠInt errupt +Ġadvant ages +uc le +Ġmechan ical +.table LayoutPanel +ĠWork ing +Ġan onymous +R ating +ig ious +_ph one +.addAction Listener +Ġfr an +und en +Ġ*) & +_ bool +ul ative +Ġcon e +ĠM ult +Ġm ö +ĠFor ward +] ):Ċ +Ġconvin ced +act ed +ãģ ĵ +ĠConfig ure +Ġce iling +D er +Ġpass engers +Group s +Ġsoc cer +/ W +avi ors +sw ith +ĠZ one +. Options +ĠM om +ied er +Array s +Ġtreat ments +Ġprotect ing +f ac +Ġpick le +Button Item +Ġblock ing +str ar +à ² +ĠEx port +Ġth rew +ott a +ĠB ASE +.w s +.LE ADING +order By +_d elay +ĠP u +.d ll +ĠCh oose +Pol ice +ĠBE GIN +box es +Ġdiam ond +, l +Ġ ĉĉĉ +Ġcur ious +t v +Ġerot ische +ack ages +ĉ Set +T ick +.b order +static method +Ġch er +in voice +Ġcr u +Ġdef ect +_m etadata +re lation +ik an +[ N +(Q t +( Base +æģ ¯ +be at +ĠEm pty +ĉ o +_sh ift +Ġreg ret +Th ose +C ent +ĠPort ug +ĠIs lands +ĠT IME +Man agement +-s p +ê me +Ġnot ion +un ifu +P K +è¡ Į +ĠCUR LOPT +\" \ +U V +ç º +d ra +c ou += ` +ĠD estroy +r p +.c ancel +G G +r untime +ĠV ue +Ġprogress ive +/s ervices +Ġrun ner +_FR AME +.ToolStrip MenuItem +Ġ' ,' +d elay += utf +Ġscreen ing +Ġpull ing +om as +Ġan th +- new +/ local +Ġi Pad +Ġt witter +Ġd ying +Ġhe aven +ĠU Int +ĠSen ator +Ġpres um +ĠWalk er +Ġover come +ete ction +Ġemb arrass +Ch ina +In clude +RO LL +Ġdata Type +D avid +ภ£ +lo p +-m onth +Ġsc ar +ĠS afe +Ġ **************************************************************** +Ġaccess ories +Ġr amp +_U SE +Ġcontr ad +)) ]Ċ +Ġpre st +ĠH R +ĠR ap +Ġus ize +Ġcap ability +Ġc ort +- next +Ġbur den +_read er +Ġ@ @ +reg ular +ĠK a +M AN +Ġa str +Ġ' ')Ċ +Ġf ed +Ġpars ing +ĠY ears +Ġbro ker +": {" +Ġa kt +In ventory +abe led +Ġarg parse +****** *Ċ +vers ation +Ġc ord +ĠT i +Ġhope fully +Ġa h +ver b +Ġst olen +. Entry +Ġexpect ing +O rientation +Ġpower ed +Ġp ersist +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +'] ); +')) ,Ċ +ĠC ash +ĉ item +gr ades +rop ol +b asic +Ġ" );čĊ +Ġaw ards +(r ange +- all +ĠIB Outlet +ĠInd eed +---------------------------------------------------------------- ------------ +Ġstom ach +Ġfl ower +Ġs ew +_t imes +av is +Q String +ĠR outes +_pro t +Ġcom edy +Ġlog out +Ġwood en +Ġpost er +p iece +.J oin +ĠP ok +cel ona +mut ex +;čĊ čĊčĊ +Ġstri kes +Load ed +) arg +es a +Un ited +E p +PE LL +ĠAtl antic +ul let +app le +Ġsett led +a con +Ġprint er +ĠG C +å® ļ +Ġrender ed +, âĢĻ +he it +s ocial +. ge +ĠR ick +ĠUt ah +g ot +on ical +ĠSc roll +ĠSc iences +Ġj ug +Ġam pl +ent i +LE FT +Ġt abs +Ġenorm ous +.get Key +loc ate +. EX +.st orage +.W e +Ġto ast +ĠAdd itionally +ĠN OW +_ UPDATE +Ġtrans ferred +th a +.D isplay +_ ui +ID EO +Ġmeaning ful +ĠMos cow +, this +ĠVict oria +æĶ ¹ +ĠÐ Ł +.st ack +ĠB arn +pared Statement +: string +Ġb ij +ĠST ATE +Ġemploy ers +ĉ input +( | +Ġle x +in voke +ĉ num +++ , +at ial +ors es +Ġfor k +_t xt +ĠAnton io +Ġ( < +aver se +Ġdev ast +ãĢ Ģ +.D ec +ĠG ard +/ ui +. % +tr i +Ġrol led +Value Pair +itt en +ĠTh er +Ġv rou +ĠFl ow +ĠFin ance +ĠCom b +H C +.set Visible +is l +Ġp k +Ġup set +( raw +ĠV ice +e atures +ĠL ang +Look ing +ĠA ST +Ġtri ps +ĠJust in +b rowser +=" '.$ +. vertices +- co +}/ { +Ġ? , +ĠD omin +ĠBel g +" < +Ġsup pose +add y +Ġwalk s +ERR U +_f ilters +Pre ferred +sc ene +е Ñģ +ĠAff airs +Ġ"# { +Ġon Submit +Ġstock s +/ view +g ree +- get +h it +J o +.get C +Initial ized +ÑĤ и +c uts +( Type +ĠAg reement +ĠViet nam +Ġ/* ! +Ġp izza +- view +_ em +Ġl hs +Ġm uy +ĠId ent +ĠF riends +Ġab und +_A D +.t imestamp +- ' +Ġd uplicate +Ġhun ting +Ġregul atory +ia o +am ous +ĠEnt ertainment +[ A +iat ric +_CL IENT +ĠK ids +/p kg +B reak +)) );ĊĊ +ĠSh ape +Ġrel ating +Int errupt +able Opacity +emb re +Ġmyst ery +Ġjournal ists +rit able +.L ink +Ġstop ping +CRE T +.D B +Ġpopular ity +Ġg ew +Ġim pr +set Value +FL AG +ĉm ax +Ġb ake +w y +ĠEcon omic +Ġen contr +Ġf name +/ de +R ank +Ġbug s +.s m +Ġmed ian +D OWN +ĠS ure +At Index +ĠD ick +Ġ( __ +.d elta +F r +Ġsuggest ing +ĠRec yclerView +, e +ST ART +/************************************************************************ **** +xf ord +Ġrece ipt +CL AIM +read only +Ġeng aging +C a +as ma +Ġens uring +Eng lish +ĠV ancouver +hy th +Ġpurch asing +ĠP I +. word +(s p +.h ome +: def +Ġg ig +ĠV e +for um +ĠM itch +B ay +_F L +Ġs oll +_column s +Ġminor ity +b ird +Ġhand ed +SS L +ST AT +Ġnerv ous +ĥ ½ +Ġfile Path +CRE ATE +A w +Ġp ens +se ed +ĠCom pute +ol k +ĠAs set +re ach +'), čĊ +n avigation +L F +/ util +ĠP ub +Ġâ Ķ +c ion +## Ċ +II I +Tag Name +Ġam id +per mission +if iable +xFFFF FFFF +н и +.B uffer +_ irq +d ark +Ġret val +.f ire +produ ction +.list en +ĠWe ather +Ġbuy ers +. ne +er p +ĠP ent +Ġw elfare +Ġpage Size +ĠSt adium +ert a +Ġle v +amp a +P ager +Ġcharg ing +ĠNet flix +| null +_r andom +.x path +Ġst ere +ĠIS IS +pons es +( loc +ey ond +ĠOff icial +ĠMary land +Data Type +_p ar +{ }, +ĠEn joy +_SH IFT +ĠA wards +_ENT RY +Ġseem ingly +entic ate +Ġheart s +_ ;ĊĊ +ĠH IV +Ġindiv id +ĠFl ag +_ ctrl +ĠC allback +, z +ĠG PU +ĉ obj +ĠPh oenix +ĠB US +Ġrub ber +_A UTH +ĠSol utions +( location +Variable s +.set Enabled +_h igh +W O +G esture +Ġre try +Ġobject ForKey +allow een +Ġm os +ĠC ele +Ġik ke +(c ell +ĠM ODE +ren a +Ġdescri bing +Ġph i +Ġr d +Ġdes erve +Ġwhe els +å¸ Ĥ +Ġcrit ics +N amespace +ĠF ra +Ġ ĊĊĊĊ +Ġall a +Ġrequ iring +æľ Ł +ut ation +Ġdelay ed +Ġadministr ative +Ġb ay +.h idden +T ex +Ġbound aries +Ġ] );ĊĊ +ĠFollow ing +~ / +F i +_con v +_T ITLE +Ġdes de +ICollection View +Ali as +Ġb ite +pat ient +_COMM AND +Com pleted +ĉ elif +( < +B usiness +ĠP ool +Ġpurs ue +ĠB an +_st eps +_DE CL +um ble +Ġcom bo +ĠL ayer +.x r +Ġd up +-------- - +Ġmod ifier +ro b +re z +Ġath letes +Us ed +w ear +Ġlegit imate +Ġ" ĊĊ +Ġh v +St d +ĠH old +Ġsurv iv +ĠAll iance +ĠEar ly +Beh avior +(f ont +/lib s +Ġrect angle +Ġs inger +Ġam p +Equal To +Ġ" ." +Ġgirl friend +å ± +line ar +obs erv +Ġpi ù +Ġcomple ment +With Value +(p assword +t ake +Bl ank +ĠCom par +' ", +_p olicy +m ongoose +_FA ILED +.re port +R atio +.Perform Layout +us able +m ers +_re nder +PE ED +Ġles b +ĉ E +_t ool +Ġl adies +о Ñģ +)) ))Ċ +;; ;; +.d ot +Ġn est +pe ak +uk kit +ec a +_S W +Ġ& ( +ĠOk lahoma +Ġbank ing +ĠN intendo +Ġreprodu ce +_element s +_m ac +pro xy +Ġremark able +}/ ${ +Ġout s +.has Next +M ODE +Ġan ime +.con n +Un ique +D om +Ġimportant ly +itt y +Ġju ice +T w +ĠPart ners +Ġattack ing +Ġport able +am iento +.P ictureBox +.g en +Ġopt imal +Ġre cre +Ġjournal ist +ĠEx tract +ĠMore over +Ġmargin Top +.A p +Ġf iring +Na N +ĉ template +аР´ +. En +Ġdef ence +ĠT el +il en +j an += data +ĠU rl +ĠRe uters +(t otal +ĠFif th +Ġess ays +Ġinterpret ation +Ġchar ity +ĠR ules +Ġsub section +st yled +az er +l ags +L IST +Ġupload ed +Ġtr ash +Ġreg istr +Ġsell er +>' ;čĊ +Ġstart Time +ç Ļ +s y +(Http ServletRequest +Ġtr ap +G C +Ġembed ded +Ġsurround ed +im its +T X +yl inder +ĠF al +Ġsent ences +ĠJ a +IF ICATION +we apon +ov ation +Ġco at +Ġinter pol +Ġl ips +ĠK y +Ġv ectors +_ am +Ġint ake +.w orld +Ġin box +ĠM AC +_ ab +(name of +Ġent ert +Ġgather ing +ĠS IM +++ . +ny a +' }} +ĠUP DATE +Ġp ac +( html +ĠS ant +i ating +ĠIde as +Ġspr ay +ĠH art +Ġver ification +ades h +/ modules +ĠM ind +ĠSized Box +Ġsh elter +Ġher oes +att y +Ġcert ified +s j +Ġê tre +ÅĤ o +Ġpublish ing +ĠMal ays +.get User +ĠPro vider +ĠLinked List +ĠB or +RO UND +d id +t ain +p ire +ĠJ enn +t el +and e +_f ront +ĠMc G +Test Method +à¸ Ń +Ġoccasion ally +ĠW ales +Ġexerc ises +ĠÐ Ĵ +- plus +Ġvalid ator +Ġpr ayer +L ATED +_ author +Ġlab our +++ Ċ +-e quiv +ĠG PL +Ġface book +s imple +g ly +Process or +ip y +Ġ* > +Ġcle ared +ĠP ush +Ġpen is +Struct ure +li j +ĠM organ +Ġhand ful +" .Ċ +| \ +Ġ ******************************** +ĠA qu +_ IC +.load s +Ġm eter +ĠMar ine +:: { +ĠT S +ĠArray s +.T itle +GR AM +ter min +Ġco inc +El se +_st ates +-r un +m embers +ast ro +Ġon Press +Ġbe ings +Ġabandon ed +Ġtax p +own ers +.m ode +Ġdiagn osis +Ġ_ Ċ +ĠK night +ĉ A +Ġob serve +), ' +! ")Ċ +ĠPar a +Ġvari ation +( False +ĠAnt i +Ġg ri +Ġhome less +? v +Ġbe z +.S erver +re lease +ĠP atri +Ġchar s +Ġrank ing +activ ation +Ġw ides +q r +.S ql +ac ular +ĠB ot +_s ync +Ġhapp iness +Ġvolunte ers +Ġs its +/ < +[ e +(file Name +Ġcap ac +ĠMar ia +f ather +Ġgr am +* i +Ġcas o +_d raw +ĠR aw +ĠIter ator +ĠP adding +P D +BO X +ĠS PECIAL +Ġfe cha +Ġv ide +ĠLe ader +ä» ¥ +$ (". +Ġdiam eter +Ġm ild +Ġrock s +app ings +d irectory +.fl ush +ĠJ ess +UN IT +ĠP ear +Ġmand atory +S ur +q t +Ġstream s +Ġco operation +ĠS ac +Ġche aper +ĉ ch +an imation +f are +( height +( True +N Y +Ġw rest +Ġpoll s +Ġencounter ed +ĠMarket able +_P ASSWORD +_SE LECT +ĠArab ia +_c lock +Ġv oy +Ġи з +Ġst ir +is ible +-e ffect +.c reated +Ġto ys +ĠTrad able +Ġr ust +Ġstr cpy +_t imestamp +Ġtalent ed +, null +ĠJ obs +ĠPort land +Ġweak ness +Th row +ĠAng el +ä¿ ® +Ġun cert +ï¼ī Ċ +ĠìĿ ´ +Wh ich +Ġ[- ]: +S omething +Ġconv icted +k le +ed ium +Ġbranch es +Ġb ases +ç ® +Ġcomplex ity +ĠF ig +. reshape +$ db +_CON ST +ĠT es +.r untime +Ġden y +ĠB SD +Ġk r +h att +ĠSt atic +Ġunivers ities +Re place +Ġdro ve +Ġad oles +_pl ugin +ĠL GBT +Ġt ex +du ction +ED I +ĠT ed +_ URI +Ġre ception +art en +.S ingle +r ice +sc ious +_b g +Ġw ages +ĠS ervlet +UIL ayout +Ġform atted +.M od +< class +is en +Ġrepresent atives +"] = +Ġport al +ĠHun ter +Ġh iring +__ )Ċ +ric ulum +u o +li est +Ġt ears +L at +Ġliter al +.In sert +Ġc urs +ĠCom put +Ġterror ism +Ġswe ep +Ġ[] čĊ +Ġpass enger +Ġeast ern +Ġtwe ets +Ġoper ated +w nd +ĠS yn +.t ools +ĠW M +ul ates +Ġbacter ia +( bytes +.set Data +Ġvis ibility +// ================================================================ +el m +Ġgener ating +Ġm v +Ġk h +j en +/ search +Ġaccount ing +se gment +act ic +. ip +Ġdeploy ment +Ġfoot er +> ',Ċ +Ġexpand ing +ĠHam ilton +ĠCon trib +.T ables +Act iv +H H +ocom merce +_ ; +Ġamong st +ow ing +ĠC old +AP H +Ġpsych ological +_t ensor +Ġpack aging +ĠSw eden +Ġp are +Ġag gregate +Ġmoder ate +_h and +Ġdesign ated +Ġdr um +Ġget User +ĠC reek +_s cope +ĠTrans fer +ĠM arg +Ġfight ers +W nd +ĠS el +ĠLa unch +Ġemerg ing +if rame +ĠAdd itional +Ġf ears +Ġsat ellite +_ : +Ġdis posing +Get Value +Http Post +AT IVE +ul ary +View s +Ġatt ending +ĠT ennessee +ĠM ission +Ġmedic ation +ĠW y +ĠAn na +Ø ¹ +ĠVert ex +.t ypes +O rgan +.DataGridView TextBoxColumn +ĠR S +Ġtemp o +( App +Version UID +.p oint +ĠD utch +H ours +L U +Ġqu oted +.b uilder +ĠPer fect +ĠAl ways +_t wo +Ġexclus ively +ĠC ra +ific ar +ĠA WS +ing ham +com plex +k ernel +Ġgr avity +Ġw i +Ġover view +ĠW ant +ĠW P +( sh +. rotation +St ates +ĠTe en +_com ponents +ì Īĺ +Re ceived +Ġly rics +rit es +ĉĉĉĉĉ Ġ +-A merican +[ num +/ python +ĠU ART +Ġapp le +ĠJon athan +Ġmoment um +ภ± +Ĥ ¹ +Ġm ich +and ra +Ġbi ological +ĠM ens +Ġ% % +else a +ĠMex ican +.rand int +Ġt ale +ĠValid ate +Ġdefe ated +.ht m +Ġcop per += / +cos ystem +Ġr ip +dec imal +.V ISIBLE +ĠT a +ĉĉĉĉĉĉĉĉ ĉĉĉĉĉĉ +Ġdownload ed +en vironment +Ġnom ine +build ing +ĠSp ot +ipher al +Ġal to +qu et +ĠF T +/ get +/m aster +W IN +åħ ĥ +W est +arg c +Ġprodu cers +ĠM uch +_st orage +cred it +CON T +Ġv et +Ġvo ices +(' ', +Ġinstr uments +ĠM SG +es se +re pository +om ics +Ġdeal er +St ill +Ġb anner +asc ii +Ġrem arks +[ js +Ġshort er +g ulp +Ġmyst er +Ġk un +ĠB ird +Ġti ene +n ut +ĠU m +Ġw ise +Y eah +INE SS +_b egin +- heading +C ourse +Ġ čĊčĊ +omb ie +grad ed +ĠG PS +Ġ że +F it +c aption +ö n +/ image +l ia +(m od +Ġle ak +en za +/ H +ĠH appy +D ist +n x +ĠGovern or +(l ast +te acher +ĠS ent +s upport +ject ory +Ġ Ùħ +Reg istration +ĠGr ay +, false +Ġadjust ed +( settings +< R +ĠM age +Ġpl aint +_ )Ċ +ĉ it +omet ric +. bootstrap +Ġcar ries +I p +Ġ! $ +Ġswim ming +ĠMar io +ĠQuest ions +P ACE +æĸ ¹ +e or +}} " +Ġo ven +ĠK on +Ġwis dom +Ġac quisition +ess ment +ag ine +Ġexpress ions +Sequential Group +F ront +ul pt +aw k +'] )ĊĊ +_ AR +Ġanal og +ul in +_PR INT +ĠL G +Ġb lob +ĠFurther more +_com ponent +ĠC ole +L AN +SCRI PTION +Ġl ap +icens ing +_TIME OUT +ĠF ro +Ġli ability +Ġcom posed +.create SequentialGroup +_p erson +Ġbe am +ĉ ĠĠĠĠĠĠĠĠ +ĠNot Found +. 'Ċ +ÃŃ s +.Text View +P DF +Ġk ar +__ (' +Ġ" :" +_m essages +Ġhar vest +.h istory +> 'Ċ +-f old +æ Ĭ +ĠBet ter +Ġ"\ < +sp acing +Ġfurn ished +os er +] }Ċ +Ġ$ " +p ull +.P ost +( ip +Ĺ ı +.f ront +nt e +ĠF M +g uid +Ġnegot iations +agon al +Ġtrem end +unge on +Ad v +car ousel +ÃŁ e +_DE SC +Ġham mer +áº Ń +ĠĠĠĠĠĠĠĠ ĊĊ +-c ore +-s ervice +Ġcorn ers +ĠS F +p red +> A +ĠJ Label +Ġrom antic +Ġtestim ony +os c +ĠGener ation +as ures +_int ernal +Ġprint s +Ġ] )Ċ +ĠC leveland +re po +D isc +Ġ" >Ċ +�� �� +Ġne arest +_t b +( require +EO F +- child +Ġbu dd +.Xtra Editors +alt ies +\": \" +W ords +Ġloc ally +Ġpurch ases +Draw er +ex tract +Ġexec ut +} '. +user data +Ġfocus es +-min ute +ĠP ublish +og o +Ġmount ains +B ot +} >{ +Ġt ension +ro d +m esh +Ġtransform ed +, R +() }Ċ +.l ong +Ġg orgeous +ĠS chedule +Ġol dest +Ġsub process +( IN +y ect +ĠCo oper +arn ess +ĠMon itor +.p art +ĠN BC +Ġc otton +Ġh ol +Ġrg ba +ĠB io +Cont inue +P od +Ġparticip ating +clus ions +(By Val +à ¬ +ĠH OW +_set opt +Ġaccompany ing +at on +Ġ/ \ +ĠAuth entication +i én +ĠBar ack +/* . +Ġe ager +ĠC ancel +< lemma +ep h +ĉ window +Ġinc idents +), ( +.D es +ib e +ĠFunction s +Ġhosp itals +Ġo xygen +root Scope +Ġd rew +ĉ request +not ice +ak u +am ents +f ar +Ġprec ise +_w rapper +Ġlisten ers +A Z +.b ounds +ĠA verage +field set +_ axis +Ġexam ination +' .Ċ +mon s +++) {čĊ +ĠForm s +íķ ľ +Cpp Method +_tr ace +Ġengine er +ĠFl at +Ġrev ision +Ġhe ating +/ profile +.r u +p riority +Ġin fer +_ST REAM +Ġ* )( +> $ +OLE AN +OK IE +IB ILITY +U AGE +ĠSur vey +Ġres ign +w ing +Ġsecre ts +Ġch ips +JSON Object +Des ktop +_SY MBOL +(res ource +ĠĊ +Ġnew est +ul i +Ġdes ert +Ġd ip +ĠP ow +Ġequ ation +Ġposs ibilities +ĠF ed +os ph +Ġ[ % +Ġb ubble +ether lands +Ġc ement +. auto +_ AN +âĢĻ . +se lection +ĠB ond +D en +- O +.get Type +.W indow +p res +Ġsw inger +" })Ċ +Ġp ip +Ġm ice +Ġcomp ound +- plugin +ik o +Ġcent uries +ic ular +-in line +ĉ key +> \< +EN SION +Ġ[ čĊ +Ġprecis ely +Ġét é +ĠP ast +ĠCam bridge +-f ull +Ġanaly ze +ĠSte ven +Ġn em +d ue +ore n +Ġmus cles +ij ing +/ - +ĠKenn edy +R M +oss ible +Ġact ress +Ġd olor +å½ ķ +Ne ed +.t oggle +ĠR ace +w ers +.m aterial +ĠD ue +ĠP el +# print +Ġindepend ence +ex us +Sh adow +Ġenc oder +( level +ĠSw ift +.d oc +_se lection +Ġserial VersionUID +Label s +Ġperform ances +.T ag +ĠN HL +iz en +/ UIKit +_CONT ROL +Ġearn ings +ĠAl t +_H ANDLE +C tx +Ġpers u +Ġtr an +ç ¨ +_CH ANNEL +Ġsatisf action +ĠG P +io x +m itt +land o +Ġp ig +inal s +ê ncia +S urface +ĠU UID +Ġbenef icial +Ġsequ ences +ĉmem set +Ġmag ical + « +Ġw orn +AS C +pop up +COM P +_b efore +en ess +U i +L es +.re quire +.Serial izable +add Gap +Ġauthor ization +.py plot +urr ay +lat itude +fr ames +aj s +Ġcomp ass +Ġobserv ations +_s up +.en viron +Ġtri ple +ĠRub y +Ġdr ain +_F ILTER +S an +UM P +Null Exception +ĠG ab +ow e +ĠTurk ish +_se quence +ĠGr ant +uel a +Ġw o +Ġc ube +i q +Ġdis orders +Ġextra ordinary +Ġc trl +ĠSe q +ent r +Ġsan ctions +uts ch +Re ports +Ġin herit +Per iod +Ġphot ography +ĠF ramework +Ġspecial ist +Ġ? ĊĊ +_ selected +.P layer +Ġal location +( account +Ġstruct ural +v able +- offset +.App CompatActivity +аР¼ +.Add WithValue +Ġicon s +Ġshut down +_l ow +ĠCom pare +ĠC e += head +l am +.p redict +_DE C +ĠS leep +ĠGr atis +Ġsuggest ion +ĠD EL +ca ff +av irus +No thing +ŀ ĭ +Ġwides pread +Ġmechan isms +Ġtext Align +occ up +ĠR ail +: NS +Ġf iber +Ġm k +Ġv intage +-l ong +.re duce +. Entities +( record +Ġple asant +FR ING +.C ells +OT T +ĉelse if +_con firm +ĠView Group +s ym +Ġpr ay +Ġsus pected +Cont ains +Ġb orders +Ġcomponent Did +ASS ERT +Ġinf inite +- order +Ġh ello +ĠGr ade +.currentTime Millis +apol is +z h +ĉ Object +: \\ +H O +val uation +Ġvoc ab +Ġcou pon +atab ases +.Get Type +L earn +] =" +ĠG ary +ot ive +Ġas h +Ġb ib +XX XX +Ġbal anced +VAL UE +ĠN at +_A d +< E +åĮ º +ĠMethod Info +L IB +Ġconsider able +ĠInd ustry +test s +.set Title +ĠBl uetooth +Ġm apped +ĠBru ce +ĠMain Window +ĉ status +Ġr az +ĠM and +Ġclass ification +Per missions +Ġ---------------------------------------------------------------- ------------ +Ġcontain ers +: set +_x ml +Ġwh ilst +Th rough +Ġval ign +Ġworld s +C ORD +ED IA +ÑĢ Ð¾Ð² +Ġsp are +ĠH ad +ĠDE F +(p tr +Ġwarm ing +ठ¾ +Ġcons ensus +ag ne +CT L +Ġì ķ +.M ain +web Element +Ġp ist +Fl ash +App end +.tw img +T ap +Ġveget ables +al g +.s ample +Ġcoach ing +( ind +Cell Value +Check Box +ĠH ell +RO OT +Ġst adium +Ġinvestig ating +) % +st ed +ĠW riting +Ġê ² +Ġun o +Ġ{{ -- +Ġco ords +Ġun ser +organ ization +ĠCr ime +ĠDemocr at +Ġv in +/ file +- api +ĠA y +Ġfund ed +ĠBre xit +ĠG h +ent ina +c ases +Ġd ash +Ġ!! }Ċ +H I +Off ice +Ġcapt ain +Ġwor ship +\ C +Ġglo be +_ board +Ġbab ies +Ġconsec utive +Ġenh anced +ere um +ĠAd vis +Ġgr ain +Ġc raw +ancell ationToken +. alpha +_W ITH +ĠO tt +ĠC ool +.b atch +Ġver ified +(c allback +Ġreg ards +ĠInt Ptr +ouch er +Ġk in +Ġtou ched +it Ãł +ath on +Ġadj acent +Ġaccom panied +LE AR +Ġim plies +Ġh ill +ĠBalt imore +=" - +Fin ally +S am +ic opt +Ġs od +Ġm aj +ĠSh ipping +Ġget All +Ġcoach es +Ġdon ations +il ot +ĠT ar +c err +Ġbad ge +Ġmark ers +ĠR and +ais ed +iss ance +Ġexpl oring +uc ed +ĠIndones ia +Ġbene ath +Ġmagn etic +Ġm useum +match Condition +Ġdis rupt +Ġrem ind +ĠT M +Ġ/ >< +Ġf ool +Ġes k +.N ull +ĠD ies +_OUT PUT +_TYP ED +Ġpaint ed +Ġsoph istic +ĠB ear +* n +_P ACK +Ġdeliver ing +ĠC OUNT +åį ķ +Ġj eg +-c ar +f name +Ġr anging +ĠN eg +/ ******/ +ĠCH AR +Ġul tra +Gr ad += t +Ġjud ges +ĠD ise +ann ers +Ġsc al +_c al +ĠCON NECTION +_ embed +(f n +ĠC raft +ĠP as +") -> +.con vert +.res ource +ĠST ATUS +ô ng +ĠT it +Ġclass room +ĠArch itect +ĠK ings +Ġstead y +/* !Ċ +ĠG ene +) ";Ċ +ic ia +st an +ĠCon struction +um per +w c +ĠC BS +ing ing +-p arty +(d river +M ARK +Ġn ested +ew ard +Ġdepend ency +Ġm ales +ĠO NE +ĠProdu ction +][ $ +ãĥ¼ ãĥ +_LO AD +ĠB ol +el ry +ł éϤ +ĠRe quire +Ġpl acing +xx x +CA LE +Ġth umb +Ch oose +Ġprot otype +VO ID +Ġles bian +Ġtra its +Sh arp +Ġconsum e +Tr uth +Ġaction Performed +ĠEnvironment al +ĠDe an +Ġest ado +s ame +Ġnumer ic +Ġtrans it +. Email +-s ide +_R UN +ĠVill age +_OP EN +è ¦ +.re m +-w arning +any a +Property Changed +Ġ(! _ +( check +il ia +ĠSo ft +st eps +ĠMad rid +Memory Warning +Ġhand lers +Ġexperi encing +Ġins pect +button s +Receive MemoryWarning +chem y +Link s +Ġur llib +.System Colors +ĠE igen +Ġpun ishment +:UI Control +bar a +- set +Ġ}čĊčĊ čĊ +Ġtoler ance +Ġinter faces +. redirect +ighb ors +cs rf +_back ground +. Utils +_H T +ĠInter est +im os +Ġgr ants +Ġexam ined +Ð Ķ +Ġc f +for ge +back s +ĠObject s +_s ent +. entry +ĠTH EN +ell ido +c ia +, res +/std c +. nd +( Int +ĠAuth ors +ĠApp CompatActivity +' { +Ġmed i +M usic +ig m +ce ipt +Ġa uss +Ġtarget ing +ĠKe ys +h n +: ]Ċ +Ġmin eral +à ® +.c a +om ed +Ġshe ets +Ġc amb +Ġdead ly +.in ject +( unit +ĠSe lection +.g ms +( connection +Ġ$ (" +é mon +ĠCurrent ly +pt e +_path s +le af +Ġimp lications +pos al +ä½ į +[ / +anc ia +é Ľ +m ul +c ie +Ġge ile +im als +UI View +Ġs urre +serial ize +IS O +Ġarbit rary +Ġsock addr +.f n +ĠM erc +Ġcast ing +Key Down +Ġnew Value +op ens +T odo +Ġflex ibility +ĉĉĉĉ ĠĠ +V elocity +ú n +row ing +Ġcomput ed +` )Ċ +st atement +Ġr i +_c art +L ow +trans fer +.n av +Ġgr ave +ĠDo or +ĉ alert +.sub scribe +- profile +ĉb ase +ĠâĪ Ĵ +__ ĊĊ +Ġengine ers +Ġexplos ion +Ġd ari +ĉ Log +on al +Ġisol ated +{ i +ĠM sg +F uture +Ġrac ist +-w rap +ĠV ers +b org +IS ION +Ġ ÑĢаР+ĠY an +init With +Ġn omin +( empty +ÃŃ n +ãĤ ¤ +ĉ width +Ġch amber +/ ajax +EM P +Ġnec es +iv os +log ic +*) & +cript s +Row At +ib lings +Ġe ars +Ġcomput ing +Ġm aker +ĠNe ither +b readcrumb +Ġserial ize +ĠWith in +Ġd ell +_TR ACE += a +Ġwish es +-in ch +ĠD or +Ġinnoc ent +ĠD ol +Ġint ens +for ced +ĠB IT +Ġphotograph s +Ġcas a +ĠL en +\F ramework +.S imple +Ġde ar +)/ ( +ip pi +Ġown s +Pl ayers +Ġpropos als +.p i +us alem +D amage +Ġcal ories +ĠCreat ive +Ġ[ $ +Ġ// čĊ +And View +è me +.c ustom +_f actory +command s +_lo ok +Ġstr cmp +Y N +a ired +Ġaud it +о ÑģÑĤ +ĠRe verse +ropri ate +et ics +< vector +.s elenium +. or +Ġpred icate +Ġfinish ing +Ġk le +ĠRep os +ĠK han +ĠM aking +ĠF S +Ġp ute +ĉ state +_S UPPORT +' - +orient ation +Ġexist ed +atur a +Ġexpect s +ĠSh adow +Ġorgan iz +å ŀĭ +Ġsusp ension +Ġu it +Ġsimult aneously +ĠAff ero +: ");Ċ +Ġro cket +c as +eter mine +ace ut +x l +ĠA MD +( graph +ass oci +_C R +.ar ange +(j Label +Ġbe ef +Qu ick +.c ard +] ): +- gr +.G ONE +_C LOSE +ĠNe v +ÃŃ as +Ġste pped +ĠFre edom +ĠW R +NS Array +_r x +_d ialog +Ġhot els +Ġ( \< +ĠD iamond +Ġassum ption +um i +( items +č ččĊ +æ³ ķ +Ġn el +Book s +åİ ¿ +us b +ĠF IN +æ ¬ +Ġcorpor ations +US A +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +.p roperty +ew ise +_ plot +"> ';Ċ +Ġpe pper +Ġsh ed +ĠMed ium +ĠC ookie +Ġoverse as +ed or +asure ment +åŃ ĺ +Ġ' .' +Ġph p +ĠPRO C +Ġexception al +( th +ĠJ et +Ġoccup ied +.set Image +ĠRel ated +uck er +M embers +PR INT +ĠG lo +_V IEW +} ",Ċ +Ġad option +[] )Ċ +ĠMiss ouri +ĠLin coln +eral d +Pop up +Ġf ate +- bootstrap +fe ctions +ĠP oll +_ARG S +in ance +-h ome +. ), +_d one +: ĊĊĊ +Ġdiscuss ing +ĠSQL Exception +Ġelect ro +ĉ req +Ġz w +Ġl ui +Ġover night +$ user +ĠW AY +Ġall erg +Ġdisappoint ed +Ġradi ation +Ġimpress ed +ific ates +Ġto b +CL ASS +Ġc uda +_d et +- post +ul u +Trans lation +-h and +.y ear +ĠM ongo +Ġun clear +. engine +WEB PACK +r ices +_AC CESS +Ġh olidays +per cent +.Id entity +ĠG ov +Ġpassion ate +!! . +ĠGree ce +plus plus +')) ; +G P +Ġexc it +.tab Page +_ cond +Ġspons or +M ODULE +_pro c +Ġ$ Ċ +Ġr ational +.T ool +Ġi hr +cc a +åĵ ģ +ĠE state +IB UTE +Action Performed +ĠS olar +¦ Ĥ +Ġequ ity +t id +Ġrec ip +.s imple +m k +ĠL uke +ĠGuard ian +Ġenc rypted +Ġdomin ant +. place +ĠN V +Ġtong ue +( Get +Ġst ainless +.P lay +Ġe b +ac i +.b uffer +readcr umbs +Ġvacc ine +p rom +Ġuser Info +Ġsl ug +Serial izedName +-w ide +Ġre actions +ĠY ang +ĠAdd s +(user Id +Ġpl ates +ĠM EM +Ġb ail +In side +et ed +Ġels if +Ġs ake +Ġc ycles +Ġì Ĺ +ĉ I +-c ollapse +ĠG MT +De claration +Ġg ros +Ġreach es +Ġcust ody +Unt il +t u +ĠCh en +Ġn x +( addr +ĠO ffer +Ġcol leg +ass ador +Ġm apper +ĠS IGNAL +ĠB loom +ĠH oll +ĠIm per +-d es +_s ite +Pro c +E qu +Ġat omic +ĠW oman +s ent +sc ar +Ġint elligent +ĠGet ting +ĠReg istration +ĠPh ill +Ġkill er +unic ode +Ċ ĉĉĊ +ĠJac ob +ĠCon st +Ġloc ate +Ġca us +ĠSch olar +Ġconstitution al +Ġinfl ation +ĠG ot += array +end um +Ġtransl ated +Ġdiv orce +En tries +Ġs or +ĠQu ote +irl ines +U K +Ġexc el +( opt +ĠAD V +,: , +Ġcontact ed +ĠD A +Ġr ings +ĠIndust rial +.get Context +Ġforg otten +ĠT an +Ġp ants +Ġo v +Ġdec oder +ĠPart ial +Ġv c +Ġbatt les +A rial +FRING EMENT +ir ates +, w +aint enance +ĠO d +ĠTechn ologies +åī į +ĠCar ter +.find All +N ome +B en +ĠUs age +ĠP icture +Ġbad ly +_p anel +Ġpat ent +ĠProt ocol +lot te +ĉ player +je ctions +Ġd ou +_re lease +urn iture +_t ax +ĠF ields +.d ataset +_m aster +CLU DE +ĠPh arm +b st +Ġoper ational +.c ell +Ġident ifying +Ġj wt +t uple +ĠT C +ĠC ro +ix map +- components +gener al +Ġo z +_D e +_d ouble +ĠTo o +.View Group +g ate +d ings +ph otos +Ġgrand e +ol lect +_l in +Ġaw ful +f ilters +Ġaltern ate +es p +Ġcomp ress +e o +ĠS cale +Ġind irect +Ġinv oice +ĊĊĊĊĊĊĊĊ ĊĊĊĊĊĊĊĊ +Start ing +ĠPl ayers +ie le +. then +Or d +ĠT uple +Ġb out +ĠStat istics +Pre view +Ġp uzzle +ĠW idth +ST ATE +Ġover lay +ĉ on +Ġin fr +Ġsm allest +lock ed +ÑĤ о +ss l +Ġde emed +Ġs co +re ck +Ġj Button +Ġmiss ions +ç§ ° +.Selected Index +T ABLE +Se pt +Ġacknow ledge +Ġstrt otime +ĠT ell +ĠD ak +Ġal uminum +Ġf ence +ĠSt ars +CON FIG +Ġretro fit +Ġemph asis +/ header +ĠS omething +in ished +=' ".$ +ĠValid ators +Ġpol ar +section s +.as px +Ġas pir +.M ock +Code Gen +Ġpe ut +Ġaccept ing +Ġback ing +P icture +/ ap +еР³ +_SE C +- use +annot ation +Ġcogn itive +Ġg rip +h our +ĠLeg al +Ġep ic +.t oolStrip +.not ify +.L ast +OR IZ +M iddleware +cri ptions +l ash +_F OUND +ĠLiver pool +Ġ{} ", +Inst all +Ġn it +Ġfig ured +[ len +.W in +.pl atform +Ġgam bling +(d t +av ery +ĉ include +Wh ether +R outing +Ġther ap +Rem ote +ĠL oss +y ll +Ġappro ached +ĠV ehicle +ĠAl pha +Ġvoc ê +ans wers +NS Dictionary +cons ider +un used +ĠF an +or able +f re +ĠDIS CLAIM +ĠAct or +. ] +to Have +.user Id +Ġspeed s +ew ay +Ġrec urs +ĠÐ ³ +_pr iv +! âĢĿĊĊ +Ch oice +Ġsett le +Ġplan es +' }, +T om +IT ER +! "Ċ +å » +achel or +Ġsepar ation +Ġd al +ad j +Ġreg isters +r iz +ĠNot ice +Ġl u +Ġcour age +Ġax es +cell ent +.as ync +Ġcompat ibility +ç « +Ġ! ĊĊ +ĉ title +Y LE +ĉ message +U UID +OLD ER +ĠH H +ĠStyle Sheet +Ġaccess ed +. validation +t asks +Ġpoll ution +.c anvas +Ġing redient +ĠC abin +A h +old own +ĠNO I +ĠÃ Ĺ +[ f +ed uc +y alty +(n ot +_ State +am en +Ġda o +ud ad +ell ers +} & +lic ity +_W INDOW +Ġt atto +val or +.R ange +Ġrefer enced +ĠRes erve +M oney +SCRI PT +/ product +cho ices +Ġt in +ãĤ ĵ +Ġsepar ator +Ġp kg +am med +ĠM AT +! !ĊĊ +Ġr aid +Ġmotiv ation +ĠX P +ĠBack ground +ĠQu aternion +.define Property +ik er +ĉp arent +ĠOrigin ally +ant age +ĠH ans +Ġtim eline +.c ur +op ic +ĠSe qu +m ust +ĠCo al +Ġform atter +_R GB +Ġ_ (" +'} ),Ċ +Ġ= ================ +ĠF UNCTION +Ġl ng +ic ates +l ive +_ engine +Ġtown s +')) ĊĊ +ĠP K +( api +ĉs canf +pack et +.ph one +á Ģ +ĠAnd y +_N AMES +PL Y +Ġmin s +im i +Ġbr ick +Ġbl ade +.std out +}` ;Ċ +Sh ift +ĉs b +ĠCheck s +Ġphenomen on +Av atar +Ġmin istry +ro se +ĉ File +Ġtit led +( LOG +Ġg an +des ign +(), čĊ +Ġb ones +st m +ÅĽ Äĩ +ĠInput Stream +Ġvol unt +ĠSerial izable +Ġfight er +ĠDr ag +T witter +Ġsubs id +ç ¼ +Ġfor ums +.load ing +log ged +_ this +Ġterr ain +Ġir re +ĠIn g +ĠC N +_object s +. uid +Ġconscious ness +T INGS +ĠG all +Ġport ray +ĠDevelop er +Ġparticip ant +Ġ" ;čĊ +/ model +ĠOper ations +^ \ +ĠL ater +Ġrais es +-n one +.m eta +=' .$ +Fin ished +Ġrepl acing +Ġsam pling +ĠJ en +" There +RE AL +A LE +ìĬ ¤ +Or ders +_param eter +ĠOlymp ic +Ġtr ès +Ġare na +i ol +; ?> +Ġimpact s +ĠW S +: get +Ġfl ights +ĠRuss ell +c amera +F n +s igma +Ġfor cing +Ġloc als +Ġdepart ure +Ġcelebr ation +ĠS ay +ï¼ Ĵ +ĠH ills +.has OwnProperty +Ġtyp ings +.A PI +Ġdon ation +Operation Exception +.Act ivity +c plusplus +ĠChar lie +Ġimport ed +Ġd ann +Ġoccas ions +Ġimplement ing +Ġpur ple +.d ialog +SQL Exception +ern o +Ġw ars +Ġpast e +Ġdecre ased +Ġhar sh +Ġel abor +input s +ĠView s +Ġerror Message +_m ul +ĉ write +ĠC op +ĠAnn ual +(b utton +Ġv ida +b ars +ĠHar vard +ĉex pect +Ġindex es +Ġdocument ary +Ġf lesh +OR LD +ĠD elta +M AND +Br ush +-c olumn +Ġdevelop ments +method Visitor +s lice +ĠP DO +Ġinvest ing +ir able +Ġxml ns +ï¼ Ľ +art a +Ġthe ories +_c ity +Ġ$ __ +Cre ating +( pr +D ropdown +ism atch +ĠN ET +'] )){Ċ +ĠVal ues +ĠSE O +ĠST AT +Ġe cosystem +Ġtem pt +Ġ\ \ +Ġ// {Ċ +ĠChrist opher +ĠKent ucky +ĠHttp ServletResponse +Ġhy brid +y on +Ġfeed ing +ĠEx tra +N orm +IT CH +ĠSe an +ĠUp load +m un +p ur +Ġp ersistent +ĠID C +ĠPer form +.m erge +_ room +Mean while +! =' +ĠW el +Args Constructor +.D atabase +Ġcount ing +() * +Ķ åĽŀ +ĠT OP +m ill +ĠD T +IGN ED +ĠK B +Ġcomp ly +S outh +_c ollection +Ch apter +Ġexpl aining +_ AM +_t s +c ards +Ġqu el +Ġp ole +Ġtouch down +ĠO thers +Ġpe ers +ĠType Error +Ġsix th +Ġche er +Ġdis pute +us c +) ], +th umb +Ġh iding +ĠS IG +lik es +ĠP AGE +.Ref lection +Ġhead quarters +T ING +ĠG host +M LE +$ Ċ +Ġcontr ary +ext end +'] ). +FF ECT +ĠP interest +úmer o +ric ane +ĉs ession +Ġcr ystal +- Control +overn ment +og raf +- action +v olume +ft en +Ġun con +Ġan imate +Ġle ase +sc r +Ġref use +ãĢ ĭ +ft p +in formation +Ġeval uated +Ġin jection +Ġj ack +Ġwork shop +æ³ ¨ +PT H +ĠT s +off er +ĉ os +Ġking dom +M issing +Ġlaw makers +ext Field +Ġsing ing +ab i +/ client +.m edia +ATEG ORY +Sign ature +% ',Ċ +ĠF uck +][ : +Ġsens ors +/ com +ĠPr imary +.S QL +_pro gram +Ġp ills +Ġinteg ral +Ġfle et +Ġdro pping +.s l +Be en +Ġp ets +Ġadvis ed +Ġdr agon +_ EDIT +( im +F ER +ĠDr ug +(r andom +Ġcomp ression +ou st +[ % +Ġbuy er +h op +R oles +man age +Ġpain ful +ĠBr anch +-mod al +en ant +ĠM esh +/ font +ĠG raham +Ġâ ĺ +Ġn c +ĠFranc is +Ġspec ification +Ġdam ages +- config +Ġthe oret +sec ure +_m ulti +aceut ical +Ġdemand ing +en ne +IST S +() ));ĊĊ +Re ason +Re cent +ph ase +Ġps y +_M AN +Ġvolunte er +å ¿ +istrib uted +li o +Ġproduct ivity +_com m +S pring +n is +. weight +ĠC ancer +Al loc +ĠT weet +Ġsepar ately +ĉ check +_p roperties +. Unit +_CL K +Ġg t +Ġ( );ĊĊ +Ġhand y +ĠThom pson +Ġunn ecessary +ĠRe ader +G N += request +ĠU tility +.Re pository +ĠA x +hy dr +ie u +Ġth y +Ġl t +_m ail +ä¿® æĶ¹ +ail and +ĠPhil ip +Ġbit ter +Ġbet ting +Ġtim ed +ock s +' a +Ġal gorithms +Ġre interpret +Ġto ss +ro gen +Ġhop ed +( selected +Ġvent ure +TE X +ĠLe ave +.Sub string +Ġgr ateful +uk a +ĠCon sumer +Ġag greg +C ircle +ภģ +_block s +Ġleg ally +Ġ" | +ãĥ ĥ +. board +.A b +Function s +rec ipe +è ĩ +ĠO xford +Ġwho les +.B uild +_ch anged +h ai +Ġdepart ments +I mp +Ġcoal ition +IN FRINGEMENT +Ġemp ower +itch es +N orth +Ġinfl amm +ON SE +Ġmiss ile +ĠR aj +ĠIss ue +Ġat oi +ca led +.Cont rollers +ĠW olf +Ġcrush ers +á» ĩ +.A uth +.add Attribute +h is +Ġbo ots +.c lean +c amp +Ġten ant +Ġt une +Ġ{} '. +Ġwork out +Re po +Ġpartial ly +MI SSION +j amin +ĠS B +Ġdetermin ation +Ġ' ');Ċ +ĠB eng +Ġv os +Ġin hab +/ lang +s burgh +Exec utor +h one +ĠCh allenge +_link s +.Le vel +Ġunder ground +-c ode +Ġoptim ization +log ging +_de st +Ġsn ake +Ġchemical s +_IMPORT ED +ado op +ĠTH AT +man aged +Ġredu ces +ĠRE AL +ĠG uy +_GENER IC +/ ******************************** +. amount +Ġd ere +get Time +Ġp ant +an onymous +Ġharmon y +ĠAl an +Ġscen arios +Ġd irt +ht ags +M c +Sh ell +r in +{ čĊčĊ +.p ow +ĉ client +Ġconspir acy +Ġad mission +ĠReg ional +ĠView Controller +ĠPhilipp ines +Ġde pos +Ġp ap +ĠP ad +P aul +.Com boBox +Ġt utor +ĠRec ipe +w riting +Ġcontrib utor +OT H +Sm all +V I +Ġh acer +e qu +ĠEx amples +h uman +.m essages +ĉt yp +Ġ( čĊ +ĠS SL +LE N +ĠRom ney +( grid +ĉ min +Ġ> ĊĊ +Ġfr uits +Ġvot er +In line +pan e +ĠC ollections +char set +Ġsp am +z b +item ap +Ġsucceed ed +_C OL +Ġel apsed +im eter +Ġrecover ed +T ensor +hatt an +.set up +ist o +( head +ĠS IZE +Ġtact ics +Ġdist ur +Ġpre val +ici os +( Value +_c ols +ĠF at +Ġse al +Ġs ons +Ġens ures +Ġpress ing += & +igen ous +Ġharass ment +_ JSON +Ġign or +yn omial +om er +_st atic +Ġsignific ance +Ġcirc les +_S ystem +Ġdiscipl ine +Ġdress ed +Ġs phere +Ġclim b +_ actions +ĠB ab +Ġ' =', +_s chema +" use +Ġund ers +Ġc ups +.s creen +/ new +Ġappe aring +T OP +vis ed +cl ang +Ġinvestig ators +Ġmyster ious +Ġprom ising +Ġqual ify +Ġc ave +Ġequ ip += x +G T +( link +. velocity +. erase +ot er +++++ ++++ +pro fit +Ġz ones +_ uid +- ser +Ġobject ives +Ġmil f +web kit +(m atch +ne h +ĠAssoci ated +ĠT odo += d +C am +Ġv ocal +Ġs udo +( EX +Ġtr ou +AB C +.b ean +ĠG round +ĠRE ST +we ets +In g +im on +_b us +ĠC OLOR +un to +Ġf oss +ĠLink s +ä ng +/ forms +pr ises +Ġachie vement +C ALL +ел ÑĮ +ĠVer ify +_S OURCE +apt cha +ID D +_re ference +G old +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĊ +Re ceiver +Ġa j +_d irection +} ] +ĠCom pet +Ġb ang +ĠC ass +- url +te chn +ĠJer usalem +long itude +' );čĊčĊ +Ġwin ners +T asks +ĠD MA +Ġtool tip +İ · +ĠB ra +_d uration +cur y +parent s +---- >( +ĠK ir +Ġint ros +Ġsk etch +Ġsk illed +Ġim mer +Ġade quate +_re p +( header +_ like +Ġper ceived +ss h +Ġassum ing +Ġf f +_u uid +ul as +Ġdemocr atic +. entities +S eries +aph ore +Ġnew er +} ( +SE C +ai ro +Ġcomm od +Ġprivile ge +Ġde ux +ĠH op +.' / +ct ic +. ';Ċ + C +ĠWar ren +Ġoptim izer +ĠSER VICES +_ oper +get Attribute +ĠMc K +_s elf +.r s +" )ĊĊĊ +Get Component +er ce +Ġt ous +un its +'] );čĊ +Z oom +/ E +Ġobs c +Ġfast est +on line +Ġpeace ful +ff en +Ġc argo +ĉ pr +Ġseek s +z u +Tr im +Ġw ard +Ġver d +Ġblog s +.exception s +ĠPrem ium +ĠN etherlands +S afe +Fin ish +ĠAl bum +_A CC += this +v irtual +] > +_L ABEL +ĠN ich +_w in +ĠA aron +W P +; $ +aim s +ĠImage View +Ġend less +ER A +_DIS ABLE +Ġcancel led +- us +Ġins pection +em in +ĠG rey +- open +Ġiter ations +. owner +Ġk eras +.P assword +ĠR y +ĠIN S +A ir +ĠSe veral +.Tab Stop +ING LE +ĠH air +ĠCan vas +AA AA +Ġfl aw +ced es +.Re port +í Ĭ +ĠT ips +cript ors +.trans action +.S pring +Ġview er +Ġins ights +è¾ ĵ +ord ion +U INT +se ek +ĠA uf +ìŀ IJ +Ġstr ain +To oltip +Ġd z +ign al +ad t +Ġu c +fin ite +Ġn m +.c md +ĠMy Sql +[ data +.j ackson +.t ree +Request Param +_ agent +") ]čĊ +Ġass ass +( Constants +: ss +ĠM AN ++- +- +ĠB ottom +print s +ĠS ame +@ Autowired +sw ap +ici ón +Ġprotest ers +Ġh oney +ĠV eter +(C alendar +- ad +ĠBrook lyn +L ife +_V AR +ze ch +ĠC ALL +_C AST +ĠE lection +Ġthick ness +V ery +_IN TEGER +- dev +)) )) +ap at +oo oo +d emo +Ġparse Float +ĠR ather +ST IT +m aker +[ current +chron o +Ġch rist +ãģ ª +ĠD etail +ư á» +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġs ul +id ency +Q ue +Ġeleg ant +ap ons +Ġdish es +Ġinteg ers +( read +find ViewById +ĠAm ount +ĠSk ip +Ġhab its +* )( +Ġmon sters +M AC +: end +Ġfr ank +As sembly +Ġd fs +Ġne ut +_TYP ES +e qual +loy d +( uri +Ġch i +Ġdefend ant +Ġconflic ts +Ġv il +- js +ĠPe ace +Ġmut able +) sender +ĠF ocus +å» º +Ġapprec iated +s leep +ĠR ED +C ulture +Ġdesign ers +_g enerator +c odes +/ ex +.Get Value +umb led +.scal ajs +per or +Ġveter ans +Ġ} )čĊ +Ġun fortunately +_C REATE +M ass +ĠCL AIM +ĠMe et +_s upport +B ank +() .Ċ +D ark +_LO W +ĠMin ing +ĠO wner +ier a +Client e +Ġencour aging +> S +Ġboy friend +ĠH alf +ĠA CC +A ff +_ ar +-l ife +c x +.J Button +iz ado +.z ero +.open qa +ot on +.text Content +Ġto ll +at ie +Ġball ot +- number +. Exception +ĉ params +c ircle +-m ap +Ġn ap +ĠRob ot +ĠI ch +reg istration +Am azon +roll ment +( exp +Ġt anks +ĠG ordon +Ġmach inery +Ġbas eline +æ ĭ +Ø © +ĠCon vention +ĉ config +ook ies +m ult +Rec ords +ĠE ST +Ġgar bage +Ġcon form +id al +Ġb arg +Ġsurv ived +Ġinvestig ations +.contains Key +---------------------------------------------------------------- ----------Ċ +ort ion +Ġhor r +_ http +Ġm ant +] ;čĊčĊ +b inary +em pl +Ġin quiry +ĠMean while +Ġcollect ing +.Entity Framework +", ĊĊ +ĠP ic +@ Inject +ick ness +ĠB inding +Ġcont rolling +re verse +Ġch airs +semb led +( add +Dis abled +an as +.trans late +-------- ---Ċ +Ġref lected +"] ĊĊ +Ex ternal +Ar row +Single ton +% x +Ġ Å +Ġan cest +ĠOr leans +ĉc md +Ġprohib ited +ith metic +(ch annel +_c ss +For ward +.s ocket +Ġl uc +â Ĩ +ĠFire fox +ĠM ovies +) _ +. ends +( shape +Ġde alt +Ġs aves +Ġgl ory +Ġmej or +Ġbreath ing +Ġ eller +get Data +Ġang les +Ġtool bar +Ġsp acing +IP S +Ġflo ors +_ACT IVE +Ġsh uffle +/ shared +ĠE le +ed ish +Ġweb cam +.ex pect +il oc +ĠIn cludes +Ġtweet ed +Ġ: ) +ĠEss ay +F ix +-b etween +_ web +.con v +Ġrac ism +Ġreflect s +um m +иÑĤ е +_f ooter +/d ocs +ĠP our +Ng Module +.initial ize +pattern s +_ In +ĠAb b +* čĊ +Ġsent iment +b uff +_count s +Ġre use +ch unk +Ġim posed +Primary Key +Fore ground +Ġconsum ed +? ! +Ġd ick +Ġch ron +ĠF ern +Ġrespons ive +Ġin sect +icult y +Ġr w +Ġal ike +Ġsub set +ĠCook ies +ĠP air +Ġt ier +IF O +av our +ĠQ U +, sizeof +Ġmerg ed +m v +it ol +yl on +Ġjump ed +. role +ens aje +R ules +Ġb rowse +An imator +Ġy oga +Ġvari ants +Ġcour tesy +ur an +p bs +else if +Al t +ĠL ane +CL K +IM ARY +_PRO PERTY +ï¼ IJ +Ġch an +Ġgrad ually +Ġsh ake +Ġbl onde +... ");Ċ +-se x +Ġgame play +ac ies +.ref resh +US B +ĠPl ot +W as +iss ippi +ĠT ensor +Ġcryptoc urrency +Ġdifficult ies +De leted +With out +_ append +_ ver +")) čĊ +Ġhonest ly +Ġp ivot +Ġtem ps +_p s +ĠUn like +[: - +V S +_in f +Ġjun ior +Ġanim ations +Ġfile path +? {{ $ +Ġun icode +pl aces +ĠC offee +.S E +ĠP AR +(t xt +ge bra +Ġf ires +Main Window +med ium +Ġ( âĢľ +Ġl g +Ġc mp +/ base +_l ayers +_ entries +Ġadmin ister +ĠSU CH +B P +ĠScott ish +ĉčĊ ĉčĊ +gu ard +ĠStr ong +In sn +ĠC AP +as ury +ĠSE E +C lock +er ie +\ models +Ġ$ $ +ĠC ab +Ġwur de +Ġsold ier +Ġcl ips +Ġarrang ement +ĠW onder +ĠH orn +Ġsc ared +Ġc ure +m kdir +Ġal igned +ĠP ink +Ġland ed +Dim ension +Scroll Pane +.ch at +.W ith +ĠTr ain +] .Ċ +Ġth irty +Ġdur able +Ġl d +Ġlate init +Ġch arts +Ġins ult +.F atal +_ ct +Ġm asks +CLU DED +Pres ident +Ġcol ours +g ments +.at tributes +ĠF lex +ĠC lock +ÃŃ cul +im en +J O +ĠReg ex +_L INK +Ġc ouch +ĠIN PUT +Ġbe ating +b usiness +pre ced +. unit +ĠF el +N ever +osp el +.start swith +ĠE PA +. only +Ġprevent ing +y er +Column Name +Ġelev ation +fl u +icy cle +Ġoff line +Tool bar +Ġcompet ing +) ]. +Ġm og +Ġis Valid +As k +_ av +_l at +AN C +ĠJ oh +k ers +Ġgu ards +Ġch ains +ĠSimple DateFormat +.st atic +Ġvess el +Ġm ud +Ġst abil +Ġst ret +g m +am ation +ç ľ +-w ith +Ġro s +_P A +Ġresult ado +Ġconf idential +ĠTok yo +ĉ using +ĠMath f +omb ine +ĠESP N +Ġdeal ers +Ġdismiss ed +TR Y +Ġte ens +rec ords +Ġw ings +g allery +account s +_L IB +Ġj acket +ĠNS Object +Ġst ones +ĠDel ivery +ĠD iet +/w atch +Ġto ilet +ĠG uest +.d ay +Ġint val +Vis it +Ġinvestig ated +Ġpent ru +ĠThe atre +andid ates +L ang +ĠS erv +Ġcont rollers +Ġset Title +N P +am y +fl at +( ui +_d ocument +è ĥ½ +ĠC oin +ĠAd ams +pt ic +Ġproduct ive +Ġaccompl ished +čĊčĊ čĊčĊ +Ġdefer red +ient es +Ġs inc +ol ars +Right arrow +Ġvari ations +( offset +.Layout Inflater +Ġsus pend +Ġprevent ion +_pr ivate +_ js +âĺ ħ +Ġw ieder +at um +Ĵ Į +Ġappear ances +.D ocument +Ġvalid ates +cal endar +} ";Ċ +.d emo +con ut +Ġcorre ction +ĠDe al +Ġbatter ies +.d uration +, \ +_m arker +m ulti +Ġh alt +Ġc ms +Ġsh aped +B ro +re duce +Ġ #### +CT OR +ĠBen ef +Ġicon ic +Ġp iano +Ġeffect iveness +| .Ċ +Ġa jax +Ġv olumes +ภ¡ +Ġcl js +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠ Ċ +ath s +ra its +å¤ § +Ñ ĸ +_m ult +Ġfasc inating +A verage +Ġpr é +ĠChair man +.find Element +_p in +Ġcomp aring +Ġdark ness +-F i +- server +Ġselect ing +ster dam +ĠPart s +FORM ATION +Ġnot ing +Ġp ile +og s +Ġpa lette +_d o +it ize +() ( +Ġdef ining +Ġremain der +Un its +_T ASK +Http Client +S ocial +Ġfund ra +N R +ch est +C urrency +.ad apter +Ġd op +un ting +ANG UAGE +" He +ĉ index +_p ackage +.I con +Ġrep et +m ass +=" .$ +ĠS ud +Ġl id +pro vince +ì ľ +G PIO +Ð ļ +ĠMy SQL +Ġdoc s +ĠG A +Ġip sum +K ernel +Ġaccept s +Ġfit ting +Ġcu ando +Ġd uplic +ĠBro ther +ĠK le +num s +Ġmor ph +Ġ ######## +ĠCG Point +< unsigned +ä¾ ĭ +ĠD uke +.set Bounds +q s +or ic +j er +Ġregard ed +Http Request +Ġbond s +Ġthorough ly +enc ent +Ġhighlight ed +Ġac res +Ġwork place +ĠL ux +Ġqu ot +.in flate +Ġdocument ed +Ġadd iction +Ġmut ation +.c ity +Ġbott les +ĠRepos itory +on n +err no +ARI ABLE +åº ¦ +_B EGIN +gl as +' })Ċ +ĠMass age +ĠWh it +reg ex +W A +Ġout let +- head +Ġexp ired +ĠTh ai +/ include +grad ient +scan f +Ġse am +w al +ĉb uf +B earer +Ġprec ious +if acts +co ord +Ġexpl oration +.get Y +(h andle +Top ic +ĠV ent +r hs +---- --Ċ +ĠB right +Ġg uild +m other +st orm +Ġmunicip al +Ġin k +.T YPE +w l +... manual +ĠTechn ical +Ġcorpor ation +ĠH W +ank a +T AIL +ist as +Ġperform s +ĠBeh avior +.F or +_ ORDER +ĠK ick +Ġcallback s +_d r +ue go +h ub +uff icient +sk y +Ġb p +ht able +ĠON LY +ĠAUTH ORS +.Arg ument +" };Ċ +ĠTh under +ĠK om +.Sh ould +A UTH +ah u +_p ayment +Ġst arter +ìĦ ľ +ìļ © +B log +.p atch +Ġgovern ed +ass y +-f ound +Ġthe ater +ĠFont Weight +ĠBat man +" If +.R andom +_d elta +ĠC E +Auth enticated +Ġdr one +Ġc ous +r adius +M er +( None +ĠN J +_ headers +Ġam er +py test +ĠA ctions +ĉĉĉ ĠĠĠĠ +Ġet t +Ġh oly +Ġun comfort +ĠN in +ĠDec imal +ĠM essages +.s ender +] ])Ċ +Ġembr ace +Th ough +/ sp +Ġcult ures +Ġhigh way +t ar +.f ail +_h idden +ĠcomponentDid Mount +ĠW right +Ġj ag +_ il +../../ ../ +ig u +F ood +Ġa ce +Ġa ños +US D +Ġmut ual +Log ic +Ġtem ple +Ġbrief ly +ĠT rip +class method +default s +Ġch unks +,, ,, +ĠRe ason +$ id +-up s +Ġdam n +Ġtruck s +Ġun limited +Ġsc ulpt +ĠC ards +Ġaut or +ĠTest ing +Ġdies e +sh ops +ç ´ +(p ayload +ĠP ATH +ĠMem orial +Ġridic ulous +eg ree +-w inning +Ġre hab +Ġsophistic ated +wp db +ĉ path +! ";Ċ +_S YS +.s peed +Ġso ap +s uffix +W rap +Ġenh ancement +à ī +ú b +Ġplay list +Ġmix ing +ant idad +=" ";Ċ +ĠRev ision +ĠBe at +.in c +-w ay +enc ias +ul ers +C at +id el +ĠSh ip +.set Color +Ġthreat ening +.mod ules +Ġafter wards +ĠD ashboard +Ċ ĠĊ +Sign al +Ġpr imer +orne ys +ici ary +Ġl igne +_p redict +Ġa est +_ https +> : +ĠL ex +Ġrencont res +eg ral +sc ala +_f amily +ÃŁ en +_s ym +Ġuncert ainty +ĠVAL UE +Ġ} ;čĊčĊ +Ġbro ader +Ġh orses +ãģ Ŀ +ĠK al +ob a +_IN ET +ĠK ill +j query +am ination +[ @" +Ġm uj +## #Ċ +First OrDefault +then Return +C he +/ footer +Ġpark s +as je +ĠG ulf +Ġmod est +. Init +ï¼Ł ĊĊ +Ġpros pects +Ġs vg +Ġå ı +.D ialog +_N ET +Ġ( ($ +Ġe k +ĠW arning +ĠM K +< LM +Ġ' čĊ +i em +h etic +Ġi x +th ink +-sh adow +ĠE ld +ĠNev ada +ĠLe af +ĠG ROUP +Ġprom o +ent ine +ĉ Map +ĠModel s +ĠK rist +_k ernel +-m ade +Ġc err +As sets +ell ar +Ġinv oked +.v ue +Ġcult iv +C losed +Ġgener ates +ffff ff +thes ize +s qrt +ĠCast le +.c ar +Ġke en +und a +ĠC row +ĠSing h +y thon +Ġbe ans +l arg +æĸĩ ä»¶ +Aw esome +unc ate +Path s +o ji +(c urr +CON DS +Ġm im +Ġshould ers +H ard +ast es +а еÑĤ +Ġconv ince +de cess +m ade +ĠC MD +. Im +Ġcha os +ens ively +Ġcool ing +Ġbur ied +(' @ +_S e +ĉĉĉĉĉĉĉĉ ĉĉĉĉĉĉĉĉ +.com pany +.sub mit +ph ant +Ġboot strap +_h elp +à § +.d ump +Ġdif er +_m apping +Ġcirc ular +Ġescort s +Ġb ere +Ġgrad u +ĠLeg end +im edia +ĠBar celona +Ġbed s +åĪ ° +ãĢ Ĭ +_v olume +Ġtremend ous +Ġsc aling +Ġp ins +en as +type param +D ashboard +render er +Ġsp i +Ġ& $ +ĠSk in +alm art +Ġh ockey +Ġ'" .$ +Ġerr no +Ġb ew +Follow ing +.M odule +er able +ĠM ilitary +ĠR io +_ available +ĠSur face +Ġst ab +IF IER +ĠL IST +Ġd ashboard +Ġcl usters +.pl ugin +Ġj ou +ĠDec or +F our +Ġdel le +****** /Ċ +ia z +in de +ch ing +Ġget Item +.Add ress +ment ed +A meric +Pl ain +Ġus b +ĠPract ice +_ ment +.bl ue +H int +ÑĢаР² +Ġconn ector +Ġinher ited +и в +Ġinterval s +Ġc ere +Ġu d +Ġin con +.Ex ists +ĠM ic +F K +(c ard +.Set tings +Ġexhib ition +Ġon Pressed +Ġrest ored +eng u +. def +Ġrec v +." );čĊ +enc oder +ather ine +( dest +az ed +# endregion +sem bl +, M +ob y +Ġп еÑĢ +.C all +Ġattend ance +-b order +Ġaddress ing +ê n +ĠLe v +Ġb ash +ben ch +C redentials +Sp acing +( of +_RE SET +ig uous +Ġcr uel +Ġcross ed +Ġle ur +ĠG olf +or rect +Ġpack ets +ĠData Set +Ġpart ly +SEQU ENTIAL +Ġindic ation +ĠS alt +ac ia +Ġ* );Ċ +ĉ info +ĠView Bag +on z +Ġeditor ial +ĠA rena +Ġs ir +_ Static +( socket +s u +cho ose +.m onth +.M y +é ri +; font +do es +Ġcon verter +Ġsal v +Ġl r +Ġinflu enced +(f eature +ĠQue ens +let t +_M ON +& amp +Touch ableOpacity +O FF +Ġmetab ol +( iter +Ġvit amin +ĠIND IRECT +aut om +_p ublic +Ġadjust ment +Ġspecial ized +w indows +.add All +Ġaccording ly +ĠJ OptionPane +Ġcell spacing +Ġqu ad +Ġcre ep +Ġout lets +}` )Ċ +Ġpri est +_TH READ +ĠMar x +ĠBy Val +Ġc ual +éĿ ¢ +Ġtempor arily +An n +ke leton +å ¥ +ĠLO C +au er +der ive +Ġbeh aviors +as ename +ĠCent ury +Ġhor rible +ME SS +_ List +we i +P at +ĠCh oice +_F ROM +ĉ line +.in voke +.B ottom +Ġnow here +." ĊĊĊĊ +_ export +Ġstrugg led +.Ap pearance +ĠJ Button +ĠJer emy +([ [ +Ġkick ed +mar shal +st aff +es ity +Ġqu iz +_e ffect +Ġ} ));ĊĊ +m el +b anner +ĠP IN +Ġin vention +Ġcons olid +Ġop s +ĠB etween +j ack +ern ational +Ġsacr ifice +ag ation +ĠJ oy +Ġam endment +ĠS old +Ġprison ers +ан нÑĭ +Doc uments +) ])Ċ +ust ed +ĠLine arLayout +os o +_E M +.s elf +.M iddle +) // +Ġ\ ' +Ġfuck ed +ĠM urray +Ġprof ound +_E LEMENT +ult a +il ers +port folio +J une +t cp +mod ified +ĠTr ace +ĠK el +aly zer +) => +ĠRep air +_B E +Br and +u art +pre view +Ġiniti atives +run ning +b ang +ĉ update +ĠCo ach +R ich +Ġy outube +Ġrit ual +app a +ĠRobin son +prec ision +//////////////////////////////////////////////////////////////// //////////// +=[ ]Ċ +Ġcelebr ated +OT O +Ġin clusion +J P +' ;čĊčĊ +Ġnot able +(_ . +Man aged +Ġgu ides +& nbsp +ated Route +ĠAd just +Ġcol ored +_s cores +ĠTes la +_pro gress +.in st +[' _ +.fl ags +Ġf close +_O PER +ż y +_n ote +Ġtrans gender +å ķ +RI PT +Ġabs ent +Ġam et +Ġoper and +ë © +Ġh ood +to LowerCase +av o +ĠCirc uit +ĠL ind +-- }}Ċ += m +Ġsup press +ĠM AP +i ang +- admin +Ġside bar +ĠB u +ĠH ex +, F +ĠSign al +Ġtrans parency +ĠFeder ation +/ V +Re q +Ġpul se +Ġt ends +Num bers +% ' +Ġde port +dat as +_U INT +_ tra +ok o +Ġ" ? +comp et +sole te +und ry +Ġover lap +}` ,Ċ +. ly +_sum mary +ĠL ost +.C enter +Ġdis ability +.Serial ization +Ġge om +Ġ? : +ĠW o +Ġsh ipped +Ĥ æķ° +Ġu gly +Ġexcit ement +Ġext erior +Ġcheck out +Ġk ur +, D +ĠAl aska +Ġsyn thetic +ĠB udget +ĠSub scribe +Ġ& Ċ +ÈĻ i +ĠY u +ĉ query +} .Ċ +Ġtr aged +ass en +Ġaccommod ation +Ġphys ician +Ġren amed +Ġtid ak +z Äħ +Ġmin us +ny ch +_EX CEPTION +thread s +Ġt ire +_c reated +ens ure +Ġworth y +Ġexc use +Ġclo th +.parent Node +/pl atform +ĠU FC +ĠG tk +un ny +Ġg ibt +ke ley +h um +(t x +ĉ dev +Ġout fit +do ors +Ġf on +ic ut +vol atile +Ġhom osex +Max imum +Ġexp end +Ġ});ĊĊ Ċ +E q +ond ers +dep artment +ĠPhys ics +" });Ċ +Ġpar ad +.S tr +Ġse le +IF IED +Ġdel ivers +iv an +Ġrespons ibilities +Ġadvoc ates +è µ +ĠR ID +.param eters +M etrics +ron ics +ĠUITableView Cell +A bsolute +ip se +yl um +MLE lement +_VAL ID +< title +D lg +p aces +Ġsynd rome +be ans +_d atabase +oz illa +ĠM eg +DB G +Ġl ub +Bag Constraints +ab ad +Ġproject ed +_BY TE +.Size F +st reet +ĊĊĊĊ ĊĊĊĊĊĊ +ĠLO SS +Ġdirect ors +/ news +Ġnurs ing +ĠD one +. HTTP +dis count +ĠR ot +To Many +Ġen abling +Ġauss i +ost a +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ čĊ +è½ ½ +Ġhel icopt +ĠIn side +ä¿¡ æģ¯ +is per +ĠAll ah +ARCH AR +Ġroll s +Com pare +X P +Index Of +S UM +Ġass ured +ĠPhys ical +End point +.G lobal +.d etail +Ġthe ft +.j upiter +Ġhum or +.R ender +A lex +.c ap +Ġbuff ers +Ġdis pose +t ion +.p resent +z el +, P +Ġdesper ate +.get Column +Ġtw in +ì ĸ +.c an +Ġf lee +ĠIran ian +Ġstick y +ĠU TC +L T +//////////////////////////////// //////////////// +Ġl icensing +_PO INT +ĠM aps +Ġl ol += models +-t ab +ĠN ash +_log ger +tor ch +ĠCON SEQUENTIAL +Not Empty +/ react +Ġp f +Ġassert ion +Ġsubsequ ently +_c an +Ġpand emic +og ue +"+ Ċ +_ ent +_P aram +.ĊĊ ĊĊĊĊĊĊ +Res earch +C apture +Ġbel oved +d em +Ġextract ed +Ġf ights +ER C +(a uth +position s +Ġrevers ed +(st ack +Ġ_ ) +uto ff +_fl ow +ç Ĥ¹ +( Game +Ġex cluded +ĠCS V +c g +ĠT itan +p ause +Ġcer ca +Ġdump ster +L ess +Ġkotlin x +aster xml +Ġpoint ers +Ġfl ows +ĠT un +ĠMain Activity +Ġdis cret +Ġcomb inations +vis it +_b ind +oot ing +d ater +_look up +.n io +Ġswe at +ĠR d +Ġscient ist +ĠP ixel +@ NgModule +Play ing +Ġunf old +Trans late +ĠLaw rence +ĠFIX ME +B ill +ĠR IGHT +Ġwhere ver +Ġo ok +vid ence +Ġ] ]; +ĠSk ill +unist d +ĠðŁ ĻĤ +Ġfem ales +-- )Ċ +İ· åıĸ +ĠF red +Over all +Ù Ĥ +Ġess ence +Ġthere by +Ġw ounded +ĠD OWN +les son +text ure +R ound +Ġautom ated +ĠÐ ¡ +ĠUp dates +Ġsh ade +p ublish +ĠG ear += lambda +Ġle ver +) +" +h ill +Ġrad ar +ry ing +Ġ" ). +f illed +Ġline up +Ġd l +Ġworks pace +V o +_d t +ë ² +_ Item +NS URL +. verify +ĠHawai i +G od +M arch +Ġ[â̦ ] +Ġpel o +ur ious +ĠPitt sburgh +. It +C lean +> \<^ +Ġi os +s ound +"] ; +Ġfre ed +rot tle +ĠL ower +[ count +å Ŀ +Ġp ale +ĠWay ne +ear th +_c ategories +U CK +.m etadata +Ġsum mon +H OME +олÑĮ з +Ġmanufact ured +Ġdo ck +Ġcompet itors +_MODE L +ok ia +ĠH ey +Î ¿ +Ġback ward +ĠPO SS +rop a +Ġc ri +_O BJ +Trans port +-h igh +Ġerot ik +_s lot +Ġart ic +_f ramework +-ser if +ĠSql DbType +') ( ++ "/ +Ġw ore +S il +Ġst oring +ĠPh ase +u ant +Ġb ump +in ho +Ġd ign +Ġback s +q q +(h ash +Ġge o +Ġt ender +Log o +! )Ċ +ĠM X +ĠAr thur +esso a +_C h +Ġbed rooms +="# ">< +Ġth roat +ins ic +.int eger +Ġpr imitive +Truth y +Ġfacilit ate +Ġcreat ivity +ĠD NS +Ġg ra +ue z +Ġcount less +ĠPol and +' M +ĠD ist +Ġv est +Ġcert ification +á» ij +h eld +ext ensions +( static +Ġgr ades +ĠU ber +ãģ Ł +Ġ[ ])Ċ +dat os +Ġget Data +ĠCh arg +ĠB S +.m icrosoft +.v ideo +.d irection +->{ ' +l ua +ape st +Ġbo iler +ere k +Ġdec ides +.j ar +IS C +ĠW ords +(C ON +EMPL ATE +ree ze +sh ots +app s +unt ed +.set Name +:: < +-b old +ê ² +å¯ Ĩ +Long rightarrow +Ġunf air +Ġear ning +Ġsh elf +URE MENT +Ġid le +_M ENU +.C ustom +AG ER +- " +_s witch +b ecause +) view +m are +_ condition +ĠStart ing +M vc +(p re +d ump +_LO CK +at etime +.c allback +ĠC er +op ol +ib rary +Ġres ervation +ĉĉĉĉĉĉĉ Ċ +lect or +grad uate +Ġgener ous +Ġ ion +ric ao +m q +_com plete +(c ursor +ĠForm Control +: center +Ġsub stitute +ĠPl anning +Ġp ension +Ġrecommend ation +ĠT ags +Ġg ef +Ġalbum s +Ġwash ing +ro c +Ġtr ains +at ings +Ġex ponent +ack bar +- ln +á g +.Data Annotations +ĠE IF +ĠMalays ia +ĉ PORT +on us +Ġcle ver +Ġpe u +> ĊĊĊĊ +ĠArg uments +Ġdebug ging +( right +' D +com pute +Ġfin est +OR AGE +Ġspect acular +ph rase +Ġind ia +Ġlegend ary +b irth +Ġcom posite +Ġg rows +ĠT D +Ġep id +Ġlaunch ing +] ][ +Min utes +ĠCh a +Ġclean ed +Ġwitness es +uk an +ĉ Type +Ġhab e +par agraph +ĠJ Panel +ĠH ann +Ġvar ied +ĠP okemon +ĠM UST +åĬ ¨ +.vis ibility +op up +^ [ +.exp and +Ġ" ', +.f asterxml +_ auto +ĠShe et +mark er +Par cel +ew s +ĠStr ategy +-m aking +Ġun ve +Ġtrail ing +Ġclick s +ĠGet Component +ĉ content +IG ENCE +ERN EL +NSMutable Array +Ġb reat +Ġharm ful +¶ Ī +Ġbes ides +Ġb oring +Ġbrut al +v ang +(p arse +qu ick +Ġpy test +Ġswitch ing +() ]Ċ +Ġì Ħ +L ER +ĉf ont +Ġnet t +) ]ĊĊ +(/ \ +æŀ ľ +to Array +Ġbre ed +ĠC AR +ĠWe apon +A bs +t ot +Ġset Name +apt ive +Ġ: , +Ġesc aped +ord en +ĠP ri +th umbnail +Ġdescri ptions +/ styles +ĠPC I +Ġal phabet +astic search +NOT E +Ġc ialis +ĠGr iff +Ġpor que +Ġprote ins +pl ays +Ġst ating +Ġimag ination +Ġfac ial +ĠMe chan +Ġarr anged +_ used +Ġarrang ements +ĠP ipe +host name +Ġprov inc +T it +.Flat Style +ĠS plit +ĠLo ader +.c c +Ġclin ic +---------------- ------------ +Ġb aking +ĠEN T +ne ath +ãĢģ ĊĊ +AN E +.EntityFramework Core +app ers +. ic +ĠNg Module +ĠF ORM +Ġ' ; +-pro fit +h w +en emy +ĠE ye +Ġca ution +t own +Ġur ged +ĠJim my +ynchron ous +-s ized +m aking +, { +] ', +_ Object +ah oma +Ġactiv ist +IN VAL +ĠCom mercial +ĠOr lando +(t ab +ĠØ ¨ +Al gorithm +Ġher itage +Get Mapping +Ġfail ures +ri os +at iva +Ġt et +Ġcar pet +( Z +th ree +Ġdisc losure +. ERROR +_c alled +Ġd ial +Ġoccas ional +.E rr +Ġfunc ion +caff old +Ġrele asing +ï¼ī ĊĊ +_ Value +ĠV ari +y ellow +Ġstrugg les +.c al +ĠDak ota +ĉc lose +Ġsand wich +Ġanaly tics +Ġ** ) +& # +ĠJ os +Ġpass ive +AT TR +Th rowable +ĠM un +ĠU int +(dis posing +ar ak +ĠLe aders +Ġaffect ing +Ġitem View +Ġeconom ics +f v +à¹ Ģ +.r b +ĠOver all +Ġwealth y +Ġev olved +nd a +ĠH us +re strict +um en +ĠA gricult +! ĊĊĊ +Ġexp ires +Ġspokes person +int erval +Ġà ¢ +Ġque en +(n il +ing o +He ap +Ù İ +Ġcompl ain +S ym +ĠCl one +ĠR u +ĠW ILL +ĠCr ystal +/ content +ing en +oint ment +Last Name +av icon +ĠIB M +ĠDim ension +an h +icip ants +ĠAn ne +.pro gress +Ġal go +ob il +ĠV oice +ĠF E +Ġg li +Ġv ed +Ġprevent s +\ Column +Ġfol k +ett i +Ġm n +ĠCL ASS +Ġdisplay ing +ĠK l +ĠF err +d uto +. ib +Ġd ados +' name +-s pace +Ġit alian +Ġin verse +Ġd ense +ut er +ĠI Enumerator +-s ign +Ġnation wide +Ġperson a +Ġsol ved +Ġdram atically +Log out +Ġgr av +Ġanalys es +ol lo +Ġl amp +. team +ĠE rot += [" +Ġd ancing +Ġ?> / +Ġc ater +ff e +ĠSh a +ĠB os +ĠRE QUIRE +ĠMon ster +ĠR B +ĠI DE +Ġsu its +Ġform Data +( theta +Ġsp atial += NULL +ĠSql Connection +Ġ à +ĠV enez +ĠMor ning +Ġpublic ations +ĠNON INFRINGEMENT +first Name +ud s +W ould +_HE AD +Ġinvest ed +st able +f red +Ġcommand er +SE S +âĢĶ a +an che +ĠM ovement +ë ³ +S uite +Ġjur isdiction +ë¦ ¬ +ĠB eth +j Query +ĠIs a +Ġd ental +, * +ĠL imit +ili ation +=" { +b ast +Ġt urb +is y +O OK +Ġadvoc ate +im ag +LE CTION +л ÑĮ +(c ategory +.de c +Ġun iqu +_s n +Ġattract ed +Ġà ī +ĠRun ning +_ edges +ĠDis able +_A S +åĽ ¾ +Ġnetwork ing +_br anch +H aving +toBe Truthy +G I +Ġcamp s +se p +-p art +Ġ)ĊĊ ĊĊĊĊĊĊ +ustral ia +ĠRe ports +rit o +Ġwa ist +_pl us +ĠW W +-p erson +Apr il +Ġs ar +.t ar +Ġagricult ural +t ic +Ġt cp +Ġset Value +agent o +ĠAp pe +p iler +CA DE +Ġan che +atch er +Ġcom ics +Ġl bs +_se gment +'] =$ +itt ers +ich er +G INE +Ġutil ize +ĠC ursor +_ex pression +Ġd ag +< long +Ġr hyth +æı IJ +Ġconsult ation +Y et +")) ĊĊ +_M AC +c ould +Ġ' \\ +ĠV o +ĉ http +Ġg s +ph er +- grid +J ames +J ul +Ġsch on +Ġtensor flow +ĠLOG GER +am as +Ġsc ipy +Ġconv iction +. ag +Ġadministr ator +)) {čĊ +Ġn un +" group +P or +Ġnur se +ex pression +ak y +ĠHe avy +. opt +.get All +Ġover l +/ ", +_c ountry +ç İ +ĠG ENER +_r oute +ĠD al + ´ +ol oad +Ġuncomfort able +(m enu +Ġhost name +' ");Ċ +Ġcalcul ations +-c lick +Ġprotect ive +ãĤ ¯ +_F orm +ung s +Act ual +m f +ĠProcess ing +ĠIn ventory +(m atrix +app ropriate +w eg +ij a +Ġch r +Ġr ifle +-w sj +k ar +Ġindepend ently +I OS +Ġconsist ency +v n +/s ystem +ĠCh anges +Ġexp ose +ici ents +Ġrel ate +ĉ next +è ¨ +ud es +Ġglass es +F XML +.... .. +ĠP df +Ġappro ve +Ġ{ \ +Ġexist e +)) ( +ARE NT +оР¿ +ĠL atest +ĠNiger ia +.Inter faces +Ġrem oves +En emy +Ġen force +vert s +ĉ pos +_text ure +W ARD +ĠINC IDENT +( container +Ġdef ending +ĠR X +ĠH ook +br is +ĠFl ask +Gr ay +. )Ċ +vis ibility +ĠRedirectTo Action +err al +_e lem +Ġres on +front end +_variable s +ater ia +Ġ+ " +ave led +RI X +Ġdef icit +_C heck +YY YY +To One +sp y +Ġun ited +end ent +Ġp ode +ãģ Į +C AT +(f mt +ĠBon us +Ġre ck + º +Mod ules +Ġvac uum +R adio +ĠDAM AGE +P en +ĠPark er +; ;Ċ +ĠRe ally +_n eg +p ending +Ġnomine e +ĠC ategories +ĠUl tra +We apon +Ġdef ender +I ss +ĠG ender +ĠD ress +Ġimpr ison +Ġbank rupt +imension al +PH A +ĠStr ateg +ĠPROF ITS +Ġp atri +//////////////////////////////////////////////////////////////// //////////////// +de legate +Ġfor State +Ġdev oted +_m ake +Ġterror ists +ĠS nap +_n av +ĠA A +ĠI an +ĉ app +Pl acement +_h dr +< K +Ġs ang +st roke +- Q +> x +.T ask +m oney +ib aba +' });Ċ +ĠSpec ific +ĠLine ar +_O PT +Hash Code +( Player +.Contains Key +Ġcoll apsed +trans parent +_R ANGE +View er +(c fg +Ġsort ing +Ġinf ected +ĠN ach +Ġaccommod ate +.element s +_P ART +ĠSex y += get +( year +Ġx hr +: ] +ows ki +Ġsum mar +Ġ ¿ +Ġint e +Ġwork flow +ĠTai wan +vers ions +åı ij +Ġsurprising ly +Ġopt ical +Ġpro ces +Ġdisag ree +Ġnue vo +ĠC AM +sort ed +le ases +ist le +Id ent +ĉ event +ject ed +Ch unk +V ars +.pro vider +Ġproceed ings +Ġin clusive +Ġart work +end ants +ï¼ļ Ċ +se en +Ġl ig +Ġm akers +_f un +Ġlength s +Path Variable +[ item +ภµ +De ad +FFFF FF +ĠUr ban +up les +ich en +(null ptr +.s pec +, System +UR ATION +(j ob +å¼ ı +Ġtrack er +Å Ļ +ĠM R +ĠSQL ite +Ġd to +Ġ; ;Ċ +Ġm int +ĠInt roduction +ca o +Ġquestion ed +Ġf itted +rev ision +s q +Ġm ig +_un its +_ async +Ġf lick +});ĊĊ Ċ +Ġnot re +}` , +F ilters +Ġm undo +_d ays +Ġfr m +ut c +Ġval s +ew idth +ĠGener ator +ĠArt ist +ĠID s +ĠArt icles +re ater +ĠComponent Fixture +. = +Ġr ou +- no +.b ukkit +eg g +ĠD iff +atic s +Ñĥ Ñĩ +âĢĶ ĊĊ +ĠChar lotte +by e +Ġ} );čĊčĊ +ĠV ik +ĠB row +Ġl v +ĠG ib +-w ing +GL IGENCE +(I l +ĠEngine er +.W ait +ĠP ictures +Ġr het +Ġth ermal +Ġpr aise +< >();ĊĊ +ĠSp ider +P ause +ĠB aker +Ġsl ower +Ġ} ]Ċ +_en queue +Ġdisappe ared +ĠT icket +IN UX +_LOC AL +аÑģ Ñģ +@Inject able +comm unity +Gesture Recognizer +åĽ ½ +Ġsca les +Ġ- ( +/ '+ +ĠS it +Ġexecut ives +ard ing +Ġad vers +Ġback wards +ĉ context +ĠH amp +ĠP F +ĠDe ck +ĠCra ig +A merican +Ġb ell +Ġpro l +uf en +Ġr ng +ar shal +ĠSim ply +first name +sh ore +J uly +Ġmort ality +ĠâĨĴ ĊĊ +Help ers +Ġbench mark +em ade +Ġorganis ations +.g son +ĠText Field +Ġciv ilians +.Array s +ĠMiss issippi +Ġinter mediate +get User +_cl uster +Rel ative +fore ign +.querySelector All +Fore ignKey +Ġreason ably +-------- -Ċ +C ards +ĠK am +ĠTh or +Ġroll er +-e lement +ĠC urrency +dd ie +ALL Y +ĠR A +Ġper met +aa aa +Ġhom ework +ĠV it +Ġm old +ĠF er +[ start +Ġstatist ical +Ġsc ary +_H OME +.B egin +Con struct +ogen ic +ĠDEAL INGS +Ġtamb ién +ix on +. ind +ac re +Ġtransform s +ĠN ap +.B lock +uss ia +pir ation +ul ent +Ġce il +Cl ause +na ire +T ES +Ġne at +ST D +ĠReg Exp +per form +: ) +Ġun ions +Ġs ublic +Ġw inds +lo ating +g lich +Ġp agination +S kill +App ly +ĠOper ator +ist ogram +Ġqual ities +C ross +Ġde com +], " +ĠJ uan +.mod al +.Ch ild +ĠRog er +STIT UTE +:CGRect Make +a lette +Ġst a +as ide +Ġbl ur +ĠW a +if etime +re ed +control s +Ġb ins +Ġп ол +*/ ,Ċ +U IS +ĠR ou +ĠDem o +- awesome +ĠCh ain +Ġh asta +ĠB art +. KEY +Ġvend ors +nof ollow +ĠD est +_b uilder +Ġarg ues +_ answer +g oto +ĠRES ULT +ĠM ON +Ġp oder +o ons +_C ASE +Ġrep lic +Ġfin ancing +ĠD ATE +c ern +_tr ack +t ies +/ logo +ĠNE GLIGENCE +get Type +> T +b et +g irl +ĠINCIDENT AL +-s ite +.tr igger +ĠL isa +_input s +Ġrel atives +Logged In +Config ure +I K +. accept +Res ume +ĠD raft +Ġ* >( +ĠW A +ed ian +ern ess +ĠLayout Inflater +*/ čĊčĊ +oth y +Ġoblig ation +Sub scribe +Ġth umbnail +ex ist +Ġins isted +ĠU ICollectionView +ĠAng ular +Ġtable ts +ĠImp act +ãĢį ĊĊ +ah o +Ġcharacter istic +g d +Ġ= ================================================ +our t +` . +App ro +Co ordinate +Rem ember +Ġmar ine +] ==' +ĠAdmin istrator +.get Default +Ġforg ot +ĠStruct ure +V ue +ars ing +m oment +k w +_c ursor +Att ack +Ġath letic +Ġdiagn osed +Ġend e +åĪ łéϤ +H ouse +ĠP ARAM +Ġw iki +ĠO pp +Ġcons ervation +Ġs nd +_t em +sub str +ĠC ape +.s im +UT ION +an an +âĢĻ un +Ġg y +- work +Ġcomp elling +=' # +ĉs ub +Ġdirect ories +íĬ ¸ +Ġtouch es +out ines +.C ollection +s chedule +.l at +ĠDo ctrine +CA A +ĠRe fer +Ġshift s +Ġlik elihood +pre ter +ĠF emale +Ġinter cept +Ġl ou +çĻ » +Ġr ug +ĠC rown +Ġ************************************************************************ **** +- product +Ġprompt ed +ung le +d ocker +ĠT u +ĠUn ique +_ Error +ul os +Ġâ Ħ +Ġ( ` +Get ting +_s cal +ĠEn h +ü t +Ġsust ained +Ġp atches +Ġpros per +ĠG aza +_l ight +Ġin cons +-------- Ċ +ĉĉ ĠĠĠĠĠĠ +S F +C N +: ";Ċ +ĠColl ins +( *) +Ġcomp ilation +'] čĊ +Ġcon sequence +, ... +Ġd m +ĠB LOCK +Cl uster +Ġsk i +(arg c +T uple +Ġjo ins +ĠSher iff +W ar +ind i +Ġcomment ed +H OST +Ġinv itation +apan ese +Ġperm its +preced ented +_z one +ĠA my +_R D +Min imum +Ġinv ocation +.en able +icht en +- owned +" id +_PO INTER +F ac +Ġspecific ations +Ġnom ination +Ġg p +< ( +Ġrob ots +ĠJ erry +Ġhold ers +Ġw and +c ms +Ġ} ))Ċ +.To ast +ĠI List +B ased +z oom +/ style +ĠBe ck +M en +Ġcontrib uting +Ġund o +ĠO H +Ġadd Object +Ġe igen +sign up +éĶ Ļ +Ġdist ant +PAR ATOR +ĠM ari +Ġm á +E mp +ó s +Ġì Īĺ +ev t ++ j +p ark +ĠSt ay +ĠD un +Ġso y +> % +az ines +Ġti empo +(m e +p resent +.Th is +Ġedit ors +F IELD +.W ork +ĠUn iverse +Ġdr unk +.t imer +Ġalter ed +ĠN ar +ëł ¥ +.Act ive +id or +ç Ń +.delta Time +Ġawk ward +& quot +ĠSaf ari +Ġtr icks +MENT S +div ision +Ġvary ing +ĠHigh way +Ġphotograph er +ĠSt ewart +Ġlast ing +.P re +.amazon aws +ĠL uck +.D escription +ĠN az +n eg +Ġc ó +<<" \ +ĠSur v +ĠU nc +Rec ipe +.Border Style +Ġmod ifications +- at +AT FORM +h dr +ak o +Ġsublic ense +ĠJ ump +Ġbe im +ĠMan hattan +. bool +_h w +ÑĤ ÑĮ +B in +Ġg ateway +" ": +ĠU IS +:" + +- def +ĠReg ular +/ testing +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +string stream +Ġdis par +Ġmob il +- read +ĠAd apter +ĠCh ampions +Ġsched uler +Ġk ills +ĠM ultiple +ir ror +Ġgod s +AD O +ak te +ĠUs uario +.c ircular +Ġre cept +ĠEx pr +Ġelder ly +Ġnic ely +Ġbest e +W ant +Ġclass ical +.s prite +obj c +ĠM ason +Ġsist ema +.Bl ack +es o +ĠZe it +Ġdiv id +Ġent ers +_sub ject +ĠPlan et +.w arning +ĠG ram +_t okens +Ġhousehold s +_c ustomer +user Name +c ross +Ġp ione +Ġass ists +_S M +ib o +Ġlo yal +Ġuse less +# elif +ĠUlt imate +C ome +g el +Ġd ich +xy z +ik el +ob ra +_s can +ĠInter ior +ĠN ice +Ġpl ac +ĉt arget +Ġvir al +ass o +() / +und e +ĠAd obe +O s +vis ited +ĠO W +ĠFe ed +ĠSe quence +Ġman ages +in son +ĠLouis iana +{ }) +ĠH ab +ĠL D +Ġb ip +pr ites +(e lem +.h ibernate +él é +Ġoh ne +_trans action +Ġann unci +P ublished +ĠH onda +ĠT am +ĠP acket +_ selector +Ġchalleng ed +Process ing +-h over +Ġtr ainer +_c ancel +ĠNS Dictionary +ab ric +ĠM LS +_s ensor +Ġshr ink +ĠF X +th reshold +ĉH X +-m ark +` .` +S cheme +(f ull +_w riter +ĠS ys +Ġf led +ĠC in +-w idget +ĠPre vious +G ender +_ question +Fe ed +Ġscr ut +(p refix +ãĢĤ ãĢĤ +Ġin fections +Part s +Ġhier archy +_DE LETE +ĠPat ient +_p ay +Ġprom oted +Ġì ĭ +Ġcivil ian +Ġagricult ure +ĠP iece +Ġst ance +uts che +Ass ign +.A CTION +F ig +_r adius +ĠS ync +du cer +f ailure +ens ed +pt ime +B M +_dat etime +qu ivo +QUE UE +èĢ ħ +Ap pear +Ġsum mit +: void +Ġv ine +è® ¤ +on ne +_TR ANS +.g reen +_ cc +Ġhung ry +Ġ" > +() );čĊčĊ +Ex tract +iz ens +Ġsol ver +Not ify +Ġeng lish +ĠSh opping +inter faces +RE Q +Ġil leg +ĠUI ImageView +Ġdis connect +ĠUnt il +ĠConserv ative +@ Column +Ġshift ed +Ġ: čĊ +Ġf ich +Ġd la +Ġsh oe +"), čĊ +ular ity +_RE SP +We ather +UI Application +. iterator +Ġag ing +.P arent +ow ie +(e qual +ĠCon v +/ default +Ġmeas uring +.pre v +.Is Valid +.F at +Ġs Äĥ +key words +with out +Ġso vere +Ġex changes +Ġm elt +Ġis lands +ĠInt egr +Ġjump ing +Ġg le +Ġjournal ism +Ġd ated +Local ized +ĠRef resh +Part icle +Ġa a +ĠSTR ICT +Ġb od +.Pro cess +_A UTO +ĠP ublished +e very +Ġtechn ological +ls x +Ġir rit +Add itional +Ġdel imiter +_l anguage +- area +bo ys +ĠT ube +Ġw at +Ġmechan ics +_ owner +Sp ell +ĠSt ories +.Append Line +Table View +h em +st ick +oll ower +I FF +ĠU V +oll ision +S UB +Ġcompar able +Ġdon de +s ales +ll vm +Ġ} ],Ċ +OTT OM +ĠPur pose +L ab +Ġinterview ed +o is +as il +.set Id +ĠIn struction +-- > +ĠMod ified +ation ally +ĠMe eting +è¯ ¯ +# region +Ġrout ing +.f ocus +ĠYou th +< D +ĠN ag +contact s +Ġform ing +Ġm ie +',[' ../ +ĠB P +Ġapp et +ĠTe acher +ĠT P +Ġann ually +outed EventArgs +ĠSpe aker +Ġre name +CF G +(" // +æİ ¥ +/p ages +Ġpr és +ĠSp ell +.All ow +ĠINT ERRU +Ġ( # +âĢĻ ĊĊ +_G eneric +.im show +_t im +- face +(& ( +atin um +Ġrevolution ary +ĠH ours +r ain +Ġany time +Ġab b +.j sp +Scroll View +ĠTr uth +Ġanticip ated +Ġacc ent +. checked +Ġspec ifies +Ġca f +Ġcell padding +Ġcook ed +ĠH ugh +pe ek +_R ATE +Ġd orm +/ čĊ +IV ITY +.Cont roller +(p art +.con straint +Ġinv asion +MO VE +Ġgl uc +l ename +Ġam en +eng lish +ĠSw itzerland +";ĊĊ Ċ +pe st +.col lect +N ib +ĠD ict +ĠE mb +(sub ject +Ġoutr age +Ġdec iding +Ġsent enced +F echa +" A +Ġqu er +Ġfont Family +Ġqu adr +- Y +_C ACHE +Ġanaly zed +Ġg aining +ĠAgain st +ĠSou l +ta u +Ġlight weight +ĠT F +ĠEffect s +.T ypes +.add Class +Ġv egan +é ģ +.' " +ĠExpl orer +.d etect +.sh ift +Ġoblig ations +last Name +Ġassoci ations +ĠTime Span +un ter +ĠF resh +Compat ible +P ub +id ges +. option +var i +.hash Code +Ġg eb +. section +- not +ĠSub mit +T N +reg istry +_m edia +Ġn aj +ff t +Ġm ate +-th ird +Ġp ockets +est a +Ġb ent +ĠN ord +Ġretail ers +ĠMor ris +."" "ĊĊ +W rong +Ġ ÅĽ +R ay +. ec +ĠB ind +_H AND +(n on +is Valid +Ġsimilar ly +_L IMIT +Ġdynam ics +Ġdist inction +ãģ Ĩ +< N +Ġor th +ĠToy ota +ĠK ate +ĠL S +or ie +ĠSpr ings +Ġf reak +last name +_M ULT +-st ep +" ( +AD DR +Ġentert aining +_CON F +Ġdec oded +Ġst reak +Ġwait ed +Ġnot ified +rodu ced +vis ual +.Layout Params +æ ° +es ian +f its +s pring +ĠBern ie +User Defaults +Ġped est +Ap pearance +ĠW iki +ĠNOT ICE +Ġs sh +Ġdur ante +ĠZ ip +ı r +ĠNAT O +Ġtw elve +Ġro yal +ï ¸ +Ġmer chant +ĠF urniture +'] ),Ċ +, X +Ġfold ers +ĠG ate +ĉf unc +p ick +_us uario +ĠV erm +ment ion +ur pose +Ġalert s +x ious +_s ig +ĠF u +Ġ( : +Ġd umb +åħ ³ +Ġaccur ately +éĩ į +R B +-s creen +ĠV ER +j our +Ġrom ance +uc ceed +. choice +Ġad ip +_d ims +Serial izable +ãĤ ĭ +.j ob +Ġpro g +uch ar +Ġg ently +ĠR SS +ict ured +_ENABLE D +ĉ label +aw ks +ĠEn sure +rem ember +ìł ķ +Ġtrans mit +{{ $ +.Trans action +ur se +_rel ative +Ġs ized +ĠX X +ĠPr incess +ĠL arry +Ġpr ó +ĠÑģÑĤ ÑĢ +Ġs isters +estr uct +Ġcheck point +: length +ĠCar los +/ icon +_T ARGET +T okens +Ġpat ience +ĠSe lected +q ty +.show Message +Ġwild life +ĠP rops +b m +- arrow +Ġpar cel +fire base +ĠBen jamin +cess o +.t im +ĠG arc +. any +ĠHOW EVER +ĠK o +Ġgrab bed +_f rames +Ġobject AtIndex +ĠADV ISED +Ġsub ur +ĉ GL +Ġ}) }Ċ +-l ength +ìĭ ľ +ĠPot ter +_b uff +.g ui +ĠEnc oding +E lect +-m essage +Ġ � +Ġ ÈĻi +ĠArgument NullException +а ÑĨи +Ġmin imize +Ġrespond ing +$_ [' +ĠInd ividual +á c +ĠIN TER +Ġmast urb +ĠB in +(' $ +ëĵ ľ +Ġopen ly +Ġ> < +Ġun to +olog ically +ĠM ul +VID IA +Ġsl im +ĠCommission er +( on +Ġunder neath +/ db +v ote +( Message +ĠP ope +Def ined +Ġsw ift +ur f +Ġadapt ed +SE L +Ġreven ues +Ġdiv ine += y +Grad ient +_ act +Ġ/*! < +Ġpoly gon +ĠF DA +ĠC arr +at ables +(std out +Ġrefr iger +Ġco ordin +avor ites +ÑĪ Ð¸ +Ġcompass ion +ĠPOSS IBILITY +- secondary +ur acy +Ġcomp romise +_A V +_ os +Ġbes ide +ĥ Ŀ +Ġl n +.pl ugins +Cap acity +al ah +.b in +ĠC RC +_b alance +Ġflex Direction +Ġam bit +Ġnick name +ĠFor ces +C LE +ĠSh ell +Ġs ail +ĠW riter +ĠA lice +d w +ĠInd ians +ĠMar shall +_S RC +Ġnormal ized +ĠJ ag +ãĤ Ĵ +ze it +r pc +ÃŃ c +.in line +Ġtrav ers +_n umeric +Ġutil ities +Ġev ac +IN PUT +ĉ register +M X +ĠCamp bell +Ġdatas ets +Ġdem anded +Ġinitial State +g an +Ġe i +Un expected +- web +tr ait +, Y +ĠT odd +Ġske leton +Ġoptim ize +ç¬ ¬ +ĠU pon +ĠSt Object +Ġap lic +.' P +v ron +. UN +Ġpaint er +izar re +Ġl av +Ġp om +p reg += function +( serial +ific a +um ing +åľ ° +ãģ Ĥ +- op +U CH +ĠH end +.prop Types +Ġy o +Ġrout ines +Ġcar ing +S em +Ġres erves +Ġprior ities +red its +IST R +Content Type +ĠSch w +/ media +Ġe str +Ġclim bing +- week +cher che +s ensor +To Array +ĠMont real +Ġcloud s +ĠInject able +ĠR ice +Ġpropag anda +_pro vider +Ġind oor +Ġin aug +Ġdipl om +Ġmess aging +_m ut +å ¦Ĥ +Ġk w +ON S +ari ans +R PC +) ]čĊ +-r ay +ĠS or +m all +Ġmarket place +Ġv tk +M a +og an +ig i +Ġspons ored +ĠD ani +.S EVER +>' .$ +m ultipart +ĠW ol +Ġtable Name +ĠUser name +Background Color +Ġf right +_E MAIL +Sept ember +_val s +op ia +Ġsp otted +- Ch +Ġdata Source +/ "Ċ +ек ÑĤ +ĠRequest Method +ĠRe place +-d o +ah n +ĠPh D +] .ĊĊ +N ON +g ement +ĠTh r +Ġquiet ly +Ġtort ure +Ġte as +ĠC Y +Ġa tr +develop ment +-d etail +Ġlight er +Ġarg uing +Ġdes erves +Ġcur riculum +_CON TEXT +ÅĤ y +H ITE +ĉ ID +/ uploads +Ġt its +re o +_d rop +. UTF +Ġpick up +Ġgro cery +ĠP ure +Ġeas iest +Ph il +.f eature +(" * +Ġinvest or +t ok +Ġj ar +L os +âĢĶâĢĶâĢĶâĢĶ âĢĶâĢĶâĢĶâĢĶ +. queue +-s peed +M al +um blr +ĠCON ST +ĠH RESULT +ĠD ance +(file Path +Ġattrib uted +ॠį +ĠB und +co ins +Ġs ão +Ġp ir +person al +Ġpre lim +Ġprop ose +ĠT L +] ]) +ĠSub scription +ĠK re +, len +.First OrDefault +) -- +_product s +.Get Bytes +Sh ip +Ġenc rypt +ĠS G +ĠM yst +h ir +Ġiter ate +Ġint end +.mock ito +Ġch apters +( angle +ĠV lad +è® ¾ +' .ĊĊ +Response Body +ĠAb d +de al +Ġbar riers +-out line +b ill +ĠF alls +_se cond +. include +. ceil +Ġoccup ation +ph ony +.move To +ĠJenn ifer +AST ER +; ">< +ĠEn abled +Ġtermin ate +ĠI o +l ations +ĠTHE ORY +Ġear liest +Ġr ack +ĠSc ar +sh ake +ch ip +Ġu v +Ġall iance +п иÑģ +ĠGOOD S +z ione +ĠV I +Ġ{ - +Ġfilter ing +Ġmis con +.Dock Style +Ġb ush +Ġj unk +æ Į +ĠQ UE +Ġhook s +Ġfirm ware +Ġmiddle ware +d ic +ĠOak land +Ġarr ives +P ayload +p ixel +] | +Ġstart Date +.P RO +_a udio +Ġmid field +igid body +ĠSw iss +ĠCl ip +ĠD ump +ĠText Box +Ġg eh +y ield +od s +Ġrefer endum +Back end +ĠC ream +Ġdomin ated +ĠArch ive +Ġrid ers +.prepare Statement +Ġqu ando +Ġche f +w iki +in el +am pling +(" \\ +Ġs ag +_pro xy +ãģ ķ +p do +.getElementsBy TagName +Ġdemonstr ation +ĠN PC +Ġarch ivo +end ance +Ġefficient ly +( actual +.t ableView +Ġm ush +Ġbe ars +_thread s +j as +ah un +Ġne ural +Ġdesign ing +ĠG DP +Ġlift ed +çĽ ® +ĠJ oint +ĠIn clude +ĠGi ants +Ġwithdraw al +ĠR ent +n ative +ĠSe ek +gress ion +_C PU +\ S +ĠSh ield +Ġsol ic +Ġbo om +yect o +Ġmanufact ure +ĠâĢ ĭ +Ġb box +Ġearth qu +ollect ors +:@" % +Ġlo ops +J e +alk ing +ĠWh ats +ĠBo ys +. book +ARG E +_p ixel +Ġsus pects +Î ¹ +us p +ĠBM W +ie ces +(p erson +å¼ Ģ +é » +ĠPod cast +Ġb ou +( Item +à » +( Input +Http Get +Ġb urg +) ^ +BO ARD +*/ , +Ġg ulp +ĠB enn +Ġdeck s +.status Code +Ġac ute +Ġh ug +ug u +Ġp led +," % +h ape +Ġз ап +ĠMain e +.re al +Ġd alam +ĠMin or +.F loat +dis p +Ġt l +Ġen count +=> $ +Ġf g +te es +ĠRec omm +ä l +Ġchem istry +Block s +O ID +Ġfore x +ĠApp end +Ġ{ * +ĠSup ply +CG Float +(b l +Ġat e +ador a +Ġg ust +Ass oci +> .Ċ +F ETCH +.s erial +widget s +ard less +ie fs +_F ULL +ernet es +ĠP red +Ø Ń +äº ĭ +ub ernetes +ĠL aura +Ġl abeled +High light +Ġanno ying +/ update +(d escription +Ġintim id +$ c +")) )Ċ +.A P +Ġ[] * +ĠEX IT +.H ost +ĠOP EN +.send Message +_c amera +_t ile +Ġth erm +onom ous +Ġdis adv +Ġna ar +index Of +ĠP P +.prot ocol +AF E +Ġtext ures +################################ ################ +umb ai +.st ats +ĠG E +Ġi e +ĠST D +ĠM ann +.ref lect +K B +Ġd ive +.w av +/* ---------------------------------------------------------------- +/ settings +.l ifecycle +Ġda ughters +or us +ub er +N ING +st ri +ĠT ip +Ġz n +Ġswitch ed +in et +uff y +ĠTransport ation +( conf +fr ica +ĠX L +ĠLe ad +_per cent +< Map +Ġthr ust +or b +ik k +Ġtra uma +Access or +ĠF it +ĠString Buffer +ex pl +(s creen +Ġaud iences +ĠO PTION +_ round +[ node +be h +-> __ +per missions +ĠD etermine +.M an +Ġadv ances +. InputStream +Ġstrong est +Ġe Bay +Ġ# - +Ġdir name +ĠS MS +Ġmedic ations +Ġam ended +Ġchurch es +ĠImper ial +$ row +ĠMad ison +ĠIn sp +Ġaff air +Ġpsych ology +v h +Ġsever ity +âĢ IJ +Ġstri ps +A H +vert ising +Ġcon se +IM AGE +ĠSt ats +ĉs c +.C ursor +Ġfree ze +ss on +(x ml +ĠSus an +.t ile +ed ed +ĠĠĠĠ ĉĉĉ +uel le +ĠMitch ell +b ased +Oper and +½ æķ° +ĠF F +ĉstr cpy +ounc es +ild o +.execute Query +Ġapproach ing +ĠSe ven +Ġn uts +Ġr ic +ass ignment +Ġcalcul ator +ĠMur phy +ĠB ou +í Ħ +Ġbut t +Ġt icks +Project s +il ib +.text Color +m ov +_log o +( template +ĠIN IT +Ġimage View +scri ptions +OR ITY +Con sumer +Ġun precedented +Ġtour ist +Ġbr on +Ġcontract or +Ġlic ence +ĠN am +æ ¯ +( transform +_AT T +P ref +ĠG am +Ġvess els +Ġh av +L ater +.To Lower +Ġurl s +Ġbreak down +Ġpen alties +Ġf oster +ĠU E +Ġcl ue +com ed +åIJį ç§° +-m ain +Ġp ts +Ġcount ed +ict s +/ post +Ġget attr +Ġp ing +ANCE L +Ġp ec +Ñħ од +ant om +ĠBlue print +ĠEvent Emitter +Ġl ä +æ ² +Ġstr aw +( comp +' une +> N +- client +es Module +-b ase +Ġret reat +_s imple +ĉĉĉĉĉĉ Ġ +fe e +') čĊčĊ +Control Item +Ġsubscri bers +ple ase +ĠE ff +Ġp ound +ĠBy tes +ĠTe a +_ activity +Ġmax im +Ġop code +B SD +. constant +; } +omb res +Ġcare ers +) .ĊĊĊĊ +Ġsp reading +-exp anded +ĠOr d +amar in +Ġmob ility +Un fortunately +ak k +N L +_ redirect +ĠP G +ĠS ensor +b ol +t ap +_MEM ORY +ĠUI Alert +plit ude +We bsite +ĠLog o +lo ve +[ ind +Ġalto gether +Ġwonder ed +Ġes per +ĠLib eral +Ġo ss +Ġel it +Ġst iff +od ox +_ment ions +ĠDou glas +_p id +ĠC K +ĠinitWith Frame +.b log +p kg +ang hai +QUI RED +u u +Ġm kdir +AT AL +Ġun h +in ces +st h +Ġhypo thesis +Ġc ata +ĠT B +ĠCl ar +Ġpre decess +Ġsitu ated +-w orld +)) / +Ġhead lines +.st at +Ġout break +sp ath +_FLAG S +ĠServlet Exception +S un +F ROM +ĠD ir +ãĥ»ãĥ» ãĥ» +_co ord +ĠOpt im +Mon itor +.b it +XX X +Ġtod as +f eld +ÑĢ Ð¸ +im ir +Ġpolit ically +Ġmolec ular +Ġtrad ed +Ġ{{ $ +ĠSw edish +Ġ'@ / +_RE AL +Ġw arehouse +t oday +, L +or p +< section +- br +ym e +ĠUser Service +Ġlib erty +Ġmoment o +( Image +< size +S ch +Ġj og +i ology +arent ly +Ġquant um +ĠAb u +Ġr im +Ġman a +Font Size +Build ing +st airs +AIL ABLE +Ġ& ' +Ġs ect +Ġs igh +(b atch +.I Container +p oll +ĠCor ps +Î µ +ar u +ĠK ay +.r ange +_click ed +ĠRobert s +.N etwork +fin ish +- Man +Ġcolleg es +ĠF ine +")) ,Ċ +f ilm +Ġrem inded +Ġgest ure +out il +Ġthread ing +Ġobj et +Ġt ours +activ ated +.m kdir += user +Ġre de +f ü +_SY STEM +p v +Ġcon gr +Ġmass asje +Ġpract ition +Un iversity +Ġtab index +Ð ĺ +S ets +Ġcount ies +g uest +f an +Ġword en +.d i +на Ñĩ + ¿ +ig Decimal +Ġsh ore +Ġg ö +Ġrep airs +Ġhelp ers +Ġcenter ed +OL LOW +Ġmap StateToProps +Ġc ents +< A +Ġexpect ation +Oct ober +Ġbg color +ca les +.C ON +ĠV el +Ġcry ing +-se ason +Ġfunction ing +_LOC ATION +ü ss +ber y +Par a +omin ator +- le +Ġeth ical +has htags +emp lo +Ġn úmero +( activity +.St op +.str ftime +IL D +Ġto e +ĉ Node +") čĊčĊ +ĠPu erto +Ġexec uting +ĠG UID +Ġoppos ing +al ph +Ġexhib it +_fl ash +Ġme ille +Ġjson Object +H ero +aint ed +_D OM +Ġw il +Ġslo pe +Ġm Ã¥ +ĠIraq i +Ġorgan ize +ĉj Query +H UD +sh ine +. we +ĠSk ills +pons or +Ġcon clusions +Ġre forms +Ġrel uct +n amed +ĠOl iver +Ġ// }Ċ +- looking +Ġf og +ĠH O +ĠF ried +Ġinev itable +ĠData GridView +H our +il les +log ical +Ġconnect ivity +.tw ig +ĠK yle +(d st +- Sh +ĠStud ios +( Level +.j et +_PRO TO +-de coration +OT HER +Ġread ily +.Param eter +Ġmultip ly +ĠL IB +ar med +Ġsoon er +æ Ħ +_ ES +Ġfoss il +ĠA nc +âĢľ This +l odash +Py thon +Ġhist ogram +west ern +Ġinf ant +Ġco ordinator +Ġn ib +: m +Ġres pected +Ġdef init +& T +_p ad +ĠTr igger +th al +Ġimage Named +Ġbeat en +ĉ rc +ĠPal ace +Ġhaz ard +Ġisol ation +_ rc +cont re +OUT PUT +Ġre ign +ĠPl ate +AT ES +Ġfl ux +Ġpack s +.get Selected +Ġparticip ated +Ġneed le +-de pth +:::: :: +-l aw +ins pace +on itor += no +ĠAt omic +ĠBr ain +Edit able +-s c +red ential +ĠP erry +k ie +Ġ ----------Ċ +.st roke +( Intent +Ġun ity +um lah +F urther +Ġpr ze +Ġs ø +ãĤ Ĭ +ĠPROC UREMENT +ĠH ousing +Ġatt orneys +Ġcomp ose +atter ing +" What +dra ul +Ġstraight forward +In stant +.J TextField +Ġtr ades +л а +Ġ{ ! +Ġl ately +IM G +ĠA ld +ĠIN NER +Ġcart oon +.S ource +F ALSE +Ġd ough +f en +( rect +Data Table +N ick +ĠBut ter +read s +_com ments +EN V +ĠConnect icut +-F IRST +ĉĉĉ ĠĠĠĠĠ +ach i +.M sg +re ction +Ġrelax ed +Ġsha ft +Ġe f +ĠAdd ing +Ġbre ach +Ġ ï¼ļ +ram a +Ġconduct ing +Ġ( ; +(g l +ĠCA USED +ash i +ĠF LAG +ĠCom merce +ĠIN TEGER +h ours +ĠSchool s +Ġn ucle +Ag ain +pro j +Ġsevent h +EMPL ARY +(m ock +'] ,čĊ +_S PEED +> false +Ġsp a +ĠN ear +ì ķ +Ġintr ig +_m embers +w ave +Ġanalyst s +_O S +ed in +ĠF ri +Ġretrie ved +Reg ular +_ obs +EX PORT +')}} " +" class +__ (( +b ucket +Ġst ro +ĠP atch +yst ick +ful ness +ap os +D a +ĉĉĉĉĉ ĠĠĠ +Ġen rich +un ordered +h ole +C ong +< Product +ĠC urt +( the +_l ower +Ġavoid ing +Ġbu zz +Ġv iable +ub a +- is +are l +Ġact ed +-d etails +ภĩ +ĠThe ory +ĠP un +ĠAn onymous +... "Ċ +è res +åı ¯ +ĠV ision +_se m +ash a +Ġcelebr ity +Ġend Date +Ġpop ulate +Ġcu is +qu ant +f loor +Ġglob ally +Ġcru ise +ĠStan ley +Ġb ikes +.get Connection +Ġpoor ly +_ other +amp ing +." );ĊĊ +od i +_A DMIN +.color s +ĠG aming +> ';ĊĊ +STR UCT +Q R +ID s +(arg uments +_a ux +( Event +_PR IVATE +ĠTre k +Ġdownload s +m utable +_STR UCT +(w x +Ġdom ains +js px +ĠVi agra +Command s +J s +.c fg +Content Pane +ĠEdit Text +à¥į ठ+Att ach +ĠAR M +posit ive +ĠGener ated +Ġse ized += : +Ġelectron ics +ĠApp Component +/ ',Ċ +.equals IgnoreCase +Do ctrine +d isk +ĠPolit ical +CH O +< F +ĉ height +ĠB ug +. le +ik h +Ġmill iseconds +Ġconstit u +m ag +.n l +-r ange +ang gal +', [ +ropol itan +Ġà ľ +ĠU C +.d esc +-L AST +f stream +ib il +Ġf ier +VER Y +Ġë ³ +IR T +_ UI +( abs +Ġkne es +Ġro okie +ĠV ac +are na +comm end +- \ +ĠSUB STITUTE +So ft +Ġpart ir +we alth +è¦ ģ +(d ataset +ĠCl imate +- show +Ġreli ability +_ch unk +ä» £ +_st ock +ĠEX EMPLARY +ï¸ ı +Ġv ÃŃ +Ġsm iled +Ġdr ill +.F unction +ĠS I +Ġreg ression +- X +ĠJ ar +p ref +ĉs uccess +ĠHit ler +Ġinst inct +Ġfem mes +Ġlo ver +< Ċ +Ġmulti plier +r il +Res ize +ĠAuthor ization +ĠK an +Dispatch ToProps +Ġc rops +t okens +ec n +ential ly +ĠINTERRU PTION +f ake +Und efined +ĠA K +ĠTest Case +Ġr ab +Ġtor rent +ĠO t +B ars +Ġlect ure +Ġen jo +Ġrespond s +Ġindex ed +Of Work +_ch ain +)) -> +ĠBeaut y +Ġ` < +Ġtouch ing +Ġ| -- +ĉf lag +normal ize +Ġtr apped +Ġestablish ing +/b uild +A J +f y +- react +av n +RI PTION +Ġk ut +ĠF ashion +ĠIn form +cur ities +< byte +ĠUkr ain +Ġs ug +Ġconsist ing +ood le +. ctx +.To List +Ġcomment ary +Ġtransf ers +Ġn ost +ih ad +ĠU pper +Ġconf using +miss ing +- cl +Ġbound ing +Ġcongress ional +Ġreve aling +d h +r up +Ġt res +re peat +, ĊĊĊĊ +_t ac +Ġexp ed +G irl +h orizontal +Ġ"../../ ../ +( option +Ġwe iter +ĉs ql +Ġ=> {Ċ +Ġgar lic +Ġre pr +Ġrepl ies +( prop +Ġspir its +Ġins pire +Ġbas ement +.re ject +Ġhint s +Ġpoll ing +ĉ ĠĊ +_r ating +Ġc ath +av ier +Ġcomp ressed +ĠV S +] ' +Ġjud icial +ĠT rend +tr aining +EST AMP +ogn ition +Ä ģ +SE NT +vent ions +Ġconsult ant +um ph +Ġuser Service +, NULL +k h +D ear +_B AD +it ations +Ġmet aph +' é +and ise +-f ont +.ch art +Ġs g +_ Controller +.j peg +ĠUL ONG +ĉg ame +( ss +ĠM aj +ĉg o +ĠS ad +ĠB erg +ĠM ine +P ack +Ġres istant +ĠR OM +Ġp eg +ĠStan ford +ĠY ahoo +Ġsca led +Ġl an += [] +"/ > ččĊ +Ġs ud +ĉ background +Ġsch olars +-m uted +ar á +Ġ= ==== +Ġ__ __ +C reat +ene ver +/w p +ĠV PN +Error Code +) ],Ċ +(b uilder +ĠEn emy +S ensor +us a +Ġtr iggers +Ġplayoff s +_RE Q +Ġ( ~ +ĠBar ry +Ġperman ently +ĠR UN +Ġb ure +.Fat alf +Ġch ick +ĉ panic +ps i +ok a +éĢ ī +> [ +Ġunderstand s +ĠJun ior +ĠIN FO += mysqli +ust ain +-s ource +s erv +ĠC REATE +. au +Ġsell s +ĠĠĊ ĠĠĊ +E urope +z w +pre h +ĠNS A +Ġx y +ภ´ +ĠB eyond +Inst ead +Non Query +Ġar ise +Ġavoid ed +.em place +_model s +} ),Ċ +Ġh id +Ġ& _ +.p oints +.get Width +.Ex ec +Ġ// // +ĠS essions +... \ +ĠCol omb +Ġacceler ation +rest ore +Ġ ile +ob ic +< Node +ĠD X +ĠBes ides +. age +ĠCont ains +N ational +ĠIm plementation +Ġeff ic +ĠR M +H y +ĠWed ding +ok ies +Ġrec ursive +Ġprosec utors +.Se lection +ĠForm ula +Been Called +[i i +ĠFr an +Ġtraged y +_F EATURE +Ļ ¨ +comp ass +ĠB h +? ĊĊĊ +.w riter +ĠH our +Db Context +io v +am on +re pr +é ĥ +ĉf i +'] ] +ĠD ry +. ro +ĠO bserv +æł ĩ +Form er +ĠB alance +ĉ json +Ġpr zy +I SS +( sock +ĠL INE +Ġde ce +Ġal ly +Ġtend ency +F un +Ġschem es +Ġinter ven +æĺ İ +Ġad verse +quote lev +Ġsacr ific +_s ide +Ġmut ex +AG IC +Ġocc urring +ĠCommunic ation +um ar +ç¼ ĸ +ĠTreat ment +.p erson +ĠL C +Ġe ch +( (" +ĠDise ase +ä d +ĠA Z +.A ccount +Ġcontinu ously +END ING +ĠRET URN +- string +.f ilename +syn thesize +Res ponder +( opts +reg s +Ġn uest +Pe er +// ------------------------------------------------ +Ġg auge +ĠK in +.s chema +Ġarr ange +ĠBl ake +_Type Info +C over +ĠHamp shire +P aper +-in ner +util ity +Ġcross origin +F OR +Ġign oring +ĠD D +av an +Ġtrad itions +Ġget String +Ġeth ics +ĠMaterial s +DE SC +Ġen zym +io let +ĠCh ip +ĠMc Donald +Ġn erve +ç Ħ +") ] +æ± Ĥ +ĠS ugar +_S IM +j peg +Ġdiscret ion +ĠT N +bo ve +ĠMin imum +ĠForm Group +Ġwork force +ĠExec ution +err er +ĉ ĠĠĠĠĉ +Ġpres cribed +.Text Align +OP EN +ĠP B +im ity +ĠEx ternal +° C +ĠApplication Controller +Ġb arr +imp licit +_d ot +ĠCol on +C OLOR +.Pro ject +* }Ċ +pl aint +get Text +Ġindivid ually +Ġcheck box +U Y +ĠL amb +Ġdys function +ĠL ar +à ° +ĠCre ating +');ĊĊ Ċ +" They +loc ations +_C ORE +Inter action +umbn ails +ĠPart ner +b rit +Ġless er +ĠSl ot +set Attribute +ĠW ave +.p o +/ store +Ġbrows ing +_p d +sum e +s ed +Cur ve +Ġpl asma +Ġsusp icious +ìĿ ¸ +ĠB ah +ĠExp licit +_C C +.Client Size +\ View +Ġsub stit +lo on +ĠG AME +ĠB rid +Ľ 建 +_ User +Ġsqu ares +f one +Ġsac red +ug hs +] interface +ĠTh row +ĠK irk +Ġemp ire +Ġassess ed +T ax +ĠHe aven +-b uffer +_STAT IC +én é +-b ordered +Ġpun ct +(m ode +Ġke ine +S ent +ĠCal cul +ĠE ve +Ġsty lish +Ġoil s +.Test Case +Ġtrad emark +Ġliter ary +Ġconcentr ations +ĠRel ations +( Class +Ġstd in +Ġv æ +back up +. VERSION +.AutoScale Dimensions +st arter +Transaction al +- panel +St udio +k c +ĠCh amber +ĠSpi el +Ġr ho +ا ÙĦ +! ' +.At tributes +Ġmurder ed +apeut ic +Ġint imate +Ġtext Field +ĠBuff alo +d ummy +" % +ĠLib erty +ob ar +ĠT ank +ĠPop ular +erv isor +ĠIn iti +ĠM all +ĠP rior +C AP +ĠCl ay +ĠCert ificate +.L ock +-st rip +-dr iven +/ all +ĠMessageBox Buttons +_SE CRET +_p b +Ġr ats +ा ठ+Ġn t +.R outer +_top ic +Ġt ennis +ĠP UBLIC +ĠActiv atedRoute +Ġ' ,Ċ +Ġcost ume +Ġj okes +. Handle +ĉ byte +Ġflav ors +( cc +Ġperson as +ĉ image +ĠN azi +Ġgram mar +Ġú lt +Ġval ve +Ġv ic +ĠR achel +_in valid +P refs +std int +(r oute +Ġhtml specialchars +Ġpe oples +pl ine +Ġn v +ĠQu ant +opp ers +Ġcurrent User +ĠC atal +Ġrecon c +Ġconj unction +l x +amb urg +Ġinflu ential +d anger +ind ers +Ġ% @", +.config uration +os ome +. identity +Ġpick er +n ost +ĠDI Y +Aug ust +ab lo +Le af +ĠRec o +ck o +DO C +ĠH erm +: any +ĠInt erview +ĠT ex +x fe +( work +Ġle ap +He ading +Ġqu arters +\ Bundle +re b +Per haps +ĠG mbH +B irth +ĉ sum +ĠWat son +.n il +ç ¡ +{ }ĊĊ +ica id +Get ter +" name +Ġ" čĊ +_n one +z m +ac ute +uest o +Ġs ous +Ġre build +Ġnewsp apers +ĠH az +Ġk its +if o +Bl ur +Ġsu ited +- In +à ¯ +ĠKe ith +ĠNor way +IN IT +ire ccion +iet ies +_us age +ĠDou g +r ise +Ġtr illion +im ited +ĠR EL +al ic +Ġcritic ized +the orem +Ġce ase +Ġsid ew +ĠT erry +Ġsubs idi +Ġfirm ly +Ġaw s +Ġh ott +Ġdress ing +bad ge +ĠApp lications +è¿ ĶåĽŀ +Ġlaugh ed +Ġh obby +Ġmus icians +Ġ* . +. placeholder +Ġcount ers +ĠCap itol +SD K +Ġhel met +and box +qu it +Ġcriminal s +Ġteen ager +( update +G l +.se lection +Ġdis charge +Ġpresent ing +ufact urer +_UN KNOWN +Ġstress ed +å ύ +Pro to +_cor rect +ha us +Ġren ov +Ġfire arms +Ġtechn ically +-b rowser +Ġc andy +St roke +Ġexec utor +Ġocc urrence +ĠIP v +_INTER FACE +ĠRetrie ve +.b ad +Ex change +Nav bar +ĠK id +(get ApplicationContext +_ST OP +ĠB oss +List eners +Ġshoot er +ĠAl b +ä ch +Ġp ix +.key Code +al one +Ġabs urd +ĠC um +ĠNewton soft +ik t +Ġlaugh ing +Ġcapital ism +ree Node +T x +_QU ERY +.S leep +( login +Web Element +Ġcelebr ating +Ġde precated +Ġma ar +Ġart istic +_ASS OC +ĠBorder Radius +ĉw p +Ġsurviv ors +In ner +- red +Ġprosec ution +_ pp +(" $ +Ġcomm a +un checked +graph ics +r ors +G ROUND +( public +Ġcustom ized +ĠArk ansas +ĠR ew +Ġexp iration +× ķ +ĠC ul +Ġn ons +.F ilter +Ġsen ator +_def inition +ash ington +ym ph +/ J +Ġf use +ram id +ĠSup plier +Ġaut ocomplete +Ġ} ), +." ĊĊĊ +_function s +ĉ to +.e val +ĠT Object +Re ferences +Ġhe ated +H AL +Ġ)) }Ċ +} $ +ĠB arr +_UN IT ++ $ +Ġget Value +ip ed +ch ied +(v m +c ue +_int eger +_c ourse +th ird +Ġrevis ed +** /Ċ +_D IRECT +Out Of +(" ( +ĠFe el +Ġre ass +Ġsub title +per i +n f +Ġenjo ys +Ġtreat s +) this +-t abs +anc ers +Ġcontin ent +Ġcard io +S er +. question +Ġph rases +Valid ators +Ġpop ul +Ġl ÃŃ +s ong +_IN TERNAL +Ġadvis er +Ġp uzz +Ġambit ious +ĠT ob +ĠD P +Ġpres idency +Ġsurre nder +Ġwatch es +_b inary +ĠSo on +Ġcan ada +(" ")Ċ +] =' +ĠBr andon +eps ilon +r w +.add Child +.C opy +Pr incipal +Ph otos +Ġmarg inal +Ġbas ics +e ing +M ust +_ String +Ġo le +M agento +.c ustomer +(p rev +ภ¥ +Ġlo yalty +C og +Ġprot ocols +ĠCom panies +Ġtheoret ical +Ġaccess ing +ĠZ en +. ones +att ice +_w orld +z es +Ġtatto o +Ġmen os +Ġinter sect +"] ;ĊĊ +bel ie +Ġin active +.read line +-label led +.d one +lick r +ĠW ORK +Ġderiv ative +Ġd atabases +âĤ Ĥ +Ġs x +.is Array +Ġy s +Ġp ada +ĠBul let +(` / +is Active +ĠCG Size +(equal To +ĠColum bus +Ġmar ry +DE V +_l imits +ron es +I AS +Ġt au +min o +_W rite +ĠW ine +Ġ[ [' +ĠP ull +rit ers +ri ents +Ġsh ifting +up p +_TIM ER +ĠCondition s +Ạ¥ +ĠOr ders +ĠSt rength +æī Ģ +Ġvalid ity +Ġf ot +et ur +Ġb olt +åĨ ħ +ĠAl ong +os hi +Ġassum ptions +Ġmag azines +_S PI +Ġp unt +_PRO DUCT +Ġrel ay +ĠJ avascript +. te +- es +Ġwidget s +(f s +< Item +_ex tra +Ġrecru iting +E t +Ġnecess ity +p w +Ġnov els +uss els +Cre ator +ĠM VP +ĠO C +th ood +cl ients +)) * +Ġcharacter ized +_SE ND +ut i +T y +.from Json +@ Service +ãĤ Ĥ +Ch ris +_ Is +ĠJohn ny +Ġclean er +ĠInitial izes +UN K +( axis +еР· +ie val +ĠWar riors +} )( +DM I +âĻ Ģ +ĠTre asury +Ġfe as +Ġsl a +_EN UM +l hs +ĠIn stit +ipp ers +Line ar +Re ading +quir ies +-c ell +ch rome +.S earch +IN A +ç±» åŀĭ +ĠĊ ĠĊ +ĠSam uel +Ġmill s +Ġdon ate +ĠGe o +( rows +Ġshe ep +Ġé l +ä½ ĵ +Ġb em +_UN USED +ĠR CC +Ġintrodu cing +att a +ĠP riority +ĠF B +ĠSer ge +> "; +atch ing +ĠKnow ledge +ĉ The +; margin +less ness +op ard +um atic +() ));čĊ +Ġf als +(c ache +Type Id +éĢ ļ +_ choice +ĠGo th +ĠS ites +M G +_b order +Ind ices +Compar er +ĠRed istribution +Ġclo set +Ġvers atile +Input s +**************** **** +Ġob esity +qu iz +gr a +(g lobal +åĬ ¡ +Ġcollect or +Ġk or +ov able +AD C +ĠEvent Handler +. nc +Ġplay back +ient os +_p erm +_W ARNING +ĠOlymp ics +.n orm +ĠBroad cast +_sm all +dr ive +. iloc +Ġtyp ed +M EM +_con s +DM ETHOD +Ġl un +.d istance +(p ar +po on +Ġb ast +activ ities +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +: čĊčĊ +S ER +) && +_l st +ĠPol ish +Ġknock ed +Ġfrustr ation +au kee +Ġph osph +iqu id +_c oeff +æŃ ¤ +L atest +ĠD ust +T ipo +Ġmaint ains +Ġmar sh +inc inn +l bl +C are +Ġneighborhood s +_g pio +ĠAr senal +D em +ĠW he +_h ook +Ġl dc +ĠHar per +ĠBer keley +Ġgrad uated +Per cent +Ġarr iving +ĠAdvent ure +(s cope +(' * +qu arter +ĠMar ie +Spe aking +_code gen +Ġimm un +c aster +ãĤ Į +åķ Ĩ +ĠDim ensions +.rec ord +Ġtext o +ĠMich elle +P ending +( by +_P AR +uch t +be e +.Th read +amp ire +k now +ĠClin ical +Ġmargin Bottom +Ġdistingu ish +.F ull +. undefined +ĠSequ elize +################################################################ ############ +Ġeduc ated +_O VER +åº ı +ĠÂł ĠÂł +_e ach +Ġur ge +de part +Ġdon ors +ĠA u +Ġbill ions +Ġbelong ing +_ age +_ Int +Ġsub stances +m achine +!! !ĊĊ +Ġjson ify +ib bean +ĠC ad +Ġend Time +Ġc ycling +ĠUIT extField +Ġle verage +Ġvan illa +e at +La unch +( pt +st ates +ĠControl s +ĠRes pons +ĠJ ake +Ġas leep +fort unate +.next Line +Size Mode +ìĿ ¼ +Testing Module +G erman +ĠInvest ig +.re verse +ĠB ACK +( DateTime +Ġnon profit +ĠEx pect +Ġt anto +'] ), +ĉ the +M ultiple +(get Activity +_W AIT +Ġj á +de cor +lev ance +ĠGit Hub +min ation +_qu antity +.Sc anner +ĠL ion +éĶĻ è¯¯ +Ġd re +Ġtan tra +Ġcontent Type +Ġf id +_ alt +NS IndexPath +- pl +åĮ ĸ +Ġantib iot +table s +ac ial +ĠReg istry +Ġol ive +ig ers +Ġsubscri ber +_p res +ĠSy ntax +Ġlo vers +. Byte +old ers +_for ward +al ways +C aption +Pr iv +ĠT ampa +is ateur +-labelled by +ĠTo String +Ġì Ĥ¬ +Ġinit iated +W F +Ġinstitution al +in ject +ĠSc r +Ġdo ctrine +Ġsp acious +is ure +ĠAn a +" time +ess aging +Ġc id +ĠN an +Ġin complete +T AG +-b uild +Dec ember +Ġres idual +(P DO +ĠList en +Ġg lyph +Ġg aps +ne a +.R ect +Ġsa u +ĠPhot ograph +Ġexec utable +ĠExp ert +Cor outine +_s izes +ĠN L +.is Valid +); }Ċ +- reg +Ġc iting +c wd +ĠOtt awa +ĠB att +Ġrenew able +Ġprelim inary +Ġas ylum +Ġw rist +Ġutil iz +Ġdet ention +F ast +Ġan ge +incinn ati +Ġste ering +ĠNa N +ios ity +/ page +Ġè ¿ +ster ol +Ġdis g +( DB +ĠDESC RIPTION +Ġ_ $ +Ġobst acle +Ġb izarre +Ġextr action +_ex pected +Ġlos es +ĠCele br +Ġhtml For +Ġexplo it +олÑĮз ов +XY Z +Ġmagn et +amp ed +Ġat oms +S ources +pect ives +Ñģ ли +Ġ= čĊ +Ġd are +ĠWal ter +Ġbright ness +Ġan notations +ë ı +is ke +S chedule +. images +ros so +Ġ" .. +g amma +Ġin structor +Ġover write +- am +Ġdevast ating +ĠSaint s +Ġh s +Ġbon uses +$ output +ij d +(Action Event +mon itor +Ġmatt ress +Jan uary +.j p +Ġcar acter +Ġim pose +_re st +ĠSign ature +Ġcoron avirus +ãģ Ĭ +_com pare +Me asure +it ated +el ijk +ig os +es ar +Ġrush ed +met ry +_SE PARATOR +_W E +_ATTR IBUTE +Ġy aml +Ġspec s +ĠR ah +ph eric +ĠInvest ment +ä ll +Ġappe aling +Ġview port +ç © +Ġmargin Left +Ġsub tract +ĠED IT +ĉ ArrayList +gr ading +ĠF ailure +as per +EE K +(n ow +< object +ĠAl ignment +ple ado +q tt +( ERROR +ĠIN VALID +Ġuser id +ra ises +ID I +Ġvari ance +ĠN il +/ delete +_M AIN +.T oken +.C ategory +> )Ċ +Coll ision +ĠGre ater +ĠR acing +al an +Ġmon etary +, new +ĠS orry +. Enable +ĠInstant iate +oll en +ë© ´ +ĠCall ing +_h our +AD A +Ġsh y +) ** +Ġ== > +Ġes pecial +Ġinterpre ted +! =" +Ġpharm acy +.s ingle +ĠC ialis +Ġpar as +.to UpperCase +ĠDem on +Pr ime +Ġrank ings +Add ing +_H ASH +ĠEx am +Ú © +ĠVict or +Ok ay +"] ;čĊ +Ġfort une +ĠF ETCH +exp and +.Inter op +Ġb arn +æ ¶Ī +ue vo +Ġspec ulation +âĶĢâĶĢ âĶĢâĶĢ +ĠN u +ĠBl ues +(f name +Ġinhab it +Ġ\" % +C ES +ular io +_c r +Ġvalid ated +Ġmid night +ank ing +Ġincorpor ate +Ġpurs uit +EX P +pr ime +P id +- US +ĠN urs +ĠW heel +é ĺ +Ġin p +Ġsupport ive +.m ember +ĠSh ot +.Check Box +Ġaff irm +T or +Full Year +Ġconsider ably +cred entials +_ opts +R oll +( round +Ġcom ent +_U ART +Ġext ending +R G +result ado +it u +.get Session +Ġattr action +& D +$ html +ĠJess ica +ĠAssoci ate +a ñ +_ ed +ĠL ag +Ġorig ins +()) -> +add EventListener +IAL OG +åIJ ¦ +.Com pare +Al bum +ĠK u +< Q +arg est +Ġpro long +Ġconfig urations +Ġaccident ally +_ph oto +Ġ'' ;čĊ +Ġver se +B ob +Ġfarm ing +del ivery +ĠM ack +Ġuse Selector +.bootstrap cdn +keep ing +en y +. upload +ĠM ETHOD +cre ator +< _ +ĠE aster +. -- +UI Button +ãĤ ī +om eters +Ġsh ine +Ġh ogy +\ s +Ġh arness +.C ell +Ġlif ting +Ġcomb ines +ĠOcc up +ex clude +pat ial +Ġres pir +_f it +Ġfif ty +ĠM ol +Ġtun ed +-d imensional +Ġq s +Ġto ps +> ";ĊĊ +quis ite +ch annels +/ res +ĠAn alytics +.app compat +/ to +Ġon Error +( attr +IR M +Ġrag az +- as +.Se cond +orient ed +Ġdon n +Ġlight ning +f id +ĠP le +ãģ¾ ãģĻ +t ro +.Tr ue +O bservable +× Ļ +umb ing +Ġpros pective +-f ilter +Ġpurs uant +(p oints +.B ind +Ġp alm +clear fix +ö s +ĠG onz +Ġwe aken +Dr ive +en ido +l ld +ob ox +ane an +G ot +ä¿ Ŀ +Reg ex +æ ĥ +Ġsal ad +ass is +" net +inherit Doc +ĠR V +qu ier +Ġcl azz +ı ÅŁ +oster one +Ġair line +.list dir +Ġdownload ing +ĠP alm +w aukee +& lt +.B L +_IN LINE +off s +<< ( +_new s +Ġch ase +/ >< +Ġeuro s +ĠEgypt ian +ĠSt ainless +_BO OL +ĠG uild +ĠD ynam +[index Path +Ġ ï +Ġmemor able +ĠCh ampion +Resource Manager +.Log in +ĠForm er +yp ed +Ġl leg +; ", +D WORD +Ġtax i +Ġbom bs +ra h +.t ags +_test s +st ones +âĢĿ ) +[ g +r type +Ġv u +Ġhost ile +Ch ars +ĠPatri ots +/ status +< B +ĠIn come +ĠD ad +Ġpat rol +_CH ANGE +Ġup graded +Ġch ina +set q +Start ed +.U ndef +Ġcheck sum +Ġfrustr ated +{ o +Ġen f +Ġwood s +ĠAny one +Enc ode +ĠQt Widgets +are as +Ġshe er +sk i +end point +_T est +S oup +~~~~~~~~ ~~~~~~~~ +(f iles +ĉĉĉĉĉ čĊ +.sp ark +Ġval ued +Ġ% Ċ +.control s +ĠXCTAssert Equal +Ġf ame +ĠR ic +D OT +ĠAlbert a +ä½ ¿ +os al +.Web Controls +Ġ ------------ +ĠM is +ĠS YS +Non null += item +Ġexp ire +Dec ode +_ operation +ĠValid ator +.C ENTER +uff s +* m +Ġav ant +æ¬ ¡ +âĢľ You +.per mission +... ) +ĠL ic +_co ords +.n ombre +c lo +.Int ernal +ĠCh o +_s w +ĉ Il +cl k +Ġcast le +(l ayer +p it +Ġgu ided +Ġâĸ Ī +Ġsuper b +Ġsup plements +_c ent +Ġpe ek +IN ARY +.Content Alignment +f alls +")) ; +W all +). čĊ +ĠD anny +irm ingham +IAL IZ +( create +" In +Service Provider +Ġpr iced +mac ro +am ac +. box +---- Ċ +ãĥ « +ĠS uit +ur st +br u +ourn als +num ero +__ ()Ċ +D as +ĠM itt +ud er +? \ +f u +[ B +Ġ: )ĊĊ +(int er +br ains +Ġatt itudes +Ver ify +Ġsign atures +ack Bar +Ġg d +J ack +.c at +Ġz z +war f +FT ER +");ĊĊ Ċ +Al ive +IC LE +ĠWh atever +Ġout lined +s prite +еР² +_A B +_DE PTH +Ġcrush ed +aa a +(e v +æľ º +Ant i +IC O +is EqualTo +.s un +ic ulo +s ale +_h ex +ĠV k +apt or +Un ion +ĠDis count +list a +.Undef Or +Ġautom ation +N or +å¯ ¹ +åı Ĥæķ° +Ġref lex +ĠLa ure +.showMessage Dialog +.t emp +Ġa kan +Ġ__ ____ +.Is True +ARE D +ag le +E nergy +Ġquant ities +âĢĻ Ã© +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġcitizens hip +m outh +Ġin appropriate +ĠOut door +White Space +An onymous +load s +webElement Properties +T en +Ġacc idents +Ġadvertis ement +ĠY emen +(c all +Ġsl avery +Ñģ п +ĠL am +_BIT S +ome ga +ĠO le +Ġkid n +_A n +ĠR aid +Cre ation +s aved +Ġpro port +W ARNING +\ P +Ġp wd +Data Reader +is cher +ade on +ĠP redict +Ġreason ing +Ġdestroy ing +H el +* d +ĠLeg isl +_P r +ĉĉĉ ĠĠĠĠĠĠĠ +Ġsymp ath +Ġch ess +Ġm am +: hover +Ġconvert s +Ġp ela +Ġprogress ion +Ġ"_ " +ĠG ill +ĉ show +Ġsupposed ly +ac curacy +el in +Ġunf olding +ĠHy per +Ġw anna +Ġup s +( # +ĠCr iminal +( Point +at Lng +act ly +Ġcontract ors +'] } +draul ic +ód igo +ĠT T +ĠW ide +ĠAR G +_ ic +FLAG S +S chool +Ġclear ing +-be ing +={ [ +, const +man ent +Over lay +(' " +éĩ ı +ĠT imestamp +Ġmail ing +ĠC ake +.Th at +Ġmed itation +q p +Ġemp resa +ĠL ions +Ġw eld +ĠLinked In +Ġc ush +Ġgen ome +.Index Of +ag ain +Ġf allback +Ġcamp ing +re dd +-strip ed +Ġd v +Fe bruary +ĠPro xy +us k +Ġdies el +W RITE +RE AK +L orem +.In voke +- div +Inter ceptor +ĠD H +ia les +Ġvill ages +Ø ´ +ĠEN V +S ys +.X R +Ġpo em +à Ĥ +c ade +pl ots +Ġ{ ( +.g it +/s vg +nc mp +ĠÄ į +ain es +åĩ ½æķ° +Ġ( )ĊĊ +ops is +ĠRel ationship +_ aut +ĠB omb +ĉ com +* sizeof +off icial +_p ayload +ĉĉĉĉĉ ĠĠ +.m anager +ĠA round +ĉs end +ĠEx ercise +ĠB illy +iv i +Ġneed ing +_url s +_t asks +ĠH em +Ġtear Down +enc rypt +.t ie +Ġas m +IC H +ĠCGRect Make +ìĦ ± +ul ong +Ġit r +ĠG ST +Ġoffer ings +ro be +EE E +oper ators +_PRO P +ind ent +A DE +or f +ë IJ +Ġbless ed +vas cular +Ġcon oc +H appy +B ridge +ilit ation +j oint +ĠAdmin istr +- transform +Ġmeant ime +/ K +ĠBed room +Ġrig id +Ġbrows ers +EM PTY +.S erialize +_ ED +Ġst itch +Ġj an +ell t +Ġbr ace +Ġtr ails +p ublished +å¯Ĩ çłģ +} ')Ċ +Ġac ids +Ġ! !! +_d irect +> ());Ċ +aj Äħ +_O CC +Ġplan ets +æ Ł¥ +ĠDub lin +Ġser ie +.print f +de ep +` ) +Ġ\ $ +ĠÎ ¼ +_V IDEO +end ors +ĠC rypto +F ar +.Trans parent +.T R +ias m +_tr aining +Ġteach es +ĠB elt +Ġlimit ing +ĠK ath +ĠIndex Path +Ġachie vements +Ġser á +interop Require +Ġdis se +.I f +arm ing +uls ion +P o +_DE TAIL +Prot otype +ĠC AL +Ġagre es +.v o +.Execute NonQuery +ĠTop ic +Ġ' {} +Ar m +Ġe cc +M ag +Ġserial ized +ĉ conn +c ached += tf +ĠByte Array +prot obuf +var char +ĉ ASSERT +Ġlist e +_tr igger +· ¸ +Fe el +T ahoma +ĠL ik +Ġstruct ured +erg us +.In itial +_ ge +cl js +.cont act +Ġand ere +$ stmt +_C URRENT +ĠDis cover +$ res +form atter +H a +vang st +Ġem erge +ãĢĤ âĢĿ +ĠCabin et +-s quare +éĥ ¨ +Ġr age +ĠA J +ĠV T +sh adow +ĠFa ith +en ames +pret ty +has il +part y +Ġvar char +Ġf otos +Ġal um +ĠBelg ium +.y label +Ġde j +_num bers +Ġh u +.set Adapter +ĠUs ually +(s ample +.Sh ared +Ġbook ed +Ġ>> = +Ġmin erals +"> +pro g +bo o +_m d +_p ack +(ex press +ut z +\ Auth +, id +ĠCh ile +act ice +Ġrecruit ment +Ġpos es +Ġvulner ability +inst anc +or um +d ess +Ġx l +%%%%%%%%%%%%%%%% %%%%%%%%%%%%%%%% +( fig +Ġdelet ing +.d el +) ')Ċ +ĠWeek ly +?? ? +(str cmp +sm ith +Ġpurs uing +- so +ĠApp s +/ 'Ċ +Ġdec is +FO RE +Every one +Ġl anes +V irtual +. attach +( Log +ĠMed icaid +( Path +ĠTurn er +/ application +Ġport rait +Ġopp ose +check out +Ġfinish es +_M E +Bar rier +S ong +V AR +Ear lier +rell a +Ġh ast +az ar +Ġpull s +ng x +Ġinspir ing +Ñĥ Ñİ +-d irection +Ġexplos ive +Ġcreated At +st o +Ġwhe at +ĠB uilt +' ai +Ġtrack ed +ham mad +RowAt IndexPath +_ heap +D ue +Ġconnect s +.p ublish +em u +Ġbul lets +B AR +ol ate +Ġintern ally +Ġcatch ing +-p assword +ou ched +æĢ § +e ous +Ġx range +Q uality +v v +Man age +( ($ +ac ements +ĠBro thers +ĠHE AD +ĠUn supported +s an +es i +** *Ċ +Ġadapt ation +ĠWork er +'] / +.save fig +( trans +Ø ¬ +ne e +Cor rect +... ")Ċ +Ġsubmit ting +-p ath +ĉ last +iss an +.x label +ĠS epar +/ no +_b est +ĠM ills +_s ock +(f lag +Ġdest inations +em ption +ĠF AIL +å ĴĮ +Ġr p +f act +ĉ len +D AY +Ġse iz +_d st +l ip +.Line ar +ĠB asket +$ t +$ i +- brand +ĠNe il +ĠE q +Ġth ou +og ene +Ġscholar ship +æĽ ´ +Ġs wo +ag inator +en i +( book +Ġbl ink +th us +Ġcancell ationToken +ĠPalestin ians +Ġprofit able +Ġback pack +ens on +< Long +Ġp ools +Ġst icks +Ġspokes woman +Be ing +ĠHer itage +ĠN ike +SH A +ĠNotImplemented Exception +$ core +ĠR ico +/ latest +ĠC zech +ner Radius +(l ines +Ġsem ester +Ġw ounds +Pro cedure +.m ail +() ):Ċ +Ġcor rid +ter ed +ĠN CAA +Ġgal axy +_k ind +il k +Ġtr as +_P OL +ĠH et +Ġrefuge e +Ġteen age +.b inding +post al +Ġiç in +ĠData Type +é ĸ +ycl erview +, value +_id entifier +< b +Ġout file +čĊ ĠĠĠĠčĊ +Ġcr é +Ġrespond ents +ĠBe ast +ce led +Ġinter f +-th eme +g if +ĠR angers +IT AL +Ġauthentic ate +Com pletion +urs ors +Ġcin ema +Ġdisc our +ĠJ aw +OCK ET +Ġpr ayers +ĠL uis +fr ag +=[ Ċ +Ġbr ave +_p ose +C ertificate +- fe +ifer ay +ĠFl ags +Container Gap +ĠC rit +Result Set +ĉc ur +Ġcorrespond s +St aff +.Http ServletRequest +Ġneur ons +ĠMain AxisAlignment +ed ar +Ġg ad +_p arts +ĠÎ ² +Ġf x +/ files +ĠB ros +hip s +Ġgluc ose +Ġfar ms +Ġment ally +rest aurant +Table Name +ĠMer cedes +. Visual +Ġan ch +inal g +_r untime +Ġpropri etary +Ġintent ions +iz i +S lice +; "> true +ĠNY C +Ġb ored +ĠD etect +Ġapp ar +Ġje ans +ĠT ak +I OD +ĠH orse +( FILE +( ? +ri que +optim izer +n at +lo ys +ĉ Token +oub ted +u ess +oco a +Data Member +_P OWER +class List +Push Button +ĠWi Fi +. Stream +.g uild +Ġn og +ĠPortug al +ĠUnt er +Pr imitive +b oss +ĠDe utsch +Ġerot ic +Ġstr conv +.Try Parse +Ġgr ams +.S uccess +_p k +ĠHar vey +-m inded +.c ountry +[] " +Ġang el +Ġbe ats +ĠV or +il io +.m aster +s omething +ĠP ACK +( if +Request Body +Ġant es +/w idget +Ġmod o +ĠA W +find er +Ġoptim ized +Ġmiss iles +N B +ĉint ernal +t ex +ĠS ri +Ġdam aging +ĠM ais +- Allow +ĠZ h +- alt +Ġ ));ĊĊ +è ī +Ġinflu ences +Ġc atal +_REG ISTER +ĠAPI s +-cent ury +Ġbi ology +ĠAct ual +Ġhe els +TR ACE +_D IG +D ataset +ĠM atter +Ġclass ifier +.w ikipedia +ĠRog ers +Ġdon ated +raw ler +en en +Ġcas inos +ort al +Ġpr ive +s pe +duc ers +. ep +Ġgr asp +ac ji +Ġd airy +Ġb uses +.com m +. ins +ĠI RS +ĠBe er +ad c +o ard +_M ET +Ġ' +' +r ans +Ġkind a +ĠâĶ Ĥ +ĠM aur +аР³ +Ġband width +ib us +ĠD ifferent +(m at +ĠRes ume +_UN S +est ablish +Ġfon ction +Sub scription +_com pany +Ġlight ly +.con firm +.y aml +ĠBo ost +Com merce +- template +_DEL AY +ĠH I +Ġn avig +(S ender +ĠH S +_ "+ +ĠRE QUEST +Ġw ifi +=" "Ċ +]) -> +Ġro pe +Ġviol ated +Ġgl ance +ĠK urd +Ġè ® +de ck +ĠIS BN +Ġin fect +ĠF oo +Ġget ter +Ġt ener +ap pe +.h h +_h ot +< AM +p oly +! ",Ċ +Ġconver ting +ĠW WE +RO S +(' { +Com mit +) L +ĠO re +Ġsp arse +Ġdis posal +Ġcan celed +åIJ İ +Ġa er +Ġvin yl +á» ĥ +rec ogn +ark ing +Ġtrick y +* s +Ġproceed s +Ġis o +Ġco conut +Ġcraft ed +IEL DS +Ġquest o +Ġcomm un +_CON NECT +Ġtraff icking +De ep +a ções +c odigo +ve au +Ġbet ray +int a +T ED +æ r +m art +_B US +/ sc +ial ly +Ġcigaret tes +è¯ ģ +(n n +Ġmodel ing +/ products +w arn +Ġmet ro +ĠI v +& ) +ĠC able +Î » +Compar ison +g ary +ĠB A +P ART +Ġp v +_up dated +C redit +orth y +observ able +Ġthe atre +B LE +; }ĊĊ +la unch +_str ings +ug o +ĠR PG +- auth +Ð ł +hol m +ĠP and +U id +Ġim ply +ìľ ¼ +'] =' +/ User +Ġstr cat +нÑĭ й +Data Adapter +Ġland sc +Ġdipl omatic +ï¼ ĵ +************************************************************************ **** +ĠCh icken +Ġbc rypt +.In f +[ col +ĠQu antity +- position +Ġdiet ary +Ġfil mm +Is rael +Pre v +ĠMill ion +Ġrem ed +Ġbill ing +Ġout doors +.t m +Ġn ad +F org +Z Z +Ġs sl +], ' +K T +f req += document +bl ur +¬ ¸ +ĠJeff erson +C s +(s ave +Ġstr ap +Ind ia +Ġide ology +BO SE +ĠF P +( ans +Ġfe ver +ĠY am +K ing +à ² +AT ING +bo hydr +roll back +Ġnew Node +ĠN VIDIA +Ġhon our +ĠCon firm +xb d +Ġsuccess or +/ u +l iv +ourn aments +Att achment +Ġgr up +Ġtri be +Ġca res +e ft +_s ame +' label +Ġ ãĢIJ +M otor +Ġin exp +Ġ" (" +_POS ITION +Ġval ley +ĠResult Set +Ġpres erved +Ġmut ations +Ġquestion ing +mun ition +parse Int +ĠS r +ĠMet adata +âĢĿ ï¼Į +timestamp s +Ġtrans itions +í Ļ +Ñ Ĭ +i om +.D o +Ġp ine +Ġf ung +Ġtrans mitted +ct ime +ĠF am +Re vision +B as +UP ER +D estination +toHave BeenCalled +Ġun fortunate +IN ES +_pro f +Am ong +ĠCy ber +ĠB attery +gen re +ĠView Model +- = +Ġutil ized +p aint +.Integer Field +ern ity +comp iler +âĢĭ ĊĊ +ĠM asters +.To Array +Ġstrt ol +ĠUkrain ian +} ));Ċ +Ġsh emale +" That +for all +/ download +Ġrhet oric +.l atitude +ĠWH EN +Ġshock ing +IF IC +.N ormal +_F OLDER +Ġdr ift +Ġmount ing +- book +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ Ċ +ĠWire less +> ".$ +Ġrel ies +( Console +Int ernational +-> {$ +M id +Ġdis sert +dd s +Ġdepos its +ĉd river +# ga +pr ising +print ln +Ġpres enter +Ġmin es +C SS +ĠD ual +(! ( +Ġk am +Ġis Loading +ĠProt ect +. upper +ar ium +]: ĊĊĊ +Y ii +-sh irt +ĠIM AGE +_color s +Ġur gent +.Cont ainer +! (Ċ +S aturday +Ġsoci eties +ĠTh an +ĠC od += @ +Ġattach ments +.m obile +Ġsp ite +Ġb ounce +raw l +instanc etype +ĠTr uck +Ġmanip ulation +( Config +-in st +Ġst or +it ution +Preferred Gap +Ġmain AxisAlignment +Ġlist ened +'' 'ĊĊ +ott age +- project +.AP PLICATION +ĉ root +Ġwh it +Ġb ilder +Ġk er +Ġappl iances +row ave +ìĿ Ģ +ematic s +ĠO rg +op ing +_SE ARCH +Ġch am +add ContainerGap +Ġ( ). +ĠAr row +Il legal +Current ly +Ġus a +Ġpassword s +Ġre nown +av ern +ĠEv il +Ġconc at +Ġdu o +Ġv ale +ĠBe an +Ġindic ators +cm ath +ĠP ump +Nov ember +ific ant +_DOM AIN +reg ar +ĠPort al +" $ +Ġformer ly +"] :Ċ +ĠVis ibility +.getElementsBy ClassName +_RE D +Ġch ampions +à ´ +Val or +_ es +* a +-re peat +B and +.st age +Ġbure auc +C nt +et en +- function +Ġm uito +P ID +_ editor +Ġcrash ed +de ad +k at +ag h +ĠEX T +ass er +-sm all +Ġreal iz +( Entity +ú s +ĠAct ually +ĠEl ite +Ġhel m +(non atomic +ash er +Comm unity +all eng +ir y +ĠG rowth +Ġs ue +Ġfrequ encies +_des criptor +.At tribute +Ġrecip ients +_N S +/ "+ +ib an +Ġath lete +ĠI gn +_D MA +(d s +ĠRequire ments +AD I +ere z +\ Admin +br aska +ĠR ust +Rel ation +C OD +ĠV ERSION +em ma +)) { +.D uration +ĠC amb +- logo +Ġread able +Ġcre ators +() ];Ċ +Up Down +-h alf +.get Month +(s f +P ic +Ġhun ger +.t x +Ġexceed ed +_se ed +( ^ +_s k +.per form +Ġ> :: +Ġm ongo += float +bind Param +Sm art +if a +Ġse curities +Ġpre jud +Ġ, " +Ġcor ps +Ġv ra +amac are +it err +(M edia +uch e +Ġc ob +Ġlib er +. geometry +Loc ator +Ġsl iding +Ġsurg ical +_C UR +Ġcon sect +[ * +ĠRes ort +St ub +_DO UBLE +ĠS oph +Ġelect oral +_dis able +ĠÑģ о +ĠLight ning +Ġment ions +oc y +Ġle aked +Ġrelax ing +Pres enter +v sp +Ġgu ilt +=- =- +.re ply +ĠMir ror +C amp +Ġ+#+ #+#+ +Ġ+#+#+#+ #+#+ +.A uthor +Ġdirect ive +-h ook +íĦ ° +}ĊĊ ĊĊĊ +@ pytest +_r and +m is +Ġcolor ful +u je +lass es +ĠClass es +.h ave +% ), +é¢ ĺ +Ġdistur bing +sub string +ĠK oh +In vest +p urchase +Ġrec ycling +ĠA RT +ier archy +Ġf ps +.check Box +íķ ´ +_m aterial +duc ation +Ġf w +ud it +Ġreview ing +ĠS id +S yntax +ĠW ritten +arg ar +UM E +/ q +Class ifier +Off icial +Ġj azz +Ġom ega +Ph ysics +Ġl ugar +_access or +.command s +Ab ility +ĠB atch +R AM +Ġencount ers +. Qu +BY TE +ĠD istribution +Ġus o +ĠReco very +appro ved +Ġden ial +/sh are +Linked List +)čĊčĊ čĊ +udd y +Ġf ines +Ġr y +Un icode +ĉ render +Ġprem ises +Ġp on +ali ases +/F oundation +c uda +ĠC ock +,: ) +(f older +Ġm éd +dr ag +Ġtal ents +ĠĠĠ ĊĊ +е ÑģÑĤв +m ob +.y ml +Ġa ster +Ġdis cre +go al +ĠGT X +ĠS UCCESS +ĠL ONG +(f ind +Ġsing ular +_s z +ĠEth ereum +.. Ċ +Ġir res +')) {Ċ +Ġmin isters +St eps +ivers al +ĠNever theless +- led +Ġ( %) +ç¡ ® +Ġtime zone +Ġstr anger +(re nder +Ġsh util +Ġm ph +Ġtri o +pp y +Ġpred omin +Ġend ors +ĠRuss ians +ĉ row +Ġw izard +.s erialize +Ġcompl ained +Ġs ido +Ġdelight ed +-m e +ĠR av +H uman +ad ays +rec v +Work ing +J ump +ĠÃ¥ r +ĠAut omatic +_B ase +æł ¼ +aur ants + ¯ +æ ¸ +(C Type +IF I +( amount +Ġbelie ving += mysql +Ġf ir +Ġrest oration +ere co +Ð ¢ +_ '+ +Ġe book +Ġde bris +(input s +AY OUT +Ġscre aming +av ia +land er +Ġdist ress +Ġas sembled +ĠA void +( thread +ĠR PC +_EX IT +( queue +и ÑģÑĤ +D ll +Ġsk ull +_p ub +che z +min ate +ens en +Ġins ane +b ounds +ĠR osen +Ġcondition ing +process ed +v ideos +f our +.Con v +| ;Ċ +Person al +cer pt +:UIControlState Normal +Ġdos es +ĠKar l +ĠFre qu +.B ASE +ĠV ote +Ġcon current +ĠMessageBox Icon +Ġà ĸ +ĠDub ai +ĠR etail +: number +ĠOb server +ĠBig Integer +_ origin +_W ORK +F rames +Ġnot ably +. âĢľ +Ġtrop ical +Ġn iche +am ina +.s ys +(t okens +mod ify +os it +st rom +ĠCom ics +O PTION +T icket +Ġfact ories +Ġdis put +_F ile +ĠFin n +ee e +ĠDisc ord +_m oney +.t pl +_s afe +L B +Ġgl ut +J K +.fl ow +- cont +g os +Ġhor izon +ĠR ush +:: * +P ipe +ull a +bor ough +he imer +(m ove +( Text +} );čĊčĊ +w elcome +ĠCom ponents +Ġgovern ance +c losed +ĉm argin +Ġla undry +ĠTerm inal +iz ards +. âĢĶ +.rem ote +.r adius +ĠQue bec +Ġd h +T ech +ĠM ist +s eller +_l iteral +Ġgen ius +Ġbr ains +g em +ĠMe asure +Ġcata st +r ance +.Text Field +Ġconsum ing +Ġ'\ '' +oubted ly +ĠC ertain +E v +ert i +be ing +Ex perience +Ġ// [ +ĠArab ic +ĠC rist +ĠAz ure +Ġhor a +l adesh +\ Blueprint +d ar +.re l +Ġsup rem +ĠRe agan +ĠAt tributes +-s idebar +Ġuse Styles +ĠA irlines +Ġh ills +/x html +v inc +_m ock +Ċ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ +ĠP ill +.Layout Style +ĠCommand er +] < +sign ature +Ġ{ }čĊ +Ġhat red +Ġë ĭ +ole sterol +Ġ ******** +ancell or +c rop +T IM +ĉĉ ĊĊ +ys qli +uit ive +ĉun set +_s el +Ġmen us +t ick +Ġconstit ute +ĠElement s +ĠRed is +agg io +_f p +_de pend +em as +CA ST +or ange +j on +ĠEm ily +Ġpot atoes +Ġre ceptor +ĠElect ronic +ĠL ights +Ġcomb ining +ĠSome one +Ġ######## . +ĠT OD +/ show +X d +." ' +af x +Ġtr agic +St yled +ĠMar co +G allery +d ale +.âĢĿ ĊĊĊĊ +é rie +/s ervice +äº Ĩ +Ġamb ient +_SET TINGS +.Ad apter +l ene +Ġtrav els +Not ice +Ġcle ans +ĠF em +ch air +Ñĥ н +/ my +_b ad +ĠEcon omics +IS A +_C NT +(M enu +äº İ +ĠR idge +Ġlength y +D ot +Ġjump s +Ġhe y +$ pdf +Ġw orm +Ġs ut +Ġsh er +iam o +ĠCal c +trie ve +Ġc ops +ĠCh rom +Ġreg ulated +reat ment +ĠHigh er +ok s +Ġde ze +LOC ATION +ongs To +Ġfin ite +Ġvar ies +Ġposition ed +' il +éĩ ij +Ġh ike +(d one +play list +Ġad a +Ġcoast al +ĠN ancy +.DateTime Field +Cpp CodeGen +ĠSimilar ly +re ur +ĠCon tr +ĠH idden +ĠB eta +atch ed +_inst all +. Output +Look up +ĠRich mond +qu ared +Ġm anga +-control s +ĠBern ard +L arge +Ġslic es +Ġoff ence +ĠM ega +Ġest ar +Ġjoint s +Ġsum m +_pl atform +B uff +.add Subview +Ġret ained +Let ter +.d im +Ġess ere +ĠS caffold +EX PECT +ĉ RE +.long itude +ü nd +Ġstat ue +.add Widget +ĠCar ibbean +add PreferredGap +il de +UIL abel +ĠOp port +Ġimper ial +urs ion +Ġmand ate +Ġpromot ional +Ġv k +ia ÅĤ +Ġp yl +ĠCre ation +оз д +Ġsim pler +. what +ĠRec ent +St orm +. quantity +ĠL ov +" - +ubb les +_not ification +(w orld +ur ger +* (- +: "Ċ +h m +ans hip +ĠAl most +Ġmotor cycle +_f ee +Ġabsor b +ĠVin cent +Ġsound ed +ÃŃ st +Ġpharm aceutical +ht ag +ĠKind le +ital ize +ĠEm peror +oust ic +Ġspecial ists +åħ ¬ +Border Style +/ \ +RE LATED +(', ', +(ex pr +Ġh t +åį Ī +_C reate +Ġspecial ly +Ġ[] ;čĊ +Ġhe el +Ġse pt +_ arch +(in itial +% .ĊĊ +\", \" +Ġdiscuss es +Ġu pt +Ġ[ & +Ġman us +.h and +ĠM AIN +ĠDen mark +Ġ], čĊ +Ġcr yst +Ġn ack +Co ords +_in ner +Ġmid st +Ġaw ake +ĠÐ ŀ +-b reak +ÃŃ vel +_P ASS +ĠParam s +Ġdet r +Ġsp ider +ĠCon cept +Ġpre nd +CH ED +.Ex it +Ġpop ulated +Ġvirt ue +_SE SSION +Ġnou vel +o auth +Ġд аннÑĭ +r ink +.Header Text +atur ated +Ġer st +Ġå ħ +ॠĩ +_vis ible +ey er +Ġli able +Ġde be +Ġb w +{- # +_W IN +df s +H over +ĠP UT +- angle +Ġnob le +Ġtr aces +enc v +Ġuser Data +_in s +ĠS uz +Ġnews letters +ĠMod i +Ġentreprene urs +Ġtrib ute +Ġrum ors +Ġr r +ĠQu arter +ê³ ł +Ġfeed s +ó g +Ġen velope +Ġle ar +Ġk ø +develop er +Sim ilar +: ")Ċ +sub scription +Mod ifier +ital ic +Ġn asty +Ġtermin ation +Ġchar ming +Ġâ Ł +ton s +.tr ace +h ots +ĠU R +M ont +Ġjust ified +ĠG ang +ine a +Ġb og +( ap +_ $ +Ġcont amin +.D ot +ĉ Debug +( exports +Ġpa ired +ĠAss ignment +Ġautom obile +ĵ į +Ġph ases +v w +@ SuppressWarnings += \ +r ant +- ed +ĉ await +Ġcert ificates +'> " +Ġint act +CT RL +M ike +greg ation +AT TERN +Ġre public +_up per +ili ary +Ġcomput ation +h ire +ĠSh in +_ ANY +ĠManufact urer +ĠC arm +Ġbear ings +_c omb +c ad +ur istic +Ġwholes ale +Ġdon or +.inter faces +press o +ĠBr un +-c lose +pro ve +_S K +ĉf rame +et ros +ĠP ain +_EX P +ĠL T +_f s +.dat as +ĉ ss +vo ir +ĠA xis +M ajor +=" < +[ h +Ġprof ess +igr ate +(s core +Key word +" os +ĠĠĠĠ ĉĊ +an alysis +Ġre play +.p ass +\ d +t ls +Ġsan ct +.l ight +_m obile +ÑģÑĤ ÑĮ +ĉt otal +u ity +Ġpa used +N AS +Ġen core +lo e +Ġ-* -ĊĊ +.h igh +am pler +ĠSec ure +Ġfrag ments +_ vel +ill ary +ĠSte in +ĠD awn +Ġmax imize +ภ¢ +Ġ/ ^ +Ġcontin ually +Ġsh adows +ĉ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +ĠI ActionResult +Ġinform ación +C HECK +.Selected Item +b undle +ol ley +< Int +AIN ER +ĠW ing +tit les +ount ain +C Y +ĠLoc ale +form er +< context +R adioButton +_s chedule +Ġfab ulous +Rob ert +_PRO FILE +Ġg ates +IM P +ĠPent agon +g old +b ach +employ ees +R otate +Ġch amp +Ġsel bst +Al tern +Ġconvert View +/ , +Ġ~ ( +St reet +_ place +Ġpersonal ized +P ublisher +ĠSO CK +_NAMES PACE +ĠStand ards +so ever +_C ENTER +Inter est +ô t +tem perature +View port +get Resource +Ġeat en +Ġsem pre +Ġab normal +Ġc ylinder +Ġtroub les +n od +Ñĭ в +g ames +_g l +Pl ane +g rey +_t bl +.Component Placement +ĠCh ase +Log ging +man y +ì Ĩ +Ġfl ame +="< +Ġtra jectory +_r ing +Ġhydro gen +tr on +Ġstat ute +Ġcondition al +Ġtr ay +-s chool +(w idget +$ config +Ġrequest ing +. uint +et on +brit ies +Of Type +AD MIN +p redict +Ġg egen +ĠH app +OC UMENT +ĠA part +Ġ---- - +ro e +u ide +just ify +ĠSqu ad +Ġprof es +.b ot +_c urrency +inn en +ĠM umbai +ĠNum bers +avana ugh +agn itude +âĢľ There += http +çī ĩ +Ġv b ++' {{ $ +Ġin ode +s il +Ġh ace +Ġsever ely +ĠOver view +Ġspr aw +Ġbeach es +: left +· » +($ { +ĠF IRST +ĠSp a +- ass +Ġb aise +ĠN ODE +ĠP izza +P et +(se q +\ ">Ċ +CppMethod Pointer +Ġv p +Ġi a +_se conds +em et +/b lob +_TH RESH +... čĊ +D est +ĠN H +.data Source +it és +ĠJ ak +s ell +Ġwork shops +< u +Ġr ivals +ĠEX ISTS +h om +-t oken +compat ible +.J Panel +Ġphys icians +art in +Ġdes irable +Ġdistinct ive +.D ep +g id +ili ate +, max +Ġprem iere +Ġq Debug +Ġadvoc acy +Ġwh isper +P t +Ġun changed +_q ty +请 æ±Ĥ +Se ason +avel ength +ĠP ul +Ġd ÃŃa +'] ]],Ċ +al is +(" & +bor o +Ġb m +ĠR adi +w rong +ĠGo ing +ime Type +ij i +- feedback +ĠN ames +ĠB apt +Ġprob able +ĠE ther +ĠPolit ics +_prot ocol +lin ing +S at +Ġcor rel +.Pr imary +(null able +RI ORITY +Ġcolor ing +Ġutil izing +d as +Ġexport ed +Ġcar riers +Con v +. editor +i ó +(h andles +Ġapprec iation +. import +ĠAust ria +ĠStr ip +il ight +Ġappropri ately +ĠP rest +ĠW ir +ĠUI Application +al chemy +ĠM ob +ĠD etermin +ergus on +register ed +_con vert +ĠVlad imir +.Show Dialog +ref lect +Ġsh ook +Ġass ure +ĠO ften +Ġcivil ization +Ġvocab ulary +fore ground +ĠS cope +Ġunw anted +act ing +Ġ( [] +Ġmark ing +. original +ĠMO VE +Ġsport ing +ception s +NS Number +S izes +Ġprovinc ial +_Tr ans +Ġproblem atic +d igit +ĠEm ma +lock s +ĠC rew +ib a +') : +ish a +Ġm amm +Ġocc ured +w cs +(r ule +Ġmerch andise +es pecially +ĠT win +Ġn aming +Ġs log +Ġimpro ves +Ġad her +: text +.h adoop +_HT TP +.to List +.dis abled +Ġl enses +.in i +ĠR are +ĠUb untu +Ġsc ram +ol ation +tit ulo +Every thing +Ġnod ded +icht ig +_const ant +z c +l ift +ĠNot ify +ond o +ĠIN F +(" + +ĠK az +Ġd read +.m apper +le ur +ĠCome y +ĠN B +ic ers +.P ush +ĠH ack +ĠBrazil ian +_pro d +Ġ// ĊĊ +Ġb icycle +Ġun available +Ġadoles cent +bl k +Ġmit ig +_bl ue +ì ĺ +fade In +ĠUtil ities +ĠM N +; k +< style +- status +ind o +Ġinn ings +Ġg j +Ġ|| = +.e u +: Number +Ġcuis ine +ĠURL s +ie k +Ġw ires +ĉ ps +ie g +.m k +so ap +Ġsom etime +Ġst ap +_s eries +.T arget +æ º +.dest ination +OUN TER +R aises +& A +Ġsmart phones +NI Env +.s dk +Ġhelicopt er +Ġim pe +ĠB irth +A U +b readcrumbs +co ords +Ġexplo red +Ġl od +ĠI p +g able +ian e +Ġart ifacts +Box Layout +ا ر +list ener +.c art +ĠH uff +ĠHind u +ĠData Types +ĠDr upal +IGN ORE +Ġoffset s +ĠR TC +- login +æ ® +ĠQ Object +Ġprosec utor +R ock +_ch at +W ay +ì ² +Ġneg lig +Ġd ude +; < +Ġdeleg ates +_f ailed +/ dev +/ work +( New +et able +() " +( Icons +Ġp ork +ĠModel AndView +ĠV IP +ĠK or +m ix +Ġox id +ĠSC REEN +ĠFour th +/ ",Ċ +Ġte e +ĠSte vens +t icks +Ġp ledge +ib bon +ĠLo an +Ġne o +n umpy +ĠShared Preferences +- oriented +ĠLogger Factory +ĠGraph QL +zen ia +" _ +W omen +.c ast +Ġdeliber ately ++ b +ĠAr n +font Size +Ġm aze +Ġbl amed +.m as +} )čĊ +eler ik +Ġsc anning +ĠWork shop +Ġfind en +Ġca ut +UI Font +( return +al in +cast le +//////////////////////////////////////////////////////////////// //////// +Ġincent ive +op ath +b lob +Ġcigaret te +Ġfert il +*/ ĊĊĊ +ĠSh ar +Ċ ĠĠĠĠĠĠĊ +Ġunc ertain +ĠS ton +Oper ations +ĠSp encer +Ġdef in +ĠS olo +on est +·» åĬł +Ġu omo +G ive +Ġdent ro +; padding +ent ai +ĠC ars +Ġenthus iasm +ĠOper ating +S kip +par ation +Ġprotect s +Ġre ver +d g +ĠC incinnati +Ġconsect etur +Ġm uss +employ ed +a uses +ink le +. Values +£ ¼ +lo v +_W ARN +Ġbook mark +ĠAp ollo +. axis +Ġm ét +Ġop ener +Ġtum or +d an +Ġelement ary +Ġsk ipped +ĠK er +as ia +_res p +Ġdem ol +ĠCan adians +Ġt astes +U Integer +Ġ' ${ +.aw s +RO ID +ri ans +M Q +ord able +Ġcous in +Prop agation +(S ession +ph alt +UL D +ĠSc alar +Ġblo ody +Ġ ঠ+.m ask +, q +ĠUn its +Ġcent res +ĠPr im +. ]ĊĊ +ĠSh aw +P rom +ĠTh ought +Check er +_output s +( chan +E INVAL +Ġb ob +_c mp +P ed +Ġmat rices +Ġvrou wen +Ġgenu inely +high light +(d isplay +) != +Ġdel icate +ĠL uther +ĠM iles +Ġuser ID +% = +ate urs +_B UF +---- ---Ċ +imit ives +Ġsh elves +sl ow +_in formation +LE G +W r +.form s +cel and +/ un +: & +.âĢĻ ĊĊ +=" % +Ġpro st +Ġfont size +uc ión +get ic +am t +=" . +Dec or +B rit +Ġ"" ). +Ġfound ing +.File Name +ĠT ier +Ġdisc lose +á m +.s yn +.View Holder +lic ant +_st age +Mon day +Ġdes erialize +t alk +Ġtradition ally +æĢ ģ +Ø ® +LE X +Ġe h +ĉ ROM +Ġ{ })Ċ +Quest ions +nc py +Ġfix ing +к Ñĥ +_ Key +: x +ĠSTR ING +ĠÑĦ ай +ĉ left +ĠBen ch +ell ij +UR RED +ĠDi agram +} catch +/ time +ĠMiss ing +db name +Ġs ore +ĠW alt +ugg ing +rep resent +ĠG S +ne ys +ĉ page +Ġvol can +(b tn +Ġexceed s +Ġ erg +Ġpil ots +ĠS ed +ers ions +Ġpat ron +R V +/ top +. asset +_c ross +. Editor +.t b +Ġwel coming +SC REEN +) findViewById +C oder + ",Ċ +_P in +ues e +Ġover rides +_ ready +Adv anced +Ġop i +-c art +("/ ", +ĠDe b +CR Y +ĠVert ical +ĠO VER +ĠCorpor ate +Ġ"" ; +Ġste pping +e j +Ġaccus ations +Ġor az +_t ail +Ġindu ced +Ġel astic +Ġbl own +, // +Ġbackground s +âĢĻ une +-s dk +Ġset Interval +Ġincent ives +Ġveget able +_ On +exp anded +p ix +_sh ader +ĠSP DX +@ example +ĠW rapper +.Z ero +Pos itive +Ġsp inner +Ġinvent ed +ĠG ates +оÑĤ оÑĢ +Ġcompar isons +è · +.pr imary +data Provider +add itional +ĉ options +s napshot +.set Horizontal +Ġ" {} +ĠFish er +hal ten +< Type +Ġmax Length +ĠM t +Ġê° Ģ +.jet brains +Ġident ifies +Ġflow ing +ĠDisc ussion +ats by +Ġsch w +ught y +Ġr ivers +.un ique +_PH Y +ed ral +( ll +Ġcs rf +pp ers +ü l +ĠEs pecially +port ed +ĠHarr ison +****** */Ċ +Text Color +ìĬ µ +w ire +Ġstatus Code +ĠFin ish +c ence +ĠMcC ain +ĠW or +( await +Ġ) -> +ĠRegister ed +IN ED +k al +par ison +Ġobj eto +V i +mand a +Ġrenew ed +ĠS of +ess el +.nd array +Ġcr ap +ç® ¡ +.ab spath +( up +Ġclear ance +ĠT W +_C OPY +ĠĠĠĠĠĠĠĠĠĠĠĠ ĉ +Ġforest s +Ġarg uably +ĠA SS +he y +am el +_f ore +ĠSou theast +Ġab used +Ġpract icing +aked irs +ä¸ » +_res ources +Ġp ond +.F ixed +Last Error +ĠPsych ology +Ġ" // +! : +Re usable +Ġmens aje +Ġro spy +Ġb our +Ġvar ieties +Ġem path +(( { +_ org +ĠM es +ĠMag ento +IST ORY +Un less +Ġh j +ĠD uty +J un +, size +Ġpaint ings +Ġdisp ens +d art +Ġbehavior al +Ġr pc +cal culate +fr uit +_m m +ĉp thread +Max Length +Ġc urrencies +_cap acity +ĠO z +Ġfire arm +Ġcoeff icient +Ġbankrupt cy +w art +Ġfat igue +AV A +Ġes pa +_p c +ĠQu otes +_L IGHT +ĠT ickets +Ġrel ates +Ġpublish ers +Ġunlock ed +Ġ// ---------------------------------------------------------------- +ĠInterrupt edException +Ġout look +r n +Ġreb els +W ritten +Ġas ian +ot to +Ġ ĉĉĉĉ +_g pu +T xt +.Image View +Ġsu is +_t ables +.Rec yclerView +Ġwhat soever +è ģ +] ++;Ċ +assert True +_ verify +ĠR ivers +Ġ ][ +J et +id ian +S ibling +Ġgen res +.A ccess +OP S +Ġtr ivial +ภª +al en +в ед +ĠS word +Ġscrut iny +(c b +Ġcomm erce +Ġguarante es +_ad v +ĠL ET +rec io +Ġh ilar +Ġback yard +ãĢ ı +Ġillustr ated +/v endor +. Util +Ġw ow +LO Y +ĠMar shal +"> '.$ +ĠB ak +Ġmod ifiers +d ictionary +ĠSt re +m ultiple +")) , +ĠC ort +'] "). +( admin +ĠCre ator +Int ernet +( ms +log y +DECL ARE +ĠMarc us +<< << +ãģ ł +_m y +(in st +Ġsc iences +ND ER +. enter +Ġit u +Ġbeh ave +P an +omb ies +=' < +')) ;čĊ +ĠM ENU +ĠWork ers +.No Error +Ġbind ings +Ġdis abilities +{ \ +ĠM unicip +Ġco res +ur ple +ĠN okia +us ions +ĠF itness +.handle Change +Ġjav ascript +ìļ Ķ +( dec +Ġpack ing +-de pend +Ġtrans cript +z eros +_ alert +? ",Ċ +lib s +± оÑĤ +Ġ| ĊĊ +tr ained +ĠG ent +ĠR ab +x p +_config uration +å¤ © +_ accept +.rec yclerview +: url +ĠMu hammad +Ġprivile ges +_b ank +uk u +w allet +ĠRO OT +Ġenc uent +? family +ĉ position +Ġc g +Ġprec ip +method s +_f ast +in crement +ĠT iger +_OCC URRED +qu ip +ĠH AS +_d om +Ġw reck +b j +Ġd ern +Ġorg ans +. entries +Ġ_ (' +ram ento +ĠJam ie +Ġp unk +IP P +Ġprogram a +Ġatt ain +Ġpro ves +/s ign +Ġanswer ing +Ġl adder +************************ **** +ĠW almart +ĠCONT ENT +duct or +Ġver bal +ĠP ID +c rypto +_CALL BACK +Ġ= ================================ +Ġpot ent +Ġshort s +.U ri +.un iform +; border +ĠW er +Ġhere in +ll a +ĠI hr +P ixmap +l iteral +! )ĊĊ +g eneric +r ust +_script s +ost o +it us +ĠCoal ition +Ġrem ot +de ploy +ĠEag le +ãĢģ ãĢĮ +Ġimportant e +ĉ object +Ġseason al +ne j +aid u +Bind View +ĠSi erra +-b g +Ġmake Styles +[ offset +G ames +Ġhorm one +AR IO +head s +( select +ĠStart ed +@ param +_de cl +_b log +Ġa ño +\ Api +ĠMil waukee +Pro vid +An imated +Ġcool er +ĠSe ed +. Edit +Ï Ħ +ĠT aking +Ġborder Color +-found er +.Logger Factory +Ġ"" ĊĊ +AL T +ĠL ate +EDI ATE +Ġ);ĊĊ Ċ +af a +Ġcancell ation +At om +ĠB irmingham +emp resa +HE MA +asc al +Ġup side +.V ersion +ĠF older +ĠE ight +ĠV intage +ĠApp Delegate +ĠPre vention +.se parator +ST M +( room +gener ator +Ġc attle +ĉ Z +ĠPart icle +' };Ċ +Ġneighb ours +ĠState less +Ġalt itude +Ġsa int +об ав +Ġconv inc +ĠCont ents +Ġje une +(t s +Serial ization +(c ollection +ĠJ azz +ĠD od +ĠR och +ac io +comm ended +DEF INE +.on load +Ġspecial ty +PL ACE +_MO VE +Ġaccount able +Re uters +Ġf icken +Ġde pr +W ow +V oid +.s pace +à¸ Ĺ +Ġt q +ĠP ets +< $ +(C urrent +ber ries +plan ation +Ġlist Of +ĠTh u +ĠPR INT +Ġm ismo +Ġdo i +ch k +ĠUn icode +( role +Ġvir gin +< Point +_RESP ONSE +-h ouse +ĠVenez uela +EM AIL +Ġp úb +_ex ist +B all +.C L +re ferences +ĠBeautiful Soup +ĉ Expect +TH IS +Ñĥ д +b ane +Ġtemp oral +ER IC +et as +Ġrefresh ing +Ġsec ular +@ synthesize +ac cur +Ġn ella +ĠS OL +.p ipe +Ch annels +èĩ ª +Ġinsert ion +á» ĭ +el ia +Ġadjust able +Can ada +ĠI TEM +Ġcur ves +ĠChe ap +let ing +Ġoptim istic +al lo +Ġpolit ician +_down load += edge +ORT H +Ġmodel o +art o +. rotate +Ġs elenium +æĪ ij +_al ias +Ġrenown ed +.' . +Ġc zy +Ġal les +.Com piler +ĠB ass +Conn ector +.R ole +L INK +Ġc riterion +lem etry +Success fully +/p ng +Ġey eb +asp berry +( gr +Ġd angers +Ġcorrect ed +Ġgl ow +Ġelabor ate +ĠB ears +aw ai +=" '+ +Ġpromot ions +Ġmathematic al +Ġ" ` +_Generic Class +ĠChe f +.S ort +table Name +R IC +Ġvolunt ary +ĠBl ade +-e lect +ĠCom bat +ĠAb ility +Ġab dom +Ġd uck +T mp +åħ ¨ +Ġer ase +.P h +ĠDefault s +p artment +_US B +ê te +; ' +Ġp ads +ĠOb amacare +.T otal +Ġdiv ert +Ġcr icket +Ġrecre ational +( red +ĠC le +R U +Ġmist aken +ĠMont ana +Ġstr ive +_sl ider +ĠPl astic +Ġdecor ated +ĠV P +lic o +ĉf alse +Ġpre fs +( \" +_f alse +i endo +Ġ@ $ +B ucket +act ical +ĠZ hang +.c ols +.B inding +Ġw ax +_ST ORAGE +Ġlaw n +Ġr f +.Sc ene +ĠCal culator +.d esign +Ġres il +л ем +E mploy +ĠPr ices +ĠP WM +ag i +.e valuate +ĉ param +Ġbr ass +bb en +Ġinflamm ation +ull ivan +Ġan not +Ġp H +iam eter +ĠB TC +( box +Story board +Ġcl ay +.assert Raises +| string +.App ly +Ġmatch er +und ed +Ġsatisf ying +Ġìł ķ +Render ing +_app ro +ind rome +AN EL +_f ix +br ush +.M atch +Ġsm iling +on aut +S unday +Ġdelet ion +Ġencour ages +P ull +Ġreven ge +Ġqu arry +tr ade +Ġc ables +(d elta +ites pace +Ġf h +.b unifu +Ġvi el +_IN CLUDED +ĠT ail +ad ar +of s +Ġmet als +g om +_method s +Ġn j +.St d +(w in +$ (' +Ġt urtle +ur on +Ġen rolled +ĠH z +ĠBox Decoration +Ġp ont +rel ationship +B i +³ » +Ġmas cul +Ġsh ades +Ġv r +ĠLog ic +Ġa in +ĠD IST +Ġcoll ar +" profile +Generated Value +ĠP ossible +Ġe ines +ĥ ģ +.time out +ĠE c +Ġjer sey +.D ouble +Ġqual ifying +v or +CRE EN +_A pp +_rec v +Ġali ens +It s +E sc +i ator +ĠE clipse +Ġg h +V ict +ĉ html +to o +. const +Ġant erior +ĠW u +(key s +Ġul tr +_p oly +ĠT ap +ĠB ud +A WS +Ġcrash es +_t ot +Cont in +-h anded +alth ough +ภļ +ific ent +Ġde ve +ut ory +ĠW orth +_M S +Ġfloor ing +Ġsell ers +ĠThank sgiving +Ġp ng +Ġval ores +Ġslee ve +Ġfil le +Ð IJ +Ġappoint ments +Ġv im +User Info +BO OST +Ġpos ed +initial ized +.product s +ĠLeaders hip +man uel +' % +em arks +Per centage +(d ist +. avatar +(h Object +ä» Ĭ +_ iff +ic one +; ) +_n il +Ġab ol +е ÑģÑĤ +Ġven ues +.Con vert +! ')Ċ +.B itmap +sk in +_C OLUMN +Re v +G RESS +g ow +Ġw ished +tract s +.assert False +Ġscreens hot +Ġfo is +Com b +Line Width +ĠGr ab +Ġint ensive +ĉ sh ++ ) +.first Name +_PRO CESS +Ġt ilt +it ored +.L OG +Ġb ak +Ġintention ally +.play ers +(c anvas +)) )čĊ +.Pro vider +_P UBLIC +T alk +ĠL iv +ched ulers +Ġl c +ad ic +feature d +.res ources +Full Name +Ġmean while +B uffers +Ġres olver +ĠS AP +_T E +G NU +ĠForms Module +_ wh +ĠS we +.widget s +Ġcabin ets +Ġsus cept +ĠB ott +activ ex +av ar +ant ics +Ġ" =" +_k wargs +Ġgame Object +ĠAng le +.I ter +mar sh +ĠB irthday +ĠC MS +request s +ĠPear l +_E OL +Ġlin ux +( org +_M ouse +.con structor +Ġz d +Ġk icks +art isan +Ġe ax +K n +pon ge +ĠFin land +Ġmet res +ĠAss essment +part ner +/ pre +! ',Ċ +[ Int +Ġos lo +date picker +/ String +op lay +ĠHe brew +, double +Ġtrab al ++" \ +ĉ EIF +/ text +_F IRST +ĠP ete +Ġe go +Ġextr as +P DO +Ġreg ulate +ĠQ Widget +st s +ĠSh ows +ĠN HS +.c ourse +p thread +ĠF uel +.t imes +Ġ ° +Ġstr ides +($ ('# +( words +Ġrhyth m +Ġsp ont +Ġsens ation +Ġsp ike +C losing +页 éĿ¢ +N umeric +Ġbreat he +Ġfin ale +_F ACT +in ion +Ġch ill +Ġform ally +ANG ED +Ġ' :' +ĠпÑĢ Ð¸ +a q +ĠFab ric +(l at +ĠPr incipal +Ġer ro +oc ale +N om +Ġf ost +_C USTOM +.int ellij +ert ools +Ġcl asse +adi ents +Ġfundra ising +EN E +_OPTION S +_ ob +// }Ċ +Ġprote ctions +.se ed +N V +term inal +;; ; +P redicate +Ġì ¶ +Ġbomb ing +G F +Ġch ew +)) ). +qual ified +] ={ +list en +C ENT +d igest +E ast +Ġd iver +Ġend points +Ġe e +Ġcolle ague +Ġdissert ation +_com mit +_D AT +. rc +Ġbre asts +ĠR ug +ĠP il +Contract s +ĠBry an +Web View +Ġconcent rate +ĠIn ner +Ġ' | +std out +_S ub +> -->Ċ +V ol +ĠS SD +)) ), +. Optional +Ġnurs es +Ġor b +_ pe +);čĊ čĊčĊ +pl aced +ess er +Ġther apeutic +Ġwhites pace +Ġa ston +Success ful +Ġpr aised +ĠW es +Ġe ighth +ir al +Ġvrou w +Ġf action +_b ias +Ġw itch +Ġnp c +(s b +ĠRod rig +_b ig +Dep endency +ĠAb raham +ard i +C AR +n os +Ġabund ance +Ġnut rients +in stein +.V ert +ĠI SS +< U +Ġsum s +_h ist +Ġfar mer +ĠA br +Sh ot +ĠBad Request +Ġh ass +ĠR ails +Ġaffili ated +æĿ ¥ +Ġer f +IN F +ĠView Holder +min i +ĠR oth +Ġfaith ful +ĠPhill ips +AND OM +]. [ +_P AY +ĠAr ctic +f aker +D igit +M ale +std err +se ys +Ġ Å¡ +_rem ote +li que +Ġin def +ĠIndust ries +it ra +_p airs +< iostream +Ġsal aries +ik en +.F rame +PL IC +_S PEC +ĠMed iterr +Ġsystem atic +Ġinter rog +Icon Button +se a +int ro +ĠIss ues +enc rypted +Ġintern ationally +Ġsn printf +Ġpast a +ĠBrad ley +_ Status +AL K +_P AD +.l aunch +< select +Ġhar dest +Ġph y +Ġ(( * +-s lide +ĠNob ody +S u +Ġas ÃŃ +close st +_initial izer +Ġsupport er +-g en +Ġt ales +Ġcor p +_f u +s at +ne ighbor +.M igrations +Ġal gun +Ġsin on +.S pec +? ,Ċ +.G L +m ale +Ġmon itors +yl an +-L icense +.m atches +ĠA BS +ĠM ast +ĠW allet +($ ("# +Dir ty +Ġco pe +Ġinterpol ation +ous ed +ĠJ ets +.F LAG +.C ancel +.Event s +ne ver +ĠM Hz +> D +Ġs ervlet +bast ian +Ġ> & +S ID +_cl k +Ġdiv isions +} ',Ċ +Ġd ildo +Ġpar ade +m ajor +Ġab oard +; ++ +Ġf usion +"}, {" +ĠDialog Result +ĉ arr +- em +_n r +(h andler +.N ET +.Xtra Reports +ĠSh ah +ĠB rief +- , +Ġprec io +ĉĉĉ ĠĠĠĠĠĠ +Ġt ant +ĠGrand e +/ xml +_IC ON +ĠR etro +un que +Ġn ag +to Fixed +X L +Ġdecl aring +ĠCon crete +ĠAm azing +ĉprint k +Ġdeb ates +D ATED +Ġaest hetic +emet ery +Routing Module +ĠNash ville +W AYS +Ġw olf +Ġobserv ers +OT A +ans on +Ġe a +Ġgreen house +ĵį ä½ľ +Ġst air +Ġimmigr ant +_app ly +pe are +ĠBloom berg +_PL AYER +Res p +æŃ £ +Cho oser +ĠI Collection +P eter +Er ro +.detect Changes +Map s +Ġs queeze +ĠHom es +weg ian +Ġformat ting +Ġnegot iate +ul d +ĠN ep +ĠQ B +Ġeconom ies +Ġ*/ , +Ġredu nd +ĠA ber +.IsNullOr WhiteSpace +yc led +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĊ +_S h +Ġske pt +Ġre created +Ġget Type +Ġmarg ins +Ġcolon ial +ch arts +// @ +Ġprocess ors +è¯ ´ +b atis +æĦ ı +ator io +mention ed +P atient +Ġpre y +Check box +_x path +.s kip +ĠMorm on +ĠMemory Stream +CRE MENT +Ġk u +m eld +\ Data +ĠK ernel +il tr +éĢ ģ +( profile +Car bon +RO LE +( pl +] *( +.m emory +Ġmed al +Ġadvis or +it ät +Ġh dr +ier ung +ĠProvid es +( alpha +Ġteen agers +- parser +.L atLng +] ()Ċ +Ġfel ony +ĉĉĉĊ ĉĉĉĊ +BO OK +Ġsl ash +Ġclear fix +ĠPro phet +å® ¹ +right ness +-f i +.k ind +ert on +J im +Ġmanip ulate +Ġworks heet +ol in +st ars +Ġart ifact +_EM PTY +ĉm ain +------------- ' ; +Ġexpress ing +ĠI Q +ĠF act +/************************************************************************ *******Ċ +_m ass +)) : +Ġcon dom +Ġcreate State +omet own +Ġir r +Ġ> ( +> B +iter ation +ãĥ ª +Ġshirt s +ount y +-> $ +_S IGN +ĠD ale +Ġj j +E asy +F re +ĠN y +Ġch lor +match ed +ĠG erm +- UA +ĠN athan +educ ation +-y ard +- che +h ouses +r itional +Ġprox imity +Ġdies em +áºŃ p +Ġd rought +.a udio +ĠLe o +Ġfavor able +in ch +ĠD aw +rib ly +_st udent +id able +O VE +Ġlack s +ounc ing +.b usiness +Ġre open +may be +_G LOBAL +Ġdress es +ĠEd wards +ens ible +ĠHard ware +ĠEx cellent +ĠTime Unit +CTION S +Ġsched ules +Ġseg ue +Op ens +am men +- Identifier +Ġst aring +Ġhapp ily +ĠH ob +' _ +Ġ" ); +ament os +et ched +Ġ/> }Ċ +. Users +Ġinterrupt ed +Contact s +Ġreg istro +in burgh +CH A +_ imp +ph is +s ay +Ġretail er +.N ODE +/ maps +_L AST +ĠCh arge +_g uard +Coll ider +ĠStateless Widget +": [" +(" ../../ +iox ide +ĠS und +Ġ'' ; +un set +add Widget +л Ñİ +el les +alk er +A rc +Ġded uct +G UILayout +ĠV illa +Ġfor bidden +_ where +Ġ\ / +ĠT ib +_A X +] čĊčĊ +ĠB ir +Ġb end +ĠMA KE +ĠM ET +Ġfut ures +Ġweight ed +"" "čĊ +Ġauthor ize +(pro gram +}, {" +Ġcoeff icients +ê s +Per Page +ĠBath room +ĠPublish ing +G PL +Ġsub missions +ĠNUM BER +j Äħ +Ġaddition ally +em pre +ĠSh el +ot yp +S olution +Ġth under +_ ec +ĠĊ ĠĠĠĠĊ +ĠF ellow +Ġk ay +Ġnew State +ONT AL +Im plementation +.L ook +Ġ ents +Ġl ors +ĠB IG +f ab +Ġaver aged +ĠFe edback +ĠW ells +Ġm artial +Ġind ul +ĠComm unist +ĠFore x +ĠAgricult ure +" [ +Ġqu ar +ĠK ont +ĉ view +. Bytes +des ktop +ĠM akes +akes peare +.Null able +Ġspot light +V B +ow y +(t orch +tr idge +_b ounds +Ġapolog ize +.add Item +ant d +* );Ċ +, u +(g en +ç» ĵ +re ator +ĠC ord +ou pper +.m etro +Ġ ew +ĠW ORD +.A fter +Ġdet ained +ĠHam mer +ex isting +Ġo st +Ġmon ument +-c ustom +User ID +ĠN om +Ġre jection +(d im +Ġsingle ton +ĉd ie +ari ance +re ports +] != +eld a +Ġpreval ence +_reg s +." . +Ġfemin ist +Code c +Ġ **Ċ +(label s +_M ARK +FA ILED +Ġadminister ed +W N +ĠĠĠĠĠĠĠĠ ĉĉ +Ġn oun +w ig +Ġg otta +Ġr if +- im +ĠPaul o +ĠCommand Type +] ))ĊĊ +-z ero +Tr aining +Ġl ord +_ art +re ddit +C ert +Ġpes o +R ot +Ġend anger +.d r +user Info +un ts +n v +ĠTrail er +-f irst +(m ake +Ġbenef ici +-bl ack +i ÃŁ +Ġund oubtedly +Ġm ex +ĠAnc ient +( as +Ġdes cent +P ick +Ġrep lica +$ obj +ä hr +Ġar rows +ft y +ĠLib ya +ug a +charg ed +T ur +Ġh omic +iss en +ĠF ake +Ġbe ers +Ġsc attered +( Time +UT IL +Ġbureauc r +/pl ain +Ġstick ing +FA IL +ĠC ovid +Th ird +_p resent +ĠPier re +Ġë ª +Ġ[... ]ĊĊ +Pro b +ĠTra ffic +ica o +do ctor +Ġ), ĊĊ +T abs +al u +ï¼ļ âĢľ +Ġinher ent +_N o +rit is +ĠPro of +.b asename +ä¼ ļ +Ġch im +ĠProt ected +c rit +Ġpr one +Ġк он +ĠHero es +Ġan xious +Ġan os +Ġweek ends +Ġs ext +Ġredu cer += UTF +h alf +ĠS aw +.m m +Ġnue va +.current Target +.l ua +_EXT ENSION +ĉ reg +ĠC trl +_ align +accept able +Ġrush ing +fr ac +Ġbo asts +F ive + ± +ĠTem perature +> ): +Ġchar ter +RE ATED +Ġsubject ed +Ġop c +health y +使 ç͍ +ĠScient ific +Ġfra u +ri ages +à¸ Ķ +.in ventory +ation ale +M ad +min utes +>> ();Ċ +ĠEn v +Ġrecord ings +Ġsusp icion +sql ite +ĉ read +ãģ ¦ +Ġwor ries +.put String +ĠSh anghai +( uid +r er +ĠvÃŃ de +") : +Ġmethod ology +Ġк оÑĤоÑĢ +cc c +av ad +Ġindu ction +ĉ Thread +, string +ạ i +neh men +u ition +Ġ* __ +.em f +Ġì ľ +/th emes +ĠN ine +. One +ĠEm bed +Ġf az +u ations +Ġpriv ately +Ġl ing +[ F +ush i +Ġlaunch es +( KEY +G MT +Ġaim ing +pat ible +ĠB iden +i w +ĠD egree +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġ$ ('< +á rios +to UpperCase +ìł ľ +ĠE UR +Ġovers ight +Ġtable sp +Up dates +.m akedirs +Ġhum idity +/ template +Al ways +( IS +_c ert +D ig +Ġunder way +ort on +ĠHur ricane +Ġsp ends +ĠSeg ment +Ġfl ies +ĠT oggle +ĠLyn ch +Ġs enses +ĠK os +set Enabled +ist ically +Ġtest er +Ġadministr ators +Ġtag ged +Ð ĵ +Ġshort cut +ĠRes olution +Ġsuperv ision +ĠAsh ley +Tr acking +ul atory +and el +ist en +Ġun re +(d iff +ANT S +Ġr ider +Ġs Äħ +.S eries +_ orders +ORIZ ONTAL +Ġret ention +ãĢĤ čĊčĊ +Ġdi agonal +ĠC ancellationToken +_ Internal +Ġru in +.Q t +ocr atic +T el +ĠAn swers +m atic +Ġx p +at em +_j obs +_ any +Ġsen iors +Ġland mark +ĠQ List +Ġman eu +ot ify +/ ";Ċ +/ server +ĠPhil osoph +uten ant +( io +h z +Ġauthentic ated +d v +- Compatible +Origin ally +, function +ãĢĤ čĊ +ĠRepresent ative +as ily +irc uit +.d t +(m ath +.M arshal +[ , +ĠC ities +_ turn +| )Ċ +Ġcant idad +al ter +ĉ ui +ĠNe braska +Ġsk irt +.b g +Shared Preferences +( style +Ġg rief +g ew +Ġsaf eg +ol ang +_l ists +ì Ľ +Ġgran ite +Ġhott est +.j dbc +.C ustomer +Ġâī ¤ +Ġwa ar +_sc ene ++' / +ĠJ TextField +Ġse ating +Ġwe ars +Ġ` / +C ases +ĠY outube +ı m +Ġbal con +, G +Meta Data +- price +SC R +Un ity +Ġtr unk +={` ${ +Ġearthqu ake +Part ial +Ġsub st +Ġelim in +=" '. +//* [@ +Ġsuperv isor +vro let +_ article +Ġp ane +b io +Ġmot ors +N M +F rank +Ġon ion +- word +Item ClickListener +Ġb rit +end encies +Com puter +_r unning +( day +- he +(n amed +ĠS ach +о Ñĩ +c ampaign +.Ab stract +(w rapper +.p ay +Ġu w +Ge o +r ails +/ select +icht e +son s +E VENT +Ġal iment +Pro viders +A wait +_INTER VAL +. off +Ġgl uten +_cl oud +Ġw en +.ex tract +ĉ button +/ MM +Part y +Ġdem ographic +_err no +Ġh iking +(' ')Ċ +", @" +Ġw it +r á +olog ie +ĠSt yles +ĠBrowser Module +.Request Mapping +ic ans +P AGE +cre ation +ĠF erguson +ud ed +num bers +ĠGT K +Ġpresent ations +ĠB obby +_s pan +est yle +Ġilleg ally +abel a +Ġbattle field +cap acity +ter ror +] ");Ċ +Ġwar rior +le ader +ĠDB G +ĠRe venue +Ġvig il +Ġcounter parts +( Error +ACT ER +Ġhe eft +Ġselection s +ze ug +t om +-t wo +. ;Ċ +_st atement +ĠA id +ĠV ul +_r gb +Ġpr izes +Ġedit able +ĉ form +ın ı +.de cor +D emo +lic es +Ġen ctype +rat ulations +ĠR OS +_ch ars +ĠJ ahr +part ial +Ñĥ ÑĤ +ĠRe ceive +ĠL ands +AP TER +Ġch opped +.. " +ĠAn aly +ĠU ID +ĠR adeon +ĠB ee +Ġun m +> M +.find all +Token izer +ĠWH AT +Ġs j +D rawing +E ss +ON D +Ĭ ¶ +(p acket +âĢĶ but +Inv ocation +ĠN uclear +? ;Ċ +Ġgrand es +ĠC rypt +rem ark +Ġ'../../ ../../ +Ġin ability +m agic +c ats +Ġsim ulate +: ${ +in flate +Ġen er +: NO +ip les +Ġmer it +ĠR ated +Ġgl ue +/b log +Ġg ren +Ġthr illed +.C H +unc an +ĠPR IMARY +Ġper sec +Ġfe ared +.M IN +ĠThe ater +é Ĵ +ategor ie +æ® µ +Ġappet ite +s quare +ĠAlex and +.User Id +_g t +_ enter +Ġgradu ates +Fragment Manager +Author ize +-N LS +(M y +Ġtri umph +ust ing +_PARAM S +Char acters +(: ,:, +_B UILD +M Hz +Ġwash ed +Ġun cle +Ste ve +ard own + ${ +_confirm ation +Ġtro phy +Work s +ĠElect ronics +ĠMediterr anean +_m etrics +Ġannounc ing +ĠD AY +_pro to +Ġp ear +base Url +ĉĉĉĉĉĉĉĉ Ċ +Ġcoord ination +: N +.an imate +ĠC otton +_h it +â ľ +Ġjet zt +if ter +(f ields +own load +ific acion +.c uda +ĠLi u +> equals +ĠA ce +ÑĢаР¼ +ĠSuper man +ĠGarc ia +Ġarrest s +ag ar +Ġ{} ) +Ġmac ros +rou pe +ê tre +Ġtw isted +str uments +_ (" +_ vertices +ĠTrans ition +и к +[ max +m ind +Ġaccess Token +Ġun le +m us +c op +ĠF actor +Ġcon ced +Ġre tr +.l inalg +-s lider +ob l +_Static Fields +Ġz ombie +s elling +Ġch ap +Ġsh aking +ĠTrans late +ĠAm sterdam +ĠE TH +_EX TERN +k d +_d isc +Ġpreced ing +Ġpri x +Object Name +_mod ified +ard ware +Ġ?> "> +ĠD W +` ${ +Ġ?> ">ĊĊ +Ġspin ning +_p ending +Match ers +. Keys +ĠP V +en us +ant is +Ġdisc ard +Ġh aul +Ġem pir +Ġpath way +Ġo ak +м ен +-ind uced +Ġimp air +ĠCal gary +.is Hidden +d z +_ include +Ġg m +Ġ' (' +P Y +uggest ions +Ġcommod ity +c ro +/ sub +Ġget Instance +ĠLeg acy +ĠK il +B al +( short +In form ++ x +* r +ĠHope fully +or ate +Ġmach en +Ġtreat y +ĠO ri +.p ublic +-h orizontal +Ġtact ic +Ġb ord +w ares +Ġam mo +ĠL ists +Ġequ ations +/ her +ĠNS W +B ounding +_C ollections +Ġav ail +.Drop Down +è ° +Ġh h +Ġl Ãł +.p b +Ġmemor ial +ĠAT TR +Ġexhaust ed +Ġt sp +ĉ redirect +Ġlik ewise +ST ER +L java +Ġcondem ned +oca ust +(str ict +Ġexem pt +Ġs ms +Ġex agger +S YS +Ġl ounge +: ^ +Ġto dd +de b +ator ial +ĠPort er +Ġtu ition +Ġexem pl +Ġp aren +.line To +Ġkid ney +Ġç a +Ġc ui +ï¼Į 请 +X C +Ġmo ż +Ġnomin ated +l ung +Im Gui +ĠB uzz +Ġstere o +port al +res as +Ġk lass +Ġdraft ed +Ġproject ile +/g pl +(param eters +* )Ċ +Ġassist ed +ĠNS Integer +s itemap +:n th +.View s +.Argument Parser +Ġme er +z ier +ĠD ig +Ċ +Ġpl ag +p ine +Ġblank et +Ġ: - +Ġl cd +------------ --- +(" " +Ġtact ical +ĠRon ald +ex tr +ĠF est +Ġf uer +-n avigation +Ġk b +gh ost +Ġhandle Change +_cl s +() != +Com parator +.v m +ĠCo x +_re view +/ @ +_c ookie +Ġrecogn ised +ld ap +Thread s +ĠSex ual +ĠB earing +(S QL +Ġx r +Ġth igh +URL Connection +ĠSU V +Ġm Context +Ġinc idence +ĠE ste +.s up +_t e +(EX IT +C MD +/ "> +Al most +ĠU ne +Ġand eren +ĠSingle ton +Ġb ore +Th ink +Ġn arc +] initWith +_sh op +(str ategy +! ', +her its +ĠDes k +_m achine +.net ty +ı nda += < +ĠQ R +ĠS idebar +.split Container +Ġon Success +Ġmon key +En joy +(n odes +pect rum +Ġ(* ( +ĉU INT +, height +ĠNetwork s +.t ail +.l inspace +Ġ" ... +List en +Æ ¡ +.Ch annel +- defined +Re peat +ad just +ER M +_ application +.assert NotNull +- stream +Ġr abbit +Ġposition ing +Ġw oke +Ġf ing +Ġmulti player +Ġregister ing +un til +Ã¥ n +( :: +uss ions +Ġpot ato +ĠE quals +.S up +/ap ache +Ġ( = +. ") +.p tr +ĠSpe ech +.cl ip +ĠGab riel +Ġmusic ian +/ issues +.sh op +ĠH ier +_RE T +_b ucket +ãĥ ¡ +av s +Ġro z +fl ower +Write Barrier +ĠMil an +Ġlegisl ature +ĠD oll +Ġprov ing +.concat enate +âķ IJ +Ġg char +cdn js +b les +ĠList ing +л о +.xr Label +ĠS ak +just ice +ĠVal entine +un less +Ġp iger +(r un +Ġtest ified +AN A +ĠRem oves +)) ));Ċ +rec ated +ĠRuntime Method +Ġcon qu +ãĤ ¢ +Ġt issues +ail er +ét é +- Star +Ġfl ames +.set Icon +Ġsup ern +Ġvag ina +- variable +Ġwell ness +C UR +Ġbel le +.get Request +Ġp oco +ben h +ag ens +Ġsp ill +ĠJ ur +Ġdispatch er +н ого +emon ic +(dir name +ĠÐ Ķ +Ġpas se +Ġg anz +ric ing +E U +Ġmuj eres +ess en +.at tribute +j j +ĉĉ ĠĊ +[ ^ +Ġstrtol ower +lex er +ect ar +hot el +.s quare +Ġr all +Ġlower ed +hand led +Mark et +ĠUs es +iv as +.B usiness +ãģĹãģ ¦ +D IV +Ġw asted +Ġav oir +ê m +_ACC OUNT +. et +ĉ SDL +k ap +Ġf ox +up pet +{ },Ċ +", ' +F avorite +P END +ĠA ES +} ), +Ġded uction +Ġpol ÃŃt +Ġcomponent Will +ĠT elerik +_SE LF +Ġm use +C raft +Ġd ens +ठ¿ +( tp +Ġt asty +Ġbal ances +Ġded ication +ĠWall ace +Ġun law +\"> \ +Ġm um +- update +ement e +Ġs oda +Re public +as mine +é ric +( Status +ĠJson Convert +ĠD isk +.Red irect +Ġfilm ing +/m ol +R o +Ġv ille +Ġtrab aj +Ġsyn thesis +reg a +Ġr l +S cheduler +ISH ED +current User +(error s +' h +_b ot +x imo +ĠUS ART +_s uper +_DEC REF +н ой +_RO W +Ġprom otes +ĠT A +Ġhor as +ĠRep resents +Ġname of +ĠEx c +ĠGar age +Ġse ine +, # +Ġher b +/ resources +Ġple aded +.r adioButton +Ġæ ĺ +O ps +ĠN est +c string +ĠDef ence +Ġref ere +_le af +Ġrevel ation +ë § +.execute Update +_W ORLD +Ġexp ans +(" \" +j ab +Ġdoub ts +ĠGe ometry +Ġintrodu ces +Ġsen ators +Ġcan al +.h elper +ĠBi ology +_SE NS +.pre vious +-t ouch +ab it +Ġimpact ed +Ġbr ackets +.d irect +acc um +Ġtest osterone +ĉ action +ĠCh ance +Ġpe aks +CppCodeGen WriteBarrier +Ġun belie +_p ress +.R el +ang led +/ templates +-- >čĊ +l ime +Ġsufficient ly +_ nt +Exp and +.is file +Ġis Empty +Ġq t +Ġmul her +ac ob +Ge orge +å¸ ¸ +Ġass im +as o +Ġcompr ised +O V +(CON FIG +ĉw riter +Ġdes p +Ġten ure +(c r +.p ool +ĠB rend +Ġc ensor +(time out +Ġple a +.W rap +Ġtight ly +ĠW ere +ĠI gnore +abe i +Ġbr idges +Ġcondem n +Ġsimp licity +Ġrout inely +Ġblack s +j b +ĠP it +U tf +Ġ/ Ċ +re load +Ġset Object +/g lobal +Ġf atty +Ġsock s +Could n +Ġerot isk +æĿ ¡ +ĠPress ure +ĠM az +n pos +tol ower +ĠE Q +ute ur +ĠM oment +Ġet a +{{ -- +Ġgraph s +ĠGu ar +r ine +( -- +ĠHttp Status +(st udent +* np +Ġrail way +Ġas ynchronous +_v m +'] ,' +, text +mer chant +(G uid +ĠG ra +ix er +fetch All +.add Listener +fl ip +* $ +> (), +Ġsun light +ass igned +Ġab c +ĠC OLUMN +ĠðŁĻĤ ĊĊ +) ... +Ġen semble +Ġnew line +_S INGLE +ied ad +Ġdark er +orm ap +Ġl ion +pl its +Ġillustr ation +ĠI EEE +Ġv ista +ous ands +****** * +ĠTom my +Ġh ue +S el +Ġa ura +ĠTher apy +Ġanim ator +.con straints +Ġv ague +(" ") +Ġvill ain +Ġbless ing +Ġstring Builder +ĠM isc +ĠD IR +f ax +- node +ĠWalk ing +ĠA U +s ess +Ġgr ill +VERT ISE +ĠF oods +Ġt ournaments +à ĵ +ĠMar sh +Ġw onders +Long itude +.Command Text += input +_enc oder +page Size +Ġget State +> >Ċ +.g rey +p od +Ġread ings +Ġre consider +Start up +Ġexc er +.b alance +_c ycle +_T ime +LOC AL +ĠE FI +ĠRe yn +.set Foreground +by n +Ġdis connected +ACT IVE +Ġembed ding +ick ers +Ġsurround ings +* c +Ġgar ant +Ġb f +Ġw ipe +Ġ ä¸ĭ +_T RA +ado x +ç ķ +Ġsu cks +ĠS ongs +ĠAssoci ates +ĠB ald +ĠB rett +ven ile +Ġv t +Ġin ade +Ġres igned +ĠGl enn +.p attern +.Data Bind +Ñĥ м +Layout Inflater +ch et +ĠTest ament +.m s +Ġp av +ĠReact DOM +ur dy +AD ATA +M u +/ actions +ĠJ s +_ex tract +ĠBr ing +: id +str t +iv ation +Ġoutr ight +az u +loy ment +и Ñı +al do +ĠP ublisher +E ducation +Pa lette +_d rv +Ġ($ ( +ĠAnd a +Ġrem edy +Ġincons istent +te ction +Ġregul ators +Ġshort est +(p air +ĠInstall ation +Ġdefend ants +Ġ( ); +-l arge +M el +Ġthreat en +н Ñı +Ġfet ish +ot ine +_d ic +Ġ< $ +Ġst agger +sp i +$ response +S erv +-b orn +j os +ĉ img +ĉW HERE +_l t +å½ ĵ +.c ost +ĠT ue +.label s +ĠL V +wcs store +ĠJes se +ภ« +Tr ade +Ġpredecess or +ë Ĥ +fin ally +_g eneral +ogg ler +_REG ION +n ement +Ġblog ger +ĠHar bor +ĠD ataset +[ w +Ġattend ees +. ico +max imum +.Un lock +_SY NC +ág ina +Ġdown s +ĠW ii +]) / +Ġkick ing +unic ation +ĠD AC +ĠID S +ĠR ental +Ġcurrent Time +Ġvacc ines +ĠDev il +Ġn ors +_m ouse +urre ction +(n o +Ġ> čĊ +Ġaggress ion +Ġbre eding +.s ymbol +im an +Absolute Path +ĠWH O +_fl ush +- root +arn a +& M +Ġf athers +ĠR ocket +ive au +Ġw ander +Ġcom pos +ĠWar rior +ĠSe at +ĠClin ic +_in voice +(dis patch +Product o +at uring +oss ier +ĠM AY +Ġd agger +Ġsanit ized +ĠR FC +Ġpro ph +Ġur ine +Ġgr ind +ĠExp anded +des cripcion +-f w +ĠK erry += name +Ġch k +Ġnation ally +Ġthe e +In c +Ġ? >> +.R adioButton +.Http ServletResponse +/ Y +ĉf ield +Ġhom me +y per +Ph ysical += v +Ġdr iv +ĠErr ors +Ġc Äĥ +De ath +ĠW INDOW +Ġpo et +ĠSh arp +ĠImm utable +ĉ create +Ġge ht +ĠRe form +ais er +ĠInitial ization +Ġimm unity +.com pose +Ġlat ency +ĠLeban on +ĠPar ad +Ġfu els +ĠEx hib +co h +% ">Ċ +ĠCL I +) initWith +-Z a +_C LEAR +reg n +Ġfin ances +.st andard +_C ATEGORY +.lib rary +Ġtravel ers +_w p +ĠE valuation +start ing +Ġ )),Ċ +ep isode +ĠV ariant +Ġda emon +ĠJul ia +ĠN R +Ġdoub les +< v +/r untime +Ġinterpre ter +ĠIN DEX +ĠHol mes +_D IM +Ġp addle +_ex ample +Ġfore ground +.r outes +Ġs owie +S UCCESS +ĠC DC +ĠB D +_ - +as ured +W riting +Ġcurrent Page +( answer +ĠASC II +à ¨ +Ġsocial ly +yy y +ĠSpecial ist +(c ustomer +ist ani +ke st +ĠM ak +Ġth o +. pt +( comment +ĠCon verter +g am +b ins +. tele +ĠVeter ans +_AL LOC +олÑĮзов аÑĤ +inn amon +; width +oh l +Ġfant as +Ġs ung +ĉ K +( Json +Ġneighbour hood +Ġv ow +Ġs ins +on acci +Ġepoch s +im agen +.Ch ange +.my batis +Se ek +W ER +管 çIJĨ +Ġinter ess +_ Event +eder land +Ġterr itor +Ġci udad +uck ed +Ġsn ack +Ġtransport ed +ĠMan ifest +ĠD AT +_th eta +Ġw ont +.ĊĊ ĊĊĊĊĊĊĊĊ +Ĭ¶ æĢģ +ĠEp ic +De ck +l tra +_Z ERO +Ġ[] ; +/ scripts +Ġ---------------------------------------------------------------- ---------------- +æĥ ħ +Ġwe ed +N BC +Ġrap ed +ĠG ateway +[ M +ĠTime out +ench mark +.View Model +Ġporn os +ĠY a +th ritis +ĠFly nn +Ġme ga +ac in +Ġtrib al +.app le +ĠB lo +â n +ib i +ro v +ĠL ives +^ . +get Request +ĠEst ablish +cont ainers +Ġst arring +Ġcele brities +ĠRel ative +ĠHe ights +Ġtq dm +ĠNorth west +iv ic +ĉ cl +Ġautom otive +ent ric +Ġfort unate +Ġfire place +se ud +nick name +; s +_C AL +h alt +(n s +_de leted +Develop ment +m ovies +Ġident ities +Ġprompt ly +ا ÙĨ +Ġant e +Ġ" ',' +åı £ +imp se +Ġy ap +Type Name +Ġb itch +Ġassoci ates +HE ME +- empty +ĠØ ª +ol vers +Ġpist ol +Sc oped +ag ner +'] ==' +ĠI MP +ex c +Ġo mitted +Ġmind set +Ġ[] ( +Ġor n +_C AM +A vg +Localized String +ĠN atur +Ġcom poser +ĠPlay ing +Ġover d +_ utf +.s k +ĠF ol +$ page +, Object +Ġbe es +al ary +bul let +_lib rary +O ffer +loc ated +Ġ(_ , +âĢľ He +ĠOwn ers +) ).Ċ +Ġb ri +.Ad min +kt ion +лÑİ Ñĩ +Ġerot ici +Cancel led +Ġa gr +re views +_d ma +RI CT +Ġg fx +mp i +pp o +Ġ// @ +Ġupper case +Ġcommit ting +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +User Data +Ġv ai +ĉs ort +Ġcongr at +Ġd ioxide +д а +. area +ĠJosh ua +ĠK och +_b reak +az ure +ist ical +_AL PHA +_ views +Ġelim inating +OM B +en umer +ĠHy dro +(* ( +ERT ICAL +Ġinev itably +Ġst ole +-e ast +ier on +Ġl inger +/d oc +Å º +ĠAl ready +as io +Ġ-- Ċ +Ġabb rev +ĠAt om +h im +ĠINS ERT +s un +âĻ ª +CON NECT +er ator +ĠM anning +Ġ: ( +g as +=> ' +Ġquery set +; }čĊ +ĠPop ulation +uted String +res ident +_F ONT +ĠRes pond +Ġobsc ure +Ġo bservable +ĠContrib utors +k on +ĠMus k +ex ao +ĠT ub +Boot Application +S OR +.H orizontal +.find By +.p ower +Ġposit ively +ven ience +ĠJ ong +Ġwh istle +Ġз наÑĩ +Ġl ending +Ġdestruct ive +Ġon Delete +author ization +(); ?> +_ original +sc ience +at ra +?, ?, +ĠAs c +Ġconvinc ing +$ a +org en +_D ate +ĠPro vide +Ġlon ely +) 'Ċ +ex change +; ?>Ċ +.f ast +S amples +L ondon +'] )čĊ +ĠI onic +Ġp esso +ĠKn ights +ĠR af +_attr s +Ġrepe al +> Main +ĠOrder ed +_N ew +=" "> ";Ċ +ĠS ERVER +ĠHE ADER +_ velocity +ĠIn voke +.timestamp s +Ġs ulf +I QUE +Ġinhabit ants +ph ins +azz o +Ġmon o +Leg end +Ġnon ce +IF E +; ";Ċ +- create +" ",Ċ +per mit +ĠImm igration +Ġpath name +ffect ive +âĻĢ âĻĢ +Ġex ams +- event +ĠT ill +[m id +F IX +; color +( Order +_tra its +Ġorder By +Ġs unt +ĠNich olas +Ø ² +Ġsun ny +in ers +Ġaccess ibility +ĠH B +.com p +ĉ op +Ġminor ities +ethe us +Ġcollabor ative +pr it +H IR +Ġwr aps +ĉd raw +g od +ĠI X +.app s +ĠN M +Ġirre levant +ĠT igers +Ġdi ag +G V +ĠAccess ories +k ont +Ġsimpl ify +ĠF avorite +_t ools +([] );Ċ +Ġtow ers +B es +Ġhun ter +Ġsal on +(b uff +ĉ debug +Ġmal ware +M oving +- options +) +' +ĠLO VE +_S OCKET +_f in +ĠDel aware +Ġsher iff +-in valid +ĠF ULL +Ġп од +el as +" strings +ĠRepresent atives +s urface +res olved +ht docs +)) :čĊ +Ġpress ures +Ġnorm s +Ġpl a +Ġs urname +Ġpost al +ĠDep art +Ġsla ughter +or ida +Ġhe bben +Ġdes ar +comp act +_L ANG +åIJ Ī +op oly +_r ad +ĠST DMETHOD +L azy +ĠĠĠ ĉ +... , +( web +ĠP ont +Ġet was +Ġup ward +_h at +Ġ], ĊĊ +Ġbase Url +Ġworry ing +-add on +(get Class +S PI +Ġcapt uring +) },Ċ +Effect s +Ġcompet ent +Ġf oul +Ġsubscri bing +ĠO BJECT +IX EL +b ucks +( edge +(p ass +ĠPet erson +Ġbo obs +ĠD elay +_s quare +el im +ot ers +_P C +% E +on click +ĠSV G +Ġto pped +Ġf ist +sm art +ĠR alph +( owner +j ours +Ġbron ze +ĠArgument Exception +( original +_S CALE +_c p +Ġrecomm ends +.set Style +S ure +L AND +Ġrepe ating +M att +. Visibility +Ġenter prises +.Set up +(sc ene +ĠRe active +ur ge +b w +.P ut +p ersist +.c ookie +ĠAud i +` s +sup plier +( Form + ¡ +_s o +Į Ģ +ĠLeg ion +t te +N d +L oss +( attrs +.sc atter +Ġg room +Ġgl impse +Ġn ails +Ġcum ulative +Ġf azer +_s ervices +.N um +ib ilit +_res olution +ĠT x +umin ium +op a +.s chedule +sm tp +ภķ +ur ry +ü k +go og +_sign ature +.int o +ĠSte ps +Ġhome owners +ĠNS URL +ĠP AC +ĠĠĠĠĠĠĠĠĠĠĠĠ ĊĊ +> ')Ċ +en h +Ġinc ap +$ MESS +Ġmo ins +ĠF i +Ġoff season +press ions +> .Ċ +ĠGr ass +ĠGo al +_p df +Hand lers +Ġstack s +.get FullYear +=[ ];Ċ +è½ ¦ +, V +(s plit +Ñĥн к +Ġbake ca +Ġ~ /. +pe z +t ails +ĠG len +Ġset Image +ĠCom ic +B LOCK +ĉ This +o ader +Ġcapital ist +_ST EP +( Boolean +ĠCor rect +r ina +Ġconc aten +å® ŀ +() :ĊĊ +Ġun anim +ll i +al ars +- ne +Ġdiv or +ĠKick starter +]. _ +< number +/m enu +GR APH +vis itor +Ġimpro per +_N EXT +Ġb isa +background Color +/ input +Ġmo i +Go al +li qu +Ġmiscon duct +Ġcompr ises +aw ns +ĠP ie +ra is +role um +Ġcur se +y u +_p oll +.current User +ES H +]) [ +Ġstory t +)? ;Ċ +* = +ĠB urg +/ layout +_back end +; ?> * '+ +åĿ Ģ +ac ency +( URL +_h alf += l +Ġlist View +( section +.to Array ++ / +ĠRodrig uez +ist ream +Ġelig ibility +:: - +.new Instance +P B +ĠAs sets +ĠCom posite +ĠL abs +ĠHam as +++ );Ċ +Ġbl k +ĠNe o +L uc +@ login +Ġun aware +.m et +_RE LEASE +( ST +AM IL +ri ke +Ġ( ){Ċ +(s printf +ĠAccount s +ĠV IEW +ĠA j +ãĤ ° +Ġwh isk +Ġid i +Ġro de +Ġih n +ĠElement ary +Q ty +Ġintrig uing +Ġå ¤ +J obs +ĉ offset +ĠAh med +ĠTal iban +Ġè İ·åıĸ +Ġinject ed +.Auth entication +_line ar +.Dec imal +Ġapp les +Ġshare holders +Ġb aked +.d iff +ĠE ddie +ok ers +Ġconfront ed +vo ices +Ġt us +ĠSp in +N ODE +_ Un +CT X +/g oogle +Tem perature +Ġ' '). +Ġmagn ificent +Ġstart Index +semb les +Any one +z k +eh en +ĠD ame +. strict +Ġrepl aces +Ġline back +Ġpush es +Ġche ek +ĠSh i +_BY TES +RE A +ả n +_CON NECTION +G ateway +ĠTr avis +ĠA X +ĠBas ically +ĠUp grade +à ª +th emes +erm o +k or +F emale +_att ach +ĠìĤ¬ ìļ© +Ġpo z +============ ==Ċ +(s ymbol +ĠS ector +__ )ĊĊ +_p adding +ï¼ļ " +Ġf abs +Ġr anged +set Name +Ġp error +â Ĺ +ĠFile Reader +Ġful filled +_C urrent +Ġdom inate +Ġsm ugg +Post Mapping +_for ce +Ġb loc +ĠG iant +(v ideo +ĠC U +System Service +Ġ elf +Ġkont akt +ë ª +ke es +gt k +Ġparam Int +Ġmark up +u ales +Ġaccount ed +Ġgang bang +RY PT +ĠW rong +Ġcred ited +ĠM ESSAGE +Ġfl aws +Ġbb w +Ġmetab olic +ĠO EM +/ event +(C ollectors +mont on +ap pear +Ġopt ed +Ġche at +Ġd av +ĠPro ceed +Ġê ¸ +ank ed +и з +ans k +ĠH ang +ĠC ler +Ġdis gu +Ġc map +.cl js +Ġa ument +le z +ĠJo ined +_re ceived +Ġa erial +ot el +Ġgre et +" s +ĠGen esis +ĠCal if +pan ion +Ġtail ored +m apping +and Expect +.tr ack +at omy +ĠO w +ull ah +.Y es +ĠSimple Name +db h +' en +Ġnons ense +Ġphilosoph ical +(get Context +Ġis so +ĠA CE +start Date +Ġb ÄĻd +ĠAUTH OR +ĠGlo be +Ġinsect s +_A l +ush ing +è® ° +/ Home +ĠLocal Date +need ed +hes ive +Ġill usion +äº Į +Ġtr at +x o +/d etail +_M ATCH +Ġbroad band +Ġw al +ĠIllegal StateException +IRE CTION +Ġnor theast +es ium +ĠClient e +ul ance +nt y +Ġt ecn +Dev ices +Ġgr ains +ĠO g +ĠS EL +ud iant +Ġ++ ;Ċ +Ġexplan ations +oc co +Ġdi ets +Ġco hort +( controller +.Iter ator +-r ich +ro cess +G D +Ġcar bohydr +Ġfri ed +ĠEmploy ment +ìŀ ¥ +ĠLeon ard +_ ${ +qu ares +Ġcompan ions +Ġpar is +Ġstim ulation +ĠZ oo +Ġre levance +ĠCol our +Ġspe ar +ot ional +ĠL ite +ĠK osten +Ġà ³ +_att achment +orph ic +Ġdam it +Ġd lg +Ġthr ive +CH ANGE +ĠApp arently +Ġat ual +Ġroot ed +( images +aw i +ari at +Ġch erry +STAT IC +m nt +ĠUser Id +il let +ĠHis panic +Ġn ak +Ġcent ro +Ġdim s +_initial ize +ı k +ĠCent ers +RE N +Ġevolution ary +ĠTop ics +_d amage +em er +Ġr und +Ġpun ished +Ġcub ic +f air +[] ;ĊĊ +Ġinstant iate +Ġover see +- delete +unte er +start Time +ĠP ipeline +_G AME +ĠC ir +ĉ Null +.Format ting +uc umber +ĠR ide +Ġz oo +Ġcheck er +åIJ Į += C +Ġg rit +"); // +_x y +ĠDe claration +Ġcall able +F oo +ĠList Item +Ġin accur +ml in +ĉ Data +Ġev olving +aw an +Ġca fe +fol k +_ID X +ĠAny thing +ĠPalest ine +ĠGrid View +Ġcol ony +ĠGerm ans +( + +.p id +.js x +ĠSuper ior +Christ ian +ĠL ect +ĉ Game +Ġinstrument al +Anim ations +д ал +ĠMos es +ĉĉčĊ ĉĉčĊ +z s +k te +ä¸ ļ +_D IST +bit map +d B +Ġp ersistence +ÑĢ Ð¾Ñģ +$ l +B ron +Ġ{ | +_ch art +ĠCon sum +Ġh emp +Ġ" ))Ċ +Ġattack ers +Ġknowledge able +Ġc et +Ġvir uses +' I +Ġpitch er +Ġsweep ing += list +apt ops +.de pth +Ġinstruct ed +ĠR us +benh avn +Ġи н +S ports +Ġon set +æĿ ĥ +. RED +_s i +ĠP ST +.on Change +> tag +ĠR oh +_char acter +ĠLaw s +ĠB achelor +_s wap +.re activex +Ġreward ing +Med ium +- [ +ĠRec ently +J oint +part ition +ĠMin utes +Ġind o +Ġabsor bed +ĠG N +_IN D +Ġsab er +Sp awn +output s +ĠJeff rey +Ġmed ieval +h ed +Gu ide +Ġpsy cho +Ġgl am +E lim +äd chen +_pl ain +ĠS au +-f our +Ġanaly zing +QU ERY +Ġtom ato +_button s +V EN +.set Status +. Url ++ ĊĊ +Ġcompl aining +deg ree +conf irmed +Ġsub t +p arsed +Ġtor que +Ġtroub led +ĠT ARGET +Ġtrad emarks +ĠCo ordinate +ĠV iv +Ġ// }ĊĊ +Ġapr ès +.get Position +(Key Code +ĠSil va +Ġmet eor +Ġendorse ment +Over view +ĠP oss +.In ject +Ġeven ly +Ġvisual ization +Ġw char +ĠH DMI +Ġfun ct +ick name +',' ',' +Ġfor wards +Managed Object +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠ +ĉ server +ĠOut look +ĠChron icle +Ġdub bed +Ġd ok +ĠW ear +.A L +pare n +. Interface +Inter faces +.c od +Ġd ib +.Global ization +ĠAcad emic +Ġass ms +Aut om +Ġl w +ĠN W +Ġ&& čĊ +Ġproble ma +ĠManufact uring +lim its +-m obile +Ġfil me +/ map +Ġdo it +ĠIn k +Ġsu ed +. arr +Ġunder min +ĠPro c +croll View +__ $ +Ġsidew alk +( that +ภ· +[ q +gram mar +Ġt ë +qu ito +Ġspir al +ext ended +Ġf ocal +Ġdig ging +p as +ĠT all +.pro xy +it ures +TR ACT +ĠRe alm +Ġf eder +Ġorient ed +ĠAltern ative +Ġo we +Ġsour ced +ink er +.d et +S ep +ĠQ ui +ĠPal mer +(_ , +s amples +oy er +ull an +que z +Ed ges +Ġsh out +ĠA chie +Ġha ar +_Con struct +Ġprem ature +Ġre vert +'). Ċ +Ġs chn +filter ed +null ptr +S aved +itect ure +CL A +Ġv l +st ell +ĉ Me +ĠL ip +n ational +Ġwh olly +Ġspr ings +.T imer +ĉs rc +els en +åħ ¶ +Ġcommunic ating +ĠQu iz +Ġt eng +Ġge z +ĠOut side +.S ign +(c s +Ġdisput es +ĠWe iss +ann es +> No +ĠB ach +.remove All +re fer +/d ashboard +ĠA jax +Index Changed +ĠWe ak +' "Ċ +Ġs ights +access Token +ĠJ oi +(d omain +ĉc v +Ġcontin uation +Ġpl um +ad ir +.set Message +Ġ ï¼Į +Ġsw allow +ĠL amp +Ġq w +Ġu u +C oin +ub ic +ĠDe als +r ace +Ġdict ator +Ġmem e +turn ed +ĠJul ie +.grid Column +Ġpup py +Ġp am +Ġ) {čĊ +Ġinv iting +Ġf rench +v im +Ġwr apping +Ġ#- }Ċ +([ - +Ear ly +Ġsh iny +.f aces +Ġreb ell +abc def +ä lt +Ġest imation +ph ys +los ures +_RE L +Ġex clusion +ĠSk ype +we ise +-st op +no thing +ĠE gg +is ors +Rich ard +Ġcounsel ing +Ġcomm em +ĠQ MessageBox +ĠSy nd +ĠFro st +ĠCompet ition +ĠAw ake +Ġt ed +ic iones +ĠDev Components +VERTISE MENT +ott i +.run ner +Ġuniqu ely +.fl ag +ĉ rs +_g eneric +Ġ`` `Ċ +ACH INE +Ġme in +( Application +( br +Ġrat ios +: , +ĠXCT est +ustain able +- www +it les +_T EMP +Ġs yst +umeric UpDown +ĉassert True +Ġw f +. peek +ĠBul g +Ġterr ifying +.M ODE +ĠG W +á r +Ġf ic +Ġcommit ments +- tech +ĠL iquid +ope z +z heimer +a ña +-m edia +( animated +_go al +Ġg um +yst one +.S ET +ĠW end +set CellValue +Ġmsg s +c ash +AL LOC +/ aws +Ġmic rowave +.Point er +ĉ Console +_s orted +ĠFil ip +Pro d +Ġ//! < +ing roup +Ġk s +_T RI +Ġteas poon +ĠAT T +Ġrecover ing +ĠG LOBAL +.P ar +Ġ/> ;Ċ +Ġmar ble +ul ators +ĠC ycle +Ġher bs +_m etric +) ! +_C LOCK +_ Button +H arry +è¿ Ľ +Ġstr ains +ĠApp Bar +ĠCh an +/v ideo +Ġb am +.Pro gress +$ f +lem en +Ġir regular +ĠD uncan +ĠM int +-v ideo +ঠ¾ +ó wn +ĠEM PTY +Ġstack ed +ĠH A +_c ut +Ġwhere in +ĠW ays +(count er +è¯ ķ +Form Group +Ġble w +c ourses +Ġproduct os +ry s +ĠRest r +Ġsty ling +> s +Ġp iv +Ġit ertools +get Repository +ĠI k +_dev ices +lay ui +Ġhalf way +Ġfran ç +Ġtun ing +O A +_N ode +ar de +Ġfier ce +lic ted +# čĊ +Ġbreak through +ĠE rik +Ġb ride +Ġ. " +cul us +ins ide +ĠIndian apolis +ĠE E +Ġy og +urre t +.f s +. grad +_c ards +_ac curacy +_ep i +qu eda +/ org +é ªĮ +Ġcom pte +)) [ +Out side +G reater +ĠRender er +. actor +Account s +Id le +_h ours +ern er +Jo ined +Ġmen j +requ ires +ĠO PER +.remove Child +ĉs p +Ġes se +r ift +xF E +ĠSh akespeare +________ ____ +Ġbudget s +Model State +fill able +- component +oc os +ĠBUT TON +/ io +, out +s ms +Th omas +ĠAr med +res ume +Ġrot ating +ĠV ault +Ġse us +. (* +Ġa mino +Ġ[] );ĊĊ +Ġprov oc +no x +.Get Enumerator +==== ===Ċ +æĸ Ļ +_sc roll +Ġfil med +ĠS oci +g ap +g ro +V ote +" But +_R C +An imal +Â Ģ +ib ile +Ġaw aken +ore st +in ja +ĠI van +( Command +Ġ ***** +Î · +Ġkv inder +/h elpers +_c ases +t g +ìĦ ¸ +Register ed +ĉp ass +_d igits +Ġcont our +Ġinf ants +Ġjust ification +ĠFort unately +Con tr +ĠonCreate View +_S AMPLE +Ġallow Null +Ġn ud +Ġfet ched +_e qu +ĠUn able +=\" " +> {Ċ +Ġcommit tees +ist ema ++ ". +ÃŃ an +m ant +Ġsou theast +ï¼Į Ċ +dialog s +PRO JECT +charg er +- port +(u uid +. export +S ix +ĠR P +P rem +Ġconsc ience +Ġmargin Right +_d istribution +y aml +res izing +D ock +ĠLoc ations +G Y +Se ed +B UFFER +oss ip +ull en +Th ings +- self +.p oll +PL AYER +Ġå ® +G ROUP +ĠA way +Ġg ospel +xf d +M ary +ĠPort able +T URE +Ġutil is +Ġse it +Ġstr and +Ġtrans c +Ġ( ^ +ĠAl fred +.m em +.c ircle +Ġ~ / +for cing +Ġr iot +pro x +TH ON +iz ación +ĠN I +ro st +Ġdis pro +_in stances +ï¼Į âĢľ +ograph er +end as +ĠIsa ac +ĠP ine +/d is +Ġcolor With +iter ate +_str ide +Ġpun to +.Event Args +( center +Ġneighb oring +ĠPr ison +ĠMess enger +Ġepid emic +da o +_com plex +Ġgr avel +_D IP +é ment +ĠA ri +_bit map +.qu it +( valid +Ġp end +Ġrespir atory +Ġre bound +Default Value +ãĥ Ń +Ġcomm its +.test s +_f r +it et +.s f +Ġspace craft +c ritical +Ġde pressed +ĠAny Object +Ġun b +Ġdisc ern +(m ysql +L atin +ĠB og +ĠWild life +To File +iox id +@ RestController +Ġ"$ ( +Ġ<< " +Ġdefect s +Ġdat um +h in +Ġreal izar +any ahu +ĠS ig +@ Data +ad aptive +ĠC atherine +.c r +ĠCO OKIE +Ġp ictured +ĠFight er +Query able +ĠAny way +ĠGL FW +_n amespace +_ ft +Ġ] ) +Organ ization +Ġconstit utes +Ġqu and +(ch unk +"/ >čĊ +ĠL akes +main window +Car thy +sp in +(c sv +: red +-com merce +ภ¹ +Ġdiscover ing +Ġe co +_f ac +inc eton +ĠGre ens +j wt +Ø µ +ĠBron cos +ĠGood s +(G TK +Ġreturn Value +Ġsi empre +Ġneut r +w ent +ĠN atal +Ġenthusi astic +á» į +F N +/d atabase +C atalog +Ġbr un +ĠK ash +_P l +isc rim +, width +Ġin mates +Ass ignment +ĠH aven +Ġplay ground +ex am +@ Controller +ul iar +.get Parent +Ġ" ;ĊĊ +: size +iss ors +Ġf is +Ġal c +ens ation +ĠN ixon +Ġmight y +- str +_s pecial +_A DC +ĠTw ig +um bling +- address +Ġher oin +Y TE +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĊ +F riend +Ġa ve +ĠP NG +ĠKurd ish +DataSet Changed +Ġbl ades +br al +St eam +Ġsig u +IRT UAL +ac os +UD P +(d atabase +he c +ĠString s +_scal ar +ĉd esc +ĠT LS +; "Ċ +ĠCor byn +Simple Name +u ell +ĠEnt re +ell ites +- place +Ġfrank ly +ĠE rf +CE L +Ġpa ÃŃs +Ġh edge +Ġlat ent +ĠIR Q +ĠH erald +ĠP rec +ë³ ´ +.T EXT +Sal ary +Ġaut umn +Ġtrav ail +.S um +Ġc ared +M or +Ġint uitive +Ġj ournals +_ IT +ĠT rou +ä¼ ł +Has ColumnName +Com posite +Ġsp ice +_d isk +_CODE S +ĠInt roduced +ion a +Ġnue stra +o ct +ĠĠĠĠĊĠĠĠĠĊ ĠĠĠĠĊ +(param eter +Ġstud ios +Ġproject Id +Ġbd sm +.Sql Client +im izer +ĠC ARD ++ t +a an +.s ol +_Ad just +Ġright eous +ĠLog ging +.f ilters +_T AB +ĉs ys +roph ic +other apy +ĠB rowse +key board +R ON ++ \ +ro pped +Ġext ensively +f k +Ġl ime +year s +Ex c +Ġs ph +Ġche ating +and ro +ÃŃ o +Ġpr ince +o ire +ĠD estination +ĠConvert s +Ġup stream +o led +Ġserv ants +Ġsem antic +Ġcr unch +Ġevent ual +run ner +/ error +Sp in +Ġsecret ly +Ġas semble +.P erson +end error +_ < +Ġp endant +S leep +ĠChem istry +Ġboss es +l k +)) ),Ċ +Block ly +DE VICE +Ġreflect ing +Ġam ple +Mill iseconds +ĠPresident ial +Ġus uarios +ĠN Z +ĠSal ary +ĠA manda +_n p +j ury +Ġkö n +Ġtherap ist +Ġhomosex ual +ĠDr ake +-w indow +ĠLoc ated +.D river +ĠV IDEO +Ġmerch ants +ĠC hest +- lock +/ php +Ġmil ano +_ST YLE +arg er +ide a +G UID +adv anced +me al +Options ItemSelected +=' % +ĠCh am +: data +(st at +Will Appear +Ġinform al +aj i +Ġre productive +ĠC AS +ãģ £ +F UNC +ĠR uth +)+ ( +CON ST +ĠF ans +Ġgroup Id +xffff ffff +Ġsam pler +Ġ}} "> +. the +Ġh ollow +W AY +ĠFac ulty +Attrib utedString +ĠLook s +ĠR ex +j k +ĠM IL +Ġb ard +.L ong +Ġliv est +Ġsk al +ic ism +MA IN +Ġmu cho +B ODY +Ġes e +ĉ use +F oot +.SQL Exception +Ġinherit ance +re ceived +Ġput as +ed is +als a +ĠError Message +Book ing +Ġtr act +ac z +ĠC ant +_reg ex +Ġide ological +Ġj ihad +h os +/s ys +col m +(p ool +Ġest án +ĠP ending +em ás +Ġktó ry +));ĊĊ Ċ +trans actions +Ġw ield +it ere +ert ure +_s s +Ġstretch ing +Ġprison er +.Read All +Ġbes ch +-- ;čĊ +Ġcr isp +_SC AN +Ġa e +Str ict +ĠMin neapolis +ĠBo eing +ar is +re k +_p ipe +Ġpri ests +(E IF +eh icles +ĠInter active +b etween +ĉNull Check +ĠBl air +ĠL t +_in line +eth yl + ¼ +_p ackages +Ġbarrel s +_ he +Ġreg exp +_ pts +_H andler +ing ular +ĠN issan +ĠR anch +Ġper ch +Un supported +Sm ith +ĠLeg ends +M i +Ġg f +st eder +Ġacqu iring +Ġsim ulator +() ," +re ceive +Ġin place +A CTION +ĠWeb Driver +files ystem +< Order +lo pen +ĠHE IGHT +.set Border +į ° +__ [" +Ġcl amp +Seg oe +b ands +to List +amb a +>' +Ċ +Ġcred ible +am at +play ing +.setImage Resource +qu el +Ġpod r +ge om +E k +ĠQ atar +Ġg eld +? ',Ċ +Ġc yl +( ax +ĠW I +ur ally +ĠBr asil +Ġsen za +ale y +on en +Ġb ah +Ġmolec ule +R ad +è¿ ° +AN CH +- background +- agent +Ġprol ifer +: boolean +Ġt ide +erial izer +_ ;čĊ +F ee +** ) +erg y +ĠHon or +.Log ging +ir is +Ġunder mine +ĠD y +Ġt yr +Ġde que +Ġdam er +([] )Ċ +.layout ControlItem +pe ated +C AN +rag ments +L and +) ]);Ċ +ĠS ah +ĠDE CL +With in +ĠN amespace +an other +sem bling +.des cribe +Con sum +ĠF ear +g iven +Or ange +< boolean +Ġstead ily +pa Repository +Ġresult Set +_ ENTER +_re peat +Ġt ones +ĠPRO P +n al +part icle +Ġsign aling +Ġaccess ory +ĉĉĉĉĉĉ ĠĠ +Ġvie le +ĠNo ah +- ag +Ġmur ders +Ġa ired +ĠPL AY +ĠS ullivan +_C ore +Ġul ong +Ġblog ging +> This +Ġdata Index +Ġprint able +ĠE yes +_target s +(P y +. over +Ġbr u +am pton +Ġplaint iff +< Key +b ull +Ġ⣠¨ +Iss ue +.cor nerRadius +C ritical +_p hi +. angle +Ġdynam ically +! ");čĊ +> );Ċ +in vest +.* ĊĊ +Ġt élé +Ġsuper f +Ġcas cade +DT D +Ġviv id +Ġsubsid ies +ĠH ass +Ġcoll aps +Ġcer amic +{} ". +ĠLeak age +-tr ash +coll apsed +-s ocial +ĠCh ad +Ġincl ined +Ġst o +Ġstory board +.p ayment +stack overflow +ĠRaid ers +Ġ# ' +olic ies +ìľ¼ ë¡ľ +em ap +Ġk j +Ġqu ota +ĠGard ens +ë² Ī +ĠAng els +Ġof t +Ġlower case +Ġi Param +Ġche apest +un ta +_p kt +ic ators +Ġle urs +Ġdecre ases +ĉ define +PRE C +amm ers +ĠPre paredStatement +(d irection +Ġcre ws +ark ed +ĠMem phis +ĠS ell +G TK +Ġm aid +: disable +éĽ Ĩ +ĠP f +Ġal beit +open h +?> ">Ċ +.get Source +(s cale +D u +ĠP IL +_ref resh +Ġbet s +(c ar +ĠV on +| --------------------------------------------------------------------------Ċ +ĠGr at +M uch +( Dialog +.stop Propagation +Ġte k +Ġex its +'], $ +Ġphone Number +uc s +ec imal +------------ -- +in p +.po jo +Ġcor pus +Ġpractition ers +.p ic +" testing +Ġstring By +.Not Null +Ġr ang +.D ynamic +_R ender +аÑĤ а +Wait ing +ĠW ik +Ġoverwhel med +% "> +ĠA E +}} >Ċ +u w +_t yp +Ġbuck ets +Ġgre eting +Ġla ughter +Ġant agon +uggest ion +- email +ĉt op +Ġer os +_tr i +Ġiss uing +Ġh á +Ġisol ate +Over flow +, E +Ġnut ritional +ĠAbb ott +Ġn f +.t ouch +.fetch all +_z ip +") }Ċ +Ġam at +ĠC isco +Ġn Ã¥ +PLE X +Ġse i +f oto +.to Json +å¤ ļ +ĠKle in +Ġlib c +Ġmin ers +å ¢ +- print +ĠP ride +T odos +Ġmask ed +Ġset Data +Ġtele fon +Ġunh appy +ĠT ables +ge b +( debug +_all owed +- access +Ġlog istics +Ġg ems +ĠM ature +Ġr sp +ĠAl le +.get Bytes +\ web +ynchron ized +Par agraph +Ġth rottle +.sql ite +cons ulta +ĠSe ah +C e +Ġsub mar +ER E +V ous +Ġre ddit +Ġsql alchemy +-m ile +oc ide +P our +}} ">Ċ +st ead +Ġ@ ( +Ġ[ ]) +ĠAd s +Ġover load +r idden +ĠDes ert +ĠW rap +ĠPortug uese +et z +ĉf irst +Ġmile stone +æĹ ł +Ñĥ Ñī +(s uccess +< Vector +co ol +Ġ[ ]);Ċ +erv als +Ġin vert +" io +cur so +fr agment +Ġfeas ible +.set Position +Ġel m +Ġimag in +@ Spring +Ġb ats +pu és +ga lement +ns ic +gi ene +ell ation +ĠBa iley +Sh ar +ĠT ul +ĠH K +Ġfree zing +gl m +ce ans +-c ut +_c ircle +åij ĺ +n egative +Ġind ian +s alt +Ġt ing +ĉm od +Ġs int +ak in +um l +ĠText Input +Ġpop ped +T MP +Ġpark ed +×Ļ × +ĠF usion +Ġhe ater +ET F +ro zen +h all +ĠM ik +lev ard +- heart +ĉ order +M aking +Ġpled ged +Ġdir s +$ post +ĠH err +stant iate +, "Ċ +.get Color +ĠS AT +Ġtimed elta +ĠM ai +ĉm ethod +Ġid iot +ĠTr av +ident ified +ĠDiv ine +.get Path +D ash +Ġinf iltr +Ġhandle Submit +bro ok +.g eneric +.short cuts +................................ ................................ +Ġdat ings +ĠM V + # +} "ĊĊ +Ġimprison ment +ason ic +rou d +uc ion +æĬ ¥ +Ġdia lect +Ġon Mouse +const expr +.label Control +Ġwe aker +Ġman kind +ĠRE CE +Ġd iz +Ġapp Bar +Ġqu é +f ra +_default s +Ġal iqu +_at om +: indexPath +Ġmiss es +Ġvis ually +ĠH ands +STR U +i ates +_ asset +F inder +mid t +Ġsn acks +(__ (' +. uri +ĠIn strument +ven ir +($ __ +.Dot NetBar +Ġconfig s +Ġguess ed +ि ठ+Ġinitial izer +Ġ? ", +ĠVer izon +man ifest +ge ben +.d etails +G ate +pons ible +ĠEl im +, str +Ġwrit ings +ĠD erek +ĠCo ordinator +Ġpill ow +Ġnotice able +R s +Ġduplic ates +ern els +k J +.z z +oll and +ĠSE CTION +_f name +uff led +'].' ")Ċ +ĠD ollar +Ġem oji +Car ousel +- player +Ġadjust ing +Ġjug a +alleng es +g ene +(body Parser +lop edia +ĠBeh ind +Ġslee ves +Ġdrag ging +ĠChe vrolet +Ġb iz +iv ities +ĠFrequ ency +, char +.W HITE +_pre view +) ';Ċ +_ ax +ION S +.c pu +.input s +UB E +_fe ed +ĠSup plement +! ). +es us +ĠU DP +Ġmicro phone +Ġconf irms +.is NotEmpty +":" ",Ċ +_S CREEN +ĉ expected ++-+- +-+- +ĠH ait +fast call +Ġdep ict +v b +_p icture +ĉd escription +ĠW ife +uc i +Ġv icious +ä» ĸ +ue ba +Ġset User +ãģ ¡ +Ġd iving +Ġoper a +user content +ar ah +) }, +y un +vel t +Ġun covered +Ġh ips +Ġosc ill +Ġassert ing +ĠX i +.re store +ke a +Ġsp elling +Ġder ive +ab we +ĠD ow +.set Type +_v s +Ġco zy +.c ategories +O rg +_m gr +Ġd ungeon +collection View +ĠBl ank +ac ias +ä ä +_clean up +_ACT IVITY +Ġtri angles +.Menu Item +Ġip hone +ĠW on +] ]ĊĊ +ĠCompar ison +.D oc +Ġcan onical +ĠSud an +') { +Up Inside +b uiltin +ENC Y +x be +Ġch uck +Ġcontrad ict +Ġnuest ro +Ġarchitect ural +ĠF ib +Ġcomp ares +* k +C fg +çĦ ¡ +nt en +Match es +ĠDOWN LOAD +_HAND LER +man agement +[ S +EN G +ÂĢ Â +f ang +Ġsl ipped +ĠL anka +esc aping +Ġtack les +ĠPed ro +.P rop +.' ' +.G enerated +.New Guid +at rigesimal +ill on +Ġstat istic +spec ies +hold ing +Dr upal +Ġfundament ally +Ġbond age +Ġres olutions +Inline Data +\ Type +est ion +.w rap +Ġwar riors +ĠLOC AL +Arch ive +Ġembr aced +á» § +.V er +ĠAff ordable +oles ale +ĠAp plied +ĠCon version +m ega +_c am +Ġcer emon +aur us +ĠVol k +.op ens +/ about +ĠSt d +j ournal +()) {čĊ +," \ +( Arrays +ĠD ense +ase ña +än ner +/ stat +user Data +Ġg erman +Ġt z +worth y +Format Exception +ph erd +Ġsm iles +ĠWh enever +( adapter +.bad logic +Ġbrief ing +.Grid Column +- char +dim ension +ĠC opper +Ġnin th +Ġ' {{ +Ġr av +_T able +Ġderiv atives +ĠR aise +ĠF ut +arm or +-p adding +Ġre min +ĉ style +ĠMembers hip +Ġspread s +Ġgall eries +ĠClar ke +Ġcon ception +min ute +Ġab usive +_ad j +Ġterr ific +Ġover t +our cing +Ġentr ada +level s +Ġcrit ique +Ġrespect s +ĠM MA +i ene +Ġenc aps +ĠRay mond +Div ider +iv able +b az +Ġ@ _;Ċ +ĠCl aire +Ġur ging +CE E +Ġtransform er +disc ord +ĠJ ourney +t os +Ġcompet itions +ĠO BJ +ĠB is +Ġrelax ation +id y +_IN STANCE +ĠP ref +d ados +ici encies +ĠMedia Query +ĠC ube +ĠStr ange +g pu +(d ays +_Init Struct +Ġfinger print +em at +ĠGe cko +Ġr ails +ĠL um +str action +ig ung +(m ovie +_d ictionary +_int errupt +ĠQ C +ik ed +append Child +rec ipient +r é +V e +Ġtow el +.last IndexOf +Ġplace bo +ĠW ie +.es p +( Debug +oper ative +Ġdece ased +& id +ĉm utex +el ic +Ġb apt +ĉ čĊčĊ +Ġfar ther +H alf +.dis able +.menu Strip +le ccion +Ġresult Code +Ġc ans +-e lection +f emale +_F IX +aus ible +ĠP OWER +Ġrecon struction +Ġsc ans +.Xtra Bars +âĢĺ s +Rem oved +Ġparagraph s +_m argin +Ġl ymph +Ġb os +ling ton +ĠBapt ist +Ġadvertis ements +ĠMan age +/ yyyy +IO US +ENC ES +ĠF iction +ĉm enu +ĠFile OutputStream +ov an +ĠF eng +Ġsk ipping +get Class +ann i +Ġreb ounds +Ġpublic ity +Ġing res +use ment +Ġthought ful +.Ch art +Ġhat te +pass port +Ġhook ed +ĠL ens +Ġflag ship +Ġst ip +ĠG EN +Ġcl ues +ip v +ĠR ise +ĠG ew +tab lename +Ġfore most +_ validate +_an alysis +oll a +Ġqual ifications +Ġdistrib utions +ĠFl ower +Ġt ense +Ġthank ful +Ġcl utch +Ġun ified +ro ads +Ġsit i +Ġst all +_P RIORITY +c stdlib +_USER NAME +.by tes +? page +ermal ink +ĠVe get +/v nd +- author +.N ONE +ĠCon current +ĠC ry +Ġstart ers +ĠInter action +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠ +ĠLE VEL +E ll +Ġcom boBox +ĠTh eresa +te k +_H andle +Ġab y +.g dx +, end +(L ocal +O l +kn ife +ar ial +ĠH off +Ġprostituer ade +Do ctor +Inst ances +.Set Value +ĉf rom +Ġlux urious +Ind ent +Alloc ator +_D RAW +(", ", +ĠFr ances +Ġgroup Box +(s chema +Print f +OR IES +- gradient +Ġre put +ar in +_D ONE +in cre +ig nty +Ġex ert +Ġ- . +/ App +-th rough +Ġdecl ining +Ġdess ert +Ġinc umb +Ġdesign ation +.P ORT +, strong +Ġsand box +Ġw ines +ĠP av +$ str +ask ell +Ġh ö +ĠP Y +Get Instance +Text Input +game Object +/ events +created At +Ġlocal Var +ĠWH ITE +per ed +ile ge +eff icient +, color +c ate +ĠC afe +Ġsimilar ities +Ġp umps +ĠHung ary +.User name +Ġsk ate +Ġtouchdown s +Ġacceler ate +ĠH elen +OM EM +ĠK un +_v ol +Ġfind All +ĠMens chen +a head +); " +kom men +Ġpossess ed +.arg max +.trans ition +AR P +OLUM E +(s cript +ĠÐ ĺ +ĠF inding +on ces +I o +B old +Ġrenew al +_D IALOG +Ġdis reg +INT ERN +Ġt oute +Ġelect r +ĠG ross +ĉ true +.F ields +ĠW IDTH +ĠD ent +Ġà ģ +NS Notification +Ġa os +Ġme lee +. Validation +ĠDE C +-depend ent +Ġsu ic +T raits +$ message +ĠD ear +ĉ FILE +l anguages +.P rot +.add r +-g eneration +IC ON +Ġtrans plant +-d escription +Ġch asing +Ġche es +Ġ} */Ċ +Tr ad +qu eries +/widget s +sub package +Ġes pec +Ġcr acked +Ġcompet itor +P urchase +- team +olec ular +or Thunk +& P +Ġrel ent +/ #{ +Ġproduct Id +Ġè ¾ +ĠL av +ĠAl ter +.M ode +AD IO +gr p +æ ·»åĬł +Qu it +Ġdepth s +-c ategory +ĠD ATABASE +S PELL +ĠFal con +ĠQString List +Ġ'' . +ĠIn stitution +d amage +az or +bel ongsTo +ver ages +ĠN ONE +ipp ets +, \Ċ +Ġfoot print +_ archive +n ak +.get Field +ĠRef lection +Ġ' ] +ĠH BO +_dis count +Ġin cest +ĠD odge +ĠW ade +.N O +" encoding +ĠBlock chain +Ġlaws uits +ĠM aint +ch ten +Ġét ait +Ġktó re +_ ctl +(t imer +B attle +iz o +ay ed +I OR +ĠGlas gow +Ġsyn th +_log s +.p ose +_Adjust orThunk +(( & +Ġuns ure +yst ate +íķĺ ëĬĶ +O ULD +. ng +Ġdefault dict +work space +Ġselect ive +Picker Controller +YNAM IC +.method s +Ġpath ways +ĠF ew +K G +CRY PT +follow ing +ĠD LC +ĠS ara +Ġpres et +estruct or +ĠK urt +Ġair plane +Ġo mp +ĠParent s +ĠMart inez +.com plete +Ġbroad ly +Ġsc are +ĠM é +Ġelim ination +Ġpou red +/ sw +Ġcom un +Ġm asc +ĠOrgan ic +ĠString Utils +il ateral +Ġreluct ant +- age +Ġn z +." \ +Ġpast or +ale z +Ġe fect +pro v +/ init +Ġp enn +und s +Ġs size +ĠPro j +bas ename +Ġsh ells +ĠNe ck +ĠEn forcement +vid ed +st own +S phere +$ r +uss en +af il +ĠTele gram +Ġanaly tical +нÑĭ е +us ually +x n +Ġhistor ian +ĠGreg ory +ol ph +ĠUn a +Ġcon tributes +% - +anti ago +ÑĢ ÐµÐ´ +.reg ion +Ġab rupt +ĠUnsupported OperationException +ĠT ASK +_f inish +Ġnot orious +ĠV s +ĠM Q +Ġsun set +Ġun acceptable +ar cer +Ġill umin +ĠOr b +Ġb h +E ste +_dis patch +Ġr ipped +Ġtou jours +ĠPar cel +_ ll +.user Name +.class es +S OURCE +( Number +ел Ñı +Ġhead phones +(s ide +const itution +ann ah +čĊ ĠĠĠĠĠĠĠĠčĊ +Ġcl iff +- ref +Ġmo strar +ĠPow ell ++ y +ĠB G +_f ragment +.P ort +Ġreal izing +param ref +Ġh ometown +@ Table ++" --}}Ċ +F rench +Entity Manager +ĠPl ain +//////////////////////////////////////////////////////////////// //// + ³ +( RE +c apt +Ġorgan isms +Ġj ets +ol ocation +ĠApp RoutingModule +Ġgl orious +æľ į +Ġdisc arded +ĉĉĉĉ ĠĠĠĠĠ +ĠArn old +l ug +Ġpar l +Ġhorm ones +Ġm ah +ĠSon ic +Ġorgan izers +_PL ATFORM +.in v +Ġch ord +vent ional +ĉ of +Ep isode +. Enum +unk t +ĠD h +ĠJ ared +ĠN ak +Ġint ends +End ian +Ġa ustralia +_c v +(res olve +Ġclin ics +lik ed +ASH INGTON +in ha +' * +ĠN P +_b eh +Ġh f +Ġw ür +c ategoria +$ form +Ġsub way +Ġis Active +pop ular +C our +Ġco oldown +Ġa insi +ĠGL uint +ere al +Ġarray Of +Ġh atch +======== == +ress es +_P P +. ^ +_dec ay +ĠB less +met rics +ĠCOPY ING +ĠDump ster +ĠJos é +ĠDesign s +< +Ġ" }Ċ +time zone +Ġe er +max cdn +ĠE SC +ig aret +_conn ected +_re verse +Ġquestion able +ĠUS C +Ġtut ti +Ġdrop out +ĠActiv ities +ĠW inds +')) );Ċ +Ġcon gest +ÄŁ ı +Ġprolong ed +è¿ Ļ +ĠCross AxisAlignment +LE EP +ĠVAL ID +ĠG az +Ġdepend ence +ĠP rix +.Compiler Services +j ump +Ġstr at +c irc +ĠC USTOM +x aa +Ġb mp +Ġb ureau +Ġw aren +N X +( Window +ĠChrist ie +_F E +Ġt n +ĠOm ega +communic ations +Home Page +com pletion +Ġsupply ing +YP ES +á vel +åĪ ¶ +(c lick +\ Contracts +/ questions +Ġe z +AM S +.m esh +Ġ' \Ċ +Rob ot +Json Object +ĠD F +ĠProcess or +_sh ould +.prot obuf +- users +Ġemb ry +F ONT +Ġstart ups +ĠData Source +) # +uro s +_C olor +Ġstand alone +} [ +j d +Ġforg ive +Ġng x +ĠGener ally +Ġconfig urable +/ order +Ġv as +') ";Ċ +ĠR R +ĠT roy +Ġcomprom ised +ĠSw an +int endent +Cent ral +_ keeper +Ġar quivo +ĠRead Only +_cur ve +k v +ent in +è ± +ĠE y +.im read +ĠP am +if fe +at ivity +xb c +Ġgr im +-f illed +names e +'] : +Ġa ur +ĠGib son +.Mouse Event +Ġl ado +avad oc +Ġfam il +ĠM oder +f ps +ãĢĢ ãĢĢ +- example +ĠAl zheimer +ĠU tf +_arg uments +Con clusion +text Content +rem aining +Ġinterrupt s +ĠBack up +ĠM ong +Ġrecept ors +h istor +.cor outines +Ġsh outed +Al arm +Ġcomb ust +Ġg rote +ult ural +( ids +---------------------------------------------------------------- ---------------- +ipl inary +O pts +ĠY ale +local Storage +Ġequ ival +ĠF leet +\ b +* pi +ĠQ Label +æ ¡ +Ġv x +ĠA CL +Ġsu cesso +Ġper c +ĠNot re +Ġan arch +R ing +sp b +Ġstr pos +st ores +ĠMap le +(Main Activity +(" ")) +Ġview Holder +Qu ad +Ġig ual +ors che +.m argin +Ġind ie +Ġfr anc +ĠForm Builder +ĠPart icip +.fl ash +Ġstorm s +U lt +Ġf en +[ new +E ver +=" Ċ +Ġlocal ized +_f ollow +Ġn ave +Ġdomin ance +(t ile +J ournal +ĠV C +Ġpenet ration +ï¼ ķ +Ġcomp artment +Ġb ids +Form atted +****** /ĊĊ +(c ity +âĢĶ it +[ C +Ġuse Callback +a ub +) ?. +ĠV AR +ĠSe bastian +ĠM oss +Ġabund ant +G reg +ÑĤ а +_c i +Ġbib li +CR M +ĠAt tempt +ism e +d ash +ãĢ İ +_m u +.Formatting Enabled +Ind eed +-d irect +Ġsuck ing +Ġp ne +ocab ulary +ĠPack ers +.N avigation +Ġp ied +cri bing +ĠSt uart +.To Double +ĠSecond ary +S aving +ĠD ut +ĠM add +M agic +, H +.document Element +ĠB ST +Ġdiff ers +Ġmore over +_ nd +SE ARCH +п ÑĢав +æ ´ +to Match +Ġdecre asing +-m ember +amp us +( boost +D aily +Data GridView +ĠHttp Context +Ġh ipp +_work ers +-l anguage +é ĵ +Ġconsist ed +ath ing +ĠMer cury +$ content +Ġpract iced +ĠMod ules +_D AY +Ġweakness es +ĠL odge +Ġn ar +ĠM ate +Ġj p +ĠHttp Headers +Ġsm o +ĠT OKEN +] )( +Ġaqu i +sw agen +Ġs rv +ĉ ans +A round +ĠMan uel +Ġfiction al +ĠIM G +Ġ. ' +ĠB erry +Ġwall paper +sex ual +ier o +Ġ çļĦ +ìĨ Į +Backing Field +ĠAd rian +BASE PATH +Ġrepe ats +Ġbl ues +Ġunp redict +_c oll +st acle +ĠT umblr +ĠEl f +Ġass urance +Ġc ensus +ĠIM PORT +END ER +an os +Ġ= ( +ĠEll is +" ĊĊĊĊ +.w in +ĠA bove +al on +_t ick +Ġrepresent ations +Ġæ ķ +w id +ĠAr ms +List a +_f ailure +_c m +.Flat Appearance +Ġthr one +P atch +ĠV oy +eng l +Ġnegot iating +> ` +Ġshoot s +ĠF PS +.Y ear +ĠK iss +enc ión +reet ing +From File +Ġresign ation +Ø · +Ġtw ins +ưỠ£ +Ġge bru +.get Content +.T ree +ĠEmploy ees +ĠF IFA +Ġcert ainty +(C l +Ġtot als +edit able +à¥ Ģ +.Report ing +M as +qu iet +.r ules +ĠV O +con exion +, K +Ġalloc ator +ĠPow der +\ Repository +Be at +_t ipo +Ġ[' ', +_IN TR +Ġ<< < +< hr +") == +ugg age +ĠC raw +Ġé galement +Ġg inger +Ġprim era +Ġprod uto +lt k +.User Name +Ġstr error +m ith +_n b +Ġdis comfort +']; ?> ");čĊ +drop IfExists +ĠB eg +_H AL +Ġcross AxisAlignment +ĠE vidence +Ġpec uliar +Ġinstit ute +ve is +Ġf ft +à ģ +Ġzo ekt +an aly +ĠHom eland +Ġpen etr +udden ly +ĉ element +ĠB ren +ĠTr udeau +ĠCub an +j am +us lim +_e v +Ġst ems +} % +Ŀ å§ĭ +Ġbrand ing +Ġcorrespond ence +.j query +¢ åįķ +ĠRead s +(Http StatusCode +ass in +(s lot +ĠGrad uate +/// < +Ġinform ations +EN ABLE +Ġp uis +Ġfind er +ĠBr is +Ġnett steder +_m id +Ġo gs +ĠSter ling +Ġar rog +str ftime +| ĊĊ +Ġvo x +ĠReg ardless +Ġes o +ĠCom fort +.Boolean Field +Ġu h +AC Y +Ġsque ez +ĠV ic +cont ro +. lo +Ġ ire +ĠCom edy +ë ¶ +Ġorigin ated +Ġsh ipment +| max +_g uid +lev ation +на Ñı +( undefined +ĠD DR +Ġshoot ings +ĠLat ino +END OR +Ġaver aging +Ġgre eted +Ġthe aters +о е +Ġd B +Ġg st +Ġdef inite +. Storage +.h er +Ġa fore +ĠRe ality +ĠGod s +vers ed +Ġhands ome +Ġex cluding +( ad +Qu otes +ĠS cheme +? q +ĠT amil +T icks +Ġp est +' n +Ġporn ography +_mod al +Ġ ---------- +Ġdis posable +F REE +Ġsh ark +C HE +Ġdep icted +Ġdemonstr ations +ĠK illed +ĠR ULE +Ġobs essed +Ġsimpl ified +Post al +Ġconcept ual +Ġp st +L as +_PRO JECT +ucceed ed +ol u +ÄŁ i +Ġpersonal ities +Ġres hape +Ġenc losed +ĉp tr +Ġtutor ials +Ġexpl oded +_DIRECT ORY +åĨħ 容 +Ġcan on +Ġrecogn ise +P AD +ĠAppro x +ĠRest ore +ĠImport ant +Ġheav ier +.Se quential +Ear th +ĠMil k +.set Request +.t em +Ġre construct +Ġskept ical +_Pr ivate +BU F +qu a +: a +Ġse k +Ġd well +oss a +Ġreward ed +и й +(top ic +_part ition +Ġ__ ________________ +Key words +ĠFr anco +L ite +Ġn aken +Ġз а +O BJECT +Ġcraft s +ĠSw ap +.X na +.Con nect +Ġbalcon y +(re al +ĠBarn es +b ir +ĠTw enty +ay an +at ars +ĠProp el +ĠIh nen +Up grade +Ġcur b +- second +Ġn eph +.p res +ìŀ ħ +.se q +Ġp added +" ? +j l +ãĥ ¬ +') a +Co ordinates +Ġen acted +ENT S +Ġl ac +.f inal +ĠPhp Storm +c alled +Ġin quiries +.m iddleware +ĠD owntown +/ ';Ċ +Ġkil omet +ac cel +Ġqu ien +w string +set Data +Ġman era +Ġmod ular +rim p +Ġtar iffs +âĢĻ il +_TH ROW +/c olor +ĠHT MLElement +Ġcar ro +Ġpr ere +Ġplot ting +ĠPos itive +ĠMach ines +OT ES +á» Ľ +ple asant +Ġal te +Ġa inda +th ese +Ġc ors +ip ay +ĠAdvis ory +ĠRub io +j q +Ġl imestone +Ġdet ached +设 ç½® +ten ant +ĠDep th +al ore +ĠÑģÑĤÑĢ Ð¾Ðº +ĠF ORE +ĠL ay +p resentation +) ');Ċ +.sub plots +Ï ĥ +N OW +G ar +hand les +ab ra +put ies +ĠElect rical +M iddle +rop ic +ĠJ D +ĠD yn +ĠB ristol +ĠMc Carthy +Ġstri ker +Ġenumer able +ĠEv an +.default s +qu ences +) || +ĉt oken +â Ĺı +-d ropdown +ST ORE +ĠGraph ic +( pp +Ex pl +Ġup wards +ĠD istributed +ĠW EB +J er +is NaN +çĶŁ æĪIJ +> R +üss en +ef s +Ġun cover +Ġl ud +.cal culate +Ġint ptr +Ġmidfield er +. Headers +Ġm f +ere f +.M etro +ĠSpe aking +: b +Ġcryptoc urrencies +Ġdem ons +ĉ EXPECT +Ġw icked +y outube +: Int +ĠHind i +ĠC AT +ĠØ ¹ +r ar +om ore +/ per +/lic ense +Ġre im +Ġawait ing +Ġle thal +ĠE F +round ed +ĠPl atinum +ĠвÑģ е +.co ords +.De vice +/ item +ĠW enn +compile Components +ĠK inder +.remove Item +Ġand a +bn b +Ġpr a +( transaction +Ġembarrass ing +ĉ BOOL +.content View +Ġevent data +at ore +Ġprovided In +ir ma +Ġz ona +_H W +æ Ļ +Ġst ove +Ġcounter part +_Pro duct +_MAN AGER +Ġinfr ing +ĠE RA +_p arty +Ñ ij +Ġin ici +_ Request +Ġmir acle +Ġcancel Button +S py +at ó +Ġpol ish +ĠNic ole +.display Name +\Request s +Ġuse History +Router Module +Ġst ared +ID ER +Ñĥнк ÑĨи +Ġnot a +$ arr +pec ified +Ġto pp +_DR IVER +/ ng +å ł +_t m +% timeout +< s +Ġ( *) +ĠHttp Request +_TR ACK +(n ote +ĠExp lore +_s erv +Ġç » +B inder ++ ", +. att +ĠEth i +Ġc ódigo +=' \ +.l ines +( Of +å° Ĩ +miss ible +Ġv é +Ġac oustic +Ġcraft ing +n it +.b a +ĠLuc y +Ġi Pod +Ġpup ils +-m ax +_w r +(c p +ĠRE PORT +Ġd ns +ĠRe ferences +Ġundert aken +Ġkø benhavn +Ġch ai +ĠC roat +_ Log +rown ed +_m ed +ĉ date +# __ +Ġcost umes +ĠRe quires +aff le +ç Ĭ¶æĢģ +-S emit +ela ide +еÑĤ од +Ġp estic +Ġd ra +DOC UMENT +Ġ... čĊ +}` }Ċ +ĠA uction +ĠD ock +xxxx xxxx +(get String +ħ į +Ġborder Width +ĠMach inery +Ġpredict able +.S H +Ġam plitude +.for Root +IN avigation +Table Model +at trib +Ġmaneu ver +Ġexc av +B ERS +Ġd apat +Ġinstall ations +.A sync +Ġr ays += âĢĿ +; ččĊ +.c rypto +_db g +ĠEnum erable +Of Size +_epoch s +m w +M ENU +out line +ĠP apers +============ Ċ +Ġuniform s +ĠG ig +- package +ĠJen kins +ĠHome Page +.is Selected +Ġmechan ic +M K +ĠS ounds +//---------------------------------------------------------------------------- -Ċ +Ġresearch ing +Ġinf os +ograph ics +ers et +([' / +ĠTim ber +. agent +.to JSON +_command s +par ing +_ad just +.n ome +(g lm +Status Bar +file path +? âĢĻ +Ġdetect ive +Ġunser er +ĠTib et +EN DED +(se ed +Ġsne ak +Ġam or +=" // +ĠPan thers +all ax +ĠL IVE +ĉD WORD +]= - +Ġtorn ado +/ min +Ġlung s +-c urrent +ĠBook ing +åĪĹ è¡¨ +Ġenjoy ment +ठ° +J A +typ ed +.B tn +f at +ug al +ĠSh ares +Ġdis gr +ĠB AR +ĠFO X +Op code +ĠS z +key down +iction aries +Ġdetail ing +} ))Ċ +Ġp ok +Ġdemonstr ating +Ġnot ation +l ayers +@ if +ĠN PR +.strict Equal +ĠRec ipes +.T ensor +Ġliqu or +Ġdeb ts +.ends With +W heel +.P os +CS V +$ arity +Ġun stable +( loss +ENS OR +Ġele ven +ĠL opez +ĠHop kins +con om +ĠS eth +Ġpo ems +Qu ant +Ġg sl +Ġsy rup +Ġs ibling +Ġc ass +-v ous +ö t +_P ATTERN +_SE CTION +est imated +up grade +.m ongodb +ĠBo at +_C TX +Ġfetch ing +ust in +pi el +M arg +Ref lection +Ġd uct +ĠMunicip al +Ġb x +.Get Current +ml ink +ĠAccount ing +ĠGene va +_P os +Ġpass er +Ġhear ings +com pan +Ġfrag ile +Initial izer +walk er +.M aterial +ĠHun ting +trys ide +Ġk at +Ġcl erk +á Ł +do ing +ĉg roup +Ġsan ction +.l b +ĠL azy +ĠCon straint +P agination +Ġpou vez +ĠInd icates +M ER +Ġcour s +Ġyear ly +Ġgros se +abb rev +ĠD ON +Ġproceed ed +ent lich +Ġproperty Name +ĠTe aching +st adt +Ġc utoff +orn ers +Ġa frica +Ġrend ers +ĠYan kees +ĠTool bar +sp aces +.fill Style +Ġseg undo +_str len +.F irebase +å¤ Ħ +Ġmention ing +\ ( +ĠVal ve +Set ter +Ġsp ans +ĠAl cohol +ĠLet ters +\x e +ĠT K +_B LE +.get Result +< Player +ĠP att +Ġeas ing +Ġtur key +ĠF en +') " +Ġconf ined +Ġin clus +Sup erview +(with Identifier +enc ial +Ġstuff ed +Th eta +Ġeconom ists +} ));ĊĊ +co okies +ĠRo ose +ĠChe ese +Ġfich ier +Ġen forced +AB B +no ÅĽci +_AL LOW +Ġrecru ited +Ġexpend iture +-n ight +Ġassert NotNull +_ex ecute +ĠØ ¯ +IN DEX +_F MT +Ġresc ued +ĠMonth ly +ĠCons ervation +ĠG eb +Ob ama +Ep och +ic ies +ĠOr t +Ġso it +( icon +F riends +m ol +Ġground ed +ĠC ause +ad ena +WE EN +ĠL un +IT IVE +. loop +_un til +Ġcor r +.ed ges +Ġhyp oth +ched uling +trans lator +ĠÐ ľ +R om +ãĢij ĊĊ +ĠX amarin +Ġviol ating +. anchor +--- ĊĊ +Ġtr ader +AD VERTISEMENT +Ġuns ere +ĠD AO +Ġbl ond +ĠP AT +.g lob +Ġè¾ ĵ +Ġsplit ting +Ġun subscribe +Ġatmos pheric +ĠTr im +Ġcit ation +Ġin ference +ĠF t +ĠDar win +find One +ĠG el +( Convert +Ġaccess or +; text +(s orted +Ġjud ged +); \ +: p +Ġme ine +ĠS lim +.Command s +Ġper ceive +coh olic +< Data +.entry Set +Ġassert False +ĠPat rol +ense m +ÅĤ Äħ +¨ ¡ +W IDTH +ĠRes cue +ĠU IF +_THRESH OLD +ĠMich el +ATER IAL +opens ource +ĠD iana +Ġinv ites +_B ODY +Ġreserv oir +Ġro i +c ust +(t c +ï¼ģ ");Ċ +Ġfest ivals +Ġperform ers +Ġclim bed +Ġj ungle +String Length +Ġunlaw ful +ier re +vertis ement +Ġst akes +Ġh ats +Mod ify +ĠLET TER +.H ide +Ġstat utory +_ white +ĠPer l +uten berg +em ple +.W orld +Ġoverlook ed +Ġcon cludes +/* ================================================================ +-w ise +ĉ stream +pop ulation +Ġevent o +Ġillustr ations +ft s +Ġaut of +ĠPro cedure +Ġdes erved +-t imes +Ġg ol +N SError +cre st +ĠPak istani +any ch +get Current +Ġl ar +nt l +ĠRe becca +Ġm ateria +Ġfind By +/ ad +Callback s +ĠAl s +ĠKat ie +ĠObservable Collection +ĠDocument ation +Typ ed +ĠCulture Info +ĠTim othy +Ġlater al +" type +Ġun authorized +Ġteach ings +Ġdebug ger +[ value +Ġal ors +Ġu z +Ġsc atter +Ġdown ward +Ġmig li +status Code +Ġ( )) +ĠM W +Ġм ож +RO SS +.b uf +Ġfair y +ĠInf rastructure +=> " +t lement +$ (" +From String +ĠB ild +Ġconvent ions +_n ative +ĠIns pector +ĠP ist +ub ar +Ġreg s +ĠP ilot +Th us +>' + +Ġc ela +.new s +( Product +L iving +R ussia +Ġfac et +et ical +Ġ[' $ +/ [ +ĠD ire +Ġg ases +ĠIN FORMATION +ĠE at +ĠFor ums +ĠChar acters +_m et +Ġìĭ ľ +Ġk ings +ach ie +ĠL ambda +Ġtim ers +ĠLight ing +ĠCase y +add ir +and ex +. answer +ĠH ip +ĠPr incip +Start Date +Ġ ãĢĮ +t res +Ġ& # +.Max Value +ĠPro blems +Ġlat ex +Of Class +ĠLyn n +// ' +Ġvoy age +Ġshut tle +ĠRoll er +ĠRuntime Error +uy a +D ic +ĉb uilder +Ġbul lying +Ġsimple st +.c alled +ĠL R +Ġmor ality +Ġst urdy +tr acking +.sw agger +_B IND +IT OR +-url encoded +ĠÑ ħ +ĠTr inity +Ġtr aps +Ġ| - +Ġset Text +Ġbarg ain +Ġbr akes +.get Code +Ġmigr ate +Ġrib bon +) return +Ġcharg er +ac om +ADI US +ĠAmb assador +-a fter +Ġann i +ĉs pin +Con cept +ĠHend erson +ĠH OST +.r ank +ĠNor theast +Ġber lin +Ġrequ is +.f eed +Ġsource Mapping +ĠRen contre +. ajax +nest js +Ġtre k +ĠN acional +Ġ& [ +Ġpay able +ort ex +Ġde pt +field Name +Ġcomple tes +ĠR VA +Ġon ions +al ignment +Form ats +Ġ' {$ +Hash Set +ĠB od +.Invariant Culture +Ġsettlement s +Ġhy dr +. updated +vent h +( seconds +="/ " +Ġweb page +( ĊĊ +Ġt ir +Ġto es +ĠBr ick +Ġamb ition +P ot += max +ET IME +Ġdep ot +c alls +ĠNor wegian +` : +Ġbur ger +Ġprofess ors +ĠAl locate +-third s +-ch art +Ġfor d +* N +.k otlin +Ġpaper work +ĠDE VICE +% @", +res pect +(m p +é «ĺ +- if +Ġcush ion +ob ot +Ġpar c +SP ACE +ĠNet anyahu +Ġself ish +fe at +Ġclient es +-to ols +Ġpor ch +Ġj q +. verbose +Ġlib erals +] )ĊĊĊ +p ies +Not Blank +( term +ÈĽ i +_Param s +.normal ize +B ullet +AS IC +(h ex +_client e ++ , +_D I +Ġforth coming +} ")]Ċ +se o +U m +> Name +Ġcomfort ably +irection al +W ITH +/ pr +ĠP oor +ĠVit amin +v ic +G H +Ġprior it +ĠN N +ĠC losed +¤ í +Ġis Open +\ Console +And Feel +.S UCCESS +_OPER ATION +pol ation +ĠT as +ps z +> '. +C URRENT +V endor +host s +ĠE rd +>tag ger +ĠsourceMapping URL +Ġmar athon +_c losed +Ġexem ption +Ġrecogn izes +ides how +' $ +('/ ');Ċ +m its +war z +ĠCh erry +µ ¬ +n or +port e +Ġw l +_back up +.get Boolean +.get Resource +Ġdefinit ive +. EditText +Ġs ÃŃ +.C ONT +ĠPL AYER +.c ards +ĠSh ore +('/ ')Ċ +cl uir +Web Driver +(m onth +-re lease +Ġins pector +å £ +ĠN F +_cl ip +åŃ IJ +Ġinteract ing +.t mp +Ġ'' 'ĊĊ +Ġde e +Ġfro st +"] ))Ċ +ĠPl aces +Th rows +f ork +/ day +i Phone +ĠM IC +Ġfold ing +Ġcro re +ĠCh iefs +pher ical +( price +.Write String +Ġexit ing +] ',Ċ +ight ing +Ing redient +( vertex +Ġscroll View +h f +: new +SE N +se ctor +Ġsp ins +ĠS cheduler +ote chn +sem icolon +Font OfSize +ĠSpecific ally +fl amm +.Object Id +Ġcont a +_per missions +ĉF ROM +IC ODE +/ kg +ĠHot els +-m ed +ĠD in +Ġn avy +get Param +Ġm end +Ġportray ed +ĠMet ropolitan +Paint er +Ġref erral +_g ood +Ġmar vel +osa ic +> (& +. ur +Ġest os +Will iam +Ġtim ber +Ġquel ques +ĠDoc uments +.X aml +Ġbatch es +éģ ĵ +ĠRe leased +T ail +CO OKIE +he id +_st ation +ĠV ia +S ale +ĠRe peat +Ġprom in +ĠZ o +- forward +ĠI on +it ary +Ġj us +- request +Ġproud ly +ĠStream ing +(Mouse Event +ĠS print +_ rotation +Re positories +Ġt art +ĠÑģ в +Ġm appings +è ª +C u +C ycle +Ġb un +ĉl ua +ãĥ ī +Ġ(( ! +Ġcollect ively +ĠCon d +Ġwsz yst +(l ib +openh agen +_s kip +.Column Header +é Ĥ +peri enced +ı è¿° +_p rops +Ġcontr ace +Ġmatch up +ab etic +.m embers +RE CT +(d at +Ġs og +ren om +_M ethod +Custom ers +full name +Z N +re try +Ġk ap +ĠNe u +è Ĭ +add Child +will Return +_p ermalink +Ġener getic +ĠW et +ĠMor r +Ġg cd +count s +, type +d ig +( Login +Ġcr acks +Ġbacter ial +ĠMe at +ĠArm strong +ĠBron ze +Ġapprox imate +_dir s +lig a +ÅĤ ad +Ġkind ness +Ġcont re +ĠE VERY +M ET +Ġannounc ements +g pio +ĠWaitFor Seconds +ĠPhotos hop +Ġdis contin +/ dd +Ġtop ology +an ical +. interface +auc oup +.Hash Set +ARI ANT +(r outes +ĠT eh +Ġh ype +] "). +Ġsl am +Ġbro th +- inter +ĠR id +-m anager +Cancel ar +ĠP agination +Ġsound track +Ġpost erior +Ġscr ub +cre ating +- * +ir teen +.d y +.s ymmetric +Ġ"" . +============ === +Ġch assis +ĠnumberOf Rows +Develop er +_b ins +ĠO UR +ri eb +Pro s +Ġwi ÄĻ +" d +Ġasync io +ze igen +_s pi +.A LL +Ġscre ws +Ch inese +Ġapi Key +Ġun successful +ĠSeah awks +OR G +ç« ł +Ġprofession ally +ĠCou pon +åŃĹ æ®µ +Con vention +Ġpol ym +æī ĭ +Ġsalv ation +Ġengine ered +ĠW rest +ĠG CC +Ġwar mer +Layout Constraint +Ġag grav +Script s +vent ure +Ġrefriger ator +Ġinnov ations +ĠRun ner +N IC +ĠRoll ing +Control Events +Ġlo os +p ac +ĉ panel +ef e +ĠBudd ha +------------ --Ċ +åº ĵ +(for Key +Ġl umin +Ġ( ? +ĠA IDS +, user +im ientos +content Type +ant lr +é ¦ +ĠW elt +Produ ction +m ight +ĠV II +", ( +Ġobserv ing +Ġdeliber ate +( control +Ġwith d +Ġsem ana +ST ACK +uch en +N ice +ĠDeutsch land +ĠSpec ifies +d ma +iz io +ĠF acts +_pop up +ĠDirect ors +{ : +[ R +ĠÑį леменÑĤ +Ġpl at +Ġdirect ing +ä¸ ī +ĠGil bert +â̦ .ĊĊ +.q ml +Ġthere after +Ġdis position +d raft +Ġsurge on +ĠIns ider +Bl end +ĠT rev +tr insic +Top ics +rie ve +_FILE NAME +Ġaut res +J ose +Produ cer +er us +Ġpet it +ĠN EXT +ĠF ilters +Ġreplic ate +"] ). +Ġl enders +] ",Ċ +; charset +Cpp Object +Ġfl oral +ĠT ipo +Ġcirc uits +e asy +(& $ +itt a +ery l +_COMM ON +'}} >Ċ +-back ed +(var iable +( Index +Ġvo ir +_loc ations +++) { +ĠLouis ville +Ġgrat itude +.Mock ito +ĠP owers +ie urs +Ġge ographic +ra le +Ġc ra +ĠSp urs +iph ertext +AC ION +- common +Ġvict ories +ĠFinal s +.sh uffle +-m illion +_PRO C +ass ume +Ġil s +DB C +Boot Test +Ġl avor +.test ing +. ast +"] / +m oid +Ġqual ification +ges ch +ĉ put +Ġair ports +J I +Te acher +_un iform +Ġn ama +ĠB ast +ert ype +c apture +get All +ĠReyn olds +oo led +.com ments +Ġch in +). * +Ġи ли +t gl +ud os +Ġd ÃŃas +ch ai +.pro gram +Ġps z +ĉ icon +ph il +ent ral +_WR AP +ov i +Ġnost alg +In finity +ĉy ield +Ġvit amins +Qu aternion +S ink +_g oods +Ġ ........ +ĠW ings +ur idad +-st ory +"] )ĊĊ +idel ity +Type Def +G tk +Ġí Į +_M ain +Ġche z +ĠR aven +Ġpay roll +Ġfreel ance +LL U +ĠM end +ed ay +Api ModelProperty +.Form BorderStyle +Ġeconom ist +stan bul +Ġfre ight +-A gent +(m eta +Ġsym metry +Ġ' .. +.C alendar +- aut +g f +p ent +yc lopedia +Ġwish ing +ĊĊĊĊĊĊĊĊ ĊĊĊĊ +Ġgentle man +Ġê ³ += # +Ġlect ures +âĢľ In +Ġ! _ +Ġh b +ĠV endor +Recent ly +_n otes +æıIJ 示 +" My +Headers Height +_S O +Ġunw illing +Ġsuper hero +g io +ps y +ĠPe er +j avax +& apos +ĠCr isis +ord inal +Mem cpy +++++++++ ++++++++ +- val +Ġwork book +- ap += k +Ġmetal lic +_ peer +By PrimaryKey +_S D +u ator +_SH ADER +) Math +.Trans form +Ġc ows +Ph i +ĠC lem +(_ (" +ĠL ud +-d elay +ĠSec urities +ĠOrth odox +Sym fony +(re port +Ġent ertain +E PS +iz oph +ex ual +IR D +ä» İ +Ġl ith +Ġsanit ize +Ġfemin ine +IS BN +.auth entication +_p ipeline +/ constants +ĠCON F +Ġluc r +ric ia +.t tf +.set Content +Ġst an +ore an +ĠL loyd +.raw Value +Ġg or +ĠBrow ns +Re gression +Ġlower ing +na issance +Ġbl ows +Ġam azed +Ġun related +Re views +Ġrub y +ĠMod ifier +Ġgi ants +. thread +Ġcontain ment +ĠStart Coroutine +um at +ore lease +ĠR andy +@ endif +D igest +Ġsubur ban +=" );Ċ +Ġann once +. variable +\F oundation +Ġa cre +V an +Ġt uples +d ns +ĠStand ing +_l arge +Ġbox ing +Support ActionBar +ĠFort une +ĠR um +_m ultiple +arch ical +Ġf write +_ quote +Ġfool ish +Ġcompr ising +Ġо п +- selected +v f +ma id +N ama +(d atetime +Ġindirect ly +g art +fix tures +ch os +ĠH alo +Ġrec urring +- news +v il +ĠNurs ing +- produ +ĠH Q +\Http Foundation +enc i +au en +Ġv y +ocr acy +Ġdeleg ation +Ġas phalt +Ġset Selected +k ok +/ rest +met ics +ĠNS Date +Ġtravel led +Ġrec ib +Ġm ime +CL IENT +ĠG U +ĠH ANDLE +/ Q +[ z +Ġbother ed +ĠBB Q +ç as +_ex amples +_F IN +Ġwhite Color +Ġastr onom +-d ir +Ġsovere ign +Ġb reeze +Ġin ning +ĠEd monton +g li +.blog spot +js x +Ġvers a +ĠMoh ammed +.J ob +-t oggler +Ġп олÑĮзоваÑĤ +ard on +Ġnew born +Ġnav al +note q +Ġtum blr +Ġh entai +ĠTyp ically +Ġlo ot +.S prite +Fl ight +Ġw avelength +-s k +ĠEl le +_ exports +Ġ Ñı +ĠI H +izoph ren +Ġí ģ +_pr imary +Ġmo is +ĠB N +Ġsystem ic +Ġdifer entes +IN CT +Ġ'' ĊĊ +$ q +Widget Item +cl ide +$ file +L emma +/ table +ag rid +ĠMongo DB +int e +Ġapp rent +ÂŃ ing +.D b +Ġà Ĥ +ham mer +=' ';Ċ +Ġbro kers +it lement +sembl ies +E le +{ x +Ġlast name +< - +Ġfl atten +_b and +.R oot +.read FileSync +==== == +.r x +? čĊ +Ġmetaph or +T i +con te +Ġdeb it +Ġcont empt +Cpp Type +æĶ ¯ +Form Field +r atio +os opher +Ġimpl ant +P URE +Ġal ta +_man agement +Ġref ine +ĠCheck Box +ĠChar l +- version +cond itional +ven ues +Ġrif les +Ġoff spring +Ġmill ing +Ġshar ply +Ġunder water +( origin +_ Control +Ġ. $ +Pl ugins +Ġdry ing +Ġillustr ates +- u +Ġveget arian +n pc +He art +; ',Ċ +com ma +te enth +as an +/s pec +_m oves +-m argin +Ġing en +³³ Âł +Ġpro jet +Ġo tra +Ġbr as +. utc +Ġsle pt += sub +ab ilit +post er +Ġs dk +ounc ill +Ġw d +Pre paredStatement +ĠDr um +( attribute +ĠEther net +ĉ DB +Cal ifornia +c ube +[ I +.C reated +ĠH M +Ġtr acing +Forms Module +- you +.c urrency +feed ing +Ġt body +L i +acc ion +n as +Ġtr ouver +N ONE +"} ,čĊ +Ġf tp +With Identifier +pol ate +File Info +Ġpurs ued +ĠĠĠĠčĊ ĠĠĠĠčĊ +DE SCRIPTION +} */Ċ +From Nib +Ġdecor ative +_S SL +(ch at +T LS +Ġsurpr ises +al culate +ĠS plash +( Configuration +ĠS EM +im son +/lib rary +< Double +. robot +³³³³ ³³³³ +ĠCP F +ĠUnder standing +Ġcos metic +ĠX t +t ips ++ k +(" ' +ĠP DT +W AR +.get Object +ĠTrad itional +.sl ug +ĠDi pl +=" ", +ĠFil ms +ĠAn im +.h elp +Ġemb assy +ĠBoot s +Ġb unk +-r isk +Ġp ci +Ġ/ \. +ĠI PT +Ġcrash ing +Ġip v +_ ke +ĠRES P +.Log Error +Ġinade quate +I on +ĠF ür +ric ula +Ġshould Be +al ready +']." +G ED +fa q +Ġoption ally +_D is +ĠSuccess ful +ĠC ensus +Ġinc arcer +_C ARD +Ġav iation +ĠG ym +Author ity +.B ean +sh ader +Not Exist +_Text Changed +ĠST OP +( team +" H +w g +Ġgr inder +Ġstri pe +Ġpres ervation +Cl aim +avers al +ware house +target s +Tr ust +Ġal lev +, www +ous se +_ch an +_S ize +system s +Ġobj ection +ĠK ane +Ġcor ros +ĠD SL +Ġu a +ĠM H +ĠStrateg ic +_t cp +Ġê° Ĵ +Ġborrow ed +ĠA ch +ĉ command +Ġg ps +le ston +iche ver +ĠU A +Ġassault ed +Ġspecial izes +ĉ search +Hot el +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ čĊ +ĠP itch +Ġ Ùģ +READ Y +Ġparent al +Ġg éné +Ġdonn ées +Ġdet ain +T ARGET +Ġprotagon ist +Ġclear Interval +ĠIcon Button +ĠGet All +Type Info +E H +âĢľ They +Ġ{ [ +Ġg ag +Ġ Ú© +ĠD ropdown +.f ree +g one +im ens +Ġinst al +ĉc url +_C AN +ĠB one +ï¼ Ķ +ony ms +-g overnment +.binding Navigator +ĠD ans +ĠMc L +( en +>( _ +ÐĴ Ñĭ +.* ;čĊ += j +-c or +S on +.ToolStrip Item +- around +_X ML +end Date +Ġsl ack +Ġrot ated +Ġno qa +Ġc ottage +Ġencontr ar +_s kill +hou ette +! čĊ +. weather +Ġemphas ized +å® ¶ +ĠÑģ пиÑģ +ĠComp iler +( android +ĠâĢ º +. turn +Ġsup pression +_c alls +Ġ* @ +(str len +.h ex +ĠB ills +ĠR SA +Ï Ĥ +ĠEs cape +ement ia +Ġfront end +Ġp int +_ex c +zz o +[ ],Ċ +Ġ"',' " +. Environment +Ġafore mentioned +Ġend ure +prot otype +ther apy +ss i +D eg +_pl ugins +.user Info +Print er +ĠPRO GRAM +Ġru ins +Ġempir ical +Ġcraw l +ĠBo iler +- comment +.sub plot +_ et +Ġ'. ', +min or +ĠCustom s +Ġy aw +under line +ĠCom o +( (' +(m ean +Ġcha que +ĠBlock s +.r ad +ilib rium +Ġweb driver +Ġmel hor +d ana +ĠAb use +ĠSouth west +ĠP aren +PERT IES +ĉ IL +Ġscre am +v u +Ġin comes +Ġn im +Ġl ace +Ġcompens ate +Re verse +D at +_att ack +Ġn our +ach en +ce k +< Func +w ie +com pressed +-m atch +(" ")]Ċ +im ized +. orientation +.compare To +Ġmass aggi +Ġìľ Ħ +Ġel bow +Ġant ioxid +undred s +/ tools +ĠR OW +an mar +ĠW ow +_t icket +Program ming +Ġthe or +-re view +() )));Ċ +ĠRichard son +ĠP ocket +] [] +am pp +_ health +ĠP OP +ĠNav al +Gu ess +Ġancest or +.Get All +.local Scale +ĠM apper +Ġaccum ulation +Ġsim ulated +ĠDr ivers +Ġd és +cur ring +Ġele phant +Ġadvert ised +Ġmail box +SH IFT +ĠMon ica +Ġan c +Ġward robe +Ing redients +Ġ|| čĊ +ipp y +Ġantibiot ics +av ings +(c x +ĠFerr ari +ĠAn imator +.d type +rem oved +order by +Ġc res +oc ê +Ġp ym +ĠCirc ular +@ index +ĠW arm +S ay +ĠAss istance +Ġcur tain +ĠMont e +IL ER +ĠC VE +ĠD uck +ĠAll ows +_f ire +ĠDer by +Ġre pos +Ġhttp Client +Ġpsych iat +Ġnow adays +Ġcaut ious +ĠComput ing +Ġcompletion Handler +ĠWel sh +ĠB EST +Ġstress ful +_P E +æĹ¥ æľŁ +ĠData Frame +ĉ Integer +_P rint +M oves +Ġtransform ing +.B atch +y ahoo +Position s +ze j +Ġno od +io res +_ * +Ġcl k +ĠF loyd +Ġh ap +font size +Ġn az +.not ification +ĠDep ression +Ġac ne +*** ĊĊ +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĊ +.cont ents +yn th +ĠStra ight +')}} "> "+ +Ġtoken izer +Ġsovere ignty +ĠP ence +() ");Ċ +Ġpesso as +.G e +ĠIn cluded +Ġpag ina +Ġex posing +е ÑĪ +_SC RIPT +/$ ', +Th umbnail +× Ķ +webElement X +webElementX paths +press ure +ĠCur ry +_C P +OL UTION +ILE S +prot ect +ool a +Work space +{ };Ċ +ĠU NS +Ġsymp athy +ro ker +Ġrem odel +ĉc ell +Ġat op +.Full Name +Ġfa ut +ĠE asily +_d ynamic +Ġfr amed +Ġmot ive +è· ¯ +s am +Ġmar ca +ĠText EditingController +Ġde structor +cre am +Ġr ude +ĠB old +ĠInd igenous +Ġg ens +Ġrel acion +(s ystem +ĠUIF ont +_char ge +UST ER +E V +.N amespace +Ġmer ger +Ġcal loc +g ang +Bad Request +Ġs per +-d esign +Ġâ ĩ +Ch an +Ġorgan ism +, ) += id +_pl ane +ĠC ases +elf ast +ĠLegisl ature +ĠF aker +Ġinv oking +- utils +(). ' +.f ace +Ġguard ian +my Modal +Ġclip board +ĠAT M +Ġpe as +ĠS ylv +.c alc +ĠContact s +int Value +Ġmodify ing +ĠBar b +. loss +_per centage +Ask ed +(l st +ategor ical +- files +ĠRoman ia +.A c +Ġh ai +ĠF lying +Ġ ż +j p +ĠTr ainer +. arc +_de g +Ġtrace back +Or Fail +F LOW +. old +oy a +g mt +is empty +Ġvacc ination +Ġob solete +recogn ized +Ġru ined +ĠRe in +ĠTr acking +xf b +ا ÛĮ +Ġvæ re +Ġbr yster +ĠIT S +Ġdest iny +Ġsw ear +Ġred es +Ġcl f +Ġfl ipped +ĉ head +Bl uetooth +ĠOver rides +: Boolean +_ = +_l r +sp awn +: index +VAL UES +is key +? ");Ċ +.syn thetic +ĠCheck ing +struct ures +ip ing +Ġvoc als +- Up +ĠManufact urers +ĠMar riage +代 çłģ +Ġgar ner +_C lient +par allel +RI END +Ġvine gar +seg ue +J B +Ġcontact ing +ĠCar roll +Ġout reach +t ensor +_var iant +Ġthe at +lic able +{ | +t iny +_ letter +Ġp encil +HeadersHeight SizeMode +ilt ro +.auto configure +.d rag +.use State +ĠB MI +h int +Com pile +* \ +en ary +Ġl vl +.C ache ++ =" +_t v +ruit ment +Ġf read +Art icles +f ila +Ġpack aged +âĺ Ĩ +AT HER +ĠPl anned +s cheme +Ġdi ary +Ġoff enses +/ F +ĠSt ick +Ġc erc +ĠS lee +ĉĉ ĠĠĠĠĠĠĠĠ +< Image +Ġè® ¾ +- editor +pie ces +ĠD rama +Ġ// //////////////// +ĠT asks +AR C +g ateway +.get cwd +.M etadata +Ġguess ing +åľ° åĿĢ +Ġsm arter +ĠGet Enumerator +Ġe fter +/ operators +ĠGL float +Ġf ør +Ġop aque +ä¿Ŀ åŃĺ +Sp read +SY STEM +Ġinv ersion +ĠBasket ball +Ġsim ulations +Ġden ies +Ġa vez +_list ener +Ġenh ancing +ĠMy th +ĠL akers +_M D +Nd Ex +D ATABASE +Ġt á» +ar th +[ left +Ġcontest s +st ile +(K ERN +_f c +_p m +Ġpres idents +Ġhospital ity +Ġfade In +RO PERTY +_m aps +ĠDefinition s +Ġassess ing +Ġus ar +Ġquant itative +mo z +Be autiful +[ (( +b ons +f requency +Cont ain +Ġpuzz les +ĠCast ro +Ġv illa +Ġkind ly +Font Awesome +ern a +epoch s +_dat as +ĉ ip +.p adding +ĠCont est +Ġed itions +Ġdispro portion +ĠI CO +Ġcome back += value +ri ad +-s ort +Sub mitted +(n etwork +ĠC el +Ġinstall ment +l ashes +.List View +ĠV atican +(Media Type +IV ED +reach able +: Is +ĠC ITY +äº ¬ +ĠHelp ful +Ġba ÅŁ +% čĊ +Ġpsych iatric +Ġrec ycled +FORM AT +ĠG row +b ine +G it +.s s +ĠWe apons +ĠSt y +_ arrow +* self +ire ment +Ġdeg li +App Delegate +_b anner +Ġcoordin ated +ĠWeb cam +Ġcelebr ations +. act +******************************** **************** +( show +Ġweek day +Ġconc erts +ол н +cl in +Ġcr on +ĠN im +.set Vertical +ĠEll en +س ت +ĠS AM +E ff +g z +ste am +Ġant ique +ph ysical +ĠForm Data +.set ter +ĠPO INT +B on +Ġflav our +erv ention +_ENT ITY +ĉ ĠĠĠĠĠĠĠĠĠĠĠĠ +Ġintr insic +Ġæ İ +append To +aram el +) ]) +ĠRecomm end +) m +OutOf Range +Ġkn ight +Ġsat ellites +ĠTit ans +Ġweigh ed +ĠD ana +e ase +Ġs ip +S IM +ĠDevelop ers +mal ink +/ check +_P LL +n ung +Ġdry er += A +.d w +_S QL +Ġsub plot +D ROP +Ġprot otypes +Ġhour ly +display Name +Ġas i +ĠViol ence +Ġastr onaut +Ġdat atype +Ġinformation al +Ġinvestig ative +etermin ed +ren al +; '> +ĉc ol +V G +_ boolean +re cent +Ġ* )ĊĊ +ĠRain bow +om men +Ġl ur +Ġopp ression +(", ");Ċ +ĠFac ility +DEF INED +Ġne on +Ġoff ender +AF P +ĠClean ing +[] ): +Ġund ocumented +.Re positories +ĠG uitar +аÑģÑģ ив +Sk ills +Ġtestim on +rypt ography +ĠAm ber +ĠSt alin +Ġl one +Ġap enas +Ġdies es +ĠAr duino +è½ ¬ +== - +_A ct +Ġc oded +âĸ ł +amb urger +-link s +Ġarm our +.H igh +get Content +st ag +Ġhe ck +ĠìĹ Ĩ +ĠMc Connell +ĠCon cert +ĠAl loc +ä re +.replace All +Ġpart itions +rot t +ĠF le +_T REE +reason able +ĠReport ing +Ġbillion aire +s cores +min s +- eye +M ORE +ab ort +ĠSW T +Ġin verted +ĠTe achers +; n +Ġast ro +н ов +ани ÑĨ +product o +c ountries +ĠO wen +Ġcont amination +Ġv ibe +ĠEll i +.s cript +ĠOl ive +D MA +v ier +: semicolon +-m odule +gress ive +ag u +_ players +Ġresult ados +start ed +scroll Top +==== = +Ġweigh ing +Ġ[[ [ +z ahl +( NS +ĠAssert ion +le ague +.setText Color +ĉ Message +Ġmom s +_A F +. wh +AL S +Ġaut re +] ĊĊĊĊ +.op acity +ĠBudd hist +Ġde af +ĠOrgan isation +(G lobal +ens ch +Ġhead ache +ĠAli en +_in ode +ĠSt ark +Ġæ ī +-l nd +ore f +_fe at +Ġpedest rian +Ġnom inal +Ġbal loon +Ġspr ites +Prototype Of +ĠA post +ĠF EATURE +O H +Ġre cess +ĠDon na +con sumer +$ GLOBALS +ĠG IF +- frame +In icio +Ġpass ages +Date String +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠ +.by te +B ug +initial izer +p kt +od ium +ĠD ER +. ops +ler i +Ġgift ed +Ġdet ach +ter rain +elt ers +ãģ ı +. loader +ĠN GO +str ncmp +K h +(font Size +ro cket +Ġpreced ent +ĠAur ora +ĠEx periment +is phere +Enc oded +ĠâĢĵ ĊĊ +Ġpy ramid +ĠAnn iversary +of il +ë Ł +( plugin +C oeff +Ġcooper ate +Ġpredomin antly +IS M +Ph rase +_DEF INE +Fl ip +AMIL Y +ĠMark ets +ĠStream Reader +ĠComb ine +Ġmanus cript +z za +, tp +Wh atever +IT ICAL +ighb our +Data Provider +.Text ure +priv acy +.S DK +Ġre charge +Ġc pp +ĠC FG +(h older +(p y +m ot +Ġsav oir +ĠR osa +ĠPC s +Ġí Ļ +.her oku +Ġf ren +ĠR iley +ag ate +Ġs ond +.x lsx +Ġh acked +st ad +G i +Ġsan ity +ĠSql DataAdapter +... ", +ĠP ussy +Ġ **************** +Ġhass le +_P ARENT +ĠU AE +Ġbegin ners +( Client +Ġstatist ically +.h our +ed elta +Ġtr action +uel ve +ar at +Ġsa una +IN VALID +Ġindict ment +AL LE +Ġdiss ent +ĠTyp ography +Ġintention al +s it +ĠAn imals +Ġcoun tryside +Ġu art +} \" +Ġseam less +¾ 示 +Ġaut os +Ġ"' ";Ċ +Fl ush +ANN OT +Ġal gebra +ass oc +ĠW aters +Ġprepar ations +ron ym +[, ] +S ans +Ġarm ies +ipe g +Ġcream y +. art +et re +ĠAn imated +Ġun pleasant +eme an +g reat +i Äħ +ĠEar lier +Ġch ic +Ġpres erving +(ex ec +ĠInvest igation +ĉG PIO +Ġrig orous +ij o += num +Ġtool Strip +) set ++" & +ĠAcc eler +Ġdevelopment al +is posable +Ġflaw ed +re ne +Up dating +Ġwatch dog +Ġden ominator +Ġsubur bs +Ġ... ) +Ġconv ictions +c losure +.I P +Ġtransl ates +.sw t +.Tr ace +Ġmet tre +.is Enabled +ĠEffect ive +.to Int +Ġen chant +Ġst unned +Ġpo i +/ code +ad m +.datab inding +ĠL orem +________________________________ ________________________________ +Ġled ger +Ġcar a +ĠG ir +Ġwa its +Un o +Ġc wd +è¾ ij +ĠT Result +Ġre jo +Ġem itted +ĠWest minster +ä¸Ģ 个 +ne k +_T is +Ġen act +ĉ with +org ia +Ġj ue +Per form +SP ATH +.top ic +ĠD aten +Ạ§ +Ġsit io +_M M +" So +b ial +Ġsc oped +Re quires +ĠT OTAL +ĠCh ancellor +( contents +Ġste alth +dev ices +-p ass +ili h +ĠMal colm +ĠDep ot +Ġconfig ur +a ussian +_con straint +в еÑĤ +G RA +ĠR ates +.dataGridView TextBoxColumn +ĠNob el +it ics +Ġignor ant +ĠReport er +ĠEb ola +ĠSh ock +_re lation +ĠNin ja +) c +Ġt icker +.is Checked +ĠSup pliers +ĠRap id +Level s +âĤ¬ âĦ¢ +ĉ queue +Ġch op +ĠUn ix +re ject +-c alendar +(s ort +è ne +erc icio +Ġh ect +CALL TYPE +rou pon +Ġrent als +auth ors +{ name +ĠF IFO +Ġl assen +ĠN ous +Ġsn apped +Ġfert ility +" log +click ed +Ġplant ing +Ġg b +/ output +PE AT +Ġc ategoria +Ġb ach +Prof essor +in th +"] čĊ +Rec order +ser de +ĠTrans mission +tr ad +Ġtur bo +_VER TEX +\ Event +il ver +Ġbod ily +ĠS ources +Ġkill ings +.xr TableCell +Ġfold ed +/ legal +un er +ĠR ifle +ĠM IDI +_Selected IndexChanged +.Size Type +ĠWeb Socket +Ġsele ccion +S and +ot ros +Ġenv ision +/ etc +ĠMel issa +Sp ot +но е +_ ARM +At tempt +ĠB I +ãģ Ķ +ĠD U +Ġback lash +str ide +/ classes +Ġtext Color +_st aff +ob lin +agent a +.c ollections +ill age +' čĊčĊ +fl atten +_s ales +_M ASTER +T W +_d a +P itch +ph ies +Ġz ombies +ĠV ERY +ĠPharm acy +Ġprogress Bar +Ġhas htag +S idebar +@ stop +(p c +ол ж +MA KE +ĠCor on +Ġkv inner +ĠM aid +b ob +.title Label +Ġsuccess es +ĠDemocr acy +ĠSurg ery +Ġcou gar +Ġcur so +Ġl oro +ist ency +Sen ior +æ k +ĠA AA +ĠBO OK +к о +W STR +Ġ*/ ,Ċ +oy al +.v ector +ĠS PEC +SS F +Ġcomp uls +ĠAppe als +ĠW inston +ĠMock ito +con trib +. available +entity Manager +ari as +_s ale +_r s +Ġdec oding +Ġloc ator +ol ith +Ġk ol +Ġasc ii +ĠR ut +/ interface +ĉĉĉĉĉĉ ĠĠĠ +ĠN umer +.fl ip +-d el +Ġbol ster +on omic +Ġz m +L G +Find By +Ġadapt ive +lo o +Ġv ue +(re verse +_c anvas +. roles +ific ado +ven ient +" As +ĠEn tr +al igned +Ġbere its +/// ĊĊ +.g wt +. employee +_cl i +Ġanticip ate +éĻ IJ +Ġp ik +Ġmush rooms +(t t +Ġo ma +ĠSan chez +_g oogle +. Valid +ĠFile Name +iv ative +k ed +-w ar +Ġm aturity +и д +Ġmin er +Reduc ers +ĠLat Lng +_ST D +D igits +Cal c +-up load +Ġhand ic +ี à¹Ī +egr ated +ĠST M +C lients +ĠTur bo +SY NC +Ġphotograph ers +. Out +.char acter +B UILD +.un lock +Ġar ises +ĠCommand s +(" ");čĊ +_F ORE +; ', ++" ' +. Images +") { +ĠM eyer +Ġneg atively +ĠD LL +Ġex e +Ġdef iciency +Ġwild ly +-s witch +con struction +Ġexception ally +ĠL iz +/j ava +Ġtheir s +ĠCont emporary +l is +.fill Rect +ĠN FC +Ġre he +(num bers +Ġr aster +Ġfig uring +Ġshow c +ĠJ ill +Ġarc ade +ĠConstruct s +md l +(' | +Ġident ifiers +Ġst ellar +( Connection +Ġ" {{ +y or +(m ysqli +Ġdo ve +Of Birth +.dis connect +_h i +Ġzw ischen +ĠGr und +i ros +_A rray +.on click +ans om +An swers +ĉ remove +F a +Ġhur ry +-in f +Ġget Class +ĠReg ulation +ĠFLAG S +m isc +K en +_ heading +G Hz +- entry +Ġbi ography +S ig +-m f +Watch er +âĢľ A +} px +Ġsp icy +_s q +L ost +(tr ack +а ли +Desc ending +< bits +qu ine +ĠAdv oc +_S N +ĠHann ah +PO P +Ġem itter +Ġc yn +ĠC AD +? ). +/ set +ĠS ister +ĠEnd point +Ġmen or +Ġinter p +r k +id le +Ġout fits +. vertex +Ġc lic +ARE N +Ġpost ure +ĠOpport unity +v x +ĠFor bes +.D irection +Ġres ide +Ġremember ing +nest y +Auto resizing +pro viders +ĠA H +Ġhur ting +ĠL ily +eval uate +lij k +p apers +ĠSm ash +ĠL AST +Ġwell s +w asher +_RO LE +ĠD anger +* (( +_re pository +ĠRes olve +ĠRoom s +_R G +ĠQ T +o op +ĠHe ap +Ġslow ing +Ġgrat uite +_c atalog +Ġpol ynomial +L y +pc s +F ox +ĠC yr +Ġdim in +/ month +S alt +Ġh ind +.P ER +For um +c en +_p ol +íĺ ¸ +Ġin ser +( ~ +@ test +ĠGold man +Ġupload ing +F c +Ġkom mer +Ġm itt +_log ged +Ġbu cks +-l ayer +) };Ċ +ĠO M +Ġv eg +col our +Ġоб ÑĬ +Std String +_ que +ĠT ian +Ġspecial ize +и п +Ġк л +tr ial +- edge +Ġm ars +OG LE +Ġempath y +ĠB om +Ġcoll isions +Ġcart e +ĠTe il +ĠM PL +Ġporn ô +Ġa irlines +A ws +N s +ĠSp awn +( use +é» ĺ认 +Ġy acc +st or +Ġconf ess +Ġpe que +r age +? "Ċ +/dat atables +ĠSh ower +__ / +Ġcryst als +Ġbus car +ĠH aus +iz ação +_ entities +ķ Į +ļ Į +x cc +v irt +-che vron +( Result +c ake +COM E +Ġprohib it +ĠCh ess +Ġbe aucoup +ĠÑĩ ÑĤо +R UN +ĠI K +ó ÅĤ +_ Update +Ġsle ek +ĠSpec ify +_c redentials +ÅŁ t +ĠUser Name +ĉ Value +Ġarray List +Ġex changed +ips is +.re lated +ĠSe ite +_B AR +ĠL em +ĠW ATCH +ĠC lients +Ġ. * +ĠEar l +-re port +Ġforeign ers +Ġstrengthen ing +ĉ Description +(g o +.tool bar +Ġcalcul ates +ĉs ource +Ġcz as +Ġre cl +ab o +Ġlocal host +Ġ^ {Ċ +.P op +ĠDes igned +\ Abstract +H old +ĠGuid elines +ipl ine +Ġc aching +.Re ader +_ext ernal +.str ptime +ĠWeek end +-M ar +ĠBe i +Ġ{* } +ĠR ud +Ġexpl or +ĠBou levard +C ash +Ġprep ares +Ġserial ization +ew ater +Ġad c +: ĊĊĊĊĊĊ +Re fer +Ġsc anned +} }ĊĊ +ĠF ul +Ġtour ing +ãĥĥ ãĤ¯ +> (( +sur vey +Ġí ĺ +... ')Ċ +ĠDiv ider +os l +_C ANCEL +_pre pare +st in +ĠHe ath +.Primary Key +ĠâĨ IJ +ĠLocal DateTime +Ġcooper ative +L earning +.en queue +Ġgo og +ĠReg ression +im ates +Ġvoy eur +ĠDr ink +pl ug +Ġl ender +man a +Ġperson nes +yp se +Ġun link +ĠRav ens +Ġhur d +Ġperiod ically +ARG S +ĠG H +char acters +... "ĊĊ +- establish +Ġd n +( condition +ĠGr avity +Ġest as +_f ocus +Creat ure +(s ite +Ġc arr +ĠR L +ĠR I +ĠM oto +AS F +ĠLuck ily +ĉ Route +Ġent ropy +(" ," +Col lect +( contact +ĠFlo rence +Ġpremium s +Ġlif ecycle +Ġb ans +x ef +Web Kit +ĠFlo ating +Ġcos a +Spec ific +ĠLo ans +b read +Ġdes criptors +Ġ{ :. +TH READ +ĠT rent +Ġsc op +Q A +ĠAnt ar +p el +_d ifference +_ch anges +(... ) +ĠR otation +ĠLG PL +ĠJ UST +(T ask +_sub set +ĠTR ANS +åĬ Ľ +ĠSc out +-p opup +Ġsm oked +_C lass +Ġturn over +br akk +ĠRock y +t as +.Regular Expressions +ĠElli ott +ĠSp inner +DU CTION +Ġlib re +Ġmol to +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠ +ĠF TP +m peg +(f eatures +Ġb ald +ĠV id +Ġsh outing +L int +Ġsock ets +Ġpro w +Ġnouvel le +isc ard +ĠS ponsor +Ġconsult a +)) ); +Ind ian +ĠR aspberry +Ġteam mate +ĠJ WT +ĠGh ana +Ġc akes +pr imer +form a +erg arten +_M anager +Ġpre season +G AME +| " +ĠBro ck +Ġoccup y +Ġdecor ations +á nd +Ġc ot +Ġpar an +D isk +rem ain +> ? +Str ong +Ġfr ance +ĠE ra +-c r +.Buffer edReader +ĠParad ise +ĠV AT +ĠAnd ers +Ġlim b +amp oo +Ġimper ative +UT ILITY +ĠRec ognition +Ġragaz ze +Ġpop s +yp ress +Ġemb argo +// {Ċ +Ġsy ll +P TR +åŃĺ åľ¨ +Ġdid nt +Mail er +Ġacad emics +ĠFra uen +ne ider +- rel +Ġrain bow +( In +Ġslic ed +============ =Ċ +(s end +NSMutable Dictionary +v os +(p ackage +Ġord inance +view er +ĠSant os +-s elling +Ġgo v +ett le +Ġfound ers +Ġw aking +sl ashes +-p ound +re cht +ا ت +.on Click +Ġn ord +st änd +_ when +UT ERS +ic c +Ġcaps ule +ĠW id +M arc +ภ¸ +ro red +UG E +LO UD +ĠAud it +ip ients +op ian +ĠS ue +Ġwur den +.H elpers +Ġf actions +[ np +-th an +Ġre co +Ġk as +Ġcmd s +/n etwork +xb f +get Color +Ġbi ased +ĠL ak +D atas +vent s +Ġë ² +_P S +. Validate +Inv oker +Ġne uen +Ġju venile +V ISION +Ġdev ote +Ġlin ha +Ġdiscount ed +\ Config +Ġworth while +Ġskin ny +ĠC ourses +le ys +ĠMort gage +K evin +Ġannounc es +]) * +res ervation +Ġæķ ° +Ġprejud ice +ĠString Comparison +Ġbe ard +-w in +ĠS ão +ĉ ms +j al +ĠE arn +_ ports +ĠN ombre +_C OR +ĠB UILD +.s ound +Y ellow +Ġlineback er +Ġchar itable +j ug +_NON NULL +ĠD ental +"> ${ +ĉm atch +R ussian +Ġvers ch +Ġp inned +Ġadopt ing +Options Menu +P ag +Ġpair ing +Ġt read +erc ises +ĠSp read +) i +ĠB AD +_t f +UI ImageView +pop ulate +b ab +ĠÏ ĥ +[ ++ +Ġopi oid +Ġ## Ċ +d type +ĠStart s +('/ ') +Ġperson als +-mark et +Ġredund ant +ĠEss ential +Ġscrap y +Ġи м +a cl +Ġcre ar +ĠB end +Ġrel ieve +- room +w ife +Ġv Ãł +ĠQ Point +Ġqu asi +Ġmethod Name +\x c +ĠPer u +/ The +. orm +Ġv iz +/p df +Loc ated +Ġconfront ation +ĠChampionship s +Ġhyp ert +Ġd j +ĠUser Info +ĠåĪ Ľå»º +\x b +(s im +Ġ== Ċ +Ġst aging +Ġdr astically +åŃ ¦ +l ords +. less +вед иÑĤе +ĠB ucket +ĠM am +. term +_p i +c zy +.p ub +prec io +ĠV irt +Ġrom an +it at +L ex +_inf os +Ä ° +. other +VE LO +Ġp onder +Ġh anno +( Page +do i +Ġpol ite +Ġprogram mer +D ies +$ d +Ġrep lication +add Column +fr ican +Ġl eng +be er +o it +Ġw asting +yl im +me asure +N eg +Ġpart ie +.con sole +ĠGu inea +TE L +_f act +.ch unk +Ġl ent +Ġall er +Ġठķ +_id le +Ġad missions +JSON Array +Ġv ibration +.h elpers +å¤ ĸ +Ġh en +j ohn +Ġì ĥĿ +Ġjud gement +Ġge en +ter ra +^ { +ĠI z +Ġc â +inst ances +Ġthreat ens +Ġm üssen +Kind OfClass +Ġstoryt elling +_d emo +ri as +Priv acy +h ift +ĠY i +es or +íķ ł +ens itivity +.W riter +ภĤ +D istrict +.get JSONObject +Im pro +(get Resources +ĠS PELL +rodu ce +Ġslow ed +Ġlin ewidth +Ġhonest y +ĠCo ord +ĠF ork +ĠDispatch Queue +ĠCl iff +ĠW iring +_TIM ESTAMP +oll ah +av oid +++ ];Ċ +sem antic +-c ss +Ġv eto +ĠM err +Ġlegisl ators +CEE DED +Ġquestion naire +ĠP ills +Cal culate +(c ore +' e +Ġdis like +ĠPre ferences +_EX TERNAL +è° ĥ +Ġd odge +æľį åĬ¡ +.n ames +.draw Image +_p rom +uck land +Ġ<$ > +ı z +/s ite +é¡ ¹ +rop he +Ġcomp elled +Ġl aptops +Ġun i +C LOSE +Ġcasual ties +ĠUn iform +Term inal +. "," +D AT +(T reeNode +ĠGand hi +(st mt +AX B +* M +Ġumb rella +an imal +Ġgr pc +Ġwhere by +Ġfloat s +ĉ arg +Ġdb g +Ġexceed ing +Event Type +.SaveChanges Async +Ġ{ {{ +Ġow ed +ahren heit +Ġì § +Ġequ ipo +ur ai +Ġid ol +] ")Ċ +_m ajor +Ġentire ty +inger print +ç os +/ account +ĉ right +urs os +ĠE DT +_INS ERT +Ġsh ining +Ġ< : +Edge Insets +Ġcolon ies +. IM +ĉĠ ĉ +RO AD +CC CC +pl acing +Ġget Activity +em acs +' %( +.click ed +ĠTh em +is ia +Bus car +.re name +Ġo ath +Ġafter ward +ĠU FO +AP S +ĠJackson ville +.s ome +Conf irmed +.s can +ig Integer +Decor ator +sh ield +ress ive +.d id +请 è¾ĵåħ¥ +Ġsh utter +D am +Ġparent ing +ey ed +$ item +-de velop +Ġextract s +Ġdecentral ized +ĠEl sa +_sp in +]) + +-in itial +Ġmult itude +Ġsens ory +ĠMODE L +Ġsafeg uard +ì ¹ +Ġhunt ers +ĠT iny +IN O +decor ate +ĠNo Such +H o +( Response +Ġr uler +ĉ short +Ġc aster +Ġclient Id +Ġp db +ëı Ħ +it ic +ĠGame State +Ġnew Item +)ĊĊ ĊĊĊĊ +ou is +n oc +.BL ACK +_V ECTOR +---------- (); +.get P +any e +Ġneur on +if old +ĠK nown +Bit coin +Any way +ay ette +Ġ' [' +Ãł nh +m gr +Ġcor related +Ġn ause +Ġment ality +has Many +ĠF G +amp ie +IT U +F s +.S p +_b etween +Dep endencies +ou g +Place holder += text +ĠMan aging +ocal ypse +åĮ Ĺ +_m ag +f ld +â ij +C AM +ĠHelp ers +Ġd ost +/ out +Ġassass ination +.get Image +ĠKenn y +.' )ĊĊ +){ // +ĠR anger +Ġg ek +Ġsinc ere +< Value +ĠD OT +ĠVict ory +Ġleg ends +Ġpr isons +(ex pression +ĠR abbit +_s entence +Ġbit es +Ġon Failure +ĠâĪ Ī +K im +.g ender +ĠÎ » +Ġ[ . +"] ); +land ing +-d igit +TE MP +ĉ entry +Ġstrt ok +Ġdesc endants +um no +Ġlean ing +Ġspecific s +q n +ĠSp art +Ġpor r +EDIATE K +Ġse per +' aut +ĠSTE P +ĠBorder Layout +Ġret ros +ĠSalv ador +ĠEN GINE +x dc +T weet +v k +Ġì ² +] << +het ics +c oding +Re ach +.re q +gu ide +.s cope +sh irt +rog ate +SET TING +ĠProte in +Ġe ing +. EMPTY +.d f +Ġclear er +Ġc rossover +ĠTo ys +Ġco ated +.M onth +ĠAtt ach +/ run +.t abs +Ġogs Ã¥ +B rown +.D ATE +Ġf os +åŃŠ符 +W ood +-th ree +her ited +Ġ rop +( ac +Ġembod iment +ĠKenn eth +Ġcan non +Ġb idding +čĊ +.get Resources +Ġl ump +_const s +( ext +ĉd ir +â Ŀ +Ġpadding Top +Ġobs ession +Ġb anning +ĠApp Module +Ġpart isan +Ġcatalog ue +Ġmin ors +Ġpitch es +we ep +Ġundert ake +Ġthem ed +aud it +.scroll Top +Ġr er +Ġsympt om +Ġopen ings +.block s +open id +Ġas sh +-s ave +ĠP ig +Ġreg ain +Ġin icial +/f avicon +ĉ exp +Ġsp ices +isk a +claim s +m ak +definition s +Ġcorrespond ent +ĠCann abis +__ ,Ċ +ĠL ucky +ĠGa ussian +ĠN early +C AD +'] ]Ċ +Ġadequ ately +ĠT ITLE +constitution al +-m m +_ override +Ġbl as +.ready State +Ġremin is +Ġrein forced +ĠColl abor +Ġdecor ating +Ġb achelor +ERRU PT +Ġup right +ip ation +ĠNob le +Ġvalue ForKey +Ġset Loading +.I gnore +å ģ +G lobals +ĠM ent +AS SES +Ġlim bs +ĠH UD +inc i +. iv +ĠQ ModelIndex +F use +Ġped al +_F REQ +( verbose +Ġlong itud +ĠChar ter +ê ·¸ +Ġbund les +. ignore +um bo +EM A +.... ... +s x +.C ard +Ġhe ute +Ġste er +j umlah +Ġ{ _ +_Check ed +Ġf ax +ĠG ust +itch ens +Ġ ))ĊĊ +Ġremark ably +/ XML +- remove +_b t +Ġinc ub +.p ackage +.current Thread +ĠHigh lander +.s ide +s plash +Ġ ici += D +Ġp uck +Ġball ots +Ġhug ely +co eff +Ġp Data +.C OLUMN +ĠHe aling +Ġord in +! ), +Ġ' ',čĊ +(m d +ĠS ask +< strong +Ġsurviv or +.s eries +Ġcaffe ine +Ġ` ( +.TRA ILING +_ Input +(" ^ +z d +& );Ċ +ĠP ing +Ġv oucher +.r ating +-sh irts +ĠRetrie ves +.al ibaba +Or acle +_MO V +Old Data +Ġ/* čĊ +Ġg boolean +Ġ=> čĊ +Ġr á +Ġbl unt +ĠImage Icon +if ik +RT C +Ġfib ers +Ġto ile +.s ent +ĠPy Qt +$ app +Ġmed io +Ġgrant ing +Ġtsl int +ĠM ö +(fig size +Ġhur ricane +Ġlif es +Ġà Ħ +rocess ing +_st andard +- option +')) ) +Ġvac ant +å· ¥ +ĠH ollow +handle Change +Ġdiv ider +ĠEngine ers +Ġsv ens +Ġcompl iant +t anggal +ĠC redits +ĠEm irates +Rule Context +Ġreal ization +Ġdistr acted +]+ = +Ġaug ment +ĠD w +ot p +or rent +Edit ar +.st ock +St udy +pe ctions +ĠGame Manager += cut +Ġf lock +ĠRom ans +th em +-h op +Ġscreens hots +Ġ/* !Ċ +Ġconvers ions +Ġnormal ization +(config uration +Ġa eros +_se curity +! 'Ċ +B onus +ĠDR IVER +ĉ Date +t ie +ĠWy oming +St and +it re +Ġsh oppers +Ġdisadv antage +Ġlik ing +ç¬ ij +Ġunderstand able +SE E +Ġh oy +Ġnin ete +Ġcon fer +Ġnow rap +ĠV ern +, čĊčĊ +imest ep +Layout Manager +à · +ĉw ait +PLE TED +J apan +Ġindu ce +Ġå ¯ +оз в +_END POINT +.h orizontal +Ġacceler ated +rim on +IV ES +Trans actions +Le an +ĠSO UR +wh ether +y g +Ġo id +ĠEntity Manager +OUN TRY +Ġfil a +OLUM NS +IN UE +ĠAn chor +TR AN +wo o +block quote +ĠN urse +ĠCar p +Ġrede em +. try +ĠJ P +Ġtimestamp s +Ġ?> ">< +ĠREM OVE +ĠStar bucks +Re ally +Ġflood ed +.C allback +Drop Down +ip ro +Ġt ended +l te +Ġproport ions +- te +ĠR ena +lic ate +for ces +.ex tra +.auth enticate +в од +¡ ° +Ġfor ControlEvents +Ġsen ha +Ġke in +Ġmin ist +ĠPre ference +ĠTele graph +Ñĥ п +str pos +Ġillness es +Ġp igs +Ġget Intent +S ol +Ġ ¡ +(c pu +[ prop +s creens +'); ?> +ĠAct s +Ġstr dup +Ġaver ages +an al +ĠCas ual +Group Box +ĠHand book +/ comments +Ġnumber ed +Ġbroadcast ing +çĽ ij +.native Element +.m u +Ġupdated At +ĠDoes n +.A C +.c oll +Ġrec order +_sh a +B g +b il +Ġbol ts +Ġç ¬ +Ġim posing +ĠInformation en +_flash data +e conomic +Rem ark +uc as +ĠOff icers +ĠT ER +W alk +Ġmerc ado +_g enerate +H Y +Call ing +s nap +script Id +. operation +ĠFl ame +l iness +Ġrent ed +_t oggle +-ch anging +ĠT Y +' util +EE P +Ġgraph ql +ĠUn i +Ġimp ulse +.B asic +Ġenerg ies +M ARY +ĠMar cel +Ġmort al +Ġf res +m ens +m otion +Ġsample d +âĢľ That +id ay +qu ipment +get Int +ĠA bsolute +,' " +un ed +.sh are +Ġ} )( +mm m +ĠR ising +ä» » +Ġun employed +x fa +.f ollow +ĉĉĉĉ ĠĠĠĠĠĠ +sl t +.P hone +Ġkn ives +Ġe ve +on Click +] ))čĊ +ĠW itness +ĉ NS +ĠE OS +ĠSte fan +ĠPri est +âĢĶ which +Get String +. By +Ġup stairs +Ġdetr iment +bro ken +emb ro +Ġnic otine +il ion +Ġaston ishing +_ aff +ĠLess on +Ġaccident al +od or +Ġdec ir +Ġnew Name ++ . +çĽ ¸ +igs list +ĠG ithub +Ġsuccess ive +rac ial +Ġen viron +éªĮ è¯ģ +Ġredirect ed +T OTAL +Ġgrab bing +ĠL ance +Ġfor fe +_C B +å¾ ® +El apsed +_w ay +(Dialog Interface +_me asure +x bb +D og +Dep art +-s rc +res olver +with standing +_sh ell +ĠLast Name +ĠAv iation +Ġbegin ner +("% . +(to ol +Ġн ов +: init +(A PI +ĠMorr ison +vt Color +Ġstap le +/ INFO +Ġsupern atural +Ġste ak +tim eline +zz le +" `ĊĊ +Second ary +ĠNep al +.String Utils +Ġad am +Ġ( ... +Ġsub stitution +Ġboard ing +ĠKey word +ĠAss ault +dbc Template +Ġorder Id +( engine +.assert That +ĠVen us +Ġhomic ide +ĠA val +Ġg utter +ĠSupport ed +/p art +Ġac claimed +H istor +Ġmes es +ü ber +ĠRen ew +Ġgr as +ĠE k +Ġin file +ind y +.m usic +.S croll +ĠA ges +ĠNar uto +ĠG ather +Ġconfirm ing += (" +Ġpitch ed +ole y +Fr ance ++' " +$ total +Ġon de +Ġd itch +_s igma +Ġcontinu ity +re ward +- load +Ġproces o +Lock ed +st aw +Ġsp inal +l azy +! == +j est +Ġd un +ĠRod gers +ĉ grid +Ġlog os +ĠBeng al +.s uper +Provid es +Ġnut rient +.T imestamp +IZ ATION +åĨ Į +Ġf ats +ĠX xx +ct ica +Target s +Ġcont ours +Ġre ordered +: Array +Ġtoler ate +V ir +Ġter ribly +Ġbr icks +(& _ +h b +Port al +ĠB read +. which +ÂŃ t +as InstanceOf +Ġj object +ĉ length +_M T +; ">čĊ +_EX IST +Ġmat ernal +RE L +Ġê²½ ìļ° +he e +Ġlayout s +ĠL ap +ais y +Ġst umbled +ĠU IG +ĠS co +Ġimp aired +RES SED +Ġab uses +V F +AR B +.N AME +r ch +prim ir +_com pleted +Ġp enny +Ch rome +(b egin +ern en +- checkbox +Plain OldData +ĠL PC +r ade +sp ir +Ġcon ceived +T ips +ĠIo T +ĠG an +èģ Ķ +Ġbi ases +Ġconsult ants +ple d +_ ht +associ ated +], ĊĊ +Ġdelight ful +ĠÑĤ ек +Hel vetica +( load +-exp and +_W IDGET +to a +ĠA kt +Ġom n +Ġcl auses +Int el +*/ }Ċ +_reg istration +Ġold Value +Ġrest oring +Ġun real +O VER +ĉĊĉĊ ĉĊ +AT S +_pro be +Ġdiv isor +.update Dynamic +å¹ ³ +Produ ces +st amp +.j boss +ĉt ask +! (: +Ġpsych ic +@ class +M artin +ĠPass ed +clar ations +h el +а Ñĩ +ĉc opy +-b in +z an +ig ram +া ঠ+(s ig +ĠC aval +_ ## +Ġ% = +out lined +ĠAc id +Ġunpredict able +-d ashboard +Hex String ++ c +.P ublic +Ạ© +Ġconvey or +ĠE B +Ġselect s +Ġknock ing +ĠC ec +IBUT ES +owa Äĩ +g atsby +* v +ent ropy +Ġdispatch ed +Ġcam el +ĠSat urn +Ġover weight +( phone +par able +% B +_v ectors +Ġbrew ing +ĠT k +ĠDownload s +ĠS aved +.Pr ice +Ġcur ved +ĠParen thood +è ¶ +.p nl +plet ely +.D ay +Ġadvertis ers +Ġej ec +Ġpr zed +ë ¯ +! ';Ċ +ĠK ush +ĠT AB +Ġquest s +Ġcoinc idence +umm ies +ĠKash mir +ĠEth ics +_g rowth +Ġakt iv +Ġgroup ing +å¢ ŀ +_tr uth +åIJ ¬ +t odos +is et +Tex Coord +ä tt +ĠZ ur +ro ys +_M AGIC +Ġbrew ery +( State +ĠSM ALL +ĠPl ants +it bart +each er +ĠAd elaide +L u +Ġf ick +und les +_load ed +и е +P oll +rit ic +EL Y +Ġ+ ' +ĠProf ession +Ġst amps +ĠS ew +scroll View +Ġcomm unist +/pro blems +}čĊčĊ čĊčĊ +, o +Ġu dp +Ġob ese +appro ve +ancell ation +_G ame +ĠHas htable +adaptive Styles +Ġpossess es +.match er +function al +M rs +ĉs ave +ĠDb Type +Ġk en +get Context +Ġm ans +( rel +ĠBrother hood +) `Ċ +è§ £ +.In formation +OutOfRange Exception +ĠS ek +C as +Ġblog gers +E ither +(" "" +Ġpin ch +Ġco arse +) p +ĠP ulse +Ġlear nt +Ġdent ist +Ġon change +Ġdirect ives +( actions +ny der +ĠSh ir +T rait +_de p +ĠP ET +ĠRE P +.App Settings +cu ador +iden av +Ġenv i +Ġsl ammed +ĠSh oot +Ġdate Format +.j oda +ve ys +Ġ) .ĊĊ +Ġcare g +ĠPar allel +_ translation +.function s +. obs +Runtime Exception +[] = +over view +ĠSch l +Ġno isy +ĠOn PropertyChanged +S ending +Ġunf amiliar +U pon +ĠPrint s +.t yp +Ġflee ing +ĉm ove +( Un +Ġq r +× ľ +_b eta +Ġsk ies +ĉm e +W ND +Ġstick ers +bl as +Ġinsert s +Ġvers es +ĠD ew +Ġtang ible +Ġhe cho +P OL +Ġte ardown +om nia +IB E +.c over +_str ategy +^ - +set Position +u ale +S igned +Ġif ace +as eline +.set Time +ĠMin eral +ĠFight ing +sk ins +Ġdiscrim in +Ġdans k +ĠPr inceton +ac ist +Ġ( ));Ċ +tr acks +imon ial +ad ecimal +EP ROM +ugg le +.Not ification +$ mail +c antidad +ĠJ ung +Ġseek ers +Ġpl ausible +t ier +еР¶ +Ġr apper +ĠMan a +ĠHttp StatusCode +Ġburn t +los es +ĠF oto +ĠJson Object +Inst agram +Ġsys call +Ġreal ities +ĠMAT LAB +:^ {Ċ +TER M +ĠC bd +ĠPar agraph +Ġtrav és +Ġconstruct ing +Ġsw al +Ġp ige +LL LL +-ex isting +G ets +Ġmelt ed +Ġmitig ate +H en +Ġh m +im as +ĠA o +ĠP erez +ĠD AL +Ġëĭ ¤ +Ġdiv is +Storyboard Segue +ĠMod ify +ĠÃľ ber +_O VERRIDE +.p em +unt os +Ġespa ñ +Ġ{ ? +ĠP AY +_ip v +ĠF ury +__ .__ +el ow +-center ed +check s +_ Reg +-J avadoc +ĉ load +ĠLik ewise +ا Ùħ +UN E +.se m +x cb +ĠC ave +_s leep +Ġsil ently +ĠExt reme +.To Upper +ĉC HECK +Ġc ue +ĠQ ByteArray +Ġcorrupt ed +ĠD é +Ġimp ed +Get Name +Ġinaccur ate +Ġso ber +е е +Ġbar code +-- ){Ċ +ink i +Ġé p +Ġd ri +ĠAL T +>>>> >>>> +ont a +[ L +Ġinter es +ver ting +Ġdi agnostics +p dev +è © +ĠIntegr ated +). ' +_g c +$ text +.g ames +ĠT erra +' Re +.trans fer +_F IFO +get Model +Ġbl and +ĠCole man +Ġpr imes +Ġæ Ī +Ġcross es +n k +G ING +Ġ' ^ +ĠB lob +Ġinter course +ĠBl vd +Ġweigh s +_reg ular +ĠPer th +Ġsepar ating +Ġb illed +.tab Control +Ġpup pet +Ġutil ization +Ġâĸ ł +Ġsucc es +Ġl amps +_pro j +E ric +Ġren ovation +ĠFam ilies +ĠB its +part ials +-M en +s olution +Ġd warf +.IN TEGER +ĠLO CK +. ct +Ġexcer pt +ĠP ix +ĠFirst Name +ANT ED +ĠAd mir +-h elp +P rior +ĠAl ign +.IN STANCE +Line Edit +('/ : +Ġin et +od us +.p kl +ĠK Y +up ert +Ġn erves +_grad ient +} ',' +_un ref +Ġs aturated +ĠConn ected +ĠF N +EX IT +Ġtele port +Ġav ait +Page Route +Ġdivor ced +(l ang +f st +ĠT yr +Ġmess enger +if stream +X S +ĠBank ing +Ġinfect ious +ĠM ons +_LO OP +Ġzur ück +Ġobt ener +/re pos +V el +ac ro +Ġuser Repository +style Type +ĠS RC +VML INUX +rec ursive +/ bar +_ch ip +omin ated +ĠN it +âĢĶ to +ĠBudd h +ом еÑĢ +ĠM AG +ĠC HE +_d en +. raises +_de gree +Ġpump kin +_tem plates +_M EDIA +ĠTim eline +Ġb ots +Object Type +Ġbu ys +.post s +C AL +wait ing +ĠDani els +Ġd abei +ĠS igma +il or +ig el +, W +AD S +( panel +ì² ´ +it ating +.p alette +Ġmos quito +Ġt ego +(parse Int +Ġdes pués +p romise +Ġw ij +types cript +ĠT v +_IDENT IFIER +).ĊĊ Ċ +_fl at +its u +US R +ex perience +-f it +ph inx +_th resh +Ġide ally +ĠFre eman +, DB +_r w +çŃ ī +U b +_stat istics +=" ">< +Ġch ore +Ġy ork +inst alled +Add itionally +Ġp stmt +yl ko +:: Ċ +Fore st +Ġhead set +Ġgall on +ÑĢ ÐµÐ¼ +Ġwithdraw n +ĠC andidate +Ġmel ting +Ġfree zer +Ġh l +_HE LP +m ime +( /* +Ġth irst +$ return +member of +еР± +ĠHttp ServletRequest +( ob +_ Result +Ġassert ed +Ġfulfill ing +Ġstret ches +par ated +-f unded +Ġå Ľ +ing les +_c a +. condition +ĠDis plays +Ġor ang +ĠC RE +Ġgl Bind +ĠSelect or +/ type +ĠAlex a +ched ules +ĠPen insula +Ġpar ity +ĉ dest +ĠDo ors +čĊ ĉčĊ +_dim ension +Ġa load +.St oredProcedure +(p aren +ĠBur ke +') ]Ċ +- engine +Ġqu ir +ĠHy brid +ĠDo e +Ġout lines +ĠTrend s +_N V +per iments +ĠH in +? ', +ĉ Text +F UL +Ġsm ells +Ġs lick +Ġmis erable +ĠArray Adapter +Ġparam String +H om +_l iterals +us uarios +Ġprompt ing +_l azy +ĠActiv ation +_ oc +We ak +Ġan ecd +ĠU CLA += re +isse ment +ĠEsc orts +Ex cellent +ĠP ause +Ġre positories +T OR +ari ate +_is o +up dates +hal b +udi ante +ë¡ Ŀ +Ġna ive +ĠP eg +ĠL ounge +ARG IN +(b in +On ClickListener +ĠFA ILED +Ġl ite +Ġd zie +ĠL iteral +iv or +fc ntl +Ġe ats +Ġq ed +Un lock +rid ing +und ai += M +AT TER +Configure Await +ici as +ustom ed +Ġsuccess ion +end Time +ĠJ upiter +Ġjud ging +d ration +_d ocs +.m o +Ġeduc ators +ĠV ine +Con d +[ out +q b +\ Validator +Ġmean ings +Ġpresent ly +Ġdiv iding +otten ham +asc ular +Ġtrail ers +ĠC LOSE +ам и +âĢĻ ai +ĠG ain +w or +Ġpl anner +Ġdistrib uting +v at +month s +x label +H F +V iol +.BASE LINE +еÑĤ ÑģÑı +ĠR otate +Ġtx n +: bold +Ġb loss +Forg ery +( embed +Ġjak o +s printf +the ir +Ġexhib its +- static +he cy +get ActiveSheet +.c lients +ãģ į +_h ide +[ word +C b +add Item +ax e +_r adio +al ion +mod ifier +Ġsat uration +Ġden om +_p ixels +m ess +(f l +at if +Ġse cs +Ġpro stitution +Ġgrand children +Ġparad ise +ĠF eld +_B INARY +it ous +๠Ħ +Ġflash ing +-s ided +Ġcontrad iction +/* ĊĊ +y label +ĠT et +Ġadm ire +res o +Ġlet z +ĠSE ARCH +sl ots +ĠRew ards +ĠH og +ĠNS Data +st ash +F all +ĠA mer +Line arLayout +/ photos +Ġfe ather +Ġ| čĊ +Download s +.Start sWith +Ġ// # +ine Transform +Ġaff id +V tbl +ĠRog ue +scri bed +Ġfa uc +ĠMon roe +Ġdecl ares +mod ern +re on +ay be +P ASS +f ers +_MULT I +ĠMath ematics +Ġsud ah +_ATT ACH +Ġnumber With +ĠSol omon +j in +ograf ia +ö l +_d esign +cul ated +ĠL una +ies z +Ġ=> ' +Ġrevel ations +Al ong +( ed +ĠF ilename +Ġy label +Sec ure +Ġbus ca +agn osis +_RE CE +Ġoverl apping +Ext ent +Ġanticip ation +Check s +ĠALS O +or c +iling ual +it ational +Ġadv ancement +ou ro +ĠP redicate +å¾ Ĺ +er ia +ĠPier ce +or io +Ġmer its +Ġpe anut +.P ackage +ĠCon duct +_SENS OR +Ġbo iling +Ġin tra +ĠI GN +ĠF ur +.Ref resh +ĠRe ach +_dec oder +.Ex p +ĠÑĤ ак +p ill +, Q +ĠGr ill +Ġpop ping +.A g +Ġpro yecto +Ġmile age +Ġec ological +] ]);Ċ +ĠÂ Ń +sub plot +ac ad +ĠTry ing +rec ipes +$ criteria +ĠPers ian +-b ound +M ASK +ĠG esture +Ġk k +ĠP VC +Ġprohib ition +Ġcom ando +ĠLO OK +Sh opping +Ġdist ortion +< Boolean +.Get Length +um pt +\ Product +ell ery +Ġfire wall +form atted +.red is +Ġes a +ĠRh ode +S om +.n on +Ġ' ). +Ġget View +ạ n +pr us +Mat thew +Ġs ia +ĠF ors +G PU +ient ras +_IN ST +Ġol arak +Ġimport ing +T CP +/ ");Ċ +e ither +Ġfresh ly +c ascade +(char acter +ĠJe ep +ot ics +_ UTIL +.Xtra Printing +.first Child +ĠEx cell +Ġd vd +Ġt aller +Ġr as +yp ass +Ġassign s +Ġgri ev +-m ore +J D +ĠBurn s +' >čĊ +.D ependency +.Query String +.O wner +Ġexp iry +Th u +( Vec +Ġhazard ous +Ġr pm +AP ON +Ġadd Target +sv ille +p Net +ĠIm g +ĠTIM ER +.An imation +Ġbe k +Ġass ort +Ġle bih +Ġbody Parser +Ġvibr ating +ID L +Ġbutter knife +int ers +Ġpersu ade +ĠLGBT Q +è ĭ +.s oft +Ġbe ams +_s ur +.D ef +Ġl abs +ĉ plt +Ġsk ins +Ġtransf erring +Ġimag inary +_E nd +; background +Ġl aps +_COM MENT +(S DL +ond s +.Rec ord +ĠIm plements +_t icks +() ))ĊĊ +Ġa rose +] ? +ĠM p +ĠI Command +Ġsculpt ure +Ġcontract ed +< HTML +Ġcal end +at y +/ Sub +Ġkv inn +_ IGNORE +ĠSh ane +ML S +Ġstim ulate +Part ition +Ġm un +ó m +eral a +- account +.B inary +c é +Ġse ize +connection s +ĠĊ ĠĠĠĠĠĠĠĠĊ +ĠDi agnostic +V ISIBLE +ĠRun s +Ġimpress ions +s uite +ob le +~ - +ak ukan +< Person +ĠN os +ĠG ui +.wait For +RE SET +Ġpost pon +Dis cover +arr ison +sh aw +b lood +AJ OR +æĽ´ æĸ° +ĠM use +æĶ ¶ +Ġret aining +ot te +Ġmos que +ĠS ne +Ġstandard ized +Ġmain land +_th ree +unge ons +get Doctrine +Ġwh ale +Ġag g +ĠP orsche +now led +lat ent +ĠRel ation +Ġ// ' +Ġshut ting +ĠRem ix +_c ov +Ġs ailing +Ġv owed +Ġp ots +out u +Ġhair y +cast s +Rel oad +Ġre connect +ter a +.child Nodes +ĠR ack +Ġcurrent Index +Ġall en +Ġ ç͍æĪ· +ĠC ubs +[ X +_SE Q +_RE MOVE +.get Action +(/ ^ +err ar +Ġ ether +cur ve +Ġsl ap +Ġu om +O thers +Ġen gr +Dis position +Ġst aged +E ye +ĠA ux +auth enticate +Ġ$ ? +ĠAndre as +Ġset w +.A rt +Ġforecast s +Ġa unt +-m iddle +Ġmis d +des k +Ġescort e +ĠCas a +rop ical +Ġexem ple +plan et +(U INT +Ġwh ip +ĠPC B +clide an +=" \ +Ġox ide +Ġsucceed s +der ived +ĠEcon om +_co ordinates +ir as +D raft +Ġvisual ize +B rian +_ASS UME +ĠObject Id +Ġtrain ers +_FOR CE +Ġcon soles +- process +lic her +ĠSim mons +T aking +ĠCl aims +Ġdiffé rent +Activity Result +Ġsn s +éĢī æĭ +ĠCr us +Ġll am +r ab +ĠJo an +AA A +ĉf ilter +ish ops +get ting +à µ +Ġquant o +P ast +ov ich +Ġin justice +ĠF LOAT +Ġal right +\ DB +( GameObject +u ish +(b ot +Ġgall ons +ĠR é +ĠS aid +ĠSTDMETHOD CALLTYPE +ais ing +_process or +ell idos +ter dam +ĠBe am +Text Area +Ġret orno +.M ake +Ġ$ ("< +Ġlock down +Ġremed ies +Ġve el +x ee +do ctype +F il +ĠExp and +Ġemp loys +Ġsession Storage +Ph p +P ublish +Ġret al +f abs +ynam ics +Ġtoss ed +ĠnumberOfRows InSection +x path +\ modules +Ġdis astr +ĠM ULT +.M esh +-st age +Ġs df +it ung +ug es +Ġ?> ">' +kin son +Ġк ол +ogn itive +_ li +Ġim minent +Ġaff inity +.sign al +Ġnot ch +ĠSteel ers +max length +K K +ĠEug ene +_P WM +ro i +Ġâ Ĺı +ĠH amburg +.M ust +Ġax e +en ef +Ġamb itions +ĠSpec ies +ĠSt ress +Ġa while +Ġб Ñĥд +Ġwith stand +ĠDec oder +_in ventory +Ġ{ ččĊ +Ġt gt +Ġrail road +W ASHINGTON +Ġnegot iated +N ST +- phone +, U +Ġexerc ising +á» ¥ +_P IXEL +av ors +iter ated +Ġv ampire +ad al +In grese +Ġun g +ject ive +.c ells +Ġn ano +Ġmark down +_R ULE +(event s +Ġl uggage +MESS AGE +ig keit +$ count +Attribute Name +IG INAL +_E nt +ĠB F +ĠCOM MENT +_in i +ĠEurope ans +ĠB elle +åij ½ +) [' +åº Ķ +ĠUse ful +.re ference +() ", +_ grade +ĠK aw +Ġsent encing +Ġsocial ism +mon ster +_L AYER +Ġdee pest +w k +ĠNo ise +### ĊĊ +Ġpr éc +ot le +ÑĤ е +a uf +ib al +Ġcon quer +> Email +Ġamb ulance +O AD +Ġ(" % +ĠF I +.f ixture +Ġter se +ĠĠĠĠ ĉĉĉĉ +Ġsanct uary +ug i +ĠCom parator +Definition s +Ġast hma +Ġl act +Ġhard wood +.c lock +Ġattract ing +ĠM our +(d istance +ic its +Ġbon ne +ĠAC CESS +.Deserialize Object +ĠTyp ed +Ġje u +Ġapp Id +ĠCl ara +ĠH F +ĠRe ich +ipp les +//---------------------------------------------------------------- ---------------- +_del ivery +erial ization +Ġplaint iffs +Sc ient +sh opping +ĠD ummy +ĠW ald +Group Name +Ġins cription +el og +:::: :::: +_ ld +Back Pressed +.R aw +ĠOn Trigger +Ġmuse ums +ĠBe en +ĠAdvent ures +Ġsl ate +Ġlet t +Ġsu nd +ĠG in +ĠMechan ical +.s hip +App Component +Ġdest ined +Ġdw elling +Prof iler +Pre pare +ze ich +Ġsil icon +(h as +Ġ# % +VID EO +Ġcollabor ate +L in +Ġsc opes +( className +(s d +and in +.h am +Service Impl +-des cribed +Ġiron y +st ial +ĠHu awei +(re po +Ġunexpected ly +ĠK ai +.inst all +\x f +Ġexhib ited +_T CP +ĠO x +_CH O +Ġprostitu erte +Ġv ä +Ġsit o +Ġconstitu ents +ĠContin ued +ĠS AVE +r ss +/ message +ub es +Ġmisd emean +Ġtax ation +Ġstory line +h air +ĠFind s +S IG +ver ification +~ = +.h p +Iter able +Ñĭ е +ator i +Ġc tr +R x +_ );ĊĊ +d ag +.p in +Ġp seud +Ġinv o +ÑģÑĤ ÑĢ +_p ix +为 空 +Ġsw orn +âĢĶ or +_reg istry +Ġdis asters +ĠRO I +ĠâĢ ķ +akt u +fore st +be iten +âĢĶ I +ue va +eg t +Ġsp ikes +URE S +ĠRecomm ended +Ġexplo ited +ĠFreder ick +_COMP LETE +ĠDr ugs +!!!! !!!! +ĠR iv +ST OP +RO OM +ĠP ASSWORD +C ookies +.E l +á» Ń +ĠB ert +Ġhash ed +ic ester +Ġdecor ator +Ġquery String +: ;Ċ +Ġ" [" +oto pe +-A meric +ĠMatthew s +UR AL +âĢľ , +Sum mer +f os +_CONT AINER +_A CK +Ġfil tr +_dis p +_ Re +Ġfac ile +а ÑĪ +Ġìķ Ĭ +Ġe ben +Ġspr ink +ĠQ uint +> V +Ġhistor ians +our met +ĠMonitor ing +led ger +c ott +Ġw are +GG LE +c ars +ĠM EDIATEK +Ġvol upt +_ View +HE L +(c opy +(st ats +Ġchrom osome +ĠCurt is +- conf +( asset +Ġhv or +File System +< >();čĊ +oc oder +ĠC annon +) x +ĠSm ooth +ĠS AS +_ ce +ĉ prev +_m ovie +E c +_w all +< Button +ĠF AST +Ġon View +ul an +ĠS UPPORT +Ġgesch ichten +ĠS ons +Im m +$ IFn +Ġfair ness +Ġd pi +ats u +J osh +Equal ity +Ġ} ()Ċ +_ less +ĠR atio +ĠC ats +ĠS tern +Mon ster +Ġmer cury +ü hr +Ġplus ieurs +.des erialize +sc opy +.F alse +) animated +ĠExp erts +Ġ"") {Ċ +.W hen +see also +.un pack +LE M +.select All +Ġperception s +ud ing +ir ling +ĠPrint ing +gram s +ĠFile Stream +erv ille +il og +ic mp +_C ount +Ġlivest ock +- ca +doc uments +Ġpo les +ĉw ant +Ġflu ores +Ġstand point +ĠH uge +Ġradi ans +ĠUIB ar +EDI UM +ĠHistor ic +_h older +ĠMar ines +Ġt ä +.L ight +quir er +ason ry +div ider +ĠFl utter +_f b +restrict ed +ĠEvery body +N ão +Ġkn ot +ĠT witch +Ġhall way +(C ollider +Input Element +? )Ċ +/ off +/ ) +play ed +[ OF +Ġbat ting +_d l +Ġcom edian +Ġé v +ĠD EM +ĠEd en +: white +' ', +Con struction +acer b +Ġtask ed +.man age +Rel ationship +Ġph on +n z +_B GR +Validate AntiForgeryToken +_ air +âĢľ When +Ġgl fw +ĠCon versation +_T OTAL +, Z +Ġg raz +Ġiter able +ĠP ASS +Ġadvert ise +Ġmö glich +/ train +ĠVolk swagen +Ġcreep y +Ġ" )čĊ +QU ENCE +Ġalt ar +Ġed its +comp iled +aw ning +ĠD ungeon +Ġo sg +Navigation Bar +Ġtrend ing +ĠE co +ogg les +cd ot +| - +S ie +ec ret +ĠN egative +ĠL ing +ĠD IM +ĠC WE +ĠCar rier +Ġcar tridge +_us b += os +ĠJack ie +Ġo tras +Ġcommod ities +ĠP resentation +)&& ( +ĠMar tha +ĠCath olics +ĠM ond +об Ñĭ +_ absolute +Ġash amed +pons ors +t al +Ġsad ness +Ġpu ò +F ade +-pre view +ĠRequest s +ĠCal vin +h orn +Reuse Identifier +(pro vider +/app s +ime o +ĉ Class +S amsung +ĠW ORLD +Ġc innamon +dot env +ĠI User +ĠDE V +_C har +.ib atis +et i +/ me +s st +.s ym +ĠRug by +-m aster +aj ar +ĠY EAR +Ġo dp +ĠR oles +Ġbip artisan +ail le +Ġblock er +Ġgre ens +.SE CONDS +Ġbelie vers +ĠL ikes +F LOAT +Ġm ak +Ġg cc +âķIJ âķIJ +(" ~/ +SCRIPT OR +Ġton nes +ĠS ang +Ġtrans pose +enn ai +P red +Ġsoll te +.github usercontent +( print +ĠH ole +çľ ĭ +ad get +Ġprompt s +Ġgen etically +ĠH od +Ġvert ically +_control s +ÑģÑĤ ан +") {čĊ +$ title +Ġ} ),ĊĊ +Ġstate wide +ĠCor respond +ĠAt tr +it ant +Element Type +Ġout ward +Ġfam ilia +( article +Ġbl at +Âł Ċ +Ġgl Get +ĠRe ceiver +Ġ% - +ad am +W inner +Ġtail or +_p wd +ert en +St an +ĉ all +al ive +strt otime +� s +s essions +$ conn +ass ist +Ġchat ting +ĠM ant +Ġ% @ +Ġ"" );ĊĊ +Ġd gv +Ġíķ ¨ +.re peat +_M essage +Ġadvis ers +/ path +Ġk es +) } .ĊĊ +ogen esis +ĠOPTION S +upt ools +Ġmilit ant +Ġex ited +ig ar +ĠCOM M +ĠDis posable +ay cast +Ġrow span +Ġsyn thes +Ġsond ern +ĠĊ +ĠJ acket +R ATION +.getSelected Item +- init +ĠReg isters +_se p +ĠTool kit +.d ict +Ġx label +\ Table +t oc +_com bo +ĠComp act +Ġr ugged +à¥ĩ ठ+-man agement +')}} ">Ċ +ĠSt amp +ı l +ro x +Ġlandsc apes +_NOT E +mon ary +c ab +Ġmo et +x af +rc ode +- cli +_g ate +[ event +SP ORT +g ia +ĠS UPER +/ Login +_sh utdown +int errupt +Ġpret ending +Ġfr inge +ĠRed s +ĠC UDA +ĠUN IX +v it +Ġbr ig +dr v +ĠConn ector +There fore +Ġl ia +D etection +_ actor +Ġtemp file +Ġecc entric +- role +Ġpad x +d ent +West ern +Ġê ·¸ +ĠApplication Record +Ġcampaign ing +_run ner +ĠC ivic +ale igh +Ġdire kt +.s ul +ĠĠ ĉĉĉ +ant en +Ġiss uer +Ġassert ions +( orig +AT IO +Ġlean ed +ä s +.D TO +expl ode +.O bservable +Ġstagger ing +Ġkidn apped +Ġprogram mers +ĠInn ov +.param eter +Ġdom ination +Ġske ptic +Ġæĺ ¯ +Ġavoid s +.Ver ify +ub by +ĠAS N +Ġformat o +ĠBeat les +_b rand +Ġin set +y outu +Ġto c +-f inal +Show ing +ĠD oub +ĠM esa +Ad j +_m edium +Cre ates +(end point +ĉ UP +bb ie +Ġst alk +.datab ind +.S can +ag ents +$ , +ind ividual ++ )/ +ĉv m +(not ification +Ġin ex +ĠClass ification +ren o +Ġo lig +-r ated +Ġform ulation +', { +Ġa cept +_un pack +_C A +.P ow +ĉ im +Ġal uminium +AN O +Ġx n +Ġcó mo +ĠIng redient +Ġseiz ures +åħ ± +ific ador +Ġsigu iente +ĠIn fragistics +Ġduplic ated +ĠDe e +Ġn ø +ĠAC CEPT +(c rate +иÑĤ елÑĮ +- less +Ġinf inity +An alyzer +-D ay +rit t +(c in +ĠG y +Ġmulti plied +uch i +ĠBald win +/ ip +Ġshort cuts +.A DD +Ġvig or +_in struction +( ; +_ eta +è¿ ŀ +utor ials +Ġboost ing +b v +Ġacknowled ges +List ening +FA Q +; b +(( - +Ġarchitect s +Ġz we +Ġpul s +Ġget Count +ver bs +ãĢ ľ +(C ollection +k re +Ġjuris dictions +_b ridge +ĠCr ack +ĠDiff iculty +K O +Res ervation +_re quires +T our +ãģĹãģ Ł +.set Current +Ġk y +ĠAlb any +Ġè § +ll er +agn a +work ers +.bl ank +ĠPr ayer +M IC +Ġresil ience +Te X +ĠL anguages +st udy +ĉc urr +Ġenzym es +Sl ug +ĠíĮ Į +str al +Ġtum ors +Ġseg unda +=' { +in struction +ĠL isp +/ info +Ġ" {$ +,: ), +Ġg v +( ErrorMessage +Ġ' = +}- ${ +.Doc uments +" Well +Ġreminis cent +Ġg az +iro pr +eh r +Ġsup pressed +ers h +.scroll To +Ġcad ena +Ġgame State +ÃŃ m +( conv +ĠTom orrow +ĠC CT +M ongo +ul g +.C amera +.hand lers +m ph +Ġst k +Ġgen etics +AC ING +Tr ivia +ĠB am +(m arker +.St retch +ĠSun ni +ĠBet ty +.t olist +un likely +.Rect angle +ob solete +IL ON +inner Text +emb ourg +a N +ĠV ehicles +un lock +: utf +n ob +ĠSee ing +ĠNE VER +Ġt ls +Ġfil les +Ġbenef ited +ĠCl int +*/ ), +.f old +Ġpos ible +A DED +th ouse +.D AL +ĠO dd +ro kes +ĠSun ny +ĠPartial Eq +_B uffer +ĠLe vi +long rightarrow +eld on +g ages +_w arn +.Create Table +ĠD ip +_ questions +.log ic +Ġ# " +={() => +Ġt ep +Ġju icy +ì Ĥ¬ +en ko +ia lect +Ù ī +Ġon board +Ġæ ı +ĉ rt +_ UTF +ĠQ Action +âĢ ŀ +( Component +(a udio +.h it +g te +Ġprogram med +state Params +Ġpoly ester +f ires +by ss +] =( +_ quality +Of Day +ĠFair y +Ġy elled +op l +(user Name +ĠD ifference +Ġevalu ations +iff any +Ġcycl ists +Ġc idade +Ġtext book +Ġprof iling +__ ), +de a +. activate +Ġindic ations +Ð ķ +Touch UpInside +Ġinval uable +ĠM ASK +Ġcont end +F req +Ġrecru its +(int erval +ĠUser Profile +Ġ'./ ../ +ed u +_C allback +Ġanal ogy +ĠTro phy +app hire +V ideos +ĠCh er +ĠH av +â̦ " +. validator +g fx +ĠU Object +class names +tri angle +ĠEnc oder +.s py +Ġpred ators += status +-s afe +: ",Ċ +ĠIn cluding +Ġ{} ;čĊ +* cos +Ġend ured +.sul ake +Ġnurs ery +Ġfrag rance +Ġre building +Ġn th +ĠFr aser +.set Date +ĠV ince +_RE ST +Ġvent ilation +æµ · +cri bes +.as m +lp Vtbl +ĠA be +uis ine +, array +ĉ className +err als +Ġ' ĊĊ +Check out +Ġsol icit +A ux +_c apture +Ġrib s +rag on +vi ol +top ics +Function Flags +ĠM arty +b ike +ĠT ucker +(k ernel +ĠO ps +Close Operation +/d emo +ild a +ĠlÃŃ nea +APP ING +Ġsu ites +.visit VarInsn +ur us +ĠMin ute +(m anager +Ġbutter fly +Ġap are +Ġw olves +J WT +ĠSal on +ĉd elay +-es lint +is ations +.r pc +)| ( +ĠSnap chat +/m m +M N +cer ies +.text Alignment +ĠFrank furt +Ġad o +(new Value +( access +( Expression +ĠSign In +ĠHait i +_t p +.set Parameter +Min ute +Ġmanual s +ric anes +ĠP TR +ĠOut er +Ġget line +oc ations +_C D +ĠLy on +/g ui +_l ive +id an +.ge om +Ġborder Bottom +im uth +_check point +Ġme u +ĠIr ving +Ġpeu vent +(M AX +ĠAR CH +Ġp ov +.source forge +Ġjam ais +Ġar k +ĠBaghd ad +ĠC LEAR +Menu Bar +Ġtro is +CHED ULE +Ġ# čĊ +(C all +$ order +(M aterial +Ġencontr ado +$ list +ĠMETHOD S +.begin Transaction +_M AG +Style Sheet +Ġmaj ors +Ġindef initely +clean up +Ġhom eland +(d to +D ates +P resentation +ĠD K +={` / +ĉ Key +( Block +_check box +ne eds +Ġon Complete +ric o +Ġgle ich +Ġx m +O OD +B etter +ĠSQL ITE +. Book +x ad +ĠG one +ĉd p +Ġdev otion +Ġst m +Ġobs ess +ĠBack end +Qu eries +I k +// **************************************************************** +Ġdivid ends +.parent Element +} ")ĊĊ +ĠMaterial PageRoute +: num +Ġexp lic +ĠO L +le ast +O ops +iment os +Ġins urers +Ġhero ic +ĉf ields +.img ur +.btn Cancel +ĠDetect ive +(s m +ĠMutable LiveData +.l ab +(( [ +Ġha irst +ĠTrans actions +å¼Ģ å§ĭ +Ġstd Class +uent o +G IS +_c od +Instruction s +C alls +Pointer Type +ĠR w +Ġassort ment +ĠD IG ++ r +_C ERT +Ġinst ability +Ġv ib +on as +Ġro ku +ap ellido +Ġan gl +prene ur +Ġfluid s +ise ase +Ġde ed +qu ist +_CONST ANT +Ġequ ilibrium +_de legate +ĠQuant um +re i +Cap abilities +rect angle +? >< +al ien +ĠJ ug +D NA +T ickets +Occ urs +ĠHaw k +.setHorizontal Group +\ Collection +ff iti +Ġre arr +.setVertical Group +Ġc avity +Ġadult e +Fac ade +- wh +ĠL OL +Ø ° +Ġgrand parents +Sw ift +ĉw x +æīĢ æľī +if en +ff set +B eyond +// }ĊĊ +Ġw ager +Ġb ury +Ġcomm ence +reg istro +sc ient +ĠPer cent +Ġд олж +( identifier +.set Model +Ġs eldom +nt on +Ġappl iance +am us +rys ler +Ġpant ies +engu ins +Ġmim ic +Ġon Changed +Ġal coholic +.reload Data +Ch arge +ĠF ax +Ġj ScrollPane +Emp resa +Ġsh attered +x ba +Font s +? s +Ġpost season +ret ain +_r ates +Ġrequest Code +.t odo +´ s +CH K +ĠKeep ing +enge ance +Ġvs code +IPP ING +Default CloseOperation +_ raise +ĠO culus +ogram s +ra j +pc i +Ġcorros ion +.handle Submit +Access ible +ĠP iano +l ittle +AC L +Äĩ e +.un wrap +ĠCon vers +ĠLe ben +ione er +ĠMer chant +ĠJ orge +Ġembr acing +Ġvent a +á st +Ġvi ene +< QString +Ġexplos ions +Ġdistur bed +." < +m emo +ĠAb original +Ġcomple to +Tex Parameter +Ġuom ini +( agent +Ñĥ ÑĢ +ĠWh olesale +/ am +ĠBook mark +dr agon +Ġglo ve +Ġ" "));Ċ +iv ariate +now rap +In Children +.B r +Ġcon exion +Ġback bone +Ġe clipse +Ġpersec ution +': ĊĊ +/ link +ĠP ero +and as +ĠT ek +. "); +-an alysis +Ġer ad +Mar shal +Ġanch ors +og er +Ġconver gence +st icky +Ġnave g +int ern +_DE SCRIPTOR +ĠConsult ant +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ Ċ +ĠA uch +Ġer re +ÅĽ li +ĠHor izon +col a +Install ation +hot mail +C NN +.C ollectors +ch s +(tr ace +ĠEnc rypt +Ġ---- -- +ĠBase Controller +Ġag ua +Ġre active +id l +Ġclass Names +ĉ Session +ĠDod gers +H ad +_l v +Is Valid +ĠHEL P +ut to +ĠVer ification +Ġget env +_p a +.b mp +: f +ĠLou ise +(' ; +/ socket +Gr anted +.c alendar +( IP +ĠP X +.R oom +Ġprogram m +ens i +Ġtablesp oons +Ġle ve +Ġmo str +.t ipo +/ an +(d i +Ġb iod +Ġdb Context +ĠJS X +ĉ results +. END +ht e +l ify +P recision +èĬ Ĥ +ARS ER +)did ReceiveMemoryWarning +at tempt +IS P +& a +_P OP +ĠT ac +Ġprepared Statement +Ġзап иÑģ +Ġow ing +, start +Ġreview er +Ġr st +Ġprop Types +Ġrock y +_lo cale +ĠStrateg ies +ĠWe ber +.C ascade +_equal To +Ġcos as +ĠDe letes +ĠMax im +Ġsh rimp +re trieve +.In clude +IG IN +ĠO E +] );čĊčĊ +.en umer +Ġco ef +_N ull +R a +ty ard +ĠSh awn +keep ers +Ġq q +_s b +om ens +ĠExec utes +# " +TT Y +ĠValue Type +); */Ċ +ĠAbs olutely +ĠT ottenham +/ art +Ġbless ings +Ġswift ly +b uster +Ġa vid +COM M +, temp +Ġ} ?>Ċ +-g rowing +Ġdeep copy +A ck +egg ies +Ġ__ (" +Ġno ir +terror ism +Ġanth em +ag ency +_PACK AGE +ĠC losure +.reg istry +Ġmamm als +< L +U ICollectionView +ĠLED s +Ġvol ley +( Buffer +_N ATIVE +lib c +impl ode +Scroll Bar +ĠMar ion +.Con tracts +_A t +ĠWe instein +compare To +ĠH ose +en ity +.create Query +_r outer +Ġstim uli +Ġ++ ) +ĠCh amp +ĠBay ern +ass a +.v a +Ġdistrib utors +Ġfile private +Ġdepart ed +cc cc +@ click +ĠL unch +> L +Ġbl uetooth +.De ep +- standing +ác il +Ġro oft +ĠPath s +_iter ations +Invalid ArgumentException +.s pi +ĠUIAlert Action +uy e +sign in +.p riority +ĠEss ays +=' {$ +Ġè¿ ĶåĽŀ +_s igned +.p ersist +Ġred esign +To Lower +ĠNew man += start +ĠIsrael is +asis wa +Spe ech +Ġnum eros +hand lers +ĠW ong +Ġм еÑĤод +We ights +ĠGu jar +te il +ĠNon etheless +_E FFECT +Ġv ect +ĠO sc +Ġco ats +ĠW heat +Ġge ek +ĠPRO PERTY +w orm +_const ants +ĠB oulder +ĠP arm +co le +Ġdefault Center +ĠRou ge +: A +xc f +ĠVen ice +med ian +Ġred emption +F resh +Ġcos m +Ġfig ur +Ġref urb +CO PE +.c d +Ġch ords +ĠS gt +Å į +VP N +ĠS END +ain en +_account s +Ġtent h +Ġdiss olved +< App +ĠCover age +use State +é ro +.. < +Ġì £¼ +Ġdream ing +ĠFore cast +.C ursors +Ġvis as +/ script +_start ed +Ġga str +(P RO +]; // +.T ile +* sin +( Adapter +ĠSand ra +_S IG +ard ash +ĠO val +Ġdescri pcion +(s l +ĠDes criptor +Ġ` $ +/f ree +ĠKey words +Ġt udo +ion ale +(f ound +.x yz +ĠGeneration Type +_DISABLE D +( area +Ġel ites +Ġh ombre +(m essages +ĠR ac +Ġext ingu +ĠEst a +op o +. vel +mouse out +Ġconv olution +ĠHand ling +Ġceil ings +T ek +ĠAre as +.writer ow +< View +ĠCorn ell +_B IN +.in valid +'' 'čĊ +ie ż +_P osition +Ġk idding +PC ODE +Ġwatch er +lo x +Ġâ Ĺ +D ave +_all ow +Ġbis exual +Ġun ordered +ĠSch we +_se gments +Ġt earing +IN LINE +Ġund es +.g oods +.c am +ĠL W +ĉ where +Cal culator +-th reat +- alert +ĠSuz uki +ĠIP A +ĠAtt achment +AC CESS +(d type +O pp +_s ymbols +Ġdans ke +l age +or get +res olution +е Ñĩ +ĠQ Color +ĠBar rett +аÑĨи Ñı += \' +ĠNav Controller +/ ref +(c ountry +_H DR +Ġterse but +pet ition +Ġsu f +cred its +๠Į +x m +ĠDav ies +.re ddit +Ġw oven +ĠO bl +ĠK M +ĠConsider ing +ens ored +.per iod +Ġd dl +$ wp +Ġextrem ist +; \Ċ +Ġk im +al ers +Ġspan ning +Ġco herent +Ġconse gu +.text Label +.g eneral +_d ashboard +л ение +k ick +_P ID +ĠExt ensions +reg exp +ĠCl ause +_m ov +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠ +ĠR eward +ĠLEG O +A k +=-=- =-=- +ĉ parser +Ġon ze +éĢ Ģ +âĢĿ ãĢĤ +_b all +(r hs +Ġch orus +< count +as urable +Ġwirk lich +ĠEr in +ĠMS NBC +Ġet ter +ĠC ron +_F LOW +Ġ, čĊ +Ġcal idad +ĠFile Writer +ĉ stmt +( Byte +_p at +Ġte lescope +Ġgre ed +ĠT ort +(w rite +\ application +ĉRT LR +ĠConfiguration Manager +Un ix +End Time +In cludes +ĠHar vest +en berg +ĠAustral ians +Ġë ĵ +Ġr n +Ġreput able +Ġbl ending +UL ATION +ĠBrend an +d ad +Ġm ø +ĠW oo +_d c +U ne +Ġr ue +with in +ang ep +Ġp ouch +\" ", +ĠS ic +âĢĿ ), +aly ze +ĠG ef +c overs +Ġd bo +replace All +ĉ Logger +Try ing +[ state +-p iece +éĸ ĵ +beh avior +all ows +l rt +_p ython +ert ura +-c ountry +ĠT G +.UI Manager +b ens +ale x +ĠBre itbart +b ac +Ġpredict s +Ġg ab +Ġcard inal +.Time Unit +ĠVis itor +ĠM ing +Ġliv re +Ġparent Id +port un +Ġdimension al +ĠV est +en ic +à ³ +Ġ Ùĩ +ĠBL UE +Ġitem Count +Ġfe athers +ĉp stmt +ĠPol ar +{ // +und i +Ñĥ ж +z ar +Error Response +ì ĥģ +Rep resentation +* _ ++ ] +pre pend +Ġ' > +Ġlegitim acy +Ġo o +S linky +Ġnation als +. words +; p +tr ap +oman ip +Ġc ues +Ġgradu ating +Ġsem aphore +"] );ĊĊ +ace y +RE ET +Gr ab +ĠFel ix +( Id +_ne ighbors +Ġmeaning less +(d el +Ġj eder +ĠContent Values +.abs olute +/ cl +Ġx b +dat um +Ġtort ured +Ġrub bing +S cores +ĠðŁĺ ī +Ġav ons +Ġam sterdam +E OS +H al +Ġtrust worthy +# = +.EX TRA +Ġman o +is icing +-s upport +ĉc ursor +ĠSp o +aim assage +M ission +[] {" +Ġprint ers +G REEN +Ġt eg +Ġabdom inal +! ĊĊĊĊĊĊ +.Sh ort +аз в +ĠGift s +} ") +(b inding +x ce +âĢ ij +inf os +Form Data +Ġd art +Ġele ms +(in v +Y L +t in +GEN ER +á» ¯ +ĠT aken +uck le +: e +Ġspect ral +.b aidu +/ ');Ċ +Ġgre edy +es ion +,,,, ,,,, +Ġ/> ,Ċ +Internal ServerError +NSNotification Center +ĠA i +Ġsp it +Ġaug mented +Ġstandard UserDefaults +FIN ITY +R ace +: C +ĠRE CORD +ĠHigh light +Ġ' ` +Ġdef icits +Ġne i +Ġresearch ed +T a +Ġc opp +.Get HashCode +): čĊčĊ +On Click +ĠWell ington +Ġrev ival +æ¯ Ķ +éĹ ® +ĠN SS +Ġfor n +Ġint é +ĠKu wait +_fl ip +_ bo +_ \ +Ġocc urrences +ĠScient ists +S RC +og ens +igr ant +RE MOTE +ĠS ID +. opts +u ve +() ])Ċ +Ġlibert arian +ĠGl ide +les en +Ġform e +ow ania +Ġannoy ed +Def s +ĠExec utor +Ġcast s +.set Checked +ĠSh aring +.Serialize Object +Ġselect ors +_ OTHER +ë¯ ¸ +(s uper +( OS +_VER IFY +id unt +< header +Ġ/> ';Ċ +Ġvidé o +ĠNeg ro +ĠL ords +ĠT ours +Ġsoft ly +.re ceive +ĠE RC +Ġdata Set +Bad ge +ĉ Event +Ġper l +Ġ{} \ +(s entence +Or Update +Ġdim inish +P IN +(d raw +.To DateTime +.Equal To +(p in +-p encil +lu ent +ĠCall er +Ġplay ful +- '+ +x ca +sw ick +){ }Ċ +}: ${ +ĠM eth +.get Cell +.b reak +Ġy max +=' Ċ +ĠH iro +( TRUE +as urer +Ġcu er +U ber +. Operation +Ġol an +Ġthr illing +< Response +ĠF emin +Ġtravers al +Ġp oc +Ġset Status +decl ar +std afx +Ġaddict ive +ĠB tn +Ġexplos ives +ĠCook ing +ĠPl aint +Ġaccum ulator +ĠApp ointment +, password +ĠF AR +lu et +Further more +decl spec +_Static s +.D ictionary +"> '. +ĉ valid +" ", +In strument +> J +Ġno str +ĠR ift +_P ort +Ġvec es +[ [' +Ġrall ies +- series +Ġv v +. uc +Ġr tn +State Changed +( ins +ĠCl a +------------ Ċ +c us +ĠRel oad +//---------------------------------------------------------------- -------------------------------- +.se conds +_dest ination +Ġscrew ed +> c +Th ickness +Design er +Ġgr ids +n Äħ +( cookie +T rip +-M obile +Ġv oll +Ġgen ital +Ġconf isc +ĠConfeder ate +Ġweb View +Ġm ise +Ġcl er +(se lection +$ date +Ġshar pen +rag en +And Update +Ġrem ix +Ġh tons +R W +M PI +Ġretrie val +Ġric hest +.Dec ode +:init Components +ĠT Value +S aint +@ include +ĠPER SON +.se p +ĠLD AP +g ba +Ġgro ÃŁe +Ġreli ably +ĠD FS +.getItem Id +Ġprés ent +.get Token +Ġch inese +ĠMe al +Y OU +"> >ĊĊ +b ower +Ġsw apped +/ install +Ġs inks +etr ize +Ġdecl ines +ĉm ysql +ĠC String +ĠMotion Event +.L anguage +R oad +ÑĤ еÑĢ +asc imento +')) -> +. about +( editor +ĠR atings +in come +Å¡ e +.de queueReusableCell +ĠAust rian +Ġs ulla +ĠTrib unal +ĠDid n +ов аÑĢ +Ġins pections +B oss +Ġcock tails +Ġapolog ized +_sub plot +op al ++ =( +Ġreson ance +ib u +Ġë ¦¬ +rom a +res erve +pl s +ĠT ah +ax ies +OP LE +ĠDar ren +ĠZ ombie +_M ap +Ġ] )ĊĊ +ĠQ i +ĠS ail +Ġrestrict ive +Ġeros ion +- par +WH ITE +Ġold u +Ġap erture +Ġbit coins +text o +ĠCom cast +Ġtime less +en kins +Ġfeed er +/ tmp +res den ++' _ +.D estroy +Ġç ok +ĠD OCUMENT +.l ng +.tag Name +Ġk ullan +eg rate +Ġ(* . +ç¼ĸ è¾ij +Ġhand shake +s oc +_ geometry +ĠDam ascus +Min or +ĠK afka +ìĹ ¬ +Fl orida +_com pute +.ex pr +Ġpar alle +ĠD iaz +c ir +[ target +Ġj oking +Ġgl or +(set q +_hand lers +H ang +Ġf err +rim inal +ĉĠĠĠĠ ĉĉ +ent ies +def ines +-t ax +json p +ĠU PS +met ro +__ ;Ċ +ĠUg anda +])) :Ċ +_t d +x ae +l w +. OS +ĠLog ged +ac id +ĠMay o +as pect +Ġvag inal +Ġinitial izing +Ġster oids +f iction +G RE +g end +Ġli abilities +ĠL ets +M ech +( nc +( change +Ġconnect ors +: k +Ġt ast +! ");ĊĊ +th ings +ro phy +luet ooth +ĠSign Up +. ctrl +Ġthere in +ord a +. escape +ig ator +Ġpet rol +Ġspec imen +Ġdeb uted +- Pro +Ġcr ises +.add View +ëı Ļ +-d oor +Ġmon et +Ġmill is +Ġv ier +Internal Enumerator +Ġadmin s +ĠL air +z in +get Query +umb les +L IMIT +ĠV ig +_s ong +< Character +:: . +_h om +_b p +ĠSup ervisor +sub mission +ab ile +Ġno i +Or Create +Ġpe el +Ġon Start +Ġsent iments +veh icles +Ġclass rooms +Ġs zer +Ġb ending +Ġlong evity +Ġa cl +ĠAle ppo +ĠU M +ĠR icht +Ġmultip rocessing +DOM AIN +"," + +_Y EAR +Ġsc rape +Ġsol itary +Ġ"] ";Ċ +/ errors +ìŀ ¬ +ľ ëł¥ +b etter +ĉ number +ĠL F +ĠAc ross +Pub Med +\" " +ĠExcell ence +Ġus ando +ĠU IP +Activity Indicator +_V OID +Ġbre eds +ï½ ¥ +uest as +ĠTre asure +ustral ian +(f ace +ĠT ennis +ĉ Int +ĠHans en +ç µ +: I +Ġâľ Ķ +GR AY +O USE +Ġhe pat +ł í +A IR +ó ż +Ġque ued +vinc ia +ĠChrom ium +Ġcompet ence +ung al +ill i +Ġget By +ĠF inder +Ġincap able +Ġs add +Ġc ites +ĠChurch ill +S dk +More over +As pNet +( Float +$ password +ĠConn or +-s ession +_d m +* )) +Ġde utsch +ĠN X +Ġper ks +_S ORT +_TO OL +_V ISIBLE +.as p +æĪ ĸ +ĠBre ath +D etect +ĠD uel +.c mb +[ it +.Set Bool +Ġnarc iss +Ġab ide +Ġej emplo +ĠâĦ ķ +Ġm ornings +Ġcomput es +.s sl +j t +Ġmuch os +_S S +[ end +Ġbas in +Ġalgun os +ĠCroat ia +lin ewidth +(t ags +(h idden +ÃŃc io +Ġap ar +ĠÐ ¶ +ä¸ İ +. food +ĠR ural +Ġbread th +å½ ± +(s ess ++ ") +ĠP aste +Ġserv idor +ĠBit Set +ĠTr an +la us +v ette +ey es +ĠCL ICK +ĠV III +ĠTurn s +ĠLe Bron +ĠM uj +ĠD eg +ĠAdult s +_s uite +process able +ĠPH Y +g hest +.F ail +ĠSl ack +ce j +\ Carbon +Ġsuper star +Ġhold ings +( forms +Ġ'# ' +M ultip +("[ % +-s olid +/ url +-t ier +[ length +ĠStream Writer +ĠMarket place +get text +_T ICK +ĠFor ge +Ġblack jack +ĠDO ES +ĠM atters +w aves +Ġwhisper ed +Ġl ush +ìĺ ¤ +d igital +Ġwr ink +ĠH ogan +Ġrust ic +.Apply Resources +ĠHard y +os omes +A UT +.ST ATE +Ġnarr atives +ĉ store +b ib +ĉ Scanner +ĠC ody +\ Repositories +Ġre union +and um +âĢĻ h +Ġsn iff +NS Bundle +Ġcompreh end +_US AGE +_ occ +URRE NCY +J NI +Ġspecial izing +Ġvis ions +Ġdol ore +Ġv á +ĠChe vy +ĠSt yled +imp act +all en +Ġk art +ĠTable t +st uff +re esome +аÑĤ оÑĢ +//---------------------------------------------------------------- -----------Ċ +_Ad min +Ġcell phone +Ġaut oplay +Ġcamb io +Ġmar itime +_BO OT +- quarter +Ġlat ina +ĠAJ AX +e quiv +ĠFront ier +ĠX Y +} ]Ċ +ĠR ough +.pro to +Ġcorrect ness +Ġfac il +ĠRe ached +ãģĿ ãģ® +V IS +.p s +Ġstr ncpy +Ġdiff usion +.start Activity +�� � +Ġaccom p +AMES PACE +imon ials +ĠBl ast +aby rin +Ġd ome +Ġextr av +Ġy en +Ġcul inary +P RI +ĠComm unities +n id +_oper ations +.h s +ĠMil ton +Ġno ises +Autoresizing Mask +(c id +}ĊĊ ĊĊĊĊ +] },Ċ +ĠD etection +tab la +Ġlib erties +_D YNAMIC +w get +ĠT ür +ĠP ascal +Trans parent +Delay ed +] () +ĠHer bert +< ActionResult +ch allenge +Ġmush room +.insert Before +ĠR in +Ġhum our +Ġf ø +api Key +alloc ated +Ġconf ession +. ",čĊ +ĉassert That +ĠS ORT +ĠL ORD +Ġexport er +.set Level +p okemon +ash tra +Ġf é +ur ator +(M SG +Ġt up +ĠH ull +Ġyield ed +.Sub ject +\ Route +! ? +ĠÑĥ дал +\ Security +- ar +Ġalleg ation +( Settings +ä nder +Ġell ipse +ĠRetro fit +Ġregul ating +ĠM olly +ĠL ok +_C ustom +ĠProm o +is in +Ġres umed +Ġmet ropolitan +.error Message +: ------------- +Ġpas ado +th ank +_De lete +ĠBright on +, unsigned +ä½ľ èĢħ +Ġaspir ations +-h ow +R ose += (( +_ne eded +_pl ural +< Application +ĠW EEK +ĠUn lock +ĠT EMP +S ou +Ġschizophren ia +Ġt roll +Ġcomplement ary +ĠNET WORK +Ġbl ir +Ġprogress Dialog +" %( +ĠAttribute Set +ĉ ts +.iter items +è¯ Ŀ +Ġesc rit +v ous +_pl aces +H K +Ġseg uir +_f w +ĠR ounded +Ġdis posit +è§ Ĩ +par m +w ow +STRU CTION +. allow +ĠChar Sequence +ĉ extern +Ġprosec uted +Ġmort ar +ĠJ uda +- msg +Ġest ud +.get Description +Ġs ow +amb re +Ġrom a +En h +bon us +Ġsqu at +Ġdist ra +ed Image +Ġpe ppers +-per formance +, ĊĊĊ +, file +ĠM IME +_con cat +AB S +-f ashion +Ġunder cover +One ToMany +Ġre claim +C OPY +Ġb inds +ĠT ape +Ġg ossip +ĠEqu ity +/ Card +. activ +' am +Ġdrain age +< Scalars +ĠonBind ViewHolder +() ?. +Ġs orrow +ĠI b +up y +_U UID +ĠCh arm +ĠElection s +.on Destroy +ĠInterest ingly +ounding Box +_d etection +-h eld +_ unknown +Ġrefr ain +Ġmét odo +Ġe Book +EN OMEM +Ġd ang +Prof essional +Ġd ictionaries +/m ysql +ĠST UD +Ġmas se +s cape +Ġdre i +: name +.log o +Sign Up +Ġt ahun +( theme +ĠFem me +Ġbom ber +ĠJ ade +ĠT ay +Ġsubmar ine +_cl ause +zy ch +Ġsimult aneous +Ġcas os +. boolean +(l hs +Ġcontin ental +-s ale +ĉ env +ĠC ute +ĠFactory Girl +ab us +/ value +Ġj adx +Ġst ern +> >ĊĊ +Ġsurf aced +Ġìł Ģìŀ¥ +pl atz +ĉ email +cept ors +"> ( +Ġep ile +è¯ » +ĠDe bt +åij Ĭ +N OP +" https +: j +Form Item +_L ICENSE +.get Double +ĠAg enda +ĉf inally +(f ilters +( av +ç¾ İ +AP ER +Ġl ava +еÑĢ Ð¶ +)) ))ĊĊ +Ġfault y +_n m +Ġtr ava +(B itmap +Ġspeed ing +> '). +Ġscreen ed +_ roll +ĠMac Book +ĠA UD +Ġdiagn ose +.G enerate +Ġ^ ^ +Ġstr s +[ Test +Ġr ansom +ĠDH CP +eld en +Ġinterpret ations +() ]. +flat Map +Ġline Height +_m ount +ĠW izards +Ġsl uts +eh ler +od al +Ġmilit ia +å ² +earn ed +Ġmis ery +int val +f und +Ġh ides +Ġdi arr +ĠWes ley +Ġx mm +Ġqu em +ĠAr abs +if th +ategor ized +Dis posable +P ure +_NOT IFY +sn ippet +ĠGar rett +.run ning +. weights +Ġ( -- +Ġin variant +äºĭ ä»¶ +ĠAll owed +dir s +Ġpass ions +Ġl ad +ĠFl ush +men us +: block +Ġcompr a +.ch omp +alloc ator +Ġcur ated +ĠKnow ing +ĠPatt erson +Ġtel ah +' ex +Ġdo omed +Ġphil anth +ott y +.st yles +Own ed +Ġallerg ies += params +oc ese +it elist +ĠS ending +b ef +orr ar +ĠN ão +ĠF argo +ĠL ub +ĠComb ined +_g iven +ĉĉĉĉĉ ĠĠĠĠ +Ġreconc iliation +Pattern s +az ard +Ġbiom ass +ĠH ouses +resp uesta +cc o +/top ics +ĠY uk +Ġweaken ed +_c alendar +Ġmulher es +ĠMar l +Ġs ine +ĠT il +ĠSou ls +ĠDe utsche +ĠF OLLOW +Ġpip elines +ĠBever ly +_DIP SETTING +" # +ĠPro to +.b ig +ĠSav ings +ĠT anz +j un +ĠG amma +ĠS add +Ġadvis ors +Ġro ast +Ġun ters +ud ies +_l on +-point er +ĠElement Ref +\ Builder +example Input +.web driver +data Type +ĠQu ite +ĠCelt ics +u il +-def ense +b ish +ĠUI Window +ĠS uddenly +.h ot +.re ason +Ġg ör +AM D +.M ulti +auth enticated +reg ions +; ( +а ÑĢам +ĠKir by +$ route +PREC ATED +ĠDur ham +ow o +ĠPer forms +Ġdisreg ard +n st +ĠP ols +Ġget P +"] : +-col ored +( Keys +ĠAl leg +_mod ify +_ loading +str ained +Ġat roc +_p hr +< Sprite +Ġsatisf actory +m anship +.p ipeline +T ony +Ġth ief +pol ator +( lock +bur st +ĠOptim ization +Ġsurf ing +" Yes +Ġdesc ended +æ Ĵ +_C lear +Ġc ries +ĠFro zen +D IRECT +- Con +ĠLe icester +å¥ ³ +O OM += db +Ġget Message +< Student +_b atches +.M ask +_ eth +\ ) +Ġsom a +C atch +[ ch +Own ers +ind le +: auto +. vert +iv r +.set Location +Ġfl uent +_END IAN +ĠCar lo +cept s +add Action +.o auth +< UnityEngine +re ements +.S kip +? )ĊĊ +.default Props +Ġc abe +ĠSh en +eros is +ĠPro fit +Ġpo is +_C REATED +Ġremove From +(w s +? action +( Field +Ġerr one +.min imum +ĠRetrie ved +Ġd ado +ĠPR IVATE +-s pec +Ġg zip +p data +Ġpos Y +(l ow +Ġqual quer +/ cloud +ê² Į +( common +ĠAr beit +organ isation +Ġtid y +ĠRol and +( ph +.z one +Ġgent lemen +ượ c +å± ± +Ġenc losure +ĠMan afort +ĉ Color +St encil +N ic +Ġthe orem +ĠV G +Ġcol oured +V BoxLayout +uls ive +Drag on +c ff +et est +ens a +of day +.A zure +:UIControlEvent TouchUpInside +_up dates +Ġtrend y +ug as +weak Self +Ġr idge +ib ri +Ġì¶ Ķ +(C G +ĠMon key +.write Int +.tim edelta +ViewController Animated +ĠProvid ence +ãģ Ī +Ġbl ends +/Sub threshold +ĠAp pl +Ġat an +Ġreload Data +umb otron +st üt +O Auth +ĠG iving +ĠìĦ ¤ +ĠFinn ish +check ing +. Embed +sequ elize +Ġinitial izes +ĠOs lo +Ø ¶ +get Extension +_AL T +(bl ank +Ġfatal Error +Ġdem ise +**** *Ċ +ĠX S +(A F +ĠEn s +an tha +ĠP OR +Ġn ich +.N amed +Ġgig antic +ĠObserv atory +.Res olve +ĠPay ments +g uild +Ġcurrent State +============ ===Ċ +ĠS ey +p Data +Ġdead lines +Ġcentral ized +ĠScholar ship +_s upported +.ch rome +() ]);Ċ +Ġc yan +ĠC age +Auth ors +_ čĊ +/ os +k im +de e +.t ex +Ġyours elves +Ġm gr +Ġal k +-inst all +Ġdraft ing +Ġrum or +Ġstat ues +Pool ing +ol ina +AAAA AAAA +/* ---------------------------------------------------------------------------- +Ġextrem ists +Cal cul +ighth ouse +In set +(IN PUT +Ġsynchron ization +iv irus +. axes +ĠG ap +- An +_T emplate +Ġgam er +ĠCr icket +Ġl int +Ġauthor itarian +NS UInteger +Ġred o +Ġadip iscing +_F ETCH +che id +ĠF ang +. indices +t one +д ел +Ġ{{-- < +bra him +Ġsal a +get Code +Ġcommunic ated +start sWith +ert z +Read able +Item Id +oref errer +cred ible +á ria +Ġcombine Reducers +** /ĊĊ +Ġbl iss +Ġad orn +dep ends +ĠRO OM +Ġfr aming +Ġ? ', +aut y +_p ot +_t abs +Ex act +, ", +Ġ'} ';Ċ +Ġarbit r +ahr ain +.getString Extra +Ġ$ \ +Ġoutput Stream +Ġcomm enc +an us +ch y +< Employee +Ġhex atrigesimal +Ġn acional +(serial izers +_put char +_S AFE +ential Action +ItemSelected Listener +.Dis patch +Conf lict +_ about +os aur +Bound ary +Ġclear Color +( Location +ĠMON TH +ĠT aste +- General +ĠW AR +Ġer halten +-s aving +Ġcou pling +-tr igger +m otor +Ġy yyy +ĠPat ent +pt o +Ġmisdemean or +vas ion +ĠAdmir al +à¹ī า +_P WR +Ġdevast ated +fol ios +ITU DE +urre ct +Ġrobot ic +ĠSan ct +ĠHawai ian +.R oute +- condition +Ġr k +/**************************************************************************** Ċ +create Element +ĠK op +ign ant +. rollback +Ġsal ud +_ ', +ĠAN SI +Ex cept +ĠDraw able +.Utc Now +":[ {Ċ +Ġk ole +L ua +ĠBel ieve +Com put +Ġhall uc +ĠSign s +r st +.h u +ĠKN OW +W i +ĠBr ass +ĠR as +@ hotmail +Ġsed iment +Ġap k +Ġì ĥģ +_reg ions +Ġpod ium +< Book +ж е +Ġsix teen +ĠAli as +Ġinfr ared +ĠV ander +ĠLe ading +uc ing +,: ,: +_h or +w at +Ġdé cou +_W idget +S ounds +_n avigation +Ġschn ell +(g enerator +uc ene +Ġrem ake +IP v +Ġré al +_IN CREMENT +Ġhypoth etical +_ ang +Ġof s +Ġ! Ċ +.com pleted +Get Type +Ġkom men +ál ido +add On +Ġz ÅĤ +UL A +_ind icator +'] ĊĊĊ +ap ache +_S elect +ĠGre ene +Wh ats +_an im +Ġrepet itive +m uch +ĠTh reshold +Ġl f +(C ategory +con e +M ix +_MET ADATA +ays ia +Ne ighbors +ĉĊ ĉĉĊ +IP HER +ĠFr ag +ĠC ells +Ġnames paces +( back +ĠRest aurants +sv c +Ġл и +ote ch +-s l +¥ ¿ +ĠW T +ĠRed uction +Ġd otted +ĉf ound +ĠTE AM +B orn +ĠM ush +ĠCompar able +Ġh itch +AT O +Ġmax Height +begin Transaction +ÃŃ v +_b n +Ġher d +Ġrevers al +ĠH ond +del imiter +Ġconf use +Ġh ops +Ġcent roid +Ġcourt room +.decor ators +Ġm pi +ĠImpro ved +IN NER +ĠBang alore +ĠT amb +Ġbo ast +() ))čĊ +Ġil licit +ĠMor occo +greg ator +_res ume +Ġcrack down +Ġport raits +/h igh +( \' +Ġay ud +_fe edback +Ġc ate +/ avatar +Ġhe b +Point Cloud +Ġå ĴĮ +Ġ< ![ +Ġget Resources +} :{ +Oper ating +ĠF og +ĉt ab +ĠResearch ers +Ġfabric ation +.datas ets +ĠCamp o +ĠKa uf +Ġd ll +lig t +] ));ĊĊ +st ellen +ACK ET +l vl +ĠGl ory +.date Time +Ġcomm ute +ĠonCreate ViewHolder +ĠX Element +ĠT okens +< thead +_p ick +ì ¤ +v on +depart ure +(render er +phone Number +(P erson +gen es +ĠL ars +Ġ) {ĊĊ +ĠJson Result +Ġmet odo +VO KE +.get UserId +Acc eler +ĉ required +Ġchampionship s +Build Context +/t ask +/re leases +C ategoria +_over lay +Ġscar ce +_l im +n gr +ah len +ĠArt ificial +sp read +Ġbow ling +.an alysis +SM TP +ĉp assword +Ġbath s +] )){Ċ +current ly +ac iente +_se parator +Ġde ber +ĠDis abled +i ères +Ġâ ķ +_process ing +Ġprotest ing +ĠR OT +gr ab +Ġз ак +Ġpro active +word press +ĠSe ver +ind en +Ġw ikipedia +){ čĊčĊ +_w indows +is lation +Ġun rest +Ġdismiss al +.N UM +_F AST +iss ued +ĠF ACE +_u nder +Ġpl ugged +Ġå ° +ĠbÄĻd zie +ĠI CC +Ġcombust ion +Ġkiss ed +Ġstar red +ĠW atts +Ġspi elen +-p urpose +ĠE val +arg es +, result +techn ology +Ġnational ity +ic us +ĠN ug +ĠÑĤ о +ĉĉĉĉĉĉĉ ĠĠ +col o +Ġg astro +ante ed +OL ID +.b ias +_t ele +.ins pect +Ġve il +. footer +Ġneglig ence +Ġjud gments +Room s +yn n +ĉcount er +occup ation +Ġ çĶŁ +un as +Ġ(^ )( +L ambda +f el +.Param s +Ġд обав +set Layout +Ġdeport ation +Ġlocal Object +ĠPharm aceutical +cept ive +ĠN ome +Equ ipment +F an +Un iversal +ĉ socket +Ġgr in +Ġex poses +Ġhab er +Ġsincer ely +Ġc ams +Ġm ü +en ia +E mer +C rypto +Sl ow +(x hr +! =( +-s ervices +ĠP W +Ġprend re +Ġm ädchen +em ons +озв ÑĢаÑī +.M anager +ì Ļ +Ġg raf +- ra +met rical +/ fl +Ġc emetery +g ens +Ġp ÅĻ +ĠMySql Command +- To +Ġv Ã¥ +Ġa irst +oment um +Ġserv o +m illion +ĠMir anda +" She +Ġadvoc ating +-c aption +ĠAt tribution +Ġwel che +_v endor +ĉ Status +arr is +Ġprint k +"," # +Ġrel ativ +if ferences +izz es +Ġdec imals +ĠPro v +.max imum +Ar n +Ġhelicopt ers +_B OTTOM +ch ure +od ings +' ( +")) );čĊ +( bean +.f d +F und +Ġhang s +app id +/k ernel +.p oi +.Min Value +- validation +L uke +c df +ĠFun eral +ĠS amples +ĉ de +Ġto astr +Ġtax able +Ġcl ustering +Ġ'\ ' +Ġre straint +ec ed +ch ains +ãĢĤ ï¼Ī +_GR APH +Ġfue led +éľ Ģ +H p +å¤ į +T iles +Ġa unque +J C +Ġhost age +ĠE sk +Ġm av +Ġgest ion +Ġb anners +} {$ +.int Value +.' "ĊĊ +_M ATRIX +Ġce ased +ĠG OD +_CAM ERA +.Allow User +tr acked +C ook +b airro +( company +Ġview point +.get Writer +ĠN ets +w ives +Ġ( ))Ċ +example Modal +ĉ child +Ġmyth ology +Ġ// " +_ axes +ib old +.D ark +ĠMax well +Ġg pointer +olic itud +B at +ul ner +bal anced +mail er +Ġcont empor +æīĭ æľº +(" __ +Ġ" )" +re ar +ĠHu ang +] ')Ċ +× © +FT A +ĠCalling Convention +ĠOutput s +P k +.Re ference +lect ual +Ġ) :ĊĊ +Ġbrace let +ug er +ĉ Error +S weet +("/ ");Ċ +h x +Ġun reasonable +Inter preter +Ġlo ft +_product o +Ġsoci etal +.P arser +ĠAd apt +. foo +( where +.F eature +ĠYam aha +g lass +For ge +Ġprohib its +Ġcapac ities +Ġíķ¨ ìĪĺ +Ġper mutation +Ġih m +F ld +el ial +======== ===Ċ +@ Configuration +Ġge ared +ios o +iest a +trans lations +Input Change +Pop ular +ĠPL US +Ġv f +_F ree +b box +Ġcaus al +PI LE +Ġsch ö +Ġiron ic +M ir +. @ +åį Ĺ +Ġè ĩ +R ew +ul ence +fl en +Ġcan Activate +- response +Ġacc ents +ign ored +° F +.Dependency Injection +ĉ point +Ġconting ent +Ġsqu ash +Ġpar ms +ĠC emetery +Ġdelta Time +ĠD OS +Ġvan ished +аÑĢам еÑĤ +ĠD PS +t foot +ĠZ us +_IN STALL +G AN +Ġar b +Ġmunicipal ities +Into Constraints +AutoresizingMask IntoConstraints +, image +_ ignore +Ġdanger ously +quis a +pl uck +Ġhar us +up pe +Http Exception +Br acket +.' 'ĊĊ +ĠT ol +ĠView er +zb ollah +.Code Analysis +ì nh +Ġcorrect amente +.d a +ĠAl ger +× IJ +ba um +ĠPan ther +part icipant +å¿ ħ +-s up +Ġem ulator +Ġf ading +ĠW olver +cre ates +Ġbook ings +.Q uestion +§ è¡Į +Ġstress es +Ġre written +.PI PE +ed es +Ġc bd +": "/ +Ġenh ancements +_s y +B IN +ĠSl ip +Ins pect +ĠW eg +Ġcon gregation +Ġ_ : +_r m +Frame buffer +Ġ'& # +ĠFall out +Is Required +ĠPear son +ĠF ACT +Ġrel ie +ĉ box +ĠShe pherd +ĠWiki Leaks +ĠCollect or +Ġres ized +method Name +Ġevent Type +ĠA then +Des criptors +Ġb ers +- oper +ĠInitial ly +å ¡ +_B TN +ĠĠĠĠĠĠĠĠĠ čĊ +á b +_c ampaign +_w atch +F ord +-date picker +Ġvis c +Ġsat u +_s ms +Ġcont ador +-s vg +ĠDO I +$ args +Ġkn ob +.B OLD +Ġdeb ated +img s +sock opt +tr uth +ĠFe es +Ġh Wnd +_f ood +Ġab ras +Ġnot ions +ĠT od +: create +ĠConf lict +Us uarios +OT OS +Ġm sm +K HTML +([ ( +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠ +Ġ} ] +w izard +Ġm ientras +Ġdata List +Ġemerg es +Äĥ ng +.Read Int +PG A +ILL ISE +I Enumerator +(t uple +Christ mas +Look AndFeel +og enerated +Ġ# ĊĊ +control led +Ġex quisite +Ġa cest +Read Write +G ain +ãĢį ãĢĮ +Ġcopyright ed +Ġdo om +.Table LayoutPanel +ĠD ort +Ġch ili +Ġwer k +ĠEVENT S +ĠBe acon +Ġship ments +Ġse bagai +up on +ut om +.con verter +.Drop Table +={ }Ċ +f ic +~ ĊĊ +Ġlesb ians +_n a +Fore ign +ĉ then +/ ms +Ġor i +get Property +ĉsn printf +hes ion +ãģ ¤ +"} ," +Ġac rylic +P ers +@ Enable +I sl +(C ard +. Stack +L icensed +_G UID +: title +Ġh ust +Ġprincipal Table +an itize +/ embed +Ġens ured +ĠE GL +ÙĪ Ø± +ĠåĪ Ĩ +/ ,Ċ +Ġfundra iser +Key Name +Ġmarch ed +_VAL UES +ĠSc enario +Ġmet ic +_ass oci +ĠPast or +ĉĉĉĉĉĉĉĉ ĉĉĉĉĉĉĉĉĉĉ +er ate +Ġinv itations +quo ise +Ġbl aming +Ġd aring +UM MY +Ġrich er +em aker +ĠIdent ification +ĠìĿ ¸ +ĠBinding Flags +ch as +Ġresil ient +_p g +Ġre leg +ĠI RA +ST E +Ġtr actor +- loading +ĠPre viously +ĠV acc +/ be +Ġn Ã¥r +Ġurl encode +ĠNor folk +.Re lease +ĠNe utral +ä¸Ń åĽ½ +ĠAr lington +Ġalleg es +ĠW riters +Test er +ĠR ally +Ġc á +ĉ Print +Ġâĩ Ĵ +ĠUser Controller +ĠSeek ing +.V AL +List Node +_ ff +ĠPhill ip +FA CT +Ġc aramel +ĠM ultip +ĠCom pared +ĠSer bia +Ł ³ +Ġrev ive +ĠK anye +Ġver ge +ĠBulg aria +get Body +Ġ| > +ce ph +.DateTime Picker +." ;ĊĊ +ĠT ie +, item +Ġm enn +G as +och a +_v irtual +Ġmaster piece +_se quences +L TE +ĠSub mission +Call er +$ \ +S port +ag us +Constraint Maker +Ġcol oc +Ġw ig +ĠÐ £ +ĉ Array +Look s +ĠGT A +.st eps +atch ewan +_r anges +ext Alignment +ĠBren nan +Ġab straction +uler Angles +.m isc +Ġantib odies +Ġexponent ial +ĠCH ANNEL +exp ense +' y +Ġdetect ives +Ġpur ported +Y STEM +Ġradio active +ĠLat ina +.Enc oding +.T AG +x in +D egree +ur acion +pr ices +ĠRefer entialAction +Ġr arity +Ġp iles +g ende +_project s +_g lobals +.start Time +Ġê µ¬ +SE CTION +_p ublish +F ault +DD L +_p rior +M om +Ġth icker +Ġsequ elize +Ġessential s +str as +in tr +>( () +.man agement +e il +éĹ Ń +A ware +.C ity +ĠAr bit +_D M +_key board +L Object +- webpack +ĠNew port +Ġprincipal Column +leg ant +Ġp allet +Ġfract ure +Ġg mail +.M eta +A bove +.Key Event +j it +_mac ro +_P USH +á» © +/ controller +åĬł è½½ +Ġsuperf icial +exter ity +Ġmens agem +W ind +ist on +.open api +и ÑĢов +ĠSerial izer +uct ive +Ġz ar +Pl aces +.St atic +B a +Ġin advert +ĠIndones ian +_IP V +(h orizontal +Ġget Title +ide press +ĠConsole Color +ip ers +$ out +Ġfest ive +Ġeven ings +.Get Data +uit ka +ĠManual s +uss ed +_M ax +.Ch at +ĠA ircraft += com +FO UND +ap ro +Ġtre asures +_al ive +Ġgad get +ek ing +Button Down +B rowsable +.PER MISSION +P ASSWORD +ĠH ASH +f é +\ TestCase +LO SS +o thers +, J +Ġassh ole +wer k +Ġm ã +. ie +ev il +kont akte +//////////////////////////////////////////////////////////////////////////////// Ċ += sys +ĉ lock +-- ;ĊĊ +_F UN +Fill Color +ó a +pre nd +Ġcompress or +M other +ĠAr cher +.g oto +Ġwür de +Ġbam boo +ï¼ İ +ĠT rees +Ġb umper +Ġsa usage +ĠEl asticsearch +Ġhor izontally +ĠG ul +Im mutable +Ġlos er +Ġabort ed +-d emo +ĠH atch +Ġund e +Ġprocess o +-c all +In come +å ĥ +_ returns +']." ' +(s w +C BS +am ilies +ĠYour self +ĠH olt +.M ON +à§ ĩ +ÑĪ Ðµ +an on +ĠFont Awesome +produ cer +j r +Ġm au +ĉint er +Ġdish onest +Ġmagn a +ĠCollect ive +Ġvra iment +Ġcho ix +st ay +Ġweld ing +r ising +, min +ĠF ate +g lob +RGB A +Ġdet te +V en +Ġembarrass ment +.DE LETE +greg ar +-re nder +(b ucket +"> ĊĊĊ +.wait Key +Bus y +Ġdifferent iation +ĠC ST +.Con stant +Ġline Number +(m atches +Ġweb socket +Ġbar red +Ġpued es +M ono +C ORE +I ID +ĠĠĠĠ čĊčĊ +Ġpúb lico +lean ing +Ġcleans ing +Ġcr is +ĠDev ils +_SET TING +unt ary +. );Ċ +Ċ ĠĠĠĊ +[ curr +ts y +ĠAlex is +rit el +Ġpet roleum +.pre processing +m atter +For Result +- license +Ġtrav ellers +ĠDispatch er +enn ifer +Ġdigest ive +P ED +hib ition +MAS ConstraintMaker +ĠW att +Ben ef +.set View +d to +TE E +ĠPel osi +_EX TRA +Ġmed als +x hr +fore cast +Ġn argin +oun s +-f ill +_CUR SOR +Ġsuperv ised +Ġtur f +ĠEd gar +POS ITION +Ġcategory Id +â ī +_ ER +á»§ a +Sh own +. ll +_POL ICY +(), ' +ĠPre v +ĠString Field +ĉG lobal +ass ed +Through out +o stringstream +.awt extra +Ġslo pes +ĠSe quential +Ġgi orn +Ġz elf +Ġvers atility +lene ck +.c gi +Ġdou bling +ĠBang kok +Ġbu urt +Ġusu ário +st udio +Ġje unes +Ġm uted +Ġ ips +_f raction +&& ( +Ġst unt +'); ?>čĊ +Ġev apor +b able +ĠPR ICE +Ġæ ³ +lu cent +Ġv amp +ĠTechn ician +Ġuniqu eness +M es +ur ban +.param etrize +ĠRe play +S essions +em br +-Americ ans +_PRO XY +Ġp ian +Ġtri e +ĠD estructor +Game State +ĠIM F +ch in +Ġport e +ĠSw al +åŁ İ +Sub string +im ing +/L ibrary +Ġfright ened +w rites +Ġrecurs os +ar Result +_INIT IALIZ +ĠBad ge +_c rc +E ight +ĠDIST INCT +Ġth ro +@ Xml +ĠLegend ary +-t witter +_e asy +Ġ+ ++ +(D ATA +.L ocale +Ġk ä +Ġn urt +Ġcr uis +_ ios +Ġsens ing +_L ine +Ċ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ +pon g +ole on +Ġwild card +ç͍æĪ· åIJį +Ġbeg ging +R od +ĠÃ İ +_C ELL +Research ers +. selector +_ ing +Ġaspir ing +Ġimm ortal +Ġy min +_ robot +Ġpl ur +B TC +ĠD ID +Ġpier cing +* u +_DEFIN ED +ĠTh i +ita ire +(m edia +- ons +Ġche fs +Ġ"* . +/ AP +Ġraz or +Ġsearch Data +Ġ= & +Ġ ãĢĤ +Ġm ourn +ting ham +Ġo li +ĠVern on +_R S +ŀ æĢ§ +Ġf ácil +ang en +cel ain +Ġa il +le st +ĠQ COMPARE +g ain +ĠÎ µ +ĠK ob +ĠF ault +_config s +ç»ĵ æŀľ +. + +cal ar +(color s +M ul +_ ART +Ġexperiment ing +erm en +ĠAng lo +.Fixed Single +Se a +Ġc txt +.s lider +C ollapse +G rey +Ġf ld +-pro of +.cap acity +get Parent +ĠCom pliance +Ġburg l +- rec +Ġover written +M U +Ġrout ers +ĉ Model +Ġfantas ies +av ian +_p rec +ĠSc andin +Ġ// < +/o ct +Ġceremon ies +Month s +und y +Ġqu ed +ĠN ou +ĠV ibr +.r gb +Ġcit rus +Ġbr aces +-upper case +get Table +Ġdop o +ĠK err +_CH ILD +- cloud +ĉ Matrix +Ġgard ening +S ing +al most +Require ments +ugu ay +( Property +sub scriber +FA ST +re action +(l p +) })Ċ +` ). +.w allet +_ex change +.Max imum +ĠVer b +âĶ ģ +() < +ï¼Ľ Ċ +RO T +C ARD +ub it +{ @ +_k el +ĠTool tip +My SQL +Main Activity +ar f +Ġm align +Ġse inen +ap ist +Ġ< % +Method Impl +M il +ĠM ick +.de pend +< ID +Ġpredict ive +ĠAP PLICATION +le f +dim ensions +Ġconoc er +/ conf +ĠTr acy +F oto +_rem aining += file +Ġpage Index +ĠPar ish +Ġt exas +ĠM AGIC +ĠH ew +d ifference +Ġalt ura +c um +ĉdata Type +Ġcaracter es +avi ours +ĠV OID +è¿ ij +P UBLIC +B io +ĠstringBy Appending +Parse Exception +ĠS uff +ĠN orton +/d etails +.n ull +>> & +ĉ ok +-l ow +. usuario +n ested +X B +OUR S +.Border Color +Ġb row +ĠÐ ķ +cor r +ĠRed skins +.get Tag +.get Transaction +Ġst igma +hard t +ĠPlayer Prefs +als y +uc son +L anguages +ĠOl ivia +Ġt ac +Ġb li +Ġc aval +Ġconsolid ated +Ġper il +Ġde le +Ġform ulated +Ġhigh ways +.sp awn +== $ +ĠN iet +Ġv eggies +yp o +-r ule +ĠV ie +/e pl +Ġenf ants +string Literal +Ġtou ghest +buy er +Ġcov ariance +Ġil i +ĠSoph ie +ĠB AB +Ġ" ), +ĠU k +current Index +_user data +.code c +ĠPun jab +ĠSN P +l ol +adv ance +Ġcom fy +Json Ignore +Ġfashion able +ĠI CON +Ġor a +ĠP ricing +< num +ĠI RC +ER V +ĠMe in +ĠID ictionary +AD OW +is New +ĠDev on +at l +(request Code +ĉ PreparedStatement +IM PORT +Ġmar ital +_SELECT ED +get Response +ar Down +B V +ib Name +ĠP ATCH +ä än +Ġda ar +ĠFile Mode +Ġm arty +.Spring Application +c ene +amp oline +get Size +Rest art +æķ Ī +.project s +ĠEthi opia +Ġstatus es +T ION +(b g +ĠX unit +Temp orary +ĠEng agement +Ġx f +Ġprox ies +Ġgen esis +Pager Adapter +ĠSl ave +Ġsung lasses +ĠCh loe +Ġko ji +ad em +ĉ JSONObject +Î ³ +Ġh ors +* w +ó r +es ch +Ġcritic ised +z ial +ĠSale m +.Vert ical +ĠR ash +> E +ter ing +/s creens +Ġheight ened +аÑĢ ÑĤ +Author ities +_b box +ün st +.font Size +ĠBO OLEAN +div ide +ĠSlo ven +uc er +Ù Ĵ +st ub +Ġnavig ating +: animated +_N OW +_v ect +} {Ċ +@ ( +Ġtele com +Ġcontract ing +ĠAss ange +Ġextract ing +Ġgr ö +c obra +.D IS +Ġcr ab +Ġtw itch +Ġvert s +Ġreject s +ĉ format +Ġreg eneration +.S ys +s olve +ĉd ialog +sh i +m eter +(b est +valid ators +Ġon wards +Ġg uru +Ġmoder ator +ow ied +ex periment +r ub +Ġm qtt +ĠCa ucas +Ġnational ism +Ġm ange +ĉ ImGui +/ Edit +Ġin h +Ġint ellig +ero kee +ĉ export +Ġdiscrim inate +sub tract +ĠM oodle +ens er +ĠGuid es +R AP +-h ot +_gr p +.p icture +X A +Ġinit View +_Com m +Ġoverd ose +Ġ+ ĊĊ +ĠSil ent +show s +Ġinterpol ate +Form ation +Ġb isc +mark ets +( SC +Z e +ĠNetwork ing +Ġad renal +ĠG uns +ete or +Decl ared +orget own +Ġk arena +/ password +_address es +ITER AL +B uzz +ĠCon way +(c ase +P WD +he iro +( act +** čĊ +());ĊĊ Ċ +Ġan v +Ġ. .ĊĊ +(Menu Item +(m ail +_section s +ĉ net +Ġpl ut +Ġw rench +/ object +ĠI st +ĠV IS +/p ub +al ten +Ġguit ars +Ġantibiot ic +ï¼ ĸ + ¹ +Ġ" +" +form ula +Ġbab es +ĠP rompt +Ġen im +/ player +ĉ ref +Ġby Äĩ +Ġconsum es +ĠH ast +ĠT ao +Ġ' ))Ċ +Ġcl am +Ġthigh s +Ġmot if +Api Operation +ĠW L +get C +ĉf lags +oint ments +Ġeconom ical +need le +x ls +pr actice +ut zer +time ofday +- output +Ġfind ById +ĠBudd y +Ðŀ ÑĤ +Se ven +ĠB ark +Ġenv oy +_al gorithm +åĪ © +Ġball istic +ç§ » +r ades +ĉd oc +rodu cing +ĠE ating +Un mount +/data Tables +_b onus +Ġl itt +pp s +) localObject +per f +ĠHel vetica +sh utdown +/ ml +.t okens +ĠHard core +, row +/b g +Sc aler +âĢĶ as +_log its +âĢĻ int +ĉ App +Imp licit +.F printf +ET O +Ġterr a +Ġpossess ing +.r strip +, ), += yes +ĠStr ipe +? = +ne utral +.g ood +Ġk ennen +ĠS ung +f ault +ystate change +Can adian +',' ".$ +ĠM its +æ nd +ĠSTR UCT +ĠURL WithString +ĠCom pass +Ġ-- ĊĊ +ĠNS LayoutConstraint +| min +-ad just +Ġreb uilt +L IGHT +/ se +-m ount +vp n +valid ated +(Q Object +Ġign ition +ĠCharg ers +RYPT O +]initWith Frame +ĠFl uid +Ġcad re +Ġnomin ations +Ne ill +ĠH ou +Ġcurrent s +_g ene +(in p +Par is +z ÄĻ +ag gregate +Ġass oc +weet ed +err at +âĢĵ ĊĊ +Ġ'/ ',Ċ +fix ture +ĠH ighest +amb ient +Ġch mod +Ġcon te +Ġsens ual +Ġgar ment +z ers +ĠPower ed +dom ains +R eward +i omanip +Ġcock pit +out file +Ġbuilt in +Ġins isting +. vars +zip code +Ġ ���� +f ails +Ġconsolid ation +_ oid +Plan et +Ġ= ", +ĉ el +UIL T +ät z +af ari +ĠMc Cl +Tim eline +Est a +Ġfr am +Y E +Ġcere bral +Of Month +ĠP regn +Ġкл аÑģÑģ +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ +ĠF res +Appro ved +.S pecial +ĠProtest ant +Ġallerg y +_p cm +ĉC opyright +Ġsuper Class +" strconv +ĠMoh amed +Ġ' // +Fore Color +Ar thur +ĠJ ungle +Ġve ins +S ad +Ġback ups +ĠOp inion +û t +Ġinter mitt +ody n +ĠChrist ina +Ġand re +Ġevac uation +pa lette +h orse +ĠRes ident +ĠHass an +.N il +Ġa isle +ĠG rowing +Ġblog info +/s ql +_io ctl +Sc aling +ĠMon ad +_c pp +ĠH utch +ĠApple WebKit +Exp ense +_J OB +Ġpoint less +From Body +ant al +Ġdepict ing +ĠC ELL +Ġref in +ĠC NC +ì¹ ĺ +_dim ensions +ĠS AN +Ġa ft +Ġfoot steps +cc oli +_PH ONE +/m ath +-k ind +ĠMe ans +ich ael +.g una +Ġinaug uration +-dr iving +( delete +Ġtotal Count +_M C +.Ext ension +Com mercial +Ġz Index +< Customer +" g +-sh are +Ġp act +ag ara +ĠS IL +_m odes +ĠM olecular +Ġsystem atically +< G +_s cr +ĠO ro +as ers +Ġb ic +Ġdest roys +PI PE +.Start Position +Ġc á»§a +ire z +.B unifu +_F unction +Ġs ü +_f uture +ĠWe alth +ĠNatur ally +æĢ » +_y es +Ġabrupt ly +String Encoding +ĠCGPoint Make +Ġz h +Ġimp erson +Ġpiv otal +ĠSom alia +Ġsegment ation +_AN AL +ĠLogin Component +Cons ult +Ġtr uncated +] ";Ċ +.get Config +Ġintern ship +B aby +ê° ľ +Ġstrengthen ed +_M I +b asket +Ġnicht s +ĠTV s +ĠSh an +ãĤ µ +rac use +.Re LU +/ interfaces +ĠgetItem Count +Ġret iring +Ġspecial s +Ġentity Manager +bel ief +Ġs older +da ughter +ij kl +Ġutil izes +.f ixed +S U +Ġdr astic +Ġh acks +gr und +ĠM U +ĠSt arter +.Com ponents +_m otor +Gold en +Ġl odge +Ġ )); +ĠCor inth +иÑĩ еÑģÑĤво +ón ico +gre SQL +ĠFl uent +Ġmar c +.Load Scene +.Group s +Ġer h +ĠAut umn +St opped +Ġitalian o +Ġmin ions +ĠAssert ions +Ġm ux +B u +Ġ---------------------------------------------------------------- -------------------------------- +ĉ up +read ystatechange +_M eta +Ġcurrent Date +ĠChap man +Und o +Se an +ap r +Ġpar m +_ icons +ĠSt a +á z +Ġsub division +Ġalter ing +P NG +ponent ial +Ġpost gres +ĠB DS +-ex istent +ĠBrad ford +ĠO MX +_W HITE +_PRO GRAM +q c +Ġtypings Slinky +ĠP ics +_M ETA +IT TER +_sub scription +IRON MENT +ĠHy undai +();ĊĊ ĊĊ +ĠØ ³ +Ġj ac +Ġelimin ates +) });Ċ +Ġcomp rend +ĉ insert +_f aces +"> $ +Ġeb ay +Ġcapt ive +pl iant +ĠCalcul ates +ol ta +est ing +_re vision +Ġm ús ++ m +"," "," +WH AT +Ġcompassion ate +h arga +[ random +Ġmod ulo +(s n +Ġoccup ations +//// Ċ +ĉ board +ĠB alk +wi Äħ +ĠW ifi +.Pro file +:m aj +ĉm at +LOCK S +(j Button +Ġ(' $ +M ur +æĮ ī +b ble +Ġf rog +-h ide +Ġbroad caster +ภŀ +ha led +Ġam using +_predict ions +_in tr +Ġe agle +аÑĤ елÑĮ +Ġget List +ps ilon +Ġcharacter ization +AR DS +Ġre location +Ġr ulers +P AY +ĠDef initely +_A ction +Ġclos ures +Ġfact ual +odyn amic +Ġpreca utions +nie j +ĠPart ies +ĠSub aru +Ġcous ins +ar beit +.m oney +gun ta +( and +get item +.Style Priority +Ġsl id +single ton +Ġg arn +ĠP AS +Ġd azz +a ż +Ġbog us +ĠM og +Ġrival ry +is ol +Ġland marks +ñ as +B ern +ĠSach s +Ġ" )ĊĊ +Ġhost ility +_m ex +m ere +M ot +p ictureBox +Def ense +Ġaffid avit +other wise +.d irectory +_ UnityEngine +-b log +.s kin +ph em +Ap ellido +er chant +[ class +Ġw art +." [ +ale ur +/ back +ĠĠĠĠ ĉĠĠĠ +Ġprecip itation +Ġob struction +Ġp Obj +Ġr upt +UCK ET +ay e +æİ Ĵ +g x +Ġe cl +Ġsecre cy +/ Header +ĠLes b +Ġle i +ĠBullet in +Ġgive away +.H ome +_RO OM +" W +Ġcow ork +_ ra +ĠC ycling +ĠP aw +Ġpup il +/ arch +ĠFile Utils +é¦ ĸ +r sp +Ġfreed oms +ĠL ear +}` ). +Ġbow ls +/b lock +_log ging +Ġmeth ane +Ġhorn s +Ġwonder fully +Ġalter ations +Ġex ile +ls en +_p ause +_L ANGUAGE +ĠUS DA +_m ysql +_AM OUNT +ĠL IFE +Ġyoung sters +Ġri ots +[ E +Ġun forgettable +, },Ċ +Dis posed +ĠAss assin +UN G +ĠNew sp +User Service +: aload ++ ', +Ġsett lers +Ġscre ams +Ġincon venience +.R otate +Ġj ars +ĠP uzzle +Ġm est +ars i +ĠSh arma +| ( +.d s +ĠSac red +_e vt +Ġexpress es +Ġh och +ĠD uch +.c alls +th r +ĠShe ffield +.Alert Dialog +Ġrad ically +Ġtr ous +Ġprev ailing +ĠWW II +âĢĻ n +ens ely +ĠY esterday +ĠSir ius +Ġkill ers +ĠF FT +Ġo val +') :čĊ +Ġìłķ ë³´ +our age +ĠCheck box +Work book +.def er +_f loor +Ġc ouncill +Ġnors ke +mo il +ore a +Ġmarket ed +_S UR +x AA +Ġst ained +e ut +ĠM eng +Ġi eee +. extern +eg ie +Ġr app +ĠPy ongyang +' class +M ob +Ġinitial Value +_w ave +Ġj ab +Ġmascul ine +Ġampl ifier +Ġt ty +Path Component +_ xt +ĠG FP +/ sec +ĉdis patch +mark down +ĠS chn +bo le +· · +mouse move +Ġerr Msg +Ġas ign +_m ono +To Selector +ĠZ u +(R ect +ĠError Code +lat in +ang ible +v tk +CG Size +P okemon +Ġclass mates +Ġattract s +ĠT atto +ult an +ol óg +Ġhalt ed +ठ¨ +ĠK art +Ġ ue +_Init Structure +Test Class +ĠAir bnb +_ ", +Ġchar coal +Ġip c +ĠSt retch +.g lide +lates AutoresizingMaskIntoConstraints +Ġpot ion +ITT LE +Ġcount ert +_h d +pre pared +Ad s +ĠV ampire +rob ots +.Create Index +Status Label +Ġt ucked +af ür +U t +Ġswe ater +_F N +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĉ +ata ka +Ġeyeb rows +ac oes +ud en +.LinearLayout Manager +Ġsw ay +Ġmult in +() )))Ċ +ĠNS UInteger +ĠMy Base +Part ner +uts chen +ĠC ater +.setBackground Color +Ġaccompl ishment +_pro blem +.d td +Ġpage Number +Ġj ackets +Ġcro pped +u els +ĠH ep +Ġc apped +* Math +_callback s +Ġpub b +ĠBrun swick +.res pond +[" _ +Ġbed ding +hyth m +O X +(s peed +Ġpestic ides +Ġ---- --- +.Bl ue +Ġnood les +ĠGo es +Ġs aver +o xy +_com pletion +ĠSw inger +Ġget Date +Ġmind ed +int egration +ĠLot us +(st op +(', ');Ċ +Ġflood s +ĠWork flow +Ġerupt ed +Mac ro +ĠSau ce +Ġevent Name +\ Input +Break ing +ĉ when +_p w +IND ER +ĠWell ness +Ġvox el +ĠM ell +ĠM EDIA +SE NS +ĠFund s +ĠM ild +< Array +- this +ump ed +/f w +ĠDb Context +W I +girl s +H OW +'); ?>Ċ +Ġtempt ing +Ġtest ament +Ġb ible +Ġconsult ed +ĠIndex Error +è¨ ĺ +Ġkey pad +izz o +( ok +Ġwhats app +ĠRemote Exception +Ġteam ed +âĢĶâĢĶâĢĶâĢĶâĢĶâĢĶâĢĶâĢĶ âĢĶâĢĶâĢĶâĢĶâĢĶâĢĶâĢĶâĢĶ +» , +Ġget Time +di ag +iss y +Ġh ed +Ġkn ots +j om +Ġfun nel +-m ails +Ġexport ing +ĠV L +ĠK arn +ĠBuddh ism +ĠAll an +_R ADIUS +Ġw ording +ĠFor get +ĠCor ona +ip hy +Ġlim burg +ugg y +ĠUser Repository +im in +(e le +Ġlabel led +ç¤ ¾ +ĠH erman +.q q +Ġ" ));Ċ +ie ber +.Trans late +ry n +Ġdes env +um d +Sim ply +ĉm ode +R pc +ĠVal encia +Ġstaff ers +Ġsel v +ĠSpi ke +Ġdel ic +Ġer u +_D T +J udge +á» ķ +ĠBas in +.m utable +" url +Ġtar iff +ĠSlee ve +Ġfl are +.drop out +Ġbr ides +)) ,čĊ +_con straints +de struct +Out line +Ġdisappe ars +_lock ed +ĠNS LocalizedString +ck e +ĉ null +ad resse +Ġto pping +ĠJ oker +b ishop +но ÑģÑĤÑĮ +and ering +_ amp += time +_S pace +_P ULL +' = +Ġant iqu +Ġc ach +___ ĊĊ +ON ES +о Ñı +Ġun read +.p olicy +oooo oooo +ëŁ ¬ +Ġu sted +ĠRe ce +Ġal lem +ãĥ¼ ãĤ¹ +ĠThought s +ve illance +istr ate +_l ane +Ġfam ed +.Get Name +Ġsmo other +ĠQual ified +az ers +_ geo +F ax +ĠM inds +ĠR aises +Ġtrans cripts +Con versation +Ġremark ed +ëĤ ĺ +d ling +Ġdeploy ing +Ġshared Application +Ġk p +FontAwesome Icon +_d ummy +reib en +ĠJane iro +Direction s +.get Bean +s ass +Ġcommand ers +v ation +error Code +ĠAl loy +.local ized +Ð ij +Ġdish washer +ĠSou p +N u +_D efault +Ġune ven +Ġ/> ";Ċ +-B ased +Ġseam lessly +- null +ĠX C +Ġst ew +(d elay +AT ORS +ĠWhe eler +" H +e ast +. air +âĢľ But +Object Context +success fully +_l and +Ġfold s +_CO ORD +Ġsub po +.get Address +in str +Material s +Ñĥ ÑģÑĤ +de posit +-l ast +_GR AY += find +Ġmut ant +Ġlesb ienne +let cher +RO UGH +ure ka +.c apture +Ġen n +Ġ([ [ +ĠFl u +Ġtask Id +ĠHus sein +.f older +Ġa usterity +ISTR ATION +_ Impl +注 æĦı +Ġdec ree +- chat +Ġimp lication +Ġguess es +ul kan +An alytics +. plus +COM MAND +е ли +» ĊĊ +_S ITE +Ġequal To +Support FragmentManager +ĠRec ording +å®Į æĪIJ +Ġbag gage +Ġpitch ers +ĠE h +o que +ĉc nt +Ġ=> $ +/ foo +IR A +ĠSat ellite +bor ah +Ġ}} "Ċ +ĠEnd s +ĠSpr ay +, param +.Ch rome +* q +th ought +ibr ated +Ġth ieves +Ġbenefici aries +Enter ed +ottes ville +Ġveter in +By ID +qu ipe +um ption +- unit +Execution Context +@ s +ĠG iov +.Tool Tip +_f riend +( attributes +Ġdump ing +ĠJ C +_D OCUMENT +ĠArm our +( insert +.Horizontal Alignment +ĠQ ed +ãģĦ ãģ¾ãģĻ +/g it +ĠY YYY +ĠCard iff +Ġap a +organ ic +ĠWhere as +Ġæ Ŀ +ĠM ia +Ġdemol ition +Ġsc ars +Ġp ai +Ġre tries +Ġr q +ĠDen is +( Utils +Ġallev iate +ĠP IC +id ue +Ġacknowled ging +Ġ// //////////////////////////////// +ç¡® å®ļ +Ä « +\ Json +.b inary +Ġx type +sign als +ĠAp pearance +& r +} s +C i +ĠI llum +por ate +h og +Ġindex Of +\ Command +_par allel +ĠSher lock +í ĥ +Ġ" ")čĊ +//////////////////////////////////////////////////////////////// //////////////////////////////// +Ġcritic ize +ĠSo ap +ĠMatch er +Ġgr illed +* T +Ġad ore +ull ing +Ġjed och +_ref s +lean up +ĠJ AXB +Ġro ses +ĠL iam +size i +Ġget char +Ġtar de +-to oltip +Ġqual ifier +ĠInter mediate +_W indow +ĠMal ta +Dis connect +ew here +Camp o +Ġirr ational +led o +ĠD N +ARG V +Ġout ro +Ġth irteen +Jose ph +M AR +/g l +J ess +ĠPsych iat +Ġpadding Bottom +- loop +/ fonts +_se en +Te ams +React DOM +(m an +(x path +.get SimpleName +>( * +ĠP vt +Ġel ders +Ġp ies +.user Agent +- region +ĠGree ks +(f ragment +st u +Ġcouncil s +Ġst amina +ĠGod dess +è ¥¿ +Ġphilosoph ers +Ġpers one +ĠL ose +ĠCL R +ĠD ocs +Ġso ak +ĠHOLD ER +Ġb ells +hash Code +R ATE +_WE IGHT +in ous +end ra +oph obic +Ġpro se +Ġfin ely +/o auth +(s pace +ad ge +ĠM ama +Ġstring Buffer +Ġst int +Ġmis ma +Ġvill ains +ĠCrime a +Ġdipl oma +Ġпо Ñģл +ĠBe a +(j oin +Ġíķ ´ +CH AT +per ing +ĠC ros +Ġmon keys +Ġpred s +yl a +,, , +Ġvibr ator +ĠN U +åħ Ī +f ant +z et +Ġb ietet +un ft +sw orth +.F low +Ġpsy ched +ĠContin ental +> t +Ġqu ilt +. UP +Ġexpans ive +Dis pose +(l anguage +C aps +_Z ONE +Ġrec ycle +ĠMan aged +current Color +.b roadcast +sign In +.p rom +ll u +ue blo +Ġpunch es +Ġautom at +Ġassign ing +Ġcreate User +ĠAll ied +Ġconduct or +Ĥ ¨ +Ġs addle +Ġd ni +omed ical +-W est +Positive Button +Ġit alic +? [ +(tr igger +Ġele phants +":" "," +Ġcal iber +raft ed +d igits +Ġmar shal +mill iseconds +mark ers +m om +/ place +Ġhol istic +: t +# , +Ġb oto +Ġnause a +ĠSh ooting +ite ch +Ġtext Status +< Class +ĠDes cribe +Ġbuff et +g il +Ġlog its +std call +mod s +ĠSk ull +ĠB are +h ope +ĠIn tr +F air +ĉ pt +Ġacompan h +Ġf kk +_r pc +Inst alled +_ ans +.get Minutes +â̦ "ĊĊ +- thread +Ġpres chool +AIL S +Ġdiff ic +( convert +ĠN ath +ĠDO J +Ġreg imes +Ġenthusi ast +Ġwarrant ies +Ġfasc inated +_b inding +_N ot +oft en +_R W +/m ail +Ġtitle Label +Ġvill agers +ĠJ iang +Ġsw agger +.Row Index +_img s +rap y +VER AGE +. Up +Ġno op +c io +ĉ ST +Ġdecre ment +Ġmagn esium +_ rotate +S it +Ġnieu we +Ġter med +íķ ©ëĭĪëĭ¤ +Ġur g +_t ouch +Ġsw arm +Ġcl ave +th est +ĠL af +H X +ĠH ulk +Ġplaint ext +ĠSof a +get Session +L ed +Ġecosystem s +he i +ĠK ills +Ġhus bands +Ñħ ÑĢан +(d om +_t iles +Nib Name +Ġdon ating +. acc +Ġlifes pan +.b n +_RG CTX +æ ¥ +ans en +Ġmod elling +Layout Params +ĠonChange Text +rs a +- location +.P e +(b us +(s ong +Ġprodu k +ĠSH OULD +ĠC J +Ġs os +ĠHome Controller +.load ed +(D ocument +.s ocial +t iles +Ġl ame += df +.parse Long +Ġpr ac +Ġdet ox +ĠV E +Ġpunt os +Ġdo ctr +Ġan cor +CA PE +Ġc mb +çĦ ¶ +*) " +:// / +Value Type +Ġmort gages +; q +ĠRock ets +s port +UG C +ct s +ãĤ ģ +ie ur +ĠAppe al +(n b +//////////////////////////////////////////////// //////// +IM ATION +ĠC res +ĠMan ip +C ause +at ypes +man ufacturer +# ---------------------------------------------------------------------------- +Ġsp or +es on +Ġpun ched +Ġbook marks +ĠBul k +Complete Listener +ĠTalk ing +ĠEr nest +Ġrub bish +k ills +ĠDE FIN +Ġneighbour ing +ar lo +ĠP CA +ĉm atrix +lo k +Ġat las +ĠG ur +Ġw yn +-n egative +Ġt ul +Ġre lic +ĠV oltage +ĠPre is +ĠJ NICALL +ĠPM ID +ak et +ĉ attr +Ġet iqu +ĠM J +ĠG mail +cl r +_exec ution +éĶ ® +pos itor +. af +N r +Ge orgia +Top ology +Ġperch é +Ġmus lim +Ġepid emi +Ġsab ot +act us +Ġë ĮĢ +ĠIO Error +. est +p refs +ĠKr ish +.Read Key +NAS A +u ção +_D b +umer ator +W ide +(st atement +.end point +.... ..... +Ġ[ * +stream s +m time +P x +at r +Ġt pl +R oman +Ġscen ic +.n z +ĠSe conds +sub menu +Ġìĭ ¤í +_b undle +Ġde ÄŁ +ĠS isters +pre ferences +Ġport a +Ad visor +max Length +ĠG REAT +__ (Ċ +ole st +ĠLabel s +Ġen fer +ĠĠĠĠĠĠ ĊĊ +ĠThe ft +_F ILL +ĠW ise +) application +un ami +> ())Ċ +ADD RESS +B ST +et zt +ĠQ gs +S ense +Exception Handler +ĠCh u +.get OwnProperty +Ġexerc ised +iot ic +ĠRe leases +Ġp interest +ol ie +is oft +Ġsequ encing +Ġpad re +] ));čĊ +(r adius +.m ed +aint ies +.Object Model +Ġem ple +Ġseg uro +St ars +Ġqual itative +lem n +á» ± +> "). +Ġg x +-c ert +ĠAST M +Ġfull name +Ġte lemetry +ĠCamb odia +_ ul +ĠCl are +C USTOM +Q C +ĠUn s +ĠHTTP S +ĠPark inson +ancy box +',' . +T ue +.get Last +Ġab i +Äħ d +A st +ĠEd iting +.Un ity +j mp +Ġm ats +Ġshared Preferences +Capt ain +.page Size +Ġr tl +Ġan meld +Runtime Object +Ġdemand e +(" ; +se ite +-head ed +ĠK ra +ĠF ONT +` \ +Class NotFoundException +. avg +atic al +A j +Ġpermit ting +Pro j +ERR Q +Ġcre ampie +ĠBuy er +-mod ules +ĠSund ays +| `Ċ +Ġday time +Ġ+ ( +Ġgl itch +ĠOper and +Ġtox ins +iny a +D NS +ĠS as +C ake +ĠNation als +.add To +Ġs inking +Ġcompreh ension +Ġsc or +ag ements +Ġt ard +Ġmarch ing +ĠM TV +Ġs ane +Create Info +Ạ¯ +Ġend Index +ĉ layout +ĠåIJ į +S ITE +ĠT HERE +Ġ[ {' +opath ic +Ġtrans mitter +/ body +Ġp und +ĠC losing +Ġset attr +Ġbound ed +At las +sum ing +(t imes +par er +yn om +fe it +Ġf rem +- leg +ĠBr as +> # +Ġì¶ ľëł¥ +ĠIN STANCE +ĠC ouch +_host s +lik elihood +.M arker +ĠM asks +Ġcere al +util ities +Ġelement al +Ġdist orted +in active +c ry +W L +UPPORT ED +.Th rows +/s chema +ser ie +." ', +ĠBened ict +-p icker +ig gs +ĠPir ate +åij¨ æľŁ +ĠTh ema +ĠSouth ampton +Ġarray With +ĠPaul a +Ġpredict or +- Ass +.user id +Ġper i +Ġexagger ated +ur ate +arse ille +ĠCon cent +ĠP ik +Ġ@ _;ĊĊ +Ġform ations +Ġden omin +"/> .Ċ +ended or +Ġpan cre +Ġam t +Ġon Resume +on Delete +ĠB CH +) (" +m ovement +Ġpot assium + čĊčĊ +ĠMah m +} ";ĊĊ +Ġd q +ĠPublish ers +ĠAm pl +ĠDani elle +Ġt ern +èµ · +no ÅĽÄĩ +e in +ĠAsync Storage +un ger +rou w +Ġsc issors +/ assert +.b ucket +/ archive +_M an +Ġint oler +Ġ() => +ĠÐĴ Ñĭ +Ġsa i +.x y +." čĊ +Ġur inary +es ub +IST ICS +ĠÎ º +Ġcompl iments +Ġtypings Japgolly +ih ar +Exp ansion +ĠS erving +_st udents +ĠX BOOLE +( il +Ġì² ĺ +Ġj ó +(t ol +( JS +ĉC G +ĠD RAW +tw ig +Ġo at +_sm ooth +ĠC SL +Ġos ob +Ġens uing +Ġbank er +ĠBack pack +_p ing +Ġwish list += ax +ĉĠĠĠ Ċ +Dis ney +stead y +"> % +Ġproph ets +ĠZ X +Ġminimal ist +.PL AIN +Se attle +. ordinal +ĠPI PE +Ġret orna +Ġjug ador +ĠB ret +ĠâĶ ľ +Ġpl ush +UL ATOR +Sort ing +.grid y +ect omy +_ activ +r ack +Inter active +ĠAntar ctica +Ġv engeance +en so +_k nown +up plier +.Mod ules +ĠConnection State +éļ IJèĹı +@ FindBy +Ġpl acer +\ model +< ()> +.is Successful +-g ood +b z +ĠDr aco +Ass istant +-ex tra +аб лиÑĨ +Ġhyp ocrisy +Ġt st +ĠA gr +$ txt +Ġlog istic +lic ensed +ĠH of +Ġt at +( iv +Ġinto xic +post Id +_st rike +Ġhum iliation +pc odes +" sync +(rec ipe ++ N +rent e +ĉ Client +ycop g +ĠZur ich +ĠPro files +C ountries +Ġp ict +Ġroll out +requ encies +Ġpatch ed +Ġcar tridges +Ġsh ading +J ar +Ġsalv age +ĠTax es +Ġstand by +apor an +E igen +. angular +ĠN ested +äº « +Ġis Visible +ĠDw ight +_BR ANCH +.D elay +Ġk end +Ġfacilit ated +.flat Map +Ġs anta +ĉS end +/m essages +Ġof Type +ĉs wap +# plt +ĠTur ks +N ES +Ġprogress ively +ĠRes idence +ĠT REE +Ġno en +d io +Ġn elle +Ġsog ar +itt i +week ly +Ġambigu ity +_Set tings +W are +.ne o +_D ST +Ġæĸ ¹ +pre p +lob by +@ email +/m ovie +Ġfun kc +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ Ċ +ÂŃ s +Ġguard ians +- pos +Ġconfig uring +ĠC PS +ĠDe us +Ġvidé os +_ empresa +Ġsl apped +< Model +Ġunders cores +U h +.access Token +SET S +ĠS parse +ĠCal d +: path +ĠS ervers += batch +Ġkn itting +Ġx a +Ġsearch Bar +Ġsn ag +Ġinf used +.b am +le ver +Ġtax onomy +Ã İ +Ġatt aching +Ġh ern +_N OP +Click able +(P arse +ĠDynam o +-b uilder +Ġdere g +Ġsc attering +è¿Ľ è¡Į +an zi +ĠShe pard +"> ',Ċ +_X DECREF +ĠBuzz Feed +_M ARGIN +P LOY +.sm all +Ġm imeType +Ġh olog +ĉc amera +li as +Ġsusp ense +ody nam +b au +Ġgrave yard +_n amed +":" ' +Ġ******************************** **************** +Ġgame Over +ĠLENG TH +ĉs creen +Ġdo InBackground +_depend encies +Ġr tc +/ up +_ ROM +H all +Ġdef iciencies +( te +' # +_e quiv +Ġpre order +ĠA xe +ом Ñĥ +.send File +Ġfil t +ĠLim its +ĠCaval iers +.dis count +âĨ IJ +ĠW it +QRST UV +Ġi j +Ġt egen +Ġ: ", +diff iculty +p unkt +ĠEmail s +ch lor +(f un +.U int +ĠSt all +_ verified +u D +File Type +Ġple asures +Ġjud iciary +Ġsh am +ip ur +_PL US +off ers +( foo +_G T +ĉc ore +ENT ION +ĠLib eration +Command Line +_de partment +.A r +_ne ighbor +ĠSub mitted +ĠĊ +Ġdro its +Ġhomosexual s +Ġab duction +ĉw idget +$ headers +ĠD AR +Ġfl a +th reat +Ġlou is +.Get Property +" Just +(f rames +ry o +prof ession +| i +íķ´ ìĦľ +(s v +Ġun recognized +I onic +F ashion +Screen State +ĠIn coming +Not Nil +Ġsync ing +em ie +Ġtherm o +_pro cs +Ġincons istency +rel igious +.m j +Ġperson n +Ġmoment os +or arily +Ġæ Ĭ +_ne urons +Ill ustr +im oto +il ik +ĠW oj +Tr ading +Ġapp are +Ġentre prises +ach at +Ġ ¬ +Ġne igh +BUTTON DOWN +ĠMah er +ag han +-h ash +" f +Ġclient ele +.add Button +ĉ SP +Q i +Ġgr ated +POS ITE +: > +ĠHow ell +ĠCompar ative +ĠIS C +ÂŃ i +O cean +D avis +ĠFil me +W ins +ĠJ IT +oc cer +ĠC orm +ENCH MARK +rch ive +ica ção +Ġm ata +Ġchild birth +ĠOption ally +En s +Ġx http +Ġel ucid +_Osc InitStruct +)) ):Ċ +Ġint uit +ĠDon ate +Ġcorrel ates +> Delete +Ġequ ipe +Ġb oca +Ġinfl atable +er ah +ĠDateTime Kind +Ġcal ves +\ Lib +Ġem lrt +ĠTr ilogy +ĠP anc +ĠD uis +ĠpelÃŃcul a +WAR DS +_DE TECT +-section al +dh cp +For Row +-de struct +ĠPres enter +/s lick +, on +ĠCit adel +logged in +_sub type +Ġsig ue +Ġc uring +ĠFire wall +Ġfluores cence +ĠItal ians +иÑĤ ÑģÑı +.get Style +In Seconds +j ie +-S mith +Ġx link +Ġsub missive +он ÑĤ +arbon ate +ĠF aul +_go als +ĠCommission ers +chart Instance +_POST FIELDS +Ġmed ial +Ġman os +Ġdel t +sv m +.Ap is +ep hy +Ġasym pt +Ġapp Delegate +Ġimpro bable +ck a +sim d +/ Error +. âĢĵ +ĠP TS +de er +Ġs ina +m agnitude +ID ADE +'] }' +Ġmay ores +ĉ comment +/ console +" @ +v olt +.s ell +ĠM acy +Ġmel od +Ġim ágenes +_ch g +Ġin out +ident e +) '),Ċ +d ni +.b lob +Ġtyp ography +Ġe erie +_O ID +pes an +aj an +Ġch opping +Ġbl uff +ad f +_b ases +.Form atter +Ġ\ % +ĠPage Info +Car rier +ĠCal ibration +com o +-b odied +Ġfinanc ier +ĠIN A +. ERR +Ġhood ie +ĠSan ity +gu arded +.opend aylight +ISM ATCH +High lights +ün k +ani em +anger ed +assign ments +Ġregistr ado +ĠU PPER +ampil kan +ash ire +ĠNik ola +ĠC FL +ĠH DC +Ġp oids +ĠIP s +Ġprevent ative +ips oid +if ix +.c amel +.g a +V olumes +- ste +Y ahoo +_s ibling +H ighest +opt group +Ġkvin na +âĢĿ ãĢĤĊĊ +ĠAppl iances +Ġ" >< +') ")Ċ +ht t +ĠIdent ified +Ġpenc ils +Ġmember Id +Ġappend String +.load Data +Ġmock Mvc +Ġj ub +ĠSl ut +ĠTai pei +st att +Pol it +Ġpart ager +Did Change +Incre ases +) }. +ĠB aba +_CL IP +[ unit +Ġк лÑİÑĩ +Ġalc uni +ĠL ola +Ġcl inging +@ PostMapping +(con cat +Ġss id +ĠFa uc +ok it +ĠRecord ed +á lez +($ ('< +.assertIs Not +Ġk ali +V olt +Ġwarm ly +Ġsca res +get ti +füh rt +_d oes +. EMAIL +im ations +Ġspring fox +ĠDec om +arc y +Ġgl itches +ĠM off +ĠV oll +.b etween +Ġcoord en +ĠPart icularly +GB P +Ġsem ble +East ern +_M SB +]) {čĊ +m organ +ĠE VAL +d ere +HO USE +mo ire +ist ique +_l stm +-com mit +yster ious +Ġtw ink +-th umbnails +en ÃŃ +:' ', +Ġblack out +ĠFlo ors +Ġso fas +Ġou i +lesh oot +ĠRa q +- abs +Ġk ra +M ining +sha ft +.set Columns +Cl azz +PRE TTY +.play list +éĸ ¢ +-Sah aran +M ING +ĉ bl +è® ® +j f +DO CKER +hope fully +( ignore +ĠUsers Controller +ĠMitar beiter +ĠL ES +Ham ilton +-m etadata +ĠK K +ikt ig +Ġwoll te +egr ator +] bool +, current +Ġvalue Type +Ġexcav ation +ol and +Ġv erv +/file path +Auth Provider +Ġpro crast +ĉ ULONG +_MEM BERS +Ġup lift +ĠAut onomous +Ġart works +ĠOut reach +Ġp ore +Home page +Dialog Title +ĠGener ating +PAR SE +Ġsem anas +Ġhuman o +JSGlobal Scope +Ġvol te +Ġb ella +(is instance +Ġpl c +\C atalog +Ġeste emed +éĽ · +(s uffix +Ġswe eps +ĉ ORDER +Ġdo ivent +ĠSw arm +ĠComp iled +get Page +AD R +.R ichTextBox +ĠN aming +ag ged +ĠG ANG +r asing +ode led +Ġg ala +ĠJS Name +dd f +Ġill ust +ĠLans ing +[ port +-de ath +Ġdin heiro +ĠE ighth +Ġb ian +st Ã¥ +Ġvers ión +ĠLinear Gradient +ĠHard ing +. *) +ec zy +$ header +Ġv Ã¥r +Un checked +Ġko je +ĠPal adin +() )), +G iving +() })Ċ +Ġd ips +F riendly +Ġport rays +Ġhel ium +Ġinsurg ency +_ex piry +ĠstringByAppending String +Ġa antal +s lope +m ast +.get Integer +Ġ################ ######## +_PIPE LINE +Ġdens ely +Ġmut ating +m idi +ĠSe it +ay ne +NOW LED +ĠDes mond +ĠF Name +ĠN airobi +\ Context +Ġcalc ular +-d en +Ġc ott +] ):čĊ +ĠRecommend ation +ĠRole x +Ġvalidation Result +.p at +Ġn Ãły +ĠRest Client +ĠG PI +ĠAshe ville +ĠO SP +ĠPER MISSION +ÐĶ Ð°ÑĤа +/ notification +K night +_W ord +ĠB ender +rank ing +Ġpart ida +_res ervation +Ì Ģ +Ġm Name +Ġget ch +Ġb orr +Ġdilig ent +Disc uss +æŃ£ åľ¨ +ape ake +ion ed +-N azi +.c um +ĠK ron +=$ ('# +/s ingle +Ġerot isch +ĠV ib +Ġrat ified +Ġconcert ed +ĠREG ARD +Ġdo br +.Driver Manager +' r +Port able +ĉs uite +Ġrel aciones +ĠD op +emplo i +DO B +Ġcr umbs +Ġx ls +_App lication +(': ', +Ġ---------------------------------------------------------------- --------Ċ +m se +Ġber k +ĠReturn Value +ĠBel ly +Ġcam ar +ĠPe ek +els ing +Ġnot ifies +ĠTr istan +ĠG AR +em me +ĠElev ated +_C SV +(ch alk +Ġtw enties +ĠSearch Result += search +ĠMix ing +ý t +Ġrecru iter +ĠIDE OGRAPH +ĠA go +( Operation +$ values +Ġworld ly +ĠRosen berg +ĠConfigure Services +>* Ċ +Ġsn ork +_op acity +ĠinitWith NibName +i ado +A AC +Ġ] ). +; z +_par agraph +Ġnos es +stand s +if r +_m E +I raq +.P redicate +ena ire +]] ];Ċ +Ġun idad +Ġretire es +_h ello +Ġmode le +ĠUIT ableViewController +f write +_num ero +_vis ited +Ġrece be +( Notification +Fant astic +_sub menu +ĠP EM +ĠCup ertino +approx imately +class ed +.Read String +Ġdomic ile +_P W +Ġball park +ĠK ale +con tra +_f avorite +/ of +Qu ite +ĠOT A +Ġacceler ometer +did n +| ^ +ĠRohing ya +ivic rm +ann abin +обÑĭ ÑĤи +or ado +') + +Ha unted +, ID +( UIAlertAction +ur v +_b el +ĠMex icans +/ terms +ĠPaint er +Input Label +ĠV inci +ĠRos ie +\ uc +< Menu +Ġcool ant +(current User +_d ual +) "},Ċ +& p +Ġconver ged +Ġrestr ain +ĠYugosl avia += target +Ġimp uls +ds a +Search Tree +Ġh box +ĠImp ress +§ Ãĥ +get FullYear +(d a +ĠY YS +.al ignment +.Get Text +.token ize +ĠOlymp us +Ġmur ky +ore station +Ġdiss atisfaction +ĉT Array +_ kses +.Add Singleton +ĠStart Time +Ġfan atic +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĉ +Ġentity Type +. override +Ġ ------------- +ĠDat agram +f out +(with Id +Ġ# __ +Ł èĥ½ +ek yll +.f riends +ame leon +Ġz ach +.simple Button +ret orno +Ġkon k +/s mall +ĠQuick ly +un read +Don ate +Detail View +Ġdu a +Ġpenetr ated +OM UX +Ġn ir +_p data +"], [" +Ġlow es +Ġdop ing +Ġas ymmetric +Ġneed less +our cem +Ġup ro +ĠGu zzle +af b +Ġsext reffen +-c ollar +Ġcol ossal +Mon key +n ish +Ġhandle Message +Incre ased +* dx +ĠChatt anooga +f org +ĠOr den +Ġsh ri +ĠV and +Ġ" @" +Image Sharp +ĠWild cats +pon ible +.sc enes +Ġpaint ers +ĠPf izer +ĠZ ah +To Local +ĠFl am +Ġé taient +)) ^ +ĠSand box +ĠTR ADE +Ġchrom ium +Ġac claim +Ġpac man +´ t +) reader +M ari +.Dispatch er +.A DMIN +ĠRem ed +Sw eden +Ġoverl ays +. er +Ġp ang +Ġclean ly +aven port +Toy ota +patch es +Ġv tx +ĠE is +cl ado +ĠR itch +RO LS +Ġh ade +Ġconspic uous +Ġdo cks +(j q +ĠPrem iership +ĠBe z +ĠâĦ ĸ +ĠÑĥ Ñģл +_tot als +Ġprov a +ĠC ue +Ġsa úde +ĠGame Controller +IM IZE +, port +ãĢĤ ( +.C decl +Instant iationException +Ġcoll age +ĠIO C +Ġb ais +Ġon Finish +-st ars +set Size +Ġmog ul +Ġdis illusion +Ġche vy +(S chedulers +( IR +_loc s +Ġcann ons +Ġcancell ing +/b us +Ġbuf io +ĠY ours +ĠPik achu +Ġter me +r Ã¥ +f ahren +Ġowner Id +Ġoblig atory +Ġcul p +Ġacid ity +-m ult +ĠBam boo +Ġ' "> +_g s +Ġcomp il +n ard +-ex c +Ġrh yme +Ġbut to +s ays +ant asy +ë ¸ +Ġcitt Ãł +Ġche g +Time String +Ġpos itivity +ĠD abei +Ġw ang +Ġes cre +" c +ĉv ideo +ĠRank ed +.str ings +>> >( +Ġин ÑĤеÑĢ +Ġrest a +[: ,: +Ġrend re +Ġdes er +J os +Ġdis ruptions +Ġоп еÑĢ +s ampling +sup press +Ġcontainer View +ĠSeam less +Ġair y +Ġon load +.Window Manager +ĠPL A +br aco +.set PositiveButton +Ġp du +Ġg si +ĠC li +_gr adients +Ñı д +ĠWh isper +c stdint +Ġl äng +Ġform ulations +én om +ourn emouth +[$ _ +Ġordin arily +.set Username +Ġfacult ies +MIT TED +/ values +Ġwe ir +ĠA pt +M Z +ĉc f +uck en +ĉĉĉĉĉĉĉĉ ĉĉĉĉĉĉĉĉĉĉĉĉ +def ense +[i Var +ĠBusiness Exception +Select ors +(co ordinates +ĠRes ets +ĠDr inks +ole ans +(st ypy +_IO C +.x xx +ĠSl ater +ĠBel ize +Ġ/ ************************************************************************ +add in +_ep isodes +Ġis chem +legal ArgumentException +D anny +Ġp ared +.code haus +ĠAss y +ĉ Rect +â ŀ +.list a +Ġв аÑĪ +Ġv ets +HW ND +ison er +Ġx o +Ġor ally +ĠSt mt +.r nn +ĠD PI +ĠStr ikes +.setViewport View +Ġèĩª åĬ¨çĶŁæĪIJ +Y ELLOW +GL enum +part ners +ĠImp licit +Ġtak o +âĢĻ elle +Ġerm ög +total Count +G il +ĉ work +Ġpr atic +in ati +ab ies +ĠSk inner +Ġspir ited +Ġpancre atic +Ġh df +' em +Ġpsych osis +olic it +Ġ" {" +_at ual +Ġé lect +TE AM +Ġd ak +ĠSW AT +.Fragment Manager +Ġprovision ing +l ifetime +_EXTENSION S +ĠC ASCADE +Ġ! [ +(K P +Ġv em +ĠInterr acial +'] },Ċ +sp acer +_k v +W arehouse +R DD +_f sm +.Stretch Image +, Yes +ĠRefuge e +ĠBr inging +Ġv álido +.inter section +Ġsp ooky +_port al +Ġmo th +ĠZ odiac +ĠSOC IAL +M imeType +'] }} +_Bl ue +Ġbot anical +Ġfr ags +Ġfamil ial +- du +Ġse izing +(block s +.r d +.check NotNull +Ġmis er +Ġmax x +ĠK nee +View Item +Inner HTML +D anger +(( __ +Ġprz ypad +create Url +** , +ĠDecor ating +ATEG Y +?> / +.Design er +hex digest +ĠEvery where +all eries +.TEXT URE +.Block s +z ell +Ġpre ço +S uddenly +input Email +(s ync +.b d +gold en +> '); +ĠDick inson +>> (Ċ +ĠQUE UE +Ġget Column +ĠS AND +.p iece +lic er +Fl utter +Ġget Version +Ġresource Id +og l +ÅĤ aw +.Br anch +ĉ web +Ġfr amerate +PP P +Ġfr ay +C NT +Ġinformat ie +'] čĊčĊ +ne as +Header Code +Ġæ ¸ +Ġtr g +raw types +H onda +Ġmark eter +Ġrequest Data +ĠP g +ĉ not +Ġpage Info +Ġakt uellen +ãģķ ãĤĵ +ĠA MS +push ViewController +ĉ AL +Ġv ests +produ ce +-m ême +ĠRah man +F unny +E Z +_ Valid +Ġsquad ron +Ġl ash +Ġ irm +ias co +ĠPar an +Ġpet ites +ĠDec ay +Ġun initialized +priv ileged +Ġm bedtls +å¤ĩ 注 +Ġ^ . +Ġec static +D etroit +Ġpart en +Ġsou venir +.get Login +моÑĤ ÑĢ +en ção +ĠmÃŃn imo +ĠAccess ed +ri ó +M ic +ĠV ocal +.Set String +Ġmens ajes +åĢ į +Ġattr avers +ĠA ph +Ġ' );čĊ +ünd e +Ġench anted +ĠRoot State +ĠCLOSE D +ĉĉĉĉĉĉĉĉ čĊ +Ġcal iente +or ris +Ġphysic ists +h wnd +_v i +Ġráp ido +Ġcapital ized +ed By +Ġmach ining +Ġhub by +ĠSt acy +.B us +dr ink +H ur +Ġprop ia +Unit Test +Ġmiscon ception +__ ));Ċ +/d c +ĠMay weather +_m C +.create From +ĠQ Painter +rops ych +inn itus +ay as +Ġg eg +(d w +Ġus ado +Ġtrick le +Ġann ihil +ĠP asta +Ġ++ Ċ +(Expected Conditions +.post Value +ic ap +ĠDon etsk +_s oup +-p ublish +ĠP b +ment ions +AC CEPT +.P ull +,âĢĻ âĢĻ +Ġret arded +_AT OM +ĠTermin ator +-c ourt +ĠCLLocation Coordinate +Ġrever ence +ĠS SC +ut ely +ĠW ON +ĠG SL +fre i +.get Longitude +Ġopen FileDialog +.B utter +- important +_M ANY +ĠG ong +âĢľ How +Ġg orge += msg +ĠEz ek +create Command +: checked +Ġinf ographic +.W EST +Dir s +Ġguard a +Ġbeet le +< small +- android +Ġcred itor +ĠM éd +Ġfinal ist +Ġab l +ne v +_inter action +ĠMonter ey +j ah +Ġcand ies +ĠQu incy +èª Ń +Ġbatch Size +ak it +Ġo be +(p ara +Ġexperiment ed +Ġcouncill ors +Ġcl ashed +s qu +-st rokes +ĠG K +ĠEx pires +Ġprosec utions +ĠCreat ures +Ġy ö +x lim +_IM P +Entry Point +ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ +.Default CellStyle +Ġbre ve +ĠBrit ann +Ġsweat y +Ġle th +Ġflash back +per manent +ĠJ DK +_D etails +E uro +p pt +Ġrich TextBox +/ board +Ġtr ance +.c ycle +'); ");Ċ +Ġtox in +_de init +Ġover arching +Ġconfig parser +ĠKaw asaki +.th umb +Ġplay a +ĠJose f ++ _ +Ġzero es +Ġa up +ĠH ari +comm itted +N it +.file Path +ĠDis abilities +man ufact +-al igned +.RE SET +Ġrust y +E y +Ġou sted +cos a +Struct ured +.get D +Ġs ábado +> Loading +_m A +.get Random +bl ings +Ġchees es +tt i +. âĢ¢ +ĠBurg ess +ender it +. ',čĊ +(" "+ +ac b +% p +index ed +_pred icate +nes ia +Ġb ied +ĠC IT +( Pos +_r adi +ä»· æł¼ +B iz +ĠAdoles cent +Ġvi ên +c ycl +_C ancel +Ġcon clusive +Ġappell ate +inform atics +S J +Ġelect ive +role Id +Fetch er +ĉ Command +(" (% +Ġf art +IL A +get Block +A USE +Ġд ан +ĠAr te +Ġnot ifying +Ġge le +.s ame +ĠReg el +ĠBa ÅŁ +.c reation +ĠV N +_comm unity +Ġuns ustainable +SE X +Ġgrid Size +res cia +avers able +(', ')[ +ĠPh elps +á»ķ i +ANCE LED +- IS +.run ners +ĠSt okes +.P rodu +Ġwh ipping +_ac quire +Ġinvestig ación +f ried +.copy With +ĠHard cover +- Se +áŀ¶ áŀ +inv itation +les ai +ĠD orm +ĠÑģпиÑģ ка +Ġconcaten ated +oph il +Ġthink er +/font awesome +ĠLe opard +Ġ"/ ");Ċ +Ġresidual s +ĠMic rowave +Ġconform e +th rop +Ġdis emb +ĠO MG +ĠDisc ipline +ĠAc robat +/re pository +df a +_M ED +buf io +Ġméth ode +_H OLD +ias i +_ legacy +) ččĊ +æ£ Ģ +Get ProcAddress +Ġy ay +ot ence +order id +-t w +Ġdear ly +In coming +/ il +Ġneu rop +uc z +); čččĊ +ĠInnov ative +Ġprof und +ig mat +Selection Mode +re levant +.G O +Ġbru ises +Ġs ach +ode f +Ġre imb +/d esktop +-s pot +und ance +Ent ropy +\ core +Ġsug er +ĠM vc +ĠGN OME +_ind x +ĠYY STYPE +ĠMat lab +ĠC IF +Ġ* )) +Ġproduct List +ĠAl right +ac emark +ÑĤи в +mod ification +int ernational +Ġhom ers +Ġdict s +ĠQ Font +.SQL ite +Ġtransplant ation +ĠMessageBox Button +ĠEl ves +'] ])Ċ +(Q Icon +Ġcin emas +CO ORD +- China +Ġkh ẩu +æĪij çļĦ +Ġskull s +Ġpain staking +f ce +.XR Label +Ġspec ifier +Ġpref erring +/ activity +( Photo +á lt +.l ot +' '. +ann once +.google code +-p df +ĠP oke +_A CL +Ġend owed +dis cover +.om g +Ġwood land +.M agic +Ġvol ont +Not Allowed +Ġch ave +BM W +',' =', +ĠS IX +æĪij 们 +Ġkos her +Ġaspir ation +int l +_ref ptr +'+ Ċ +ment or +.cl ub +Window State +.A RR +Ġz za +Ġmessage Type +.e qu +Th or +Ġin just +Ġg ums +Ġborder Side +//// / +ĠTrans mit +Ġbuf size +Ġh ak +Ġell as +R ANDOM +ĉm c +Ġpe a +ek o +document o +Ġhyster ia +Ġaren as +Ġgun men +Ġm ike +Ġimp unity +atis ation +_Z ero +_COMP ANY +ĠG ors +Ġuse Class +( redis +ĠRUN NING +ĠB air +vel te +Ġ',' . +аÑĤÑĮ ÑģÑı +ö st +encode URIComponent +_re strict +Ġdec als +ĠPed ido +Ġalter cation +Dis plays +ĠApp licants +C US +Text area +ĠAng ola +.f uture +ĠUS HORT +Ġsuppress ing +Ġset zen +AP olynomial +Ġto ch +Ġhall mark +Ġ$ $$ +ĠCHAR SET +.r pm +ĠD ich +---------------- ---- +_p arm +è¿ ĺ +acc iones +h ait +WAR DED +_r outing +ĠN OM +Ġen clave +ĠLot to +ĉf r +complex Content +ĠBall ard +k ube +/w in +.getColumn Model +_RE PLACE +Header Value +Ġest udiantes +Ġap is +Ġb pm +ĠType Name +And Get +rit a +Pl ans +> Note +Ġfet isch +Ġton ed +_g oto +ons ense +Ġm olds +Ġinfiltr ation +ĠGuerr ero +ub bo +ck i +($ (". +_ activities +(ch anges +Ġof App +ĠKe pler +ĠD emp +ĠCont inent +.T icks +ĠUn signed +ĠJah res +Ġfresh men +ĠArch ived +ĠкоÑĤоÑĢ Ñĭй +Ġ' :: +T utorial +C c +Ġtable LayoutPanel +from Json +.level s +_trans ient +Ġendors ing +ĠD IC +la uf +Ġsh red +_E MIT +ific antly +AL A +/ proto +Ġnarrow ing +U tc +Fact ors +Ġsent ient +æŀ IJ +lix ir +ĠC ROSS +met eor +Ġgro in +Ġm db +ĠRot terdam +Ġcom ida +ĠOp Code +ĠDefault Value +Permissions Result +Ġheter ogeneous +Ġm oot +Ġde ceived +-in dependent +ĠObject OutputStream +Ġover power +.d up +Ġl db +Ġdomest ically +Ġbest ellen +Ġlo v +ĠContract ors +Tri angles +Ġfod der +Ġfilm es +ä¼ ģ +Ġrev olver +Startup Script +/ validation +ĠResource Type +i ÅŁ +ĠL az +f ef +Ġlst m +{ * +. attachment +.h its +ew ith +DO G +Al abama +Ġmedium s +.m Context +-c ols +åı ĭ +.not ice +Ġat tn +ĠP acking +ĠL n +_COM PLEX +/ Users +.sav etxt +ĠR ounds +?,?, ?,?, +Ġing l +ĠR OC +_f emale +ĠSt ard +]] ; +Ġwrest lers +Ġtorrent s +Ġsin h + ĊĊ +ë³ µ +s ense +how ever +.Ph ysics +Inf rastructure +ĠSac r +F el +ĠD ISTRIBUT +é ments +ĠValid ates +################################################ ############ +Ġ| / +Ġes l +Ġré seau +ĠB ip +BY TES +_W ATER +Turn ing +EL S +Ġj uxtap +Ġlesb ische +ý ch +( Unknown +Ne o +@ JsonProperty +Ġal umnos +ĠRaq qa +ime i +.get Bounds +.Mouse EventHandler +#### ### +Generic Type +/c ms +Ġturn o +Ġм ин +Ġfolk lore +ĠE vo +Ġconduct ivity +Ġle ben +Ġgear box +-v s +ĠÏ Ĩ +Ġdrink ers +Ġcon exao +ĠTe eth +Ġget Arguments +ĠR AT +ent ious +E duc ++ W +ĠInstitution al +ĠB ord +is Equal +(p wd +Ġign ited +ĠR ousse +Ġimpact ful +ĠM alk +Ġg eral +ĠP ivot +Ġa zt +Ġcsv file +ĠR ope +ĠSOL UTION +ĠArbit rary +Ġlet to +.Mouse Adapter +Ġ} }} +ĠSail or +der a +Put ting +Ġconcentr ates +Ġauth Domain +âĢĿ çļĦ +-f inals +, strlen +Mu on +ĠOrd inary +fire fox +ĠLa TeX +ĠH und +engine ering +/ blue +ed TextBox +(" "); +ĠC DDL +ke pt +ĠGet String +K ir +() =' +ĠO CD +ant ium +$ menu +ĠAppalach ian +Secret ary +ë¥ ĺ +ี ย +Sem antic +Ġ* [ +est one +ung kin +Max Y +-t one +"} ;čĊ +_P art +< Member +tr am +Ġtrans istor +Ġ---------------------------------------------------------------- ----------Ċ +ĠDes de +Ġright ful +ĠCorn el +æ ij +.H OUR +Ġsidel ined +ref errer +m aze +Ġhol ster +Ġcripp led +ĠDate Formatter +oph age +_m D +Ġdes elect +ra ud +ĠPK K +row Data +Ġlock smith +.res ponses +(product Id +_ST MT +Key Type +.Th en +z ee +Ġcr t +ĠGrand ma +@ Resource +Ġbit wise +-c mpr +ãĢĤ www +zeit ig +& display +Cart Item +- No +Ġnum éro +Ġm aur +Ġinst ancia +ĉd t +_n pc +Ġskate board +âĢľ All +ĠCrow d +Ġä n +Ġb raz +ca e +yn et +/p m +/s creen +OPT ARG +ĠV Box +Ġle opard +_g reater +c pt +< dd +Ġmechan ically +osp els +) f +.l wjgl +.get Port +ĠP REF +.Add Transient +pp ard +Ġí ļĮ +Ether net +Ġsal ine +(level s +Ġservice Provider +.A ngle +alt itude +illa ume +Ġs cape +_CAL C +_ quest +ĠDiss ertation +ĠE DM +-C ds +Ġhon orary +st ops +Ġsub dir +ĠV H +ĠChe at +Ġright fully +Q E +.Write Byte +fig ures +enn ie +( DBG +Ġvoks ne +Ġexp ended +UN ICATION +il inx +ĠRec ap +_ verts +Ġtra umat +Ġget Player +Ġverb ess +Ġcultiv ating +Ġiniti ator +Th ông +find First +_per ms +Ġbu c +Ġ""" čĊčĊ +T YPES +object Manager +(Configuration Manager +Ġtim id +Ġsnap chat +Ġcon seg +ĉd istance +_right s +_D es +ĠF lesh +- ver +Ġa fl +fra uen +Ġblas ph +ĠQual ität +ma f +Monitor ing +.D iff +Ġshore line +Ġresponse Body +mem set +< decimal +Smarty HeaderCode +Ġin sets +ĠBinary Tree +amed a +Ġn ihil +ĠN ay +ym ology +ĠW G +Ġt api +ĠInst alled +m aintenance +)} "Ċ +ĠX O +-per iod +s ar +Ġning una +ORM AT +.set PrototypeOf +ĠK b +ĠHen rik +ét ique +ĠLah ore +ĉ Address +Ġmel ts +N y +_adv ance +Ġveloc idad +Ġalum no +Ġsanit izer +Ġph ishing +ĠCom et +Ġch iar +ĉs pec +trim med +(state arr +on nen +Re venue +L ens +Ġcha ired +ĠAss umes +Tr ash +_un set +\ Bridge +Point Size +ĠPol ic +Ġsex uales +ĉd fs +ĠWide String +Ġaccru ed +Y W +_S CHEDULE +Ġk ite +Ġparach ute +[ table +Ġactive ClassName +.Qu ad +Israel i +ĠÅ ĵ +Ġho og +Ġch á»ī +ew ear +Ġtire lessly +set Error +.get Amount +.set Items +ĠM anson +ĠBay esian +_F lag +AC HER +/ original +Ġimm ac +ĠLos ing +' >ĊĊ +L ic +ĠMir age +ĠAssembly FileVersion +Te V +ĠValue EventListener +-s olving +Th o +rou lette +_W P +Ġunint errupted +Ġfield Type +.T yped +Ġam our +Ġmock ery +(v ol +ĠSub committee +ĠR uf +ero x +:UIButtonType Custom +ĠBl ur +Ġwy kon +nc es +ASH BOARD +!! ");Ċ +Ġmurder ers +.d aily +ĠDI AG +j ing +Ġdol phin +Ġl òng +Ġb ö +ĠV ocabulary +.St Object +') "> +Ġz un +Ġscrim mage +tr éal +ĠL ig +[ vi +C ole +Ġfrost ing +.Pl ayers +- translate +Fe els +=\" / +.Butter Knife +Ġ?> ;Ċ +Ġav i +inn ie +.F ailure +Ġsp indle +Configuration Exception +_h op +Ġpos ição +ĠA wait +UIImage PickerController +ĉ day +Ġgen om +C ab +ĠÑĢ ÐµÐ·ÑĥлÑĮÑĤаÑĤ +OR IGINAL +Ġejac ulation +(t cp +SE COND +Ġton ic +ĠList Box +Ġ ĉĉĊ +() >Ċ +Ġqu atre +ượ ng +with Errors +.M aybe +, â̦ +token Id +_UN DEF +Ġfresh ness +ĠAmend ments +.map box +.C V +(b log +_get time +. quest +s parse +Ġres ale +Ġenthusi astically +ĠProstit utas +W a +C argo +.Parcel able +SENS OR +ĠRy u +La ughs +_N ative +/ pg +yst s +Ġphot oc +ç® Ģ +ado pt +.spec ies +conc iliation +Adjust ed +.Firebase Auth +ut tle +ord ination +Ġm unch +ĠSt ake +.p ing +ank er +(QString Literal +Ġsub script +ĠĠ ĉĊ +ĠM CC +_C md +se xy +i ou +ĠM ANY +Ġn anny +TR AIN +Ġflour ishing +ĠW atches +ĠQ Map +ĠF erm +Ġwas m +ĠA bed +_ UD +ĠGlass es ++ v +Att end +.Ch ain +Ġdec ency +ĠSupplement ary +h unter +-t xt +Ġ" }";Ċ +.set WindowTitle +(" +Ġmasc ara +( Profile +åĬŁ èĥ½ +imit é +Ġwild fires +- ROM +.is On +(group Id +Re pair +accum ulate +Ġ< ", +Ġhand written +Ġach eter +ĠM GM +ĠIr ma +->{ _ +ge e +cr iminal +Ġèĭ¥ è¦ģ +Ġmoment arily +") != +_l it +Ġexpires In +." ). +éķ¿ åº¦ +Ġfr ække +vl c +Ġor bs +), $ +Ġvent ured +/ >\ +char m +N uitka +eld ig +aton in +W itness +-l at +Ġset Hidden +Ġrelic s +Ġcons ulate +. IGNORE +" After +Ġset Address +Ġbeste ht +Ġ'' )ĊĊ +.x axis +Ġser ão +Ġmis led +_UN IFORM +ĠV IA +inc r +Ġzen ith +Ġvis cosity +Ġthin ly +.get SharedPreferences +.Error Code +"), " +ĠMillion en +Ġ/> )Ċ +Scroll Indicator +-se eking +ĠPOLIT ICO +as ca +_r l +N avig +(full file +Ġsol itude +Ġju ven +Ġhaul ing +ĠMac ros +ĠG ry +Ġexerc itation +ĠATT ACK +Tick Count +Ġr ites +Ġdo e +Particle System +Ġsl u +Window Text +ĠClass Name +Ġsl ander +ĉ Port +j ong +? a +.D ial +âĢĶ at +$obj PHPExcel +Ġso ar +EN N +appe ared +Ġquot id +em achine +Ġn ip +Ġmicro time +ĠAl ma +; ! +---------------------------------------------------------------- -------------------------------- +ĠPass age +Ġdump sters +ĠEx clude +Ġsuggest ive +ĠCircularProgress Indicator +_cl r +Array Type +ILL A +Elapsed Time +Dr iven +Ġresource Name +ĠG arrison +ser ir +-a head +Ġp innacle +ĠEs presso +S parse +Ġass ays +ĠGirl friend +im id +]=' \ +ONGL ONG +Ġportray ing +L ane +Ġb úsqueda +Ġrein forcements +ĠSpread sheet +ĠArray Collection +, arr +light box +ic ana +< " +build ers +K id +ĠMat SnackBar +EX PR +od cast +ĠFound ations +Ġind s +=' ${ +F izz +-function al +(work space +Ġstem med +_p atches +ĠJar vis +READ ING +Ġdisrespect ful +ĠQ Dom +Ġ$ {Ċ +est atus +Re ached +! .ĊĊ +IL T +ĠN DEBUG +ĠCour age +birth date +ĠT ing +Ġutil izado +án chez +Out door +Ġhand guns +Ref Count +É Ļ +rom o +Ġt ts +.S he +ĠP ane +ãĢij, ãĢIJ +ĠIO CTL +/ black +ins cription +Ġbi opsy +ĠTime Interval +.Test Check +ĠGUI Style +ĠCap ability +ĠBeit rag +don nees +T reatment +.back up +Ġsign ings +ĠB oca +dr m +.M AIN +Ġgo ede +ĠMark up +G REE +ĠBase Service +.C reator +Ġj ails +ĠK ahn +Ip Address +ACH I +Ġinhib ited +Ġ@ $_ +ĠAss ass +Ġenvi ado +Hero es +ÐŁ еÑĢ +ĠM aven +.l s +Ġ ive +| RF +Ġresize Mode +Ġrum pe +_attach ments +T U +Ġtact ile +Attempt ing +Ġro bin +y aw +Ġmerc enaries +ĠHab itat +end date +Ġo xy +ĉR andom +oh on +Is Null +ĠValidation Result +ãĥ ļ +um bed +pp v +Ġar p +ich ick +_r nn +ĠT FT +Tex Image +" On +ĠSam pler +top l +Ġj ane +y ling +ĠUN ICODE +Tab Index +< {Ċ +s uspend +uv ian +, application +ол иÑĩеÑģÑĤво +y at +ez ier +ĠCH UNK +ĠAd ler +/ Add +ĠKey Value +Ġspos ób +Sam pling +ch ers +_AM D +R u +.Must Compile +N ation +Ass oc +Man aging +ĠEng l +_G B +Ġsucc inct +Ġdis liked +ĠI ke +Bullet in +_ARCH IVE +Prop osal +Ġjog ging +.C REATED +Ġch ol +è£ ħ +Į ¨ +-p ush +Ġreserv a +core v +è tre +TH R +Ġincompet ence +Ġchar isma +æĦ Ł +Ġ" == +BT N +ĠLoc ator +iv et +('. ')Ċ +Ġfor IndexPath +ô me +Ġcapac it +w aters +ĠWR ONG +ho a +ĠM IPS +Ġem iss +ĠJacqu eline +(c mp +Ġe ens +Le o +.tim ing +CLUS ION +Ġ(" - +åĵ Ī +.k ode +ĠUnd ert +Ġbew ild +ĠEss en +.h d +Ġren egot +Ġm ower +Ġl sp +Ġpen chant +Ġman oe +Ġag li +Ġrec al +ĠOPER ATION +(^ )( +ĠÎ ½ +ĠSc oped +Ġ@ "Ċ += label +[ loc +Int l +ĠN z +table t +.Column Name +Ġscreen Size +DB us +co oked +- registration +âĢľ One +-n on +ĠwiÄĻ c +Ġcost a +.add Tab +. conditions +ĠH ess +MEM ORY +ĠAval anche +() }}Ċ +Ġtri plet +Ġl abyrinth +ĠNode List +ĠNY T +Ġy eni +d ff +.Html Controls +AV IS +/ Math +Ġmem cmp +Ø§Ø ¡ +оÑģ ÑĮ +c rap +(p ages +Ġl xml +ĠQ DateTime +_t cb +Ġopen id +Ġsyn aptic +ĠMD MA +(s lug +igm atic +en or +Ġcr amped +G OP +Ń IJ +.is File +ĠD ifferential +Ġ=" ";Ċ +ĉĉĉ ĠĠĠĠĉ +ĠC ooke +ĉU FUNCTION +Ġpersever ance +Relative Layout +IMPORT ANT +Ġex on +Ġо н +ib ase +(C ONT +n ovation +ä½ ķ +[ sub +Admin Controller +HTTP Header +cre ar +ĠN IR +ĠDrop DownList +Ġval ide +Ġde hydration +. '] +(W IN +Ġ... \ +Ġphotos hop +ĉ Init +_c ou +Ġtime Zone +dar win +rom atic +Navigation ItemSelectedListener +br ates +] --;Ċ +Ġtraged ies +ĠPed iatrics +SM ART +-A PI +ĠMessage Lookup +ĉ vo +Ġprejud ices +Ġm A +U ps +ĠMISS ING +ĉ ad +C ream +ĠT b +ĠMon a +_ ghost +ĉt ypes +Em b +ĠDocument ary +');ĊĊ ĊĊ +Ġl up +_ Reference +ĠB ATCH +Ġintertw ined +< Cell +ĠCab r +n ation +Ġis Connected +.remove Listener +Ġcon g +_t i +ĠSil icone +Ġê²° ê³¼ +ĠW AN +ĠG ibraltar +/ response +ĉp erson +ch ants +V IP +em ergency +Pixel Format +- Am +Ġsouth western +_pl l +if ers +_ON CE +ĠF ayette +.nc bi +_P anel +.Q ual +Ġpol ys +Ġcreate StackNavigator +� t +Ġlay offs +ĠBl anco +Fe at +ĠV imeo +_ch i +_l ifetime +POINT S +, private +Ġunb earable +print ing +Ġc gi +.B ACK +Ġintern s +ĠNew ly +inf eld +( IB +ĠK ata +ĠDef endants +Th r +é¢ Ħ +_V F +FFFF FFFF +Ġdavid jl +Ġbitter ly +S uggestions +.set Cancelable +FIN AL +ason s +_rw lock +_WRAP PER +Ġhapp iest +(row Index +ós ito +TOT YPE +Autom ation +Log File +Ġcons olation +ãĥ Ģ +Ġt êm +Ġpr er +rg yz +ĠG eg +ĉd to +.default Value +ĠK ami +ĠA SE +optim ized +Ġíı ¬ +Ġorigin ates +err Msg +Ġespa ço +(S YS +ĠMc B +d ance +_det ected +Ġfr ü +ĉĉ ĠĠĠĠĉĉ +< Date +(com b +ĠDec ide +\ Field +ĠProp osed +R ib +Ġdis likes +ĠW ien +ĉ Document +Ġtr af +Ġst oria +ĠT ells +') == +C ri +( VALUE +ĠBurn ett +, void +Ġdan h +Ġc cp +Block chain +:"- "`Ċ +IC lient +IS ODE +Iss uer +) }čĊ +, but +ĠU ph +( Sub +Ġtélé phone +ĠonData Change +Ġmarsh aller +-an alytics +, content +Ġdeb acle +_Value Changed +Ġfa una +Ġ# => +Ġf oyer +'util isation +ĠMü ller +ĠFet ish +Ġdefault Manager +Ġback track +B ah +Exp licit +_A SCII +Ġm Activity +(M sg +Ġê² Į +ĠTER MS +ĠAng ie +HS V +ĠMos que +.N ames +íĬ ¼ +rest e +_p arms +Ġgap ing +Ġcro pping +Data Frame +Ġrespons iveness +_ undo +_tr an +. terminate +Ġitalian e +Ġwalk through +Ġattract iveness +д е +_ST S +_ learn +Ġchocol ates +ier archical +-th inking +Ġ ))) +ish ments +.Log f +ĠTM Z +ĠCan ary +fo il +ĠVacc ine +.v x +ĠSur round +Inter mediate +Ġi ov +v ais +'; ";Ċ +ï½ŀ ĊĊ +éĢģ æĸĻ +â̦ it +Se ats +Cl ar +W ars +ĠHutch inson +ĠHas an +! ')ĊĊ +ĠRich ie +che iden +($ (' +Y ork +Ġl ids +Ġal phanumeric +ĠG lock +.sh apes +Ġspark ing +_ epsilon +uplic ated +.dir ty +]) == +ĠìľĦ ì¹ĺ +Ġsc n +Ġ/ **************************************************************** +_PRE VIEW +_H C +ield ing +f gets +ĠAdd ison +Ġproduct Service +- figure +(ret val +z ano +Ġaut ob +ĉs d +_n umer +ĠSet LastError +ĠF ior +ific ance +Unt itled +Ġin field +Ġ{} ));Ċ +Ġsp ac +Ġro okies +(des cribing +ng en +ி à® +.r df +.M utex +Ġkne eling +ĠQ E +set Max +Read Stream +Ġvent as +s ut +cm peq +.WriteAll Text +ĠEx perienced +$ __ +Ġka um +ĠL IS +Ġdocument os +_HE ALTH +icont ains +Ġart isans +OWN ER +Ġblink ed +get Display +Ġto en +Ġrow Num +Ġav ril +Ġinv is +ĠK ear +toBe InTheDocument +ap ur +Ġr acked +ĠMc Master +_ATTR IB +H az +Ġfact ura +/ ts +ĠÑĢаз меÑĢ +Ġz f +Ġshort fall +.f asta +ĠCONST ANT +.man aged +g ems +Shared Pointer +Ġblur ry +b rightness +( components +Ġ... "ĊĊ +SE LL +ĠIllustr ator +.get Channel +Ġtrou vé +yst ers +Ġvo is +ĠLind en +Ġem ojis +Ġb rawl +ĠMS R +ĠE lo +ĠCroat ian +Popup Menu +L ewis +.J WT +Ġaston ished +B ush +(item Id +Ġdet achment +ĠEnc ore +å° Ķ +Ġre kl +Ġcr am +)$ / +.get Host +_re commend +- HT +_cal ibration +Auth enticate +.firebase app +UN IX +ĉC amera +ĠHE AP +I deal +. office +Ġgoof y +(S ymbol +Ġjou er +_part itions +Ġrapid ement +ĠGN UNET +id User +Ġsuperv ise +( Contact +AW N +ãģ ĺ +Ġna am +Ġa ust +åľ¨ 线 +_soft max +Allow Anonymous +amm able +RO UTE +* D +Ġad en +ĠCrist ina +ĠCrist iano +Ġblood stream +sub class +_person a +CH ILD +-k now +Ġnavigation Options +ĠZuk unft +ĠPix ar +Ty ler +Ġunder world +Ġsincer ity +Ġdispens er +Ġk ter +idd ers +.add Node +- checked +Ġke yst +ĠW TO +.sign als +Ġadvent urer +ĠP ang +\ R += pos +Ġdispens aries +ĠClo set +("{ \" +ide on +Ġnécess aire +() "Ċ +_RECE IVED +Ġrésult ats +Ġmod en +ĠIceland ic +; d +. allowed +(new User +Ġmerc iless +.Wait For +Ġday care +ĠCon veyor +ç ĸ +ð ¬ +ç ĥ +ç Ĺ +ç ł +è Ħ +é ² +å ¦ +çĿ Ģ +å¾ Ī +é ħ +ç ĭ +é ª +æ Ĥ +é ¥ +è ħ +æĥ ³ +å ¨ +é ¹ +ç Ĥ +å Ĵ +ç Į +è´ ¨ +æ ¢ +æ° Ķ +ð « +æķ Ļ +ç Ł +å Ħ +åıij å±ķ +åĪ Ľ +è ij +æ ħ +å ŀ +åģ ļ +æĪ ĺ +æ IJ +å¼ º +æ· ± +åĩ ł +ç ¿ +å © +è ŀ +å§ Ķ +åIJ Ħ +è İ +é ¸ +é º +åı Ĺ +èģ Į +å ĺ +æ ½ +é£ İ +èIJ ¥ +åħ ļ +è ľ +éĤ £ +é¢ Ĩ +ç ij +é ³ +æľ ¯ +ä» Ģ +æĪ ¿ +ç² ¾ +å ª +é Ĩ +å¤ ª +èĤ ¡ +è Ľ +åħ ī +æŀ ģ +åĬ ŀ +è ĵ +ç ĺ +å ´ +å Ĺ +èĬ ± +çł Ķ +å¿ « +å¸ Ī +è¶ Ĭ +è§ Ĥ +æ ¤ +æ ¦ +ç ŀ +èĤ ² +çĪ ± +çĻ ½ +ä¸ ĸ +ä»Ģ ä¹Ī +çľ ¼ +å ³ +è Ĵ +æ ĵ +è¢ « +å¹ ² +çĹ ħ +å£ « +ç Ĵ +è ¸ +æ ¾ +å·¥ ä½ľ +è® © +çĥ Ń +è¾ ĥ +åĦ ¿ +åĬ © +ç§ ¯ +ç ³ +ç ĵ +ç £ +å Ĥ +è ¹ +è ļ +å· ± +çĻ ¾ +åĬ ¿ +èµ Ľ +æ ¨ +æ ¿ +è ĸ +æĿ ij +å¸ ¦ +å¢ ĥ +æĬ ¤ +é Ń +å « +èĩª å·± +æµ İ +ä½ İ +åĮ » +éĺ ² +åĨ ľ +è Ĩ +ç Ĩ +é « +åĨ Ľ +æĪ ı +åį ĩ +æĸ ¯ +ä½ ı +èIJ ½ +åħ » +èĩ ´ +ç Ĭ +ç ĩ +ç ħ +è Ķ +ä¼ģ ä¸ļ +åĽ ¢ +æī į +æł ¡ +åĩ Ĩ +å¥ ĩ +åī ¯ +é ¼ +æ¼ Ķ +é© ¬ +èµ ° +ç¥ ŀ +åħ ĭ +æľ Ľ +æ² ¹ +è¾ ¹ +åį ĥ +å¾ Ģ +åĪ ĩ +æ © +ç ¶ +å Ļ +éĻ ħ +çī Į +社 ä¼ļ +游 æĪı +æĸ ½ +ç ħ§ +æİ § +æ» ¡ +è¯ Ĩ +éĩį è¦ģ +è¶ ³ +çķ Ļ +ç» Ĩ +åį ı +éĢ Ĥ +æ ĩ +æ § +é Ħ +è Ŀ +å¸Ĥ åľº +ç»ı æµİ +ä¹ ł +æĸĩ åĮĸ +éļ ¾ +ä¹ IJ +åĨ ³ +æ¬ ¢ +è§ ī +åĽ Ń +åħ ´ +åħ ħ +ä¸ ¾ +æī ¹ +è ķ +æĬ Ĭ +æĬĢ æľ¯ +ç© ¶ +第 ä¸Ģ +ä¾ ¿ +åĵ į +çİ © +åĿ ļ +èŀ į +åį Ĭ +åĸ ľ +å± Ĥ +ç¦ » +ä» ħ +é Ł +åij ³ +å¿ µ +åŃ £ +ç´ § +ä¹ ħ +é ¤ +é ŀ +è ¤ +åĢ Ļ +åĨ µ +ç Ł³ +åģ ¥ +æĢ İ +å® Ŀ +è¡ Ģ +åŁ Ł +æĹ © +çŁ¥ éģĵ +è´ Ł +åį ļ +å· ´ +äº ² +å± ŀ +ä¸ ¥ +äº ī +å¯ Ł +è º +ç ° +建 设 +产 ä¸ļ +åIJ ĥ +åŃ © +æĹ ħ +æł ¹ +æĿ IJ +ä¼ Ĺ +éļ ı +å® ĺ +åº ķ +å½ © +å¯ Į +æ¸ © +åį « +åī § +çĽ Ĭ +æĬ Ĺ +è´ ¢ +çº ª +æ Ĩ +çĶŁ æ´» +çº ¢ +çĶŁ 产 +è¿ ľ +éĴ ± +åĶ ® +ç¾ ¤ +çı Ń +æ¥ ¼ +éĩ ĩ +èī º +å± ħ +åģ ĩ +è° Ī +æĻ ļ +é ¬ +èĪ ª +å® ³ +è Ĺ +ç į +å µ +çİ ĭ +åº · +è İ· +ç» Ń +äº ļ +é£ Ł +åİ ĭ +æĭ Ľ +èĮ ĥ +è® ¸ +åĽ ´ +é ½ +éĻ į +çº ³ +åĵ ª +æķĻ èĤ² +å·² ç»ı +å¾ · +æŀ Ĺ +å®ī åħ¨ +é¾ Ļ +大 å®¶ +éĿ Ĵ +åº ľ +æ² ³ +åı ¤ +èį ¯ +åĿ ĩ +æĻ º +ä¹ ¡ +çķ ¥ +åĨ · +ç¦ ı +å® ¤ +ç» ´ +æī ¿ +å± Ĭ +è¯ ī +åĪ » +è Ł +æ ª +å°± æĺ¯ +è¿Ļ 个 +ä¸Ń å¿ĥ +ä¸ĸ çķĮ +åŁİ å¸Ĥ +éĿŀ 常 +åĪ Ĵ +åı Į +æĢİ ä¹Ī +åΰ äºĨ +æľ ĥ +åı ² +ä¾ Ĩ +å¾ ĭ +å¥ ĸ +ç» Ī +åª Ĵ +å® ģ +è¯ ¾ +èģĮ ä¸ļ +åħ į +æµ ĭ +æĢ ¥ +æķ ij +çĭ ¬ +èŃ ¦ +é¤ IJ +æĦ ¿ +è´ « +çĸ ij +å ļ +å¥ ¹ +åı Ī +åĽł 为 +ä¸į æĺ¯ +å¤ Ł +æĸ¹ éĿ¢ +éķ ĩ +äº Ĵ +éħ Ĵ +è® ² +çĸ Ĺ +æĺ ¥ +æ¹ ĸ +å¤ ľ +è´£ ä»» +人 æ°ij +åħ ° +çŁ Ń +æķ ħ +åĩ ı +æĻ ® +äº ® +ä¾ Ŀ +åį ° +éĿ Ļ +åĢ ĭ +å¾ ģ +åIJ ¸ +ç¼ º +æĶ » +åĩ Ģ +åħ ¸ +åĽ º +è® ¿ +ç ¹ +ç Ģ +æıIJ ä¾Ľ +ç» ĩ +å¾Ī å¤ļ +çłĶ ç©¶ +è· Ł +主 è¦ģ +æĥħ åĨµ +çŃ ĸ +æŃ » +大 åѦ +æĶ¿ åºľ +å½± åĵį +ä¹ ° +åħ Ń +éĻ © +åħ « +æŁ IJ +è´¨ éĩı +åį ł +å· ® +æĽ´ å¤ļ +æľ ĭ +éĿ © +å® £ +çł ´ +è½ » +åº § +æĺ ¾ +ç¨ ³ +è´ µ +èĥ Į +èī ¯ +çĸ « +æ¯ Ĵ +ä¹ İ +åĢ Ł +è¿ · +çŃ Ķ +æ¿ Ģ +åij ¼ +äºĨ ä¸Ģ +è¶ £ +ä¼ ´ +ä¼ Ļ +è ¼ +ð¬ Ń +åĽ½ å®¶ +æ´» åĬ¨ +çݰ åľ¨ +ç§ij æĬĢ +åį ¡ +ä¸į åIJĮ +个 人 +è®° èĢħ +ä¸į æĸŃ +éĹ » +ä¹ Ŀ +èij Ĺ +ç» ¼ +ä¸ ĥ +æł ij +æľĭ åıĭ +åį ĸ +ä¼ ¤ +æ² Ļ +åĸ Ħ +å¥ Ĺ +è½ ® +ç© ¿ +è¡ ¥ +ä¸Ģ å®ļ +çª ģ +çĿ £ +è¿ ½ +å¨ ģ +åı ¦ +åĽ ° +æŀ ¶ +ç» Ŀ +æķ £ +æİ ¢ +æ´ Ĺ +ä¸ ´ +ä¼ ¼ +è´ ¸ +ä¸ ° +æĺ¯ ä¸Ģ +ç« ŀ +è¿ İ +èģ ļ +è « +æį Ł +æī § +é© ¾ +è¿ Ŀ +è ¥ +è ł +ä»ĸ 们 +æĹ¶ åĢĻ +å® ĥ +人 åijĺ +è¿Ļ æł· +å·¥ ç¨ĭ +åĪĽ æĸ° +åŃ© åŃIJ +å¸ Į +éĥ¨ åĪĨ +éĵ ¶ +代 表 +é¦ Ļ +å¸ ® +æİ¨ è¿Ľ +çĽ ĺ +积 æŀģ +éĥ¨ éŨ +åŁ ¹ +æŃ ¦ +ä¸į ä¼ļ +çŃ ij +éĢ Ļ +çİ© å®¶ +æĭ ¿ +åİ Ĥ +æ¯ Ľ +çģ µ +æŃ Į +ç »¿ +å¦ Ī +çĽ Ľ +é¦ Ĩ +é¡ º +èĦ ¸ +å° ¼ +ä¸ ½ +å¥ ¥ +éģ ĩ +è¯ į +å° ģ +ä¸ Ŀ +好 çļĦ +æĭ ħ +èĦ ± +æģ ¶ +åİ ļ +åĬ ³ +çĽ Ł +æĬ ĺ +åı ¥ +æĢ Ģ +æŁ ĵ +书 è®° +åĨ ł +é² ľ +æ ¦Ĥ +éļ IJ +å¹ ħ +èµ ŀ +å¹ ķ +æ¥ Ń +éģ Ĺ +åĪ ¤ +è ĺ +å ¶ +æĬķ èµĦ +è¡Į ä¸ļ +äº ij +çݯ å¢ĥ +åѦ çĶŁ +åIJĪ ä½ľ +åģ¥ åº· +é£ ŀ +ä¸Ģ æŃ¥ +ä¸Ģ 缴 +åıij çĶŁ +éĺ ¿ +é¢Ĩ 导 +åĸľ 欢 +åºĶ 该 +çĤ º +è® Ń +æĿ Ģ +æ¸ ¯ +交 éĢļ +éĺ ¶ +éĴ ¢ +ä» ¤ +å° ½ +æ¯ į +è¡ £ +ç² ī +é¡ ¶ +ä¹Ł ä¸į +æĬ ĵ +èĭ ¦ +å¹ ¸ +ç¤ ¼ +第 ä¸ī +大 çļĦ +éģ İ +çĥ Ł +éģ ¿ +ä» į +åº Ĩ +æĢ ķ +è° ¢ +çĽ ĸ +å° Ħ +éľ ² +æĸ Ĺ +ç Ĭ¶ +åŃ ¸ +æ¯ ķ +å· ¨ +çŁ ¿ +çļ ĩ +å¸ Ń +çĹ ĩ +æī ¬ +å» ¶ +ä¾ § +æ· ¡ +çļĦ ä¸Ģ +ç¶ ² +æ´ ģ +ç ¸ +è§ Ī +çŃ ¹ +ç§ ĺ +è¯ Ĭ +çı ¾ +èª ī +æ¯ « +ð ¨ +åį ´ +æĪIJ 为 +èĥ½ åĬĽ +é» Ħ +æĹħ 游 +èĪ ¬ +æ¯Ķ è¾ĥ +èµ· æĿ¥ +äºĨ è§£ +èĩª çĦ¶ +ä¸Ģ 次 +åŁº æľ¬ +æĽ ¾ +综 åIJĪ +èı ľ +è§ī å¾Ĺ +第 äºĮ +è· ij +æ³ ¢ +åĢ Ĵ +ç¡ Ģ +åħ µ +èį ī +çĶ ³ +çĶ ° +æĤ £ +è§Ħ å®ļ +èĥ ľ +èµĦ 产 +æ¢ ¦ +æľ Ŀ +è¿Ļ éĩĮ +å¤ « +æĮ ¥ +ä½ Ľ +å® Ī +éĽ ¶ +æĸ ¼ +ç¯ ĩ +å² Ľ +åĵ ¥ +éŃ Ķ +ä¸į åΰ +æī ĺ +åº Ĭ +æ¬ § +èį £ +æ± ĩ +æī © +åģ ı +å¢ Ļ +è® ¯ +å© ļ +æĥ ł +æ´ ĭ +å® ľ +æ¶ ¦ +æħ ¢ +éĢ ı +å® ½ +é¡ ¾ +ç´ ¯ +æ± ¡ +çĪ Ĩ +ç§ Ł +æĥ Ĭ +æ¶ ¨ +é¥ ° +éĺ µ +é¥ ® +æļ ĸ +åº Ł +æĹ Ĺ +éļ Ķ +ç¶ ĵ +åĭ Ļ +å¯ ¦ +éĢ Ķ +æī « +çĥ Ī +éĽ » +åĪ ij +éĹ ľ +éĹ ª +å¥ ĭ +å Ĥ¨ +ç¼ © +ä¾ µ +å ¬ +𬠶 +åĽ½ éĻħ +ç»Ħ ç»ĩ +ä¸ĵ ä¸ļ +åıij çݰ +å¸Į æľĽ +ç»ı èIJ¥ +åı « +æĿ¥ 说 +éļ ľ +ä»» ä½ķ +交 æĺĵ +éĩį çĤ¹ +çļ ® +ç» į +æ´ ¾ +ç§ij åѦ +åºĶ ç͍ +建 çŃij +èĤ ī +æĶ¹ éĿ© +åŁº ç¡Ģ +æ± ī +åĩº æĿ¥ +è¿Ļ ä¹Ī +åĪ ļ +åĿ IJ +ä¸į ä»ħ +ä¼ļ è®® +éĿ ł +åªĴ ä½ĵ +æ° ¸ +åĨ ² +èĭ ı +å¤ ® +çĪ ¶ +åł Ĥ +å®ŀ éĻħ +è¡ Ĺ +ç« ¥ +éĺ ħ +äºĭ æĥħ +åİŁ åĽł +éħ ¸ +以 æĿ¥ +å¨ ± +å® « +åĿ Ĺ +ç» © +éĩ İ +ä¸į å¾Ĺ +ä¼ł å¥ĩ +ç¡ ¬ +åİ ħ +æĹ ¢ +ç» ĥ +èĦ ij +å¼ ± +æİ Į +è´ ´ +æĮ Ĥ +åħ³ éĶ® +å° ļ +é¥ Ń +åº Ħ +çĻ ¼ +åľ ĭ +æİ Ī +个 æľĪ +äº Ī +å¸ ģ +è· Ŀ +æ² ī +ç« Ł +åĨ ¬ +æĬ ½ +éĨ Ĵ +å¼ Ł +è§ ¦ +èģ ĺ +è± Ĩ +æļ ´ +åijĬ è¯ī +è± ª +èµ ¢ +è· ¨ +è³ ĩ +çĪ ¸ +æĬ ± +æµ ª +éº » +ä» ª +è¡ ¡ +å¥ ¶ +çģ ¾ +èµ ¶ +èĤ ¥ +å§ IJ +åĢ º +éľ ĩ +è® ¢ +æ¬ Ĭ +ç · +å» ī +ä¿ Ĺ +å¿ ĺ +å¦ ĩ +ç¼ ĵ +åŃ ķ +æ¼ « +è£ ģ +çĩ ĥ +é» ĺ +çī ¢ +çĪ · +æĬ µ +å® ¾ +æľī ä¸Ģ +è¿ ¹ +è¿ « +è² Į +æľī çļĦ +ð¬ ĺ +è¿ĺ æĺ¯ +æīĢ ä»¥ +ä¹Ł æĺ¯ +è¿Ļ äºĽ +对 äºİ +åIJ § +缮 åīį +èĩªå·± çļĦ +èĥ½ å¤Ł +å¦Ĥ ä½ķ +æľº æŀĦ +åıª æĺ¯ +ç½ij ç«Ļ +åħ¨ éĿ¢ +为 äºĨ +å¼Ģ åıij +æĸ° éĹ» +éĩij èŀį +ç» § +客 æĪ· +ä¸Ģ èµ· +èĮ ¶ +åħ³ 注 +æ°´ å¹³ +åİĨ åı² +å¢ŀ éķ¿ +é ± +åŁº éĩij +åº Ń +åı ¶ +ä¿ ĥ +éĽ ¨ +æ¶Ī è´¹ +èĪ ¹ +çŁ¥ è¯Ĩ +æĪĺ çķ¥ +ç»ı éªĮ +å³ ° +æĽ ² +èĦ ļ +åĨ ° +å¤ ı +å½ Ĵ +ç¬ Ķ +èĻ ij +çĶ ² +åľ Ī +è¯ Ĺ +é½ IJ +容 æĺĵ +çłĶ åıij +éª ¨ +çº ¸ +è· µ +æĹ § +çķ ¶ +åĪ ¸ +è´ · +åı ¬ +ç§ ĭ +æ¶ ² +è¡Į æĶ¿ +çĮ ® +èĤ ¤ +éĢ IJ +è¶Ĭ æĿ¥ +è¶ĬæĿ¥ è¶Ĭ +æĦı è§ģ +èĪ ŀ +åī Ĥ +æ¶ ī +ç¨ĭ 度 +åħ¬ åħ± +æ¢ ° +æľ « +çº ¯ +åĶ ± +æ´ ² +æĬ ¢ +æ¤ į +å¿ Ļ +ä¼ ° +å¼ ¹ +æ³ ī +æľĢ 大 +è¶ ĭ +å· § +ç¦ ģ +æī ¶ +åį ± +çı ł +çĨ Ł +æĭ ľ +主 ä¹ī +æĿ Ĥ +éĻ Ħ +éģ į +æIJ Ń +æĮ ¯ +å¤ļ å¹´ +æķ ¬ +æij Ħ +çº · +å¼ ĥ +æ¹ ¿ +å¨ ĺ +æ¡ £ +é© ¶ +æľ Ĺ +æ® ĸ +æ¦ ľ +åĵ ¡ +ä¸Ģ ä½ĵ +æŁ¥ çľĭ +ç¹ ģ +æµ ĵ +åħ¬ å®ī +æ½ ľ +è´ ¯ +éª Ĺ +æ IJľ +å· ¡ +è ¬ +é Ĭ +å§Ķ ä¼ļ +æĤ ł +åī © +æı Ń +åŃ£ 度 +ð «ĺ +𬠬 +ä ´ +ð ª +ä½Ĩ æĺ¯ +éĥ½ æĺ¯ +å¹³ åı° +åѦ ä¹ł +åĵģ çīĮ +ä¸ Ķ +è¿Ļ ç§į +æĶ¿ çŃĸ +æĭ ¬ +认 为 +ä¸Ģ èά +æłĩ åĩĨ +æĶ¯ æĮģ +模 å¼ı +åħ³ ç³» +çļĦ æĺ¯ +è¿Ļ ä¸Ģ +ä¸į è¦ģ +çĶ ļ +ç²¾ ç¥ŀ +æĭ ¥ +åĪ© ç͍ +ä¿Ŀ æĬ¤ +ä½ľ ç͍ +èĭ ¥ +åĽ½ åĨħ +ä»ĭ ç»į +ä¸Ģ ä¸ĭ +å·¥ ä¸ļ +缮 æłĩ +æľĢ åIJİ +ä»· å̼ +å° į +éĵ ģ +è° ģ +ç»ĵ æŀĦ +éĽ ª +æĻº èĥ½ +ä¼ł 绣 +ä½ĵ èĤ² +çĶŁ æĢģ +æĭ į +æİ ª +åĨľ ä¸ļ +çī¹ èī² +è§Ħ 模 +æĹ¶ 代 +è¿ĩ ç¨ĭ +éĴ Ī +æĿ ¾ +åĶ IJ +åĮ» çĸĹ +çģ ¯ +åζ éĢł +æł¸ å¿ĥ +ä¸į åı¯ +ç³» åĪĹ +åIJ ī +åľ £ +åĢ ij +ä½ ³ +æĿ¥ çľĭ +æ¯Ķ èµĽ +ä¸ĭ æĿ¥ +åĩº äºĨ +å¹² éĥ¨ +å¾® ä¿¡ +å½ĵ åľ° +åį · +åį« çĶŁ +ä¼ Ł +çĸ« æĥħ +è° · +åĩł 个 +éĺ ´ +çĶŁ çī© +å° ¤ +ä¼ Ĭ +èĤ ¯ +éĿ¢ 积 +åĪĽ éĢł +æı ¡ +åľ Ĩ +æĻ ĵ +æĪIJ äºĨ +åĩ ¡ +çĸ ¾ +ç«ŀ äºī +è® ¨ +主 é¢ĺ +é² ģ +è¿ ª +ä¿ Ħ +æĢ ª +ä¸ ¦ +èĻ ļ +æ½ ® +çĥ § +èĢ ³ +æ± ł +éĢĤ åIJĪ +æł¹ æľ¬ +åĬł 缣 +ç͵ è§Ĩ +æ· · +ç¼ ĺ +çª Ĺ +çĬ ¯ +æĥ ¯ +æĦı ä¹ī +åĬŀ æ³ķ +ä¼ ij +æ» ij +åĭ ĩ +æķ ¢ +å¯ » +è¦ Ĩ +éĢ ĥ +ç»ı çIJĨ +åĿ ı +æ³ ½ +ä¹ ĺ +åĪ º +å± ı +é¡ ¿ +äº ¡ +éĤ Ģ +åħ ¼ +åĭ ¤ +æ® ĭ +æĺ ł +æ¯ķ ä¸ļ +æĪ ª +è· Į +å£ ģ +åı¦ ä¸Ģ +羣 å®ŀ +ç£ ¨ +è¯ ļ +å¿ħ è¦ģ +æģ ĭ +æĩ Ĥ +å¾ Ĵ +è° ĵ +æķ ı +æ ύ +èĥ ¸ +æĭ ¼ +å¦ Ļ +è¯ ¸ +èģ Ĭ +æĤ ī +éº ¼ +åĩ Ń +èĪ Ĵ +æ¶ Ĥ +è¿ ģ +æ² ¿ +å¡ ij +æĽ ¿ +æ¾ ³ +å¿ į +èĢ Ĺ +éľ ¸ +åĩł å¹´ +åĪ Ĭ +èĦ ī +èħ IJ +æ¡ Į +çº ł +æ» ļ +æĤ ² +åĨ Ĵ +å¦ ¹ +çķ ħ +çº µ +æij ĩ +å¤ º +è·¯ ä¸Ĭ +å¿ ½ +èĸ ª +æģ IJ +æĦı æĢĿ +å« Į +æı ´ +æ° § +èĢ Ģ +éĺ » +è½ ¨ +å¹ » +æį ķ +åĿ ¦ +åĵĪ åĵĪ +çĭ IJ +æ» ¨ +è² » +è¿ Ł +人 éĥ½ +ç» ĺ +åı ¹ +çµ IJ +æī ° +æ» ĭ +å¥ ij +åĭ Ł +ç¢ º +ð ¦ +éĽĨ åĽ¢ +æĿ İ +å¼Ģ å±ķ +æıIJ åįĩ +åħ¨ åĽ½ +æ±½ 车 +åѦ æł¡ +æł¹ æį® +è¿Ļ æĺ¯ +åĩº çݰ +éĻ Ī +ç½ Ĺ +èİ· å¾Ĺ +åĪ ĺ +éĶĢ åĶ® +æľª æĿ¥ +éľĢ æ±Ĥ +å®ŀ æĸ½ +åĿļ æĮģ +åħ¨ çIJĥ +éĵ¶ è¡Į +æİ§ åζ +é¡ » +åľ° åĮº +æīĵ éĢł +çļĦ è¯Ŀ +帮 åĬ© +ä½ĵ ç³» +è¾¾ åΰ +è§Ħ åĪĴ +åŁ¹ è®Ń +两 个 +æĬ¥ åijĬ +åľ° æĸ¹ +å®Į åħ¨ +æİ ī +ç»ĵ åIJĪ +宣 ä¼ł +æ³ķ å¾ĭ +èīº æľ¯ +ç͵ å½± +èª ª +ä¸Ģ çĤ¹ +è¶ħ è¿ĩ +ç͵ åŃIJ +æĢĿ æĥ³ +æķĻ åѦ +éĺ¶ æ®µ +åķĨ ä¸ļ +çī© æµģ +åĪĽ ä¸ļ +æĸ¹ æ¡Ī +çݰ 代 +æ¡ ¥ +èIJ½ å®ŀ +带 æĿ¥ +产 çĶŁ +ç§ Ģ +æ³ ° +ä¹ ± +åħ· ä½ĵ +åĸ Ŀ +èĵ Ŀ +å® Ĺ +åįĩ 级 +æ·± åħ¥ +ä¿Ŀ éĻ© +ç®Ģ åįķ +çĹ Ľ +稳 å®ļ +è¾ Ĩ +å±ŀ äºİ +å· Ŀ +ä¸į å°ij +åĴ ¨ +举 西 +å½¢ å¼ı +娱 ä¹IJ +æŃ£ 常 +é¸ ¡ +åħħ åĪĨ +å®ŀ è·µ +éĩĮ éĿ¢ +è· ³ +èĻ İ +æĪIJ éķ¿ +æļ Ĺ +çĿ ¡ +ç½ ª +çIJĨ 念 +æĮ ij +èµĦ æľ¬ +å¤ļ å°ij +ä¸ĭ éĿ¢ +å¸ Ŀ +åħ¬ å¼Ģ +æ¸ IJ +éķ · +å± ĭ +欢 è¿İ +å¿ĥ çIJĨ +çĤ İ +æ¹ ¾ +è® ĵ +éĤ Ħ +ç³ ĸ +ä¹ Į +åĬ ± +çī Ļ +èħ ¿ +å² Ĺ +ä¼ į +æĪIJ åijĺ +åŃ Ķ +å°ı ç¼ĸ +èij £ +æ³ ¡ +åħĪ è¿Ľ +åħ § +åĺ ´ +è´ Ŀ +è » +æIJ ŀ +æ³ Ľ +é¸ Ł +ç½ ² +èĽ ĭ +主 ä»» +缮 çļĦ +ä¹ ı +æ´ ¥ +æĪ ´ +严 æł¼ +çħ ¤ +çĮ « +åĶ ¯ +å° Ĭ +çĶ ľ +åŀ ĥ +åľ ¾ +æĭ Ł +çĦ ¦ +é« Ķ +å® ı +æ© Ł +é© » +æĹ ģ +å½ » +éĥ½ ä¸į +æij © +ä» ĵ +ä¹ ³ +å² ¸ +è° ĭ +大 å¤ļ +çģ Ń +èħ ¾ +æŁ ľ +èĪ į +åħļ çļĦ +å° ĺ +åįģ å¹´ +æĭ Ĵ +è£ ¡ +æŁ Ķ +å¹ ¼ +éĶ ģ +ä¸ĵ 项 +æī İ +驾 é©¶ +ç¢ İ +è¢ ĭ +éĶ ĭ +å£ ® +å° ĸ +ç͵ æ±ł +è¿ Ķ +æ¼ ı +å¾ ª +èı Į +èĥ ĥ +è¾ ħ +éĢ Ĵ +èĥ İ +éĻ ª +å¯ ¿ +å¥ Ķ +çĮ Ľ +çº ¹ +çŁ¥ åIJį +å¿ Ĩ +æ¡ ĥ +æ£ ĭ +éĢ Ĩ +çĤ ¼ +ç± į +çī § +æł· çļĦ +è¾ Ľ +åł Ĩ +å®ŀ åľ¨ +ä¼ ı +å® ¿ +èµ ı +è£ Ĥ +åįĬ å¹´ +åĢ ¾ +满 æĦı +æ¢ ¯ +æĦı åij³ +åŃ ¤ +ç¥ Ŀ +æĻ ¶ +èµ Ķ +åģ ¿ +èĦ Ĥ +ç½ ļ +ç¢ į +æ² ĥ +æ ĵį +å´ ĩ +æļ Ĥ +è· ĥ +æIJ ¬ +å© Ĩ +é ī +éī ´ +åħ´ è¶£ +èIJ¥ ä¸ļ +è® Ĭ +èĦ ı +è¾ Ī +å·ŀ å¸Ĥ +è´« åĽ° +ç© · +ä¸Ń å°ı +æ¼ Ĥ +çĻ Į +èľ ľ +ä¼Ļ ä¼´ +çī µ +æĤ Ł +éĻ · +èµĽ åŃ£ +æ¨ £ +åģ ¶ +æĺ Ĩ +è¢ Ń +æį IJ +èī ° +æ Ĥ¬ +çĶ ¢ +èij ¡ +çĽ Ĺ +å© ´ +å° İ +çº ½ +åĢ ¡ +æī ® +è¨ Ń +æĬ ij +ç¡ ķ +è¾ ĸ +éĥ ģ +è¾ © +éĤ » +çݰ åĩº +è¦ ı +å½ ¹ +éĺ Ķ +åī µ +è¯ ± +æĥ ij +æ· Ģ +é¢ Ī +ä¾ ¦ +æģ ° +æ£Ģ å¯Ł +éĨ « +çĦ¶ æĺ¯ +åĭ ĥ +èĮ « +ä ĵ +𠬸 +ä½ľ 为 +çļĦ 人 +éĤ£ ä¹Ī +ç¾İ åĽ½ +è¿ĺ æľī +æıIJ é«ĺ +èĻ ½ +åħ· æľī +åĮħ æĭ¬ +æĪĸ èĢħ +ä¸į è¿ĩ +ä¸Ĭ æµ· +åĮ» éĻ¢ +èµĦ éĩij +çĶļ èĩ³ +åζ 度 +è§£ åĨ³ +èģĶ ç½ij +ç»§ ç»Ń +建 ç«ĭ +è¿Ľ ä¸ĢæŃ¥ +æĿIJ æĸĻ +ä»Ĭ 天 +å¿ħ é¡» +åIJĦ ç§į +çݰ åľº +ä»ĸ çļĦ +å¢ŀ åĬł +é¢Ĩ åŁŁ +åıĤ ä¸İ +æĮģ ç»Ń +ä¹ĭ ä¸Ģ +çī¹ åĪ« +é± ¼ +åħ± åIJĮ +åĬ ª +çİ ī +人 们 +åħĪ çĶŁ +ä¼ĺ åĬ¿ +ä¿Ŀ æĮģ +ä½ľ åĵģ +çī Ľ +æĪIJ æľ¬ +æĶ¶ åħ¥ +åıĬ æĹ¶ +è´Ł è´£ +æİ¥ åıĹ +èį IJ +åıª è¦ģ +羣 çļĦ +导 èĩ´ +æľº åζ +è¡Į åĬ¨ +æĸ° çļĦ +å®Į åĸĦ +为 ä»Ģä¹Ī +ä¸Ń 央 +æĪIJ ç«ĭ +æĦŁ è§ī +åıĺ åĮĸ +åıĹ åΰ +å¹¶ ä¸į +åŃ Ļ +æĸ½ å·¥ +æĺİ æĺ¾ +è¿ĩ åİ» +åıij æĮ¥ +羣 æŃ£ +åŁº åľ° +æĺİ ç¡® +èĥ ¡ +许 å¤ļ +ä¸Ģ å¹´ +æĸ¹ åIJij +æģ © +缸 ä¿¡ +åľ ³ +详 ç»Ĩ +äºĭ ä¸ļ +çĶŁ åij½ +åĴ¨ 询 +æĸĩ æĺİ +çij ŀ +绿 èī² +èİ « +æĦı è¯Ĩ +æĬķ åħ¥ +åĬł å¿« +æ¢ ħ +ç¿ » +å¼Ģ æĶ¾ +æĻ® éĢļ +åįı ä¼ļ +æĪIJ 绩 +ä» Ļ +å¯ Ĵ +è¯ģ åΏ +认 è¯Ĩ +ä¸ ¹ +大 éĩı +è¿ ħ +åģļ åΰ +设 æĸ½ +è´¸ æĺĵ +èĥ½ æºIJ +æĹ¶ æľŁ +ä¸Ģ 天 +æ²» çIJĨ +åĺ ī +å® ĩ +丰 å¯Į +举 è¡Į +æĪIJ æŀľ +èĤ¯ å®ļ +çĭ Ĺ +åĬ¨ åĬĽ +æ£ ® +åĩł ä¹İ +åĽł ç´ł +æ°ij æĹı +æ´ ŀ +ç½ij åıĭ +åIJĪ çIJĨ +广 大 +æ® Ĭ +æ´ Ľ +æĿ ¯ +èĴ Ļ +ç͍ äºİ +èŀį èµĦ +ç¥ ĸ +æľº 械 +举 åĬŀ +èĩª åĬ¨ +åĬŀ åħ¬ +é» ŀ +éĽ Ħ +å̼ å¾Ĺ +çĮ ª +以 为 +æĺ Į +è·Ŀ 离 +åIJ¸ å¼ķ +ç» ķ +éļ Ĩ +计 ç®Ĺ +éĺŁ ä¼į +大 ä¼ļ +å¼ķ èµ· +çī¹ çĤ¹ +èĥ ¶ +å¹´ è½» +æľ¬ 身 +æľº åħ³ +å®ĺ æĸ¹ +éĥ ij +æµ Ļ +è§Ĵ èī² +èij£ äºĭ +为 主 +æĹł 论 +ä¹ł æĥ¯ +æ¥ ļ +æĭ ĵ +绣 计 +åħ Ħ +广 æ³Ľ +åį Ģ +污 æŁĵ +è« ĭ +èĬĤ 缮 +ä¼ ¦ +è¦Ĩ çĽĸ +èĢ IJ +æī¶ è´« +ç»ı åİĨ +éĩįè¦ģ çļĦ +èĤ¡ 举 +æĭĽ èģĺ +åĽĽ 个 +æĩ ī +èĥ ŀ +æij Ĩ +é«ĺ éĢŁ +éº ¦ +åİŁ åĪĻ +èİ ± +æĽ´ 好 +éķ ľ +åĩ Į +åŀĥ åľ¾ +éĢ ² +çģ ° +éĵ º +äºĭ æķħ +çĶ ĺ +空 æ°Ķ +é¾ Ħ +èı ² +çĵ ¶ +æĺ ¨ +æĹ¥ æĬ¥ +æµ ® +åľ° åĽ¾ +åij Ī +大 åĬĽ +ç» ª +å¸ ħ +æľį åĭĻ +ä¸į éĶĻ +乡 æĿij +å± ¥ +å¹³ æĸ¹ +éĹ ² +æī £ +ç´ł è´¨ +èµ ´ +éģ Ń +èIJ ¨ +èĩª 主 +éĩij å±ŀ +èī¯ å¥½ +两 å¹´ +æ³ ¥ +é¢ ľ +ç²¾ 彩 +ä¸Ń åįİ +æĻ ĭ +ä¹ł è¿ij +ä¹łè¿ij å¹³ +æĪĺ 士 +åģļ çļĦ +éª ij +æ» ´ +çĵ ľ +çīĪ æĿĥ +èĤ ł +æľĥ åĵ¡ +çı į +ç¨ ® +ä »¿ +çī© ä¸ļ +åĢĭ 人 +å¦ » +ä¼ ¸ +æ± Ĺ +æĹ º +çIJĨ æĥ³ +æij ¸ +è¿Ŀ æ³ķ +å®Į æķ´ +åİ ¦ +è¸ ı +æĸ ij +æ¡ Ĥ +ä½ĵ åζ +å¸ « +æĿ Ĩ +æ® ¿ +æ¯ ģ +é¦ Ī +è§Ĵ 度 +æ¬ £ +çĥ ¦ +èĤ º +éĩĩ 访 +æij ĺ +æĮ ¡ +æ· ĺ +åħ» èĢģ +çĤ ¸ +è¿ Ī +åİ ī +åĿ Ĭ +è¾ £ +åĩ Ŀ +æ³ ª +çĸ ı +æİ ĺ +åĥı æĺ¯ +éĽ ķ +ç¼ Ŀ +èį · +æį · +åł ¡ +åı¥ è¯Ŀ +çĸ ¼ +æł ı +éģ µ +ç¢ ³ +å·¥ åķĨ +æIJ º +åĪ ¥ +ä¹ Ļ +æĹ ĭ +æĥ ľ +ä¸Ģ 大 +å±Ĥ 次 +èµ ĸ +æĬ ¬ +æ¨ Ĥ +è¯ ŀ +åħ Ĵ +ç¯ ® +èĤ ĥ +å§ ¿ +æĬ ļ +çĵ · +ç͵ åĬ¨ +æĸ° åĨł +æ¶ µ +ç¢ ij +æ· ® +æĹ ¨ +è¸ ª +æ¸ Ķ +æĦ Ī +åı Ķ +åįĹ çľģ +ç¾ © +å§Ķ 书记 +è² ¸ +æ¶ Į +è« ĸ +èIJ Ħ +æı ı +å¿ § +è¾ ¦ +å¦ Ĩ +æī Ń +åij µ +éģ ¥ +è¨ ± +ä» ĩ +åįģ ä¸ī +åī ² +èª į +èĪ ° +é¢ ĩ +é¥ ± +çĭ ł +é«ĺ çļĦ +çµ ± +æħ İ +é¢ ģ +åIJĪ éĢĤ +æµ ´ +èµ ĭ +æĬ ¼ +å¦ ¥ +éĻ¢ éķ¿ +èĢ ķ +è¾ ¨ +æħ ° +åįģ åĽĽ +æľ µ +èĵ Ħ +æŀ ¢ +å» · +æĤ Ħ +æ¶ ¯ +çŁ © +åŃIJ éĩĮ +çĬ ¹ +å±Ģ éķ¿ +é IJ +å¥ ł +ä¼ļ éķ¿ +æĵ ļ +ä¸į åıĬ +åįģ ä¹Ŀ +æ¬ º +èº º +éĺ IJ +çº Į +è¨ » +åĨ Ĭ +èŃ ĺ +é«ĺ çŃī +èħ º +å¤ ķ +ç» ij +åĶ ¤ +èķ ´ +çķ ľ +æħ ĭ +åı Ļ +åı ĥ +å³ ¡ +人 大 +éħ ¿ +éģ © +å¥ ¢ +åı£ æ°Ķ +éĮ Ħ +é ı +åĭ ĺ +è´ ¿ +éļ ª +é ĭ +éļ ¶ +ð ¥ +𬠣 +ð £ +ð« į +𬠳 +ð« ĵ +ð« Ħ +ð« Ł +𨠱 +ä Ĺ +以 åıĬ +æľī éĻIJ +åij ¢ +åIJ Ĺ +çľĭ åΰ +计 åĪĴ +è¿Ľ åħ¥ +缴 æİ¥ +åĪĨ æŀIJ +åıª æľī +设 å¤ĩ +åħ¶ å®ŀ +åĬł 强 +ä¸Ń çļĦ +ä¿Ŀ éļľ +èĢģ å¸Ī +人 æīį +å¾Ĺ åΰ +é£İ éĻ© +ä¸Ģ ç§į +空 éĹ´ +æĪij åĽ½ +ä¹ĭ åīį +ä¸ĵ å®¶ +æĿ ¨ +æĹ¥ æľ¬ +群 ä¼Ĺ +åıĤ åĬł +æķĪ æŀľ +æľī åħ³ +å®¶ åºŃ +åĮº åŁŁ +åĬª åĬĽ +éļı çĿĢ +æĹł æ³ķ +交 æµģ +è¡Į 为 +æ£Ģ æŁ¥ +æľŁ éĹ´ +å¦Ĥ æŃ¤ +èĤ¡ 份 +å½ĵ æĹ¶ +è£ħ å¤ĩ +åĩĨ å¤ĩ +éħĴ åºĹ +è¿IJ åĬ¨ +æıIJ åĩº +å·¦ åı³ +æİª æĸ½ +é£Ł åĵģ +æ¶Īè´¹ èĢħ +åѦ éĻ¢ +æĮĩ 导 +è¿IJ èIJ¥ +éĩį 大 +åĨľ æĿij +éĢł æĪIJ +æĶ¿ æ²» +éĴΠ坹 +æŃ£ å¼ı +åıĸ å¾Ĺ +éĤ£ 个 +éĽĨ ä¸Ń +åıª èĥ½ +å¿« éĢŁ +身 ä½ĵ +åħļ åijĺ +èģĶ åIJĪ +åĬĽ éĩı +éĥ½ æľī +æ ħ§ +å¡ Ķ +åĪ« 人 +表 çݰ +æķħ äºĭ +ä¸Ģ åĪĩ +å° ĩ +èµĦ æĸĻ +åŁ¹ åħ» +éĺħ 读 +æľī 人 +èIJ¥ éĶĢ +çĽij çĿ£ +çݯ ä¿Ŀ +èĢĥ èĻij +æ·± åľ³ +严 éĩį +èĮĥ åĽ´ +å§Ķ åijĺ +çĽij 管 +ä¸ī 个 +è£ħ ä¿® +åħ¬ éĩĮ +åĪĨ åĪ« +çIJĨ è§£ +éŁ © +åĬł å·¥ +认 羣 +ä¸į 好 +åİ» å¹´ +éĻį ä½İ +æľº ä¼ļ +åįı è®® +符 åIJĪ +å¢ŀ 强 +æĬĢ èĥ½ +é¦ĸ åħĪ +ç§ ¦ +ä¸ ģ +å° ¾ +æľī äºĨ +åľ° 产 +æ¸ ł +æĸ¹ 便 +ç§» åĬ¨ +éĢŁ 度 +å°¤ åħ¶ +éĢļ çŁ¥ +åĿ Ľ +éģ¿ åħį +æģ ¢ +è´ ¡ +èģĮ å·¥ +å®ŀ åĬĽ +æĺ¯ä¸Ģ ç§į +åIJ¯ åĬ¨ +çĸ¾ çĹħ +æĿ¥ äºĨ +缸 对 +çݰ å®ŀ +èŀį åIJĪ +åIJĮ æł· +åħ¬ åijĬ +çī¹ æ®Ĭ +ç´ « +ä¸ĭ åİ» +ä¼ł æĴŃ +æľĢ 好 +ä¼ĺ è´¨ +æ² Ĵ +æĮ º +æĹ ¦ +è¯ º +ä¸Ģ åIJį +éģĵ è·¯ +示 èĮĥ +è¿ĩ æĿ¥ +åIJĮ åѦ +é¼ ĵ +æĿ Ń +æľ¬ 次 +åIJĮ æĦı +ä¸ĸ 纪 +ç¾ Ĭ +æ¬ ² +å·¥ èīº +çĵ ¦ +人 士 +æľī æīĢ +ä»İ äºĭ +æľī å¾Īå¤ļ +ä¸į äºĨ +å²Ĺ ä½į +åıĺ å¾Ĺ +åĬ³ åĬ¨ +å¤Ħ äºİ +å¹³ åĿĩ +å½¢ 象 +å¡ ŀ +åħ± 享 +çĿ Ľ +åĪ© 润 +æŃ£ æĺ¯ +å¾Ģ å¾Ģ +缸 æ¯Ķ +æ¨ ª +åĪ · +æµĻ æ±Ł +大 éĥ¨åĪĨ +å¤ļ 个 +æĤ¨ çļĦ +ç͵ åķĨ +å¾® åįļ +å§ĭ ç»Ī +çĬ¯ 罪 +æĺ¯ åľ¨ +ç»Ħ åIJĪ +åİŁ æĿ¥ +æ¸ħ æ¥ļ +åIJĦ åľ° +æĦŁ åıĹ +å½ĵ ä¸Ń +è¶ĭ åĬ¿ +æĻ¯ åĮº +羣 æĺ¯ +ä¾Ľ åºĶ +转 åŀĭ +çĭ Ĥ +èĨ ľ +èĭ Ĺ +å¿ ł +å¾Ī 大 +èĤ¡ æĿĥ +ç¾İ åħĥ +æİĴ åIJį +åĬ¨ çī© +éĶ ħ +å¢ ¨ +主 å¸Ń +å¾Ī 好 +ç»Ŀ 对 +æĿ ľ +转 è½½ +çĴ ĥ +æĿij æ°ij +åIJ ¨ +åĽŃ åĮº +é«ĺ 度 +çī© è´¨ +è¾ ī +æĹ¥ 常 +æı Ĵ +ä¸ī å¹´ +ä½ĵ çݰ +æīį æĺ¯ +代 çIJĨ +ä¸į 管 +æģ Ĵ +åľ° ä½į +ç² ® +èĸ Ħ +æĺİ çϽ +ä¸Ģ èĩ´ +æĽ ¼ +åĵ Ń +åĩ ¤ +åĬ ² +æķ Į +æĪĺ æĸĹ +主 ä½ĵ +åħ¬ å¸ĥ +åıĤ èĢĥ +èĪª 空 +å¯ º +åѦ ä¼ļ +åıį æĺł +ç¾İ 丽 +太 éĺ³ +建 æĪIJ +æħ¢ æħ¢ +åIJĦ 个 +éĤ ¦ +ç»Ħ æĪIJ +ä¸ī 大 +éĶ ¦ +大å¤ļ æķ° +æ¦Ĥ 念 +éŃ Ĥ +åħ¬ çĽĬ +èį Ĵ +身 份 +æ·± åĪ» +åħ © +ç»ı åħ¸ +åIJĦ 项 +èĻ ķ +è¿Ľ æŃ¥ +åįģ äºĮ +æī§ æ³ķ +æĥ³ åΰ +æĦŁ æŁĵ +åķĨ åĬ¡ +å°ı ç»Ħ +èĶ ¬ +çıŃ åŃIJ +åIJĮ å¿Ĺ +éĿ¢ 临 +çĤ Ĵ +å¤ļ ç§į +è§Ĥ çĤ¹ +åĵª éĩĮ +å° Ŀ +å§ Ĩ +èħ ¹ +åŁİ åĮº +太 å¤ļ +çĹħ æ¯Ĵ +åľ¨ äºİ +æīĢ è°ĵ +æĻ ° +æŀ Ŀ +æĭ ĸ +å® ħ +æķ´ æ²» +ä½ı æĪ¿ +åģ · +çĨ Ĭ +èµ ģ +æ° Ľ +æł¼ å±Ģ +åŁºç¡Ģ ä¸Ĭ +èĥ Ĩ +åħ ½ +鼶 åĶ® +åĿ ¡ +女 åŃ© +æĴ ŀ +åħ¨ åĬĽ +åĴ ĸ +èĤ © +çľ ī +èĩ³ äºİ +åħļ ç»Ħ +ä¸Ģ ä»¶ +æĭ Ĩ +äºĭ å®ŀ +åĤ ³ +æ¹ ĺ +ç¶² ç«Ļ +循 çݯ +åIJĮ æ¯Ķ +æĭ Ķ +åĮ» èᝠ+åħ» æ®ĸ +åĽº å®ļ +å®ŀéĻħ ä¸Ĭ +è®° å¾Ĺ +åĪ© äºİ +æĤ ¦ +æĭ ³ +èĤ Ŀ +æķĪ çĽĬ +è© ² +æ°ij 主 +çĹĩ çĬ¶ +é¢ ¨ +å¹¼ åĦ¿ +å§ ij +æĪ Ĵ +ä¸ĭ çļĦ +æ¸ ¡ +å¹´ åºķ +è®° å¿Ĩ +åIJ IJ +大 å¹ħ +å¾ ½ +åħ¬ ä¼Ĺ +ä¿¡ å¿ĥ +çİ Ľ +ä¼ļ ä¸Ĭ +ä¹ Ķ +æijĦ å½± +æ£ĭ çīĮ +éĻ ķ +åºĶ æĢ¥ +æĶ¶ è´¹ +æİ§ èĤ¡ +仪 å¼ı +çŀ ¬ +æīĢ åľ¨ +ç¢ ° +å§ ĵ +é¡ Į +æĶ¯ éĥ¨ +使 åij½ +çĤ ī +å¯ Ħ +ç¿ ¼ +åľ° ä¸ĭ +è¾ ŀ +ä¿ ± +主 æĮģ +è´§ å¸ģ +æģ ¨ +èĤ Į +çĽ Ī +éĶ » +å¿Ĺ æĦ¿ +ç±» ä¼¼ +æĮ ĸ +éĢ » +ç¸ ½ +纪 念 +åķ ¥ +å¼ ¯ +åIJį åŃĹ +åģ¥ èº« +çļĦ å¿ĥ +é© ± +èĥĮ åIJİ +æ³ķ å¸Ī +ç² Ĵ +èĥ½ éĩı +è¾ ° +èī ³ +å½ ¼ +段 æĹ¶éĹ´ +åIJĪ æ³ķ +æĵ ¦ +ç¾ ½ +åİ ¨ +æĪij 说 +äºĭ åĬ¡ +åĩł 天 +åħ ģ +ç¼ ´ +åį ĵ +两 ç§į +çĭ¬ çī¹ +å¸ ¶ +éĴ » +æĥ © +é¢Ĩ åħĪ +è¶³ å¤Ł +å£ ³ +æĦıåij³ çĿĢ +åĪĨ å¸ĥ +ä¹ ĥ +éģ ĭ +ä½ © +è° ± +çģ £ +èį ¡ +è´¯ å½» +å¹ ¾ +ç£ ģ +åħ¸ åŀĭ +åī ĩ +åĨ » +æ¬ ł +ä¸į ä¹ħ +æµ ¦ +éŃ ħ +å¼Ģ äºĨ +使ç͍ èĢħ +è¿Ļ 款 +å° Ī +èĦ± è´« +æĶ» åĿļ +ç®Ĺ æĺ¯ +ç¨ Ģ +æĹł 人 +åł µ +å¥ ı +éĥ½ å¸Ĥ +åı¯ è§ģ +ä¸į åĩº +æ ·» +äº ı +ç¾İ 好 +èĥ ĸ +éŁ µ +æłĩ å¿Ĺ +èĬĤ èĥ½ +æĬ « +å° º +å¯ ¸ +ä¸Ģ 代 +é¢ Ĺ +èĢ ¶ +èĴ ¸ +åĸ ® +æ »¿ +çĮ ľ +æµ Ĩ +åŁ ĥ +åįĥ ä¸ĩ +èµ Į +èģ ² +ä½ľ é£İ +è³ ª +å¯ ¨ +å¹´ 人 +åį° è±¡ +æ¡ ¶ +æĴ ¤ +åįģ äºĶ +æ¯ ħ +æ² ª +åĽ½ æľī +大éĩı çļĦ +å¾ ¡ +å¯ ĵ +è¦ ĸ +æ¼Ĥ 亮 +çľ ł +ç ĤŃ +é» İ +èĻ ¹ +åĪ© äºļ +èŃ ī +æµ ı +åįģ åħ« +ä¸ ¢ +è¾ ½ +æľīä¸Ģ äºĽ +æħ Ī +åģľ è½¦ +å® ł +è§£ æĶ¾ +æľī å¤ļ +éĤ Ĭ +常 è§ģ +æĬ ¹ +çº ¤ +è¦ ª +æ¡ Ĩ +èİ ŀ +æ°§ åĮĸ +è¿Ļ ä»¶ +åĩ ° +æŁ ´ +åıij ç͵ +é¼ ł +转 åĮĸ +å¨ ĥ +æĮ ¤ +ç½ © +å¯Ĩ åĪĩ +æĪij ä¸į +é«ĺ æĸ° +ä¸Ģ ç¯ĩ +è¿Ľ ç¨ĭ +è¡ ° +è¿ĺ ä¸į +ç ħĮ +æĸ° åįİ +èĤ ¿ +æ» © +ä¸Ģ æµģ +è¯ Ī +å®ŀ ä½ĵ +å¤ĸ åĽ½ +èº ² +èµ ł +è¦ º +æ¢ Ŀ +ä¸į è§ģ +è¨ Ĭ +åĮ ¹ +åį µ +çĩ ¥ +æħ ķ +é½ ¿ +å® ´ +é¥ ¼ +èij¡ èIJĦ +å°ı å¿ĥ +æģ ¼ +éĻ Į +æĺ Ĥ +åĥ ¹ +èĬ Ŀ +æ¯ı 个人 +åīį æıIJ +ä½ĵ ä¼ļ +æ¨ Ļ +æIJľ çĭIJ +对 åħ¶ +ä¸ § +èľ Ĥ +æµ ¸ +èª ¿ +åĿ ª +é¢ ĸ +åIJį 为 +ç¬ ¼ +èĪ Į +æľ¬ 书 +èģ ¯ +çº º +ç®Ģ 缴 +éĽ ¢ +ç¾İ çļĦ +éļ ¨ +é«ĺ å³° +è¿Ļ å®¶ +å Ĥ¬ +å° ¸ +ç¡ķ 士 +èŃ · +è° ¨ +æĺ ı +æĶ¿ åįı +è¡ Ķ +ç¿ Ĵ +åľ Ĵ +åĽ½ æ°ij +主 è§Ĵ +è£ ķ +ä¼ ª +åº ŀ +æ°ij èIJ¥ +æĥ § +ç§ĺ 书 +çĹ ķ +çϾ åĪĨ +æº ¶ +æĹł çĸij +çļĦ çľ¼ +æĵ İ +ä¼Ł 大 +å½ ° +åħ¬å®ī å±Ģ +ç³ ķ +å¼ ¥ +åĤ Ļ +ä¹ ¾ +毫 ä¸į +注 æĺİ +åī¯ æĢ» +æĦ ī +æķ ¦ +é¦ ¨ +æĶ Ģ +éĢ Ŀ +åı¯ éĿł +å¤ ¸ +åľ ĺ +éĿ¢ ä¸Ĭ +æĬ ĸ +èĦ Ĩ +é© ° +ä¼ IJ +å¦ ¨ +å®ļ äºĨ +ç³ Ĭ +æŃ ¡ +éĥ¨ éķ¿ +ç§ ī +èĪ Ĩ +åĪij äºĭ +åIJ µ +æ¤ Ĵ +è¡ ĵ +è± « +èı © +åŃ µ +é¥ ² +å°± 好 +åł ª +ä¸ī è§Ĵ +åľº æ¯ĶèµĽ +ä¸į åģľ +æĵ ħ +åħ¨ æĸĩ +æ³ ģ +åѦ ä½į +æ± ° +éł ĺ +åı ł +éļ Ľ +å¸ IJ +çľĭ åĩº +åĮ ł +å±Ģ éĿ¢ +æ³ Į +è° Ĭ +åIJĮ æľŁ +æĬķ æłĩ +å¥ ´ +æĿ¥çľĭ çľĭ +èĦ ¾ +èŀ º +æŃ ī +çĽ ¯ +ç¨İ åĬ¡ +å» Ĭ +æİ © +æħ ¨ +çĽ ¼ +èĬ Ĵ +è® Ģ +æĮ £ +èĮ ħ +æĸ ¥ +æ¤ ħ +åΰ æĿ¥ +èijĹ ä½ľ +çĭ ± +äºĮ æīĭ +ä»İ æĿ¥ +çĸ ² +åºĬ ä¸Ĭ +æĸ° 浪 +æ³ Ħ +å¢ŀ å̼ +ä¸ Ľ +æļ ij +ä»İ ä¸ļ +æ· ĭ +å¤ļ æł· +æľ ´ +份 é¢Ŀ +æŀ £ +西 çľģ +æľ¬ è´¨ +æ·± æ·± +èī ĩ +ç» µ +产 å̼ +æ¼ ł +èħ » +çŃ Ľ +åİ Į +æģ Ń +å«Į çĸij +æĪ ¶ +æ» ŀ +èĨ Ģ +åĬ £ +座 è°Ī +常 æĢģ +çļĦ æĥħ +è¦ ½ +å¯ Ĥ +åĮ Ĩ +èĩ º +é¡ ¯ +çķ ı +éģ £ +åį ľ +çŃī å¥ĸ +è² ¬ +æº ¯ +é İ +çĤ¹ 头 +èĵ ¬ +æ± º +éħ ¬ +éģ Ĭ +è³ ¼ +註 åĨĬ +æľ¬ æĬ¥ +çµ ķ +æ´» æĢ§ +åħ ij +éĮ ¯ +åĨ ¶ +åĸ » +æº ĸ +èĤ ¢ +æº ĥ +æĹ ¬ +åī Ĭ +çIJĨ äºĭ +å± ł +æ² § +èļ Ģ +鼻 åŃIJ +为 æŃ¢ +常 å§Ķ +çµ Ĥ +éĬ · +çĭ Ģ +ä¾ £ +èĥ Ģ +èŃ ° +ç͍ 车 +åĻ ª +æŃ · +åį Ķ +åĪ ¹ +竣 æĺ¯ +é© Ĺ +èIJ Ŀ +çĻ « +çĹ « +æŃ § +å¼ Ĭ +åª ½ +çı Ĭ +è¡ · +éľ ī +åŁº çĿ£ +éļ ± +æ° ¨ +ç» ¸ +å°¼ æĸ¯ +çĥ ĺ +æľŁ åĨħ +è° ħ +éĽ ĩ +éļ Ļ +å ĸī +åī ¥ +çĹ ĺ +æĮ ½ +çĵ £ +æ¹ Ľ +æ¨ ± +æ¾ İ +æ¹ ĥ +åĨ¬ 奥 +æ£ µ +å® ° +åŀ Ĵ +æ§ ĭ +ä¾ Ī +èĮ Ħ +åĺ ¿ +èı ĩ +ç ĻĤ +åĬ ĥ +é į +èĶ ½ +çŀ Ń +æķ ŀ +ä¹ ĸ +éŁ § +è¾ ľ +æĩ Ī +ä½ £ +çŀ » +åŁ Ķ +èĪ ħ +å®ŀ äºĭ +é ¨ +å§ ¥ +çµ ¡ +åĺ » +çķ ¢ +æ²ĥ å°Ķ +è¿ Ħ +èĤ ĩ +æħ ij +ã § +ä ı +ð ł +ð¬ ĩ +ð« Ń +ð« IJ +ã ³ +© ½ +ð« ł +ã Ľ +ð¬ į +é ¿ +ð¬ Ĵ +ã Ļ +𬠤 +ð ¬´ +ð« ĸ +ð ¤ +ã ¬ +ä ² +ð« Ķ +ð« ļ +è¦ģ æ±Ĥ +ä¸Ģ äºĽ +å®ŀ çݰ +èĢĮ ä¸Ķ +åĽł æŃ¤ +çͱ äºİ +åħ³ äºİ +çĦ¶ åIJİ +æİ¨ åĬ¨ +ä¸Ģ æł· +æĮī çħ§ +è¿Ļæł· çļĦ +å½¢ æĪIJ +æľī äºĽ +æĽ´ åĬł +ç»ı è¿ĩ +建 è®® +æ²» çĸĹ +ä½ł 们 +æīį èĥ½ +ä¿ĥ è¿Ľ +åijĺ å·¥ +ä½ĵ éªĮ +èĪ ĩ +åģļ 好 +ä¿Ŀ è¯ģ +æķ´ 个 +æĺ¯ ä¸Ģ个 +éĩĩ ç͍ +çIJĨ 论 +æ¯Ķ å¦Ĥ +ä¸Ĭ çļĦ +æİ¨ èįIJ +çͳ 请 +天 空 +éĥ¨ èIJ½ +åįģ åĪĨ +æĿ¥ èĩª +ä¹ĭ éĹ´ +è°ĥ æķ´ +æ¯ı 天 +è°ĥ æŁ¥ +æĤ£ èĢħ +è¿ĩç¨ĭ ä¸Ń +é¦Ļ 港 +广 åijĬ +éĿ¢ 对 +满 è¶³ +éķ¿ æľŁ +è§Ħ èĮĥ +æķ´ ä½ĵ +æĶ¹ åıĺ +æĻº æħ§ +å¦Ī å¦Ī +å¦Ĥ ä»Ĭ +åIJĪ åIJĮ +éĥ½ ä¼ļ +åĦ¿ ç«¥ +åĩı å°ij +éŁ³ ä¹IJ +ç»ı 常 +ä¸Ĭ å¸Ĥ +ä¼ĺ ç§Ģ +çļĦ éĩįè¦ģ +ä¸Ģ æĿ¡ +æµ· å¤ĸ +åı¦ å¤ĸ +ä¸Ģ å®¶ +åİĭ åĬĽ +大 åŀĭ +çľĭ çĿĢ +åĪ Ģ +幸 ç¦ı +æİ¨ 广 +åIJ Ľ +å¾ IJ +æī¾ åΰ +äºİ æĺ¯ +èĩª 身 +ä¸Ģ ä½į +åľŁ åľ° +åĬł åħ¥ +æİ¢ ç´¢ +æ¢ ģ +主 åĬ¨ +å°± ä¸ļ +女 æĢ§ +çªģ çł´ +ä¸įåIJĮ çļĦ +è¿IJ è¾ĵ +èĩª çͱ +å±ħ æ°ij +æŃ¤ 次 +çļĦ æĹ¶éĹ´ +å®¶ éķ¿ +ä¸Ģ个 人 +æ£Ģ æµĭ +åĨħ éĥ¨ +广 å·ŀ +缴 æĴŃ +ä»İ èĢĮ +è´· 款 +åı¬ å¼Ģ +æĶ¹ éĢł +人 çĶŁ +å±ķ 示 +æ¯ı å¹´ +女 人 +çļĦ æĸ¹å¼ı +æķĪ çİĩ +å±± 举 +æ¸ł éģĵ +ä¼¼ ä¹İ +æ¡Ī ä»¶ +åĪ© çĽĬ +çľĭ çľĭ +å¿ĥ éĩĮ +ç»´ æĬ¤ +å®Ŀ å®Ŀ +ç½ij ä¸Ĭ +论 åĿĽ +å°± åı¯ä»¥ +ä¸į è¶³ +æģ¢ å¤į +å¸ĥ å±Ģ +è´¡ çĮ® +ä¸ĭ éĻį +æİĮ æı¡ +çļ® èĤ¤ +å·¥ åħ· +éĩį åºĨ +åĵģ è´¨ +æİ¨ åĩº +çĶ· 人 +æī¿ æĭħ +çªģ åĩº +èĢĮ è¨Ģ +æ² Ł +åįı è°ĥ +æĺ¯ ä»Ģä¹Ī +æ± ¤ +æĴ ij +çĭ¬ ç«ĭ +çݯ èĬĤ +æī© 大 +æ´ ª +æĿ ° +çĽ IJ +ä» ģ +æ¶ī åıĬ +èĢģ 人 +åį³ ä½¿ +åįĹ äº¬ +éħį åIJĪ +é¬ ¼ +çζ 亲 +ç½Ĺ æĸ¯ +å°ı åĮº +æķĻ æİĪ +åĨ³ çŃĸ +é¢Ħ 计 +æľ¬ 人 +ä¼ ¯ +ç« ¹ +åΰ åºķ +å¸Ĥ æ°ij +åĩº åı£ +éĩĩ è´Ń +æĢ» ç»ĵ +æŃ¦ æ±ī +åĬł 大 +广 举 +æµģ ç¨ĭ +人 åı£ +å¦Ĥæŀľ ä½ł +åĩº åİ» +åĩ ī +åĨľ æ°ij +çݰ 象 +åĬĽ 度 +ç»Ļ äºĪ +åħļ å§Ķ +è¯Ń è¨Ģ +线 ä¸Ĭ +æĢİ æł· +åĦ¿ åŃIJ +ç¡® å®ŀ +ä¹ĭ å¤ĸ +éĥ½ åľ¨ +èī ¾ +çļĦ æĥħåĨµ +éĩĮ çļĦ +åĽ´ ç»ķ +æĽ´å¤ļ çļĦ +ä¾Ŀ æ³ķ +åħ¬ åĽŃ +å®¶ éĩĮ +æ¯į 亲 +ä¸į åĨį +èĭ ¹ +æ³ķ éĻ¢ +飩 åĽ½ +缸 å½ĵ +ä¸į çŁ¥ +è¯Ħ ä¼° +ä¸į ç͍ +顺 åĪ© +éĩį è§Ĩ +è´¢ åĬ¡ +ä»ĸ åĢij +åıij è¡Į +ä¸ĵ éŨ +åħ· å¤ĩ +å¹¶ ä¸įæĺ¯ +è¶³ çIJĥ +é ŀĭ +åıij 表 +æ°¸ è¿ľ +èIJ¥ åħ» +éħį å¥Ĺ +æķ´ åIJĪ +è´ º +åĽŀ çŃĶ +æĶ¶ çĽĬ +ä¹Ł 许 +è» Ĭ +æİ¥ 触 +æĶ» åĩ» +åĽĽ å·Ŀ +æĢ§ èĥ½ +åĽŀ åΰ +èħ ° +ä¹Ł 没æľī +å¼ Ħ +设 ç«ĭ +éĺ² æİ§ +æĬĢ å·§ +éĢļ 常 +è´¢ æĶ¿ +éĥ¨ ç½² +åľº æĻ¯ +æ±Ł èĭı +表 è¾¾ +åĸ · +女 åĦ¿ +èĪ ¶ +çµ ¦ +ä¼ļ åijĺ +æĪĸ 许 +äº © +举 æĸ¹ +天 æ´¥ +è¿ij å¹´ +çľĭ æĿ¥ +æ¯Ķ ä¾ĭ +å² © +éĵ ľ +çİ » +å®ŀ éªĮ +æĢĿ ç»´ +æĭħ å¿ĥ +æ² Ī +身 è¾¹ +æ·± åĮĸ +ç²¾ åĩĨ +ç§ģ æľį +æ¶Ī éĺ² +åİ» äºĨ +ç»Ĩ èĥŀ +çIJĥ éĺŁ +æĺİ æĺŁ +é£Ł çī© +å¾Ī å¿« +让 ä½ł +ä¿¡ ç͍ +å͝ ä¸Ģ +åħ¶ å®ĥ +çŃī æĸ¹éĿ¢ +å¾ĭ å¸Ī +æŃ» 亡 +æ Ł³ +ä¸Ģ æī¹ +ä¸Ĭ 涨 +æľº åľº +å½¢ åĬ¿ +æĦ¿ æĦı +éĽĨ ä½ĵ +æĸ° åŀĭ +æį٠失 +æĽ ¸ +ä¸ĭ åįĪ +æ¯ı 次 +æĪIJ å°± +åħ¬ è·¯ +èĻ « +åĴ ± +西 å®ī +æľĢ ä½³ +ç§ij çłĶ +å¤į æĿĤ +æľº åύ +çα æĥħ +çħ§ çīĩ +å¹´ é¾Ħ +è³ĩ æĸĻ +ç² Ĺ +åĩĨ ç¡® +åĬł ä¸Ĭ +åĩº çīĪ +è° IJ +å®¶ å±ħ +èĥĮ æĻ¯ +ä¸Ģ 线 +äºĭ 项 +åĬ¨ ä½ľ +ç¥ ¥ +æĢ» ä½ĵ +æĪ¿ åŃIJ +ä¹Ł å°±æĺ¯ +大 æ¦Ĥ +é«ĺ æķĪ +åIJ ¹ +æİ ĪæĿĥ +éĻĦ è¿ij +æ¡Ī ä¾ĭ +éĹ ¹ +çΏ çΏ +彩 票 +æĢ Ĵ +举 æĬ¥ +æĻ® éģį +çķĻ ä¸ĭ +è¡£ æľį +æĹłè®º æĺ¯ +åħħ 满 +æ·± 度 +æ¡ ij +æĪª èĩ³ +带æĿ¥ çļĦ +éĻ µ +æĦŁ æĥħ +èµ ļ +åĵª äºĽ +æķ´ æĶ¹ +æĪIJ çĨŁ +å¨ ľ +é¼ » +çŁ Ľ +çĽ ¾ +好 好 +第 åĽĽ +åĨł åĨĽ +è´¢ å¯Į +æľĢ 好çļĦ +车 åŀĭ +éĸ Ģ +åį³ å°Ĩ +åĪĨ 为 +éĿĴ å²Ľ +纷 纷 +ä»Ĭ æĹ¥ +å¹³ è¡¡ +å¹³æĸ¹ ç±³ +éĤ£ ç§į +åĩº çĶŁ +éĿĴ æĺ¥ +人 群 +人 å·¥ +ä¹ĭ ä¸ĭ +æ¹ĸ åĮĹ +åľ¨ æŃ¤ +åįļ 士 +æĹ¶ åĪ» +æ²³ åĮĹ +æĶ¾ å¼ĥ +éĢļ éģĵ +森 æŀĹ +çĸ Ĩ +æķ ¸ +èĬ ³ +æīĵ åĩ» +æĽ ¹ +åĮĸ åѦ +æĥ³ 象 +ä¸ĩ 人 +è´¢ ç»ı +åħĥ ç´ł +ä¼ļ 计 +åħ¨ ä½ĵ +æĦ Ľ +é«ĺ ä¸Ń +æľº éģĩ +声 éŁ³ +æĹħ è¡Į +æµ © +æŁ ± +å°ij å¹´ +åĽ½ å¤ĸ +èijĹ åIJį +çĶŁ åŃĺ +å§ ľ +带 é¢Ĩ +é¢ľ èī² +ä¸Ĭ ä¸ĭ +产ä¸ļ éĵ¾ +æĽ´ 好çļĦ +å² Ń +ä¼ĺ æĥł +便 æĺ¯ +åħ§ 容 +ä¸Ģ åıª +çIJ ´ +梦 æĥ³ +ç§Ł èµģ +å¼Ģ åIJ¯ +è´Ń çī© +åĮħ åIJ« +åĪ© çİĩ +èµ· äºĨ +æľī åĬĽ +éĤ£ éĩĮ +审 æī¹ +对 æīĭ +çݰ éĩij +天 çĦ¶ +çĽ Ĵ +çĪ ½ +å¿ħ çĦ¶ +åĮĸ å·¥ +ä¸ĵ åĪ© +åķ ¡ +å¼Ģ å¿ĥ +人 ä½ĵ +éģĵ 士 +æĢģ 度 +空 è°ĥ +æĭĽ åķĨ +å§ » +第 äºĶ +æ£ Ĵ +ä¸Ģ ç³»åĪĹ +åį± æľº +转 åıĺ +åľº æīĢ +é¸ £ +æĪ¿ éĹ´ +éĢ ¼ +è¯ķ çĤ¹ +对 å¤ĸ +åĩº åı° +åľ¨ è¿Ļ +åİĤ å®¶ +å·¨ 大 +ç®Ģ ä»ĭ +çľĭ äºĨ +åħļ 建 +æĮĩ æĮ¥ +çŁ³ æ²¹ +ä¸į åı¯èĥ½ +èİ ² +ä¸į 太 +åĪĽ æĦı +第 ä¸Ģ个 +è´µ å·ŀ +è¿ĩ äºĨ +æľ¬ æĿ¥ +éģĵ å¾· +çŃĶ æ¡Ī +éĻ ¶ +ä¸Ģ è·¯ +èĤ ĸ +æ¸ħ æ´ģ +æľī æľº +åIJį åįķ +æĿ ± +åij¼ åIJ¸ +ä¸ Ī +ç¦ı 建 +è¯ķ éªĮ +å¼ķ åıij +ä¹Ł 没 +ä¸į ä½ı +çĨŁ æĤī +èIJ ¬ +ä¸į èī¯ +çł ĸ +èĩ´ åĬĽ +çѾ 订 +åIJ Ĭ +ä¾ ¯ +çĺ ¦ +å§ij å¨ĺ +æĸ ¤ +妻 åŃIJ +æĺ¥ èĬĤ +çĪ ¬ +æĽ Ŀ +çĥŃ æĥħ +éķ¿ æ²Ļ +èIJ¥ éĢł +éħ · +éĵ Ŀ +åŁºæľ¬ ä¸Ĭ +åij¨ åĽ´ +ä»Ģ 麼 +认 åı¯ +åĪĨ åŃIJ +ä¸Ģ æĸ¹éĿ¢ +è½ ´ +å¼ · +马 ä¸Ĭ +éĽ ¾ +èĩ £ +å° ¿ +çĶŁ æĦı +å®ī å¾½ +ç¥ŀ ç»ı +åĩº å¸Ń +èᝠåĵģ +çIJĨ çͱ +åįı åIJĮ +æµģ åĬ¨ +åıij åĬ¨ +åĿļ å®ļ +表 æĺİ +åIJİ éĿ¢ +ä¹ī åĬ¡ +å¦ ĸ +æľī åı¯èĥ½ +å¹´è½» 人 +大 éĻĨ +å² ³ +ä¸į èµ· +çŀ¬ éĹ´ +ä¸įå¾Ĺ ä¸į +çѾ 约 +åIJĪ æł¼ +åħļ æĶ¯éĥ¨ +æµİ åįĹ +便 åĪ© +éļı æĹ¶ +å¥ ī +ç§° 为 +产 æĿĥ +åIJ ķ +çĽ Ĩ +课 åłĤ +ç· ļ +æ£ ī +线 ä¸ĭ +èĩª è¡Į +举 æİª +åݦ éŨ +èĩª ä¿¡ +å½± è§Ĩ +ä» Ķ +çĶŁæ´» ä¸Ń +æĿĥ çĽĬ +çϽ èī² +å°± ä¸į +è¿Ľ å±ķ +æ¯ı æĹ¥ +ä¾Ľ ç»Ļ +æĿĥ åĪ© +æĹł æķ° +çIJĨ è´¢ +ä¾Ŀ æĹ§ +ä¸Ĭ åįĪ +è¯Ĩ åĪ« +çĽĪ åĪ© +çł Ĥ +许 åı¯ +åIJĮ äºĭ +åĺ Ľ +éģ ¸ +çĿĢ åĬĽ +éŨ åı£ +ä¸į å¤ļ +åħ¶ 次 +ç¢ § +çī© çIJĨ +åĨħ å¿ĥ +çϾ å§ĵ +æĢ» 绣 +å¹² åĩĢ +积 ç´¯ +åıį é¦Ī +æłij ç«ĭ +社 交 +ç§ © +åįģ ä¸Ģ +éĤ ĵ +驱 åĬ¨ +å±ķ è§Ī +èĪĴ éĢĤ +åŁº åĽł +å·® å¼Ĥ +转 让 +å°ı å§IJ +æł· åŃIJ +ç¿ Ķ +é«ĺ åħ´ +å½±åĵį åĬĽ +æīĭ ç»Ń +缸 åIJĮ +缸 åºĶ +æĻ Ĵ +è§ Ģ +å¸Ĥ å§Ķ +èĬ ¯ +å±ķ çݰ +åľ° çIJĥ +éĤ ª +ä¸Ģå®ļ çļĦ +åħģ 许 +ä¿¡ ä»» +æī ij +éĻ¢ æł¡ +ç®Ģ ç§° +åģļ æ³ķ +ä¹ĭ è·¯ +æĹĹ ä¸ĭ +èħ Ķ +æ¶Ī 失 +ä¸ĸçķĮ ä¸Ĭ +åŁİ 乡 +èĪŀ åı° +å¾Ī 大çļĦ +绣 çѹ +åħ¬ å¹³ +èĤ ¾ +çļĦ 好 +æ± ģ +çľ¼ åīį +éĽ £ +å¹ ½ +åħ± 产 +主 åĬŀ +å¤Ħ ç½ļ +åº Ļ +éģĵ çIJĨ +å¼ µ +æİ¥ çĿĢ +çĮ İ +çģ Į +çͱ æŃ¤ +人 åĬĽ +æµģ è¡Į +ä¾ ł +åı¯ä»¥ 说 +èĴ ĭ +å½¢ æĢģ +æĹ¥ åŃIJ +æ¼ Ĩ +çķĻ åѦ +缸 éĹľ +æľĢ å¤ļ +åĩŃ åĢŁ +åħ¬ 交 +æĮĸ æİĺ +æĿĤ å¿Ĺ +主 人 +éļľ ç¢į +æł¡ éķ¿ +æĸ¹ ä½į +ä¸Ĭ çıŃ +å¤ļ åħĥ +è ĥģ +éŃħ åĬĽ +èĮ Ĥ +åħħ ç͵ +强 大 +çĥ ¤ +å¥ĭ æĸĹ +å®ŀ ç͍ +éĺ ģ +ç»Ļ äºĨ +æľ¬ ç§ij +æł ĭ +æĭ ¨ +æķĻ ç»ĥ +éĥ½ çŁ¥éģĵ +æ¯ķä¸ļ çĶŁ +ç¢ Ĺ +åŀ Ĥ +è® ¼ +å®ģ æ³¢ +åѦ èĢħ +è°¢ è°¢ +åŁİ éķĩ +æĢİä¹Ī åĬŀ +éģ Ķ +æĪIJ 交 +æ½ľ åĬĽ +åį § +æĸ° å¼Ģ +éħį å¤ĩ +主 åĬĽ +åij³ éģĵ +çĥ Ĥ +é£ŀ è¡Į +å« ģ +大 大 +ç»Ļ 大家 +å¤ĸ éĿ¢ +éĨ ī +åıij è¨Ģ +æĹ© é¤IJ +åIJĦ èĩª +å® Ļ +èᣠèªī +æĬ« éľ² +é¡ ŀ +åĨħ çļĦ +èĤ ª +è¾ IJ +æ³ µ +æĬ Ľ +æĺŁ æľŁ +ä¸Ģ 带 +çĶŁ ç´ł +ç»ı éĶĢ +åĩ ¶ +åľ° ä¸Ĭ +åij½ è¿IJ +åĵ ² +ä¸Ĭ åİ» +æĸĩ çī© +è¯ ij +æĮ¯ åħ´ +éķ¿ æĹ¶éĹ´ +ç¥ Ń +åIJĪ èĤ¥ +è¿Ŀ è§Ħ +èģ ª +ä½İ äºİ +éĢĤ å½ĵ +æľī åºı +æľ¬ ç½ij +çķĻ è¨Ģ +æĥ³ æ³ķ +çѾ ç½² +å§ ļ +æĢ§ æł¼ +èĴĻ åı¤ +æŁ ı +åŀ « +åѦ åİĨ +ä»ħ ä»ħ +讲 è¯Ŀ +éĶ IJ +æĢ ĸ +åī ª +èĭ į +åIJ ĵ +强 çĥĪ +åģ¥ åħ¨ +çĸ ¯ +åı¤ 代 +å¥ Ī +ä¸į çĦ¶ +乡 éķĩ +æľĭåıĭ 们 +åĤ ħ +èģ ½ +个 æĢ§ +æ³ķ è§Ħ +å°ı éķĩ +çĶ» éĿ¢ +第 åħŃ +ç¶² è·¯ +åīį æĻ¯ +åIJ¬ 说 +ä¼ł åªĴ +æĿ¡ ä¾ĭ +åĪ« çļĦ +ä¸į æĩĤ +顾 éĹ® +强 度 +éĺ¿ éĩĮ +èµ° åĬ¿ +å¸ ½ +çļĦ ç¡® +åĮº åĪ« +éĮ ¢ +主 管 +ä¸Ģ çľĭ +æĸ ľ +åŃĺåľ¨ çļĦ +ä» ² +åᱠ害 +éĵ Ń +游æĪı ä¸Ń +éħ ± +é¾Ļ 头 +人 å¿ĥ +éĢĢ ä¼ij +æµı è§Ī +åĬ « +éĺ² æ²» +ç® Ń +å± Ī +è¾½ å®ģ +å£ ¤ +è¿İ æĿ¥ +éŀ į +ç͍ æĿ¥ +大 åľ° +ä» ° +éĢļ 讯 +å¼Ģ å·¥ +è£ ¤ +å¦Ĥ åIJĮ +éª ¤ +éĺŁ åijĺ +è½ © +ç¾İ æľ¯ +èĻ Ł +åIJĮ ä¸Ģ +åľ ĸ +书 æ³ķ +æīĵ åį° +åIJ« æľī +éĽĨ æĪIJ +éĹ · +å¸Ĥåľº ä¸Ĭ +æĹģ è¾¹ +åľ° æĿ¿ +产çĶŁ çļĦ +ç² ¤ +éĩį ç»Ħ +è¡Ģ æ¶² +çŃ ĭ +åĬŀ äºĭ +常è§ģ çļĦ +ä¸Ĭ åįĬå¹´ +å±ı å¹ķ +åIJī æŀĹ +å· © +åĸľ çα +ç¿ ł +ä¸ī ç§į +æ¡Ĩ æŀ¶ +举 èİŀ +çĶĺ èĤĥ +èĬ ¬ +åĽ¾ 书 +åĩ¤ åĩ° +æ°Ķ åĢĻ +å° ´ +å° ¬ +两 天 +è¾ħ 导 +åĢŁ 款 +æĹ¥ èµ· +æ´ Ĵ +ä¸Ģ 度 +è¹ Ī +æ½ Ń +æī ĩ +çĻ ľ +æĸ° åħ´ +åĤ ² +诸 å¤ļ +è´ ª +éĻ· åħ¥ +èĪ Ł +èĤº çĤİ +ä¸Ģ æł·çļĦ +åİ ĺ +åľ° çIJĨ +æĬķ æ³¨ +éļ Ĭ +åħī ä¼ı +ä¿Ŀ åģ¥ +åħ Ķ +åħ¬ åĬ¡ +æīĵ çł´ +çĶ· åŃ© +åĬ³ åĬ¡ +ä½ł ä¼ļ +ç͍ åľ° +æº ¢ +åıij è¾¾ +èĤ ļ +è¿ĩ äºİ +èĩ Ĥ +éĢĻ æ¨£ +è½» è½» +ä¸Ń åħ± +åIJĦ åĽ½ +åĶ ĩ +å®ŀ ä¹ł +èĻ ¾ +æ§ ½ +ä¸į ä¸Ĭ +åħį çĸ« +åįł æį® +å·¥ ä¼ļ +åĽ Ĭ +èĪª 天 +åı¯ çα +æĸĹ äºī +çĺ ¤ +å¦Ĥ æľī +éĽ ĸ +对 æĪij +åĩº ç§Ł +好 çľĭ +太 大 +æ°´ åĪ© +åĬ¿ åĬĽ +åħ¨ æ°ij +ç½ ¢ +èµ¢ å¾Ĺ +ç͵ ä¿¡ +车 éĹ´ +æĻĤ åĢĻ +å°ij æķ° +éĵ ¸ +åħ³ èģĶ +ä¸įä»ħ ä»ħ +为 æĤ¨ +åĴ ¸ +æľº åĬ¨ +è£ Ļ +åĵį åºĶ +éģ ł +è² · +ç© ´ +å¢ ħ +éĶ ¡ +çµ Ħ +çģ« è½¦ +è³ĩ è¨Ĭ +åĨ³ èµĽ +污 æ°´ +èª ŀ +å´ Ľ +ç´§ å¯Ĩ +缺 å°ij +å¤ļ 人 +æĢ» 书记 +éĶ Ī +èij Ľ +å¿ĺ è®° +éĻĮ çĶŁ +éķ¿ å¤§ +åħĪè¿Ľ çļĦ +ç¡ ħ +åıij æĺİ +å©´ åĦ¿ +æīİ å®ŀ +èĽĭ çϽ +ä¸Ģ çϾ +缮 åħī +æ ħĮ +åĬł æ²¹ +åIJ ŀ +ä¸Ģ 群 +ä¸Ń ä»ĭ +å¸ ĸ +å¿ Į +èģĮ èĥ½ +广 æĴŃ +çĽij å¯Ł +ç§ĺ å¯Ĩ +çĭ ® +è¿Ļ æĿ¡ +éĢ ¢ +æĢ ¨ +åįģ åħŃ +è© ¦ +说 åΰ +åĩĿ èģļ +æĮĩ 示 +æ° ¢ +å¼ ĺ +éĺ Ģ +æĸ © +éł ħ +ä¸Ģ å¼Ģå§ĭ +æİĴ è¡Į +åľ¨ æĪij +纪 å½ķ +æĬ Ħ +æł ª +说 æ³ķ +ä¸Ń èᝠ+好 å¤ļ +åıª ä¸įè¿ĩ +çķĻ åľ¨ +个 å°ıæĹ¶ +认 çŁ¥ +çķ « +è§ģ è¿ĩ +å°ı å¾® +ä½Ľ å±± +çľ ¾ +讲 è¿° +æ¢ ³ +ç§° åı· +æĹ¥ æĻļ +è¢ ĸ +åķ ¤ +æľª ç»ı +æľĢ æĹ© +æī® æ¼Ķ +è¡Ģ 管 +çº ± +æĥħ èĬĤ +第 ä¸ĥ +æį § +ä» Ĺ +æ¿Ģ çĥĪ +æĹł 线 +ä¸į 容æĺĵ +å¼Ģ å¹ķ +æĸ° çĶŁ +ä¸ĵ 注 +èij ± +åįĹ æµ· +çĩ Ł +èµ· ä¾Ĩ +æ´¾ åĩº +åĦ Ĵ +ä¾ ¨ +è¼ ĥ +åįļ è§Ī +éĢ ¾ +åĮ Ģ +ç»ıæµİ åѦ +æ¸ Ĺ +ä¿Ŀ èŃ· +çī º +çī ² +çİ « +çij ° +æľĢåIJİ ä¸Ģ +æĶ¿ åĬ¡ +æ§ Ľ +èĻķ çIJĨ +éļIJ æĤ£ +æī¿ åĮħ +æ¥ µ +æ¡ © +çĽ ² +导 åIJij +èĩ´ å¯Į +ç¼ Ĩ +æģĭ çα +ä¸į åĬ¨ +ç»Ļ 人 +å· ¢ +表 æĥħ +举 åįĹ +åĨħ å¤ĸ +è¾Ī åŃIJ +åı ī +åįļ ä¼ļ +åĬŁ æķĪ +æ¸ ´ +å± ¬ +æİĴ éϤ +éĢ Ľ +ä¸Ģ ä¼ļ +ä¸į å¼Ģ +å¼Ģ å¥ĸ +é»ij é¾Ļ +é»ijé¾Ļ æ±Ł +å¿« ä¸ī +度 åģĩ +åĿ ¤ +éĤ® ä»¶ +æĩ Ĵ +ä¾Ľ ç͵ +å» £ +好 è¯Ħ +ç§ĺ书 éķ¿ +æĪĺ åľº +好 å¥ĩ +ä¾µ æĿĥ +æĨ ¾ +æľĢ åĪĿ +æī¹ åıij +åİ ķ +è¼ ķ +æŀ ¯ +ä¸ļ åĨħ +è´Ń æĪ¿ +ä¸į åľ¨ +纪 å§Ķ +æīĢ éľĢ +å¸Ĥ éķ¿ +è³ ½ +å¼ķ æĵİ +çģµ éŃĤ +éĬ Ģ +æ» ¤ +çĿ IJ +å¤ļ 项 +åĽŀ 头 +èī ĺ +å¤į å·¥ +éĥ¨ ä»¶ +ç´§ ç´§ +æŁIJ ç§į +使 åħ¶ +æĸ° 人 +æŀ ļ +æ³ķ å®ļ +å·´ å·´ +æ¶µ çĽĸ +ç¨ » +æĭ ¾ +æĻ ķ +è½ ¿ +éĢļ è¡Į +åĵ Ģ +æ³ Ĭ +温 馨 +éĽĨ èģļ +çĨ Ļ +åĩ ij +åįģ ä¸ĥ +æ°Ķ æģ¯ +æıIJä¾Ľ çļĦ +æ³ ³ +奥 è¿IJ +çģ¾ å®³ +åĩĢ åĮĸ +è·¨ è¶Ĭ +åĵª æĢķ +éŁ ¿ +å¢ŀ æ·» +çĦ Ĭ +æ®ĭ çĸ¾ +ç¢ Į +æĤ Ķ +è§ģ è¯ģ +è¾ĸ åĮº +å¿ĥ èĦı +éļ § +åį ¸ +åı¯èĥ½ æĢ§ +æľī è¶£ +åī¯ ä¹¦è®° +åĮĸ å¦Ĩ +ä¿ Ĥ +æ£ ļ +éĨ ĩ +带 头 +éł Ī +追 ç©¶ +æij Ķ +è¿Ļ éĥ¨ +ä¸į 论 +ç¥ ¸ +å ³» +éģ ķ +çĶŁ èĤ² +å¤ ł +å¤ĸ 交 +è¯Ħ 为 +ä»İ å°ı +å°ı å°ı +é ¥¿ +æĴ ¼ +è·¨ å¢ĥ +被 åijĬ +åįĹ å®ģ +身 å¿ĥ +åĨį çĶŁ +æīĢ è¯´ +æĹ¶éĹ´ åĨħ +åĪĹ åħ¥ +éĿĴ æµ· +çα 好 +çª Ħ +èĪ Ī +è¿ĩ 渡 +æ¿ Ł +éĽ Ģ +审 è®® +åĽ½ èµĦ +æŃ¥ ä¼IJ +轨 éģĵ +ä¿¡ 念 +ä¸ī åĪĨ +çĨ ¬ +åѵ åĮĸ +ç¼ ł +éĥ Ĭ +èĪĴ æľį +纪 æ£Ģ +ä¸Ģä¸ĭ åŃIJ +鼻 話 +è² ł +éĴ ¥ +åĮ Ļ +çĹ ´ +è¶ ģ +ç» £ +çĪ µ +è½ ° +éª Ħ +å§ ¨ +æĭ ĺ +çĮ ´ +è® ¶ +è¿Ļ 座 +çį ¨ +æ·ĺ æ±° +çĹħ ä¾ĭ +æ²Ļ åıij +è§Ĩ 为 +头 æĿ¡ +å¿ħè¦ģ çļĦ +åı¯ è°ĵ +è¯Ŀ 说 +ç¯ Ħ +æĹ© çĤ¹ +æŀ¢ 纽 +ç¾ ¡ +çα åĽ½ +çªģ åıij +éĢ Ĭ +æ½ į +èᣠèĢĢ +èŁ ¹ +æ¦Ĥ çİĩ +å¾Ī ä¹ħ +æĥ ķ +è¨ ´ +åľĨ 满 +çļ ± +åĪĨ æ³Į +åħħ è¶³ +çľĭ æ³ķ +è¾ Ł +æĭ ¦ +æĭ © +对 åºĶ +为 æł¸å¿ĥ +èħ Ĭ +å¤ļ ä¹Ī +æµ ij +å®ı è§Ĥ +èĦ ĸ +åIJĪ èµĦ +çĶŁ 涯 +å®ŀ è´¨ +ä¼ĺ çĤ¹ +ç͍ æ°´ +寿 åij½ +æ² « +åIJ ģ +è© ¹ +åĽ½ éĺ² +å´ © +åĿ İ +èĨ ı +ä¸Ģ è½® +éģĹ äº§ +æ¹¾ åĮº +ç» İ +åįķ 纯 +æ¾ Ħ +åīį åĪĹ +身 å½± +é»ĺ é»ĺ +æį ī +çĴ ° +èı Ĭ +æĢ ľ +åħĭ æĢĿ +æĢ» å±Ģ +çĩĥ æĸĻ +ä¸ļ æĢģ +åIJĦ æł· +åĴ ½ +åĩº èī² +åĪĿ å¿ĥ +åı Ľ +çłĶ 讨 +è¡ « +åİĨ ç¨ĭ +ç¦ ½ +è¶³å¤Ł çļĦ +èį Ĩ +çľĭ å¾ħ +è´ © +åĨ³ å¿ĥ +è£ ¹ +å¸Ī èĮĥ +åŀ Ħ +æĿ ł +åĩ ¸ +çĬ¹ 豫 +çĥŃ è¡Ģ +åIJĪ ä¼Ļ +éħ µ +èIJ½ åľ¨ +åįł åľ° +è¡ ¬ +èĵ ī +æĦ ¤ +æ¸ Ĭ +åĪĨ æķ° +ç¬ij çĿĢ +太 å¹³ +çĤ « +æİ¨ ä»ĭ +æĸ¯ åĿ¦ +å½¢ 容 +æĵ Ĭ +æĦŁ åħ´è¶£ +åĨĽ 人 +åĩĮ æĻ¨ +对 çħ§ +åıij çĹħ +å· ¾ +èĪ ī +æª ¢ +ç¬ij äºĨ +ç¡® è¯Ĭ +è´Ł åĢº +壮 大 +æĪ ļ +äºĴ èģĶ +èª ² +èħ ¦ +æĹ ± +åıĹ æ¬¢è¿İ +åį ī +éĻ¢ 士 +æ© ¡ +ä¸Ģ 对 +è¾ ± +æ² Ĥ +åı² ä¸Ĭ +æIJ ı +å´ ĸ +代 è°¢ +ç£ · +é¡ ĺ +æµ ĩ +常 ç͍ +åį ij +åĩº åĽ½ +è¯ ł +稳 æŃ¥ +ç»ı 纪 +å¤ļ å¤ļ +æīĢ å¾Ĺ +为 主é¢ĺ +ä¸Ģ åĪĨ +æł ½ +é¡ § +çº ² +åĥ ħ +å£ ĵ +åĦ ª +ç¿ ° +æİ Ģ +人 为 +åª ³ +æ´ ½ +èĿ ¶ +å¤į åħ´ +ä¼ļ å½±åĵį +åIJĦ çķĮ +éĤ£ ä¸Ģ +é¢ ¤ +çĢ ı +çĢı 覽 +å¯ ŀ +åı¯ æĢķ +åį³ æĹ¶ +çķ ´ +ä¸ĭ åįĬå¹´ +ç¬Ķ è®° +éĻĦ åĬł +çĥŃ æ°´ +å¥ ¸ +ç£ ħ +æĿ ī +æ¸ħ åįİ +éĸ ± +ç° ¡ +å¤Ħ å¤Ħ +åIJĪ éĩij +æ²³ æµģ +ç´ ° +è´Ł éĿ¢ +çļĦ 羣å®ŀ +åύ 械 +èĴ IJ +西 äºļ +å· ħ +ç² ¹ +åİŁ æĸĩ +æŀ ķ +è¡Ģ åİĭ +åļ ´ +å¸ ĺ +åĨ Ģ +æĮ « +ç͵ è·¯ +å°ı ä¼Ļä¼´ +èĿ ´ +æľĢ å¿« +æĭ Į +å® ª +æĸ · +ç¿ ħ +åĴ ³ +åĹ ½ +ç¾ ŀ +躺 åľ¨ +èµĽ 车 +æ² IJ +éĻIJ 度 +为 ä¸Ģä½ĵ +èĴ ľ +å¹ « +æIJ ħ +åĭ ĭ +åī ĸ +纳 ç¨İ +éķ¿ æķĪ +ç½ ķ +åī¯ æľ¬ +ç© į +éĴ © +ç¹ ¼ +åĽ½ åľŁ +è¼ ī +ä¸į å¿ĺ +èѦ 示 +çģ ¿ +å¿ĥ å¾Ĺ +æĦ ļ +忽 çķ¥ +åĽŀ äºĭ +åįł æľī +æ· Ħ +çī ¡ +çĽij äºĭ +ç¿ ¡ +éĴĪ对 æĢ§ +çª ĥ +è£ ½ +èĨ Ŀ +ç³ Ł +港 æ¾³ +太 太 +æ¾ ¡ +ç»Ĩ åĮĸ +åĶ® åIJİ +å®ŀåľ¨ æĺ¯ +ç« £ +çį ² +å̾ åIJij +å¼ķ ç͍ +é¹ ħ +ç¬ij 容 +ä¹IJ è¶£ +æ°ij æĶ¿ +éŨ æĪ· +å± ģ +è¿· 失 +éĶ Į +å°ı 康 +åĭ ī +æ³ ¼ +ä¾ĭ åŃIJ +ä¸ī ä½į +å» ł +èĶ ĵ +广 éĺĶ +èĢ į +èĢģ èĻİ +åĭŁ éĽĨ +èĦļ æŃ¥ +æĭ ¯ +åŃĹ åı· +çĦ ° +é¢ ł +èļ Ĥ +èļ ģ +é£ ¯ +人 æĢ§ +æĴ ° +åİ ¢ +å±Ģ éĻIJ +æľª æĪIJ +åĵª åĦ¿ +大 åıij +ä¸į å®ļ +å¾ģ æ±Ĥ +éĥ µ +åĢº æĿĥ +çα ä½ł +èº ģ +ä»ħ ä¾Ľ +è¿ľ å¤Ħ +éĨ Ľ +åĥ µ +积æŀģ æĢ§ +æİ ¡ +åīį ä¸ī +äºİ ä¸Ģä½ĵ +çŀ Ħ +çĿ ģ +æ² ¸ +åħ± èµ¢ +éĢĢ å½¹ +è´Ŀ å°Ķ +æİ ı +æĪ ² +è¡ į +éĶ Ĥ +ä¸ĩ ä½Ļ +ç§ij åĪĽ +æ¼Ķ åͱ +欧 åħĥ +æ·¡ æ·¡ +éĿĴ å±± +èĹ Ŀ +ç» ½ +令 çīĮ +éĽĨ 群 +ä½ľ çī© +çĢ ij +å¤ ¯ +ç½ij 游 +åħ« 大 +éª ļ +èª ĵ +ä¼ļ å±ķ +åħļ åı² +æ£Ģå¯Ł éĻ¢ +åĸ ĺ +éĺ ± +èĢĮ åĩº +éĢļ 车 +éĴ ĵ +æĥħ 人 +æ¸ Ľ +ä¸Ń ç§ĭ +çĪ Ń +åıª åī© +æĺ Ķ +éĩİ çĶŁ +ç¡ « +èIJĿ åįľ +æĬµ æĬĹ +çĻ« çĹ« +éĻ Ģ +èĶ ļ +å¸ ľ +满 满 +èı ± +éļĨ éĩį +æĺŁ çº§ +æ½ ĩ +åħ¬ åħĥ +è° £ +æ¯Ķ äºļ +æ¡Į åŃIJ +èµ £ +è² ¼ +æĦ¿ æľĽ +é¡ ½ +æ´¾ éģ£ +ç¥ Ľ +åª ļ +éĺ ľ +èij « +èĬ ¦ +æ³ » +å¡ Į +çĭ Ń +å»ī æĶ¿ +å¥ij æľº +æĹĹ èΰ +æĥ « +严 åİī +åıĭ æĥħ +å¦ Ĭ +å¨ ł +åĵª å®¶ +èĨ ¨ +è¶ Ł +æĮ ª +èĻ IJ +é łģ +çŀ © +éº Ł +ç¨ £ +èģĶ éĢļ +åı ® +çİĭ èĢħ +ä¸į ç¡®å®ļ +ç ijľ +è° İ +çī¢ è®° +ç¢ ¼ +æĬ¤ èĤ¤ +é¡ · +çĦ ķ +åģļ 强 +éļ± ç§ģ +éļ±ç§ģ æ¬Ĭ +åıĹ å®³ +ä¸į çͱ +çĥ ¹ +é¥ ª +é© ³ +ä¼ ½ +ä¸Ŀ 绸 +è¥ Ħ +åįģ ä½Ļ +éº Ĺ +æ¬Ĭ åĪ© +èģ ŀ +åı¤ èĢģ +éģ ı +åIJĦ å¼ı +å°± è¡Į +åħ¥ å¢ĥ +ç ĥģ +èľ ĺ +èĽ Ľ +çº ¬ +çŁ « +è» Ł +æ´Ĺ è¡£ +æĦ § +é¢Ħ æ¡Ī +éľ Ĩ +æ·± åİļ +éĺ¿ æĭī +åĨĻ åŃĹ +åį ¦ +éķ Ģ +模 æł· +åĤ į +æIJ į +èĸ ¯ +åł ħ +åħ¬ 积 +è¨ İ +ä¼ł æŁĵ +æ¯ ¯ +çIJĨ å·¥ +åĨ· éĵ¾ +ç«ĭ æĸ¹ +æ¢ Ń +åľ£ è¯ŀ +综 èīº +çİ© ç¬ij +æĥ³ ä¸įåΰ +æijĩ 头 +æ· ¹ +åģĩ æĹ¥ +åĢ ĺ +èĢ ½ +èİ ĵ +åŁ · +èĩª è´¸ +åįĬ 天 +æª Ķ +æ¾İ æ¹ĥ +éķ ij +ä¸ « +éĩĮ ç¨ĭ +å¼Ģ èįĴ +èı ı +å®Ŀ è´µ +èŃ ¬ +åķ Ł +æŁ ł +æª ¬ +é© Ń +æ± Ľ +çĨĬ çĮ« +èķ ī +éļı ä¹ĭ +å± ij +è¾ĥ 强 +èĥ ³ +èĨ Ĭ +éĿĻ éĿĻ +åĴ ª +æĭĽ åij¼ +代 è¨Ģ +ä¿¡ ç®± +è£ħ éħį +æĤ į +åįķ 车 +èIJ İ +å¤ļ 彩 +éĻ ¸ +ä»İ 严 +æ© Ħ +æ¦ Ħ +éĢ ® +éĩĮ æĸ¯ +å§¿ æĢģ +太 æŀģ +éĩ Ŀ +æº ī +è¿ Ń +ç§ ¸ +ç§ Ĩ +å·¥ å§Ķ +æ± ķ +èģ Ĩ +ä½ ¬ +ç¼ ħ +çĶ ¸ +åī¯ å±Ģéķ¿ +éĹ º +èª ¤ +è¤ IJ +ä¸į éĻIJ +èħ ķ +åij ķ +çŁ ¶ +åĨľ å®¶ +管 å§Ķä¼ļ +é¥ º +èĬ ľ +æ¾ Ī +è© ¢ +å¨ģ å°¼æĸ¯ +ä½ķ åĨµ +å°ı ä¼Ļ +奢 ä¾Ī +è¿Ļ ç¯ĩ +è¯ µ +竳 ç¨ĭ +ç´ Ģ +éIJ ĺ +éĤ ¢ +ç³ Ļ +ç¼ Ģ +ä¹ Ĵ +ä¹ ĵ +çī¢ åĽº +åĿ ŀ +å¼ Ī +ä¾ĭ å¤ĸ +å» ³ +è§Ħ 竳 +èĬ Ļ +ç¯ · +èº ¯ +æł Ī +åĿļ å®ŀ +åŁº 建 +çĿĢ çľ¼ +ç· ´ +èij © +ç¼ ļ +æ¦ Ĩ +主 åĭķ +ç¥ Ģ +äºĴ éĢļ +å°¤ 为 +å® Ľ +éª ¼ +æ± ² +ä¾ ĥ +æĤł ä¹ħ +æij § +æĭ ĩ +é« ĵ +éº Ĵ +éĻ Ľ +æŀ ¸ +æĿ ŀ +è´ ¬ +å°ı é¾Ļ +åĵ ® +èĵ¬ åĭĥ +åĮ Ī +çķľ çī§ +å¨ © +个 å¤ļ +æ² ¥ +æĺ § +çĦ ļ +æĬij éĥģ +çĸ ¡ +èĺ ij +éģİ ç¨ĭ +æ© ± +éĿ ĵ +大 çIJĨ +é« ¦ +åĪĨ 辨 +æ¸ ¤ +çĸ ¤ +åĬ¨ èĥ½ +å¼ł å®¶ +ä¸ĩ åįĥ +æ» ¥ +é¥ ¥ +åºŁ å¼ĥ +å¸ ³ +æ¼ ³ +è± IJ +ä» ij +å« ī +å¦ Ĵ +çŀ Ĵ +è¡ ħ +çĭ ¸ +å¾ģ ç¨ĭ +éĤ ¯ +éĥ ¸ +ç¥ Ī +ç¥ · +è¶ ´ +ç»ĵæŀĦ æĢ§ +è§Ĩ åIJ¬ +è¬ Ŀ +çĴ Ģ +çĴ ¨ +åĩº å¤Ħ +è¯ Ģ +å¾ ĺ +å¾ Ĭ +çľ ¨ +åĸ ĩ +åı Ń +åĺ ² +çķ ¸ +å¹² äºĭ +æļ § +æ² Ľ +åĦ Ħ +å» ĵ +åİ¿ éķ¿ +èĥ ļ +çIJ ¢ +çŃ · +éĩ ĭ +ä¾ ® +åIJ © +åĴ IJ +åĮ ¿ +æĬ¬ èµ· +æ³ £ +æ¶ ¤ +éº ½ +æĽ Ļ +åī¯ éĻ¢éķ¿ +åħļ åĴĮ +æķ£ åıij +润 æ»ij +åĵ º +æĥ ¬ +漫 éķ¿ +ä¸į æĩĪ +åŁ ł +åĹ ĵ +èĢģ çĪ· +è® ½ +æĪĺ ç»ĦåIJĪ +æ£ ł +åħ¨ åŁŁ +èł ¢ +è¯ ¡ +åīį çŀ» +æķ Ľ +ä¸Ģ å°ģ +å¹ Ĥ +èİ Ĩ +è¯Ŀ è¯Ń +ç»Ĩ åĪĻ +å± ¿ +åµ Į +éĢ į +åĺ ± +æ¸ ² +çĥ ¯ +çĿ ¹ +é¦ Ĵ +èħ ¥ +æĬĹ åĩ» +çĿ « +èį Ķ +éļ İ +æ³ī æ°´ +è¬ Ĥ +ç Ĥ¬ +åĩı æİĴ +è¸ Ĭ +è ·» +æ· Į +éľ ¾ +å¥ĩ 纳 +å¯ Ŀ +æ¤ İ +æŁ ¬ +æĸ¯ åŁº +åħ¬ ç«ĭ +è¨ ĵ +é£ Ļ +é© ¿ +åĤ µ +èĽ Ļ +ç¯ĩ 竳 +åĪĨ æĶ¯ +ä¸Ĭ å¹´ +çŃ Ŀ +ç¼ ¤ +èĢģ æĹ§ +åĻ ¬ +æľ ¦ +èĥ § +æ¶Ī è²» +æĵ Ķ +æ¦ ´ +æ¿ Ĵ +ç³ ¯ +æ³ ¸ +æį Ĩ +ç» ļ +èµ İ +çIJ IJ +èµ Ĥ +æħ ® +æ² Į +çĦ Ļ +æĴŃ æĬ¥ +æ· ĩ +åĪĩ åħ¥ +çij ķ +çĸ µ +éģ ´ +ç¨ ļ +ç© © +èŀ ĥ +æ£ ķ +æĨ § +æĨ ¬ +ä¼ º +æ¯ Ĺ +æį į +æĬ ī +ç´ Ĭ +å¼ Ľ +æĭ Ń +æĹı èĩªæ²» +åĿ · +ç« ¶ +è© ³ +è¿Ħ ä»Ĭ +è° ´ +çŀŃ è§£ +æŁ ¿ +é¢ Ĭ +ç° § +çĥŁ èĬ± +ä¾ ¥ +çĿ ¦ +éħ Ŀ +æ° ĵ +çIJ ī +å§ Ĭ +æ² ® +æħ · +èľ ķ +çij ļ +éĩĩ çŁ¿ +åł ° +åºķ èķ´ +èĨ ³ +è¾ ķ +éŁ Ń +åĴ Ļ +ç² ½ +åī Ķ +æ² ¦ +èĤ ´ +éķ ¶ +æĺ ¼ +è¾ Ĺ +å© ª +åĮ ® +æĸ ĵ +æ± ¶ +éĥ ´ +éł » +çª Ĵ +è¢ ± +åĽ ± +èĢ ĺ +è ļĮ +çĭ Ļ +çĹ ¹ +ç¥ ī +æı ® +æ· Ĩ +ç£ ĭ +éĺ ª +æ « +ã ¸ +Ļ ¶ +ã ij +𣠲 +ä ¢ +ã Ń +𬠨 +ð¬ Ģ +𬠮 +𬠯 +ð¬ ľ +𪠨 +ð« Ĺ +ð¬ Ĭ +𬠱 +ð¬ Ł +ä İ +ð ¡ +ä ĥ +ã ł +ð © +ð© ¾ +𬠺 +ð¬ Ļ +ãĢ Ķ +ãĢ ķ +çļĦ æĹ¶åĢĻ +æľīéĻIJ åħ¬åı¸ +ä¹ĭ åIJİ +ä¸ļ åĬ¡ +åķ Ĭ +èϽ çĦ¶ +æĭ¥ æľī +äºĴ èģĶç½ij +éĤ£ äºĽ +ä½ł çļĦ +åĨ³ å®ļ +éϤ äºĨ +åĽ¢ éĺŁ +åı¯ æĺ¯ +以 åIJİ +社 åĮº +çļĦ éĹ®é¢ĺ +å¹¶ ä¸Ķ +æķĻ å¸Ī +å°± ä¼ļ +天空 éĥ¨èIJ½ +æľĢ ç»Ī +å½ĵ çĦ¶ +ä¹Ł æľī +ç¡® ä¿Ŀ +æĥ³ è¦ģ +è´Ń ä¹° +人 çļĦ +åIJ ´ +çļĦ åıijå±ķ +ä¸į çŁ¥éģĵ +软 ä»¶ +æĪij们 çļĦ +çζ æ¯į +åī ij +èĢĮ æĺ¯ +å®ī æİĴ +åIJİ æĿ¥ +çļĦ åľ°æĸ¹ +èµ µ +èĢĥ è¯ķ +çªģ çĦ¶ +ä¸Ģå®ļ è¦ģ +åζ ä½ľ +è¯Ħ ä»· +åħį è´¹ +è´¹ ç͍ +绣 ä¸Ģ +çĦ¶ èĢĮ +è¿Ļ 次 +éĿĴ å¹´ +人 ç±» +äº ¦ +让 人 +è´Łè´£ 人 +éĩĩ åıĸ +çļĦ äºĭæĥħ +ä¹Ł ä¼ļ +车 è¾Ĩ +æĽ´ æĺ¯ +强 åĮĸ +æĪij åĢij +以 åīį +ä¼ĺ åĮĸ +å§Ķåijĺ ä¼ļ +åĽ° éļ¾ +å¹´ 度 +ä½į äºİ +æĮĩ åĩº +åĨį æ¬¡ +åĬŀ çIJĨ +æ¯ı 个 +对 æĸ¹ +è¿Ľè¡Į äºĨ +æľĢ é«ĺ +课 ç¨ĭ +身 ä¸Ĭ +æĽ¾ ç»ı +åĮ» çĶŁ +å®ī è£ħ +æľ ± +è¿IJ è¡Į +åıĮ æĸ¹ +æľĢ 大çļĦ +æŀĦ 建 +è¿ŀ ç»Ń +çļĦ å°ı +她 çļĦ +çŃī çŃī +æĶ¹ åĸĦ +åIJĦ ç±» +éģĩ åΰ +æľī çĿĢ +人 çī© +æĢ» æĺ¯ +è¿ħ éĢŁ +åζ å®ļ +å®ĥ 们 +å®ĺ ç½ij +è¿ĺ è¦ģ +ç»Ī äºİ +æĪ¿ åľ°äº§ +è¯ģ æĺİ +èĤ¡ 票 +åºĶ å½ĵ +èĭ± åĽ½ +è¿IJ ç͍ +æľĢ æĸ° +享 åıĹ +让 æĪij +æĻļ ä¸Ĭ +å¾ ŀ +å°ı 说 +å°¤åħ¶ æĺ¯ +è®Ń ç»ĥ +åħ¨ å¸Ĥ +æĮij æĪĺ +æľī çĤ¹ +带 çĿĢ +çļĦ ä¸ľè¥¿ +é£İ æł¼ +é»Ħ éĩij +å¼ķ 导 +æŃ¤ å¤ĸ +æľĢ è¿ij +追 æ±Ĥ +强 è°ĥ +ä¹Ł åı¯ä»¥ +æĦŁ åΰ +èĩª æĪij +çī¹åĪ« æĺ¯ +æĪIJ éĥ½ +éĢIJ æ¸IJ +å¿« ä¹IJ +ä¹ĭ ä¸Ń +æĬķèµĦ èĢħ +ä»ĸ们 çļĦ +æ° ı +å·¥ä½ľ 人åijĺ +äºĨ ä¸Ģ个 +åķ ¦ +ä¸Ģ åĢĭ +åŁº å±Ĥ +æ²Ł éĢļ +第ä¸Ģ 次 +å¹¶ 没æľī +çļĦ å·¥ä½ľ +åľ¨ è¿ĻéĩĮ +æŀ ª +æĶ¯ æĴij +æĹ¶ å°ļ +æĿ¥ åΰ +æĶ¶ è´Ń +éĿ© åij½ +æĺ¯ ä¸įæĺ¯ +讨 论 +ä¸ļ 绩 +å°± èĥ½ +ç«ĭ åį³ +è¡Ĺ éģĵ +åľ¨ ä¸Ģèµ· +æľĪ 份 +é«ĺ 端 +å¾Ī éļ¾ +ä¿Ħ ç½Ĺæĸ¯ +æīĭ 段 +åģļ åĩº +ä¼Ĺ å¤ļ +å®ŀ è¡Į +æīĵ å¼Ģ +游 客 +ä¾Ŀ çĦ¶ +å°± åĥı +离 å¼Ģ +说 éģĵ +æĸ° èĥ½æºIJ +æº ª +äº ķ +令 人 +ä¸Ģ åľº +æĪij æĥ³ +两 人 +èĩ³ å°ij +çļĦ çĶŁæ´» +æĺ¯ 个 +èĭ± è¯Ń +æ²Ĵ æľī +æĢĿ èĢĥ +éĻIJ åζ +åı° æ¹¾ +ä¸Ģ æĹ¦ +çļĦ ä¸Ģ个 +é«ĺ 级 +åĬŀåħ¬ 室 +å¾· åĽ½ +æĪij å°± +å®ļ ä½į +éĢĤ åºĶ +æĮĩ æłĩ +åħ¨ çľģ +ä¸Ĭ è¿° +å®ĥ çļĦ +åĽŀ å®¶ +欧 æ´² +éĵģ è·¯ +é¼ĵ åĬ± +çļĦ å½±åĵį +é«ĺ æł¡ +天 ä¸ĭ +é«ĺ è´¨éĩı +æĿŃ å·ŀ +èµĦ 讯 +æĶ¾ åľ¨ +æľī ä¸Ģ个 +å°± è¦ģ +ä¸Ĭ éĿ¢ +è§£ éĩĬ +éĢIJ æŃ¥ +å°½ 管 +æľī ä»Ģä¹Ī +çļĦ äºĭ +çĻ» è®° +人æ°ij å¸ģ +è§Ĥ ä¼Ĺ +è§Ĥ å¯Ł +ç͵ èĦij +çļĦ åIJĮæĹ¶ +ä½ľ ä¸ļ +宣 å¸ĥ +çļĦ ä½ľç͍ +åĽŀ æĿ¥ +éļ¾ ä»¥ +æīĢæľī çļĦ +å°ı åѦ +æıIJ åīį +æ¤į çī© +åĩ ¯ +ä¸Ĭ äºĨ +å°± åľ¨ +åħĪ åIJİ +æīĭ æľ¯ +éĥ Ń +éĿ¢ åīį +æ¯ķ 竣 +äºĮ æĺ¯ +红 èī² +éĺ³ åħī +èĭ¹ æŀľ +å¾Īå¤ļ 人 +ç»Ļ æĪij +åĵ ¦ +çľ¼ çĿĽ +éł Ń +ä¸Ģ æĺ¯ +åıijå±ķ çļĦ +åıį åºĶ +æĪ¿ å±ĭ +æľŁ å¾ħ +ç§į æ¤į +æĸĩ åѦ +åį³ åı¯ +é¦ĸ 次 +èĭ± éĽĦ +å¤ļ 次 +åĮħ è£ħ +æ²³ åįĹ +ä¹ĭéĹ´ çļĦ +ä»į çĦ¶ +åIJ¬ åΰ +èij£äºĭ éķ¿ +è§Ħ åĪĻ +ä¸Ģ 份 +大 ä¼Ĺ +使 å¾Ĺ +è¿Ľ åı£ +ä¸Ģ çīĩ +æĢ§ çļĦ +çļĦ 大 +æĪij æĺ¯ +äºĴ åĬ¨ +æ° £ +çļ Ĩ +åħ¬åı¸ çļĦ +ä¸Ģ è¾¹ +åıĬ åħ¶ +èī¯ å¥½çļĦ +æĭĵ å±ķ +å½ĵ å¹´ +广 åľº +åģļ äºĨ +åŁº äºİ +æıIJ éĨĴ +åħĦ å¼Ł +èĢģ æĿ¿ +è¿ij æĹ¥ +çĬ¶ åĨµ +注 éĩį +åĪļ åĪļ +è°ĥ çłĶ +å¿ĥ ä¸Ń +æĬĬ æı¡ +éļı åIJİ +ä¸į å¤Ł +åĪĽ ä½ľ +ç«Ļ åľ¨ +缸 äºĴ +çĸ«æĥħ éĺ²æİ§ +å¹´ 代 +带 åĬ¨ +伤 害 +竣 çĦ¶ +å¼ķ è¿Ľ +ç´¯ 计 +让 æĪij们 +åĽŀ æĶ¶ +æĬ¥ åIJį +åĬ© åĬĽ +èģĶ çĽŁ +çŃĸ çķ¥ +åij¨ è¾¹ +åĭ Ĵ +è¿ĺ åľ¨ +æµģ éĩı +寻 æī¾ +ç͵ åĬĽ +èι èζ +è¿ĺ èĥ½ +æĭħ ä»» +çļĦæĥħåĨµ ä¸ĭ +çļĦ åİŁåĽł +缺 ä¹ı +çIJĥ åijĺ +å²ģ çļĦ +çĶ· åŃIJ +å·¥ èµĦ +è¿ijå¹´ æĿ¥ +åij Ģ +æıIJä¾Ľ äºĨ +她 们 +å®¶ åħ· +çĩ ķ +è½» æĿ¾ +æł¡ åĽŃ +èĢĥ æł¸ +åį± éĻ© +åħļ ç»Ħç»ĩ +æĢ» ç»ıçIJĨ +çļĦ æĸ° +çİ» çĴĥ +è¿Ļ ä½į +对 æŃ¤ +å®¶ 人 +çļĦ è¦ģæ±Ĥ +温 度 +æĮĩ æķ° +缴 åΰ +æŃ¤ æĹ¶ +æ¹ĸ åįĹ +éĥ½ è¦ģ +ä½ľ åĩº +åIJĦ ä½į +èĢĥ çĶŁ +ä¾Ŀ æį® +说 è¯Ŀ +æĪij ä¹Ł +å·¥ åİĤ +åıĺ æĪIJ +ä»ĸ 人 +æĪij è§īå¾Ĺ +åIJĦ 级 +ä¼łå¥ĩ ç§ģæľį +ä¸Ĭ åįĩ +好 åĥı +åĬł éĢŁ +äºĮ åįģ +è¢ ģ +è£ħ 饰 +éĥ½ èĥ½ +ä¸Ģ å¼ł +åĬ¨ æĢģ +å¹´ çļĦ +è¿Ļ å°±æĺ¯ +ä¹Ł è¦ģ +èµĦ æł¼ +æĪĺ äºī +æĦŁ è°¢ +åŁ¹ èĤ² +天 æ°Ķ +女 士 +åı¯èĥ½ ä¼ļ +çļĦ 产åĵģ +ä¹Ł å°± +主è¦ģ æĺ¯ +åĪº æ¿Ģ +ç»Ļ ä½ł +大 æķ°æį® +åĮ» åѦ +åĪ ¤æĸŃ +ä»ĸ 说 +表 æ¼Ķ +äºļ æ´² +ä¸ĵ é¢ĺ +ç«ŀäºī åĬĽ +éĤ£ æł· +å±ķ å¼Ģ +å¹³ æĹ¶ +æİ¥ ä¸ĭæĿ¥ +æī¿ 诺 +æ³ķ åĽ½ +åħ³ å¿ĥ +ä¼ļ æľī +éĤĢ è¯· +é¢Ħ éĺ² +对 æİ¥ +好 äºĨ +åĴ± 们 +çļĦ æĦŁè§ī +æĢĿ è·¯ +éĥ½ 没æľī +çļĦ æĸ¹æ³ķ +女 åŃIJ +åı¸ æ³ķ +è¿ĺ ä¼ļ +è¶ĬæĿ¥è¶Ĭ å¤ļ +åĽł çĤº +æµ· åįĹ +人 æķ° +å°Ĩ ä¼ļ +ä¸ļ 主 +é¤IJ 饮 +å±ħ ä½ı +åıij åĩº +è¿ij æľŁ +å¼ķ é¢Ĩ +æľºåύ 人 +åĩºæĿ¥ çļĦ +çľĭ è§ģ +ä¿ Ĭ +让 ä»ĸ +ä¸į æĥ³ +å·¥ä½ľ çļĦ +è¡¥ åħħ +æµ ħ +çī¹ å¾ģ +ä¸Ĭå¸Ĥ åħ¬åı¸ +ç¾İ é£Ł +广 西 +æ¯ı ä¸Ģ个 +èIJ½ åľ° +åĵģ ç§į +åĴĮ è°IJ +å½» åºķ +é«ĺ èĢĥ +æĺ¨ 天 +åīį å¾Ģ +çĽij æµĭ +çϾ 度 +åľ¨ ä¸ŃåĽ½ +çļĦ éľĢæ±Ĥ +亿 ç¾İåħĥ +åѦ æľ¯ +æĶ¶ åΰ +æĿ¿ åĿĹ +ä¸Ģ 段 +æŀĦ æĪIJ +ä¼ģä¸ļ çļĦ +表 éĿ¢ +æķ´ çIJĨ +ç»ĵ å©ļ +人 å®¶ +åģľ æŃ¢ +åѦ ç§ij +æĺ¾ å¾Ĺ +ä¼ij æģ¯ +é¢Ħ æľŁ +æĪĸ æĺ¯ +çļĦ 主è¦ģ +åºĶ 对 +èµ° äºĨ +ä¸Ń éĹ´ +èµ° è¿Ľ +åijĪ çݰ +æIJŃ éħį +é¹ ı +æĺ¯ åĽłä¸º +æĥħ 绪 +å®ļ æľŁ +社ä¼ļ 主ä¹ī +çŃī 级 +磼 çĽ¾ +é£ŀ æľº +èĩ³ ä»Ĭ +æĶ¶ éĽĨ +çļĦ æķħäºĭ +åĪĩ å®ŀ +å®ŀçݰ äºĨ +å½¢ æĪIJäºĨ +åįĹ æĸ¹ +ä¸Ń åѦ +æµ· æ´ĭ +åIJ¦ åĪĻ +æĭį æijĦ +大åѦ çĶŁ +åĩºçݰ äºĨ +æĦı å¤ĸ +ä¹Ł èĥ½ +çļĦ èĥ½åĬĽ +åĿIJ åľ¨ +åĪĻ æĺ¯ +èĢĥ å¯Ł +å°Ĭ éĩį +éĺ² æŃ¢ +ç´§ å¼ł +读 书 +åĩº è¡Į +å°± æľī +å±¥ è¡Į +çݰ代 åĮĸ +åĽ½ åĬ¡ +åĽ½åĬ¡ éĻ¢ +ç»´ ä¿® +åİŁ åĪĽ +æĺ¯ æĮĩ +ä¼ij éĹ² +çĤ ® +æĸ° æĹ¶ä»£ +éĢĻ åĢĭ +ä¸į æķ¢ +å®Į ç¾İ +ç»Ĩ èĬĤ +éŃ ı +èͬ èıľ +é¢Ĩ导 çıŃåŃIJ +è¶ħ 级 +è¡Į æĥħ +人工 æĻºèĥ½ +åį° åº¦ +åŁºç¡Ģ 设æĸ½ +åıĪ æĺ¯ +èᝠçī© +åIJ¸ æĶ¶ +åį´ æĺ¯ +éĥ İ +å¥ĸ åĬ± +çļĦ æľĭåıĭ +ä¿Ŀ çķĻ +è§Ħ å¾ĭ +æĸ° çĸĨ +è¿ĺ åı¯ä»¥ +æİ¥ è¿ij +æŃ¤ åīį +æī¹ åĩĨ +æĢİä¹Ī æł· +çļĦ ä½įç½® +ä¸Ģ åĿĹ +æĭĴ ç»Ŀ +顾 客 +ä¹Ł åľ¨ +ä¸Ģ çĶŁ +éĥ¨ éĺŁ +å¹´ åīį +æĸ¹éĿ¢ çļĦ +å°Ŀ è¯ķ +羣æŃ£ çļĦ +ç¦ģ æŃ¢ +è¿ĺ 没æľī +æ°ij çĶŁ +èµ° åIJij +èĦ¸ ä¸Ĭ +å½ĵ 天 +éĽĨåĽ¢ åħ¬åı¸ +çļĦä¸Ģ ç§į +西 æĸ¹ +åĽŀ åºĶ +ä¸Ģ 声 +常 常 +æıIJ åΰ +èħ¾ 讯 +æľį è£ħ +为 ä½ķ +äºij åįĹ +å°± ç®Ĺ +ä¼ł æī¿ +åıį èĢĮ +ä¸ĩ åIJ¨ +è´¢ 产 +å¦Ĥ ä¸ĭ +æĹ¥ åīį +åİŁ æľ¬ +æľĢ éĩįè¦ģçļĦ +认 è¯ģ +ä¸Ģ éģĵ +ä¿¡æģ¯ åĮĸ +å¾Ĺ åΰäºĨ +é̲ è¡Į +æĪij è¦ģ +éĢļ ä¿¡ +室 åĨħ +èµļ éĴ± +æĶ¶ èĹı +è§£åĨ³ æĸ¹æ¡Ī +æĪ¿ 产 +çĭ ¼ +æ´» åĬĽ +ç»ıæµİ åıijå±ķ +çŃī å¾ħ +ä¹Ł å¾Ī +åĿ ij +å¾Ī 好çļĦ +éļ¾ åº¦ +ä¸į å¦Ĥ +人æ°ij æĶ¿åºľ +åĩº åıij +åīį æľŁ +æ¼Ķ åijĺ +女 çĶŁ +èģļ çĦ¦ +审 计 +é¢Ħ æµĭ +ä¾Ŀ æīĺ +äºĶ å¹´ +è¡¥ è´´ +æ¸ħ æĻ° +éª Ĥ +çľĭ èµ·æĿ¥ +çļĦ åŃ©åŃIJ +é¢ij éģĵ +ä½ı å®ħ +éĿ¢ åIJij +æľĢ ä½İ +æĹ¢ çĦ¶ +ä¸Ģ å¥Ĺ +æķ° åѦ +群 ä½ĵ +åĮĹ京 å¸Ĥ +å±ħ çĦ¶ +æ°Ľ åĽ´ +éĢĶ å¾Ħ +çļĦ åŁºç¡Ģä¸Ĭ +èģĮ è´£ +åı¯èĥ½ æĺ¯ +åĨĽ äºĭ +æĪIJ æķĪ +åŃ©åŃIJ 们 +计ç®Ĺ æľº +èµ ¤ +产ä¸ļ åıijå±ķ +å·¨ 大çļĦ +å·¥ 人 +çĶŁ éķ¿ +éĥ½ åı¯ä»¥ +çļĦ æľºä¼ļ +èµĦ è´¨ +çĹĽ èĭ¦ +ç²ī ä¸Ŀ +å¢ ĵ +å¹³ å®ī +管 éģĵ +è·Ł çĿĢ +饮 é£Ł +åķĨ å®¶ +å¤ļ å®¶ +åı¸ æľº +åºĶ该 æĺ¯ +éĢı éľ² +认 å®ļ +è¡Įä¸ļ çļĦ +çļĦ ä¼ģä¸ļ +æ¯ı ä¸Ģ +èĮĥåĽ´ åĨħ +è¾ĥ 大 +è´ ¤ +大 èµĽ +å¤ļ äºĨ +é¸ ¿ +临 åºĬ +åľ¨ è¿Ļ个 +çļĦ åĨħ容 +éĶĢ éĩı +å¾Ī å°ij +åŃ Ł +ç»´ æĮģ +åĴĸ åķ¡ +æľ¬ åľ° +èī² å½© +å¹¶ éĿŀ +èĢĮ å·² +温 æļĸ +èIJ § +æĬĵ ä½ı +èĢĮ ä¸įæĺ¯ +åĸ Ĭ +çļĦ åħ³ç³» +çī© åĵģ +éĤ£ æĺ¯ +åĨľ 产åĵģ +è¿Ļ æĹ¶ +å©ļ å§» +æ°´ æŀľ +æĶ¶ èİ· +ä»ĺ åĩº +客æĪ· 端 +æ¼Ķ åĩº +åħ¨ æĸ° +è¿Ļ ä¹Łæĺ¯ +æĺ¯ çͱ +è§Ĥ 念 +æľī 个 +éĢł åŀĭ +èĥľ åĪ© +ä¸ī æĺ¯ +è¶ħ å¸Ĥ +åħļ建 å·¥ä½ľ +æĶ¾ å¿ĥ +线 è·¯ +æĭĽ çĶŁ +åIJĥ é¥Ń +è½ ī +å°½ éĩı +è§ģ åΰ +åIJĮæ¯Ķ å¢ŀéķ¿ +åįİ ä¸º +æĪij å¸Ĥ +æıIJ åĩºäºĨ +æ°ij èѦ +åįļ çī© +åįļçī© é¦Ĩ +è¯ļ ä¿¡ +åīį éĿ¢ +å±± 西 +è¾ħ åĬ© +转 ç§» +æĽ´ 为 +丰å¯Į çļĦ +åį ¢ +å¿« éĢĴ +æĺ¾ èijĹ +çī© èµĦ +åΰ è¾¾ +æľī åĪ©äºİ +åij Ĩ +åŃ©åŃIJ çļĦ +ä¸į ä½Ĩ +çłĶç©¶ éĻ¢ +çͳ æĬ¥ +æļ ¨ +æ°ij éĹ´ +åį » +çļĦ å£°éŁ³ +å¸Ĥåľº çļĦ +ä¸Ģ åı¥ +çľģ 级 +æĿ¥ çļĦ +åĵª 个 +æīį ä¼ļ +åĪĨ éħį +èĶ ¡ +ä»ĸ åľ¨ +åħ± æľī +å¡ ĺ +èĴ Ĥ +éľ į +åıĤ è§Ĥ +ä¸Ī 夫 +ä¾Ŀ éĿł +æľī æĹ¶ +äºĨ å¾Īå¤ļ +ä¸ĸçķĮ æĿ¯ +å®¶ æĹı +ä¸į éľĢè¦ģ +大 å¸Ī +èŀį åħ¥ +éĿŀ æ³ķ +çĹħ 人 +åIJİ æľŁ +大家 éĥ½ +ç½ij åĿĢ +åİŁ æĸĻ +便 å®ľ +æ¶ Ľ +仿 ä½Ľ +å·® è·Ŀ +åı¦ä¸Ģ æĸ¹éĿ¢ +产åĵģ çļĦ +èµ « +æĥħåĨµ ä¸ĭ +éĴ¢ éĵģ +æľ¬ ç«Ļ +纳 åħ¥ +å·² æľī +æľī 没æľī +ä¼° 计 +é£ ĺ +æľŁ è´§ +åĢĭ人 è³ĩæĸĻ +ä¸ĵä¸ļ çļĦ +çĪĨ åıij +èĩ´åĬĽ äºİ +çİ°åľ¨ çļĦ +æľī åĵªäºĽ +çł´ åĿı +æķ°åŃĹ åĮĸ +åľ° éĿ¢ +é»ij èī² +å¹¼åĦ¿ åĽŃ +çļĦ ç²¾ç¥ŀ +äº Ń +导 æ¼Ķ +çݰ æľī +æŃ¦ åύ +èĭı å·ŀ +çİ Ħ +æ±Ł 西 +å»¶ 伸 +论 æĸĩ +è¾ĥ 为 +çİ© æ³ķ +é¼ İ +åIJĮ æŃ¥ +éĩĬ æĶ¾ +æĽĿ åħī +åĿļ åĨ³ +å§Ķ æīĺ +å°Ĩ åľ¨ +äºĪ 以 +ä½ľ æĸĩ +èĢĮ åľ¨ +ä¼ĺ åħĪ +åĽŀ åİ» +ä¿® å¤į +åĽ½åĨħ å¤ĸ +çŃĸ åĪĴ +åıij æĶ¾ +å¿ĥ æĥħ +çļĦ åİĨåı² +éĿ¢ è¯ķ +举 åĮĹ +ä¿¡ åı· +ç²® é£Ł +è¯ģ 书 +æŁIJ äºĽ +è¿IJ ä½ľ +åĨ² åĩ» +çĥŃ çĤ¹ +æĹ¶ æĹ¶ +æĹ¶æĹ¶ 彩 +åľ° çĤ¹ +ä¸Ģä½ĵ åĮĸ +éļ¾ é¢ĺ +æĽ ° +ç«ĭ åĪ» +æĺ¯ éĿŀ常 +åħ± åĴĮ +åħ±åĴĮ åĽ½ +æ¿Ģ åĬ± +æľīæķĪ çļĦ +å¤Ħ ç½® +该 åħ¬åı¸ +æ£Ģ éªĮ +èѦ æĸ¹ +è´ ¾ +äºĨä¸Ģ ä¸ĭ +ä»Ĭ åIJİ +çħ ® +ç͍ åĵģ +读 èĢħ +æĪij åľ¨ +åĽŀ å¤į +ä¸Ģ 座 +è¿ĺ 没 +å®ļ åζ +没 æĥ³åΰ +å¤ ¹ +ä¼ł éĢĴ +ä¸Ģ 款 +强 大çļĦ +çļĦ è¡Į为 +å¤ı 天 +åıijåĬ¨ æľº +é¢ĨåŁŁ çļĦ +å®ŀéªĮ 室 +ä¸Ģ æĬĬ +æĺ¯ 为äºĨ +éĻķ 西 +æĭħ ä¿Ŀ +è¾¾ æĪIJ +è¦ģ æĺ¯ +æĺİ å¤© +ç»Ļ ä»ĸ +建ç«ĭ äºĨ +ä¸į è¡Į +ä¸Ń æĸĩ +åľ° 说 +åIJİ çļĦ +çĽij æİ§ +éĢ ¸ +æĢ» éĥ¨ +æľ¬ æĸĩ +é¹ ¿ +æĻ¯ è§Ĥ +çļĦ 缮æłĩ +èĽ ĩ +åĨ ¯ +ä¸Ń åĮ» +æķĪ åºĶ +产 éĩı +åŃ Ŀ +è´¦ æĪ· +è¿Ŀ åıį +èij£äºĭ ä¼ļ +京 举 +责任 ç¼ĸè¾ij +åķı é¡Į +çα å¿ĥ +èѦ å¯Ł +é¤IJ åİħ +å¸Ĥ æĶ¿åºľ +天 天 +æĸ° é²ľ +éĥij å·ŀ +è¶ħ è¶Ĭ +å½ Ń +çŁ¥è¯Ĩ 产æĿĥ +åĽŀ å¿Ĩ +è·¯ 线 +å»ī æ´ģ +éĿĴ å°ijå¹´ +åıĸå¾Ĺ äºĨ +çľĭ åΰäºĨ +é¦ ¬ +ç²¾ åĵģ +åľ° éĵģ +æĮģ æľī +ä¸ĭ äºĨ +æľī æĹ¶åĢĻ +ä¸Ģ 人 +æĴ Ĵ +ä»Ķ ç»Ĩ +èĢģ åħ¬ +äºĭå®ŀ ä¸Ĭ +èģĶ èµĽ +ä¾ĽåºĶ éĵ¾ +é¢Ħ ç®Ĺ +åζéĢł ä¸ļ +å®īåħ¨ çĶŁäº§ +俱 ä¹IJ +俱ä¹IJ éĥ¨ +çļĦ æł¸å¿ĥ +æīĵ ç®Ĺ +å½± çīĩ +æIJŃ å»º +ä¹Ł ä¸įä¼ļ +æĭħ å½ĵ +å±Ĥ éĿ¢ +åѦ åijĺ +临 æĹ¶ +缸 ç»ĵåIJĪ +对 æ¯Ķ +ä»ĸ æĺ¯ +æĸ° åĮº +è¿Ľ åİ» +çϾ å¹´ +ä¿ © +å°½ å¿« +ç͵åŃIJ åķĨåĬ¡ +æĽ´ æľī +æ¸ħ çIJĨ +åı¦ ä¸Ģ个 +åĤ » +ä»Ģä¹Ī æł·çļĦ +æĺ¯ æľĢ +åij¨ å¹´ +å¾Ī 容æĺĵ +åĽ¢ ç»ĵ +ç´ Ħ +æĹ© å·² +çļĦ åıĺåĮĸ +éľ ŀ +æĹ¥ ä¸ĬåįĪ +失 åİ» +ä¸Ń åľĭ +çļĦä¸Ģ äºĽ +å°ı åŃ© +ä¸ĭ è·Į +éĶ» çĤ¼ +é ij +éij « +å¿ĹæĦ¿ èĢħ +èĤ¡ å¸Ĥ +èµĽ äºĭ +许åı¯ è¯ģ +åı¯ æĮģç»Ń +åijĬè¯ī è®°èĢħ +éĢ» è¾ij +å¼ķ åħ¥ +çļĦ è¿ĩç¨ĭä¸Ń +è§Ĩ è§ī +èĩªæ²» åĮº +è¯ģ æį® +è£ħ ç½® +第ä¸ī æĸ¹ +å¹´ æĿ¥ +å¹¿ä¸ľ çľģ +带æĿ¥ äºĨ +éķ¿ æ±Ł +访 éĹ® +å·® ä¸įå¤ļ +æĺ¯ æĪij +éģŃ éģĩ +æĬĵ 好 +é«ĺ è¾¾ +å¹¶ åľ¨ +èĩª è§ī +ä¾ĽåºĶ åķĨ +æĥħ æĦŁ +ä½ı äºĨ +çļĦ èģĮä¸ļ +çļĩ å¸Ŀ +西 éĥ¨ +åĴĮ å¹³ +çļĦ åĬĽéĩı +æ± ª +åħħåĪĨ åıijæĮ¥ +æĬķ è¯ī +èµ· åΰ +äºĴ 缸 +æ¾³ éŨ +æİ¥ åΰ +æ°´ æ³¥ +模 åŀĭ +ä¸Ģ åįĬ +ç§© åºı +æĪij们 åľ¨ +æī¿ 认 +ä¸Ģ éĥ¨åĪĨ +åįł æ¯Ķ +å¦ĩ 女 +ç² ĺ +äºĨè§£ åΰ +ä¸Ģå®ļ ä¼ļ +åIJĦ 大 +èµ° åĩº +为 大家 +é«ĺ éĵģ +åı¯ä»¥ åľ¨ +ä½Ĩ åľ¨ +çĶŁæĢģ çݯå¢ĥ +èı ¯ +çļĦ ä»·æł¼ +麻 çĥ¦ +æ¿Ģ åıij +éĤ£ å°± +çļĦ æł·åŃIJ +为 æŃ¤ +天 åľ° +çļĦ 缮çļĦ +åĢº åΏ +å·² ç¶ĵ +åĽĽ 大 +åIJĮæĹ¶ ä¹Ł +å½¼ æŃ¤ +æĭ¿ åΰ +åIJ« éĩı +åįģ 大 +éļ¾ éģĵ +å¼ Ĺ +ä¸Ģ 段æĹ¶éĹ´ +çħ§ 顾 +æķ°æį® æĺ¾ç¤º +æĪIJ为 äºĨ +èµ° åΰ +æľ¬ åħ¬åı¸ +ç»Ī 端 +ä¹Ł ä¸įæĺ¯ +头 åıij +大 约 +é£İ æĻ¯ +æ¶Ī èĢĹ +审 æŁ¥ +äºī åıĸ +æ³ķ æ²» +äºĭ çī© +ç¼ĵ è§£ +æĥ ¨ +缸åºĶ çļĦ +çļĦ æķĪæŀľ +åıį å¤į +åıijçĶŁ äºĨ +éĢĻ äºĽ +ç»ĥ ä¹ł +åݨ æĪ¿ +å¼Ģ æĭĵ +欣 èµı +夫 妻 +ä¸į ä¸Ģæł· +产 èĥ½ +èĬ¯ çīĩ +è¦ģ ç´ł +åıį 对 +çİĩ åħĪ +è´§ çī© +æĹ¥ ç͵ +ä½ľ å®¶ +æĶ¹ è¿Ľ +æĪIJ åĪĨ +åĽł èĢĮ +åĩı èĤ¥ +æ½ ĺ +å±±ä¸ľ çľģ +åĬ Ŀ +åŁ ĭ +æŃ¦ è£ħ +æ±ĩ æĬ¥ +ä¸Ģ个 æľĪ +çĥŃ éŨ +大 éģĵ +æ´» åĭķ +éĥ½ å¾Ī +ç͵ 梯 +ç´§ æĢ¥ +åĢº åĬ¡ +客 æľį +ä¸Ģ éĥ¨ +ä½ł æĺ¯ +çݰ çĬ¶ +æŃ£ç¡® çļĦ +ä¹ĭ å¤Ħ +ç¼ĸ åζ +ä½ł åı¯ä»¥ +çŃī åľ° +èİ ī +对 è¯Ŀ +æ·ĺ å®Ŀ +è°ĥ èĬĤ +æİĴ æĶ¾ +åºĵ åŃĺ +ç´ ļ +çļĦ ä¼ĺåĬ¿ +æĿĥ å¨ģ +以ä¸ĭ ç®Ģç§° +ä¸Ģ 项 +èģļ éĽĨ +ä¼łç»Ł çļĦ +æ·· åIJĪ +è¿Ļä¸Ģ çĤ¹ +ä¸Ģ çľ¼ +æĹł éĻIJ +èİ·å¾Ĺ äºĨ +éĢī æīĭ +åζ åĵģ +åįı ä½ľ +çĭ¬çī¹ çļĦ +ä¸Ģ 级 +è¿Ļ个 éĹ®é¢ĺ +æĸ Į +æĺ¯ æĪij们 +æķĮ 人 +æ¸ħ æ´Ĺ +ä¸Ģ缴 åľ¨ +å°ı ç±³ +çļĦ è¿ĩç¨ĭ +åľ¨ åĮĹ京 +ä¸Ģ æĶ¯ +æĹ© ä¸Ĭ +æĸĩ èīº +ç¦ı åĪ© +é£Ł ç͍ +æĦŁ åĬ¨ +åħ¨ ç¨ĭ +æĶ¯ åĩº +æĸ° 建 +å¸ ķ +æĺ¾ çĦ¶ +羣 çļĦæĺ¯ +æĸ°éĹ» ç½ij +èĥ½ åIJ¦ +åįı åĬ© +亲 èĩª +å¾Ī æľī +çϼ å±ķ +æĦı 大 +æĦı大 åĪ© +ç͵ ç½ij +æĹ¥ çĽĬ +çĨ ± +èĤĮ èĤ¤ +çĶ· æĢ§ +ç»Ħ 建 +çŃī éĹ®é¢ĺ +æ¶Ī éϤ +æĬ¤ çIJĨ +å¡ij æĸĻ +ä¹Į åħĭ +ä¹Įåħĭ åħ° +åķĨ æłĩ +çIJ ³ +æĸ° æīĭ +çļĦ çī¹çĤ¹ +åĴ ¬ +å½ĵ ä¸ĭ +设计 å¸Ī +èµĶ åģ¿ +第 åįģ +æĻºèĥ½ åĮĸ +å¼Ģåıij åĮº +åı¯ä»¥ éĢļè¿ĩ +åħ±äº§ åħļ +åİī 害 +çģµ æ´» +æĹ¶ åħī +éĥ¨ ä½į +人 æĸĩ +è¿Ľ æĿ¥ +ä¹ĭ æīĢ以 +ä¸ī åįģ +çļĦ åѦçĶŁ +éĺ² æĬ¤ +åĽ½ 产 +æ·±åľ³ å¸Ĥ +éĤ£ å°±æĺ¯ +åΰ ä½į +çī¹ æľĹ +çľĹ æĻ® +å®ŀ æĹ¶ +åı° çģ£ +èĢĮ ä¸į +æĮĩ å®ļ +åĿ Ŀ +èħIJ è´¥ +çī¹ å®ļ +å¢ŀ éĢŁ +æłĩ çѾ +æĪ¿ ä»· +æĦ ģ +贯彻 èIJ½å®ŀ +æĢ§ è´¨ +çłĶç©¶ çĶŁ +ç¾İ 容 +æī¹ è¯Ħ +ç©¶ 竣 +人åĬĽ èµĦæºIJ +éĸĭ å§ĭ +åĽŀ å½Ĵ +èIJ¥ åķĨ +èIJ¥åķĨ çݯå¢ĥ +ä¸ŃåĽ½ 人 +çļĦ åŁºæľ¬ +è¯Ŀ é¢ĺ +æłĩåĩĨ åĮĸ +西 èĹı +åĭ ¾ +çļĦ 设计 +ç®Ģåįķ çļĦ +å¤į åζ +æ¸IJ æ¸IJ +以 å¤ĸ +èģĶ åĬ¨ +两 次 +æĢ§ åĴĮ +æĽ´ 大 +çļĦ åIJįåŃĹ +éŁ ¦ +ä½ł è¦ģ +å¢ĥ å¤ĸ +æĹ© æľŁ +åĪĿ æŃ¥ +è´¦ åı· +害 æĢķ +æĺ¨ æĹ¥ +åĪļ æīį +ç¥ŀ ç§ĺ +ç²¾ å¿ĥ +æµģ éĢļ +åħ¨ æĸ¹ä½į +以 å¾Ģ +ä¹Ł å°Ĩ +æĺ¯ ä¸ŃåĽ½ +åĽ½å®¶ 级 +å°Ĩ åĨĽ +æij Ĭ +æľĢ 为 +第ä¸Ģ æĹ¶éĹ´ +æ¶Ī æ¯Ĵ +å°Ĩ äºİ +å¨ģ èĥģ +èĭ± æĸĩ +æīĭ ä¸Ń +çIJĥ è¿· +è§Ĥ çľĭ +离 å©ļ +æľ¬ åľŁ +åĪĨ æķ£ +æĻ ´ +è¦ģ 注æĦı +浪 è´¹ +管 æİ§ +åĩº åĶ® +æĢ» è£ģ +ä¸Ģ éĺµ +å¨ ĩ +äºĶ 个 +å½ĵ åĪĿ +çºł 纷 +ä¸ĵ ç͍ +å¤ĩ æ¡Ī +åĪĿ æľŁ +å®ĥ æĺ¯ +åĮº åĿĹ +åĮºåĿĹ éĵ¾ +大 è¿ŀ +è¿Ļ ç±» +åıĺ æĪIJäºĨ +éĤĦ æĺ¯ +åįļ 客 +çı¾ åľ¨ +ä¸Ģ æĸ¹ +å®ĮæĪIJ äºĨ +è¿Ļ个 æĹ¶åĢĻ +åħ¨ å¹´ +ä¸Ĭ 线 +ç½ IJ +ç«ŀ èµĽ +åĩºçīĪ ç¤¾ +åĵ¥ åĵ¥ +å¯ « +å¾Ĺ 以 +èĬ± åĽŃ +äºĨ èµ·æĿ¥ +èĦ±è´« æĶ»åĿļ +çļĦ åİŁåĪĻ +讲 è§£ +æ¶Ī åĮĸ +æį٠害 +æļĤ æĹ¶ +å¾Ĺ çŁ¥ +éĢĤ ç͍ +éŨ åºĹ +è§£ 读 +æĻ® åıĬ +人æ°ij æ³ķéĻ¢ +åī¯ ä¸»ä»» +å¿ĥ çģµ +è¯Ĭ æĸŃ +ç¾İ 女 +æŁ ¯ +å¹´ 以æĿ¥ +æ´» è·ĥ +åĢŁ åĬ© +åħ± 建 +è¯ī 讼 +æĶ¾ æĿ¾ +çªĹ åı£ +ä¼ģ æ¥Ń +åĬł æĭ¿ +åĬłæĭ¿ 大 +ä¹° äºĨ +主 æµģ +æĩĤ å¾Ĺ +å°Ĩ åħ¶ +éĢı æĺİ +å·¥ä½ľ ä¸Ń +èĤ¡ ä»· +æ¡£ æ¡Ī +没æľī ä»»ä½ķ +åijĬ çŁ¥ +å¹´ åĪĿ +æĹ¥ ä¸ĭåįĪ +åİĤ åķĨ +èĬĤ å¥ı +主 导 +è£ Ŀ +åħ³éĶ® è¯į +èģĬ 天 +åĨĻ ä½ľ +æĶ¹éĿ© å¼ĢæĶ¾ +æľī æľĽ +éĢļ æĬ¥ +èIJ Į +æĢ» é¢Ŀ +çŁŃ æľŁ +ä¸Ģ çķª +çĶŁæ´» çļĦ +åĮĸ çļĦ +æĺ¥ 天 +è¿Ļ åľº +æĸ°å¼Ģ ä¼łå¥ĩ +æĺ¯ è¦ģ +å°ļ æľª +åıĺ æĽ´ +ä¸Ģ åij¨ +客 è§Ĥ +æĹ¥ èĩ³ +é¹ ° +çİ ² +å°Ĩ æĿ¥ +客 人 +åıĺ éĿ© +说 äºĨ +åİŁ çIJĨ +èģĮ åĬ¡ +åıĪ æľī +ä¸Ģ åı¥è¯Ŀ +æĦŁ åıĹåΰ +ç¬Ķ èĢħ +ç§» æ°ij +西 åįĹ +ä¹ĥ èĩ³ +æŃ£ è§Ħ +åĪĿ ä¸Ń +çĬ ¬ +å½ĵ äºĭ +å½ĵäºĭ 人 +æĪij们 è¦ģ +åħ¥ åı£ +éĤ£ æĹ¶ +æľīéĻIJ 责任 +å°ij 女 +è¿Ļä¹Ī å¤ļ +åĪĨ åħ¬åı¸ +å®ĩ å®Ļ +çļĦ éĢīæĭ© +å§IJ å§IJ +åıij èµ· +è» į +æĽ´å¥½ åľ° +éĻĨ ç»Ń +æľ¬ æľįåĭĻ +å« © +èµ¶ ç´§ +èĦĤ èĤª +第äºĮ 天 +æĪij ä¼ļ +两 ä½į +æķ ² +åħ¬å®ī æľºåħ³ +ç§ijæĬĢ åĪĽæĸ° +å°º 寸 +è¾IJ å°Ħ +å®Ĺ æķĻ +转 æį¢ +åĩº çİ°åľ¨ +ä¸Ģ é¢Ĺ +æľŁ éĻIJ +åIJĮåѦ 们 +åĮĹ æĸ¹ +ä½ł å°± +ä¸Ģ带 ä¸Ģè·¯ +èĢģ å©Ĩ +游æĪı çݩ家 +çļĦ ç»ĵæŀľ +è¡¥ åģ¿ +å¤ĸ è´¸ +对 å¾ħ +ç»´ çĶŁç´ł +ç»ıéĶĢ åķĨ +è¿ĺ å°Ĩ +åŃIJ 女 +æĽ´ é«ĺ +ä¸į 大 +éī´ å®ļ +让 ä»ĸ们 +æīĢè°ĵ çļĦ +æŃ» äºĨ +帮 æī¶ +åĵ² åѦ +以ä¸Ĭ çļĦ +çļĦ åħ³éĶ® +æĹ© å°± +æĬ¥ ä»· +éģµ å®Ī +æī© å¼ł +æĺ¯ å¾Ī +å¼Ģ éĢļ +æĸ° åĬł +æĸ°åĬł åĿ¡ +ç¿» è¯ij +询 éĹ® +é¸ Ń +ä½ĵ åĨħ +两 个人 +çĪ ¹ +éľ ľ +乡æĿij æĮ¯åħ´ +çĿ¡ è§ī +å®ĺ åijĺ +åĪĽ å§ĭ +åĪĽå§ĭ 人 +ä¼Ĺ 人 +åį³ ä¾¿ +çĸ« èĭĹ +ä¼ģä¸ļ å®¶ +æ¸ £ +ç²¾ åĬĽ +å¤ĸ éĥ¨ +èģª æĺİ +è¿Ļ ä¹Ł +å½ķ åıĸ +åĨ² çªģ +åħ¨ 身 +åŃ£ èĬĤ +忽 çĦ¶ +çļĦ æĢģ度 +åĤ¨ å¤ĩ +ä¿Ŀ åħ» +çļĦ æĥ³æ³ķ +ä¸Ĭæµ· å¸Ĥ +æIJº æīĭ +çļĦ ä¿¡æģ¯ +åķĨ åľº +çļĦ æĢĿæĥ³ +æĿĥ åĬĽ +毫 æĹł +æĢĢ åŃķ +硬 ä»¶ +åĨħ èĴĻåı¤ +æİ¢ 讨 +åħ» çĶŁ +çļĦ 表çݰ +空 ä¸Ń +æģIJ æĢĸ +å¾Ī é«ĺ +ç»ıæµİ 社ä¼ļ +ä¸Ĭ æĿ¥ +å»¶ ç»Ń +éĩį å¤į +éĺ² èĮĥ +çļĦ å½¢å¼ı +æľĪ åºķ +èĢģ 年人 +绿 åĮĸ +å±± åĮº +æĭ¿ åĩº +æĹħ 客 +æĽ´ æį¢ +åħ¬ 主 +èĬĤ 约 +åħ¨ åİ¿ +åĽŀ æĬ¥ +çIJĨ æĢ§ +çĸ¯ çĭĤ +æ¶ī å«Į +åī§ æĥħ +åĨ¬ åŃ£ +åIJİ ç»Ń +è¿Ļæĺ¯ ä¸Ģ个 +æ¼Ķ 讲 +ä¸Ģ å±Ĥ +æľīåħ³ éĥ¨éŨ +æĹł å¥Ī +ç§į ç±» +缸åħ³ çļĦ +æĪĸèĢħ æĺ¯ +æī¶ æĮģ +å¤ļ æķ° +çļĦ ä½ľåĵģ +ä¸ĭ ä¸ĢæŃ¥ +å¸Ī åĤħ +é«ĺéĢŁ åħ¬è·¯ +好 åıĭ +ä¼ĺç§Ģ çļĦ +è¿Ľ äºĨ +æģIJ æĢķ +äºĨ åIJ§ +大 è§Ħ模 +çļĦ ä¸ĸçķĮ +æĢĢ çĸij +å· · +åħ´ å¥ĭ +æĪ ° +æĿij éĩĮ +æľĭåıĭ åľĪ +åĨ¬ 天 +ä¸Ńåįİ äººæ°ij +åįı åķĨ +è¯Ħ éĢī +æĹ Ń +å¢ŀåĬł äºĨ +åıĹ ä¼¤ +ä¸Ģ èĤ¡ +便 æį· +ä¸ ij +é¹ ¤ +å¤ĸ è§Ĥ +å·¥ç¨ĭ å¸Ī +åĴĮ åħ¶ä»ĸ +è¿Ļ å°± +ä¸Ńå°ı ä¼ģä¸ļ +西 åĮĹ +åĽ½æľī ä¼ģä¸ļ +èĭ¥ æĺ¯ +åı¯ æĥľ +çĶŁ æĹ¥ +åĩ ½ +ä¹° åįĸ +ç¥Ŀ ç¦ı +人æ°ij 群ä¼Ĺ +åħī æĺİ +åħ¬ å¯ĵ +æĺ¯ è°ģ +æĪij çŁ¥éģĵ +è¯Ń æĸĩ +æķı æĦŁ +ä¸įéĶĻ çļĦ +æĿ¥ 讲 +æ³¢ åĬ¨ +çļĦ 第ä¸Ģ +åľ° éľĩ +åľ¨ åħ¨åĽ½ +骨 å¹² +å®ī ç½® +å®¶ ç͵ +ä¸İ æŃ¤ +ä¸İæŃ¤ åIJĮæĹ¶ +åıĹ çģ¾ +çĥŃ çº¿ +çļĦ æĬĢæľ¯ +æµĭ éĩı +ä¾Ŀ èµĸ +ä¸ŃåĽ½ çļĦ +çī¹ æĢ§ +è¾ĥ é«ĺ +è¸ © +ä¼ļ åľ¨ +建 éĢł +导 èĪª +æĥ³ èµ· +åħ¨ ä¸ĸçķĮ +建 æĿIJ +ç¯ Ģ +çļĦ åŁºç¡Ģ +èĩªåĬ¨ åĮĸ +åīį åIJİ +çĿ¡ çľł +æİ¨ è¡Į +æį® äºĨè§£ +ä»Ģä¹Ī æĹ¶åĢĻ +ä¸į åĸľæ¬¢ +çħ¤ çĤŃ +éĤ£ä¹Ī å¤ļ +å¸Ĥåľº åĮĸ +ä¸į管 æĺ¯ +ç«ĭ åľº +éĥ½ 没 +课 é¢ĺ +æĪij们 å°Ĩ +è¿ĩ çļĦ +åĨį åĬłä¸Ĭ +çĪ ¾ +身 æĿIJ +çĶ· 女 +è¿ľ è¿ľ +çĶ· çĶŁ +èĩªèº« çļĦ +è´Ł æĭħ +çϾ ä¸ĩ +西 çıŃ +西çıŃ çīĻ +åĩĢ åĪ©æ¶¦ +æ¾³ 大 +澳大 åĪ©äºļ +ä¸į åİ» +æī¿ åıĹ +楼 çĽĺ +å¢ĥ åĨħ +æ·· åĩĿ +æ··åĩĿ åľŁ +æĢĿæĥ³ æĶ¿æ²» +å¸Ĥ åĮº +æĭĽ æłĩ +åĽ¢ ä½ĵ +è¿Ľ 度 +åĨĽ éĺŁ +åıį å¼¹ +äºĨä¸Ģ äºĽ +æİ¥ å¾ħ +çļĦ åŃ¦ä¹ł +éħį éĢģ +é£Łåĵģ å®īåħ¨ +æĽ¿ 代 +æĺ¯ 以 +éĢļ ç͍ +çłĶç©¶ æīĢ +ç¦ ħ +æī Ķ +éļĶ ç¦» +ä¸ĩ å¹³æĸ¹ç±³ +çļĦ è§Ħå®ļ +ç»Ļ æĪij们 +æ¿Ģ åħī +ä¼ļ åĩºçݰ +çŁŃ ä¿¡ +ç©¿ çĿĢ +æ²Ī éĺ³ +æķĻ æĿIJ +éĺ² çĸ« +ä¼ĺ èī¯ +约 å®ļ +æĪij çľģ +åħ¬ æ°ij +éģ¸ æĵ +é쏿ĵ ĩ +å·² æĪIJ为 +ä¸į å¿ħ +ç¥ĸ åĽ½ +å¹¶ æľª +åľŁ 壤 +å¾® ç¬ij +äºĭä¸ļ åįķä½į +çļĦ 游æĪı +åħ¬ 示 +åIJĪçIJĨ çļĦ +çª Ŀ +æ°Ķ 象 +å®¶ ä¸Ń +亮 缸 +åį« æĺŁ +è®° è½½ +è§Ĩ éĩİ +åľ°åĮº çļĦ +ä½Ĩ ä»ĸ +èĤĮ èĤī +äºı æįŁ +åĬŀ åѦ +ä¸Ģ è¡Į +è¯ŀ çĶŁ +åıijå¸ĥ çļĦ +çļĦ æľįåĬ¡ +çļĦ çłĶç©¶ +åij¨ æľ« +产ä¸ļ åĽŃ +é«ĺ 温 +æĪIJåĬŁ çļĦ +æŃ¥ 骤 +åŃĺ åĤ¨ +åŃIJ åħ¬åı¸ +让 她 +ä¸Ń æľī +åĺī 宾 +å¦ ® +æĺİ å¹´ +äºĨ åIJĹ +äºī è®® +æĪ Ī +ä¸Ģ æľ¬ +ç¾İ丽 çļĦ +ä½ł 说 +大 人 +æĶ» çķ¥ +ä¸į æľĥ +å¾ħ éģĩ +ä¸Ģ è¾Ĩ +çīĪæĿĥ æīĢæľī +æ°ij ä¼Ĺ +åĬ٠夫 +å±ķ ä¼ļ +大 èĦij +æ¯ı æľĪ +å°ı 麦 +æµĻæ±Ł çľģ +çļĦ æīĢæľī +ä¸ĭ æ»ij +èĵĿ èī² +è¦ģ æĥ³ +åѦçĶŁ çļĦ +å½ĵ ä½ł +ä½ľ æĪĺ +å®¶ 乡 +å¤ļ åIJį +é«ĺ äºİ +åĿļ 强 +è¿ŀ éĶģ +åIJİ æŀľ +人 äºĭ +ç´ ħ +æ¿Ģ åĬ¨ +è¿Ľ æĶ» +ç© Ĩ +ä¸ ĺ +让 èĩªå·± +以 æŃ¤ +夫 人 +å¼Ģ 设 +æ°Ķ è´¨ +鸡 èĽĭ +çĦ¡ æ³ķ +åIJĥ äºĨ +åĪĨåĪ« 为 +èģĶåIJĪ åĽ½ +å½ĵ 代 +å¦Ĥæŀľ æĺ¯ +è¿ľ ç¨ĭ +åĸ Ĥ +è®° ä½ı +æ¸ħ åįķ +åIJĪä½ľ ä¼Ļä¼´ +åİ» åģļ +æķħ éļľ +模 æĭŁ +å¸Ī çĶŁ +åīį æĿ¥ +ç͵è§Ĩ åī§ +çĥŃ çα +éľ² åĩº +é«ĺ å±Ĥ +ç͵ åύ +纪 å¾ĭ +å¼Ģåıij åķĨ +éķ¿ å®ī +è½½ ä½ĵ +çļĦ å°±æĺ¯ +被 人 +åıĹ çIJĨ +篮 çIJĥ +èİ İ +交 ç»Ļ +æľªæĿ¥ çļĦ +两 大 +åIJķ å¸ĥ +çŃī 人 +çļĦ æĹ¥åŃIJ +åIJĪä½ľ 社 +æĮij éĢī +åŃĺ æ¬¾ +ç³»ç»Ł çļĦ +æĬĬ å®ĥ +没æľī ä»Ģä¹Ī +ä»İ æŃ¤ +ä¸Ń åįĪ +çĸ¼ çĹĽ +å·© åĽº +浪 漫 +缸åħ³ éĥ¨éŨ +éķ¿ åŁİ +纤 ç»´ +ä¸Ĭ éŨ +çĪĨ çĤ¸ +èµ· çĤ¹ +çļĦ éĢļçŁ¥ +èĢĮ æĿ¥ +çļĦ èĢģ +æīĭ éĩĮ +è¯Ń éŁ³ +è¾Ľ èĭ¦ +æ±Łèĭı çľģ +ç͍ äºĨ +身份 è¯ģ +æľī åĬ© +æľīåĬ© äºİ +çī© èģĶç½ij +åĩº éŨ +å¼Ł åŃIJ +æĥ ¹ +è¿Ļä»¶ äºĭ +æĪij们 åı¯ä»¥ +çļĦ çĶŁåij½ +æľīä¸Ģ ç§į +åºĹ éĵº +åıĮ æīĭ +çļĦ æ¶Īæģ¯ +èĢIJ å¿ĥ +å°´ å°¬ +éĤ£ 天 +é¦ĸ æī¹ +æĺ¯ä¸Ģ å®¶ +人 æ°Ķ +åıį æŃ£ +æĪij åĴĮ +å®ł çī© +ä¸į 对 +寻 æ±Ĥ +缸 ä¼¼ +åľ¨ ç¾İåĽ½ +åı« åģļ +åĹ İ +ç«ĭ è¶³ +ç͍ éĢĶ +åħ Ĩ +大 æ°Ķ +åIJij ä¸Ĭ +ä»ĸ å°± +é¡¹çĽ® 建设 +èĭ¥ å¹² +æĺ¯ æľī +æ¿Ģ æĥħ +çļĦ æĦıä¹ī +æĺ Ń +严éĩį çļĦ +å¯Ĩ éĽĨ +èĪŀ è¹Ī +èᣠèİ· +èİ· æĤī +æ±Ł åįĹ +åģĩ å¦Ĥ +æĪ· å¤ĸ +线 ç´¢ +ç§ģ 人 +转åŀĭ åįĩ级 +çļĦ ä»·å̼ +åįķ çĭ¬ +èĢģ çϾå§ĵ +å°į æĸ¼ +åĽ½éĻħ åĮĸ +ä¼° å̼ +æľįåĬ¡ ä¸ļ +èĩ Ń +æİī äºĨ +è§£åĨ³ äºĨ +ä¹Ł ä¸įèĥ½ +åħ ¹ +æĸ¯ çī¹ +æķħ æĦı +è¿ĩ 度 +èĬĤ æĹ¥ +çϽ çĻľ +çϽçĻľ é£İ +ç»§ æī¿ +äºĨ ä¸įå°ij +äºĮ 人 +è§ģ éĿ¢ +æĥ³ æĥ³ +å¤į åIJĪ +康 å¤į +åİ¿ åŁİ +åľ¨ åĽ½åĨħ +åľº åľ° +é϶ çĵ· +è¿Ļ 项 +çľ¼ ä¸Ń +çł ¸ +æĦŁè§ī åΰ +æŀľ çĦ¶ +æĶ¾ åħ¥ +约 æĿŁ +æİĴ æŁ¥ +车 主 +çļĦ æĦıæĢĿ +æĸ° åŁİ +æĥ³ çĿĢ +éģ Ĥ +èĮ¶ åı¶ +ä¹° æĪ¿ +åĨľ æĪ· +é«ĺ æīĭ +çİī ç±³ +æĸ°åĨł èĤºçĤİ +çħ§ æĺİ +æĮĩ åįĹ +è¸ ¢ +æķij æı´ +æĻ¯ çĤ¹ +ç¨İ æĶ¶ +çļĦ æīĭ +æŃ£ 好 +è¦ģ æĬĬ +éļı æĦı +åħ¶å®ŀ æĺ¯ +ç»Ļ èĩªå·± +è°Ī åΤ +æ¯ı天 éĥ½ +æĢģ åĬ¿ +é¢Ħ 约 +åİĨåı² ä¸Ĭ +å®Ŀ è´Ŀ +åīį è¿Ľ +ä¹Łå°±æĺ¯ 说 +çļĦ æĦıè§ģ +åı£ 罩 +åİĺ ç±³ +èĬ± è´¹ +ä½ĵèĤ² æĬķæ³¨ +åħ¬ä¼Ĺ åı· +èijĹåIJį çļĦ +å¼Ģ æĪ· +æĭį åįĸ +å²ģ æľĪ +åĨħ æ¶µ +å®Įæķ´ çļĦ +é«ĺ åİĭ +åħ¬åĬ¡ åijĺ +使ç͍ çļĦ +çĶŁäº§ 线 +妹 妹 +èµ° 访 +æĺ¯ åı¯ä»¥ +åľ¨ å®¶ +æļ´ åĬĽ +æ³° åĽ½ +è´¨ çĸij +ä¸į éģİ +天çĦ¶ æ°Ķ +缺 çĤ¹ +å°ı åŀĭ +ä¸įä»ħ æĺ¯ +é»ij æļĹ +æ¢ ¨ +æĸĩ æĹħ +è¦ģ æľī +ä¸Ń å±± +çļĦ æķ°æį® +å¾Ĺ å¾Ī +以 便 +对 ä»ĸ +åĬł 以 +çϼ çı¾ +设 å®ļ +èĤļ åŃIJ +éĿ ĸ +å¥ī çĮ® +ä¸į åıĺ +åı£ ç¢ij +åľ¨ åĵªéĩĮ +ä½ IJ +è¿Ļ 两个 +çļĦ æĸ¹åIJij +æŀ « +äºĮ 次 +çīĩ åĮº +éł IJ +ç£ Ĭ +æĭ¿ çĿĢ +å·²ç»ı æĪIJ为 +ä¹ĭ ä¸Ĭ +å®Ĺ æĹ¨ +奶 奶 +é«ĺæĸ° åĮº +社 æľĥ +è·Ł 踪 +æľįåĬ¡ ä¸Ńå¿ĥ +æī ¯ +æīĭ æĮĩ +礼 çī© +宿 èĪį +ç͍ å¿ĥ +æıIJé«ĺ äºĨ +亮 çĤ¹ +ä¸į æĦ¿æĦı +æĴŃ æĶ¾ +å¤ļå°ij éĴ± +没 ä»Ģä¹Ī +æķ° åįģ +æĢ» çĽij +çļĦ åŁİå¸Ĥ +æī¾ åΰäºĨ +åĨħ åľ° +åΰ çİ°åľ¨ +æĪĺæĸĹ åĬĽ +åİŁ å§ĭ +åĥ § +åĢĴ æĺ¯ +æľĢ åħ· +è´«åĽ° æĪ· +éĢģ åΰ +级 åĪ« +åĩº èµĦ +æĪª æŃ¢ +ç§į åŃIJ +èĥ½ ä¸įèĥ½ +幸 è¿IJ +èĸ ĩ +项 éĵ¾ +æĮĤ çīĮ +ä¸Ģ 樣 +ä¹ĺ 客 +èIJ½ åIJİ +ä½Ĩ æĪij +æĹ© åľ¨ +åĬ¨ 漫 +å¹³ çŃī +对 ä½ł +ä¸į æĢķ +å¤ĸ çķĮ +å¤ļå¹´ æĿ¥ +é¦ĸ 个 +æ²³ åįĹçľģ +æĪĸ åħ¶ä»ĸ +éķľ å¤´ +åįĹ æĺĮ +ä¸Ģ éĿ¢ +éĢłæĪIJ çļĦ +å´ Ķ +çŃ Ĵ +æķĻèĤ² éĥ¨ +åľ° åŁŁ +æĺĨ æĺİ +å·´ é»İ +æīĭ 游 +ä¸Ģ æĹ¶ +çł į +é¡¶ 级 +åħ± 计 +åİŁ æ²¹ +è¾ī çħĮ +说 æĺ¯ +æĸ°åįİ ç¤¾ +ç»ıåİĨ äºĨ +ä¸į æŃ¢ +è¦ģ ä¹Ī +èĢħ çļĦ +æĢ» æĬķèµĦ +è¡Į é©¶ +ä¸Ĭ å¸Ŀ +å¹´ 纪 +çIJ ¼ +ä¼ł 说 +ç²¾ èĭ± +æĸ¹ éĴĪ +æ±Ł æ¹ĸ +æĪIJ çĤº +æĢ» éĩı +æĬķ æĶ¾ +åĬ¨ çĶ» +èĹ ¤ +ç͵ æºIJ +éĴ Ļ +åIJĮ è¡Į +æĻ®éĢļ çļĦ +åĽ¾ä¹¦ é¦Ĩ +è¯Ī éªĹ +æħĪ åĸĦ +è¿Ļ 份 +主æĮģ 人 +å°± è¿Ļæł· +èĢĮ æĪIJ +èĩªè¡Į 车 +ä¸ŃåĽ½ çī¹èī² +èĤ¿ çĺ¤ +åIJ ¾ +å¼Ł å¼Ł +åıĹ çĽĬ +éĢīæĭ© äºĨ +æĺİæĺ¾ çļĦ +æĬ¥ èĢĥ +ç¬ij éģĵ +éĽĸ çĦ¶ +温 å·ŀ +éĿŀ æ´² +ç§į ç§į +åıĤåĬł äºĨ +è´§ è¿IJ +éļı 便 +å°± 没æľī +ç¸ £ +央 è§Ĩ +ç©¿ è¶Ĭ +çļĦ çݰ象 +åĩł 次 +çļĦ é£İéĻ© +æŃĮ æĽ² +æľ¬ å±Ĭ +å¹´ åĨħ +ä¸į è¶ħè¿ĩ +è¿ĩ å¤ļ +å¿ħé¡» è¦ģ +ç»ĵ 论 +åĢŁ éī´ +ç¥ŀ å¥ĩ +æľŁ æľĽ +ä¸ĵ 享 +éĿŀ常 éĩįè¦ģ +æĦıè¯Ĩ åΰ +åIJĪ å¹¶ +æĬĬ èĩªå·± +å¥Ĺ è£ħ +éŃĶ æ³ķ +å¤ı åŃ£ +ä¸į åĥı +å¢ĥ çķĮ +æĥĬ åĸľ +æľīä¸Ģ 天 +çĦ¦ çĤ¹ +æĪij 认为 +åħ° å·ŀ +ç͵ æ°Ķ +èģĶç³» æĪij们 +ç§ij æĻ® +她 说 +çļĦ æĸĩ竳 +å¥ĩ æĢª +åıĭ 好 +饮 æĸĻ +çļĦ æĶ¯æĮģ +çŃĶ åºĶ +éĩį éĩı +çij ¶ +åĩı è½» +ç§ijåѦ å®¶ +å·´ 西 +éĩijèŀį æľºæŀĦ +åħļ å§Ķ书记 +貸 款 +ç²¾ èĩ´ +ä»İ æľª +åį° åĪ· +åĽŀ 顾 +é¦ĸ éĥ½ +åıij èĤ² +éĹ® éģĵ +è¾¾ åΰäºĨ +å¿į ä¸įä½ı +æīį æľī +æįIJ èµł +ä½Ľ æķĻ +ä¸į æ¸ħ +éĺŁ éķ¿ +缸 åıį +æĬ¥ èѦ +大 åħ¨ +欧 缣 +帮 å¿Ļ +çļĦ æĻĤåĢĻ +缮 å½ķ +è¶³ 以 +èī° éļ¾ +ä»ĸ ä¹Ł +å·¥ ä½ľèĢħ +头 èĦij +缺 éĻ· +æĪIJç«ĭ äºĨ +å°± å¼Ģå§ĭ +认 åIJĮ +é»Ħ èī² +çĹħ æĥħ +覺 å¾Ĺ +è¿Ļ 两 +ä¿¡ ä»° +åľĭ å®¶ +ä¸įä»ħä»ħ æĺ¯ +çĭ¬ å®¶ +èά çļĦ +æĿIJ è´¨ +æµ· ä¸Ĭ +çĤº äºĨ +æľºåĬ¨ 车 +缸å½ĵ äºİ +å¤ļåħĥ åĮĸ +æĽ´ 大çļĦ +èĽ ® +åģĩ æľŁ +å¼ı çļĦ +交éĢļ è¿IJè¾ĵ +çľģ å§Ķ +ä¸į ç®Ĺ +æĶ¾ ä¸ĭ +éĹ ¯ +人 åľ¨ +港 åı£ +æĹ¨ åľ¨ +åij½ 令 +æŁIJ 个 +å¹³ 稳 +åıª 好 +人 人 +äº ŀ +äºĮ ç»´ +äºĮç»´ çłģ +æŀģ 为 +åĪ« å¢ħ +åħ¶ ä½Ļ +大 äºĭ +主管 éĥ¨éŨ +æĹł éĶ¡ +éĹ µ +éģŃ åΰ +说 è¿ĩ +为 ä½ł +è§£ çŃĶ +éªĮ æĶ¶ +çļĦ ç»ıéªĮ +åĮ¹ éħį +çģ« ç®Ń +豪 åįİ +æŁIJ æŁIJ +çļĦ æĹ¶ä»£ +书 éĿ¢ +æģĴ 大 +å»¶ éķ¿ +ä¸Ģ åIJĮ +æľª èĥ½ +交 æį¢ +çĶ¢ åĵģ +çŃī åΰ +åĪĨ 离 +æīĵ ç͵è¯Ŀ +å¹² çĩ¥ +è¾ĥ å¤ļ +å¤ļå¹´ çļĦ +èĥĮæĻ¯ ä¸ĭ +为 ä¾ĭ +æijĺ è¦ģ +å´Ľ èµ· +æŃ¤ åĪ» +æľī æľºä¼ļ +æĿ¡ 款 +é¢Ĩ导 å°ıç»Ħ +çļĦ 身ä½ĵ +åįķ ä¸Ģ +央 è¡Į +ä¸įæĸŃ æıIJé«ĺ +ä»·å̼ è§Ĥ +èĬ ½ +èIJ į +æ³ķå¾ĭ æ³ķè§Ħ +ä¸į éĶĪ +ä¸įéĶĪ éĴ¢ +åĩº äºİ +èĻļ æĭŁ +æį® æĤī +çĥ¦ æģ¼ +åħ¨ æĸ°çļĦ +æī« æıı +çĻ» éĻĨ +èīºæľ¯ å®¶ +çļĦ é£Łçī© +çļĦ åŃĺåľ¨ +客 åİħ +æĪij们 å°± +æŁ¥çľĭ æĽ´å¤ļ +è¯Ħ 审 +å¸Ĥ åł´ +è¬ Ľ +å·¨ 头 +ä¸ŃåĽ½ ç»ıæµİ +äºĨ èĩªå·±çļĦ +åĨ³ è®® +çĽijçĿ£ 管çIJĨ +æĬķ 票 +åĨį 度 +è¡Į çĤº +注 åħ¥ +ä½ľä¸º ä¸Ģ个 +æ¯ı个人 éĥ½ +åįķ åħĥ +è¦ģ çŁ¥éģĵ +被 称为 +ä¹ĭ éĻħ +è§£ éϤ +ä¸ ¸ +æº « +ä¸ī æĺŁ +é²ľ æĺİ +ä¹Ł éĥ½ +æĹ¶ æľº +åĩº æīĭ +æĥħ å½¢ +åķĨ è´¸ +éĢī 举 +对 èĩªå·± +çĶŁ åĬ¨ +åħĭ æľį +个 ä½ĵ +èĭ ij +ç¨ ± +大 åݦ +æĺ¯ 对 +åĪ© æģ¯ +è¿IJåĬ¨ åijĺ +åĮĸ è§£ +åīį æ²¿ +æĦŁ æģ© +æĢ» ä¹ĭ +é«ĺæĸ° æĬĢæľ¯ +åĿĩ 为 +åħ¨ åĮº +æ°Ķ æ°Ľ +åı¯ä»¥è¯´ æĺ¯ +ä½ı 宿 +åħļåijĺ å¹²éĥ¨ +åĹ ¯ +è·µ è¡Į +çļĦ ä¸ĵä¸ļ +èĢĥ éªĮ +èķ ¾ +åħ¬ åŃIJ +çļĦ çĬ¶æĢģ +æ½® æµģ +ä¿¡ æīĺ +è´ ¼ +åIJĦ æĸ¹ +æķij åĬ© +éĿŀ常 çļĦ +æ¡¥ æ¢ģ +åħ¬ æĸ¤ +ä¼¼ çļĦ +çľĭ 好 +å±Ģ éĥ¨ +å®ī éĿĻ +éħį ä»¶ +常 è§Ħ +å¼Ģ 车 +第äºĮ 次 +ä¸Ĭ 级 +åıĤ èµĽ +å®¶ å±ŀ +强 åĬ¿ +åľ¨ ä»ĸ +åIJij åīį +ä¹ĭ åľ° +éĥ ¡ +è¡Į ç¨ĭ +èѦ åijĬ +è§Ħå®ļ çļĦ +åķĨ åŁİ +äºĶ 大 +æķĻ å®¤ +åįģ è¶³ +æīĢ以 åľ¨ +å°Ĩ ç»§ç»Ń +çŃī æĸ¹å¼ı +å®¶ ä¼ģä¸ļ +交 ä»ĺ +çĤ¹ è¯Ħ +ç»ĵ ç®Ĺ +ä¹Ł åı¯ +å¤ĸ æ±ĩ +è¿Ļç§į æĥħåĨµ +æİĪ äºĪ +å¸ĥ ç½® +æĪIJç«ĭ äºİ +é¢Ħ èѦ +管çIJĨ 人åijĺ +å©ļ 礼 +ç»ĵæĿŁ åIJİ +åħ¥ éĢī +æĹł æ¯Ķ +åĴĮ åıijå±ķ +çϽ éħĴ +çİ© åħ· +ä¸ĩ ç¾İåħĥ +çļĦ æĪIJ绩 +æĭį çħ§ +èĢĥèĻij åΰ +ä¼ģä¸ļ åıijå±ķ +äºĨ 个 +çĶŁ æ°Ķ +çļĦ 女人 +äºĶ åįģ +çĪ· çĪ· +纽 约 +éĥ½ 被 +ä¸Ĭ 课 +çĽ ¡ +ä¼łç»Ł æĸĩåĮĸ +æ½ľ åľ¨ +åıij å°Ħ +ä¸Ģ 身 +éĺ² å®Ī +åĪ ® +é¢ĺ 缮 +åľ¨ åĨħçļĦ +ç¾İ 好çļĦ +è¿ĻéĩĮ çļĦ +ä¸Ģ ä¸Ŀ +人 åĿĩ +åĢ¡ 导 +身 åIJİ +æī© å±ķ +大 éŨ +å°± 被 +该 é¡¹çĽ® +æŀ¶ æŀĦ +ä¸Ģ åı£ +ä¿¡æģ¯ æĬĢæľ¯ +å¼Ģ ä¸ļ +æĶ¶ åıĸ +ç½ij 页 +æĶ¯ æı´ +å°ģ éĹŃ +å¡ij éĢł +大 èĥĨ +å¿«éĢŁ åıijå±ķ +çľĭ ä¼¼ +æ¸ Ŀ +è¿Ļæł· ä¸Ģ个 +模 åĿĹ +注æĦı åΰ +çł´ è§£ +èĩª ä»İ +åijµ åijµ +ä¹ĭ å¾Į +ä¹ĭ æĹħ +è·Ł æĪij +æ³ķ 人 +æİĴè¡Į æ¦ľ +åĿļ å®Ī +好 å¤Ħ +çŁ³ 头 +å¹¶ å°Ĩ +èĪ ± +æŃ ĩ +两 岸 +å¤ļ ä¹ħ +象 å¾ģ +个æĢ§ åĮĸ +çļĦ è§Ĵ度 +å¸ Ĩ +ç¦ı å·ŀ +æŁ¥ å¤Ħ +两 åĽ½ +åIJ¸å¼ķ äºĨ +é¦ĸ å¸Ń +大 åĵ¥ +é¤ Ĭ +涨 å¹ħ +éĢī ç͍ +許 å¤ļ +èIJ½ æĪ· +åĵĪ å°Ķ +åĵĪå°Ķ 滨 +åģļ ä»Ģä¹Ī +以 åħį +é¾ į +æĹł éľĢ +åΰåºķ æĺ¯ +æĢ ¡ +åijĬè¯ī ä½ł +éĺ² æ°´ +è¿Ļ æĹ¶åĢĻ +欢 ä¹IJ +转 åIJij +è¿Ļ个 åľ°åĽ¾ +åħ¥ é©» +èįī åİŁ +æĹ¶ä»£ çļĦ +åıĺ åĬ¨ +åĬłå¼º 对 +åģ¶ å°Ķ +å®Ī æĬ¤ +æ°Ķ 温 +人 éĹ´ +æľĿ é²ľ +ç»ı è´¹ +åĽŃ æŀĹ +å·¥ åľ° +è§Ħ æł¼ +åĩł åįģ +è¯ķ åĽ¾ +å¦ ĥ +éĤ£ æĹ¶åĢĻ +å¼ĺ æī¬ +ä¸ļ çķĮ +çļĦ éĢŁåº¦ +ä¼ļ ä¸įä¼ļ +èIJ¥ æĶ¶ +å°ıå¾® ä¼ģä¸ļ +çľĭ è¿ĩ +æĬĬ ä»ĸ +éģµ å¾ª +è¿Ļ è¾¹ +没æľī 人 +å£ ¶ +æ¹ĸ åįĹçľģ +æŀģ åħ¶ +çļĦ人 çĶŁ +ä»ĸ è¿ĺ +转åĮĸ 为 +èµ° è¿ĩ +æĬ± çĿĢ +çīĽ å¥¶ +ä¸ĩ 亩 +å¿ĥ æĢģ +æĹ¥å¸¸ çĶŁæ´» +ä½ĵ æ£Ģ +æĻ ĥ +çŃī é¢ĨåŁŁ +æĩī 該 +åı¯ä»¥ çľĭåΰ +æī¾ ä¸įåΰ +èĢģ å¹´ +æĬĬ æĪij +积 åĪĨ +梳 çIJĨ +ç» ³ +çļĦ æĶ¿æ²» +å¸Ŀ åĽ½ +éĻª ä¼´ +æ´Ľ éĺ³ +åħ¬ æŃ£ +å¼Ģ åı£ +çī¹èī² çļĦ +åĽ° å¢ĥ +ä¸Ĭ æľī +ç«ĭ ä½ĵ +æīĵ å·¥ +åķ¤ éħĴ +åľ¨ éĤ£éĩĮ +éĤ£ è¾¹ +个 åĪ« +ä¸Ģå®ļ æĺ¯ +çļĦéĩįè¦ģ æĢ§ +主 å¼ł +åĴĮ æľįåĬ¡ +ä¸Ĭ ç½ij +è¡¥ åĬ© +åıª éľĢ +å¼ ¦ +éģ ® +åĬĽ äºī +度 è¿ĩ +èij ¬ +é¡¿ æĹ¶ +éĦ ī +纺 ç»ĩ +åľ° åĿĹ +ä¿¡ç͍ åį¡ +ç½ļ 款 +åijĬè¯ī æĪij +éĽ Ļ +书 çĶ» +è¨Ń è¨Ī +æĢ» ä¼ļ +åΤ åĨ³ +ä¿¡ èªī +个 èĤ¡ +å¹³ 常 +æĢİ éº¼ +ä½ĵ çİ°åľ¨ +é»Ħ æ²³ +åĽĽå·Ŀ çľģ +羣 缸 +åIJĦ项 å·¥ä½ľ +åĬ¨ åijĺ +å³° ä¼ļ +ä¸Ģ æľŁ +æľī ä¸Ģå®ļçļĦ +é«ĺ度 éĩįè§Ĩ +ç¹ģ èᣠ+åıijçݰ äºĨ +ç½ij 红 +æīĭ æ³ķ +å®¶ åĽŃ +仪 åύ +è¾ĥ ä½İ +çļĦ å®īåħ¨ +æ¡ IJ +ä»ĺ 款 +æĬij åζ +åįĵ è¶Ĭ +æŃ£ éĿ¢ +åĵ ij +强 åζ +ä»Ĭ天 çļĦ +æĪĺ èĥľ +楼 å¸Ĥ +æĭ¿ ä¸ĭ +é¢ľ å̼ +举 éĥ¨ +çłĶ åζ +çļĦ æĪĺçķ¥ +åľ¨ ä¸Ģ个 +ä¸ī 人 +å®Į äºĨ +æĸ° æĬĢæľ¯ +ç»ıæµİ æķĪçĽĬ +å¯Į æľī +æ¾³ æ´² +åĬ© çIJĨ +é¢Ĩ åıĸ +è° Ń +çĩĥ çĥ§ +ç´ł åħ» +éĤĦ æľī +è¿Ľ èĢĮ +ä»Ģä¹Ī æĺ¯ +çłĶç©¶ ä¸Ńå¿ĥ +éĢĤ ç͍äºİ +æİ¥ æĶ¶ +失 æľĽ +äºĮ 级 +éĹ´ çļĦ +åİŁ æłĩé¢ĺ +èªį çĤº +æį ¡ +对 çĿĢ +对 éĿ¢ +ä¸Ń åİŁ +éĵ ĥ +çĶŁäº§ çļĦ +åıijå¸ĥ ä¼ļ +士 åħµ +è¿Ļ åı¥è¯Ŀ +ç¼´ 纳 +ä¸Ģ个 个 +åѸ çĶŁ +çĸij éĹ® +交 èѦ +示èĮĥ åĮº +天 使 +åľ¨ ä¸Ĭæµ· +åIJĮ æĻĤ +è½» æĺĵ +å͝ä¸Ģ çļĦ +çĥŃ éĹ¹ +ä¹IJ è§Ĥ +çļĦ 身份 +åĸĦ äºİ +大 åİħ +èĤ¯å®ļ æĺ¯ +éĺ² çģ« +å¤ĸ åĩº +æį® 说 +é¡¹çĽ® çļĦ +ä¸Ģ åı° +èĻļ åģĩ +ä¸Ģ ç¬Ķ +ç«ĭ æ³ķ +严 èĤĥ +æī¿ åĬŀ +åįģ åĩł +çļĦ 空éĹ´ +æľ¬ ç½ijç«Ļ +åģļ å¾Ĺ +ä¿Ŀ 温 +æľĪ åĪĿ +åľ¨ ç½ijä¸Ĭ +åIJĦ æĸ¹éĿ¢ +ä¸ī 天 +交æĺĵ æīĢ +è§£ æŀIJ +åħļ ä¸Ń央 +è¿Ľ åĩºåı£ +åĴĮ 社ä¼ļ +次 æķ° +ä¹ĭ å®¶ +ç»´ 度 +æ´¾åĩº æīĢ +产çĶŁ äºĨ +带 æľī +å¾Ī 强 +æľīäºĽ 人 +å¹´ åIJİ +äºĨ 许å¤ļ +å¯Ĩ 度 +åѦ æľŁ +çıł æµ· +æľĢå¤ļ çļĦ +è¾¹ ç¼ĺ +容 éĩı +第äºĮ 个 +ä¸Ģ缴 æĺ¯ +ä¸į ç¦ģ +æŃ ² +ä»ĭç»į äºĨ +ä¼ĺ éĽħ +æ¯Ķ è¼ĥ +èģĮ ä½į +温 æŁĶ +æľī éĴ± +æľĢ é«ĺçļĦ +åįļè§Ī ä¼ļ +ä¸į æĪIJ +éĶĻ äºĨ +è¯ģ çĽij +è¯ģçĽij ä¼ļ +æĪIJ 人 +åĿĩ åĮĢ +æľī åĪ© +è¶Ĭ åįĹ +æīĵ äºĨ +好 åIJĥ +ç³» çµ± +è·Ł éļı +çļĦ åľ°ä½į +æŃ£ å¦Ĥ +ç¨į å¾® +åį° åıij +åĪĽ ç«ĭ +é£İ åħī +å°Ĩ æĪIJ为 +ä¸į é«ĺ +é¢ij ç¹ģ +设 æľī +ä¼ ŀ +æĭĨ éϤ +å½± åĥı +æ¸Ĺ éĢı +å¹´ å¼Ģå§ĭ +ç½ij æĺĵ +è¦ģ åģļ +ç͵åĬ¨ 车 +羣 å¿ĥ +æµ· åĨĽ +ä¼ł æĿ¥ +å·® åĪ« +è°¨ æħİ +çĥŁ åı° +åįĥ å¹´ +è¯ģ å®ŀ +çIJ ª +çļĦ åħ·ä½ĵ +åΰ å¤Ħ +ä¸į å®ľ +èľ Ģ +èĥ½åĬĽ åĴĮ +çīº çī² +çļĦ éĴ± +大 éĺŁ +é¦ĸ è¦ģ +ä¸į æĦ¿ +çİ« çij° +人æ°ij ç½ij +è¿ĺæĺ¯ è¦ģ +åĽĽ å¹´ +æį٠伤 +çļĦ åģļæ³ķ +éĿ Ī +è¡Ķ æİ¥ +åIJĪ æĪIJ +没 人 +éŨ æ§Ľ +ä¿¡ è´· +çļĦ 缸åħ³ +举 é£İ +社 ä¿Ŀ +ä¸ĭ 游 +åĿĹ éĴ± +è¿ĩ åIJİ +çļĦ åºĶç͍ +é¥ ¶ +é¢ģ åıij +ä¸Ģ å¤Ħ +åįİ å¤ı +为 ä¼ģä¸ļ +åıª ä¼ļ +ä¾µ 害 +çļĦ åĬŁèĥ½ +åѸ ç¿Ĵ +ä¸Ńåįİ æ°ijæĹı +åıijå¸ĥ äºĨ +è¿İ æİ¥ +æĪij èĩªå·± +è¿ĺ éľĢè¦ģ +太éĺ³ èĥ½ +åİ» ä¸ĸ +æĺ¯ ä½ł +åIJĪ åĬĽ +ç»ĺ çĶ» +åı° åĮĹ +çĿ£ ä¿ĥ +åĮĹ éĥ¨ +æľī å¤ļå°ij +å¾Ī éĩįè¦ģ +åĪĴ åĪĨ +åı· 线 +æĶ¾ 大 +ä¼ļ 被 +èİ· å¥ĸ +ä¹ĭ åĨħ +失 åİ»äºĨ +çݩ家 们 +éĩĩ éĽĨ +å£ ¹ +å®¶ ä¼Ļ +çϽ 天 +åĽłä¸º ä»ĸ +社ä¼ļ æ²»çIJĨ +å¼Ģ åĪĽ +ç͵ ç¼Ĩ +æĸ° ä¸Ģ代 +å¹¶ è´Ń +å°± å·²ç»ı +çļĦ 社ä¼ļ +éϤ éĿŀ +åı¯ä»¥ ç͍ +å© ī +æ¯Ķè¾ĥ 好 +å®ŀ ä¸ļ +åĪĽ åĬŀ +æıIJ èµ· +é» ĥ +ä½ı åľ¨ +å¸Ĥ æĶ¿ +éĿ¢ä¸´ çļĦ +èĥ½ åľ¨ +çŁŃ çŁŃ +羣 人 +æĺİ æĺİ +èµĦ åĬ© +çļĦ ä¸įåIJĮ +å°ı æľĭåıĭ +é¢ĺ æĿIJ +ç¾İ åij³ +æĺŁ åº§ +ä¸į ä¸Ģæł·çļĦ +çľĭ ä¸Ĭåİ» +ä¸Ģ æł¹ +广 å·ŀå¸Ĥ +åıijçĶŁ çļĦ +é«ĺ ç§ijæĬĢ +ä¸Ģ è¾ĪåŃIJ +交 åıī +ä½ĵç³» 建设 +åĽłä¸º æĪij +çıį æĥľ +ä¸Ĭ åѦ +æĪĺ æľ¯ +æŃ¤ ç±» +交 å¾Ģ +æĮī æij© +人们 çļĦ +åħ¶ 實 +åİŁ æĿIJæĸĻ +渴 æľĽ +缸 å¤Ħ +å¾® å¾® +æ® · +ä¹ĺ åĿIJ +å¼Ģå±ķ äºĨ +é«ĺ åĵģè´¨ +æĹłäºº æľº +ä¸įæĺ¯ å¾Ī +çļĦ æĬķèµĦ +èĬĤ çľģ +èĩ ī +ç²¾ éĢī +çļĦ æłĩåĩĨ +åįĹ éĥ¨ +认è¯Ĩ åΰ +å¹³ éĿĻ +èĹ ¥ +æī« é»ij +æī«é»ij éϤ +æī«é»ijéϤ æģ¶ +éĢĻ ç¨® +建çŃij éĿ¢ç§¯ +ç¡® ç«ĭ +管çIJĨ åĬŀæ³ķ +æĦı å¿Ĺ +ä¸ ¨ +让 åŃ©åŃIJ +æķij çģ¾ +å½ĵ ä»Ĭ +çģ« çģ¾ +åIJĦ éĥ¨éŨ +ä¾µ çĬ¯ +æ¯ı åij¨ +æı ½ +ä¸Ģ次 æĢ§ +åħ¶ä»ĸ 人 +éĶĻ è¿ĩ +ä¸İ åħ¶ +åĭĩ æ°Ķ +çĩĥ æ°Ķ +é¦ĸ å±Ĭ +æľį 饰 +ç² ¥ +å®Į æ¯ķ +å°± æĬĬ +åĬŀäºĭ å¤Ħ +ä¸Ģä¼ļ åĦ¿ +离 ä¸įå¼Ģ +å¦Ĥæŀľ æĤ¨ +ä»ĵ åºĵ +导 å¸Ī +åIJĪéĢĤ çļĦ +毫 ç±³ +å®īåħ¨ æĢ§ +ä¾Ŀ çħ§ +产ä¸ļ åĮĸ +ä½ł çľĭ +羣çļĦ å¾Ī +åѤ çĭ¬ +éĺ² å¾¡ +å¾Ī ç®Ģåįķ +é£İ æ°´ +ä½Ĩ ä¹Ł +æİ¨ åĩºäºĨ +æ°ijèIJ¥ ä¼ģä¸ļ +çłģ 头 +å¤įæĿĤ çļĦ +ç»ĦæĪIJ éĥ¨åĪĨ +åħħ满 äºĨ +è¿ij åĩłå¹´ +çľģ æĶ¿åºľ +æľī å¿ħè¦ģ +éĻ ³ +ä¹ĭ ç±» +ä¹ĭç±» çļĦ +æĢ§ ä»· +æĢ§ä»· æ¯Ķ +åķĨ åºĹ +å¸Ĥ å̼ +人æīį åŁ¹åħ» +æ·± åıĹ +管çIJĨ å±Ģ +æģIJ æĥ§ +ä»ħ æľī +æĬµ è¾¾ +æµ· åħ³ +èµĭ äºĪ +äºĭ åĦ¿ +ä»· éĴ± +æīĭ ä¸Ĭ +èĩª å¾ĭ +åħ³ çα +享 æľī +éģĹ æĨ¾ +å¾Īå¿« å°± +æĽ´ å¿« +æłĩ è¯Ĩ +åºĨ ç¥Ŀ +ä¹Ł 好 +ä¸į æĺĵ +æĪij å¾Ī +æĶ¹éĿ© åıijå±ķ +å¤ĸ åľ° +æĬµ æĬ¼ +è¯Ĺ 人 +åİķ æīĢ +æĸ° åªĴä½ĵ +èĸ Ľ +è°Ī è¯Ŀ +ä¸Ģå®ļ ç¨ĭ度 +èµ° åľ¨ +æľĢ 强 +åĬŁ çİĩ +åħ± è¯Ĩ +大 æ¡¥ +ä¸ĭ æĸ¹ +å¤ĸ èµĦ +ç¢ ± +å·¡ è§Ĩ +æ¹ĸåĮĹ çľģ +个 çϾåĪĨ +个çϾåĪĨ çĤ¹ +çļĦ 责任 +çļĦ åĵģçīĮ +åĬ© æİ¨ +åĪĽéĢł äºĨ +ä»» èģĮ +å¿« æį· +æĿij åºĦ +åİ» çľĭ +æīį èĥ½å¤Ł +å± ¤ +æĪij å®¶ +æĺ¯ä¸Ģ 款 +ç¾ ħ +åĨ° éĽª +æŀģ 大 +çģ¯ åħī +éĨ ĭ +ä¸İ åħ¶ä»ĸ +æıIJåĩº çļĦ +éĿł è¿ij +è°ĥ åĬ¨ +å°½ åı¯èĥ½ +åıij åĬĽ +ç»Ļ 她 +éĢĤ éĩı +è·¨ åĽ½ +åħĪ è¡Į +æĸ° æĿIJæĸĻ +ä½ľ äºĨ +满 äºĨ +ä¸į 满 +çļĦçľ¼ çĿĽ +çľĭ å¾Ĺ +è¿Ļ ä¸Ģ次 +é½IJ åħ¨ +çļĦä¸Ģ éĥ¨åĪĨ +ä¸ Ļ +æ¸ħ æĸ° +說 æĺİ +身边 çļĦ +æīĢæľī 人 +å½° æĺ¾ +è± ¹ +åį ¿ +è¿IJ 转 +æĮĩ å¼ķ +å¸Ĥ åħ¬å®īå±Ģ +åıĤ å±ķ +ä¹ĭ æĹ¶ +éĩijèŀį æľįåĬ¡ +èµĦæľ¬ å¸Ĥåľº +èĥ½ 让 +å¿ĺ äºĨ +天 åłĤ +æ¯Ķå¦Ĥ 说 +éĬĢ è¡Į +èĽĭ ç³ķ +çĶ © +æł¸ å®ŀ +æĻ® 京 +ä¼ĺ ç¾İ +åı£ èħĶ +漫 çĶ» +çľ¼ éĩĮ +äºĨ ä¸ĭæĿ¥ +æĪij们 ä¹Ł +ä¾ į +为 ä¸Ńå¿ĥ +å¥ĩ 迹 +éĿĴ çĿIJ +æĪªèĩ³ 缮åīį +åĩº ä¾Ĩ +æĢ» åħ¬åı¸ +å¼¥ è¡¥ +ç®Ĺ æ³ķ +å·¥ä½ľ 室 +æīĢ以 æĪij +æ°´ åĪĨ +æīĢ å±ŀ +ä¸į 说 +ä½Ĩæĺ¯ åľ¨ +è¦ģ åİ» +åĪĽä¸ļ èĢħ +ä¸į æ¸ħæ¥ļ +åĽĽ åij¨ +æĺ¯ ä»İ +çļĦ æł¹æľ¬ +çģ ¶ +æ¯Ľ æ³½ +æ¯Ľæ³½ 举 +æµ· åı£ +åĽĽ åįģ +ä¹Ł 被 +èģ · +ä¸Ģ æīĭ +绩 æķĪ +çļĦ çĶ·äºº +书 ç±į +ä¸Ģ èĦ¸ +大 äºİ +鼶 éĥ¨ä»¶ +åħ³ æĢĢ +å¹³ ç±³ +æļ´ éľ² +å¾Ĺ å¤ļ +ä¸ī 级 +æľ¬ åij¨ +两 èĢħ +对 ä¸ŃåĽ½ +åıª è§ģ +欧 ç¾İ +å¦Ĥæŀľ æľī +å·²ç»ı æĺ¯ +çľĭ å®Į +çģ« éĶħ +èµ IJ +ä¸Ģ éģį +æĦŁ åĨĴ +ç»ĵ å±Ģ +ä»ĵ åĤ¨ +å®ŀ åľ° +å̻ ç»ıçIJĨ +ä¹Łä¸į çŁ¥éģĵ +碰 åΰ +åIJĪ è®¡ +客æĪ· çļĦ +ç½Ĺ 马 +æĦī å¿« +é£ Ľ +çĥŃ çĥĪ +伦 æķ¦ +åĮ» ä¿Ŀ +éĺ¿éĩĮ å·´å·´ +åĨį 说 +为 åŁºç¡Ģ +çĶŁäº§ ç»ıèIJ¥ +è¿ĻäºĽ 人 +åĪĹ è½¦ +æ²³åĮĹ çľģ +è¿Ļ 段 +æ´»åĬ¨ ä¸Ń +å© · +çĶŁ çIJĨ +ä¸ŃåĽ½ 人æ°ij +éĦ Ĥ +åIJ¬ åıĸ +å¤į ä¹ł +æľī çĽĬ +æĶ¶ æĭ¾ +å¾Ī åı¯èĥ½ +ç½ij绾 游æĪı +们 çļĦ +èµĭ èĥ½ +éļ¾ å¾Ĺ +åĪĨ æīĭ +羣 è¯ļ +åħ¬åı¸ åľ¨ +åĿĩ è¡¡ +åı£ åij³ +çīµ å¤´ +ä¸Ģèά çļĦ +轿 车 +çŃī äºİ +æ²ī é»ĺ +æĪij éĥ½ +å°ı ç¨ĭåºı +ä¸Ģ åī¯ +æī¿ è½½ +åľ° è´¨ +çķĮ éĿ¢ +ç͵ æľº +çĦ¦ èĻij +éĶĢåĶ® é¢Ŀ +æĸ° 车 +ä¸Ĭ 游 +主 æ¼Ķ +éļIJ ç§ģ +åıijå±ķ æĪĺçķ¥ +çļĦ åĬªåĬĽ +å¼Ģ åħ³ +è§£åĨ³ éĹ®é¢ĺ +çĿ£ 导 +对 æĬĹ +å¾Īå¤ļ 人éĥ½ +æĹł æķĪ +产åĵģ è´¨éĩı +å®ī å¿ĥ +åįİ äºº +ä¸į 符åIJĪ +èĩª å®¶ +éĺµ å®¹ +çļĦ åIJĦç§į +çļĦ çIJĨ念 +çļĦ æĸĩåĮĸ +为 èĩªå·± +å±± æ°´ +游 æ³³ +éľĩ èį¡ +çĶŁæ´» æĸ¹å¼ı +è¿ľ 离 +çŁ³ åĮĸ +æŃ¤ äºĭ +æĺ¯ 羣çļĦ +çļĦ æ¯Ķä¾ĭ +ç͍ ç͵ +奥è¿IJ ä¼ļ +ä¿Ŀ å®ī +èĽĭçϽ è´¨ +çļĦ å¿ĥçIJĨ +å· « +åı· çłģ +æ°Ķ ä½ĵ +åıij æĶ¹ +åıijæĶ¹ å§Ķ +åĮ» å¸Ī +æ¶Ĥ æĸĻ +æĺ Ĭ +å¸Ĥ 级 +ä¸ĸçķĮ çļĦ +åĪĨåĪ« æĺ¯ +çł´ 产 +ä¸Ģ æĿ¯ +æĭī å¼Ģ +å¹³ åĩ¡ +çļĦ åıijçĶŁ +åĬ¨ æīĭ +ä¸Ģ缴 以æĿ¥ +æīĭ å·¥ +éĩĮéĿ¢ çļĦ +æĹł åħ³ +ä»ĭ åħ¥ +èµ° ä¸Ĭ +å°±æĺ¯ è¦ģ +å¹´ éĹ´ +åĩº çı¾ +å½± éŁ¿ +å¹ħ 度 +éĽ ģ +éģĵ åħ· +缮çļĦ åľ° +åIJİ èĢħ +ä¸Ĭ æ¼Ķ +äºĨ åĩł +æ®ĭçĸ¾ 人 +å¿Ļ ç¢Į +æĺ¯åIJ¦ æľī +å¹¶ 对 +ä¼ļ 导èĩ´ +æ°´ åºĵ +ç»Ĩ èĩ´ +åIJİ æĤĶ +å¿ĥ æĢĿ +åģļ äºĭ +åİĤ æĪ¿ +çĿ ¿ +è¿IJèIJ¥ åķĨ +头 éĥ¨ +çļĦ è§Ĵèī² +æĺ¯ ä»ĸ +æĹ¢ æľī +å°ıæĹ¶ åĢĻ +强 åĬ² +主 æĴŃ +åħ¨åĽ½ åIJĦåľ° +æį ı +æįŁ åĿı +åķĨ ä¼ļ +ä¿Ŀ ç½Ĺ +çľģ å¸Ĥ +éļ§ éģĵ +æľī ä¸įå°ij +è¦ģ åľ¨ +建设 é¡¹çĽ® +ç³ĸ å°¿ +ç³ĸå°¿ çĹħ +æĿ¡ä»¶ ä¸ĭ +ä¼ĺè´¨ çļĦ +é¦ĸ åıij +å½ĵæĹ¶ çļĦ +丰 çͰ +大 çĽĺ +缸 ç»§ +å®ģ å¤ı +åħ¥ ä½ı +æĪij è¿ĺ +åħĭ æĸ¯ +å®ļ ä»· +å¹³æĸ¹ åħ¬éĩĮ +çļĦ çŁ¥è¯Ĩ +æĪij们 ä¼ļ +åħĥ å®Ŀ +ä½ĵ éĩį +è³ £ +对 æĪij们 +çŁ³ å®¶ +çŁ³å®¶ åºĦ +ç²¾ åįİ +å½¢ çĬ¶ +åıĹ åΰäºĨ +ä¿® 订 +ç¾İ åľĭ +é«ĺ æ¸ħ +çľ¼ éķľ +è§īå¾Ĺ èĩªå·± +带 ç»Ļ +åĶ® ä»· +éŨ 票 +åŃķ å¦ĩ +ç͵è§Ĩ åı° +åıij ä½ľ +çļĦ åij³éģĵ +éķ¿ è¿ľ +åħ¬åħ± æľįåĬ¡ +æŃ£å¸¸ çļĦ +æľī è¿ĩ +é£İ æĥħ +æ¯Ķ éĩį +åIJ » +管çIJĨ å·¥ä½ľ +综åIJĪ æĢ§ +å·² 被 +说 èµ· +æİĴ æ°´ +ä¸įæĸŃ åľ° +æĥħ æĢĢ +è¾ĵ éĢģ +è¿ĩ æķı +çļĦ åı¯èĥ½æĢ§ +æľį ç͍ +æľī 许å¤ļ +å§Ķ åī¯ä¹¦è®° +åĮĸå¦Ĩ åĵģ +æļĤ åģľ +æĬķèµĦ 人 +çıŃ çº§ +说 çĿĢ +åįĹ åĮĹ +åĪĨ è¡Į +çıł å®Ŀ +å¯ ¶ +å¢ŀ å¤ļ +被 åĬ¨ +ç®Ĭ çļĦ +éĹľ ä¿Ĥ +çļĦ èĦ¸ +æĥ Ł +ä¸į ä¸Ģå®ļ +ç¶ Ń +çģ« çĪĨ +ç§Ł éĩij +çŀ § +éĩį 建 +è· ª +ä¸Ģ 種 +çļĦ åIJĪä½ľ +å®ī æħ° +ä»į æĺ¯ +ä¸ĵä¸ļ åĮĸ +è°ĥ è§£ +ä¸į 妨 +éĢĻ æĺ¯ +å¿ħ éłĪ +ä¼Ĭ æľĹ +å¾Ĺ äºĨ +æľįåĬ¡ å¹³åı° +å§ ¬ +åħĪ éĶĭ +çİĭ åŃIJ +çļĦä¸Ģ åĪĩ +æĢ» çIJĨ +åĵ ¼ +çª ij +çļĦå¿ĥ æĥħ +çļĦ éĩį大 +çij Ł +ä¸Ģ ç¬ij +åıijå±ķ ä¸Ń +åģ¥åº· åıijå±ķ +åĵģçīĮ çļĦ +ç¦ ® +ä½Ļ 人 +ä»Ĭå¹´ 以æĿ¥ +æķ° çłģ +çѾ è¯ģ +åİ» æī¾ +åŁºéĩij ä¼ļ +æĬ± æĢ¨ +æŃ£ å½ĵ +çıŃåŃIJ æĪIJåijĺ +ä¸į åIJĪæł¼ +åζ å®ļäºĨ +ç¼ĵ æħ¢ +åζ 约 +æłı 缮 +å¸Ĥåľº ç»ıæµİ +ç»ĦæĪIJ çļĦ +严 å³» +æĹ¥ 讯 +ä¸ĢçĤ¹ çĤ¹ +æĺ¯ æĢİä¹Ī +çļĦ çħ§çīĩ +éĺ» æŃ¢ +模 ç³Ĭ +ç¼ ¸ +éģķ åıį +æIJ¬ è¿ģ +éĩij éĴ± +å½ ¬ +ä¸į å®ī +æĪĺçķ¥ åIJĪä½ľ +å¡« åĨĻ +讲 ç©¶ +åħħåĪĨ åĪ©ç͍ +èĥ½ å¤ł +èij¡èIJĦ éħĴ +éĩĩç͍ äºĨ +åľ¨ ä»Ĭå¹´ +ä¸Ńå°ı åѦ +åľ¨ æĦı +çļĦ åİĭåĬĽ +ä¸į 幸 +åζ èᝠ+åı¯ä»¥ 让 +被 è¯Ħ为 +ç»Ĩ èıĮ +æĪı åī§ +åįĬ 导 +åįĬ导 ä½ĵ +è§Ĩ è§Ĵ +åĸľ æŃ¡ +å¾ģ æĶ¶ +è°ĭ åĪĴ +æŀģ 大çļĦ +çĤ¹ èµŀ +è®°èĢħ ä»İ +两 åIJį +èĩª åĬ© +èµ· æŃ¥ +æĬ¤ 士 +å®Ŀ 马 +太 åŃIJ +å°ıå°ı çļĦ +温 æ³ī +åĩºç§Ł 车 +ç§Ł æĪ¿ +两 å®¶ +éľĩ æĴ¼ +ç§ī æī¿ +ä¸Ģä»¶ äºĭ +çĥΠ士 +å®ĺ åħµ +转 身 +ä¹IJ åĽŃ +çĻĮ çĹĩ +模 èĮĥ +æĦ £ +è¿ĩåİ» çļĦ +代 ä»· +çļĦ æ¦Ĥ念 +åĩł çϾ +è´µ éĺ³ +æĭħ å¿§ +éĢĤ å®ľ +çݯå¢ĥ ä¿ĿæĬ¤ +çĥ « +ä½ł æĥ³ +æŃ¤ åIJİ +ä½ł ä¹Ł +çį İ +éϤ æŃ¤ +éϤæŃ¤ ä¹ĭå¤ĸ +è°ĥ 度 +ç§ij 缮 +æīĢ说 çļĦ +åĬ ĩ +忽 è§Ĩ +ä¸ī 次 +ä¸Ģ æĹ¥ +åŀĤ 缴 +ç«ŀ æĬĢ +éĿ¢ åĮħ +大 æĪĺ +æIJº 带 +å¦Ĥæŀľ 没æľī +åħ» æĪIJ +åĩº è¡Ģ +çα好 èĢħ +æīĵ éĢļ +èµ· è¯ī +åijĪ çݰåĩº +æŃĮ æīĭ +åľ¨ å¤ĸ +é¢Ĩ导 å¹²éĥ¨ +åĨ ¥ +èĪĨ 论 +æıIJ åıĸ +éĺ¿ å°Ķ +æľĽ çĿĢ +ä¸ī äºļ +è² ¡ +åĪ ·æĸ° +æĻļ æĬ¥ +è¿ĺæľī ä¸Ģ个 +åĨ° ç®± +ç½ij çĤ¹ +åĩº åħ· +强çĥĪ çļĦ +æĪij çĽ¸ä¿¡ +å¸ĮæľĽ èĥ½ +çīĻ é½¿ +äºĭ å®ľ +ä¸ļåĨħ 人士 +代 æĽ¿ +åıĺ å½¢ +éĽ ² +è°ĥ æİ§ +åĪĽæĸ° åĪĽä¸ļ +æĭĨ è¿ģ +æł¸ æŁ¥ +éĢ Ĺ +åħ¥ åѦ +æĦı åIJij +æı Ľ +ä¸ĭ 次 +ä¼ł è¾ĵ +ä»ĸ们 åľ¨ +èĢĮä¸Ķ è¿ĺ +æĹ¥ åľ¨ +æķĻ è®Ń +æ´» çĿĢ +çļĦ æľīæķĪ +å¤įå·¥ å¤į +å¤įå·¥å¤į 产 +æĺ¯ä¸Ģ ä»¶ +çŃī çĿĢ +å¾ © +åĭĩ æķ¢ +éģŃ åıĹ +å¥Ķ é©° +讲 座 +说 å®Į +ç»Ļ åĩº +è° ¦ +è¯Ĭ çĸĹ +çĽ² 缮 +客 è¿IJ +å°± è¿ŀ +å¼Ģ åħĥ +å¼Ģåħĥ æ£ĭçīĮ +ä¸įæĸŃ æıIJåįĩ +ç͍æĪ· çļĦ +æĴ ķ +ä¾Ľ æ°´ +ç¶ĵ æ¿Ł +ä¸Ń åĮ»èᝠ+èģĶ æĥ³ +åħ¬äº¤ 车 +èĪª çıŃ +æĬĢ è¡ĵ +å¼ķèµ· çļĦ +å° ¹ +èµĦ æ·± +åĽ½èµĦ å§Ķ +èĺ Ń +é¼» åŃIJ +éĹ ½ +æİĴ éĺŁ +è§Ĥ åħī +éģĹ åĿĢ +举 京 +é¥Ń åºĹ +ä¸įæĸŃ çļĦ +å°±æĺ¯ ä¸Ģ个 +éķ¿ ä¹ħ +çļĦ è§ĤçĤ¹ +å¨ ¶ +æĪij çİ°åľ¨ +çķ ° +å¾Ĺ åĩº +å¿ħ å®ļ +ä¸į åıĹ +åıª éľĢè¦ģ +åĽ° æī° +ç§ijåѦ æĬĢæľ¯ +çīĽ èĤī +è¾ĥ é«ĺçļĦ +è·ij æŃ¥ +æ² ¾ +èı© èIJ¨ +æľĢ å¾Į +ä¿Ŀ å¯Ĩ +æ²» å®ī +éĤ ± +常 è¯Ĩ +èĦ¸ èī² +åĮĹ å¤§ +æ±ĩ èģļ +æijĨ èĦ± +é¾Ļ头 ä¼ģä¸ļ +女 åıĭ +çŃī å·¥ä½ľ +ä¸Ń ç¾İ +èģĮ åľº +èĦij è¢ĭ +åĨĻ çļĦ +饲 æĸĻ +åĬ³ åĬ¨åĬĽ +å± ¯ +æĮģ èĤ¡ +åĽ¾ åĥı +è¿ĩåİ» äºĨ +è² ¨ +è¾ ² +éĹ® æĪij +è·Ł ä½ł +çĶŁ æŃ» +审 ç¾İ +é¢Ĺ ç²Ĵ +ä¸Ń æĸ¹ +åĬł çĥŃ +æĹħè¡Į 社 +çϼ çĶŁ +ä¸į åłª +åĤ · +æ¥ ł +åĬŀ æ¡Ī +æŁ Ħ +æĹ¢ æĺ¯ +å¤Ħ åĪĨ +羣å®ŀ çļĦ +æĬ¥ 纸 +å¸Ī çζ +å®īå¾½ çľģ +åī¯ ä¸»å¸Ń +ä¹ĭ éģĵ +导 å¼¹ +åŃ¦æł¡ çļĦ +åŁİå¸Ĥ çļĦ +è°Ī åΰ +æ¢ Ĺ +å¹³ éĿ¢ +说 ä»Ģä¹Ī +é¢ij çİĩ +éķ¿ ä¸īè§Ĵ +çļĦ åĪ©çĽĬ +é» ¨ +è±Ĩ èħIJ +å®ŀéĻħ æĥħåĨµ +æŀĹ ä¸ļ +纪æ£Ģ çĽijå¯Ł +ä½ı éĻ¢ +çļĦ æķ´ä½ĵ +åīį è¡Į +æĮ ¨ +çħ¤ çŁ¿ +å̻ è£ģ +å°ı åIJĥ +æŀģ 端 +å©Ĩ å©Ĩ +çݰ è´§ +è¯Ĺ æŃĮ +éĴ¥ åĮĻ +缩 çŁŃ +ä½Ĩ è¿Ļ +æĸ° åĵģ +è¿Ļ 对 +çŁ¥åIJį 度 +å¿ĹæĦ¿ æľįåĬ¡ +大 å±Ģ +è¡¡ éĩı +ä½ĵçݰ äºĨ +æ¡ĥ èĬ± +åIJ¸å¼ķ åĬĽ +åł ¤ +æĵħ éķ¿ +åĴ Ĵ +缸 æľº +ä¸Ģ ç«Ļ +ä¸Ģç«Ļ å¼ı +æľĢ ç¾İ +æ°¸ ä¹ħ +çļĦ éĥ¨åĪĨ +åĪĨ å·¥ +å·¥ç¨ĭ 建设 +æIJŃ è½½ +æ°´ ä¸Ń +èĮ ¨ +çļĦ æĵįä½ľ +绣 æ²» +çķħ éĢļ +åħļçļĦ åįģ +è¼ ¸ +æ¸ ¬ +ç¾İ è§Ĥ +ä¸į åĪ© +åıį æĢĿ +éªĦ åĤ² +æłĩ çļĦ +æĿĢ äºº +éĺ¿ å§¨ +é£Ł æĿIJ +åIJĥ çļĦ +åIJİ åĨį +çŁ £ +两 ä¾§ +æ¸ħ æ°´ +è¿Ľ çIJĥ +å¼Ģå§ĭ äºĨ +åIJ¬ äºĨ +çĦĬ æİ¥ +çŁ ® +å¨ Ł +为 人 +éĢģ ç»Ļ +åĨĴ éĻ© +æķ · +ç»Ī æŃ¢ +æīį çŁ¥éģĵ +è¿IJ æ°Ķ +éĢļ é£İ +æĥĬ è®¶ +ç§ijåѦ éĻ¢ +æıIJ éĹ® +太 åİŁ +缸åIJĮ çļĦ +ä» ķ +èģ ĸ +æĥħ æ³ģ +é¢Ĩ导 人 +åĩºæĿ¥ äºĨ +沿 线 +éĻ ½ +æĦŁ è¦º +ä»į åľ¨ +æ© Ļ +约 为 +åĸĿ éħĴ +ç͍ èᝠ+ä¸ĭ ä¸Ģ +æ³ķ å®ĺ +顺 åºı +åģļ ä¸Ģ个 +åĭ ¢ +æŃ ª +ç͵ ç«ŀ +ä¼´ éļıçĿĢ +ä¹ĭ åĬĽ +ä¹ĭ 人 +äºij 计ç®Ĺ +åĪ«äºº çļĦ +ç§ijåѦ åıijå±ķ +第 åħ« +å¹² æī° +女 ç¥ŀ +è¿Ļæł· åģļ +å¤Ħ åľ¨ +æ°´ è´¨ +éķ¿ æĺ¥ +å¸Ĥåľº éľĢæ±Ĥ +ç»´ æĿĥ +è̳ æľµ +æĸĩåĮĸ çļĦ +奶 ç²ī +ä¼ł è¾¾ +æīĭæľº çīĪ +æĽ¾ åľ¨ +äºĮ æľŁ +åİŁåĽł æĺ¯ +æºIJ 头 +åıĪ èĥ½ +è£ ¸ +æĬĢæľ¯ åĪĽæĸ° +æĸĩåĮĸ æĹħ游 +åıij 票 +å¹´ 级 +ä½ł ä¸į +ä¹ĭ å¿ĥ +æķ° çϾ +åIJij å¾Ģ +èĢģ å®¶ +åľĭ éļĽ +çļĦ é«ĺ度 +æľĿ éĺ³ +æ¸ħ éϤ +èĩª æľī +书 ä¸Ń +游æĪı è£ħå¤ĩ +ä¸ĩ å¤ļ +驾驶 åijĺ +ä½ł çŁ¥éģĵ +åĽ½ åºĨ +é£Ł åłĤ +æİ¥ åı£ +æĢ» æķ° +åħ¶ä»ĸ çļĦ +çĶŁåij½ çļĦ +ä½ł åľ¨ +çļĦ 缮åħī +è¿Ļ æĸ¹éĿ¢ +éĥ½ 说 +çĸĹ æ³ķ +åĭĩ 士 +åľ¨ åħ¨çIJĥ +ä¿ĿéĻ© åħ¬åı¸ +çĿ£ æŁ¥ +åĸĦ èī¯ +表 å½° +è¹ ² +è·¯ 段 +æľĥåĵ¡ è¦ı +æľĥåĵ¡è¦ı ç¯Ħ +æĪ· åŀĭ +ä¿ĥ 使 +ä¿® 建 +é«ĺ æ°´å¹³ +åģļ åĩºäºĨ +主 åľº +è¡Į èµ° +空 çϽ +æľī人 说 +è¿Ļ个 ä¸ĸçķĮ +åIJį ä¹ī +å®Į ç¾İçļĦ +羡 æħķ +åıĬ åħ¶ä»ĸ +åı¯ ç͍ +æĭ IJ +è¾ĥ 大çļĦ +æĬĢæľ¯ åĴĮ +å°¼ äºļ +çϾ è´§ +æı ī +éĢī è´Ń +éĺŁ åıĭ +ä¼ł æĦŁ +ä¼łæĦŁ åύ +åıªè¦ģ ä½ł +为ä»Ģä¹Ī è¦ģ +ä¸ĵ注 äºİ +ä½Ļ é¢Ŀ +åħ¸åŀĭ çļĦ +缮åīį å·² +欲 æľĽ +èģĶ ç»ľ +æµģ ä¼ł +çļĦ å®¶åºŃ +åı· åı¬ +çıį è´µ +ä¼Ł 大çļĦ +éī´ äºİ +è·Ł ä»ĸ +产 çī© +ä¸į å·² +è¿Ŀæ³ķ è¡Į为 +头 ä¸Ĭ +åĪĨ è§£ +åı¯ä»¥ çľĭåĩº +æł¡ åĮº +åŃĹ ä½ĵ +ä¿® çĤ¼ +çĶļèĩ³ æĺ¯ +微信 åħ¬ä¼Ĺ +åıĸ 代 +èIJ¥ä¸ļ æĶ¶åħ¥ +æ½į åĿĬ +ä½ł èĥ½ +社ä¼ļ ä¿Ŀéļľ +æ¯ĶèµĽ ä¸Ń +污水 å¤ĦçIJĨ +夫 å¦ĩ +ä¸Ģ å¹ħ +沿 æµ· +åı£ æĦŁ +ä½Ĩ åį´ +å½ĵ æĹ¥ +çļĦ æľĢ大 +æ¯ı ä¸Ģä½į +没 äºĭ +çī¹ åĪ¥ +å¼Ģ åѦ +è·¯ éĿ¢ +å¿ĥçIJĨ åѦ +æĶ¾ ç½® +éĩįåºĨ å¸Ĥ +ä½ł èĩªå·± +æ¶Īè´¹èĢħ çļĦ +ä¸Ģ æ³¢ +èѦ æĥķ +å᧠室 +注 å°Ħ +é£İ 鼨 +沿 çĿĢ +åijĬ 訴 +表 çݰåĩº +åĽĽ æĺ¯ +åı¤ åħ¸ +æĽ´ éĩįè¦ģçļĦ +好 äºĭ +çľ¼ 泪 +æ¨ ĵ +审 åΤ +碰 æĴŀ +车 ç«Ļ +è¿Ľåħ¥ äºĨ +éĽĨ åIJĪ +æł¼ å¤ĸ +宾 é¦Ĩ +æĶ¯ä»ĺ å®Ŀ +她 æĺ¯ +æĺ¯ å¦Ĥä½ķ +人 次 +çļĦ æĪIJåĬŁ +æĹł åĬĽ +æµ· æĭĶ +æĺ¥ åŃ£ +éĥ½ ä¸įä¼ļ +çŃī å¤ļç§į +ä¸Ģ个 å°ı +åģľè½¦ åľº +让 æĽ´å¤ļ +è¿Ļ çĤ¹ +æĪIJ åĵģ +éĴ ī +éģĩ è§ģ +çıŃ ä¸»ä»» +æĦı æĦ¿ +çļĦ åIJĮåѦ +游 è§Ī +åİĭ 缩 +åľ¨ ä¼łå¥ĩ +å¼¹ æĢ§ +æĹ¥ åĨħ +ç¦ı建 çľģ +è§Ĵ èIJ½ +åĪĨ å¼Ģ +ä¼ļ 让 +å¤ĸ åĽ´ +çĨŁæĤī çļĦ +çĨ Ķ +ä¸ĩ è¾Ĩ +å¤ľ éĹ´ +车 身 +ä¸Ń æľŁ +å®ĮåĸĦ çļĦ +åĵģ ç±» +åıĭ è°Ĭ +éĢīæĭ Ķ +éªij 士 +å½ ¦ +çļĦ çľĭæ³ķ +åĽ½ çİĭ +è¾£ æ¤Ĵ +åıijå¸ĥ æĹ¶éĹ´ +åı¤ åŁİ +éļı æľº +ç« ĸ +å¼Ģ è¾Ł +ä¼Ĺ çĶŁ +没 åĬŀæ³ķ +åįĥ éĩĮ +æĿ¥æºIJ äºİ +çļĦ æĿĥåĪ© +æ¯Ķ åĪĨ +满æĦı çļĦ +ä¿® è¡Į +åĿ ł +大 æµ· +èİ ¹ +åĩº 身 +è« ĩ +åħ³ èĬĤ +åIJį 人 +éľĢè¦ģ 注æĦı +æĹ© æĻ¨ +å¤ĸ åįĸ +åıĪ è¦ģ +æ¶ī æ¡Ī +çĶ³è¯· 人 +éĻĦè¿ij çļĦ +åĬłå¿« æİ¨è¿Ľ +æĸ° å¹´ +大 è¡Ĺ +ä¸Ģ é»ŀ +èĭı å®ģ +æĤĦ æĤĦ +èĦ¾ æ°Ķ +å¸Į èħĬ +éļı åį³ +æķ¢ äºİ +å®ŀè·µ ä¸Ń +æĺ¯ 没æľī +æľīè¶£ çļĦ +æĿ¥èĩª äºİ +è£ģ åΤ +女 åŃ©åŃIJ +èĩ³ åħ³ +èĩ³åħ³ éĩįè¦ģ +æĻº åĬĽ +èµ° åĩºåİ» +çŁŃ æĿ¿ +大 åĽ½ +çļĦ 认è¯Ĩ +å¹´ å¤ľ +åĨį åΰ +åIJĮ æł·çļĦ +å¯Ĩ å°ģ +å¤ĸ交 éĥ¨ +çĶŁ æķĪ +æĤ¨ åı¯ä»¥ +ä½ł åĢij +è¿ĩ å¹´ +å¼ ĵ +è¡Į æĿİ +æ¯Ķ èµ· +身 é«ĺ +è¿Ļ个 人 +ä¸Ń å¤ĸ +éģĵ æŃī +çĽ¯ çĿĢ +亲 åŃIJ +éĹ ¸ +çϽ äºij +èĦĸ åŃIJ +ä¸ĢåĪĩ éĥ½ +æ· ij +è° ľ +åģ¶ çĦ¶ +éĿł è°± +é«ĺ 管 +ä¸ĭ åıij +æĶ¾ åΰ +ç±» åĪ« +ä¸ĭ åĪĹ +æ·· ä¹± +åIJĪæ³ķ æĿĥçĽĬ +çݯ çIJĥ +æľīæķĪ åľ° +åķĨ æĪ· +æ¹ĸ 人 +æµ· 岸 +æĬķ 产 +两 个æľĪ +éĥ½ éĿŀ常 +å¢ŀ强 äºĨ +æĿ¥ åΰäºĨ +åī© ä½Ļ +æĤ¨çļĦ åŃ©åŃIJ +æµģ æ°´ +æŃ£ ä¹ī +天 çĮ« +åģļ è¿ĩ +ä½ķ æĹ¶ +æĪij åİ» +çľģ 份 +å¥ĸ éĩij +该 å¦Ĥä½ķ +ä¸ĭ çıŃ +åģ¶ åĥı +æijĨ æĶ¾ +æĸ° 模å¼ı +æĬķ è³ĩ +è·¯ åı£ +åĨľæ°ij å·¥ +大 åѸ +ä»¶ äºĭ +æł¹æľ¬ ä¸į +æµĵ 度 +æµĵ åİļ +è½® èĥİ +æĪ¿ ä¼ģ +éĿŀ常 好 +ä»İ ä¸Ń +人 æł¼ +ç¿ ģ +æĹ¶éĹ´ åĴĮ +è¿Ļ ä¸įæĺ¯ +åΏ åķĨ +æĥĬ 人 +åύ å®ĺ +åĩĨ åĪĻ +æĥħ æĻ¯ +æĽ´ é«ĺçļĦ +åѦ å®¶ +泡 沫 +åľ°æĸ¹ æĶ¿åºľ +å°± çŁ¥éģĵ +åij¼ åIJģ +ç»ı è´¸ +èĬ± éĴ± +æľī ä¸Ģ次 +æĦŁ æħ¨ +ä¸Ģ åįĥ +å¤ľ æĻļ +詹 å§Ĩ +詹å§Ĩ æĸ¯ +è¦ģ éĹ» +ç» Ĵ +æºIJ äºİ +çļĦ è´¨éĩı +注æĦı äºĭ项 +æħ¢ æĢ§ +稳å®ļ çļĦ +建设 åĴĮ +æĻ¯ 象 +éĩı åĮĸ +çļĦ 話 +è¯Ħ 级 +æº ľ +红 åĮħ +éĢļ éģİ +社ä¼ļ 责任 +æĸ° 产åĵģ +åĨ· éĿĻ +çľĭ ä¸įåΰ +èģĶ éĤ¦ +éŃ Ħ +çļĦ åīįæıIJ +çļĦåīįæıIJ ä¸ĭ +è¾ĥ 好 +çļĦ æĦŁæĥħ +客æĪ· æıIJä¾Ľ +çĭ¬ èĩª +å¢ŀ æĶ¶ +æĸĩ çĮ® +æĭ¼ åij½ +管çIJĨ åĴĮ +æµģåĬ¨ æĢ§ +åħ¨ å®¶ +ä¸Ĭ æĸ¹ +æİ¨åĩº çļĦ +ä¸ī åĽ½ +ä¸Ģ个 æĺ¯ +æĸ° ä¸Ģè½® +æĸĩåĮĸ éģĹ产 +æ® º +大 æ¹¾åĮº +éĥ½ éľĢè¦ģ +çļĦ å®ŀéĻħ +ç· Ĭ +大 å¥ĸ +åħī èĬĴ +便 äºİ +çļĦ 表æĥħ +æ¼Ķ ç»İ +红 åĨĽ +å½ĵ æĪij +æ²» æĦĪ +é¢Ŀ 度 +éĿ ľ +ä»»ä½ķ 人 +è¡Ĺ 头 +çī¹ æĸ¯ +çĸ¯ æĭī +åĮ»çĸĹ æľºæŀĦ +ç»Ļ åŃ©åŃIJ +è§Ħ 磩 +è£ ľ +çļĦ 身影 +ä¸ĵ æłı +æĿ¥ 临 +ç«¥ å¹´ +å¤į èĭı +è¨ Ĥ +åŀĭ åı· +åĽ¾ æ¡Ī +ç®Ģ åİĨ +æĭ ± +èį· åħ° +ä»» æĦı +æī¿ æİ¥ +è¿Ļ æīį +客 车 +æľĿ çĿĢ +éłħ 缮 +åı° é£İ +çļĦ æĪ¿åŃIJ +éª ı +æĿ± 西 +éģĹ ä¼ł +è¶Ĭ å¤ļ +äºĨ ä»ĸçļĦ +ä¸Ĭ åij¨ +管çIJĨ åĪ¶åº¦ +失 ä¸ļ +çĶ· åıĭ +æİ¥ ç§į +å¨ģ åIJį +çĴ° å¢ĥ +åıijçĶŁ åľ¨ +个 åĽ½å®¶ +åĪĽæĸ° åıijå±ķ +æĶ¹åıĺ äºĨ +åģ¥åº· çļĦ +å̼å¾Ĺ ä¸Ģ +å̼å¾Ĺä¸Ģ æıIJ +åĽ¢ ä¼Ļ +åģĩ 设 +åı° ä¸Ĭ +è§ĦèĮĥ åĮĸ +éĻª åIJĮ +座 æ¤ħ +åı¯ æĢľ +åħĭæĢĿ 主ä¹ī +æ³ķå¾ĭ 责任 +ä¸Ģ é¡¿ +æĬ¬ 头 +为 éĩįçĤ¹ +è¿ľ æ´ĭ +éĢı è¿ĩ +åħ¨çIJĥ åĮĸ +è¶£ åij³ +票 æĪ¿ +æ¯ı 人 +åIJĦç§į åIJĦæł· +äºĨ åĩºæĿ¥ +ç»Ŀ对 æĺ¯ +ä¸ĭ å±ŀ +ä¸Ģ åıĮ +è¿Ļ åĿĹ +æĬĹ çĸ« +è¦ģ çĤ¹ +å½¢æĪIJ çļĦ +æĪij çľĭ +ä¸ĩ éĩĮ +èĢĥ çłĶ +为 åħ¶ +æ°ij 宿 +å¤ļ ä½į +大 èĩ´ +ä»ĺ è´¹ +åħ¥ æīĭ +å±ħ å®¶ +æīĢåľ¨ åľ° +人 身 +è¿ĩ å¾Ĺ +è¯ķ è¯ķ +访 è°Ī +åĬł éĩį +å°± ä¸įä¼ļ +çĶŁäº§ ä¼ģä¸ļ +åĽŀ åĽ½ +åºķ 线 +èµ¶ åΰ +æĶ¯ éĺŁ +æĪij们 éĥ½ +éĤ® æĶ¿ +缴 èĩ³ +éĴ¢ çIJ´ +åħ ľ +çłĶ讨 ä¼ļ +æľĪ 亮 +åĿļæĮģ 以 +åħ¬å®ī éĥ¨ +éĴ¢ 管 +å°ı çϽ +ç½® ä¸ļ +èģ ĭ +书 åĨĻ +æĿ ı +éħį æĸ¹ +èĢĮ åıĪ +çijŀ 士 +çķĮ çļĦ +èĢģ 大 +æĪIJçĨŁ çļĦ +å¹² ä»Ģä¹Ī +ä¸ĵ项 æĸĹäºī +çŃī å¤ļ个 +èĦ± 离 +ä¸ī 个æľĪ +çłĶç©¶ åijĺ +æĹĭ 转 +æŀģ èĩ´ +åħį è´£ +åħįè´£ 声æĺİ +å¾Īå¤ļ çݩ家 +车 ä¸Ĭ +交 äºĴ +å·² æĺ¯ +ä¸Ģ å°ı +çļĦ éĩįçĤ¹ +èĬ± äºĨ +ä¸į æĺİ +æľīåħ³ è§Ħå®ļ +çĬ¹ å¦Ĥ +çľ ¸ +å¯ ¡ +çļĦ è¡£æľį +åĮħ 裹 +身 åŃIJ +å¸ĪèĮĥ 大åѦ +äºĭ åħĪ +线 æĿ¡ +æ³ķ åζ +åħ» æĬ¤ +稳å®ļ æĢ§ +éĤ µ +åŀĦ æĸŃ +é¡ į +èĢĥ åı¤ +æĿł æĿĨ +èĭı èģĶ +æ°´ ç͵ +åħ·ä½ĵ çļĦ +æ¿Ģ æ´» +æĪij æł¡ +åĪļ å¼Ģå§ĭ +åĩ¸ æĺ¾ +ç¦ ¾ +åħ¼ èģĮ +éĢı éģİ +åľ¨ 游æĪıä¸Ń +社ä¼ļ åıijå±ķ +好 çİ© +å¹» æĥ³ +ä¸į 代表 +注æĦı åĬĽ +æ£ į +ç͍ æīĭ +ç¾İ 人 +许å¤ļ 人 +å¾Ī æĺ¯ +çļĦ çłĶåıij +æīĵ åĩº +åIJĪä¼Ļ 人 +ä¸Ģ å¤ľ +ç¼ĵ ç¼ĵ +ä¿® æŃ£ +æĦŁ çŁ¥ +ç»Ī 身 +æ¿Ģ ç´ł +çݯå¢ĥ ä¸ĭ +次 ä¼ļè®® +ç»ıæµİ å¢ŀéķ¿ +æī Ľ +åıij éħµ +åĪĨæŀIJ å¸Ī +åľ¨ æľªæĿ¥ +主è¦ģ æľī +ä¸Ģ åŃ£åº¦ +çļĦ 说æ³ķ +ä»İæĿ¥ 没æľī +è´§ 车 +缩 å°ı +太 è¿ĩ +æķĪ åĬĽ +ä¸į ä¸ĭ +æĬķ 稿 +èᝠä¸ļ +ç»Ħ éķ¿ +ç«Ļ çĤ¹ +å¾Ī åĸľæ¬¢ +éIJ µ +åĬ¿ 头 +æ¼ı æ´ŀ +æĦ¤ æĢĴ +åħħ å®ŀ +åĪĽä¸ļ æĿ¿ +çĪ ª +æľª å¿ħ +åºķ éĥ¨ +å¾Ĺ åĪĨ +人æ°ij åĮ»éĻ¢ +äºĮæīĭ æĪ¿ +å·²ç»ı 被 +大 楼 +æĸ° æĪ¿ +辦 æ³ķ +ç͍ åĬĽ +æĭĵ 宽 +åĨħ åľ¨ +æĴŃ åĩº +饰 æ¼Ķ +ä¹Ł 让 +ä½ľ çĤº +çī©ä¸ļ 管çIJĨ +åį´ ä¸į +为 ä¸ŃåĽ½ +å±Ģ åĬ¿ +ä¸į èĤ¯ +æľĢ æĸ°çļĦ +åı¯ä»¥ éĢīæĭ© +æĺ¾ çݰ +å°± ç®Ĺæĺ¯ +åľ¨ æł¡ +é¾ Ł +两 æĿ¡ +çļĦ å®ŀåĬĽ +è¶Ĭ 好 +她 åľ¨ +å¿ł è¯ļ +ä¹Ł éľĢè¦ģ +游æĪı æĵįä½ľ +è¶ħ åĩº +å¦Ĥæŀľ ä¸į +æīĢåľ¨ çļĦ +ä½ł è¿ĺ +以 åĨħ +æľī ä¸Ģå®ļ +åı¯ è¾¾ +è·ij åΰ +åī Ľ +建ç«ĭ åģ¥åħ¨ +æķ´ 车 +åīį æĸ¹ +éĹ´ æİ¥ +çѹ å¤ĩ +çĸ² åĬ³ +离 å¼ĢäºĨ +æ± Ŀ +éĿ¢ éĥ¨ +ä¹ĭåīį çļĦ +åıĺ 为 +å¦Ĥæŀľ 说 +对 ä»ĺ +åĿĩ åı¯ +被åijĬ 人 +ç²¾ ç¾İ +èģļ ä¼ļ +çĿĢ æĢ¥ +è°· æŃĮ +ä¸Ģ åı· +红 åĪ© +ä¼łå¥ĩ 游æĪı +å» ĸ +è´ ŀ +ä¹° åΰ +éŃ ļ +ä½ĵ è´¨ +å°ij äºĨ +æ³ī å·ŀ +åIJ Ł +ç»Ŀ ä¸į +é»ij æģ¶ +é»ijæģ¶ åĬ¿åĬĽ +ä¸Ĭ æĺł +çļĦè¯Ŀ é¢ĺ +ä¸ĩ人 次 +ä¸ĸ éĹ´ +ç͍ å·¥ +è´¯ ç©¿ +å®Ŀ çŁ³ +ä½ł 好 +åĪĩ åī² +强 åĽ½ +åĽŀ èIJ½ +æ°´ æĻ¶ +模 仿 +æ´ª æ°´ +éĢĻ éº¼ +åįģä¸ī äºĶ +ä½ ij +éĻ Ħä»¶ +çļĦ å¢ŀéķ¿ +éĻĦ å±ŀ +çݰ å·² +帮 ä½ł +éĩij çīĮ +é«ĺ åİŁ +åľ¨ å®¶éĩĮ +éĺ² èħIJ +ç¡®å®ŀ æĺ¯ +宣 讲 +天 æīį +ç»ıèIJ¥ 管çIJĨ +éĶħ çĤī +åIJĪ ä¸Ģ +è§Ĥ èµı +éķ¿ è¾¾ +主ä¹ī æĢĿæĥ³ +éĤ£ 麼 +é£İ äºij +为主 çļĦ +æļij åģĩ +æĮģ ä¹ħ +å¼Ĥ åľ° +å¼Ģ éŨ +模 æĿ¿ +æī¹ 次 +ä¸į 便 +天 çĶŁ +åĩł 个æľĪ +ä¸ĵ ç§ij +åı¦ æľī +åħ¬å¸ĥ çļĦ +æĩ · +åľº åIJĪ +çļĦå¿ĥ æĢģ +è¿ĺ 好 +å®ŀ æĪĺ +èĢģå¸Ī çļĦ +åħ© åĢĭ +åı¯ åľ¨ +éĤ£ ä½į +å¥ł å®ļäºĨ +ä¿ĥ éĶĢ +æı´ åĬ© +ä¸ĩ çī© +æĥħ æĬ¥ +é¦ĸåħĪ è¦ģ +æĸĩåĮĸ åĴĮ +éĥ½ å·²ç»ı +ä¸Ĭ ä¸ĸ纪 +åĨľ åľº +大 æī¹ +æĺİçϽ äºĨ +çļĦ æĪIJéķ¿ +çļĦ æ¯ĶèµĽ +失 误 +åģļ æĪIJ +ä»Ĭ天 å°ıç¼ĸ +é¢Ĩ è¢ĸ +æıIJåįĩ äºĨ +å¾IJ å·ŀ +ä»į æľī +è¿ĩ 滤 +å¹½ é»ĺ +çĥŃ éĩı +ä¸Ģ é¦ĸ +æ¼Ĥ亮 çļĦ +åĩł ç§į +åĢ¡ è®® +å°±åı¯ä»¥ äºĨ +æİĴ åĪĹ +éĩį éĩį +ä¼ģä¸ļ åĴĮ +ä¸ĵ å±ŀ +çħ İ +亲 æĪļ +çϾåĪĨ ä¹ĭ +稿 ä»¶ +è¿ĺ å¾Ĺ +人 åĵ¡ +äºī 夺 +æĽ´ 容æĺĵ +大 èĩªçĦ¶ +鼻 èħ¦ +太 空 +åľ° å¤Ħ +å¤ ¢ +ä»ĸ 对 +å¿ħ å°Ĩ +ä¸į å½ĵ +严 è°¨ +åĩº åľº +å·²ç»ı æľī +é¢Ĩ åĨĽ +é«ĺ æ¡£ +ä¸Ģ æīĢ +æł Ĺ +让 åѦçĶŁ +æĽ¹ æĵį +æŁIJ ä¸Ģ +伸 åĩº +èĬ± åįī +æ¸ħ éĨĴ +èģĶç³» æĸ¹å¼ı +åĪĨ å±Ģ +èħ ³ +æ©¡ èĥ¶ +éķ¿ å¾Ĺ +绿 åľ° +è¢ į +çļĦ èīºæľ¯ +女 æľĭåıĭ +ä¸Ń è¶ħ +离 åŃIJ +å¤ļæł· åĮĸ +éĺ³ åı° +ä½İ 碳 +ä¸Ģ ç±» +çŃīæĸ¹éĿ¢ çļĦ +å¾Ĺ 好 +模 åħ· +ä¸ĩ 亿 +çķĻ æĦı +临 æ²Ĥ +å°ij éĩı +çľĭ åIJij +ç»ıèIJ¥ èĢħ +çķĻä¸ĭ äºĨ +åĿı äºĨ +åijĬ åĪ« +羣 çIJĨ +ç¼´ è´¹ +æĬĬ ä½ł +çļĦ ä»»åĬ¡ +æĪij 对 +ä¹° åħ¥ +çĻ» ä¸Ĭ +æľī 两个 +ä¸Ģ 头 +æĵį æİ§ +åħ¨ è¦ĨçĽĸ +çĿĢ æīĭ +å¢Ļ éĿ¢ +å¤ļ æĸ¹ +åı¯çα çļĦ +ä¹Ł åı¯èĥ½ +æľĢ æľī +è¿ĻäºĽ éĥ½æĺ¯ +æĥ ¡ +å® ® +å¾Ī å°ı +éĹ®é¢ĺ æĺ¯ +åĿĩ æľī +å¾ģ éĽĨ +说 åĩº +æľī æĦı +é¢ Ĥ +æī¬ å·ŀ +åķĨä¸ļ 模å¼ı +çĶŁ èĤĸ +æįIJ 款 +å² Ĥ +ç¾İ æĻ¯ +è¿ĺ 羣 +æĭ¥ æĬ± +身ä½ĵ åģ¥åº· +æ·± å¤Ħ +çľ¼ ç¥ŀ +çļĦ 形象 +ä¼ĺ è¶Ĭ +å½ĵ æĪIJ +åĮº åĪĨ +åİ» éϤ +注 å®ļ +å§IJ 妹 +åĮº åĨħ +é© ļ +æļĹ ç¤º +æĺİ äº® +æħ° éĹ® +å¸Ĥåľº 份é¢Ŀ +çĮª èĤī +çļĦ èµĦéĩij +åİĨ ç»ı +å§ĭç»Ī åĿļæĮģ +çĶŁ æľº +ä¸į 顾 +éĩij åĪļ +大 声 +éĻķ 西çľģ +é² į +åĨľä¸ļ åĨľæĿij +æľī 害 +éŨ è¯Ĭ +æ¯ı ä¸Ģ次 +çļĦ åĽłç´ł +é¢Ŀ å¤ĸ +åİ¿ 级 +çļĩ åIJİ +åĽ½ ä¼ģ +é¦ĸ éĢī +ç¼ĸ åĨĻ +æĭ¿ èµ· +åģ· åģ· +ä¸İ ä¸ŃåĽ½ +åįĸ å®¶ +ç»Ļ ä»ĸ们 +ç¥ŀ è¯Ŀ +åѸ æł¡ +æĪij ä¸Ģ缴 +çŁ¥éģĵ äºĨ +åį Ĵ +åĴĮ åľ°åĮº +ä»Ģä¹Ī éĥ½ +çĶ» å®¶ +æľ¬ çĿĢ +ä½Ļ åIJį +审 çIJĨ +ä¸Ģ åIJij +åıijå±ķ è¶ĭåĬ¿ +åĮº éĹ´ +注åĨĮ èµĦæľ¬ +çIJ ¦ +ä¸į åı¯ä»¥ +çļĦ åĦ¿åŃIJ +å̼ çıŃ +ä¸¥æł¼ çļĦ +å®ŀä½ĵ ç»ıæµİ +æľī æĿĥ +æĪij åıĪ +éĵ¶ æ²³ +ç«ĭ 马 +æĿĢ äºĨ +åĮħ 容 +管 å®¶ +身 é«Ķ +éĵ ħ +å°ı åŃIJ +管çIJĨ ç³»ç»Ł +æľīçļĦ 人 +é£İ ç͵ +æĻºèĥ½ åζéĢł +ç²¾ ç¡® +æĭĽåķĨ å¼ķ +æĭĽåķĨå¼ķ èµĦ +äºĮæīĭ 车 +åİ¿ å§Ķ +èīº äºº +å¥ ķ +è¿İ æĿ¥äºĨ +ç»ĵæĿŁ äºĨ +çļĦ ä¼łç»Ł +æĭ¼ æIJı +奥 迪 +çĸij æĥij +ä¹ĭ æĹ¥èµ· +æłĩå¿Ĺ çĿĢ +åľ° åįĢ +è¯ł éĩĬ +åΰ æľŁ +åħ¨ éĥ½ +çŁŃ æļĤ +æĺ¯ æĪijåĽ½ +æĪij å·²ç»ı +æ»´ æ»´ +天 èµĭ +对 她 +åį«çĶŁ éĹ´ +çĶŁäº§ åŁºåľ° +æĹ¥ è®° +çļĦ æķĻåѦ +åĵ ĩ +æ°ij äºĭ +è¿ĺ åİŁ +æīĭ ä¸ŃçļĦ +çļĦ èī¯å¥½ +æ· « +ä¸Ńåħ± ä¸Ń央 +åĪ ĥ +åĵ Ħ +åľ¨ ä»ĸçļĦ +å°Ī æ¥Ń +åľº éĿ¢ +éĤ» å±ħ +çĹ Ĵ +å¦ Ħ +å¤ĸ ç§ij +ä¸į éĢĤ +举åĬŀ çļĦ +é Ĥ¹ +åħļçļĦ 建设 +çϼ 表 +è·¨ çķĮ +æ²ī æ·Ģ +大 çīĩ +è¶Ĭ é«ĺ +å°Ĩ æĺ¯ +è§ī éĨĴ +åĤ¨ åŃĺ +å¢ŀ 大 +ä¸į 让 +æķ´ å½¢ +å¹³åı° ä¸Ĭ +åĩł ä½į +è¯ī æ±Ĥ +好 ä¸į好 +åľ į +æĸĩ æľ¬ +é̲ åħ¥ +ç´ į +æł¹ æĵļ +èįī æ¡Ī +åħŃ ä¸ª +åĭ ¿ +åζ æĪIJ +饮 æ°´ +æ°¸ æģĴ +èĩª æĿĢ +åı¸ 马 +éļ¾ çĤ¹ +为 æĪij们 +å¼ § +åī© ä¸ĭçļĦ +åĩĨå¤ĩ 好 +çļĦ æľĢä½³ +èģĶåIJĪ ä¼ļ +æĤ£èĢħ çļĦ +æĪijä¸į çŁ¥éģĵ +ä¸ĭ ä¸Ģ个 +åıijå±ķ æĸ¹åIJij +ç¬ ¨ +æīĢ以 æĪij们 +åĨĻ äºĨ +éĢł æĪIJäºĨ +æ²Ļ æ¼ł +çŃĽ éĢī +çģ¾ åĮº +ä¸Ĭ çľĭ +éħ ¶ +æ»ļ åĬ¨ +éļ¾ åħį +åIJī åĪ© +ä¸Ģ ä¸Ģ +ç²¾ å¯Ĩ +伸 æīĭ +礼 仪 +åħ¨ æĺ¯ +è¶Ĭ 大 +ä¸Ń æłĩ +åıĸ åĨ³ +åıĸåĨ³ äºİ +éĢĶ ä¸Ń +讨 åİĮ +æīĭ åĨĮ +第 ä¹Ŀ +åŃĶ åŃIJ +çĦ¶ å¾Į +ä¸Ģ åħ± +æµ· æĬ¥ +款 å¼ı +æķ´ 天 +è¾¹ çķĮ +è·¯ è¾¹ +æĻĭ 级 +åIJIJ æ§½ +çļĦ åħ³æ³¨ +æĪij 没æľī +å°±æĺ¯ åľ¨ +缮 çļĦæĺ¯ +åį³ä½¿ æĺ¯ +é¡¶ å°ĸ +å·²ç»ı åľ¨ +å®īåħ¨ éļIJæĤ£ +æłĩ æĿĨ +åįĹ éĢļ +ä¼ļ 对 +座 ä½į +èµ¢å¾Ĺ äºĨ +åİŁæĿ¥ çļĦ +身 为 +书 åºĹ +è¢Ń åĩ» +ä»Ĭ æĻļ +以 èī² +以èī² åĪĹ +æĬĸ éŁ³ +åį´ æ²¡æľī +丧 失 +çļĦ å±ĢéĿ¢ +åįģåĽĽ äºĶ +çŃī 缸åħ³ +æ±ĩ æĢ» +å¤ĸ 表 +为 æ°ij +éľĩ æĥĬ +å¥Ĺ è·¯ +çĬ¯ç½ª å«Įçĸij +å°Ĩ 以 +çİĩ é¢Ĩ +éħĴ åIJ§ +è¡Įä¸ļ åıijå±ķ +å¹´ èĩ³ +åύ æĿIJ +åĴĮ æĬĢæľ¯ +æľĢ å°ı +è¿Ļä¸Ģ åĪĩ +èģĮ ç§° +å½ĵ ä½ľ +æİĢ èµ· +åĴ ĭ +ä¸Ń éĥ¨ +æīĭ èĩĤ +ç½¢ äºĨ +媳 å¦ĩ +æ´½ è°Ī +æĹ¶ä»£ ä¸ŃåĽ½ +人çĶŁ çļĦ +æŀģ éĻIJ +ç¦ Ħ +åĮº æĶ¿åºľ +æľ¬ éĴ± +礼 åĵģ +çļĦ éĤ£ä¸ª +侦 æŁ¥ +太å¤ļ çļĦ +å®ŀæĸ½ æĸ¹æ¡Ī +é«ĺ æłĩåĩĨ +æĮĩæĮ¥ éĥ¨ +å̾ æĸľ +çī¹èī² ç¤¾ä¼ļ +çµIJ æŀľ +éĴ» çŁ³ +ç§» æ¤į +çī¹ ç§į +èĩª æĦ¿ +æĭľ çĻ» +åįķ 身 +åį´ åıĪ +åĪ¥ 人 +åIJĪ è§Ħ +æľº ç͵ +çī¹ æĦı +å½ĵåīį ä½įç½® +ä¹° å®¶ +åIJĪ çº¦ +èĤ© èĨĢ +为 åĩĨ +å®¶ è£ħ +çļĦ çĥŃæĥħ +éĿŀ éģĹ +çļĦ éŃħåĬĽ +åİŁ åijĬ +社ä¼ļ åIJĦçķĮ +ä¹° çļĦ +å¤ļ åIJĥ +éĽķ å¡ij +èµ· ä¹ī +åĬł åī§ +éĤ£ä¸Ģ åĪ» +å°Ĩ è¿Ľä¸ĢæŃ¥ +æ¡Ĥ æŀĹ +æĽ´ 强 +对 ä¼ģä¸ļ +æĹł æĦı +ä¹łè¿ijå¹³ æĸ° +æµģ 失 +å¾® 软 +缸 对äºİ +座è°Ī ä¼ļ +主 èIJ¥ä¸ļ +主èIJ¥ä¸ļ åĬ¡ +ç§ģ åĭŁ +å±ķ示 äºĨ +常æĢģ åĮĸ +è² ´ +符 åı· +å¹´è½» çļĦ +å°± éľĢè¦ģ +ä¹Ł æĽ¾ +çļĦæĥħ 绪 +è¾¾ æłĩ +èĩ ¨ +ä½į å±ħ +ä»ħ 为 +é¦ĸ å®¶ +éĺ´ éĺ³ +ä¸įåĨį æĺ¯ +åĽłä¸º å®ĥ +ä¼ģä¸ļ åľ¨ +çĺ ¾ +åIJ¬ è§ģ +åİŁ æľī +åζ è£ģ +å¯Ĥ å¯ŀ +éĢļè¿ĩ 对 +æ»ij éĽª +è¿Ļ å¼ł +çļĦ çIJĨè§£ +æĸ° ä¸ŃåĽ½ +è¿Ļ åĦ¿ +ä½İ ä»· +æĥ³ è¿ĩ +çļĦ ä¿¡å¿ĥ +建çŃij çī© +çļĦ é¢ľèī² +ä¸į åºĶ该 +æĹłçĸij æĺ¯ +å¼ķèµ· äºĨ +åħ¨ åijĺ +æĿ° åĩº +è¿Ļæĺ¯ æĪij +èª ° +èĺ ĩ +éĺµ åľ° +åħħ å̼ +çŁ¿ ä¸ļ +çĿĢ ä»ĸ +ä¿¡ 访 +ä¸ĩ è¾¾ +æij© æĵ¦ +å¼Ģ 端 +èı² å¾ĭ +èı²å¾ĭ 宾 +车 åŃIJ +æľ¬èº« çļĦ +çģ«è½¦ ç«Ļ +常 å·ŀ +为 代表 +为代表 çļĦ +广 ç͵ +亲 人 +åı³ æīĭ +éĽĨ è£ħ +éĽĨè£ħ ç®± +çļĦ åį°è±¡ +æ©Ł æľĥ +åĮĨ åĮĨ +åħī ç͵ +大 æĸ¹ +è¿ĺ æľª +åĪ© 好 +ç»Ŀ 大å¤ļæķ° +åľ¨ è¿Ļç§į +ä¸Ģ ç»Ħ +æĸ° èĤ¡ +转 åıij +æ³ķ åºŃ +æĹł æīĢ +éģĵ è·¯ä¸Ĭ +çŁ¿ å±± +èij ī +æĶ¶ åĽŀ +ç§° ä¹ĭ +ç§°ä¹ĭ 为 +æıŃ éľ² +åı£ 岸 +åIJ ¼ +å¿ĥ æĥ³ +çļĦ 梦æĥ³ +éĽ ¯ +ä¹ĭ åĪĿ +å¥ĸ 项 +订 éĺħ +èĵĿ 天 +åĿ¦ åħĭ +ç«ĭ æ¡Ī +èģĶ æīĭ +ä½Ĩæĺ¯ æĪij +帮 æĪij +ä»ħ 代表 +说 æĪij +çļĦ è¶ĭåĬ¿ +æ¯Ķè¾ĥ 大 +èµ° å»Ĭ +éĩįçĤ¹ é¡¹çĽ® +èµĮ åľº +åIJį çīĩ +æĦŁ åı¹ +åľ¨ åľ°ä¸Ĭ +åıij çĥŃ +èĮĥ çķ´ +çļĦ éģĵè·¯ +éĩij èī² +ä»ĸ åıĪ +ä¼ļ 产çĶŁ +æ°ij åĽ½ +å®ĺæĸ¹ ç½ijç«Ļ +æĶ¶çĽĬ çİĩ +çļĦ åΰæĿ¥ +çļĦ åĬŀæ³ķ +æĶ¹ åζ +ä¸ĩ ç§ij +ä¸į äºĪ +è¿ĻäºĽ éĹ®é¢ĺ +çα ä¸Ĭ +çIJĥ åľº +è´£ 令 +æİĪ è¯¾ +åľ¨ é¦Ļ港 +ç»Ĩ èħ» +å¤ļ ä¸ĩ +åIJĮ å¹´ +大 使 +æĸ ĭ +ä¹Ł 为 +æĥł å·ŀ +åIJī 祥 +çͰ åĽŃ +åĽ½å®¶ éĺŁ +éĩį çĶŁ +åľ¨ åħ¶ +é¦Ļ åij³ +è´Ł èį· +亲 åĪĩ +èĩª 豪 +没 éĶĻ +åĽłä¸º åľ¨ +æĺŁ æĺŁ +éĤ ij +è¿ĺæľī å¾Īå¤ļ +æij© æīĺ +æij©æīĺ 车 +æŃ¥ è¡Į +管çIJĨ ä½ĵç³» +èĦļ ä¸ĭ +éģİ åİ» +æ±ī è¯Ń +对 ä¸įèµ· +çļĦ ç»ıåİĨ +åıĬ 缸åħ³ +ä¸įå°ij 人 +éĩį ç£ħ +åĬ³åĬ¨ èĢħ +大åĬĽ åıijå±ķ +æĢİä¹Ī åģļ +çĭĹ çĭĹ +举åįĹ äºļ +åĭĩ äºİ +åħ¬ éĸĭ +çĵ· çłĸ +åıĤ çħ§ +广æĴŃ ç͵è§Ĩ +举 åĬ¨ +æ±Ł 西çľģ +æķĪ èĥ½ +å͝ æľī +éĿ¢ è²Į +èĩªåĬ¨ 驾驶 +æ¦ľ åįķ +å½ĵ æĪij们 +仲 è£ģ +æľ¨ æĿIJ +ç±³ åħ° +çϽ éĵ¶ +çļĦ 人éĥ½ +å°± åĥıæĺ¯ +æŃ¥ åħ¥ +åįł ç͍ +åĩ» è´¥ +让 大家 +ä¼ļ è®©ä½ł +åİ¿ æĶ¿åºľ +è¦ģ ç͍ +çŃī å½¢å¼ı +åįĩ é«ĺ +责任 æĦŁ +å¤ĩ ç͍ +ä»ĸ 认为 +æ¸ħåįİ å¤§åѦ +ä»ĸ èĩªå·± +éĸ± è®Ģ +太平 æ´ĭ +éĶģ å®ļ +çŃ Ĩ +è¿Ļ çīĩ +æī§ æĶ¿ +è¿ĶåĽŀ æIJľçĭIJ +å°± æŃ¤ +éģĩ åΰäºĨ +å¼Ģå¹ķ å¼ı +管çIJĨ éĥ¨éŨ +å§¿ åĬ¿ +设 æĥ³ +åĽĽ åŃ£ +æĬĢæľ¯ 人åijĺ +å·® çĤ¹ +è¾ŀ èģĮ +èĢģ 師 +çļĦ æĦŁåıĹ +ä¹Ł éĿŀ常 +å¹´ ä¸ĬåįĬå¹´ +æĢª çī© +èĮĥ æĸĩ +æĪĺ å½¹ +åIJ« ä¹ī +åħ¨ è¿ĩç¨ĭ +èĢĮ éĿŀ +éĢļ讯 åijĺ +è¿Ļæł· æīįèĥ½ +æľº ç»Ħ +è£ ı +çķ¶ çĦ¶ +èµĮ åįļ +åIJĦ æľī +å·¥ä½ľ æľºåζ +äºĭ åIJİ +åī§ éĻ¢ +å±Ĭ æĹ¶ +åĺ´ éĩĮ +主 线 +ä¸Ģ åľĪ +主è¦ģ åİŁåĽł +å°¸ ä½ĵ +åĮ»çĸĹ åĻ¨æ¢° +ä½ł æĢİä¹Ī +ä½Ĩ çͱäºİ +æĹ¶ 空 +çĶ· æľĭåıĭ +çĶľ èľľ +é«ĺ åľ° +æĻ ĸ +èĴIJ éĽĨ +åĩĿèģļ åĬĽ +å¤ĩ åıĹ +æĸĩ åĪĽ +马 æĿ¥ +马æĿ¥ 西äºļ +æŁ´ æ²¹ +使 人 +æķĻ ä¼ļ +ç§ĭ 天 +æĺİ çıł +åħŃ åįģ +çݯå¢ĥ ä¸Ń +æ¸ħ æĻ¨ +积æŀģ åıĤä¸İ +å·ħ å³° +为 æľŁ +çѾ åŃĹ +æĦŁ æ¿Ģ +ç§ĭ åŃ£ +æĿij åŃIJ +æ¢ħ 西 +æļ´ 鼨 +çĶŁæ´» åľ¨ +çªĹ æĪ· +æģ¶ åĬ£ +纯 ç²¹ +åľ¨ æİ¥åıĹ +没 èĥ½ +è¡Į 人 +åĭ º +æĭ¨ æīĵ +ä½ľ åĩºäºĨ +çļĦ 主é¢ĺ +æľª ä¾Ĩ +ä¸Ń æľĢ +æ¾ ľ +é«ĺ è¡Ģåİĭ +åħ´ èµ· +æŃ£ èĥ½éĩı +åŁ¹è®Ń çıŃ +æİ¥ åħ¥ +çĦ¶åIJİ åĨį +åѦçĶŁ 们 +é¢ĨåħĪ çļĦ +çģ« çĥŃ +ä¸ĵ èģĮ +æĪĸèĢħ 说 +建 è¨Ń +é» ı +对 åħ¬åı¸ +çī¹ æľīçļĦ +åħī èᣠ+å½ĵ åľº +éĿ¢ åŃIJ +èµĦ产 管çIJĨ +æĹ¶æľŁ çļĦ +çŀ İ +åįİ ä¸ľ +åıĪ ä¸Ģ次 +èĥİ åĦ¿ +å®ļ çĤ¹ +头 çĹĽ +æ¶² ä½ĵ +æĺ¯ä¸Ģ ä½į +帽 åŃIJ +å¹´ èµ· +ä¸į ä½İäºİ +è¾ĥ å°ij +éĿ¢ä¸´ çĿĢ +å±Ĥ å±Ĥ +èĿ´ èĿ¶ +èī° èĭ¦ +éĺ¿ æł¹ +éĺ¿æł¹ å»· +æ¦Ĥ æĭ¬ +请 éĹ® +èµ· åºĬ +å±Ģ å±Ģéķ¿ +稳 åģ¥ +å¦Ĥæŀľ æĪij们 +éħĴ ç²¾ +æĪ· åı£ +æĦŁ æĤŁ +æĪij们 éľĢè¦ģ +æĬĢ èīº +èĩª åªĴä½ĵ +è¿Ľ åĮĸ +æ¿ĢçĥĪ çļĦ +ä½ĵ 温 +èļ ķ +èĩ´ è¾ŀ +宪 æ³ķ +ä¸Ģ çŃīå¥ĸ +çĵ¶ é¢Ī +æĥł æ°ij +èµ° è·¯ +çݰ ä»» +åķĨ éĩı +ä¸ĭ 车 +åĪ ł +責 ä»» +èŀįåIJĪ åıijå±ķ +ç´ł æĿIJ +æ²¹ ä»· +åģļ 人 +çŀ ª +æĶ¹éĿ© åĪĽæĸ° +çļĦ åĮºåĪ« +è·¨å¢ĥ ç͵åķĨ +æ¶īåıĬ åΰ +æīĺ 管 +æĪij è¿ĺæĺ¯ +åĿIJ æłĩ +ç½ij 讯 +å½ĵåľ° çļĦ +追 溯 +åľŁ è̳ +åľŁè̳ åħ¶ +åºķ ä¸ĭ +åĩł åįģå¹´ +ç©¿ è¿ĩ +çĶŁæĢģ æĸĩæĺİ +æİ¨ èĸ +æİ¨èĸ ¦ +éł Ĩ +åĴ³ åĹ½ +åĪĨ æĪIJ +çĹķ 迹 +æĪ· ç±į +éĥ½ ä¸įèĥ½ +æĻļ ä¼ļ +åĢ © +ä½ĵ åĬĽ +è¿Ļ个 èģĮä¸ļ +æĹł å½¢ +åıª æĥ³ +è¿Ľ åıĸ +æĿĢ æŃ» +èĦ Ĭ +äºij åįĹçľģ +æľª çŁ¥ +ç¾İ èģĶ +ç¾İèģĶ åĤ¨ +å¤ĸ å½¢ +诱 æĥij +çĽ £ +è¡Į 使 +åłĨ 积 +çĨŁ ç»ĥ +éĺIJ è¿° +æľĢ大 éĻIJ度 +å·¡ æŁ¥ +夺 åĨł +ä¼ģä¸ļ æĸĩåĮĸ +çĭ® åŃIJ +ä¿Ŀ å®Ī +ä¸ºæł¸å¿ĥ çļĦ +æī© æķ£ +åζéĢł åķĨ +æŁĶ 软 +为ä¸Ģä½ĵ çļĦ +游 çİ© +çĶŁ çĹħ +幫 åĬ© +åͱ æŃĮ +æīį åı¯ä»¥ +宽 æĿ¾ +è¦ģ æ¯Ķ +æĺ¯ æĢİæł· +çģ° èī² +çİĭ åĽ½ +æIJħ æĭĮ +计 éĩı +åij¨åĽ´ çļĦ +æĻºèĥ½ æīĭæľº +常 åĬ¡ +常åĬ¡ åī¯ +é© ´ +å°Ĩ è¿ij +寻 常 +ä¸ŃåĽ½ å¸Ĥåľº +容 åύ +å±± ä¸Ĭ +èĥĮåIJİ çļĦ +亲 å¯Ĩ +æīĢ以 说 +éİ ® +çļĦ çIJĨçͱ +大 åŁİå¸Ĥ +常 å¹´ +æĹħ游 ä¸ļ +å°±æĺ¯ è¿Ļæł· +åĨį æĿ¥ +é«ĺ ä½į +åĨħ 饰 +æŀĦ éĢł +ä¸Ģ èµ·æĿ¥ +çͳ è«ĭ +å·²ç»ı å¼Ģå§ĭ +çļĦ åĬ¨ä½ľ +被 è¿« +éģį å¸ĥ +åīĸ æŀIJ +å°ı äºĭ +å¿ĥ ä¸ŃçļĦ +ä½ĵåζ æĶ¹éĿ© +çļĩ å®¶ +æķĻ åłĤ +åIJĥ å®Į +åĽ½æ°ij åħļ +æĺİç¡® äºĨ +åıijå±ķ è§ĦåĪĴ +第ä¸Ģ æŃ¥ +å¾Ĺ èµ· +åľ¨ åĵª +çļĦ è·¯ä¸Ĭ +é» Ķ +çķ¶ æĻĤ +大åĬĽ æĶ¯æĮģ +åıĮ éĩį +çŁ¥éģĵ èĩªå·± +åIJĪä½ľ åįıè®® +æ°Ķ åĬ¿ +éķ¿æķĪ æľºåζ +ç½ķ è§ģ +åĽŀ æĿ¥äºĨ +ä»ĸ ä¼ļ +ä¸Ń æĸ° +ä¸Ńæĸ° ç½ij +çļĦ åķĨåĵģ +èµł éĢģ +決 å®ļ +å¸Ĥåľº çĽij管 +çķĻ åѦçĶŁ +ç͵ åİĭ +äºļ 马 +äºļ马 éĢĬ +è¿ĺæĺ¯ æ¯Ķè¾ĥ +ä¿ĥè¿Ľ äºĨ +æµģ åħ¥ +æijĦ åĥı +æijĦåĥı 头 +æıIJ åıĬ +åıij æİĺ +æī¾ åĩº +æ¢Ŀ ä»¶ +ç¹¼ çºĮ +æĪij åĸľæ¬¢ +å¥ İ +æ¦ľ æł· +å¼Ģ èĬ± +æ²ī éĩį +åŁº åĩĨ +ä»ħä»ħ æĺ¯ +轨éģĵ 交éĢļ +åĶIJ å±± +çŃī ä¸Ģç³»åĪĹ +ä¸įè¿ĩ æĺ¯ +åŃĺåľ¨ çĿĢ +èĬ± çĶŁ +å¤ · +ç»Ī ç©¶ +ä¹Łæĺ¯ ä¸Ģ个 +åįģ åŃĹ +èĸª éħ¬ +伤 å¿ĥ +æĺ¥ ç§ĭ +åĨ· åį´ +ç²¾ çģµ +çļĦ åľ°åĽ¾ +æ¯Ķ çī¹ +æ¯Ķçī¹ å¸ģ +æĢ§ åĪ« +ä½Ļ ä¸ĩåħĥ +ä¸įå¿ĺ åĪĿå¿ĥ +å¿ĥ çĸ¼ +æĽ² 线 +é«ĺ ä½İ +è¦ı å®ļ +æĻ¯ èī² +è¦ģ 说 +åħ¬åı¸ å°Ĩ +æ¶² åİĭ +è¿Ŀ 约 +åİļ 度 +åºŀ 大çļĦ +è¿ĺæĺ¯ å¾Ī +é¦ĸåħĪ æĺ¯ +çµ ² +åĬ¡ å®ŀ +並 ä¸Ķ +å¢ŀ è¿Ľ +ç»Ħç»ĩ å¼Ģå±ķ +èµ·æĿ¥ äºĨ +è¾ĥ å°ı +导 游 +两 åľ° +ç¿ ĺ +çģ¿ çĥĤ +é£İ éĩĩ +æĶ¯ 线 +æĶ¯çº¿ ä»»åĬ¡ +娱ä¹IJ åľĪ +天津 å¸Ĥ +åĮħ åĽ´ +æľ¬ èµĽåŃ£ +éĩįè¦ģ 讲è¯Ŀ +åıĮ åIJij +åįİ ä¸½ +éĶ ¤ +åĦ¿ 女 +åįĸ åĩº +ä¾Ĩ 說 +ä»ĭç»į ä¸Ģä¸ĭ +åIJ¦ 认 +åĭ Ŀ +æĻ®éĢļ 人 +çļĦ åĬ¨åĬĽ +涨 åģľ +åŁºéĩij 管çIJĨ +ä¸Ģ个 éĩįè¦ģ +è¿IJ æ²³ +çħ ŀ +è´¢æĶ¿ éĥ¨ +è¡Įä¸ļ åįıä¼ļ +éĥ½ å°Ĩ +è¨Ģ 论 +ä¸ĭ ä¾Ĩ +墨 西 +墨西 åĵ¥ +åĽłä¸º ä»ĸ们 +æĢİä¹Ī åĽŀäºĭ +åĬłå¤§ 对 +èĬ Ń +çīĮ åŃIJ +ä¼ļ 使 +妹 åŃIJ +ç«Ļ éķ¿ +å¿ħ å¤ĩ +æłij æľ¨ +æģ¶ æĦı +æ²³ éģĵ +å¯Į è£ķ +ç¹ģ åįİ +代表 åĽ¢ +æµij 身 +é¦ĸ ä½į +èĪªç©º åħ¬åı¸ +鼻 å½± +ä¸ĵ è¾ij +æ°´ æºIJ +ä¸Ń æ¯Ĵ +並 ä¸į +èĢĮ åİ» +é ĥĿ +äºİ æŃ¤ +æĸĩåĮĸ 建设 +èĤ¯å®ļ ä¼ļ +å¸ĮæľĽ 大家 +æıı åĨĻ +ä½İ è°ĥ +æĸ°åħ´ 产ä¸ļ +æ·Ħ åįļ +æĶ¾ å¼Ģ +çļĦ æĢ§æł¼ +çĸ¾çĹħ çļĦ +æķ´ é¡¿ +线ä¸Ĭ 线ä¸ĭ +éĢī 项 +çļĦ 认åı¯ +æķ´ é½IJ +çĶļ ä¹Ī +çľģ åĨħ +åı¤ 人 +æ°ij ä¿Ĺ +çī¡ ä¸¹ +éŨ çªĹ +éĤ£ æł·çļĦ +çĽijäºĭ ä¼ļ +ç¿¡ ç¿ł +ç¦ ¹ +åįĥä¸ĩ ä¸įè¦ģ +æĶ¶ 缩 +çļĦ æĸĩåŃĹ +åĴĮ å°ļ +æĮĩ 令 +åħ±äº§ åħļåijĺ +çļĦ çĪ¶äº² +å®Į å·¥ +åĬ¡ å·¥ +马 æĭī +马æĭī æĿ¾ +æµĭ è¯Ħ +å² ļ +ä¸į åģļ +ä¸ĥ å¹´ +åĿĩ ä»· +主 è§Ĥ +å¾Ī ä¸įéĶĻ +èĤ¡ä¸ľ 大ä¼ļ +äºĶ ä¸Ģ +é£İ åIJ¹ +å¼Ģ éĩĩ +è¿Ļä¹Ī 大 +èĥ½ çľĭåΰ +èĢĥ è¯Ħ +åį³ ä¾¿æĺ¯ +çݰ代 åĨľä¸ļ +æ¯Ķè¾ĥ é«ĺ +è¦ģ çľĭ +没 äºĨ +è§£ 決 +çݯ æ¯Ķ +åĨ² åĬ¨ +æ·± å¤ľ +åĩł åįĥ +ä¿ ı +ç½ij æ°ij +å°± 没 +ä»ĸ 表示 +éĩı åŃIJ +æĹ©é¤IJ åĬłçĽŁ +åįĬ å²Ľ +æIJŀ ç¬ij +ä¸Ĭ æĬ¥ +å¯ © +é¢Ħ 订 +èľĤ èľľ +æŁ¥ æī¾ +ä¼Ĺ æīĢ +ä¼ĹæīĢ åij¨ +ä¼ĹæīĢåij¨ çŁ¥ +æĹ© æĹ¥ +åıij æī¬ +åĴĮ 个人 +åĬłåħ¥ äºĨ +åĸ® ä½į +åĪĨ æĺİ +第ä¸Ģ æī¹ +ç¾İ åĨĽ +æĿĢ æīĭ +éŨ å¤ĸ +åķĨ åľĪ +ä¸Ģ åĪ» +çļĦçľ¼ ç¥ŀ +éľ Ħ +äºĽ ä»Ģä¹Ī +åĬł æ·± +æ¯ı ä½į +å¸Ĥ éĿ¢ä¸Ĭ +åıĶ åıĶ +çļĦ éĤ£ç§į +粤 港澳 +è´´ å¿ĥ +æĸĩåĮĸ 产ä¸ļ +红 æĹĹ +åĺī åħ´ +æĶ¶ çĽĺ +å®ĮæĪIJ åIJİ +ä¼ģä¸ļ 管çIJĨ +纵 横 +ä¸į ä¿¡ +æĪIJ éĥ½å¸Ĥ +æ´Ĺ 澡 +举è¡Į çļĦ +çĶ¢ çĶŁ +ç©¿ ä¸Ĭ +åĪļ 好 +åħī 线 +æīĵ æŀ¶ +è¿Ļ æľ¬ä¹¦ +åĶ®åIJİ æľįåĬ¡ +åĩł åĪĨ +ä¸Ĭ 次 +ä¸į åĪĨ +产 åIJİ +éģ¿ å¼Ģ +ç»Ī æŀģ +代表 大ä¼ļ +æ¼Ķ æĬĢ +åĽŀ è´Ń +åѦ è´¹ +éĺ» ç¢į +ä¸Ģ大 æī¹ +ç«£ å·¥ +åĨ³ å®ļäºĨ +ä½Ĩ å¦Ĥæŀľ +ç͵ æµģ +ä¸Ŀ 毫 +èĥ½å¤Ł åľ¨ +éĶĢåĶ® æĶ¶åħ¥ +åľ¨ åŃ¦æł¡ +æ°´ åĩĨ +è§Ĩ 线 +èĩª åľ¨ +åķĨä¸ļ éĵ¶è¡Į +为äºĨ 让 +çį² å¾Ĺ +çݩ家 æľĭåıĭ +éĿ¢ èĨľ +åĪĨ åī² +åī§ æľ¬ +ç« Ń +说 å¾Ĺ +æĥ³ çŁ¥éģĵ +çļĦ人 çī© +èĮħ åı° +åIJĮ ä¸Ģ个 +æķ°æį® ä¸Ńå¿ĥ +çĶ Ħ +åĸľ æĤ¦ +ä¸ĭæĿ¥ çļĦ +å®ļ åIJij +æŀģ åħ· +çļĦ åľŁåľ° +éĤ£ åĢĭ +æijĦ åħ¥ +äºĨ æĪijçļĦ +马 è·¯ +åħ¨ 社ä¼ļ +è®® æ¡Ī +å±ĭ åŃIJ +åIJį åı« +åĮ ª +åľ¨ å¤ĸéĿ¢ +åįİ åįĹ +åıij è´§ +å¯Ĵ åĨ· +é«ĺçŃī æķĻèĤ² +详ç»Ĩ çļĦ +个 é¡¹çĽ® +çĶŁäº§ åĬĽ +æĹ¶ 常 +å°± æľĥ +ä¸ĩ èĤ¡ +éĻĮçĶŁ 人 +æıı ç»ĺ +å½ĵ çĦ¶æĺ¯ +æĭī åĬ¨ +éĵ¾ æĿ¡ +æī£ éϤ +ä¸Ģ缴 éĥ½ +å°ı åŃ©åŃIJ +伤 åı£ +第äºĮ å±Ĭ +è´Ń ç½® +çļĩ 马 +æĹł èģĬ +表 åĨ³ +诸 å¦Ĥ +åĵį èµ· +é£İ æļ´ +ä¸Ģæµģ çļĦ +ç ·¨ +è§£æĶ¾ åĨĽ +室 å¤ĸ +å°± è¿Ļä¹Ī +å³ ¶ +æīĢæľī 人éĥ½ +æIJľç´¢ å¼ķæĵİ +çļĦ æĪIJæľ¬ +åħļ æĶ¿ +åıijè¡Į 人 +çļĦ äºĭå®ŀ +对 该 +åıĹ æįŁ +ä¿Ħ ä¹Į +é²ľ èĬ± +åĨľ èᝠ+æŀģ éĢŁ +æĢ¥ æĢ§ +两 ä¼ļ +ä¸Ģèά æĿ¥è¯´ +æµ· é²ľ +åĨ Ī +ç͍ 人 +çĶ¨äºº åįķä½į +åĢ ª +åĦª æĥł +æł¹ æºIJ +åĽ¢ è´Ń +ç¾İ æ´² +ä¸ĭ è¡Į +å¹´ æľ« +èľ ¡ +è¯ģ ä»¶ +åľ¨ æĪijåĽ½ +ä¸į åºĶ +æĮī æĹ¶ +åłª ç§° +åľº ä¸Ĭ +å¹²éĥ¨ èģĮå·¥ +æľī å¾Ī大çļĦ +æķ°åŃĹ ç»ıæµİ +æ¼Ķ ç»ĥ +æį® ç»Łè®¡ +å¾Ģ æĿ¥ +广åijĬ æľįåĬ¡ +çļĦ è·Ŀ离 +æŃ ¸ +è¨Ģ è¯Ń +被 èªī +被èªī 为 +åĭī 强 +å°Ĭ æķ¬ +ä¸ĩ 亿åħĥ +ä¸ŃåĽ½ åĽ½éĻħ +å¹² é¢Ħ +å¹´ 产 +èĢķ åľ° +èĮ İ +åį³ æĺ¯ +æĺ¨ æĻļ +æĪIJ为 ä¸Ģ个 +çºł æŃ£ +åij½ åIJį +é¢ģ å¸ĥ +çĮľ æµĭ +ä¿ĿèŃ· æĶ¿çŃĸ +æĭ ¢ +æ´» æ³¼ +çŃī éĥ¨éŨ +åѦ åΰ +å¢ŀå̼ ç¨İ +èĪª 线 +åĨ ¤ +åįģ åĩłå¹´ +æİ§èĤ¡ èĤ¡ä¸ľ +ä¸Ģ éŨ +个 å·¥ä½ľ +ä¸ªå·¥ä½ľ æĹ¥ +æĸ° 西 +æĸ°è¥¿ åħ° +论 è¯ģ +ä» Ĩ +åı¦å¤ĸ ä¸Ģ个 +æĶ¹ ç¼ĸ +严 ç¦ģ +åĸľ 好 +个人 ä¿¡æģ¯ +满æĦı 度 +åĵ ¨ +å¸Ī èµĦ +æĶ¹ 为 +ç«ŀäºī 对æīĭ +åĩº çĤī +åķĨ 人 +大 æ£ļ +æĮĩ导 ä¸ĭ +å¦ĩ ç§ij +è¼ ª +æī ģ +åIJĮæĹ¶ è¿ĺ +å¹¶ éĢļè¿ĩ +æĪĺ éĺŁ +èĶĵ å»¶ +ä¿ ŀ +éĢĤå½ĵ çļĦ +åīį è¾Ī +åĵģ åij³ +湿 åľ° +æĪIJ åŀĭ +ä¸į åıªæĺ¯ +æĥ© ç½ļ +åĩºåı° äºĨ +çİ© 游æĪı +æīį åıijçݰ +åºĶ èģĺ +å¤ĸ æĿ¥ +åįł é¢Ĩ +å±ķ æľĽ +å« Ĥ +港 èĤ¡ +æ¡Į ä¸Ĭ +æĶ¯ æŁ± +çļĦæĥħ å½¢ +广éĺĶ çļĦ +æĶ¯ è¡Į +å´© æºĥ +æľĪ ä¸Ń +æľĪä¸Ń æĹ¬ +ç»į åħ´ +临 è¿ij +æĬ¤ æłı +æļ ® +åįķ èģĮä¸ļ +è¾¹ å¢ĥ +æĹ¥ çħ§ +ä¸Ģ åłĨ +缴 å¾Ħ +åħ±åIJĮ ä½ĵ +æĸ°åįİ ç½ij +æīĵ 好 +ç͵åĬ¨ 汽车 +ä¸į æĺİçϽ +éĢĻ è£¡ +缼 大 +çİĭ æľĿ +åĨį ä¸Ģ次 +åĬŀåħ¬ åİħ +è´¨ æĬ¼ +åIJĪ åĩ» +人们 对 +鼶 é£Ł +éĥ½ä¸į çŁ¥éģĵ +çļĦ è¯Ńè¨Ģ +åĭŁéĽĨ èµĦéĩij +åĬ¨ èĦī +å½ ¤ +è¿Ļ åĩłå¹´ +çŁŃ è§Ĩé¢ij +太 é«ĺ +常 å§Ķä¼ļ +åĬł çıŃ +éĩį å¿ĥ +åªĴä½ĵ æĬ¥éģĵ +没 æ³ķ +éĹ» åIJį +çĥŃ åº¦ +å¹¿æ³Ľ çļĦ +åħŃ å¤§ +çī© ä½ĵ +ä¸į 该 +é¢ĺ 主 +精彩 çļĦ +为 è¿Ľä¸ĢæŃ¥ +èĻ ŀ +åĽº çĦ¶ +è´µå·ŀ çľģ +çºł ç»ĵ +代çIJĨ 人 +æ³ķå®ļ 代表 +åı¦ä¸Ģ ç§į +ä¸į åIJ« +æĭ¯ æķij +ä¼ļ ç»Ļ +è¯Ĺ è¯į +åIJĮ ç±» +å¾Ĺ ä¸įåΰ +æĬĵ ç´§ +以 åħ¶ +åħ¥ åħļ +è¿ĺ åı¯ +æľŁ åĪĬ +å¾Īå¤ļ æĹ¶åĢĻ +æĹ¥ åIJİ +åħ¬ 约 +ä¸Ģ 举 +æ¯Ķè¾ĥ å¤ļ +éĩij æ²Ļ +æį ŀ +æİĴ åĩº +æŃ¦ æľ¯ +ä¸į æĸ· +ä¸Ń èĢĥ +ä¿¡ èµĸ +ä»İä¸ļ 人åijĺ +çģ« çĦ° +éĨĴ æĿ¥ +ä½İ 温 +é̾ æľŁ +åĬ± å¿Ĺ +éħ ¥ +åı¯è°ĵ æĺ¯ +è¿Ļ æĦıåij³çĿĢ +é¢ł è¦Ĩ +åĮĹ京 大åѦ +ä¸ĵ 线 +åıĬ 以ä¸Ĭ +è¨ ª +èĢĮ åIJİ +çŁ¥ ä¹İ +ä¸Ģ对 ä¸Ģ +å¨ĥ å¨ĥ +çģ¾ éļ¾ +åħ¨ å±Ģ +æīĢå¾Ĺ ç¨İ +å®ŀ æĥł +èļĤ èļģ +ä¹Ł çŁ¥éģĵ +温 åĴĮ +èIJ½ ä¸ĭ +åŀĭ ä¼ģä¸ļ +åĨį ä¹Ł +ä¾Ľ çĥŃ +é«ĺ æ½® +çĢı覽 åύ +çļĦ 巨大 +åħΠ天 +å¹´ ä¸ŃåĽ½ +类似 çļĦ +çIJĨäºĭ ä¼ļ +空 éĸĵ +çģµ æĦŁ +åĬĽ æ°Ķ +带 ä¸Ĭ +ä¸į好 æĦıæĢĿ +æľī ä½ķ +å·² åľ¨ +åıĸ åĩº +è¿Ŀæ³ķ çĬ¯ç½ª +åŃ¦ä¹ł 贯彻 +åľ° 带 +楼 梯 +çŃī æĥħåĨµ +ä»İ åīį +çļĦ ä¹łæĥ¯ +ç³Ł ç³ķ +å°± èĥ½å¤Ł +è© ķ +ä¸Ģ å¾ĭ +æĮ« æĬĺ +åİŁæĸĩ åľ°åĿĢ +å½ĵ å±Ģ +ä¸į éĢļ +æķ° åįĥ +éĺŁä¼į 建设 +æĹ¶ èĬĤ +åģļ èµ· +çļĦ è®°å¿Ĩ +ç½ij绾 å®īåħ¨ +åĩ¡ æĺ¯ +æ° ¯ +éĽķ åĪ» +åŁĥ åıĬ +æĪij åı¯ä»¥ +çĽij çIJĨ +æĽ´ åħ· +åŁİ 管 +èĭ ¯ +åı¥ åŃIJ +èĭ¥ æľī +ä»İæĿ¥ ä¸į +缸åħ³ è´Łè´£ +å®īåħ¨ æĦŁ +æĽ´ è¦ģ +çļĦæĥħ æĦŁ +çī¢ çī¢ +è¾ĥ 好çļĦ +æ° ® +ç¬ij è¯Ŀ +车 å±ķ +ä¹ĭ ç¾İ +ç®Ģ 约 +ç±»åŀĭ çļĦ +èĢģ åĮĸ +çľĭ ä½ł +è¿ĩ åĪĨ +éŨ åīį +ä¸Ģ éĹ´ +æĥ³ åİ» +åª Ľ +åľŁ è±Ĩ +åıĪ ç§° +ä¸Ń ä¿¡ +åŃĺ éĩı +马 äºij +èĩ´ 使 +åħĪ åīį +èĢģ åŃIJ +æīĵ æī® +æ¯ķä¸ļ äºİ +æ¯ķä¸ļ åIJİ +ç¾İ好 çĶŁæ´» +å·¥ä¸ļ ä¼ģä¸ļ +就好 äºĨ +èħIJ èļĢ +çıį çıł +åΰ è¿ĻéĩĮ +æīĢéľĢ çļĦ +è¿Ļæĺ¯ åĽłä¸º +çIJĨæĥ³ çļĦ +å·®å¼Ĥ åĮĸ +é ® +é® ® +äºļ 太 +æĹł ç©· +æıIJ çݰ +ä¸ĵä¸ļ æĬĢæľ¯ +çĶ¢ æ¥Ń +åѦ åŃIJ +ç§ij å¹» +åįłåľ° éĿ¢ç§¯ +ä¸į åĩĨ +æľªæĪIJ 年人 +æĶ¶ å½ķ +è¿ĺ 款 +éĴ¢ çŃĭ +æ¼ ¢ +å¾Ĺ æĦı +综åIJĪ ä½ĵ +æŀģ é«ĺ +åįķ è¯į +é«ĺæķĪ çļĦ +骨 头 +æī§ çĿĢ +缼 ä¸ĸ +模 çī¹ +æĽ´ èĥ½ +ç»Ŀ æľĽ +对åºĶ çļĦ +æ¨ Ĭ +æĸ° ä¸ī +æĸ°ä¸ī æĿ¿ +æģ° æģ° +åIJį å®¶ +æł¸å¿ĥ æĬĢæľ¯ +个 å°ı +æĢİä¹Ī ä¼ļ +说 ä¸įå®ļ +西 çĵľ +åĵ İ +ç¢ Ł +å¿ħ ä¸įåı¯ +å¿ħä¸įåı¯ å°ij +ä¹ĭ éĸĵ +åĪĨ 管 +交éĢļ äºĭæķħ +å¼Ģ åĬŀ +å¾ģæ±Ĥ æĦıè§ģ +äº ¨ +鼻åŃIJ éĥµ +鼻åŃIJéĥµ ä»¶ +ä¿¡æģ¯ æľįåĬ¡ +ä½ł è§īå¾Ĺ +缴 è§Ĥ +å·² å®ĮæĪIJ +åĪĨ ä¼ļ +åĽŀ åįĩ +éļ » +好 人 +äºĨè§£ ä¸Ģä¸ĭ +åį« æµ´ +æľĢ çα +åºŀ 大 +客 æĪ¿ +çijŀ åħ¸ +éĥ½ ä¸įæĺ¯ +é¤ ¨ +èĹ ī +çļĦ åIJĦ项 +为 缮æłĩ +çļĦ è®¤çŁ¥ +å½±åĵįåĬĽ çļĦ +夸 å¼ł +佩 æĪ´ +æ±ĩ çİĩ +çļĦ çαæĥħ +æĺ¥ é£İ +æĺ¯ æĪijçļĦ +æ¨ ¹ +åįĬ å°ıæĹ¶ +å±± åİ¿ +å±± 西çľģ +èĢĮ è¿Ļ +æĽ´å¤ļ ä¿¡æģ¯ +è¿ĺ æľīä¸ĢäºĽ +ç²¾ ç»ĨåĮĸ +ç¾İ åѦ +çͱ æĸ¼ +ä»ħä¾Ľ åıĤèĢĥ +å¾Ī é«ĺçļĦ +åıł åĬł +è¿Ļä¹Ī 说 +å±ķ åĩº +åĽĽ å¤Ħ +ä¸ĩ å®¶ +æĭĽ åĭŁ +çļĦ 强大 +æĤ£ æľī +å°ı äºİ +ä¹Łè®¸ æĺ¯ +对 èĩªå·±çļĦ +èģĮä¸ļ æķĻèĤ² +æĿ¥ è¿Ľè¡Į +æ¡£ 次 +æīĵ èµ¢ +éĥ½æľī çĿĢ +åº ¸ +è¯Ń æ°Ķ +çͲ éĨĽ +空 åĨĽ +车 åĨħ +åĽłä¸º ä½ł +å®ŀ æķĪ +æĥħ ä¾£ +åıijè¾¾ åĽ½å®¶ +éķľ åŃIJ +æ¯į å©´ +ä½Ĩæĺ¯ ä»ĸ +积æŀģ æİ¨è¿Ľ +大å¹ħ 度 +çļĦ 女åĦ¿ +é¤IJ æ¡Į +åIJ¬ å¾Ĺ +çļĦ 积æŀģæĢ§ +好 åIJ§ +æĹ¥ æ¶Īæģ¯ +æľī ä»»ä½ķ +æ¯Ĵ åĵģ +æĹ©çĤ¹ åĬłçĽŁ +第ä¸Ģ 天 +å°½ åĬĽ +æł ĸ +主 æīĵ +æĺ¯ä¸Ģ åIJį +çĪĨ æĸĻ +äºĭä¸ļ åıijå±ķ +å¾® åķĨ +äºİä¸Ģä½ĵ çļĦ +çĶŁ çĮª +èĩªçĦ¶ èµĦæºIJ +çŀĦ åĩĨ +è§Ħ模 åĮĸ +å¹¶ ä¸İ +èĤ¥ èĥĸ +å®¶ ç͍ +大 çĪ· +é¢Ħ åijĬ +æĿ¥ åģļ +éĺ³ åİ¿ +æŀĦ çŃij +é¢ģ å¥ĸ +åİĨåı² æĸĩåĮĸ +æľįåĭĻ æĪĸ +æĢ» åĨ³èµĽ +åıij åŀĭ +æĪij 羣çļĦ +æĽ ¦ +åıĤ ä¼ļ +èĦĨ å¼± +åĩĨ åħ¥ +èħ¹ éĥ¨ +åı¸ 令 +æĤ² åī§ +天 ä¸Ĭ +åı£ ä¸Ń +ä¸ĩ 个 +åѦ ä¸ļ +æıIJ åĢ¡ +两 è¾¹ +大 èĤ¡ä¸ľ +åı¤ éķĩ +è¡Ģ ç³ĸ +çļĦ ç¨ĭ度 +æ£ī èĬ± +åIJİ åı° +å°± åĮ» +æķ´ æķ´ +èĴ ² +çĽĪåĪ© èĥ½åĬĽ +ç± ½ +èĦ « +çľĭ éĩį +å®¶ éķ· +èģĺ ç͍ +èµĽ éģĵ +åīį èĢħ +建 èѰ +å¾ĭå¸Ī äºĭåĬ¡ +èīºæľ¯ åĵģ +æľī èĩªå·±çļĦ +åIJ¦ å®ļ +社 åĽ¢ +åij¨ äºĶ +带 åΰ +å·¥ä½ľ ä¼ļè®® +èĤ¡ æľ¬ +å¤ĸ åĮħ +å®¶ åħ¬åı¸ +çĽij çĭ± +èĪ Ĭ +åIJį æł¡ +西 æ¹ĸ +è¶ħè¿ĩ äºĨ +åįĹ å±± +ç»Ħ ä»¶ +å̼å¾Ĺ 注æĦı +æĮ£ æīİ +äºĭ 迹 +ç¶ĵ çĩŁ +ç§ij 室 +好 åIJĹ +æ¤ħ åŃIJ +åľĪ åŃIJ +ä½Ĩ 她 +æµģ çķħ +åIJĦèĩª çļĦ +èģĮ åijĺ +è¡į çĶŁ +åħ¨ åľº +æĴ¤ éĶĢ +åį´ è¢« +å®ģ éĿĻ +åīį æīĢ +åīįæīĢ æľª +åīįæīĢæľª æľī +主 ä¸ļ +åĮĹ ç¾İ +è¯Ħ å®ļ +åĵģ å°Ŀ +大家 éĥ½åľ¨ +主 å¸ħ +ç»Ĩ å¿ĥ +ä¿¡æģ¯ æĬ«éľ² +çļĦ ç«ŀäºī +éĢĻæ¨£ çļĦ +ç§ijåĪĽ æĿ¿ +éĩĩ æijĺ +票 æį® +éĢIJ å¹´ +èĭ± è¶ħ +è¡Įä¸ļ åĨħ +人 寿 +åIJİ åĭ¤ +å¦Ĥ æĦı +ç¬Ķ è¯ķ +æ·¡æ·¡ çļĦ +ä¸į èĪĴæľį +ä½ĵ 积 +ä¹Łä¸į è¦ģ +éĿ¢ æĸĻ +æł· æľ¬ +ç¥ ģ +æĮī è§Ħå®ļ +大æ¦Ĥ æĺ¯ +æĥħåĨµ è¿Ľè¡Į +åIJĦ åįķä½į +çļĦ ç¬ij容 +åĩºèī² çļĦ +代表 æĢ§ +çļĦ ç¾İ好 +éĴ ¦ +å¾® çĶŁçī© +è¶Ĭ æĺ¯ +æĸ¹ åı¯ +å¹² èĦĨ +éģĬ æĪ² +çļĦ åħ´è¶£ +éĹ® è´£ +åĽłä¸º æĪij们 +èĢĥ éĩı +çĶŁ çĶŁ +éĺ» åĬĽ +ä¸į åħģ许 +æıIJ è®® +åĩı æĮģ +åıªæĺ¯ ä¸Ģ个 +æĪij æĬĬ +åıijçݰ èĩªå·± +å¢ŀ å¹ħ +å¦ į +èĹĿ è¡ĵ +ä¸Ģå®¶ 人 +åĪĨ 级 +çļĦ æķ°éĩı +è½® èŀįèµĦ +çŃī åĽłç´ł +大 夫 +èģĺ 请 +é£İ æľº +绽 æĶ¾ +ä»»ä½ķ ä¸Ģ个 +éł Ĥ +éĺ¶ çº§ +æĬĬ 她 +è¿Ľ åĨĽ +èĥ½ åģļåΰ +åŁ¹è®Ń æľºæŀĦ +çī© æĸĻ +ç«¥ è¯Ŀ +æĮĩ导 æĦıè§ģ +éĺ ® +æ·±åħ¥ æİ¨è¿Ľ +主 æľº +æ¸Ķ ä¸ļ +ä¸į æľį +æµĵ éĥģ +è¡Ĺ ä¸Ĭ +ä¾Ŀ 次 +æĹ¶ 段 +æ¢ µ +çļĦ åĸľçα +å¾Ī éķ¿ +åĪĿ 级 +æŀľ æĸŃ +æĬ¢ æķij +é¼ĵ èĪŀ +ä¾Ľ éľĢ +æ·±åħ¥ å¼Ģå±ķ +产ä¸ļ éĽĨ群 +åĻª éŁ³ +åIJ¬ çĿĢ +æ·±åĪ» çļĦ +å¿į åıĹ +ç͵ ç£ģ +强 èĢħ +æ»ĭ åij³ +æĽ¼ èģĶ +åı¯ä»¥ 缴æİ¥ +大 ç±³ +æŃ· åı² +æĶ¿åĬ¡ æľįåĬ¡ +åħ¬ å¼ı +社 群 +éģĵ士 èģĮä¸ļ +ä¹ĭ æĥħ +æµ· æ°´ +æ¼Ķ å¥ı +åºĹ éĩĮ +迹 象 +åıijå±ķ çIJĨ念 +é«ĺ 空 +åij¨ åĪĬ +åĽŀ åΰäºĨ +ä¸į éĢĤåIJĪ +åłµ å¡ŀ +åĬ Ī +æ°´ ä¸Ĭ +çĢij å¸ĥ +纳ç¨İ 人 +çĩĥ æ²¹ +å·¥ç¨ĭ é¡¹çĽ® +峡 è°· +æľī éĴĪ对æĢ§ +åľĨ å½¢ +æľ¬ å¸Ĥ +è¿Ļ è¯Ŀ +管çIJĨ èĢħ +ç¡®è¯Ĭ çĹħä¾ĭ +æĬĬ æīĭ +彩 èī² +ä¸Ĭ åīį +夯 å®ŀ +ç¾Ĭ èĤī +å¾Ģ å¹´ +æĵħ èĩª +è¿· 人 +èĪª æ¯į +ç²¾ ç»Ĩ +åľ¨ æĪijçļĦ +åĪĽ æĬķ +麦 åħĭ +æľĪ ç»ı +åĮĹ æµ· +ä¹ĭ æĺŁ +åı¶ åŃIJ +å¸Ĥåľº ç«ŀäºī +è¿Ļ äºĭ +åıĥ èĪĩ +产 åľ° +åĶ ī +åķĨåĵģ æĪ¿ +èĪª è¿IJ +ä¼ĺ å¼Ĥ +ä»ĸ们 æĺ¯ +鼨 æ°´ +è¯į æ±ĩ +åĨľ çͰ +欧 éĺ³ +çŁŃ 线 +管 ç½ij +æł¹ åŁº +åıªæľī ä¸Ģ个 +éŀĭ åŃIJ +å¸Ĥ å§Ķ书记 +åĪ» æĦı +è¡Į 车 +åıĪ è¢« +åı¯éĿł æĢ§ +è´ ± +ä»» åij½ +åºĶ åľ¨ +å°± å¾Ĺ +æľįåĬ¡ ä½ĵç³» +æĶ¿ æĿĥ +åıijè¨Ģ 人 +è¿ĩ å¾Ģ +两 åıª +èϽ 说 +éĢģ ä¸Ĭ +ä»Ģä¹Ī äºĭ +æķ£ æĸĩ +æİĮ æİ§ +èĸĦ å¼± +ä¸ĭéĿ¢ å°± +主è¦ģ åĨħ容 +å¾Ī éĩįè¦ģçļĦ +å°± 说 +çϽèī² çļĦ +éĤ£ä¸ª æĹ¶åĢĻ +ç»ı纪 人 +çļĦ æ¯į亲 +ç¬Ķè®° æľ¬ +åºķ å±Ĥ +è¿ij 代 +è§£ 说 +è²ł 責 +æľĢ大 åĮĸ +åķĨ éĵº +æł¡ åıĭ +æ² ģ +ä¸į åĩºæĿ¥ +éĻ· éĺ± +ç¨ ħ +åħ¬å¸ĥ äºĨ +åĩĢ å̼ +çĽ¸å¯¹ è¾ĥ +ç¬ Ľ +æł¸ ç®Ĺ +åįİ ä¾¨ +æĢ¥ æķij +æĮº 好 +åħĴ ç«¥ +äºĮ èĥİ +åĩº èĩª +åĿ Ł +æīĭ ä¸ĭ +å± ¡ +åĪĽéĢł æĢ§ +ä¸¥æł¼ æĮīçħ§ +åĨį åİ» +举 缣 +人 æµģ +äºĨä¸Ģ 声 +å°ıæĹ¶ åīį +è´µ æĹı +éľ ĸ +ä¹Łæĺ¯ éĿŀ常 +éĢ ± +çľĭäºĨ çľĭ +ç¹ģ æ®ĸ +èĩ³ æŃ¤ +é¢Ħ å¤ĩ +å¾Ī æĺİæĺ¾ +æ¼Ķ èīº +åĿIJ çĿĢ +ä¿Ħ åĨĽ +åľ¨ è¿ĩåİ» +ä¹ĭ äºĭ +æĬĵ èİ· +åĿIJ ä¸ĭ +çͱ ä¸ŃåĽ½ +ä¹Ł å¼Ģå§ĭ +çŃĶ å¤į +åŀĥåľ¾ åĪĨç±» +éĴĵ é±¼ +åIJĦ 種 +缸 éģĩ +ä¸įåģľ çļĦ +æī¹ éĩı +éĩįè¦ģ ä½ľç͍ +å§Ķ å±Ī +åħŃ å¹´ +ä¸ĥ åįģ +ä¹ĭ æĪĺ +é£İéĻ© 管çIJĨ +éŁ³ æ¨Ĥ +è¡ĮæĶ¿ å¤Ħç½ļ +æľ¬ äºĭ +æĴ° åĨĻ +èģļ åIJĪ +éĢĤ æĹ¶ +æIJ¬ å®¶ +ç¢İ çīĩ +缼 å®´ +ç®Ģ æ´ģ +åı¬ éĽĨ +ç®Ģ åĮĸ +åĮĹ京 æĹ¶éĹ´ +第ä¸ī å±Ĭ +æĿ¥ åĽŀ +常ç͍ çļĦ +京 æ´¥ +京津 åĨĢ +梦 å¹» +è¯ķ è¡Į +æľº åºĬ +åΰ æľĢåIJİ +åĬ© æīĭ +åĪĨ 彩 +åĩº åĵģ +åι 车 +åIJ¯ åıij +ä¾§ éĿ¢ +æ¯ı å½ĵ +缸åħ³ è§Ħå®ļ +ä¸ĸ 人 +è´Ń 车 +å¿ĥ 缮 +å¿ĥ缮 ä¸Ń +äºĶ éĩij +è¿ĺ è®°å¾Ĺ +ä¾Ŀ çĦ¶æĺ¯ +æıIJ æ¡Ī +ç͵åķĨ å¹³åı° +åģļ åΰäºĨ +æĿľ ç»Ŀ +å®ī åįĵ +ä¸ĸçķĮ åIJĦåľ° +åīį éĢĶ +æ´Ĺ åĩĢ +å¥ĭ åĬĽ +åŁİå¸Ĥ 建设 +å¤ļ åĬŁèĥ½ +ä¼ļ éĢłæĪIJ +åıijå¸ĥ ä¼ļä¸Ĭ +ç©¶ 竣æĺ¯ +åĪĨ 红 +çŁ¥ èŃĺ +éĿ¢ æĿ¿ +æĹł 声 +æĢ¥ éľĢ +失 çľł +çΏ å¦Ī +äº Ĥ +åħ¨ æĻ¯ +ç»ıåħ¸ çļĦ +åī§ ä¸Ń +é¢Ĩ导 ä¸ĭ +åħļ åĨħ +åħ¥ ä¾µ +æĭī æĸ¯ +ä¸Ģ å¹ķ +åĬł ä¹ĭ +èĤ Ĩ +èĭ± æł¼ +èĭ±æł¼ åħ° +å·§ åħĭ +å·§åħĭ åĬĽ +ä¸Ģ å¿ĥ +èģ Ĥ +å¾Ģå¾Ģ æĺ¯ +管çIJĨ å±Ĥ +çĻ» åħ¥ +建ç«ĭ èµ· +建 åĽ½ +åŃIJ 宫 +åºĶ ä»ĺ +æİ¢ ç©¶ +第ä¸Ģ ä½į +ä½Ļ å®¶ +çŃī æ´»åĬ¨ +æīĢ èĩ´ +è¾ĥ å¿« +æĺ¯ éĿŀ +æıIJ åIJį +äºĮ èĢħ +åıªåī© ä¸ĭ +åħ¶ä¸Ń åĮħæĭ¬ +ç¼ĸ ç¨ĭ +çł´ ç¢İ +ä¸Ń 举 +å·¥ä½ľ æĬ¥åijĬ +çѾ åIJį +éħĴ ä¸ļ +çŁ¥ æĻĵ +çĥŃ å¿ĥ +éĿŀ åĩ¡ +èIJ¥ä¸ļ æī§ +èIJ¥ä¸ļæī§ çħ§ +人大 代表 +ä¸Ģ个 æĸ°çļĦ +å¨ģ æµ· +éĤ£ 人 +涨 ä»· +æ¶Ī çģŃ +éļ¾ å¿ĺ +ç¶ĵ é©Ĺ +åı£ è¢ĭ +ç³» æķ° +æĸĩ ä¸Ń +好 转 +æĸ° 鼶åĶ® +讲述 äºĨ +å¼Ģ çĽĺ +çķĻ ç»Ļ +æħ¢æħ¢ çļĦ +æĤ² 伤 +æľ¬ æľŁ +äºĨ å¤ļå°ij +è¿Ļ 让 +åIJĮ çŃī +æ¸ħ æĺİ +个 åŁİå¸Ĥ +æºĸ åĤĻ +åĩłä¹İ æĺ¯ +强 åĬĽ +ä¿ ¯ +æ°´ 稻 +åĽºå®ļ çļĦ +æł¸ åĩĨ +说 æľį +顯 示 +è¿Ļ å¥Ĺ +æĻºæħ§ åŁİå¸Ĥ +å±ĭ é¡¶ +ä¸į æĿ¥ +çĶŁ é²ľ +çŁ¥ æĥħ +æĬķ 身 +åijĬè¯ī æĪij们 +ä¸ī åĽĽ +ä¸ĩ ä¸Ģ +è¾Ĩ 车 +为 ä¹ĭ +åΰ æĹ¶åĢĻ +è¿Ļ æīįæĺ¯ +åIJį çīĮ +åºŁ æ°´ +åݻ年 åIJĮæľŁ +å¹´ éĻIJ +éģĭ åĭķ +åıĮ çľ¼ +è¦ģ ç´§ +对 çŃĸ +åľº é¦Ĩ +çϾ ç§ij +è¶Ĭ éĩİ +å¯Į åIJ« +大å¤ļæķ° 人 +æľĢ å°ij +åı¬ åͤ +åħ¸ èĮĥ +åĨľ æľº +æŃ£ æĸĩ +åºĶç͍ äºİ +æ·± èĢķ +ä¿ Ń +ä»Ģä¹Ī ä¸ľè¥¿ +å¥Ĺ é¤IJ +å½ĵ éĢī +å·¦ æīĭ +è°ĥ çIJĨ +æĻļ é¤IJ +éļ¾ åħ³ +åĩŃ è¯ģ +çα 人 +æĮĩ è´£ +è´£ ç¼ĸ +çļĦä¸Ģ 款 +éĵ ² +åįģ 个 +èĢ » +æľįåĬ¡ åķĨ +åľ° çĭ± +è¿ŀ å¿Ļ +åĽ° æĥij +çļ ĵ +ä¸į åIJĥ +çİ°åľ¨ å·²ç»ı +çĽĺ çĤ¹ +ä¸įåģľ åľ° +管çIJĨ 模å¼ı +è¿Ļ 段æĹ¶éĹ´ +æ¤ ° +礼 åĮħ +æµģ 转 +æī« çłģ +éĽĨä¸Ń åľ¨ +æ±Ĥ åĬ© +åįĬ 个 +å¿«éĢŁ å¢ŀéķ¿ +å¾Ģ ä¸ĭ +è¯Ħ åĪĨ +å°± æĥ³ +åķĨåĬ¡ éĥ¨ +æľī éĹ®é¢ĺ +èİ· åĪ© +æ¯Ľ çĹħ +æĦŁ åºĶ +èī¯ æĢ§ +åĪĨ æŃ§ +åĨ ī +æĪij们 çİ°åľ¨ +è¦ģ åĬłå¼º +å·§ å¦Ļ +èŀº æĹĭ +åĪĩ æį¢ +çĭ Ħ +顺 çķħ +å°¤åħ¶ æĺ¯åľ¨ +èĬĿ 麻 +éļ¾ è¿ĩ +æĹĹ å¸ľ +å¤į åį° +å¤įåį° ä»¶ +å¿ħ éľĢ +对å¤ĸ å¼ĢæĶ¾ +éļ¾ åıĹ +åİŁæĿ¥ æĺ¯ +ç®Ĺ äºĨ +é«ĺ å±± +离 èģĮ +çµĦ ç¹ +çµĦç¹ Ķ +å±ģ èĤ¡ +çϾ å®¶ +éģĩ ä¸Ĭ +æĺĶ æĹ¥ +ä¸į 容 +çĽij管 éĥ¨éŨ +主 æĦı +æµģ åŁŁ +è·Į å¹ħ +èĩ³ ä¸Ĭ +åĪ« 说 +æĺ¯ æ¯Ķè¾ĥ +å®ıè§Ĥ ç»ıæµİ +å¸Ĥåľº 主ä½ĵ +污æŁĵ çī© +æķij æ²» +丰 æĶ¶ +åŃĺ æĶ¾ +åĩ Ħ +éĩij å±± +æį¢ äºĨ +ä¸ĵ 人 +éĹľ æĸ¼ +æĹ¢ è¦ģ +åĽ½ è¶³ +éļ ĭ +åıį åĩ» +èµ· 身 +åħĪ æĺ¯ +å¸ĮæľĽ èĥ½å¤Ł +åζ 订 +åºĹ éĿ¢ +åĸ Ģ +æķĻ ä½ł +éĻį æ¸© +åĬĽ æ±Ĥ +ä¸ī çϾ +çī© ä»· +丢 失 +å¢Ļ ä¸Ĭ +éĥ¨ 份 +æł· æĿ¿ +ä¹ĭ æĦı +ç½ij å°ıç¼ĸ +ä¸ĸ ä¸Ĭ +è°ĥ è¯ķ +污æŁĵ éĺ²æ²» +å½± éĻ¢ +å®Įåħ¨ åı¯ä»¥ +éĢļ åħ³ +ä¹īåĬ¡ æķĻèĤ² +没æľī åĬŀæ³ķ +èĢ ¿ +å¦ ³ +æĹł æĥħ +å¾Ĺ çĽĬ +å¾ĹçĽĬ äºİ +æľŁ çĽ¼ +娱ä¹IJ åľº +çͲ æĸ¹ +ä¸Ģ æ±½ +çĹ ° +çĸij ä¼¼ +æĸ°æµª å¾®åįļ +强 è¡Į +å½ĵ ä»ĸ +èĥ º +ç͍æĪ· æıIJä¾Ľ +åĮº å§Ķ +æĦ¿ æĻ¯ +æĬĺ æī£ +失 踪 +è¿« åĪĩ +åŃĹ æ¯į +åĴ ¯ +èªį èŃĺ +ä»Ģä¹Ī æĦıæĢĿ +çĽĴ åŃIJ +å½ķ éŁ³ +建设 å·¥ç¨ĭ +ä¸ļ ä½Ļ +å®ŀè·µ æ´»åĬ¨ +羣 空 +çĤ ĸ +åľ¨ è·¯ä¸Ĭ +主è¦ģ åĮħæĭ¬ +该 æĢİä¹Ī +æĢ» æľī +æĢ§ æĦŁ +æ°ij èĪª +å¼Ģ åºĹ +欺 éªĹ +çªģ åĩ» +缺 失 +æī§ ä¸ļ +åľ° éģĵ +å¹¶ æĹł +æ°ij åĬŀ +ç»Ħç»ĩ çĶŁæ´» +æĪij å¦Ī +è¨ĺ èĢħ +管 åζ +æī¾ 个 +èĹ » +çĤİ çĹĩ +äºĴ åĬ© +æµıè§Ī åύ +çݩ家 æĿ¥è¯´ +éĻįä½İ äºĨ +è£ Ķ +æĮ£ éĴ± +åķĨ æľº +æĶ¹ è£ħ +æµģ 浪 +æĶ¿ æ³ķ +èĢģ 头 +çĶŁäº§ åĴĮ +ç© Ĺ +亲 çα +亲çα çļĦ +å±¥ èģĮ +åŁİ éĩĮ +ç»Ĩ åĪĨ +åĬ³åĬ¨ åIJĪåIJĮ +åľ¨ æĹ¥æľ¬ +å¨ģ å°Ķ +åį« è§Ĩ +éĢ£ çµIJ +çĿĢ éĩį +æĬĺ 磨 +åĽ¾ 为 +çľ · +å·¥ åºı +æĵ ģ +æĵģ æľī +ç½ijç«Ļ åľ°åĽ¾ +çļĦä¸Ģ 大 +ç»Ħç»ĩ å®ŀæĸ½ +æĬĽ å¼ĥ +åĴĮ æĶ¯æĮģ +æ³ķ åĪĻ +浪 æ½® +çݰ æľīçļĦ +åĩł çİĩ +为 客æĪ· +åįģ ä¸ĩ +è ¹Ħ +çªģåĩº éĹ®é¢ĺ +åıĥ åĬł +éĥ½ä¼ļ æľī +çĽ ¤ +è°ģ éĥ½ +æīĭ åĬ¨ +缴 è¾¾ +çĤ¹ å¤ļ +éĺ¶ å±Ĥ +ä¸į ä½³ +éĤ£ 段 +滨 æµ· +æĺ¯ åĽ½åĨħ +æĪij å¸ĮæľĽ +åIJĽ åŃIJ +è§Ĥ éŁ³ +åģļ é¥Ń +æ±½ è»Ĭ +åħ³ ç¨İ +çľ¼åīį çļĦ +æ°´ éĿ¢ +è̳ æľº +追 踪 +æİ¨ éĢģ +éĴ± åĮħ +æģ¶ å¿ĥ +æµ· åŁŁ +å· į +å¼Ģ æĿ¥ +表 æĢģ +仪 表 +å¹³ åİŁ +åįģ å¤ļå¹´ +ä¹Ł æĹłæ³ķ +åħ¼ 顾 +è¡£ æŁľ +æł½ åŁ¹ +æĪ¿ æºIJ +设ç«ĭ äºĨ +ä¸ĩ åIJį +æķ° é¢Ŀ +è¦ģ åĿļæĮģ +åIJīæŀĹ çľģ +请 èģĶç³» +ç»ıåİĨ è¿ĩ +çļĦ æľ¬è´¨ +åħ¥ éŨ +æľ¬ æ¡Ī +çİĩ è¾¾åΰ +åı° éĺ¶ +éĴ ŀ +æĪij èĥ½ +èݲ èĬ± +éĴ ł +ä¸Ģ äºĭ +åİŁ æľīçļĦ +æ¯ı åĢĭ +æ¯Ķäºļ 迪 +æ£ĭçīĮ 游æĪı +ä¸įä¼ļ æľī +å½Ĵ æĿ¥ +äºĶ çϾ +è¿ĩ é«ĺ +鼷 è¾¾ +ä¸Ģèµ· åİ» +æķĻ å¯¼ +å°± è¯Ĭ +å°± å¾Ī +ä¸įåIJĮ äºİ +ä¿ º +å¸ĸ åŃIJ +æĶ¿åįı å§Ķåijĺ +çĸ«æĥħ å½±åĵį +åĪĨ è£Ĥ +为ä»Ģä¹Ī ä¼ļ +äºĶ æĺŁ +å°ij åĦ¿ +æĬ¢ éĻ© +梦 è§ģ +è®°èĢħ éĩĩ访 +å±± è·¯ +æĪij 个人 +æ²Ļ 滩 +è¹ Ń +æĶ¹ è®Ĭ +æĸ°åŀĭ åĨł +æĸ°åŀĭåĨł çĬ¶ +åĮ» æĬ¤ +åĮ»æĬ¤ 人åijĺ +æµ· å°Ķ +åħ³äºİ æĪij们 +éϤ å¤ĸ +åº ļ +宣 åijĬ +ä¸ī åįĥ +æ¦ ¨ +ç§ijæĬĢ å¤§åѦ +ä¸ĥ åħ« +顺 åºĶ +çΏçΏ å¦Īå¦Ī +éĢī åıĸ +åī§ çĥĪ +乡æĿij æĹħ游 +积æŀģ æİ¢ç´¢ +表çݰ 为 +å¾Ī æ¸ħæ¥ļ +大 åĨĽ +æĿ¥ ç͵ +å¥Ĺ æĪ¿ +çݰ è¡Į +享 åıĹåΰ +çľĭ çĤ¹ +åĽºå®ļ èµĦ产 +以 人为 +以人为 æľ¬ +ä¸į å®Į +éĻį 鼨 +åģļçļĦ äºĭæĥħ +å¹¶ äºİ +顽 强 +èĢ ¸ +åĺ´ å·´ +缸åħ³ ä¿¡æģ¯ +æĪij 没 +æĪĺçķ¥ æĢ§ +æĢĿ 念 +åĪĺ å¤ĩ +åĬ© æĶ» +é£İ è²Į +éĿ¢å¯¹ éĿ¢ +积æŀģ å¼Ģå±ķ +çĸĹ æķĪ +çľĭ 书 +缺 åı£ +åĽ½æ°ij ç»ıæµİ +使ç͍ æĿĥ +éģ¥ è¿ľ +å¡« è¡¥ +第ä¸ī 人 +åįĬ å¤ľ +æŃ¦æ±ī å¸Ĥ +æĪij åıijçݰ +ä¼ĺæĥł æĶ¿çŃĸ +é£İ åı£ +å°± ä¸įèĥ½ +为 主è¦ģ +æµģ åĩº +å´ĩ æĭľ +å¹¶ ä¸įèĥ½ +é«ĺ ä¸ī +ä¸ĸçķĮä¸Ĭ æľĢ +æĥ³ å¿ħ +åħ¶ æīĢ +åĢĻ éĢī +åĢĻéĢī 人 +ä¸į çα +åī¯ ä½ľç͍ +人æ°ij æĹ¥æĬ¥ +æĪij ä¸įæĺ¯ +å®ŀ çī© +ç͵ åİĤ +ä¹Ł ç®Ĺæĺ¯ +æľī éĹľ +æľī èĥ½åĬĽ +æĮĤ åľ¨ +çľ¼ ä¸ĭ +约 ç¿° +å°ı åѦçĶŁ +èµ· åΰäºĨ +å·¥ 夫 +åIJĮ å¿ĥ +åĿ¦ è¨Ģ +çł Į +åıijæĮ¥ äºĨ +èģĮä¸ļ éģĵå¾· +è¿ĻäºĽ å¹´ +念 头 +èĢģ é¼ł +åħ¨ èµĦ +åħ¨èµĦ åŃIJ +ä¸Ģ åij³ +å¤ļ ä¸ĩåħĥ +æł¼ æľĥ +éķ¿ éĢĶ +带 èµ° +èĭ± 寸 +æĸĩ ä½ĵ +对 ä»ĸ们 +åĵŃ äºĨ +å¡« æĬ¥ +çīĪæĿĥ 声æĺİ +ç͵ 线 +è´Ńçī© ä¸Ńå¿ĥ +饱 满 +ä½İ 头 +强 è¿« +ä¿Ŀ æ´ģ +欧 åĨł +缸 è¿ŀ +认 è´Ń +çģ« æĺŁ +é«ĺ å°Ķ +é«ĺå°Ķ 夫 +èij« èĬ¦ +æłĩ 注 +çļĦ çIJĨæĥ³ +æł¸ éħ¸ +æł¸éħ¸ æ£Ģæµĭ +åĬ ī +ä¸Ģèά æĺ¯ +æĢĿ ç´¢ +轨 迹 +çĥŃ å¸¦ +éĻ £ +åĩĨç¡® æĢ§ +æĪ´ çĿĢ +åľ¨ çĶŁæ´»ä¸Ń +æīĢ èĥ½ +æľ¯ åIJİ +带 ä½ł +ç¥ ł +æ®ĭ éħ· +ä¹Ł åıªæĺ¯ +çͳ è´Ń +举åĬŀ äºĨ +æľī æĦıä¹ī +æĹº 缼 +åľ¨ ç¶² +åľ¨ç¶² è·¯ä¸Ĭ +å¾Ī大 ç¨ĭ度 +管 è¾ĸ +çĸ«æĥħ æľŁéĹ´ +触 æij¸ +éĺ¶æ®µ æĢ§ +ä¼ļ è§īå¾Ĺ +çļĦ çĶ»éĿ¢ +æİ¥åıĹ äºĨ +表达 äºĨ +éĤĵ å°ı +éĤĵå°ı å¹³ +åħļ é£İ +åħļé£İ å»īæĶ¿ +åķĨ åѦéĻ¢ +åħij æį¢ +é£Łåĵģ èį¯åĵģ +éĿŀ常 好çļĦ +çľ ¯ +纳 ç±³ +åĬ¨ æijĩ +åĽŀ éģ¿ +çľĭ èijĹ +款 项 +åħ« å¹´ +åģļ 个 +æĸĩ æ¡£ +éĩijèŀį ç§ijæĬĢ +åħ¶ä¸Ń æľī +äºĨä¸Ģ ç³»åĪĹ +æĹĹèΰ åºĹ +ç§° èµŀ +éĽ¢ éĸĭ +åζ åĨ· +å®¶ éŨåı£ +åįģ å¤ļ +ä¼´ ä¾£ +çľĭ çĹħ +æĭī çĿĢ +æī Ĵ +çĸ² æĥ« +å°ijæķ° æ°ijæĹı +åĽ¾ å½¢ +è½ § +å¢ŀ éĩı +饲 åħ» +çģ« å±± +æ¯ı 个æľĪ +ä½ľä¸º ä¸ĢåIJį +è½´ æī¿ +æĸĩ 书 +ç¼ ķ +åħ·ä½ĵ æĥħåĨµ +çĹĽ çĤ¹ +缴 éĶĢ +å¡ Ĭ +ä¹Ł æľĥ +çĥŃ æ½® +å¹³ æ°ij +æ¼Ķåͱ ä¼ļ +æķĻ çłĶ +éĢĥ éģ¿ +ä¸Ģ è´¯ +å°± è¶Ĭ +å®ŀ å®ŀåľ¨ +å®ŀå®ŀåľ¨ åľ¨ +ä¹łè¿ijå¹³ æĢ» +æº º +å¿ĥ åºķ +éķ¿ å¾ģ +媽 媽 +第ä¸ī 次 +åĩº æ¼Ķ +çĭĢ æ³ģ +å°Ķ æĸ¯ +代çIJĨ åķĨ +çĨ ı +çļĦ 对象 +ç͵ éĩı +è¡Į åĪĹ +åĽ½ 人 +è·ij äºĨ +åįĶ åĬ© +èIJ¥ è¿IJ +å¸Ī åħĦ +æ¦ ® +æĥ³ åĥı +æĢ§ 强 +ç§ijåѦ çłĶç©¶ +å»¶ å®ī +ä¸¥æł¼ èIJ½å®ŀ +é¢Ĩ ä¼ļ +缸 å·® +è·¯ 人 +çĶ « +æľī ä»·å̼ +æľīä»·å̼ çļĦ +ç¾İ åĽ¢ +æ°ij主 çĶŁæ´» +æĪij æīį +ç¾İåĽ½ 人 +æ°Ķ åij³ +åıį å°Ħ +çļĦ åĨ³å¿ĥ +大 è±Ĩ +交 代 +è¿Ľ åĩº +åıį æĬĹ +æĮĩ çļĦæĺ¯ +ä»· ä½į +è¿Ľ é©» +ä¸Ĭ çϾ +ä½į åĪĹ +ä¸ŃåĽ½ ä¼ģä¸ļ +çļĦ好 å¤Ħ +主 ç¼ĸ +æ±½ æ²¹ +ä½Ĩ æĪij们 +æĢİä¹Ī çľĭ +é»Ħ å±± +å¤ļ åªĴä½ĵ +åIJİ åį« +èİ·å¾Ĺ æĽ´å¤ļ +åĬ¡ å¿ħ +为 å¥ijæľº +é¦ĸ 饰 +ä¸ĩ åįļ +è¶ĬæĿ¥è¶Ĭ 大 +ä¸ĵ项 è¡ĮåĬ¨ +å¥ĭ è¿Ľ +ä»į çĦ¶æĺ¯ +è´¨ æĦŁ +å¦Ĥæŀľ ä¸įæĺ¯ +ç«Ļ èµ·æĿ¥ +ä¹¾ éļĨ +åı¯æĢķ çļĦ +å¯Į è´µ +æ¸ħ ç®Ĺ +åIJij ä¸ĭ +åĢ ļ +çļĦ çŃĶæ¡Ī +èι ä¸Ĭ +çļĦ羣å®ŀ æĢ§ +çŃī åĬŁèĥ½ +åĸľ åī§ +å¨ģ åĬĽ +æĸ° é¢ĸ +æł¸ ç͵ +æĬ¥ éĶĢ +æķħ 乡 +ä¼´ éļı +éŀ Ń +å¦Ĭ å¨ł +åĪĨ åĮĸ +æľī å¾Ī大 +æĢİä¹Ī 说 +æĻĤ 代 +产 åĩº +ä»ĭç»į 说 +å¤ĦçIJĨ åύ +èĨ¨ èĥĢ +åī¯ å¸Ĥéķ¿ +çļĦ 妻åŃIJ +æł· åĵģ +åIJĮæ¯Ķ ä¸ĭéĻį +åħĥ å·¦åı³ +ç͍ èĩªå·±çļĦ +é«ĺ éĽĦ +æĺ¥ æĻļ +ä¹Ł æľīå¾Īå¤ļ +çľ¼ çIJĥ +æķ£ æŃ¥ +ä»ĸ们 éĥ½ +第ä¸Ģ å®¶ +åĬŀ 好 +å®ī éĺ² +ä¸Ģ ä¸ĩ +åľ¨ éĩĮéĿ¢ +éŁ³ é¢ij +åı£ åı· +ä¸Ģ è¶Ł +ç¦ı çī¹ +é³ ŀ +æĥĬ èī³ +æĸ° å¨ĺ +绿èī² åıijå±ķ +ä¸Ń å¼ı +ä¹Ł åıªæľī +çݰ 身 +åı¯ ä¾Ľ +æ¯ı ä¸Ģ个人 +第ä¸ī èĢħ +åľ° å½¢ +éĴ¢ ç»ĵæŀĦ +çĽijçĿ£ æ£ĢæŁ¥ +åı« æĪij +èĩ´ æķ¬ +æ´Ĺ æīĭ +ä¸ĭ è°ĥ +康 çĨĻ +æĪIJ交 éĩı +ä¹Ł æĪIJ为 +åħī æ»ij +å®Įæķ´ æĢ§ +çģ ¼ +ç¶² éłģ +éķ¿ å¯¿ +éģ© ç͍ +çļĦä¸Ģ 项 +çŀ© 缮 +æĬĬ èĩªå·±çļĦ +éĵ¶è¡Į åį¡ +å°± å¿ħé¡» +ç¾İ çϽ +éŀį å±± +æľ¬ é¢Ĩ +ä¸Ģ ç¢Ĺ +æīĵ æ³ķ +æĤ¨ 好 +对 åŃ©åŃIJ +æĬ¥éģĵ ç§° +ä¼ł åĩº +大 èĩ£ +ç¬ ĭ +çĽ ı +é¾ ļ +缴 线 +æĻº åºĵ +ç§Ł 车 +é£İ åij³ +çľĭ ä¸Ģä¸ĭ +æİ¨ éĶĢ +éĥ¨ éĥ¨éķ¿ +è´¨éĩı åĴĮ +åĪĬ çĻ» +å·¥ä¸ļ åĮĸ +çİĩ 为 +鼶 ä»¶ +硬 åĮĸ +ä¸Ĭ åįĥ +ç»ıéªĮ å̼ +å¹³ è¡Į +声 éģĵ +æľįåĬ¡ è´¨éĩı +çĶŁ çĶ¢ +æľĢ 容æĺĵ +ä¸Ģ æŀļ +å¹´ æĬ¥ +åħ¬ ç½ij +åħ¬ç½ij å®ī +åħ¬ç½ijå®ī å¤ĩ +çļĦ èĥ½éĩı +å®ŀéĻħ è¡ĮåĬ¨ +è¦ģ ä¸įè¦ģ +æĹ¥æľ¬ 人 +è̶ 稣 +ç¼ĸ åī§ +æ¶ © +åį° å°¼ +ä¸Ĭä¸ĭ 游 +åĩł åı¥ +ä¸Ń éĵģ +ç°¡ åĸ® +èĩª 带 +çĶŁ äºİ +ä¸Ģ åı£æ°Ķ +åĭ¤ å¥ĭ +éĻį ä»· +å±ķçݰ äºĨ +å¸ĥ æĭī +ä¼ļ éĢīæĭ© +çļĦ ç»ıåħ¸ +好 æľĭåıĭ +车 éģĵ +æķ´ åĢĭ +åľ ĵ +éķ¿æľŁ 以æĿ¥ +æĬķ å½± +çļĩ åĨł +è¿ĩ 大 +åijĬè¯ī ä»ĸ +ä¼ģä¸ļ æıIJä¾Ľ +æĬ½ 象 +éĢĤ 度 +çļĦ 女åŃ© +èµ· ä¼ı +çļĦ åĬŁæķĪ +ä¸ĵ项 æķ´æ²» +åı¯ éĢļè¿ĩ +ä¸įåIJĮ ç¨ĭ度 +å¼Ĥ è®® +åĩĢ èµĦ产 +åij Ĺ +ä»Ģä¹Ī åij¢ +å·¡ éĢ» +è¸ı ä¸Ĭ +ä½Ĩ å®ĥ +ç²¾ 度 +管 å±Ģ +第ä¸Ģ åIJį +åĨħ åŃĺ +æijĨ åľ¨ +åī© ä¸ĭ +主ä½ĵ 责任 +çĤ¹ åįĬ +以 èĩ³äºİ +åħ»èĢģ ä¿ĿéĻ© +æĦŁåıĹ åΰäºĨ +çŁ¥åIJį çļĦ +å¯Į 豪 +妥 åĸĦ +åŃĻ åŃIJ +éĵ Ĥ +说 èĩªå·± +让 æĤ¨ +æķ° æİ§ +çļĦçľ¼ åħī +注 éĶĢ +çļĦ çģµéŃĤ +è¿ĺ ä¸įéĶĻ +éĹ® ä»ĸ +èĩªä¸» çłĶåıij +èĵ ĭ +ç´« èī² +åĽ½å®¶ å®īåħ¨ +è¾½å®ģ çľģ +ä¹Ł æ¯Ķè¾ĥ +ç¾İ èĤ¡ +ä¸įç¡®å®ļ æĢ§ +å¿ĥ 头 +æĪ ³ +级 åĪ«çļĦ +论 è¿° +çļĦ åĽŀçŃĶ +ä¿Ŀè¯ģ éĩij +çŃī è¡Įä¸ļ +幸ç¦ı æĦŁ +æŃ§ è§Ĩ +æľº 票 +æ´¾ 人 +èĩ´ åij½ +åĺ´ è§Ĵ +æĸ°éĹ» ä¸Ńå¿ĥ +æĶ¾å¼ĥ äºĨ +å®ľ å±ħ +åĨĻ ä¸ĭ +éĹ® çŃĶ +è¿ĻéĩĮ æĺ¯ +å¤ļ åľ° +åĮºåŁŁ åĨħ +åīµ æĸ° +çľĭ ä»ĸ +æī§æ³ķ 人åijĺ +åĬ¨ æľº +éŁ³ åĵį +çļĦ åij½è¿IJ +é¡¶ éĥ¨ +åĵ Ł +éĥ½ æľĥ +æīĵéĢł æĪIJ +æĦı åĽ¾ +çļ ĸ +åĢĴ åħ¥ +å·´ èIJ¨ +åĬ© åѦ +å¤į åı¤ +åIJ¯ ç͍ +åĽ½éĻħ å¸Ĥåľº +åĤ¨ èĥ½ +é»ijé¾Ļæ±Ł çľģ +ä¹ĺ 车 +è¿IJåĬ¨ ä¼ļ +ä¿Ŀ åĪ© +çŁ³ æĿIJ +çµ ® +çĤĴ ä½ľ +çļĦ ä¿¡ä»» +å°± æĪIJäºĨ +åı¯ è§Ĥ +çļĩ ä¸Ĭ +è¿Ļ åĩłå¤© +ä¸Ģ éĶ® +åĨ· åĨ» +ä¿Ŀ åį« +æł¸ æ¡ĥ +åIJĪä½ľ åħ³ç³» +éĢģ åĩº +æĹĹ ä¸ĭçļĦ +åľ¨ ä¹İ +为 广大 +åįĪ é¤IJ +ä¸ĵ 访 +æĪĸ å°Ĩ +éĿĴå²Ľ å¸Ĥ +å¥Ķ è·ij +æĹ¥ æĬ¥éģĵ +å¥ij åIJĪ +æĸ° æĺ¥ +ä¸į å°ıå¿ĥ +两 ä¸ī +æĦıæĢĿ æĺ¯ +åĨ· èĹı +çļĦ çĹĩçĬ¶ +æĢ§ åij½ +è¶ħ æłĩ +å¯Ĩ 碼 +ç§ijæĬĢ èĤ¡ä»½ +äºĨä¸Ģ æī¹ +çĿ£ å¯Ł +åªĴ ä»ĭ +å°Ħ æīĭ +ä¿® åħ» +çīĩ åĪ» +éĢĤåIJĪ èĩªå·± +åıªè¦ģ æĺ¯ +åIJĥ è¿ĩ +éĩij éĵ¶ +缴 å±ŀ +åѦ éĹ® +åİĭ åζ +çªĹ å¤ĸ +æĶ¶ åΰäºĨ +åħ¨åĽ½ 人大 +ä½Ĩæĺ¯ 对äºİ +åľ¨ æķ´ä¸ª +çļĦ èĥĮåIJİ +åĩıå°ij äºĨ +åıį èħIJ +åıįèħIJ åĢ¡ +åıįèħIJåĢ¡ å»ī +æĹ · +åĪĨ æľŁ +åľ¨ æ·±åľ³ +æīĵ çĿĢ +æī« ä¸Ģ +æī«ä¸Ģ æī« +æĶ¿åºľ éĥ¨éŨ +æİ¥ è¿ŀ +å±ŀäºİ èĩªå·± +åŃIJ å¼¹ +åIJĮæł· æĺ¯ +æĢ» åħ± +车 ä¼ģ +æ¢ ĵ +åħ¬ é¡· +åıij 声 +éĴ Ľ +èµ°åĬ¿ åĽ¾ +主 èIJ¥ +åĸ Ķ +æķ°æį® åĪĨæŀIJ +ä¸į è¿ľ +æľī åIJį +æľīåIJį çļĦ +åģ¿ è¿ĺ +å¾Ī ä½İ +è®ĵ 人 +èĿ ī +é«ĺ è´µ +å°ij 许 +æ° Ł +å¹ ¢ +亲 æĥħ +è¿Ļä»¶ äºĭæĥħ +ç͍ é¤IJ +缸åħ³ æĸ°éĹ» +å°± åºĶ该 +ç»Ī çĤ¹ +æĺ¯ å¤ļå°ij +çĻ» åľº +è¯ķ 管 +è¯ķ管 å©´åĦ¿ +åģļ 大 +åģļ大 åģļ强 +çļĦ ä¾ĭåŃIJ +åħ« 个 +æĺİ æĹ¥ +çĤ ³ +èµ° åİ» +éģ º +å¢ © +ä½ĵä¼ļ åΰ +åĴ ı +ä¸ĭ è¾¾ +å¤į åıij +追 éĢIJ +æīĵ åĵį +çļĦ éļ±ç§ģæ¬Ĭ +åħ·æľī ä¸Ģå®ļ +è¿Ļä¹Ī å¤ļå¹´ +æłij æŀĹ +æľĢ éķ¿ +åIJĮ èĥŀ +åħī æ³½ +åŁŁ åIJį +æĮĩ åIJij +åıĹ害 èĢħ +æłij èĦĤ +æľīå¤ļ 大 +大 éĿ¢ç§¯ +æĹł ç¼Ŀ +æĶ¹ æŃ£ +æĽ´å¤ļ çļĦæĺ¯ +æľŁ æľ« +æŃ ¼ +ä¹ī ä¹Į +éĤ£ ä½ł +çļĦ 第ä¸Ģ个 +èĮ µ +å° § +èį « +ä¸įä»ħ åı¯ä»¥ +æ¶Į çݰ +æĢ» éĿ¢ç§¯ +æĸ°éĹ» åıijå¸ĥ +æ°ij ç͍ +å°± 读 +æīĵ è´¥ +å¤ĸ è¯Ń +æĪij们 ä¸Ģèµ· +é¢Ħ å®ļ +çĥ¹ 饪 +æľĢ 主è¦ģ +æľĢ主è¦ģ çļĦ +çīĮ çħ§ +åĽł åħ¶ +ä½İ ä¸ĭ +ä¼ļ åIJĮ +è§ģ è§£ +éĹ´ éļĶ +æķĻ ç¨ĭ +å° ī +å¸Ĥ ä¸Ńå¿ĥ +åħ³éĶ® æĺ¯ +æµ· åįĹçľģ +çī¹åĪ« æĺ¯åľ¨ +ä¸ŃåĽ½ 大éĻĨ +åħħè¶³ çļĦ +æĹ¢ èĥ½ +åĤ³ çµ± +çijľ ä¼½ +åħ¥ åĽ´ +æħ¢æħ¢ åľ° +æĬ¥ éħ¬ +æī¹ å¤į +å·¥ä¸ļ åĽŃåĮº +ä¸İ åıijå±ķ +èĥ¸ éĥ¨ +åľ¨ ç½ij绾 +åľ¨ç½ij绾 ä¸Ĭ +交 è°Ī +æĽ´ æĶ¹ +åįłæľī çİĩ +ä¸Ŀ绸 ä¹ĭè·¯ +è¡ Ľ +çłĶ åΤ +åĪ ª +åĪª éϤ +è¿Ļ åıª +çļĦ æ°Ķæģ¯ +åĬł å·ŀ +éĴ § +çIJĨäºĭ éķ¿ +ä¸ĸ å®¶ +æµģè¡Į çļĦ +å¾Ī æľīåı¯èĥ½ +们 éĥ½ +ç»ıèIJ¥ 模å¼ı +è¡Įä¸ļ ä¸Ń +éĢļçŁ¥ 书 +åij½ é¢ĺ +æľ¬ ç¶²ç«Ļ +æ²Ļ çī¹ +åıij åħī +é«ĺ ä»· +å·² çĦ¶ +åıĮ åįģä¸Ģ +ä¸Ĭ è¯ī +ç¿ħ èĨĢ +è¿Ļä¸Ģ å¹´ +大ä¼ļ ä¸Ĭ +éĩ ī +å®Įåħ¨ æĺ¯ +å¾Ĺ 太 +ä¸Ģèά 人 +è¿ĺ ç®Ĺ +æĬĺ åıł +æĬķ æľº +çĤ¹ çĩĥ +çݰéĩij æµģ +åħĶ åŃIJ +ç½ij æł¼ +æİ¥ è¿ĩ +ä¾Ľ è´§ +éĺ´ å½± +åİŁ åħĪ +æį £ +å·¦ ä¾§ +åħĭ æĭī +æīĵ åį¡ +ç§ij æ¯Ķ +æ±ĩ éĽĨ +åľ°çIJĨ ä½įç½® +è¯Ħ å§Ķ +ç»ĵåIJĪ èµ·æĿ¥ +è¿Ľåħ¥ åΰ +åı¯ è¡Į +åı¯è¡Į æĢ§ +让 å®ĥ +åĪ¶åº¦ æĶ¹éĿ© +çĶĺèĤĥ çľģ +åĵ Ĺ +åģı åģı +è¡£ çī© +ç¥Ŀ è´º +æºIJ èĩª +å¹¶ä¸į 代表 +åĽ½ 度 +好 åĿı +æĿ ĸ +æĿŃ å·ŀå¸Ĥ +湿 度 +é² ¸ +åįļ 彩 +æ³° å±± +æĿij èIJ½ +æĸ° èģŀ +èĤ ĭ +åı¤èĢģ çļĦ +çļĦ ç§ĺå¯Ĩ +ä¸Ģ个 éĹ®é¢ĺ +éģı åζ +åįĥ 亿 +è¿ĩ 硬 +å°Ħ åĩ» +èĩªçĦ¶ æĺ¯ +产 åĮº +çĤ¹ çĤ¹å¤´ +åı¯ä»¥ 帮åĬ© +说 å®ŀ +说å®ŀ è¯Ŀ +æĪij åıªæĺ¯ +ä¹ĭ ä½Ļ +åIJĮæĹ¶ ä¹Łæĺ¯ +ä¸ŃåĽ½ éĺŁ +建æĪIJ åIJİ +ä¹IJ è§Ĩ +åij¨ å²ģ +èᝠåºĹ +éĩij åįİ +严éĩį å½±åĵį +è´¨ åľ° +æĹħ éģĬ +åħµ åύ +æķĻèĤ² æķĻåѦ +离 åİ» +åIJĦå¼ı åIJĦæł· +ä»ĭ ç´ +ä»ĭç´ ¹ +å¼Ģ 头 +å°Ĩ èĩªå·±çļĦ +åIJ¬ åĬĽ +ä¿¡æģ¯ ç³»ç»Ł +ä»İ æł¹æľ¬ +ä»İæł¹æľ¬ ä¸Ĭ +æİĮ 声 +欢 åĸľ +å±ķ åĮº +åķ ¸ +太å¤ļ äºĨ +éĹ² ç½® +èĥ¡ èIJĿåįľ +å§Ķ å®£ä¼ł +å§Ķå®£ä¼ł éĥ¨ +åįĹ éĺ³ +å·ŀ åĮº +ä¸İ æĹ¶ +ä¸İæĹ¶ 俱 +ä¸İæĹ¶ä¿± è¿Ľ +å«Įçĸij 人 +èī¯ å¿ĥ +头 é¡¶ +è´¢ æĬ¥ +ä½Ľ æ³ķ +å¾ µ +åİŁ ä»¶ +åĭ ŀ +çĶ· 篮 +å¤ĸåĽ½ 人 +è¿Ŀ 纪 +æī¾ äºĨ +æįķ æįī +缸 è¯Ĩ +æIJľ éĽĨ +çļĦ ä¼Łå¤§ +ä¸ī ç»´ +å°±è¡Į äºĨ +çĭIJ æľĪ +çĭIJæľĪ å±± +å¸ĮæľĽ éĢļè¿ĩ +èĢĮ 对äºİ +éĿ¢ å°į +åĨĽ åĽ¢ +è¡Ĺ åĮº +æĤ¬ æĮĤ +便 ç§ĺ +æľīä¸Ģ çĤ¹ +ä¼ļè®® ä¸Ĭ +ä¸ĭ æīĭ +廣 åijĬ +äºĶ è¡Į +çŃī åĢĻ +ç´§ç´§ åĽ´ç»ķ +æĭ¿ äºĨ +æ¡Į éĿ¢ +ç¥ŀ æĥħ +éĽĦ åİļ +çŀ ³ +楼 ä¸ĭ +å½ ª +äºĭ åıij +åĨį è§ģ +é¤ ĺ +é¢Ħ åĶ® +åİ» çľĭçľĭ +æĪij们 åºĶ该 +ä¸ī å®¶ +æµ Ĭ +ä¹IJ éĺŁ +çľĭ ä¸įè§ģ +èĦij åŃIJ +æĮģ æľīçļĦ +çϽ èıľ +éĹª çĥģ +åĸĿ æ°´ +æİ§åζ ç³»ç»Ł +ä¸ĵ åĮº +æľĿ å»· +æĪij å¿ĥéĩĮ +å±ķ åİħ +èľĺ èĽĽ +åĨ» ç»ĵ +ç² ª +åº IJ +åIJij 社ä¼ļ +åĨ³çŃĸ éĥ¨ç½² +çŁŃ æľŁåĨħ +æĸ° ä¸ļæĢģ +æľ Ķ +æĹ¶ æĬ¥ +使 ä¹ĭ +åĽł åŃIJ +åıĤä¸İ èĢħ +çļĦ 年轻人 +æīĭ 表 +å°ģ éĶģ +为ä»Ģä¹Ī ä¸į +åIJ¸ çĥŁ +æ¯Ĵ ç´ł +åĪij æ³ķ +磫 æŃ£ +身 æĹģ +åİŁ è°ħ +çĽij æĬ¤ +æŃ¤ å¤Ħ +éļ¨ æĻĤ +æŀľ å®ŀ +åĮ»çĸĹ æľįåĬ¡ +ä¸į åIJĪçIJĨ +æIJŀ 好 +çļĦ èĦļæŃ¥ +å¤ĸ å¥Ĺ +ç¶ĵ éģİ +æĶ¾ ç¼ĵ +åģľ çķĻ +æĺŁ çIJĥ +çļĦä¸Ģ éĿ¢ +åĩł ä½ķ +è½® åĽŀ +æ¯Ľ å·¾ +ä¿® çIJĨ +ä¸įçŁ¥ ä¸į +ä¸įçŁ¥ä¸į è§ī +æķ´ 个人 +æ¯ģ çģŃ +åı° å·ŀ +使ç͍ 寿åij½ +é»ij çϽ +æij¸ ç´¢ +é¼ł æłĩ +éĿ© æĸ° +éº µ +ä¸ĵéŨ 为 +å¾Īå¤ļ æľĭåıĭ +å·¥ä½ľ ç»Ħ +åIJĪ å½± +çĤº ä»Ģ麼 +æŀģ 度 +çļĦ è¿ĽæŃ¥ +å½ĵ ä¹ĭ +å½ĵä¹ĭ æĹł +å½ĵä¹ĭæĹł æĦ§ +è´´ è¿ij +å°º 度 +åľ¨ çİ°åľº +éĻį 临 +åħ»èĢģ éĩij +ç£ ķ +åı¯ä»¥ 使 +管çIJĨ æ°´å¹³ +æľ¬æĬ¥ è®°èĢħ +æ³ķ 令 +åį¡ è½¦ +举 æµ· +å¤ļ éĩį +åħ¶ éĹ´ +ç´ Ļ +éĩį大 é¡¹çĽ® +æ±Ĺ æ°´ +ç»Ħ å§Ķä¼ļ +ä¿¡æģ¯ åħ¬å¼Ģ +ä¸į论 æĺ¯ +ä¸Ģ åIJ¬ +èĴ¸ æ±½ +æıŃ ç§ĺ +è¶ħ éģİ +触 åıij +å© ¦ +åħ³èģĶ äº¤æĺĵ +å°± ç»Ļ大家 +好 ä¹ħ +åĢŁ è´· +游æĪı è§Ĵèī² +å¼ĢåIJ¯ äºĨ +æİ ł +åħļçļĦ åįģä¹Ŀ +ä¸ĭ 鼨 +çŁŃ æĹ¶éĹ´åĨħ +å¯ ħ +导 åħ¥ +å·¥ä½ľ ç»ıéªĮ +ä¹Ł åıªèĥ½ +鼷 éľĨ +è·Ł è¿Ľ +åį¡ éĢļ +é¢ĩ æľī +æľº ä½ĵ +æĪĺ士 èģĮä¸ļ +女 主 +ä½ĵåζ æľºåζ +è¶³ åįı +èĪĴéĢĤ çļĦ +åĢŁ åı£ +æī¹ åΤ +æķ° å̼ +è« ¾ +éĺ¿æĭī 伯 +åĺ İ +æħ ¶ +è¾¾ 人 +å¼Ģ æ°´ +大 鼨 +温 室 +ä½İ è¿· +ä»į æĹ§ +éªĹ åŃIJ +亲 å±ŀ +çIJĨ æĻº +æľ¬ åŁºéĩij +å¨ ħ +åĨĻåŃĹ æ¥¼ +å¢Ļ å£ģ +å® µ +èϽ çĦ¶æĺ¯ +顺 çĿĢ +åħ« åᦠ+åķĨ ç͍ +ä¸į 失 +è¿· èĮ« +顺 便 +æļij æľŁ +欺 è´Ł +é¢ij é¢ij +该 æł¡ +æĸĻ çIJĨ +æ·± æĥħ +åīį éĶĭ +ä¿Ŀ èŃī +èģĮä¸ļ çĶŁæ¶¯ +åħ¬ å¼Ģåıij +åħ¬å¼Ģåıij è¡Į +åħ¥ æĪ· +éł ĵ +å̾ åIJ¬ +éŃ ģ +æĦī æĤ¦ +åĽŀ åIJĪ +åħ¨åĬĽ 以 +åħ¨åĬĽä»¥ èµ´ +åĥ¹ å̼ +èĥ½åĬĽ 强 +ç»ı å¼Ģ +ç»ıå¼Ģ åĮº +è¿ľ æĸ¹ +çļĦ éģĵçIJĨ +缴 åįĩ +缴åįĩ æľº +为主é¢ĺ çļĦ +ç»Ļ æĤ¨ +è¿ĺ æĥ³ +æ¯Ķ æĪij +åĨľ çī§ +æµ· åºķ +çŃ¾è®¢ äºĨ +对äºİ æĪij们 +æĹ¶ 许 +éĶ® çĽĺ +å®ŀéĻħ æİ§åζ +çļĦ æ¨¡æł· +åıįæĺł äºĨ +代 åĬŀ +åĮ» ç͍ +éĽĨ ç»ĵ +åıijå±ķ åīįæĻ¯ +æĮĩ çĿĢ +åįİ åĮĹ +è¿Ļ åĩłä¸ª +åIJį æ°Ķ +åĤį æĻļ +èĩª åıij +æ³¢ åħ° +大åĬĽ æİ¨è¿Ľ +èĩª ç§° +èįĨ å·ŀ +æIJį 害 +äºĨä¸Ģ åı¥ +æľĢåĪĿ çļĦ +éĩijèŀį å᱿ľº +æĢĢ å¿µ +è¡Į åĭķ +女 æİĴ +ä¸į è§£ +ä¼ł éĶĢ +转载 请 +饰 åĵģ +åıª 为 +ä¸İ ä¼Ĺ +ä¸İä¼Ĺ ä¸įåIJĮ +èĥ½ èĢĹ +èı© æıIJ +è¿ij 两年 +è¿Ķ 乡 +马ä¸Ĭ å°± +äºĮ çŃīå¥ĸ +æ°´ 管 +æ³ķ åѦ +çģŃ çģ« +大 å§IJ +åij¨ 转 +æľī æľŁ +æľīæľŁ å¾Ĵ +æľīæľŁå¾Ĵ åĪij +å°į æĸ¹ +ç¥ŀ èī² +æ²¹ èĦĤ +ä¸ī çĤ¹ +ä¸į åĪ©äºİ +äºĭä¸ļ éĥ¨ +å°± è·Ł +å¼Ģ æĶ¯ +å°ı 女åŃ© +åħ±åIJĮ åĬªåĬĽ +çĶļèĩ³ è¿ĺ +è¿Ļ åIJį +è¿Ļ ç¬Ķ +çݯ åį« +æľī ç§į +è§Ĩ åĬĽ +çĨŁ çŁ¥ +åħ¬ç§¯ éĩij +æ¶Īéĺ² å®īåħ¨ +é¢ĩ 为 +大 èħ¿ +éĿ ¶ +çī¹ æķĪ +æľįåĬ¡ åĮº +å¼Ģ åĩº +深度 èŀįåIJĪ +æĹł å¿§ +æŁ¥ éĺħ +ç»Ī ç»ĵ +ä¿Ŀ ç¨İ +è¨İ è«ĸ +å½ĵ åģļ +è·³ èĪŀ +å¯ § +女 çİĭ +è®°èĢħ åľ¨ +åħ¨ 产ä¸ļéĵ¾ +è´¯ éĢļ +åħ´ ä¸ļ +éĻį åΰ +å°ģ éĿ¢ +åħ¨éĿ¢ æİ¨è¿Ľ +奶 èĮ¶ +éĢī åĿĢ +äºĨä¸Ģ åľº +åIJĮ ä¼´ +è®® 论 +æIJ ĵ +诸 èijĽ +诸èijĽ 亮 +å¹² åĺĽ +æµģ æĦŁ +ä¸ĵä¸ļ çŁ¥è¯Ĩ +ç͵ ç«Ļ +åĩı å¼± +åĩº åħ¥ +åIJĦ çľģ +éĿŀ常 é«ĺ +åľ° 毯 +åıij æĸĩ +çĦ ī +çĥ§ çĥ¤ +å£ģ 纸 +æģ¶ åĮĸ +èĬ ¸ +èĥĸ åŃIJ +çĩ Ĵ +çľģ éĴ± +çϾ 强 +çIJĨå·¥ 大åѦ +éĴ¢ æĿIJ +åĽ½æľī èµĦ产 +æĪĺ æľº +æ³Ħ éľ² +åIJİéĿ¢ çļĦ +æ°´ èµĦæºIJ +æ¢ħ èĬ± +åĨĻ çĿĢ +ä¹ĭ 声 +æĹł åı¯ +æĺİ æľĿ +ç«ĭæĸ¹ ç±³ +ç· £ +æĶ¾ è¿ĩ +ç¦ı çͰ +å¾Ĺ ä½ı +åıĹ ä¼Ĺ +ä¸Ń 级 +çĹħ åıĺ +ä¸Ģ çŀ¬éĹ´ +æĿĥ éĩį +人æĢ§ åĮĸ +åĮ»çĸĹ åį«çĶŁ +ä¸įåΰ ä½į +æĻºèĥ½ å®¶å±ħ +饮 ç͍ +æ¼Ķ åıĺ +é«ĺ ç´łè´¨ +ä¹Ļ æĸ¹ +åģľ çķĻåľ¨ +èİ· æī¹ +ç©¿ æ¢Ń +客 åľº +æĮ½ åĽŀ +京 åŁİ +çĶŁåij½ åĬĽ +實 éļĽ +çĩ Ī +åĨį çݰ +çݰå®ŀ ä¸Ń +æľī ä¿¡å¿ĥ +çĸı éĢļ +åĺ´ åĶĩ +鼷 éĶĭ +èıľ åįķ +éħ ¯ +è¶ħ é«ĺ +å¾Ī é«ĺåħ´ +çĶŁ æ®ĸ +éĢł ä»· +误 åĮº +æĨ ĭ +好 æ¶Īæģ¯ +å´ Ń +以 èĩ´ +å¼Ģ çİ©ç¬ij +çĽij è§Ĩ +å·¡ å¯Ł +å¾· å·ŀ +æĹ© æĹ© +éĹª ç͵ +æĪª åĽ¾ +åı¯ä»¥ æł¹æį® +æīĭ èīº +æİ¥ 轨 +ç§į æĹı +æĢĢ éĩĮ +åİ» åĮ»éĻ¢ +ä¸Ģ äºĮ +å¼Ģ éĺĶ +åĩı éĢŁ +ä½Ĩ ä»İ +éĢĻ ä¸Ģ +åĩı åħį +主é¢ĺ æķĻèĤ² +å¼Ģå·¥ 建设 +è¹ ¦ +æľĪ 饼 +ä¸ĭ æ²ī +å°Ĭ 严 +éĻ ĩ +å®ŀ æľ¨ +å»ł åķĨ +声 ç§° +èĢĥ åľº +å¸ĥ é²ģ +èĩª æĿ¥ +èĩªæĿ¥ æ°´ +éĴ ¾ +å¹´ 以ä¸Ĭ +大 åıĶ +ä»ĸ å·²ç»ı +åħ¨ æĿij +èģĶç³» ç͵è¯Ŀ +为 导åIJij +åΤ å¤Ħ +对 éĺµ +缮 æ¨Ļ +åIJį é¢Ŀ +客 æ°Ķ +横 åIJij +çŃī åĨħ容 +åĩł çĤ¹ +è°Ī 论 +ä¸į ä¹ı +å±ķ çݰåĩº +è¾ĥ éķ¿ +éĢĨ 转 +å°ı æĻĤ +æĺ¯ å¤ļä¹Ī +æľ¬ æľĪ +è¿ij è§Ĩ +æĪIJç«ĭ 以æĿ¥ +代表 çĿĢ +æĬ¥ å¤į +æĪı æĽ² +è¨Ń åĤĻ +åħ¥ èĤ¡ +å¾ģ æľį +é«ĺ åĩº +èĪŀåı° ä¸Ĭ +å¿ĥ åĬ¨ +两 çĤ¹ +缸 çķ¶ +èĻ Ľ +主 页 +åĩł å®¶ +æĹł ä¸į +åįı å®ļ +æĸ IJ +å¯ĵ æĦı +åħ¨ 线 +æįķ é±¼ +åı¯ä»¥ ä»İ +æľī è¿Ļæł·çļĦ +æģ¶ éŃĶ +åĮħ åŃIJ +æģ ¤ +å¼Ģå¥ĸ ç»ĵæŀľ +ä¸į æŃ» +èĹ į +弯 æĽ² +æµ· 峡 +éĶĢ æ¯ģ +çļĦ çĭ¬çī¹ +示 æĦı +ä¸įèĥ½ åĨį +èĥ½ æĬĬ +éĺ² çº¿ +ä¸įå°ij äºİ +æ± Ģ +çļĦ éĤ£ä¸Ģ +羣 æĥħ +åŀ ® +被 æīĵ +åĽ½ å®ī +ç¾İ å¦Ļ +è¿Ļ åĩł +åĩº éģĵ +æľįåĬ¡ äºİ +æĪIJæŀľ 转åĮĸ +æīį åįİ +天 é¹ħ +åĩł 个人 +åĢĺ èĭ¥ +è̽ 误 +æĬĹ æĪĺ +è¡Į éĬ· +æĿ¥ è¢Ń +åĢŁ éĮ¢ +èįī èİĵ +ä¸¥æł¼ æī§è¡Į +举è¡Į äºĨ +å¤ĸ ç±į +å·² è¾¾ +æĿij åħļæĶ¯éĥ¨ +è¡ Ŀ +éĻį èĩ³ +æµ· éĩı +é¤IJ é¦Ĩ +æĢ¥ å¿Ļ +æ·± è¿ľ +å¾Ģ è¿Ķ +ç¨İåĬ¡ å±Ģ +å¹¿æ³Ľ åºĶç͍ +è®® åijĺ +æĹł æķĮ +çľ¼ åħī +çĥŃè¡Ģ ä¼łå¥ĩ +æŃ IJ +äºĨ äºĽ +è¿Ŀ èĥĮ +è¿Ļ æĺ¯ä¸Ģç§į +ä¸į 稳å®ļ +大家 åĪĨ享 +表 çı¾ +åīį åįģ +è·¯ è¿ĩ +æĴ © +åIJĮ æĥħ +ä¹ł ä¿Ĺ +åıij è´¢ +åºĶ æľīçļĦ +æĿİ æŁIJ +èĤ Ľ +马 åħĭ +éĢļ åijĬ +å·¨ 人 +ä¸Ģ åĽ¢ +éĢĻ æ¬¡ +ä¸į äºĨè§£ +æĸ½ è¡Į +èij¡èIJĦ çīĻ +åıĺå¾Ĺ æĽ´åĬł +æı £ +åĪĽæĸ° èĥ½åĬĽ +çķħ éĶĢ +表 æī¬ +æ¯Ķ åĪ© +æ¯ĶåĪ© æĹ¶ +åĮ»çĸĹ ä¿ĿéĻ© +æĵį 纵 +伤 亡 +æµİ å®ģ +åıĺ äºĨ +æľ¬æ¬¡ æ´»åĬ¨ +åľŁ 豪 +æĥ³ åĬŀæ³ķ +æĺ ķ +å½ĵ æĻļ +åĩº å±Ģ +çĥŃ è®® +è°Ī è°Ī +æĻĭ åįĩ +åĬ¿ å¿ħ +çĻ» å±± +éĤ£ åĦ¿ +åIJĥ åΰ +ä¹ĭ åŁİ +å¿« æĿ¥ +æ¹Ľ æ±Ł +第ä¸ī 个 +åħ¨éĿ¢ æıIJåįĩ +å¥ĸ åѦ +å¥ĸåѦ éĩij +æĬķåħ¥ 使ç͍ +é½IJ é²ģ +åı¯ä»¥ æĬĬ +åĴĮ ä»ĸçļĦ +è´ŃæĪ¿ èĢħ +æŃ£å¼ı åIJ¯åĬ¨ +åįİ æ¶¦ +ä¸įæĸŃ å®ĮåĸĦ +éĴ¢ æĿ¿ +ç´¯ 积 +满 èĦ¸ +åĽĽ æĸ¹ +è´¢ çī© +ä»ĸ们 ä¼ļ +å¤ı æĹ¥ +éĤ£ 个人 +éĿł çĿĢ +çĤ¹ äºĨ +çĤ¹äºĨ çĤ¹å¤´ +æ© ĭ +åıΠ好 +åıĪ好 åıĪ +åıĪ好åıĪ å¿« +éĺµ éĺµ +å°ģ 建 +æľ¬ çͰ +çī©ä¸ļ æľįåĬ¡ +èĩªè´¸ åĮº +åIJ ı +便åĪ© åºĹ +åĽ½å®¶ æłĩåĩĨ +éĿ¢ ç²ī +èī° è¾Ľ +æĶ» åħ³ +æīĵ åĮħ +车 éĺŁ +人 éĢī +åı¯ ä¸įæĺ¯ +äºĮ åįģå¹´ +åIJį å¸Ī +浦 举 +åħ¬ è¯ģ +è¿IJ éĢģ +æĺ¯ æľĢ好çļĦ +æŁĶ åĴĮ +çİĭ æŁIJ +çĹħ æĪ¿ +åĨ¶ éĩij +ä¸Ģä»¶ äºĭæĥħ +åį ¤ +åı¯ æİ§ +çī Ł +æĭ Ĥ +å·² äºİ +人 éĢł +çĶŁçī© åĮ»èᝠ+ä½ĵ çݰåĩº +èĤ² åĦ¿ +èĢģ å®ŀ +åľĸ çīĩ +è« ¸ +ç´¯ äºĨ +æĦŁåħ´è¶£ çļĦ +åĽ¾çīĩ æĿ¥æºIJ +ä¹Ł æĺ¯ä¸Ģç§į +æ¾İæ¹ĥ æĸ°éĹ» +æĹ¶ 表示 +åħī è¾ī +æĬ¥ åºŁ +å²ģ æĹ¶ +éħ ® +æ£Ģ ä¿® +åıĺ éĢŁ +åıĺéĢŁ ç®± +åľ¨ èģĮ +éı ¡ +æį Ĥ +çĿ£ åĬŀ +æ°¸ ä¸į +åģļ ä¸ĢäºĽ +åİĨ æĹ¶ +å·¥ç¨ĭ æľºæ¢° +æģ° å½ĵ +å°± åľ¨äºİ +ç§° åij¼ +éĢļ常 æĺ¯ +æł· å¼ı +åij¨ ä¸Ģ +èĭ± éķij +åĿĩ 线 +ä¼ł éĹ» +ç͍æĪ· ä½ĵéªĮ +èµŀ åIJĮ +骨 æĬĺ +为主 ä½ĵ +æ±Ł å±± +æ¸ħ æľĿ +æĶĢ åįĩ +ä¸į çĽ¸ä¿¡ +éĿ ´ +æŃ¦ åĬŁ +åĭ¤ åĬ³ +æĿ¥ æī¾ +å°Ĩ æĮģç»Ń +丫 头 +æ¨Ļ æºĸ +è£ ´ +深深 çļĦ +åŃķ èĤ² +è§ĦåĪĴ 建设 +æ¸ħ çν +ç²¾åĩĨ æī¶è´« +æīĵçł´ äºĨ +è¿Ļä¸Ģ 天 +å·¥ä½ľ æĢ»ç»ĵ +æĹħ ç¨ĭ +举 èIJ¥ +æĶ¾ å°Ħ +æľī åĩłä¸ª +éĿŀ çī©è´¨ +åIJĥ å¾Ĺ +åĹ ¨ +ä¼ļ åıijçĶŁ +篮 æĿ¿ +å¼Ģ å°ģ +麻 å°Ĩ +èıı æ³½ +ä¸į åIJĪ +ç³»åĪĹ äº§åĵģ +èѬ å¦Ĥ +ç¾İ èªī +èĩªå·± åĸľæ¬¢ +交æĺĵ ä¸Ńå¿ĥ +åIJĪ åͱ +使 æĪij +åĥı ç´ł +带 éĺŁ +ä½Ĩ 对äºİ +æĬĬ è¿Ļ个 +èĤĿ èĦı +åįķ纯 çļĦ +æĶ»åĿļ æĪĺ +缼 ä¼ļ +åijµ æĬ¤ +æª Ģ +èµ¶ ä¸Ĭ +æ¥ Ĭ +ä¹ħ äºĨ +ç¡ Ŀ +çŃĶ é¢ĺ +ä¿ĿæĮģ çĿĢ +è§ģ è¯Ĩ +çĤ¹ åĦ¿ +åįĬ 个æľĪ +æ» ĩ +浸 泡 +ä¼ł éĢģ +åľ¨ å¸Ĥåľºä¸Ĭ +ä¹ĭ 乡 +çī¹ éķ¿ +éĽ ŀ +èª ł +身 å¤Ħ +æŁł 檬 +身 ç©¿ +çľģ åħ¬å®ī +çľģåħ¬å®ī åİħ +åıĻ åĪ©äºļ +åĩł åĪĨéĴŁ +人 åĢij +åľ° 段 +èĩª åѦ +ä¹Ł è¶ĬæĿ¥è¶Ĭ +èģĮ æĿĥ +æĸ § +èĩ » +å½Ĵ 纳 +驾 é©Ń +éĥ¨åĪĨ åľ°åĮº +没æľī æĥ³åΰ +æĴ ĩ +ä¹Į é²ģ +ä¹Įé²ģ æľ¨ +ä¹Įé²ģæľ¨ é½IJ +èĤ² 人 +çļĦ æŃ¥ä¼IJ +å»¶ æľŁ +æ²¹ æ°Ķ +åģļ å®Į +åľ£ åľ° +丰 åİļ +宽 带 +åı¯éĿł çļĦ +åºŃ éĻ¢ +åŃ ľ +å°ı康 社ä¼ļ +å®īåħ¨ 管çIJĨ +å¹´ 第 +æİĴ 污 +èĥĮ åĮħ +å®¶ ä½ı +åħ¶å®ŀ å°±æĺ¯ +ä¼ļ è§ģ +帮åĬ© ä¼ģä¸ļ +ç½ij è´Ń +æĺ¯ ä¸įä¼ļ +飯 åºĹ +æŃ» åİ» +åħįçĸ« åĬĽ +æľ ķ +åĸĿ äºĨ +è½» å¾® +个æľĪ åĨħ +ç»Ħ åĽ¢ +åĴĮ å®ĮåĸĦ +é¸ ½ +æıIJ éĢŁ +西å®ī å¸Ĥ +ä¸Ńå¿ĥ 主任 +æĹ¶éĹ´ 为 +æľŁ æĿĥ +è¶ ķ +ä¸įä»ħ è¦ģ +æľį ä»İ +é¡ĺ æĦı +ä¸į å°ı +ä¸įå°ı çļĦ +ç° ĩ +çª ¦ +åĪĩ æĪIJ +åĵĪ åĪ© +天 羣 +ä¸Ģ次 次 +éĩij å¸ģ +æĢİä¹Ī èĥ½ +ç½ij è´· +ä¼ļ计 å¸Ī +çŁŃ 缺 +对 æłĩ +åıĺå¾Ĺ æĽ´ +åīį åĩłå¤© +éĺ² æ±Ľ +彩 èϹ +åĵģ ä½į +表 æł¼ +严 å¯Ĩ +æ¯Ľ åĪ©çİĩ +çļĦ åį±å®³ +å½ķ åζ +æ°´ åĬ¡ +èĥ½å¤Ł 让 +å¹³ æĿ¿ +ä¹³ æĪ¿ +è¸ı å®ŀ +é¦ĸ åĪĽ +é¦Ļ èķī +æĬ¥ 表 +ä¸Ģ æĬ¹ +åĩºçĶŁ äºİ +è²» ç͍ +åĩº 让 +åIJĪæ³ķ æĢ§ +å°¼ åħĭ +åĨ° åĨ· +é¦Ļ æ°Ķ +åı· ç§° +èµ· çłģ +åŁİ åİ¿ +çİ© èĢį +ä¸Ĭ éĻIJ +ä¼ļè®® ç²¾ç¥ŀ +æĹģè¾¹ çļĦ +便 ä¼ļ +æıŃ æĻĵ +çİ© æĦı +éĽª å±± +åIJij çĿĢ +ä½ĵèĤ² åľ¨çº¿ +说æĺİ ä¹¦ +åĮĸ èĤ¥ +åħļç»Ħ 书记 +åĬ¨ 人 +ä¹ĭ æīĢ +æľĪ èĩ³ +æľĢå¿« çļĦ +èĬĤ åģĩæĹ¥ +ä¸ĵ åľº +èĢĥ ä¸Ĭ +çª Ł +é²ľ è¡Ģ +è¾ĥ强 çļĦ +æĤĦ çĦ¶ +å¤ļ个 åĽ½å®¶ +çªĹ å¸ĺ +æŀģ å¤§åľ° +ä¸įç͍ æĭħå¿ĥ +è¿Ļä¹Ī åģļ +åĥ¹ æł¼ +ç¾İ丽 乡æĿij +å°ıæĹ¶ åĨħ +ç´§ è¿« +大 çģ« +èĥ³ èĨĬ +æĵįä½ľ ç³»ç»Ł +æ®ĭ çķĻ +åĨĻ åĩº +ç¦ģ å¿Į +åĬłçĽŁ åºĹ +è¿ij çϾ +便 åı¯ +æķ´æĶ¹ æİªæĸ½ +éĩĩ访 æĹ¶ +åĶIJ 代 +æ·±åĮĸ æĶ¹éĿ© +çŁ ¢ +éĥ½ åĸľæ¬¢ +è¶ĬæĿ¥è¶Ĭ é«ĺ +èĬ± æľµ +头 çĸ¼ +å®ī 康 +å¢ŀéķ¿ çİĩ +çľ¼ çľĭ +å°±æĺ¯ 为äºĨ +èĢĮ 导èĩ´ +åĬłå¿« 建设 +èĬ± æł· +åĨħå¿ĥ çļĦ +æĺĨ å±± +è³ĩ æºIJ +åĽŀåΰ å®¶ +èıĬ èĬ± +æ°´ éĩı +å¾ģ ä¿¡ +è¡ĮæĶ¿ åĮº +ä¹ĥ æĺ¯ +æĬķèµĦ é¡¹çĽ® +å«ģ ç»Ļ +ç¥ŀ åľ£ +ç¨ ł +æľ¬æĿ¥ å°± +éĢIJ ä¸Ģ +èģĮä¸ļ æĬĢæľ¯ +ä¸įèī¯ ä¿¡æģ¯ +æīĺ è¿IJ +åIJ¯ 示 +ä¹ĭ åħ§å®¹ +éŁ ¶ +奢 åįİ +æıŃ ç¤º +æĪIJ为 ä¸ŃåĽ½ +æ¶Īè´¹ åĵģ +åħ¬ ç͍ +æIJŀ å®ļ +请 ä½ł +æŁ ļ +åĨħ è¡£ +ä½Ĩ ä»ĸ们 +ä¿Ŀ 湿 +该 åİ¿ +饱 åĴĮ +æİ¨ åIJij +èµĦæĸĻ æĺ¾ç¤º +ä¸į å½±åĵį +人 人éĥ½ +åıijå±ķ 壮大 +åħ»èĢģ æľįåĬ¡ +çĶŁæ´» æ°´å¹³ +åIJĦ åİ¿ +ä½ł éľĢè¦ģ +说 çļĦæĺ¯ +å¤ĸ åªĴ +æŃ¤ 人 +次 è¦ģ +追 èµ¶ +åºĶ该 å¦Ĥä½ķ +æĹ¥ åĩĮæĻ¨ +çķ¥ æľī +éĥ½ æĥ³ +游 ä¹IJ +è¿Ļ款 游æĪı +å¹³ æ·¡ +æĺ¯ä¸Ģ åĢĭ +å¤ĩ èĢĥ +åζ æŃ¢ +ä¸Ģå®ļ èĥ½ +å¾Ĵ å¼Ł +以 çĤº +åįĥ åħĥ +äºĶ åħŃ +迪 士 +迪士 å°¼ +éĺ³ æĢ§ +åĨ¬å¥¥ ä¼ļ +å°±æĺ¯ åĽłä¸º +æĮĤ éĴ© +æ¦Ĥ åĨµ +åıªè¦ģ æľī +æ²¹ çĶ» +åľ° æłĩ +ä¸Ĭ è°ĥ +产ä¸ļ åĽŃåĮº +åħ« åįģ +æ£ ± +æ¶² æĻ¶ +æĿij å§Ķä¼ļ +çŃ¾çº¦ 仪å¼ı +è¿Ļ åħ¶ä¸Ń +åĨĻ éģĵ +示èĮĥ åŁºåľ° +éĩİçĶŁ åĬ¨çī© +鼻åŃIJ ä¿¡ç®± +åĽ½éĻħ è´¸æĺĵ +人 æĿĥ +ä¿Ŀ 管 +èĭ¥ æĤ¨ +åİĭ æĬij +é» Ľ +åľ° çľĭçĿĢ +éĻ ° +ä¸Ģå¹´ å¤ļ +ä»İ 容 +ä¸Ń æĸŃ +å¯Ł è§ī +ç§» 交 +éĶ ¯ +æĪĸ许 æĺ¯ +ç¶ ł +两 项 +æľĢ åĸľæ¬¢ +æľĢåĸľæ¬¢ çļĦ +å¤ľ éĩĮ +åIJĮ ä»ģ +åĪĽæĸ° 驱åĬ¨ +è°ģ èĥ½ +é£ ¾ +åħī åѦ +åİ Ħ +èĦ± é¢ĸ +èĦ±é¢ĸ èĢĮåĩº +è¿ ¦ +æĺ¯ ä¸įåı¯èĥ½ +çª ¥ +èĥ½ 满足 +宽 度 +伦 çIJĨ +åı¯ä»¥ èİ·å¾Ĺ +转 ä¼ļ +å±± æĿij +éĵº 设 +åĩº åĩ» +æĸĩåĮĸ èīºæľ¯ +ä¼ļè®® 室 +æŃĮ 声 +æ» Ķ +èIJİ ç¼© +æľįåĬ¡ åijĺ +åıij表 äºĨ +æĸ¼ æĺ¯ +æĺİç¡® è§Ħå®ļ +ç»´ å¥ĩ +æ°´ 产 +æĬķ ä¿Ŀ +éĺ´ éģĵ +èµ¶ å¿« +夺 å¾Ĺ +ä¸ĭ åįķ +çµģ åħ¬åı¸ +çݯ ç»ķ +å½ Ī +ä½ľé£İ 建设 +æĹħ游 æĻ¯åĮº +æľī æĽ´å¤ļçļĦ +丰å¯Į å¤ļ彩 +çIJĨè´¢ 产åĵģ +åĩº å·® +ä»İ严 æ²» +ä»İ严治 åħļ +缸 å¹² +æ»ĭ 润 +主åĬŀ æĸ¹ +åī§ åľº +æ»ļ çIJĥ +æ©Ħ æ¦Ħ +èĩªä¸» åĪĽæĸ° +éĢļ å¾Ģ +æł¼ å°Ķ +çļĦ ä¼ĺçĤ¹ +èĥĮ ä¸Ĭ +çª ľ +çĪĨ åĩº +å¹³ æķ´ +ä¸Ģ èĦļ +åħ¨ä½ĵ åijĺå·¥ +éĻIJ å®ļ +åŁİéķĩ åĮĸ +æ· ³ +éĢ® æįķ +è¡ĮåĬ¨ 计åĪĴ +æīĵ å¾Ĺ +åİļ éĩį +纪å½ķ çīĩ +åĿļ ä¿¡ +央 ä¼ģ +åĨį ä¹Łä¸į +天 涯 +åıĤèĢĥ èµĦæĸĻ +æľī æ¯Ĵ +åIJ¸ 纳 +è¶Ĭ åıij +éĩįè¦ģ æĦıä¹ī +åĽ½éĺ² éĥ¨ +è¿Ļ个 è¡Įä¸ļ +æĻ® æŁ¥ +å¼Ĥ æĢ§ +å»¶ è¿Ł +å°ı å¹ħ +èī² æĥħ +综åIJĪ æ²»çIJĨ +æŃ£æĺ¯ åĽłä¸º +产ä¸ļ ç»ĵæŀĦ +çłĶç©¶ æĬ¥åijĬ +åģľ ä¸ĭ +éķ¿ èĢģ +éĩĿ å°į +åįĹ京 å¸Ĥ +çģĮ æºī +转 è¿IJ +欺 è¯Ī +éĢł åģĩ +åĪĨå¸ĥ å¼ı +æĦŁ è§¦ +æĪij å½ĵæĹ¶ +åıij è§ī +åĽ¾ 纸 +æĶ¹ èī¯ +çĭł çĭł +åĨ² åĪº +æĸ° 京 +æĸ°äº¬ æĬ¥ +ç¥ŀ åύ +秸 ç§Ĩ +çĪ º +å°Ĩ è¿İæĿ¥ +å·¥ ä¿¡ +工信 éĥ¨ +éĻIJ éĩı +æŃ¢ æįŁ +åѦä¼ļ äºĨ +åįİ çĽĽ +åįİ缼 é¡¿ +å¾Į ä¾Ĩ +ä¸ĭéĿ¢ æĺ¯ +ä¸ĭéĿ¢æĺ¯ å°ı +æIJ¬ è¿IJ +ç¾İæľ¯ é¦Ĩ +æ¸ħ åĩī +å¤ļå¹´ åīį +è© ŀ +åįĥ ç±³ +表 è¿° +æ±Ł éŨ +åĬłæ²¹ ç«Ļ +æľ¬ èĥ½ +导 读 +åĽ´ è§Ĥ +å¹¶ åIJij +åŁºæľ¬ æĥħåĨµ +æīĵ å¼ĢäºĨ +è¿Ļ ä¸ī个 +æ±ķ 头 +强 æľīåĬĽ +强æľīåĬĽ çļĦ +è¿Ľ åľº +ä¹Ŀ æ±Ł +çIJĥ æĺŁ +好çľĭ çļĦ +大 æĪ· +æ¹ ¯ +å¥ĩ å¦Ļ +ä¹IJ åύ +æĪijçļĦ å¿ĥ +çľī 头 +åĨľä¸ļ çĶŁäº§ +ç¼ĸ çłģ +åŁº ç¤ +åŁºç¤ İ +天 æĸĩ +åĢĭ人 è³ĩè¨Ĭ +åİ» è¿ĩ +èģĨ åIJ¬ +æĶ¾ åģĩ +ä¸į åħ·å¤ĩ +æ·Ģ ç²ī +大 佬 +åħ¨ 天 +åħ¨éĿ¢ 建æĪIJ +éļIJ å½¢ +ç¼ħ ç͏ +åIJ ³ +è¡ĮæĶ¿ æī§æ³ķ +åŁİ åł¡ +èİ« æĸ¯ +èİ«æĸ¯ ç§ij +æīĢæľī æĿĥ +éĽĨ åľĺ +å±Ģ åī¯å±Ģéķ¿ +åĩłä¹İ 没æľī +æ´ģ åĩĢ +ç͵影 èĬĤ +åŃ© ç«¥ +æīĢ åģļçļĦ +æ¸ħ 代 +æĸ° çīĪ +éĵĿ åIJĪéĩij +为 æĬĵ +为æĬĵ æīĭ +åΤ å®ļ +çī¹ äº§ +æīĭ æ©Ł +ä¸įåı¯ æĪĸ +ä¸įåı¯æĪĸ 缺 +å¸Ĥåľº è§Ħ模 +åĿ ¯ +åĮ» åѦéĻ¢ +å¿« è¦ģ +èĮ ľ +æĬĺ èħ¾ +äºĨ è¿ĩæĿ¥ +æĬ¥åijĬ æľŁåĨħ +çī© ç§į +ç»Łè®¡ å±Ģ +æī© 建 +æ¶ ħ +责任 人 +éĺ İ +è¯Ħ è®® +å¾Ģ äºĭ +æīĢ ç¤º +æķ´ æ´ģ +éĹº èľľ +æĹħ éĢĶ +å®ŀ è®Ń +ä¹ĭ ç§° +å·´ 士 +éĢŁåº¦ å¿« +ä¸įä»ħ å¦ĤæŃ¤ +å®Ŀè´µ çļĦ +åºŁ çī© +æ²³ æ°´ +æİ¥ 纳 +ç²¾ æ¹Ľ +åħ¶æ¬¡ æĺ¯ +顺 å¾· +åħ¬åħ± åį«çĶŁ +è¤IJ èī² +ä¸į æĥľ +æĬĢæľ¯ æľįåĬ¡ +æİ · +æ±Ĥ èģĮ +ä¸ī 峡 +æĬķåħ¥ åΰ +太 åIJİ +åIJ¯åĬ¨ 仪å¼ı +缴æİ¥ å½±åĵį +æĸ° 款 +个 乡éķĩ +çϾ 亿 +åº « +ä¹Ł æŃ£æĺ¯ +åı¶ çīĩ +æľĢæĹ© çļĦ +æĪĺ 绩 +å·¥ æľŁ +æĻļ æľŁ +è¿Ļæł· 说 +è¯į è¯Ń +ä¾ Ħ +æķ£ çĥŃ +éĽĨæĪIJ çĶµè·¯ +åIJį è¯į +æĻº åķĨ +æĭ¥ åłµ +çĭĤ 欢 +è¿Ļ èά +æµ´ 室 +åijķ åIJIJ +æľªæĿ¥ åıijå±ķ +ä¸īä½į ä¸Ģä½ĵ +åªĴ é«Ķ +ä¸įå¾Ĺ 转载 +åĽłä¸º 她 +æĺ¾ç¤º å±ı +ä¾Ľ æļĸ +éĨ« éĻ¢ +æľī æĦıæĢĿ +æľīæĦıæĢĿ çļĦ +娱ä¹IJ åŁİ +åįµ å·¢ +åĪĽéĢł åĬĽ +竳 èĬĤ +人大 常å§Ķ +èĢĮ çİ°åľ¨ +å¤ĸ å©Ĩ +å¢ŀ æĮģ +äºĶ åįĥ +èĢģå¸Ī 们 +æ´Ľ æĿī +æ´ĽæĿī 磶 +æİĮæı¡ äºĨ +ä¸ŃåĽ½ æĸĩåĮĸ +æĸ° æĶ¿ +主è¦ģ ç͍äºİ +åıij çĥ§ +类似 äºİ +åĮĹ æŀģ +æĪij们 认为 +å¼¥ 漫 +åħ¨çIJĥ ç»ıæµİ +é¢ IJ +ä¸Ģèµ· è£ħä¿® +æĶ Ĵ +æĭī èIJ¨ +帶 ä¾Ĩ +åĨ· æ°´ +ä¸ī åĨľ +æĿ¿ æĿIJ +è¿ŀ è¿ŀ +éĵ ® +ç»ıèIJ¥ çIJĨ念 +å±± é¡¶ +å¾Ī æĥ³ +çĺ « +å§ĭç»Ī ä¿ĿæĮģ +åľ¨ 广å·ŀ +ä¸įåIJĮ æĦı +åıĺ åİĭ +åıĺåİĭ åύ +产 éĶĢ +表 éĿ¢ä¸Ĭ +æīĢ以 ä»ĸ +ç»ıéªĮ 丰å¯Į +éĥ¨ å§Ķ +åħµ åĽ¢ +æīĢ è¿° +æķ¦ çħĮ +ç»ıèIJ¥ èĮĥåĽ´ +åı£ è¯Ń +失 ä¿¡ +æ¯ı个人 çļĦ +æīĭ æĮģ +æģIJ æħĮ +åł¡ åŀĴ +é¦ ħ +éĵ¸ éĢł +æĭ¿ åĩºæĿ¥ +æİ¢ æµĭ +大家 ä¸Ģèµ· +å¥ § +å®ŀè´¨ æĢ§ +å°ı åĦ¿ +èĩº åįĹ +èĩºåįĹ å¸Ĥ +å¼Ģåıij èĢħ +åı¯ æł¹æį® +ç®± åŃIJ +饺 åŃIJ +å¿Ļ çĿĢ +æĿ¥ ä¸įåıĬ +缸 ä¼ł +åĽ½ ç½ij +èħ¹ æ³» +è¿ĻéĩĮ æľī +é£İ æĻ¯åĮº +åıĤ ä¿Ŀ +æŃ» èĢħ +æĪ´ ä¸Ĭ +æ©Ł æ§ĭ +è¯ķéªĮ åĮº +ä¼ł æİĪ +æµ· è¾¹ +泪 æ°´ +缸åħ³ åĨħ容 +éĥij å·ŀå¸Ĥ +åħij çݰ +两 åij¨ +èĬľ æ¹ĸ +ç͵åŃIJ ä¿¡æģ¯ +红 å¤ĸ +æĹħ游 å±Ģ +å¾Ģå¾Ģ ä¼ļ +è¿ħ çĮĽ +ä¼ł 羣 +æ¸ħ æ¾Ī +å°± è¿ij +微信 群 +ç³»åĪĹ æ´»åĬ¨ +ç»ı常 ä¼ļ +è§Ĥ æµĭ +å¿ĥå¾Ĺ ä½ĵä¼ļ +éĻĪ åĪĹ +åĮĹ æĸĹ +è« ® +è«® è©¢ +è¿ĺæĺ¯ ä¼ļ +æµĭ ç®Ĺ +æĺŁ ç©º +宽 容 +çī©ä¸ļ åħ¬åı¸ +æĪĴ æĮĩ +å¸ħ æ°Ķ +ä¸ĢæŃ¥ æŃ¥ +åħ± 鸣 +åĨ³ ä¸į +æİ¥ 管 +å¦ĩ èģĶ +æ¯Ķ åĸ» +é²ģ è¿ħ +æĮģ çºĮ +缸 亲 +å¨ģå°¼æĸ¯ 人 +ç«ĭ 项 +åĪ Ŀå§ĭ +èĩª åζ +è¿Ī è¿Ľ +ä¸Ĭ æ±½ +å®ı ä¼Ł +æł¹æľ¬ 没æľī +æĸ°åĨł çĹħæ¯Ĵ +åĵª ç§į +康 åħ» +è¡° èĢģ +å½ķ åĥı +é«Ķ é©Ĺ +ç»ij å®ļ +é¢Ŀ 头 +äºĶ æľĪ +èĬ± å¼Ģ +ä¸Ģ线 åŁİå¸Ĥ +åΰ åľº +æĬķ éĻį +çĹĺ çĹĺ +åıĹ ä¸įäºĨ +æīİ æł¹ +æĽ´ ä½ķåĨµ +æĬ½ æŁ¥ +åĩº è·¯ +审议 éĢļè¿ĩ +ä¸į åĥħ +èī² è°ĥ +çϾ ä½Ļ +èĤł éģĵ +æ·±åİļ çļĦ +马 åĬĽ +æĹ© æĻļ +æŃĮ èĪŀ +éĺ² æĻĴ +æľĢåIJİ ä¸Ģ个 +樱 èĬ± +å°ıä¼Ļ åŃIJ +åľ¨ å½ĵåľ° +å°ıä¼Ļä¼´ 们 +èµ· æºIJ +åħ¨ åªĴä½ĵ +ç° ½ +éħ± æ²¹ +æĹłè®º å¦Ĥä½ķ +裤 åŃIJ +åģľ äº§ +ä¸įçͱ å¾Ĺ +çīµ å¼ķ +ä¼ł åĬ¨ +ä¹Ŀ é¾Ļ +åĬł åĽº +ä¹Łä¸į æķ¢ +æĬĢæľ¯ æĶ¯æĮģ +ä¸Ĭ å²Ĺ +ç»ıéªĮ åĴĮ +æł¼ æŀĹ +åIJ¸ éĻĦ +æľªæĪIJ å¹´ +奢ä¾Ī åĵģ +追 æį§ +好 ä¸į容æĺĵ +èķ´ åIJ« +ä¿Ŀ å®ļ +æĬ¥ ä¸ļ +æµ· åĨħå¤ĸ +ä½ł çİ°åľ¨ +æ²¹ èĢĹ +è´¨éĩı 管çIJĨ +æ½ľ æ°´ +丽 æ±Ł +转 åħ¥ +è¿Ļä¹Ī ä¹ħ +æĺİ ä»£ +责任 åζ +éĩį å·¥ +大 å·´ +触 åıĬ +èµ· åĪĿ +大 å¦Ī +æĸ¯ å¡Ķ +åĨĽ å·¥ +书 éĻ¢ +å³ ¨ +æİ¨ çIJĨ +è¿Ļç¯ĩ æĸĩ竳 +è¿ģ ç§» +åľ¨ åIJĮä¸Ģ +ç»Ĩ ç»Ĩ +åīĬ å¼± +书 æĪ¿ +ç¶ĵ 常 +è¯ķ é¢ĺ +æĤ£ ä¸Ĭ +çĻ«çĹ« çĹħ +åĨ² æ´Ĺ +å¤ĸ æı´ +åħĭ åζ +åįģ æľĪ +åģļ ä¸įåΰ +ç¾İ åĮĸ +å¦Ĥ æľŁ +è¿ĺ éľĢ +天 åºľ +å°± æĦıåij³çĿĢ +çļĦç¡® æĺ¯ +éªĹ å±Ģ +å°ıç»Ħ èµĽ +è© © +ä¹Ŀ å¹´ +æĻĵ å¾Ĺ +çłĶç©¶ 人åijĺ +大 éħĴåºĹ +ç§ij åѸ +åħŃ åIJĪ +çķĮ å®ļ +车 è½½ +å¼Ģ çĿĢ +毫 æĹłçĸij +毫æĹłçĸij éĹ® +è¿IJ ç»´ +ç¦ģ åĮº +èĦ± èIJ½ +讲 å¸Ī +产ä¸ļ åŁºåľ° +é«ĺ æĢ§èĥ½ +åħī 彩 +çݰ éĺ¶æ®µ +åĩ ¿ +è¾ĥ å·® +饮 çĶ¨æ°´ +éĸĭ çϼ +ç½ij åIJ§ +çĮ´ åŃIJ +æŃ¦ æŀĹ +å®ī åİ¿ +ä¸įåı¯ æĢĿ +ä¸įåı¯æĢĿ è®® +éĬ· åĶ® +è´« ç©· +为 åķ¥ +éº ĵ +å¹¾ åĢĭ +è§Ħ模 以ä¸Ĭ +æı ļ +被 åĽ° +缺 å¸Ń +å¿« é¤IJ +æĬ¢ åįł +æĻ Ł +å¤į æ´» +æľ¬æĬ¥ 讯 +åĪĽ ä¸ĭ +æµ· 滩 +éĩı 产 +å¦Ĥä½ķ åİ» +车 ä½į +å¯ ĩ +äºĮ åįģåĽĽ +ç»ıæµİ æįŁå¤± +éħįå¥Ĺ 设æĸ½ +åŁºæľ¬ éĿ¢ +äºī 论 +就好 åĥı +çłĶç©¶ æĪIJæŀľ +éĻĪ è¿° +æīĵ åĬ¨ +ä¸ĭ å·´ +ç§Ĵ éĴŁ +对 人ä½ĵ +æĬĢæľ¯ çłĶåıij +åİŁ åŃIJ +æĺ¯ä¸Ģ 项 +äºĨä¸Ģ 份 +æĮĩ çͲ +ç͍ éĩı +è¿ĺä¸į å¤Ł +æĶ¿åºľ éĩĩè´Ń +çŁ¥è¯Ĩ çĤ¹ +ä¸ŃåĽ½ 梦 +å¾Ī å¼Ģå¿ĥ +礼 è²Į +éĿŀ常 å¤ļ +éĿŀ常å¤ļ çļĦ +åĽ ļ +æĹħ é¦Ĩ +å°½ æĥħ +æŃĮ åͱ +æ²Ļ é¾Ļ +车 åİ¢ +客 æµģ +åģı å·® +积累 äºĨ +æ¡ Ķ +çĶ» çĶ» +ä¹Ł åºĶ该 +åºĶç͍ ç¨ĭåºı +èĥĥ èĤł +以 å¾Į +豪 å®ħ +æ·± åĬłå·¥ +缴 è¨Ģ +åĮĸ çŁ³ +åĽ½ éģĵ +ä¸ĥ 个 +ä»İèĢĮ 使 +èĤł èĥĥ +æĹ¥ è¶ĭ +çζ åŃIJ +ç· © +æĭĽ çīĮ +产 å¦ĩ +çķª èĮĦ +æĪij éĻ¢ +建çŃij å·¥ç¨ĭ +å±ķè§Ī ä¼ļ +å®¶éķ¿ ä»¬ +åĨľ ä½ľçī© +æĹ¥ å¤ľ +æĶ» æĵĬ +è§Ħ éģ¿ +èĪŁ å±± +便 æ°ij +åħ« åŃĹ +ä¸į æĽ¾ +æĶ¯ éħį +çĨ¬ å¤ľ +人 é¡ŀ +ç´Ģ éĮĦ +ç»ıèIJ¥ æ´»åĬ¨ +大 涨 +å¸Ĥå§Ķ 常å§Ķ +åĪĨ éIJĺ +ä¸Ģ个 èģĮä¸ļ +çĹħ åĽł +è¿Ļ 对äºİ +ä¸įå¾Ĺä¸į 说 +åıijç͵ æľº +æľīæīĢ å¸®åĬ© +缮æłĩ ä»»åĬ¡ +åĽł åľ° +åĽłåľ° åζ +åĽłåľ°åζ å®ľ +å°Ĩ è¾¾åΰ +ç²Ĺ ç³Ļ +稳 åĽº +å« £ +çİ°åľ¨ å¾Īå¤ļ +ä¸ĸçķĮ 级 +å¼ł æŁIJ +çĤ¹ ç¼Ģ +èij µ +社ä¼ļ ç»Ħç»ĩ +å¾Ģ åIJİ +åĬł æģ¯ +åĻª 声 +æľī åħ´è¶£ +为æĤ¨ æıIJä¾Ľ +æ²¹ æ¼Ĩ +ç¬¬åĽĽ å±Ĭ +çļĩ 宫 +ä¹Ĵ ä¹ĵ +ä¹Ĵä¹ĵ çIJĥ +éļ¨ èijĹ +éģ© åIJĪ +åįĹ éĿŀ +æĵ ´ +西 æ´ĭ +åĬł å¯Ĩ +æĪIJåĬ٠䏾åĬŀ +åı£ æ°´ +æĪIJ 年人 +æīĢ æıIJä¾ĽçļĦ +éļĶ å£ģ +åľ¨ 京 +å½ĵåľ° æĹ¶éĹ´ +çŃī åIJĦç§į +é£İ æ°Ķ +å±ĭ éĩĮ +ä¸Ģ åŃĹ +çļĦæĹ¶éĹ´ éĩĮ +åĺ¿ åĺ¿ +å¿« 讯 +ä¸Ń åľº +ä¸Ģ çĵ¶ +æ» ķ +é¢Ĩ è·ij +好 èݱ +好èݱ åĿŀ +没 åħ³ç³» +åĩº å¢ĥ +ä¸įæĺ¯ ä¸Ģ个 +éĥ½æĺ¯ éĿŀ常 +éľĩ åĬ¨ +èİ· èĥľ +åįļ å¼Ī +æĬļ åħ» +对 ç«ĭ +æľįåĬ¡ æľºæŀĦ +è°£ è¨Ģ +社ä¼ļ ç§ijåѦ +åIJ¬è¯´ è¿ĩ +æī ³ +æīĵ 磨 +åı£ æľį +好 åĥıæĺ¯ +以åıĬ åħ¶ä»ĸ +çī¹ è´¨ +亲 è¿ij +ä¸Ģ ç»ı +æ¶ Ŀ +éŃĶ æľ¯ +éģĵè·¯ 交éĢļ +è§Ħ模 æľĢ大 +å®ŀæĸ½ æĦıè§ģ +ä¹ ŀ +ä¸Ģ ä¸ĸ +åŁ· è¡Į +è±Ĩ çĵ£ +åĪĹ ä¸º +æķħ 宫 +çĶŁ åij½åij¨æľŁ +ä¸īç§į èģĮä¸ļ +详ç»Ĩ ä»ĭç»į +å®Į å¤ĩ +岩 çŁ³ +éļı æīĭ +é£ ² +æķĪæŀľ åĽ¾ +ç§ĭ åĨ¬ +åĬŁ å¾· +è§Ħ竳 åĪ¶åº¦ +æĹ¥ æ¸IJ +æīĢ éľĢè¦ģ +æīĢéľĢè¦ģ çļĦ +å²Ľ ä¸Ĭ +åĩº åľŁ +åĽ¾ æĸĩ +ç§ijæĬĢ è¿ĽæŃ¥ +éĢļ èĥĢ +èĢģ 太太 +èĭĹ æľ¨ +éĵ¶ å·Ŀ +å¸IJ 篷 +éĿŀ è¦ģ +éħį ç͵ +å¤Ħ å¢ĥ +èĤ¡æĿĥ æĬķèµĦ +ä¸Ģ缴 åΰ +åĿĩ çͱ +æĬĹ æĹ¥ +æį® ä»ĭç»į +ä½ł åĸľæ¬¢ +åĪĽæĸ° åŀĭ +åıĺ è¿ģ +è§Ĩ å¯Ł +å®Įåħ¨ 没æľī +åħĥ æĹ¦ +åı¯ ä¿¡ +åı¦ è¡Į +æĿij 级 +åħ¥ åľº +æIJŃ æ¡£ +ä¹Ł åĽłæŃ¤ +æį¢ æĪIJ +ä¸į è´Ł +äºĨ 大éĩıçļĦ +éģĶ åΰ +å¸Ĥ åİ¿ +å¹´ è¼ķ +å¿« æīĭ +å¸Į å°Ķ +èĩª èIJ¥ +éĽª èĬ± +æIJ ģ +çľ¼ ç§ij +æŃ£ 確 +çļĦ å§¿æĢģ +åĿļå®ŀ çļĦ +æĮĩ 纹 +æªĶ æ¡Ī +ç½® äºİ +佩 æľį +豪 éŨ +åĵ Ĵ +æģ° 好 +檢 æŁ¥ +åĪĿ è¡· +大 åĶIJ +约 ä¼ļ +èĴ¸ åıij +çѹ åĪĴ +å¹´ ç»Ī +è¡Į æ¥Ń +åħ± éĿĴ +åħ±éĿĴ åĽ¢ +ä¼ļ å¼ķèµ· +ä¸Ń ç§ij +ä¸Ńç§ij éĻ¢ +æĮ¯ åĬ¨ +åį´ åıijçݰ +ä¸įåĬ¨ 产 +èĮ ¹ +æĪ¿éĹ´ éĩĮ +è´§å¸ģ æĶ¿çŃĸ +æ²» çĻĤ +æħİ éĩį +å¡ŀ å°Ķ +åĽ½ ç±į +åĽł æŀľ +çŃī çī¹çĤ¹ +å±± è°· +ä¸ĭ è¼ī +è®ĵ æĪij +饮 éħĴ +è¿Ļ个 游æĪı +ç»Ŀ 大éĥ¨åĪĨ +åĴ¨è¯¢ æľįåĬ¡ +å¹² æ´» +è®® ä¼ļ +æ¦Ĥ è¿° +åĪĨ åĮº +æŃ» åIJİ +ç«Ļ çĿĢ +主è¦ģ é¢Ĩ导 +åIJĮ åŁİ +大 æłij +对 åѦçĶŁ +社ä¼ļ ä¿ĿéĻ© +å¢ŀ èµĦ +主人 åħ¬ +å®£ä¼ł æķĻèĤ² +æĸĩåĮĸ 交æµģ +客 æĪ¶ +çŁ¥åIJį åĵģçīĮ +æ»ŀ åIJİ +äºĴ è¡¥ +æĦŁ äºº +åī ¿ +åIJİ ä»£ +äºī 龸 +æķĻèĤ² åŁ¹è®Ń +éĿĻ èĦī +ä¹ı åĬĽ +说 åĩºæĿ¥ +çİĭèĢħ èį£èĢĢ +åĢ « +åįĩ èµ· +éķ ģ +åĩº 游 +éĢļè¡Į è¯ģ +å·¥ä½ľ å²Ĺä½į +åĮł å¿ĥ +æĭ¿ æĿ¥ +æ´Ĺè¡£ æľº +æĪijä¸į æĥ³ +é¢Ħ è§ģ +æ¼Ķ 示 +ä¸Ģ缴 没æľī +è·Ł 她 +对çħ§ æ£ĢæŁ¥ +ç° ¿ +ä¸ĵ å¿ĥ +è®® äºĭ +åīį 端 +åį¡ å°Ķ +è¨Ń å®ļ +设置 äºĨ +å©ļ 纱 +åľ¨ åĽ½å¤ĸ +åı³ ä¾§ +è³¼ çī© +å¥ĩ èij© +å¢ŀåĬł å̼ +好 è¿IJ +åĽ½éĻħ æľºåľº +ä¸ĭ ç§° +缮åīį 为æŃ¢ +ç¥ŀ ä»Ļ +å®ĥ åı¯ä»¥ +æ¾Ħ æ¸ħ +èĥ½ 使 +游 åĩ» +游åĩ» éĺŁ +åĩ ¹ +ä¸įè¦ģ åĨį +åĨ³ èĥľ +åĨ³ æĪĺ +æĭ ½ +缼 åħ¸ +å¾Ī好 åľ° +æľĢ ç¾İçļĦ +åĥ ļ +å·´ åŁº +å·´åŁº æĸ¯åĿ¦ +æľĢ éĢĤåIJĪ +é«ĺ èģĮ +ä¿Ŀ å§Ĩ +æİĪ æ¬Ĭ +说åΰ è¿ĻéĩĮ +æİ¨ å¼Ģ +çİĩ è¾¾ +ä¸īåĪĨ ä¹ĭä¸Ģ +管çIJĨ ä¸Ńå¿ĥ +交 æ±ĩ +森æŀĹ åħ¬åĽŃ +å¾Ģ ä¸Ĭ +éªij è¡Į +æį® æŃ¤ +纽 带 +ç» ŀ +ä¸ī æĸ¹ +æĦıä¹ī ä¸ĬçļĦ +æİ¨ è¿Ł +å¤ļæł· æĢ§ +æĥ³ èµ·äºĨ +æİĴåIJį 第 +å·¨ é¢Ŀ +æĿŁ ç¼ļ +å®ī å®ļ +äºĭ 實 +çļĦ æĦ¿æľĽ +è£ħå¤ĩ åζéĢł +人 å±ħ +人å±ħ çݯå¢ĥ +å¿ĺè®° äºĨ +该 游æĪı +楼 ä¸Ĭ +å¼Ģ ä¼ļ +æģ ³ +åıĭæĥħ éĵ¾æİ¥ +ç¡ Ĵ +ç»ĻäºĪ äºĨ +åģı 好 +åĵ ī +交éĢļ å®īåħ¨ +éĽ Į +æ²» çĹħ +è§īå¾Ĺ å¾Ī +衬 è¡« +å¿ĥ æĦ¿ +æ´ŀ å¯Ł +æ°ij æ£Ģå¯ŁéĻ¢ +æıIJ çĤ¼ +è¦ģ è¿Ľä¸ĢæŃ¥ +驾 车 +æĻ® æĥł +æķ ĸ +ç¦ı éŁ³ +éĢģ è¾¾ +è§ĦåĪĴ 设计 +æīĭ å¥Ĺ +å®ī ä¿Ŀ +è¿ĺä¸į å¦Ĥ +åīį è¿° +æłĩ è®° +ç´§ æİ¥çĿĢ +æ§ IJ +深深 åľ° +满满 çļĦ +æĺ¥ è¿IJ +æĹ¥ 产 +çα æĬ¤ +åħ¨ æĹ¥ +åħ¨æĹ¥ åζ +转 åĬ¨ +ç¥Ń ç¥Ģ +ä¹° ä¸ľè¥¿ +对 æľªæĿ¥ +æ¶Ī失 äºĨ +åļ´ éĩį +ä¸ī æĿ¡ +éħ¸ 奶 +éĽĨåĽ¢ èĤ¡ä»½ +西 è·¯ +åıª å¾Ĺ +éĢģ åİ» +çĭł æĬĵ +åĪ©ç͍ çİĩ +ä¸ĭ åij¨ +å¥ĭ æĪĺ +æĺ¥èĬĤ æľŁéĹ´ +è´Ł 责任 +æĺĤ è´µ +å°¾ å·´ +ç¯ĩ æĸĩ竳 +åħ ® +è®Ĭ æĪIJ +å¹ ¹ +çĻ» éĮĦ +ä½ Ī +å·¥ åĮł +åĵªæĢķ æĺ¯ +åıį åĵį +ç§ ĥ +åĩº 轨 +æĹ¥ åĨĽ +åIJį èªī +æķı éĶIJ +æľįåĬ¡ æ°´å¹³ +çħ§ å°Ħ +ä¼Ĭ æĭī +ä¼Ĭæĭī åħĭ +åĨħ éĺģ +èĬĴ æŀľ +ä¸ĩ åĪĨ +éĢĢ æ¬¾ +缴æĴŃ éĹ´ +æĭ¿ åΰäºĨ +å°İ èĩ´ +空æ°Ķ ä¸Ń +客æĪ· æľįåĬ¡ +è¿IJ åĬ¿ +ç»ĵ çŁ³ +ä¸į å¿ħè¦ģçļĦ +èĥ¶ åĽĬ +çIJĨ ä¼ļ +æĬ½ åĩº +空æ°Ķ è´¨éĩı +æ¯ķ 竣æĺ¯ +åĨ· æ¼ł +ä¸Ģ å¦Ĥ +ä¸Ģå¦Ĥ æĹ¢ +ä¸Ģå¦ĤæĹ¢ å¾Ģ +æĤ£ çĹħ +åĬł æĮģ +èµŀ åĬ© +é« ® +åij½ ä¸Ń +æĦıä¹ī ä¸Ĭ +ä¸į èĪį +åģļ æ¢¦ +æīĵ æī« +æĺŁ åħī +æĸŃ è£Ĥ +åħ¨ å¥Ĺ +è£ģ å®ļ +马 åħĭæĢĿ +骨 骼 +ä¸Ģ è·¯ä¸Ĭ +å®ļ æĹ¶ +å·¥ç¨ĭ æĬĢæľ¯ +å½¼ å¾Ĺ +æ±² åıĸ +ä¸Ģ è§Ī +åIJµ æŀ¶ +ä¿Ĺ ç§° +æłª æ´² +åºŁ æĹ§ +è¡Į æĺŁ +åıijçĶŁ åıĺåĮĸ +é¦ĸ ä»ĺ +åįģåĪĨ éĩįè¦ģ +æĬĬ è¿ĻäºĽ +ç¥ŀ å·ŀ +æıIJä¾Ľ åķĨ +æ¥ · +å± İ +çĬ¶ åħĥ +åŁİ å¢Ļ +çľĭ ä¸Ģçľĭ +çĶŁäº§ èĥ½åĬĽ +åŁºæľ¬ä¸Ĭ éĥ½ +æīĵ æī° +åĪĿ 次 +åĩº 示 +åħ¶ä¸Ń ä¸Ģ个 +çĶŁæĢģ ç³»ç»Ł +æīĭ æİĮ +æµİåįĹ å¸Ĥ +åľĭ åħ§ +æŃ£ å̼ +å¹¾ ä¹İ +æİ¨èįIJ éĺħ读 +è¿Ń 代 +è°ĥ ä¾ĥ +饮 åĵģ +å¢Ļ ä½ĵ +åıĺ çݰ +äºĨ 好 +äºĨ好 åĩł +ä¸į çķĻ +çĪ ² +å°½ æĹ© +æŃ£åľ¨ è¿Ľè¡Į +åĩº éĻ¢ +æĿĢ å®³ +æıIJ 款 +åıijå±ķ 空éĹ´ +åīį 身 +ä¸įæĸŃ å¢ŀ强 +æ·± å±Ĥ次 +容 纳 +éĤ£ 份 +å·¥ä½ľ æķĪçİĩ +æľ¬ åĽ½ +失 èIJ½ +æŃ£ åĽłä¸º +èĬĤ æ°´ +ä¸ĭ ä¸Ģ代 +çłĶåıij ä¸Ńå¿ĥ +ä¸į çIJĨ +å®Į 好 +ä¿ĿæĬ¤ åĮº +ç»ĵæŀĦ è°ĥæķ´ +å¥ł å®ļ +宣 ç§° +éĺ» æĮ¡ +æĴ¤ 离 +ä¸į æĸ¹ä¾¿ +åĴ ķ +ç¬ijäºĨ ç¬ij +çݯå¢ĥ 污æŁĵ +ä½ı æĪ· +ç»Ŀ ç¼ĺ +éϤ å°ĺ +é«ĺ å°ļ +æĢİä¹Ī åı¯èĥ½ +éĿ¢ èī² +åķĨ æ¥Ń +çĸ ¹ +èµĦæºIJ ä¼ĺåĬ¿ +è¾ĸåĮº åĨħ +èĢĢ çľ¼ +æij§ æ¯ģ +ä¸ĸçķĮ ç»ıæµİ +å¼ķ æĿ¥ +ä¸Ģ åĪĻ +æĭĩ æĮĩ +æĬµ 御 +éĽ į +åĩĨå¤ĩ å·¥ä½ľ +çıł ä¸īè§Ĵ +ç¨Ģ åľŁ +èİ·å¾Ĺ æĦŁ +æĪIJåĬŁ çİĩ +ç½ij 约 +ç½ij约 车 +èĦ IJ +æķ¬ ä¸ļ +éĩij ä»· +ç²¾ é«ĵ +ä¹° 车 +åħ³ åı£ +åĨį å¤ļ +æŀģ åĵģ +åIJĦ å®¶ +举æĬ¥ ç͵è¯Ŀ +èļ Ĭ +æĸ¹ å½¢ +ç§ijæĬĢ æĪIJæŀľ +æľĢ好 æĺ¯ +éĹ® åĢĻ +红 éħĴ +åĽĽ ç§į +ç¿Ĵ æħ +ç¿Ĵæħ £ +åŀ ¦ +éĤ£ åıª +é¢Ĩ æĤŁ +çľ¼ éĥ¨ +æ³° å®ī +ä»» æľŁ +磨 æįŁ +æĽ¿ æį¢ +åħ¸ 礼 +符åIJĪ æĿ¡ä»¶ +è¿ĺæľī ä»Ģä¹Ī +åħ±äº« åįķ车 +åı¯ åĪĨ为 +åŃ£ åIJİ +åŃ£åIJİ èµĽ +举èİŀ å¸Ĥ +å¿ĥ æĦı +æīŃ æĽ² +ä½ľä¸º ä¸Ģç§į +è¿Ļ éĥ¨åĪĨ +åıĤä¸İ åΰ +ç½ij çIJĥ +實 çı¾ +ç»Ħ è£ħ +åIJij å¤ĸ +å·¥ä½ľ æĸ¹æ¡Ī +åįģ æĿ¡ +課 ç¨ĭ +颤 æĬĸ +åĵ © +éĤ® å¯Ħ +äº ¢ +åħį è²» +ç§ ¤ +åºĶæĢ¥ 管çIJĨ +åĽĽ äºĶ +éºĴ éºŁ +å¾Ĵ æŃ¥ +è¨ĺ å¾Ĺ +çĴ IJ +æĺ¯åIJ¦ ä¼ļ +æĦıè§ģ åıįé¦Ī +éļ¾ æĢª +çª į +交 æİ¥ +两 åįĥ +æĩī ç͍ +æľŁ éĸĵ +æIJ¬ åΰ +è®® é¢ĺ +碧 æ¡Ĥ +碧æ¡Ĥ åĽŃ +åģļ çĶŁæĦı +éĻĽ ä¸ĭ +è· ĭ +èĢģ人 å®¶ +带 åĽŀ +æŀ¸ æĿŀ +è¡Į éķ¿ +åĨħ容 ç®Ģä»ĭ +æ¢ ¢ +æĮĩ æİ§ +éĩį çĹĩ +ç½ijåıĭ 们 +çı¾ 代 +ç±» 产åĵģ +å¥Ķ æ³¢ +æ¸ º +ç²ī ç¢İ +è¿Ļ åıªæĺ¯ +æ£Ģå¯Ł æľºåħ³ +é½ Ĭ +æĪ¿ ç§Ł +å¾· æĭī +å²ģ 以ä¸Ĭ +纯 åĩĢ +åĪĨå¸ĥ åľ¨ +èĥ½ å¾Ĺåΰ +ä¸į å°½ +ç«ŀ ä»· +çļĦ 带é¢Ĩ +çļĦ带é¢Ĩ ä¸ĭ +ä¸ŃèᝠæĿIJ +æĿij éķĩ +ä¸įåı¯ éģ¿åħį +éľ² 天 +å°ı å§ijå¨ĺ +çī© ä»¶ +èijĹä½ľ æĿĥ +æĭĺ çķĻ +éĥ½ è§īå¾Ĺ +æĽ² æĬĺ +æ·»åĬł åīĤ +åı¬ åĽŀ +æīİå®ŀ æİ¨è¿Ľ +æĬĦ è¢Ń +åĮĸ 身 +缴 èIJ¥ +ä¹Ł å¸ĮæľĽ +èį£èªī ç§°åı· +åįĸ ç»Ļ +æľī ä¸įåIJĮçļĦ +å¥ĩ çī¹ +éĥ½ 认为 +å¦ ŀ +æĪIJéķ¿ ä¸º +辩 æĬ¤ +主 æķĻç»ĥ +æ³ķå¸Ī èģĮä¸ļ +æ¤į åħ¥ +ç´¢ å°¼ +åIJ¬ è¿ĩ +ä¹łæĥ¯ äºĨ +夺 åıĸ +éŁ ĵ +æľ¬è´¨ ä¸Ĭ +æİ¥ åĬĽ +äºij 端 +è¦ģ åģļ好 +è·¯ çģ¯ +åįıåIJĮ åıijå±ķ +æľī å¾ħ +æ°´ åŁŁ +æIJľçĭIJ é¦ĸ页 +è´¨éĩı å®īåħ¨ +åįģäºĮ äºĶ +åĵ® åĸĺ +èĵ¬åĭĥ åıijå±ķ +åIJį 声 +身 亡 +çİĭ åºľ +åİŁåĪĻ ä¸Ĭ +çĥĺ å¹² +éģĹ æ¼ı +éĿ¢ 缮 +åĽ½ ä¼ļ +ä¸Ģ缴 éĥ½æĺ¯ +æľīä¸Ģ ä½į +éħį æľī +éĻª çĿĢ +ä¼ģ åĽ¾ +æĮī ä¸ĭ +èĵĿ åĽ¾ +æ© ĺ +大å¤ļ æĺ¯ +辩 论 +æĹĭ å¾ĭ +æĬ¥ éĢģ +æĿ¡ è§Ħå®ļ +åĬ¨ éĿĻ +åĮΠ奴 +æĭľ 访 +ä¸Ģ åĪĢ +ä»ĸ çŁ¥éģĵ +主 æĿĥ +ä»ĸ æĽ¾ +æĴŃ ç§į +å£ģ åŀĴ +çī¢è®° 使åij½ +åľ¨è¿Ļ æĸ¹éĿ¢ +æīĭ èħķ +æĶ¯ æŀ¶ +ä¾Ĩ èĩª +éĩį å¡ij +å¤ļ å±Ĥ次 +ä»ĭ è´¨ +éĿ¢ åŃĶ +æ½® 湿 +åİ¿ åŁŁ +游æĪı å½ĵä¸Ń +å£ ŀ +åĪĹ åĩº +èµĽ åĮº +å¤ļ åįĬ +éĩįçĤ¹ å·¥ä½ľ +æĪij们 å¿ħé¡» +æŁı æŀĹ +é²ģ èĥ½ +æĸ½ å±ķ +åIJĦ åĮº +åħį ç¨İ +èµĽ åIJİ +æľĢ éĩįè¦ģ +ä¸Ģ个 好çļĦ +è¿Ŀæ³ķ è¿Ŀè§Ħ +äºĨè§£ æĽ´å¤ļ +æķ¬ 请 +ç¬ijçĿĢ è¯´ +ä¸įæĸŃ åıijå±ķ +æijĦå½± å¸Ī +以 éĺ² +çĤ¸ å¼¹ +声 åĵį +ç¤ ģ +æĩ ¿ +èĪĨ æĥħ +èĩªçͱ è´¸æĺĵ +æķı æį· +ä¸ī大 éĺ¶æ®µ +èĭ Ķ +æĹº åŃ£ +ä¸į 满æĦı +微信 åı· +ä¿® 为 +çł´ è£Ĥ +éĢĥ 离 +æ¯ı èĤ¡ +è¾¾ ä¸įåΰ +æ¯ıå¹´ éĥ½ +çģ¯ ç¬¼ +æŃ¤ åŁºç¡Ģä¸Ĭ +åĥı 个 +åĪĨ 娩 +æĻ ¾ +ä¸į èĩ³äºİ +红 线 +误 è§£ +举 è·¯ +æ·® å®ī +产 åѦ +产åѦ çłĶ +èī¾ æ»ĭ +è»ĭ çĹħ +åīįæıIJ æĺ¯ +æ¯ı ä¸Ģ天 +ä¸ĥ 大 +æłij åı¶ +èµ° å¾Ĺ +è¿Ļ 两ç§į +æİı åĩº +æİ IJ +é¢Ĩ导 èĢħ +ä¸Ģ æľµ +个å¤ļ æľĪ +ä¸Ń åħ³ +ä¸Ńåħ³ æĿij +课åłĤ æķĻåѦ +大 åĴĸ +éģĭ ç͍ +è¯ļ æĦı +ç»Ħ åĽ¾ +è¯ķ çĿĢ +ä¹Ķ æ²» +è¿ĺ ä¸įæĺ¯ +æľī æĽ´å¥½çļĦ +åIJİ å¤ĩ +æĸ°çĶŁ åĦ¿ +æ°Ķ è¡Ģ +æ²¥ éĿĴ +å±ı éļľ +æ¥Ń åĭĻ +æĪij 以为 +éķ¿ çĽ¸ +èĢģ çΏ +éķĩ æ±Ł +æľºæ¢° 设å¤ĩ +ä½Ĩæĺ¯ å¦Ĥæŀľ +åĿļå®ļ ä¸į +åĿļå®ļä¸į ç§» +åĨ² éĶĭ +ç®Ģ缴 æĺ¯ +åĤ¨ èĵĦ +纯 ç͵åĬ¨ +漫 æŃ¥ +举 èµ· +æģ¶ æĢ§ +è¨ĺ éĮĦ +èģĮèĥ½ éĥ¨éŨ +åħ¨ éķ¿ +鼻 è¦ĸ +ä¹³ èħº +ä½ķ å¤Ħ +æ¶Ī æŀģ +æŃ£ å¤Ħäºİ +å®ī å®ģ +æĪIJ éķ· +åıĻ è¿° +æºĥ çĸ¡ +ä½Ĩ çİ°åľ¨ +女 æĺŁ +å©´ å¹¼åĦ¿ +æĬķ èŀįèµĦ +éĹ® éĹ® +æıŃ å¼Ģ +è¯ ı +åIJį å½ķ +èĺij èıĩ +åIJĬ é¡¶ +æ¹ĸ åĮº +åįĸ åľº +建 ç¯ +å»ºç¯ ī +èİ ½ +åIJ¬ åIJ¬ +ç«ŀäºī ä¼ĺåĬ¿ +åĩº ä»» +æľī 两ç§į +橱 æŁľ +è¤ ª +è¯ķ åį· +ç»ıæµİ æĬĢæľ¯ +æ·± å±Ĥ +éĩįè¦ģ åĨħ容 +é£İ æİ§ +çĬ¶æĢģ ä¸ĭ +éĥ¨ éĸĢ +广 æ±½ +è§Ĥ æij© +éģĹ çķĻ +转 è´¦ +æĮģ ä»ĵ +æĢ» 计 +åľĺ éļĬ +æĪ¿ 举 +éĺĢ éŨ +åħ¬ åħ³ +åħ³ åĪĩ +èĤ ĺ +æķ¸ æĵļ +ä¸ī åįģå¹´ +è§ģè¯ģ äºĨ +å± Ĩ +çģ° å°ĺ +æ¦ľ é¦ĸ +è¦ĨçĽĸ çİĩ +ä»Ļ 女 +çĶŁäº§ æĢ» +çĶŁäº§æĢ» å̼ +æĪ¿ è´· +æ±Ł åĮº +åħħç͵ æ¡© +çϾ åIJĪ +確 èªį +转 ç§»åΰ +éĥ½ æĹłæ³ķ +纪念 é¦Ĩ +çŃ¾ç½² äºĨ +å¹¶ä¸į å¤ļ +æĮ ł +ä¸į太 好 +ä¸ĸ 代 +误 导 +é«ĺå³° 论åĿĽ +åħ¼ 容 +龸 æ°Ķ +æĿ¥ 访 +æīĢ å¸¦æĿ¥çļĦ +æĺ¯ä¸Ģ éĥ¨ +æĻļ é¥Ń +åİĨ 代 +åIJ¦ åīĩ +ä¹ħ ä¹ħ +æľīæķĪ æľŁ +诱 åıij +æĢ» èµĦ产 +æľ¬èº« å°±æĺ¯ +çĶŁäº§ åİĤå®¶ +æĹ¶ 髦 +èĢIJ ç͍ +ä»İå°ı å°± +æĿ¡ 约 +èĭ± åĭĩ +ä¿Ĺ è¯Ŀ说 +寺 åºĻ +å¿ĥçIJĨ åģ¥åº· +ä»Ģä¹Ī äºĭæĥħ +æ±ī åŃĹ +çķĻ ä½ı +åįĹ è·¯ +ä¸ī 项 +丢 äºĨ +æĥ³ åΰäºĨ +çѹ éĽĨ +éĻĦåĬł å̼ +西 è£ħ +ä¹ĭ ä½ľ +åģļçļĦ äºĭ +çķ¶ æĤ¨ +çķ¶æĤ¨ åľ¨ +é¦ĸ 款 +ä¸įåľ¨ ä¹İ +å·¥ç¨ĭ æĸ½å·¥ +éļIJ éļIJ +åıĺ 身 +沿 éĢĶ +æĤł æĤł +ä¿Ŀ æļĸ +çĶŁæ´» åŀĥåľ¾ +渤 æµ· +æŃ¦ ä¾ł +女 主è§Ĵ +举 ä¾ĭ +æ ·¨ +çϽ é¢Ĩ +è£Ļ åŃIJ +è¿Ķ è¿ĺ +è¿Ī åĩº +é¾Ļ éŨ +ç»ıæµİ ä½ĵ +æĶ¶ å®ĺ +çķĮ éĻIJ +è·³ åĩº +åįĩ å̼ +绵 éĺ³ +çĸ¤ çĹķ +çľĭ æ¸ħ +æĭĴ çµķ +è¥Ħ éĺ³ +课 å¤ĸ +åŃIJ åŃĻ +æŃĮ è¯į +æĪIJ åIJį +溶 æ¶² +åĦĴ å®¶ +åķĨä¸ļ åĮĸ +辨 åĪ« +å¤ļ è¾¾ +ç½ij åºĹ +ä¹Ŀ 大 +ä¹Ŀ大 ç²¾ç¥ŀ +æŃ¤ 举 +è¿ŀ è½½ +ä¸Ģ åĢĭ人 +èī² æ³½ +æ¶µçĽĸ äºĨ +è¦ı åĬĥ +åĽ½ æĥħ +åį«çĶŁ åģ¥åº· +积æŀģ åĵįåºĶ +æĭ Ļ +åζ åĬ¨ +æĥ³è±¡ åĬĽ +çļĦ ä¹IJè¶£ +å¼łå®¶ çķĮ +å´ İ +éĩį åŀĭ +å¤ĸ å¢Ļ +æĶ¾ åѦ +è®¤çľŁ åŃ¦ä¹ł +è´¬ å̼ +æ³ķ æ¡Ī +æĬ¤èĤ¤ åĵģ +éĻ·åħ¥ äºĨ +请 æĤ¨ +åŀ ¢ +æķĻèĤ² èµĦæºIJ +交æĺĵ å¹³åı° +æĹ¶ è£ħ +ä¼łæŁĵ çĹħ +æ¹ĸ æ³Ĭ +èµĦ 管 +åݨ å¸Ī +éĹľ éį +éĹľéį µ +åĵĪåĵĪ åĵĪ +çĽĹ çªĥ +çĶľ ç¾İ +åºĦ åĽŃ +缮åīį å·²ç»ı +è¾¹ ä¸Ĭ +çģ« èĬ± +æĬ¥ è®°èĢħ +æģĭ æĥħ +ç´§ åĩij +æ°´ æµģ +è¿Ļæĺ¯ æĪij们 +æ³¥ åľŁ +æĽ¾ ä»» +æĸ¹ è¨Ģ +åij¨ åħŃ +åı· 楼 +ä¼ij åģĩ +误 ä¼ļ +åĽ½ åĢº +åīį å¤ķ +两 å¼ł +éĹ « +éŃĶ é¬¼ +æĬĬ æĮģ +èĬĤèĥ½ çݯä¿Ŀ +æ¸ħæ´ģ èĥ½æºIJ +èĤ¥ æĸĻ +é«ĺ é¢ij +å°± æľīäºĨ +交 ä¼ļ +没 éĴ± +éĽħ æĢĿ +è¦ģ åıĬæĹ¶ +åŁ¹åħ» åѦçĶŁ +欣 åĸľ +çĥŃæ°´ åύ +é¾Ļ æ¹ĸ +äºĮ 楼 +æĸ°æµª è´¢ç»ı +æĸ° åĬ¨èĥ½ +èµ£ å·ŀ +æĭ³ 头 +æµģ åIJij +ä¹Łæĺ¯ å¾Ī +åıij åĶ® +ä¸Ń åIJ«æľī +åIJĵ å¾Ĺ +å·¨ æĺŁ +æĹł æīĢè°ĵ +æ¯Ľ åŃĶ +åħ¬åħ± 交éĢļ +çĤİ çĥŃ +èµ· èįī +åĬłçĽŁ åķĨ +说 ä¸įåĩº +大åѦ æ¯ķä¸ļ +å·¥ä¸ļ åĽŃ +éłĺ åŁŁ +åºĨ åħ¸ +æµģ 产 +èģ² éŁ³ +ä¼¼ä¹İ æĺ¯ +è´§ æºIJ +æ·± åĪĩ +æ²»çĸĹ æĸ¹æ³ķ +èµĦæºIJ éħįç½® +ç¶² åıĭ +çĶ £ +äº ¥ +躲 åľ¨ +社 ç§ij +è»Ł é«Ķ +女 è£ħ +æŃ¡ è¿İ +综åIJĪ å®ŀåĬĽ +æł¼ å°ĩ +åħļåı² åŃ¦ä¹ł +æľĢ åŁºæľ¬ +æľĢåŁºæľ¬ çļĦ +çľĭ æľĽ +åıĹ è´¿ +ä¸įä»ħ èĥ½ +ä½ķ å¿ħ +ä¸Ģ个 å°ıæĹ¶ +ç¾ Į +æĭĽ æĶ¶ +çĤĴ èĤ¡ +æĿij å¹²éĥ¨ +缸 çα +æ½ľ èĥ½ +ä¹ į +æĹ¶ è¾° +欣 æħ° +éĵ¶ è¡Įä¸ļ +çĭŃ çªĦ +éĩįçĤ¹ é¢ĨåŁŁ +çݰå®ŀ çĶŁæ´» +éĮ¯ 誤 +æĸ° è§Ħ +滥 ç͍ +æĹ¶ ä¸į +æĹ¶ä¸į æĹ¶ +帳 èĻŁ +ç¨Ģ 缺 +åIJij 举 +ä¿Ŀåģ¥ åĵģ +çıŃ éķ¿ +äºĴ åĭķ +笼 罩 +æ½ Ľ +æļĸ å¿ĥ +è½° çĤ¸ +åºĨ 幸 +è²Į ä¼¼ +æĵ º +èĢIJ 磨 +ä¸ĵä¸ļ 人士 +ä¸Ģèά éĥ½æĺ¯ +æ¼³ å·ŀ +åħ¨ èĩªåĬ¨ +å½ķ ç͍ +大 è·Į +æľīæķĪ æĢ§ +èĩª åĭķ +ä¸ī个 æĸ¹éĿ¢ +港 åĮº +ä¿¡ 貸 +éĢļ è¯Ŀ +é«ĺ 涨 +æ³Ħ æ¼ı +éħį ä¸Ĭ +åħļ å·¥å§Ķ +被 认为 +被认为 æĺ¯ +ä¸įä¼ļ åĨį +è°ĥ åīĤ +åıĤ èĤ¡ +èĦ± åıij +å¿ł å®ŀ +åĨħ åĪĨæ³Į +ç¹ģ å¿Ļ +åıĮ åĪĽ +é©» æĿij +åĪĴ ç®Ĺ +éģİ ä¾Ĩ +åľ£ ç»ı +èıľ 鸣 +æĭ¼ å¤ļå¤ļ +ä¸ŃåĽ½ 汽车 +çĥŁ èįī +缴 æµģ +äºĨä¸Ģ åı£æ°Ķ +ä½İ æĪIJæľ¬ +æī¾ åĽŀ +èĩª åįij +總 æĺ¯ +æĸĩåĮĸ åĪĽæĦı +天 æ²³ +樱 æ¡ĥ +éªij åħµ +éĩĮéĿ¢ æľī +çİ ® +èĥ½ æī¾åΰ +éĢĥ è·ij +åĪĩ å°Ķ +åĪĩå°Ķ 西 +以ä¸ĭ æĺ¯ +å²³ éĺ³ +çļĦ æ¦Ĥçİĩ +æĬµ åζ +å¸Ī äºĭåĬ¡ +å¸ĪäºĭåĬ¡ æīĢ +åĩĨ æĹ¶ +屬 æĸ¼ +订 è´Ń +åįłæį® äºĨ +ä¸Ń éĢĶ +å° ĭ +é»ij 马 +åİ¿ åħ¬å®īå±Ģ +ä¸ĥ æľĪ +èī² ç´ł +å¿ĥèĦı çĹħ +æĹ¶ éĻIJ +æ¯į åħ¬åı¸ +å¹ķ åIJİ +ä¸Ĭ æ¦ľ +å̾åIJij äºİ +纸 ä¸Ĭ +æ¡ ĵ +éĽĨä½ĵ ç»ıæµİ +æĥħ å¢ĥ +è¦ģ åģļåΰ +ç©į 極 +åıª æĢķ +æ¹ĺ 西 +çļ± çº¹ +åħ¨ åľĭ +çĦ¡ è«ĸ +好 æĦŁ +åįķ ä»· +è¿Ľç¨ĭ ä¸Ń +æĺĨ ä»ij +åĪĽ 客 +åħħ æĸ¥ +åħĪ æĬĬ +该 æĢİä¹ĪåĬŀ +åĵģ å¾· +åħ¨éĿ¢ åıijå±ķ +è¨Ī åĬĥ +æĢ» å·¥ä¼ļ +ä½Ľå±± å¸Ĥ +æĬĹ è¡¡ +å¼Ģ åľº +éĴ± å¸ģ +åıĭ 们 +å«ī å¦Ĵ +ç´¢ èµĶ +è®Ĭ åĮĸ +æĮ¤ åİĭ +æĮij è¡ħ +çŃī ä¸Ģæī¹ +æĿ¨ 欢 +ä¸ĵå®¶ åѦèĢħ +èĥ½ è¾¾åΰ +èµ° è¿ij +è´«åĽ° åľ°åĮº +éĻIJ æľŁ +ä¸į 平衡 +åĽ½åĨħ å¸Ĥåľº +èµĽ åľº +éħį èµĦ +è¦ģ èĢĥèĻij +ä¸ĩ åı° +æľĪ æľ« +éĶ ¥ +åŃ « +æİ¥è§¦ åΰ +åĩº 产 +æķĻ åѸ +ä½ľ å¼Ĭ +çļĦ æľĢåIJİä¸Ģ +ä¿ĥ æĪIJ +åIJ¸ åıĸ +æ½ľ èīĩ +被 éªĹ +è¾ĵ äºĨ +çĭIJ çĭ¸ +åįĩ éĻį +è¿ĻäºĽ ä¸ľè¥¿ +æĬķèµĦ åŁºéĩij +çĶŁçī© åѦ +ç½ij绾 èIJ¥éĶĢ +åIJij è®°èĢħ +èįī åľ° +æĢ ¯ +æľįåĬ¡ èĥ½åĬĽ +éĥģ éĹ· +åįķ åĵģ +å¾Ĺ 罪 +æĺĵ äºİ +个å¤ļ å°ıæĹ¶ +éĩį ä»» +ä¸Ĭ å®ĺ +æľ¬ éĩij +çı¾ åł´ +溢 ä»· +æĺŁ è¾° +æ´»åĬ¨ çİ°åľº +丹 麦 +å¸Ŀ çİĭ +æŁ¥ æĺİ +åŃĺåľ¨ äºİ +é¦Ļ æ°´ +æĬ½ æ£Ģ +å®ŀéĻħä¸Ĭ æĺ¯ +æĸ° å¾ģç¨ĭ +è´¢åĬ¡ 管çIJĨ +æİ Ľ +åĨľ åİĨ +éĥ½ èĥ½å¤Ł +éĤ¯ éĥ¸ +羣 實 +ç» Ĭ +åĨµ ä¸Ķ +ç½® 身 +ç¥Ī 祷 +çĿģ å¼Ģ +æĮĩ çĤ¹ +å¼Ģ æľº +西 å®ģ +åĮĹ çº¦ +积 æ°´ +åĩº åĬ¨ +åıijå±ķ 模å¼ı +转 æĬĺ +èĢĥ çĤ¹ +æľī ç½ijåıĭ +è´«åĽ° æĿij +æĪij们 çŁ¥éģĵ +åĪĨ éĶĢ +å±± èĦī +æ¯Ķ æĭŁ +ä¼° ç®Ĺ +æĶ¹ 建 +壮 è§Ĥ +ç§ī æĮģ +æı ª +ç¦ Ģ +åĮĸåѦ åĵģ +ä¸ŃåĽ½ åζéĢł +ä¸Ģ æŀ¶ +æīį è¡Į +æĭĽ å¾ħ +åıĺ æį¢ +åīį 线 +幸 好 +è¿Ļæł· çļĦè¯Ŀ +å¿ĥ è¡Ģ管 +æĢ§ çĸ¾çĹħ +åħ¨ èĥ½ +åĪij 侦 +ä¿¡æģ¯ åıijå¸ĥ +æĺ¾ çĦ¶æĺ¯ +éĿĴ éĵľ +åIJĥ ä»Ģä¹Ī +ç͵ ä»· +æ³ķå¾ĭ è§Ħå®ļ +çħ ² +çĵ· åύ +èĤī ç±» +æıĴ åħ¥ +åĹ ľ +è¿Ł è¿Ł +ä¸ĢçĤ¹ éĥ½ä¸į +è¿ĺ åĮħæĭ¬ +èĪį ä¸įå¾Ĺ +æłĩå¿Ĺ æĢ§ +æľĪ 以æĿ¥ +ç³ĸ æŀľ +éĥ½ åºĶ该 +çݯå¢ĥ åį«çĶŁ +èĪª è¡Į +éĥij éĩį +ç½ij æĬķ +åįģ ä½³ +ç§ģ ä¸ĭ +æļ´ è·Į +åĬłå¿« åıijå±ķ +产åĵģ çłĶåıij +åĪĽéĢł åĩº +æĢ» è§īå¾Ĺ +åºķ çĽĺ +èķ Ĭ +åĩºå¸Ń ä¼ļè®® +主 æĿ¿ +æĹ¥æĻļ éĹ´ +å®ĺæĸ¹ å¾®åįļ +å¼ķç͍ æĹ¥æľŁ +åī¯ æķĻæİĪ +ç͵åŃIJ 产åĵģ +è¡° éĢĢ +çķĻ åŃĺ +çģ« åĬĽ +çĴ § +çļ Ĥ +åħ¼ åħ· +éĩį è¿Ķ +é¢Ĩ çķ¥ +åĪĩ éϤ +åĨįçĶŁ èĥ½æºIJ +å®ŀåľ¨ 太 +çIJĨ论 ä¸Ĭ +ä¸ī å±Ĥ +ä¸ĸçķĮ åIJĦåĽ½ +å®ľ æĺĮ +è̳ è¾¹ +宽 æķŀ +æ±ī æĹı +çϽ çϽ +è¿ĻéĩĮ éĿ¢ +çĶŁæ´» ä¹łæĥ¯ +èµŀ èµı +çĶ· 士 +ä¸Ń ä¿Ħ +车 祸 +åīĤ éĩı +éϤ åİ» +å·¦ è¾¹ +çŃij çī¢ +çīĽ å¸Ĥ +å®¶ åĬ¡ +åķ ĥ +ç½® æį¢ +ç´« å¤ĸ +ç´«å¤ĸ 线 +å¾Ģ åīį +åĬĽ åѦ +ç´§ è·Ł +缮çļĦ åľ¨äºİ +ç» ® +ç¥ Ĥ +宣 è¨Ģ +äºĮ æ°§åĮĸ +äºĮæ°§åĮĸ 碳 +æĹł ç¼ĺ +ç²¾ éĢļ +è¨ º +å¼ķåıij äºĨ +æľĢ åħĪ +æ´¾ é©» +ä¸į å¿į +æĪij çΏ +å¹´ ä¸ĭåįĬå¹´ +æ·ĭ å·´ +没 éĹ®é¢ĺ +åºĹ åĨħ +è·Ł æĪij说 +çĶŁäº§ çĶŁæ´» +è§Ĥ æľĽ +æ¸ į +被 æī§è¡Į +被æī§è¡Į 人 +èĪ ľ +æİ º +ä¸Ģ ç§Ĵ +èįī åĿª +åij¼ åĴĮ +åij¼åĴĮ 浩 +åij¼åĴĮ浩 çī¹ +人æ°ij éĵ¶è¡Į +çĦķ åıij +è¯ģåΏ 交æĺĵ +çķ Ķ +æľº èĥ½ +å¦ ¾ +æĻļ å¹´ +å·¥åķĨ èģĶ +åİŁ åŀĭ +è§Ĵ度 çľĭ +æĬ¥ 社 +è¯į æĿ¡ +躲 éģ¿ +éĩį åIJ¯ +å¤ķ éĺ³ +èĤ¡æĿĥ 转让 +åľ¨ ä¸Ģ +åľ¨ä¸Ģ æĹģ +社ä¼ļ åĮĸ +åıijå±ķ åİĨç¨ĭ +æĭĸ æ¬ł +使 èĢħ +ä¸İ åIJ¦ +æĸ° å±ĢéĿ¢ +ä»Ĭ天 æĪij们 +é½IJ èģļ +对 æĪij说 +éĢĴ 交 +æľª æĽ¾ +èİ Ĭ +éĸ ī +亲 æīĭ +è§Ĵ éĢIJ +æľī é»ŀ +ç¨İ çİĩ +ä½İ 声 +é»ĺ å¥ij +æĻ® æ³ķ +大 ä¸ĵ +第äºĮ 大 +ä½ı åĿĢ +æĶ¾ è¿Ľ +äºĮ æĪĺ +亲 身 +åĽº åĮĸ +ä¸ĭ 乡 +åħ³éĶ® æĬĢæľ¯ +åĽŀ æĥ³ +æĬ¥ åĪĬ +æ¶Ĥ æĬ¹ +èĹı çĿĢ +ç¥Ŀ æĦ¿ +åįĩ 温 +çĶļèĩ³ è¿ŀ +åħ¬åħĥ åīį +ç¾İ æĸ¹ +è¯ļ å®ŀ +æĹł åģ¿ +åīµ æ¥Ń +å°ıå¿ĥ 翼 +å°ıå¿ĥ翼 翼 +两 æīĭ +温馨 æıIJ示 +仿 羣 +æĥ ¶ +èĥ¡ åŃIJ +å·¥ä½ľ ç«Ļ +硬 çĽĺ +ç« ¿ +åĤ³ éĢģ +åħ¨ æł¡ +é²ľ æ´» +çĴĢ çĴ¨ +ç»ĵ å°¾ +æį¢ æĿ¥ +æĪ Ģ +ä½İ ä½į +ä¸ĩåħĥ 以ä¸Ĭ +åĬł åĪĨ +æİ¨ä»ĭ ä¼ļ +çIJĨ èµĶ +å¾· å°Ķ +æĬĹ è®® +æ´ ¼ +åĸ § +åŁİ éĻħ +å¾Ī æ£Ĵ +人 æŃ»äº¡ +ä¼ļå±ķ ä¸Ńå¿ĥ +äºĴèģĶ äºĴéĢļ +èĸĦ èĨľ +éĩį é»ŀ +ç¦ģ æ¯Ĵ +åĨ· ç¬ij +大家 åı¯ä»¥ +é¦ĸ 缸 +è¿ij è·Ŀ离 +æµ® çݰ +ç§ĺ è¯Ģ +èµ· é£ŀ +æIJ ¶ +羣 åģĩ +æģ ķ +å°ı åºĹ +æ°ij çľ¾ +åıijå¸ĥ åħ¬åijĬ +ä¾§ éĩį +å¾ĺ å¾Ĭ +æĢ Ķ +æª IJ +æķ° 缮 +åī¯ ç§ĺ书éķ¿ +两 åı¥ +éļIJ çŀĴ +åıĮ åıĮ +æīĭ æĦŁ +èij¡ 京 +éģĹ å¿ĺ +é¬ ¥ +è¿Ļ个 åľ°æĸ¹ +说 çļĦè¯Ŀ +å·¡ åĽŀ +è¿Ŀ 竳 +æī¾ å·¥ä½ľ +æĶ¯ çIJĥéĺŁ +裡 éĿ¢ +æĺ¾ç¤º åĩº +èĩ³ å°Ĭ +两 级 +åīį æ®µæĹ¶éĹ´ +çĺ¦ èº« +èĤ¢ ä½ĵ +æ¯į 親 +æīĭç»Ń è´¹ +汽车 è¡Įä¸ļ +æİ© çĽĸ +æİ§èĤ¡ éĽĨåĽ¢ +åı£ å¾Ħ +æĶ¿çŃĸ æİªæĸ½ +æµ· 绵 +åħ¨ éķĩ +äºĭ åħ³ +å¸Ń æī§è¡Į +å¸Ńæī§è¡Į å®ĺ +éĤ£ 次 +åı¯èĥ½ åĩºçݰ +ä¸Ńå¿ĥ åŁİå¸Ĥ +ç¿» 身 +ä¹Ł ç®Ĺ +ä¾µ çķ¥ +åĸĩ åıŃ +æ¯ı次 éĥ½ +è§ ħ +éĻ¢ éĻ¢éķ¿ +å§ĭ äºİ +èѦ åĬ¡ +èᝠæĿIJ +å±ł æĿĢ +æľ¬èº« å°± +éļıæĹ¶ éļı +éļıæĹ¶éļı åľ° +åĶ® åįĸ +æĹłäºº 驾驶 +é¢ ħ +åĵģ 質 +åĺ² ç¬ij +è·ij åİ» +åħĭ éĩĮæĸ¯ +çķ¸ å½¢ +ä¿® 饰 +磩 éĺµ +éŁ³ä¹IJ ä¼ļ +æŁ³ å·ŀ +é½ ¡ +ä¼ļ è°Ī +æŃ£ çīĪ +ä¹Ł åIJĮæł· +æļ§ æĺ§ +è¡ĮæĶ¿ éĥ¨éŨ +ä¹ĸ ä¹ĸ +èĤ¤ èī² +æĹ¶ ä»» +羣 åĪĩ +æľĪ ä¸ĭ +æľĪä¸ĭ æĹ¬ +举æĸ¹ è´¢å¯Į +è£ħä¿® åħ¬åı¸ +éĢĢ è¿ĺ +åĭĺ å¯Ł +åĵ¥ 伦 +åĵ¥ä¼¦ æ¯Ķäºļ +çĭ¬ ä¸Ģ +çĭ¬ä¸Ģ æĹł +çĭ¬ä¸ĢæĹł äºĮ +è°ĥ åij³ +åİĭ è¿« +åħ¨çIJĥ æľĢ大 +åī¯ æł¡éķ¿ +æĽ´ ä½İ +åĪĨéĴŁ åIJİ +åĽŀ ä¾Ĩ +åζ åīĤ +åijĬè¯ī 大家 +çĤ¹ éĴŁ +åįģä¸ī å±Ĭ +åij¨ åĽĽ +è¿Ļæł· ä¸Ģ +è¿Ļæł·ä¸Ģ æĿ¥ +èĭ Ł +æľĽ åİ» +æĪIJ è¯Ń +å½ĵ åį³ +ç¬ij 声 +ä¹ĭ åĬ¿ +åĪijäºĭ æ¡Īä»¶ +æĮĤ çĿĢ +ä½ķ ç§į +å°ı 游æĪı +åĽ½å®¶ æĪĺçķ¥ +åĨ· åĨ· +å®ľ 宾 +æIJº ç¨ĭ +è¶ĭ äºİ +åıį çľģ +常 说 +ä¸ĩ æĪ· +åĥµ å°¸ +åįĥä¸ĩ åĪ« +åıijçݰ éĹ®é¢ĺ +åı¯ çŁ¥ +éŨæĪ· ç½ijç«Ļ +åģ¥åº· 产ä¸ļ +åı³ è¾¹ +æµ· è¿IJ +è¿ij ä¹İ +åĮ» æ²» +æĢ» ç®Ĺ +ä¸Ģ åĪĨéĴŁ +æĭ § +ä¹Ł æľīä¸ĢäºĽ +ä¾Ľç͵ åħ¬åı¸ +å»ī ä»· +帮 ä»ĸ +æŃ¤æ¬¡ æ´»åĬ¨ +åıªèĥ½ 说 +èĬ ĭ +çīĩ 段 +åŃĺåľ¨ éĹ®é¢ĺ +ä½łä¼ļ åıijçݰ +è½® å»ĵ +ç½ij éĢļ +滨 æ±Ł +æİĪ ä¿¡ +é»İ æĺİ +ä¸į å±ŀäºİ +约 åįł +éķ¿æ²Ļ å¸Ĥ +èĥļ èĥİ +åħĥ ä»¶ +éĻĨ åĨĽ +è³¼ è²· +æĮĩ æľĽ +å®ŀä¹ł çĶŁ +çī¹çĤ¹ æĺ¯ +çıł æ±Ł +çľĭ ä¸įåĩº +ä¸įè§ģ äºĨ +ç¼ ī +éĺµ èIJ¥ +åĶIJ æľĿ +没 å¿ħè¦ģ +åĽ½åľŁ èµĦæºIJ +ç»ıæµİåѦ å®¶ +åIJĪèĤ¥ å¸Ĥ +çIJ¢ 磨 +ç¡® åĪĩ +åŁİå¸Ĥ åıijå±ķ +çŃ· åŃIJ +人æ°ij æľįåĬ¡ +满 åĪĨ +è¿· ä¿¡ +ä½ľèĢħ æľ¬äºº +æĸĩ竳 æĿ¥æºIJ +ç«Ļ ç«ĭ +æŀĦ æĪIJäºĨ +è¾Ľ åĭ¤ +è¶ħ 强 +éĶ ļ +åīįä¸ī åŃ£åº¦ +å°± è§īå¾Ĺ +å´ĩ é«ĺ +è¶Ĭ ä¾Ĩ +è¶Ĭä¾Ĩ è¶Ĭ +å¸Ĥåľº èIJ¥éĶĢ +综åIJĪ ç´łè´¨ +åŃ ļ +ä¾® è¾± +äºĮ åŃĹ +å·¥ä½ľ ä»»åĬ¡ +åı²ä¸Ĭ æľĢ +æľĢ ä¼ĺ +åIJ© åĴIJ +表 çϽ +èİ« åIJį +èİ«åIJį åħ¶ +èİ«åIJįåħ¶ å¦Ļ +å¹ £ +åIJĮå¿Ĺ 们 +建设 çĶ¨åľ° +åĦ Ģ +éħį åģ¶ +å¼ © +åͱ çīĩ +æīĭ èĦļ +åħ¼ ä»» +åģľ æĶ¾ +æŃ£ å®Ĺ +æĸ° åĨľæĿij +åĤ¬ çĶŁ +æīĢ åŃ¦æł¡ +念 ä½Ľ +åͤ éĨĴ +åħ± åĪĽ +æĭī ä¸ģ +èĥĮ çĿĢ +çĶŁæĢģ ä¿ĿæĬ¤ +åı£ 头 +æĸ¹åIJij çĽĺ +調 æķ´ +æĭĽèģĺ ä¿¡æģ¯ +åħ¶ä»ĸ åĽ½å®¶ +ç®Ģ æĺĵ +åĮ¿ åIJį +è¯Ħ æµĭ +æĺ¯ä¸Ģ 座 +çīµ æīĭ +è¶³ 迹 +çIJĨè§£ åĴĮ +æľĢ åıĹ +å¿ĥ è·³ +çζ 親 +éĿŀ常 åĸľæ¬¢ +èĭ¦ éļ¾ +æĬĢ å¸Ī +æ°ij æĦı +æĪĺ åĽ½ +æĽ¿ è¡¥ +æ´¥ è´´ +ä¸ŃåĽ½ ä¼łç»Ł +åIJĦ è¡Į +åIJĦè¡Į åIJĦ +åIJĦè¡ĮåIJĦ ä¸ļ +第äºĶ å±Ĭ +èį· èĬ± +æĦı èŃĺ +票 ä»· +åĪĨ æµģ +æĿİ çϽ +æ±Ł åĮĹ +æİĴ æĸ¥ +ä½ĵ éĩı +åĮħåIJ« äºĨ +åĪĺ æŁIJ +çݰ å¦Ĥä»Ĭ +å·¥èīº åĵģ +è¿Ļç§į æĸ¹æ³ķ +åĬŀåħ¬ 楼 +ç͵ å·¥ +çħ Ļ +åį¡ çīĩ +å¹´ å¹´åºķ +ä¸ĵ项 èµĦéĩij +åĮ» ç§ij +åĮ»ç§ij 大åѦ +åĽŀ头 çľĭ +ä¸į å±ij +èĩª 驾 +没 æĶ¶ +æīĵ çĮİ +èĦ¸ éĥ¨ +åıĥ èĢĥ +å°Ĩ 士 +è´«åĽ° 人åı£ +çIJĨæĥ³ 信念 +é£İ å°ļ +人æīį éĺŁä¼į +çij ¾ +æĿ¥ è¿ĻéĩĮ +æ´Ĺ 涤 +å¹´ èĸª +èĭį çϽ +ä¸ĩ äºĭ +课 æľ¬ +åºĵ éĩĮ +çī¹ æ´¾ +ç´¾ åijĺ +èµŀ ç¾İ +ç©¿ æĪ´ +製 ä½ľ +èµŀ æĪIJ +ä¸Ģ ä¾§ +å½ĵåľ° 人 +æĭ İ +纸 è´¨ +ä½Ļ 个 +éĶĤ çĶµæ±ł +æľº åŀĭ +éĻ¢ éϢ士 +åģļ å·¥ +å¼ł è´´ +ç¥Ľ æĸij +æ®ĸ æ°ij +å¥ij 约 +æ¹ĺ æ½Ń +æIJ ĸ +åŃĺ è´§ +交éĢļ 大åѦ +è¶ģ çĿĢ +æĸĩçī© ä¿ĿæĬ¤ +å¤ĩ æĪĺ +éĩĩ 纳 +åįĬ æľĪ +æľĢ åħ³éĶ® +æľĢåħ³éĶ® çļĦ +æİ¥ éĢģ +æĶ¶ åī² +åıį åĢĴ +çĥ Ľ +æ ½Ķ +ä¼Łå¤§ å¤įåħ´ +çļĦè¯Ŀ è¯Ń +容 å¿į +å®ļ éĩı +æķ Ĺ +åĵģçīĮ 形象 +æīŃ è½¬ +åĽ½å®¶ éĩįçĤ¹ +èĨĿ çĽĸ +ä¸Ģ 楼 +大 éϏ +éĤª æģ¶ +åĽŀ åij³ +çĮ ¿ +çĿ¡ åīį +æĹł è¾ľ +çĹħæ¯Ĵ æĦŁæŁĵ +æľºæ¢° åĮĸ +çĤ¹ 亮 +溶 è§£ +åĩłä¹İ æīĢæľī +è·ij éģĵ +ç͵è§Ĩ æľº +åı ¨ +æijĩ äºĨ +æijĩäºĨ æijĩ头 +èĩª è´Ł +综åIJĪ åĪ©ç͍ +èĩª å¦Ĥ +åİŁ ä¾Ĩ +ä¹Łä¸į æĥ³ +èĬĤ 课 +è¿ĩ åī© +çͲ çĬ¶ +çͲçĬ¶ èħº +æĸ° ä¸ĸ纪 +èĩªä¸» åĵģçīĮ +é«ĺ å±Ĥ次 +ä¸Ģ è§Ĵ +è¡Į äºĭ +ç¥ĸ åħĪ +å©ļ åIJİ +éĹ´ éļĻ +ç¼Ŀ éļĻ +è¿Ļ æĶ¯ +ä¸įæĸŃ åĪĽæĸ° +å¾® åŀĭ +æĽĻ åħī +享 ç͍ +ä¸ŃåĽ½ ç§»åĬ¨ +éĹŃ çݯ +æī§ æĦı +åıijå±ķ æł¼å±Ģ +æł¸å¿ĥ åĮº +éªļ æī° +åħļåĴĮ åĽ½å®¶ +ä¸ŃåĽ½ æĶ¿åºľ +帶 èijĹ +ä¸ĩåįĥ çĵ¦ +åħ© 人 +äºİæĺ¯ æĪij +åĽº ä½ĵ +çªģ å¦Ĥ +çªģå¦Ĥ åħ¶ +çªģå¦Ĥåħ¶ æĿ¥ +éĩĮç¨ĭ ç¢ij +çα ç¾İ +æŁ¥ éªĮ +åıĮ èµ¢ +éĹª åħī +楼 å®ĩ +æĻ ı +æľī è¶³å¤ŁçļĦ +æŁĶ æĢ§ +ä¿¡æģ¯ å®īåħ¨ +管 线 +å¹¶ ä¸įä¼ļ +åύ ä»¶ +ä½ł åºĶ该 +çĿĢ å®ŀ +æĺİ æ¸ħ +æĬĹ çĶŁç´ł +æīĵ æŃ» +å®Įåħ¨ ä¸įåIJĮ +èĬ± æ¤Ĵ +æĶ¾ 宽 +ä½İ 端 +åĽĽ èĤ¢ +åĮĹ京 èµĽè½¦ +éĽĨ å¸Ĥ +æľª å©ļ +大å¹ħ æıIJåįĩ +建çŃij 设计 +çĭ¬ æľīçļĦ +æİ¢ éĻ© +æ²³æµģ åŁŁ +æħķ 容 +被 çĽĹ +åĵº ä¹³ +èı ģ +æĥ¬ æĦı +è¶ĬæĿ¥è¶Ĭ 好 +广大 群ä¼Ĺ +å¾· èĤ² +å¸Ĥåľº ä»·æł¼ +奥 å·´ +奥巴 马 +èĬĤ缮 ä¸Ń +两 款 +ä¸ĩä½Ļ åħĥ +ç»´ å°Ķ +çĶŁçī© ç§ijæĬĢ +åIJ¬ èµ·æĿ¥ +çł ļ +æĭŁ å®ļ +æ²¹ çͰ +声 èªī +建çŃij ä¸ļ +éĻIJ è´Ń +çīĩ åŃIJ +çķľ ç¦½ +ç½ij é¦ĸ页 +ä¼Ĺ çѹ +æĴŀ åĩ» +åīį ä¸įä¹ħ +åīį ä¸ĸ +åĽĽä¸ª æĦıè¯Ĩ +æµĭ ç»ĺ +éĺ² ç©º +漫éķ¿ çļĦ +æ²IJ æµ´ +æ¯Ķè¾ĥ ç®Ģåįķ +æµĭ å®ļ +åĽŀ è°ĥ +让 人们 +èĴĭ ä»ĭ +èĴĭä»ĭ çŁ³ +ç»ĵ æĻ¶ +å¢ŀæ·» äºĨ +æĿ¡ è¯Ħ论 +åī¯ ä¼ļéķ¿ +ä½ı æīĢ +ç»Ļ åĩºäºĨ +è°ĥ éħį +æ² ĸ +æľī ç͍ +æľīç͍ çļĦ +ä¸ĢæĿ¡ é¾Ļ +éĩİ å¤ĸ +ç¼ĺ åĪĨ +æ°¸è¿ľ ä¸įä¼ļ +æŀľ æłij +大åıij å¿«ä¸ī +麻 éĨī +äºij éĽĨ +åİ» åĵªéĩĮ +åħ¥ å¸Ĥ +ä»» æĢ§ +建 æ¡£ +建档 ç«ĭ +建档ç«ĭ åį¡ +ä¸Ģ 棵 +社 åįĢ +缸 ä¼´ +åļ · +å¡« åħħ +ä¸Ģ æĹı +ç¾ ģ +åıĸ è¯ģ +èΰ éĺŁ +åİĤ åĮº +è¡· å¿ĥ +åıijå±ķ éĺ¶æ®µ +é«ĺ 强度 +åĹĵ åŃIJ +é¢Ĩ è¡Ķ +楼 主 +大 èĴľ +æŀķ 头 +ç²® æ²¹ +é»Ħ çĵľ +æĵ Ĵ +å°ı çĭĹ +æĶ¹éĿ© å§Ķ +åįģ åĪĨéĴŁ +é²ľ èī³ +åħ³ ç¾½ +çĭĢ æħĭ +å®ŀç͍ æĢ§ +å°ij è§ģ +é£ŀ æī¬ +çͰ éĩİ +æIJ Ĥ +è¿Ļ个 è¯į +åºĶæĢ¥ é¢Ħæ¡Ī +è§Ĵ度 æĿ¥çľĭ +æķ¬ çķı +æ³ķ å®Ŀ +åĸĦ æĦı +æīĵ æĸŃ +对 åĨ³ +çµķ å°į +åĢŁ æŃ¤ +å¼Ģ æºIJ +å°ı 說 +ç¥ º +å²ģ 以ä¸ĭ +éĢĢå½¹ åĨĽäºº +ä¸įä¹ħ åīį +åĩº åİĤ +讽 åĪº +æĿ¥çľĭçľĭ åIJ§ +éŃĶ åħ½ +çķĻ ä¸ĭæĿ¥ +å±ħ 室 +åłħ æĮģ +çľĭ äºĨä¸Ģ +çľĭäºĨä¸Ģ çľ¼ +éĽĨåĽ¢ æĹĹä¸ĭ +æĪĺ æĪĺç»ĦåIJĪ +è®¤çľŁ èIJ½å®ŀ +汽车 产ä¸ļ +çī©çIJĨ åѦ +æķ µ +éĴ Ŀ +åĽ¢ éķ¿ +ä¸įæĸŃ æī©å¤§ +èĤ© è´Ł +åıijå±ķ 缮æłĩ +è³ĩ éĩij +åīį ç½® +ä¸ŃåĽ½ åı¤ä»£ +æŃ» åĪij +åħħåĪĨ ä½ĵçݰ +åħ³ éŨ +ç¾İ æĦŁ +æīĵ åħ¥ +æĬijéĥģ çĹĩ +å°ij çĪ· +æłij æŀĿ +æ¶Īæģ¯ ç§° +æ´Ľ åħĭ +åį ¯ +è¿Ī åIJij +æİ¨ åĭķ +ä»İä¸ļ èĢħ +åİ» ä¹° +欢 å¿« +æĭ¥ æĮ¤ +马 æ¡¶ +æĬĬ æİ§ +æĶ¿ åħļ +å¼ł æī¬ +客 æłĪ +红 æĺŁ +éĢģ æĿ¥ +åħ¨åŁŁ æĹħ游 +èĩª ç§ģ +åįģäºĮ æĿ¡ +åı¹ æģ¯ +ä¸Ģ èīĺ +ä¿Ŀ è´¹ +æĸ½å·¥ çİ°åľº +æľī 幸 +ç»Ń èĪª +åı¯èĥ½ æľĥ +èĥĮ åıĽ +ä½£ éĩij +ä¸ī çŃīå¥ĸ +å¾Ī 满æĦı +游æĪı åľ¬ +群 éĩĮ +æŀĦ ä»¶ +åºı å¹ķ +太 æ¹ĸ +æľ¨ è´¨ +æĻĭ æ±Ł +çµĤ æĸ¼ +è·³ è·ĥ +åĢºæĿĥ 人 +çŃī 诸å¤ļ +æĶ¾ åĩº +åħ³éĶ® æĹ¶åĪ» +æĦŁæŁĵ èĢħ +é£ŀè¡Į åijĺ +èĥĨ åĽº +èĥĨåĽº éĨĩ +æĬ± æŃī +åij¨ äºĮ +æĸ° æĹ¶æľŁ +åĨ·éĵ¾ çµģ +è¿Ļç§į æĸ¹å¼ı +该 æĿij +åĽŀ é¦Ī +åŁºçĿ£ æķĻ +人 åıĤ +æŀ¯ çĩ¥ +æī¹åıij å¸Ĥåľº +åħħåĪĨ èĤ¯å®ļ +å¸Ĥ æĶ¿åįı +äºĭ æ¥Ń +龸 çİĭ +çĥŃ æIJľ +åįģä¹Ŀ 大 +ä¼´ æľī +ç¾İåĽ½ æĢ»ç»Ł +åŁİå¸Ĥ 管çIJĨ +ä¸ĭ 令 +èĥ¸ åı£ +åıª çŁ¥éģĵ +åij¨ ä¸ī +ç͍ æĪ¶ +éŃ ¯ +å¿ĥ è¡Ģ +带头 人 +åĮ» åĬ¡ +åĮ»åĬ¡ 人åijĺ +æİ§åζ åύ +ä½ľåĵģ åĨħ容 +æĪĺ åıĭ +åİĨ å¹´ +ä¸į åħĭ +ä¸įåħĭ ä¸įåıĬ +æĹ¥ æŃ£å¼ı +è±IJ å¯Į +ç¨İ è´¹ +æĹ¶ æķĪ +å±ķ ä½į +è¡¡ éĺ³ +æĪ¿ 貸 +çĪĨ 款 +ä¹IJ æĦı +çĶ· 主 +å¯ ¬ +æľĥ èѰ +ä¹ĭ å¤ľ +åIJĮ 樣 +ä¸įè¦ģ 太 +ä¼Ĭ æĸ¯ +ä¼Ĭæĸ¯ åħ° +åŁºæľ¬ åİŁåĪĻ +åİ» æİī +ä½İ ä¿Ŀ +个 交æĺĵ +个交æĺĵ æĹ¥ +èģĬ èģĬ +åĽĽ ä½į +åħļç»Ħ æĪIJåijĺ +主è¦ģ ä»İäºĭ +å½± éŁ³ +åĨĴ åĩº +åij¼åIJ¸ éģĵ +è¾¾ å°Ķ +æľ¨ åľ°æĿ¿ +诡 å¼Ĥ +çģ¯ åħ· +çģ« çĥ§ +è§£ èĦ± +æĦĪ åıij +æ¹ĸ å·ŀ +é£İ ä¿Ĺ +æĸ° å½¢åĬ¿ +æĸ°å½¢åĬ¿ ä¸ĭ +è² Ŀ +èĦ ĵ +åĬ¨åĬĽ çĶµæ±ł +é£ŀ èι +飧 æĢ§ +åĪ© çī© +åĪ©çī© æµ¦ +ä¸į 认è¯Ĩ +ç¼ĸ ç»ĩ +ä½ľ åĿĬ +èģĮä¸ļ æĬĢèĥ½ +çľĭ è¦ĭ +åĽ´ æ£ĭ +æĺı è¿· +å½Ĵ å±ŀäºİ +æĤ¬ å´ĸ +éĨ« çĻĤ +å®ĭ 代 +åºĦ æĿij +èĹ ķ +çĮĽ çĦ¶ +çĩĥæĸĻ çĶµæ±ł +å®ŀä½ĵ åºĹ +ä¸įè¶³ 以 +æĥħ ç· +æĥħç· Ĵ +å»Ĭ åĿĬ +ç͵ åı° +åºĶ åĬĽ +ä¸Ńå°ı åѦçĶŁ +èĥ¡ åIJĮ +éī´ åĪ« +åĨħ ç½® +ä¹± 象 +æ¬Ĭ çĽĬ +å¼ĢæĶ¾ å¼ı +åįļ æĸĩ +讲 课 +çŃī åİŁåĽł +ç©· 人 +交 æĽ¿ +æĬ¤ çħ§ +åıijå±ķ æľºéģĩ +客 åķĨ +åıį ä¹ĭ +ç±³ é¥Ń +å¹¶ åıij +å¹¶åıij çĹĩ +æ±ī åŃIJ +æŀľ åĽŃ +对æĪij æĿ¥è¯´ +åģı åIJij +æī¹ 示 +读 åIJİ +读åIJİ æĦŁ +æĺİ æĻº +åĽ´ çĿĢ +åıį 转 +æĿ¨ å¹Ĥ +ä¸ĵ åįĸ +ä¸ĵåįĸ åºĹ +åıĹ éĻIJ +åºŁ è¯Ŀ +æŀģ å°ij +åįĪ åIJİ +è¿Ľ ä¿® +åīĬ åĩı +æľ¬ç§ij çĶŁ +ä¼ĺ éĢī +åħī çħ§ +åıĻ äºĭ +åıĸ æļĸ +åĮĹ è·¯ +æ¦ ķ +èİĨ çͰ +楼 å±Ĥ +天 èĬ± +天èĬ± æĿ¿ +çĤ ľ +å·²ç»ı æľīäºĨ +è¶ ¾ +çͳ åįļ +ç͵ éĺ» +åĬŁ è¯¾ +æŃ¥ æŃ¥ +éĤ£ä¹Ī 容æĺĵ +æŃ¤ æĸĩ +ä½ ° +计 è¾ĥ +çīĩ éĿ¢ +ç͵影 éĻ¢ +ä¸į åħ¬å¹³ +ä¸ī æľŁ +æĹħ游 èµĦæºIJ +å¤ļç§į å½¢å¼ı +è£Ĥ ç¼Ŀ +åIJİ æİĴ +硬 度 +åĽŀ æļĸ +éģĵ æķĻ +è´« è¡Ģ +æ¸ħ é¦Ļ +伤 çĹħ +æĦı 義 +çļĦ ç¼ĺ +çļĦç¼ĺ æķħ +åºĦ 严 +åıªæĺ¯ 为äºĨ +æīĵ æĬĺ +以 ä¾Ĩ +滿 è¶³ +çİĽ 丽 +風 éļª +æĸĩ ç§ij +éħįå¤ĩ äºĨ +è¿Ľ é£Ł +æ¶ ¡ +è·¯ ç¨ĭ +åı« 声 +ä¸Ńå¿ĥ åŁİåĮº +æľīæīĢ ä¸įåIJĮ +å¼µ è²¼ +é¢Ħ æĬ¥ +æľīå¤ļ ä¹Ī +è¿Ľè¡Į åħ¨éĿ¢ +æĽ¾ ç¶ĵ +ä¸ī 代 +å®ı 大 +æ¸ħ æī« +éĢī åĩº +åĵª ä¸Ģ个 +主 義 +ä¾Ŀ æĵļ +çļ® éĿ© +èµ¶ æĿ¥ +çŃĽ æŁ¥ +æ¨ Ł +ä¿Ŀ èįIJ +åIJĥ æĥĬ +æľĭåıĭ们 对 +ä»ĸ æĺ¯ä¸Ģ个 +åºŁ æ°Ķ +æ» ħ +è´¢ ç¨İ +æĿij æĿijæ°ij +èµĦ产 è´ŁåĢº +å®ī å¨ľ +缮åīį åĽ½åĨħ +æĦŁè§ī èĩªå·± +çµIJ åIJĪ +éͦ æłĩ +éͦæłĩ èµĽ +æĽ´ æ·± +åŁº æķ° +éħ¿ éħĴ +çī¹èī² äº§ä¸ļ +åİĭ å®ŀ +ä¾Ŀæ³ķ 追究 +æ·¡ å®ļ +ç®Ģ缴 å°±æĺ¯ +å£ĵ åĬĽ +æ°ij å¿ĥ +ä¸į åIJĪéĢĤ +çͱæŃ¤ åı¯è§ģ +èµŀ èªī +æ¾ ¤ +åĩłå¹´ åīį +åIJī ä»ĸ +çł´ æįŁ +轻轻 åľ° +å²Ľ 屿 +æĦı å¢ĥ +ä»Ģä¹Ī åı« +åģĩ è£ħ +éĢģ è´§ +å¹ķ å¢Ļ +妥 åįı +åĽ½ æĹĹ +äºĨ å¾Īä¹ħ +åĪĨ辨 çİĩ +ç´ Ķ +éĺ³ åĮº +åĩŃ çĿĢ +åģľè½¦ ä½į +京 éĥ½ +éĶ £ +æĵ ¾ +è¿Ľ éŨ +åĪĺ æµ· +åĽĽ 级 +女 è¶³ +è¡ĮæĶ¿ 审æī¹ +éģ¥ æİ§ +ä¸į éĮ¯ +å¾Ĺ å¾Ī好 +为 缮çļĦ +ä»į æľª +ç²¾ è£ħ +éĢį éģ¥ +å°½ 头 +çºł ç¼ł +éłĺ å°İ +æĭħ è´Ł +æĪĸèĢħ åħ¶ä»ĸ +åıªä¸įè¿ĩ æĺ¯ +åı® åĺ± +åģĩ åĨĴ +æļĸ æ°Ķ +çĽIJ åŁİ +被 è§Ĩ为 +诺 è´Ŀå°Ķ +ç»ĻäºĨ æĪij +è¿ij åįĥ +éĩį åĽŀ +éĨĴ äºĨ +ç͵ è§£ +忽çķ¥ äºĨ +èĥĮ éĥ¨ +æĸĩæĺİ åŁİå¸Ĥ +æº ħ +è² ĵ +æĬµ æĮ¡ +åĸľæ¬¢ åIJĥ +éĿĻéĿĻ åľ° +å¾Ī æ·± +åŁºç¡Ģ çŁ¥è¯Ĩ +è¿ĩ éĶĻ +çIJĨ ç§ij +交æµģ åIJĪä½ľ +èĪ Ķ +調 æŁ¥ +æħĪ æĤ² +éĴ ° +èĩ´ ç͵ +å®£ä¼ł æ´»åĬ¨ +åıĺ éĩı +çļĦ人 æĿ¥è¯´ +æĹ¶ éļĶ +ä¸į管 ä½ł +缸 è¿ij +è´µ éĩijå±ŀ +ä¹Łä¸į åı¯èĥ½ +ç²ī æľ« +åįĹ çĵľ +çϽ 马 +åħī æºIJ +éĩij å¥ĸ +çĭ¬ è§Ĵ +çĭ¬è§Ĵ åħ½ +妨 ç¢į +ç»Ļ åĬĽ +ä½Ĩ ä»į +å¼łå®¶ åı£ +èIJ¬ åħĥ +渲 æŁĵ +éķ¿å¤§ äºĨ +è®°èĢħ äºĨè§£ +æĢĢ çĿĢ +è¦ģ åѦä¼ļ +游æĪı 代 +游æĪı代 ç»ĥ +äºĮ çϾ +æĦıè¯Ĩ å½¢æĢģ +çİ º +计åĪĴ çĶŁèĤ² +æī¾ åĩĨ +åħ° èĬ± +è¿Ļ座 åŁİå¸Ĥ +污 æ³¥ +å®ĺæĸ¹ 微信 +å½Ĵ å±ŀ +æ°§ æ°Ķ +éģİç¨ĭ ä¸Ń +åį°è±¡ æ·±åĪ» +稳 妥 +çµIJ æĿŁ +åŃķ æľŁ +çī¹ æĿĥ +åĿļ åĽº +顺 åĬ¿ +æŀľ èͬ +éĨ« 師 +åİ ® +ä¹Łæĺ¯ å¦ĤæŃ¤ +é¦Ĵ 头 +缸 åĬ© +å¹² 线 +ä¸Ģ æľ¬ä¹¦ +ç» ¥ +æĮ¯ å¥ĭ +èĤ¾ èĦı +åĭķ çī© +é£ŀ è·ĥ +èıľ åĵģ +å¤ļ ä½Ļ +å¤ļä½Ļ çļĦ +éĢĿ ä¸ĸ +æģĭ 人 +å¼Ģåıij åĪ©ç͍ +顺 丰 +éĩİ å¿ĥ +æł¡ å¤ĸ +æģIJ é¾Ļ +éĿ¢ åħ· +éķ¿ è¾Ī +éļı å¤Ħ +éļıå¤Ħ åı¯è§ģ +ç´§ 缺 +éĩį ä¸Ń +éĩįä¸Ń ä¹ĭ +éĩįä¸Ńä¹ĭ éĩį +奥 æĸ¯ +奥æĸ¯ åį¡ +ä¸Ģ个 å¤ļ +ä¸Ģ个å¤ļ æľĪ +ä¸įåı¯ 缺å°ij +æĸ° æł¼å±Ģ +æıIJ æĮ¯ +è¡Į è´¿ +æ¼Ĥ æµģ +èģĬ åŁİ +åħ´ 建 +è´¨ æ£Ģ +ç§ģæľį 游æĪı +æĽ´ éĩįè¦ģ +è´ ® +çħ ľ +转åıĺ 为 +è¿Ļ 两年 +ä¿Ŀ é²ľ +æī§ æķĻ +çĥ ¨ +å¼Ģåıij 建设 +è¿IJèIJ¥ 管çIJĨ +误 å·® +京 åī§ +å¸IJ åı· +å·¥ä½ľ ä½ľé£İ +ä¸ĸ ä¿Ĺ +çϽ 宫 +天 åĽ½ +å¤©åĽ½ ç»§ç»Ń +å·´ æĸ¯ +èIJ¥ åĪ© +åĵģ æł¼ +æĿijæ°ij 们 +æĪ¿ 车 +çŃī çĹĩçĬ¶ +å¦Ĥ å®ŀ +å® ¸ +å±Ĥ 级 +éĶĻ è¿ĩäºĨ +ç»ĵ å®ŀ +ç¬ij èĦ¸ +羣å®ŀ æĢ§ +éĥ½å¸Ĥ æĬ¥ +é¥Ń èıľ +åºĶ 注æĦı +æĬ½ çĥŁ +伪 éĢł +åīį ä¸Ģ天 +éŃĶ é¾Ļ +éŃĶé¾Ļ 令çīĮ +约 è°Ī +绣çѹ æİ¨è¿Ľ +让 ç͍æĪ· +åħ¨éĿ¢ èIJ½å®ŀ +å¼Ħ å¾Ĺ +è°Ī æģĭçα +鸣 æĪIJéķ¿ +鸣æĪIJéķ¿ è®° +æ´ĭ æ´ĭ +çĸı æķ£ +éĿ¢ç§¯ 约 +æµĵ 缩 +æĸ¯ é¡¿ +çĶŁæĢģ åľĪ +æī§ 导 +ç§» éĢģ +齿 è½® +æł¹æľ¬ å°±ä¸į +缩 åĩı +èµ° ä¸ĭåİ» +çĿ« æ¯Ľ +ä¹Łä¸į éĶĻ +åıįæĺł åĩº +èĭ¦ æģ¼ +缸åħ³ æĶ¿çŃĸ +é«ĺ 楼 +ç²ī èī² +æĬķèµĦ é¢Ŀ +ä¸į ç»ı +ä¸įç»ı æĦı +å®ģ æĦ¿ +èĪĮ 头 +æ»ĭ çĶŁ +å®ģ åİ¿ +åīįåĪĹ èħº +åĩ ³ +é£Ł 欲 +åıĸ èĥľ +éĻ¢ åŃIJ +ç´łè´¨ æķĻèĤ² +滨 å·ŀ +æĬ¢ æĬĵ +å¼Ĥ åij³ +åĴ ļ +åĬ į +宽 éĺĶ +æļ´ 涨 +æĥł åıĬ +è§Ħ ç¨ĭ +ä¾Ľ åħ» +éĢģ å¾Ģ +å±± åºĦ +举 äºļ +å±ķ é¦Ĩ +è§£ éĶģ +æĹł è§Ĩ +éĻį èIJ½ +è¿ŀ äºij +è¿ŀäºij 港 +åıĤ è°ĭ +çİ ĸ +ç¬ ĥ +èĢĹ è´¹ +æī¿ å¾· +社ä¼ļ æķĪçĽĬ +åįĹæµ· ç½ij +åĪĽ 伤 +èIJ ± +åħħ æ²Ľ +ç½ijç«Ļ 建设 +大 åºĨ +åĨį éĢł +åŃĹ æł· +åħ¨æ°ij åģ¥èº« +èĮ« èĮ« +æµ® åĬ¨ +åīį åı° +å¢ŀ 设 +éĢĽ è¡Ĺ +åĢĴ éĹŃ +æ³ķå¾ĭ 顾éĹ® +çĸ ® +çĹħ çĹĩ +空 åīį +请 æķĻ +èĥľ ä»» +æĿĢ èıĮ +æĪĺæĸĹ æľº +ç»ĺ åζ +å¤Ħ æĸ¹ +çªģ åĽ´ +çĮ« åĴª +æĬ¥åijĬ æĺ¾ç¤º +ç¿ Ł +çķ¶ åľ° +æľĢ éļ¾ +纪 å§Ķ书记 +ä½İ åİĭ +èĻļ 空 +è¿Ļéĥ¨ ç͵影 +产ä¸ļ åįĩ级 +è°· çα +è°·çα åĩĮ +æĬ¼ éĩij +女 æĸ¹ +éĴ» çłĶ +æļĹ æļĹ +è¿· ä½ł +æīĢ è¬Ĥ +å¨ģ å»ī +å¼Ģ æľĹ +å² Ķ +çģ« çĤ¬ +åIJĪçIJĨ æĢ§ +åħ¬ åĬŀ +ä¼ļ ä¼ļéķ¿ +éĺ´ è°ĭ +å¼Ģ å±Ģ +æĻ®éĢļ è¯Ŀ +åį¡ æĭī +å°ij åIJĥ +éĹª èĢĢ +æŀľ æ±ģ +æī§è¡Į åĬĽ +è° Ľ +æĬ¢ åĬ« +é«ĺéĢŁ åıijå±ķ +éŁ ¬ +åįĹ æ²Ļ +é«ĺçŃī åŃ¦æł¡ +æį¢ 个 +åı¯èĥ½ åŃĺåľ¨ +æĬ Ĵ +è°± åĨĻ +被 æĬĵ +æĿ¯ åŃIJ +èĬĤèĥ½ åĩıæİĴ +æ°ĶåĢĻ åıĺåĮĸ +åĪĨ åĪ¥ +ä¸Ń æŀ¢ +欢 åij¼ +åħī 纤 +è¿Ļ 群 +çľ¼ çķĮ +åħ±åIJĮ åıijå±ķ +çݰ ä»Ĭ +éĹ» è¨Ģ +çī¹èī² å°ıéķĩ +æķij 人 +éĻį æ°´ +ä¸ĸçķĮ ä¸Ģæµģ +å°± é¤IJ +çŀ ¥ +å¤į ä»ĩ +ç¾½ æ¯Ľ +ç¾½æ¯Ľ çIJĥ +è´© åįĸ +æºIJ æ³ī +æĢ»ä½ĵ è§ĦåĪĴ +åĬ¨ æĦŁ +ä¸Ģ 审 +åĢŁ éĴ± +è§ģ æķĪ +èĬ± èįī +åIJĮ ä¸ļ +æŁ¥ è©¢ +åĽ½éĻħ åIJĪä½ľ +ä¾Ľ åĽ¾ +åģ ´ +æł ĵ +缸 éĢļ +è°Ī åıĬ +è¿ĩç¨ĭ å½ĵä¸Ń +é¦Ļ èıĩ +åįģåĽĽ æĿ¡ +ä¸Ģå¼Ģå§ĭ å°± +ä¸ĵ åijĺ +æĺİ é¡¯ +æīĵéĢł åĩº +ä¸ĭéĿ¢ æĪij们 +æľº æ²¹ +åı° è¯į +åŃIJ å¼Ł +æľĢ 常è§ģçļĦ +æĪij è®°å¾Ĺ +ç» ° +æĤ¬ æµ® +è¿ĺ 羣æĺ¯ +æĮĤ åı· +åıĭ åĸĦ +éĩį 伤 +çħ§ 亮 +æŃ¦ èѦ +åĩºçݰ éĹ®é¢ĺ +è¸Ĭ è·ĥ +åľ°çIJĥ ä¸Ĭ +å¸Ĥ 人大 +åıĹ害 人 +å² IJ +åIJĮ åѸ +éĩijèŀį å¸Ĥåľº +æľīçļĦ çݩ家 +å¸Ĥ æķĻèĤ² +å¸ĤæķĻèĤ² å±Ģ +åIJĦ å¼Ĥ +ç·ļ ä¸Ĭ +æģ º +æľī 大éĩıçļĦ +åķĨ æĬ¥ +åįķ åįķ +åħ¨ é¢Ŀ +ä¾ĿæĹ§ æĺ¯ +好 åĩłä¸ª +åĸ µ +éĩį æķ´ +çĶŁæ´» è´¨éĩı +æİ¢ 访 +åį° èĬ± +缼 è¡Į +å¾® è§Ĥ +èĪį å¾Ĺ +åºŁå¼ĥ çī© +积 èĵĦ +å®ļ å±ħ +æĤ ¼ +èĮ ¸ +çļĦ 帮åĬ© +çļĦ帮åĬ© ä¸ĭ +亿 åIJ¨ +åŃĶ éĽĢ +è¿ĻæĿ¡ è·¯ +é¥ µ +æĦĪ åĬł +éķ į +ä½ľ æ¡Ī +èįĶ æŀĿ +太 å°ij +è·» 身 +åħ¬çĽĬ æ´»åĬ¨ +çϽ æĸij +æĬĢæľ¯ æ°´å¹³ +å¸ § +æĹł çŁ¥ +åºĶ该 æĢİä¹Ī +éĢĢ å¸Ĥ +æ¸ Ń +åħ» çĮª +é© ¼ +群 å²Ľ +大 åį« +ä¹ĺ çĶ¨è½¦ +èı² å°Ķ +è´´ åIJ§ +åģľ ä¸ĭæĿ¥ +æľīæľº ç»ĵåIJĪ +åĪ» èĭ¦ +çļĦ åľ° +çļĦåľ° æŃ¥ +è¯Ĭ æīĢ +å¼Ģ æĪĺ +èĢģ çīĮ +çѹ çłģ +åħ«å¤§ 以æĿ¥ +楼 æĪ¿ +åŃĻ æĤŁ +åŃĻæĤŁ ç©º +åħĴ åŃIJ +第ä¸Ģ æĿ¡ +社交 åªĴä½ĵ +æĥ³ èµ·æĿ¥ +大 æ´ĭ +æĭ¼ éŁ³ +è¿Ľ åįļä¼ļ +è¿ĩ åħ³ +æ² ¼ +ç©¿ æIJŃ +éĤ£ ä¸Ģ天 +çł´ éŨ +æĬķæłĩ 人 +èµ¢ å®¶ +èĻļ å¼± +æ¿ ĥ +å®ī æ£Ģ +客 å®¶ +çĭ¬ç«ĭ èij£äºĭ +æīĭ åĬ¿ +åīµ éĢł +åľĨ满 å®ĮæĪIJ +为主 线 +好å¥ĩ å¿ĥ +é¢Ĩ åľŁ +çª ĸ +åħ¸åŀĭ æ¡Īä¾ĭ +çªģåıij äºĭä»¶ +åºķ æ°Ķ +头 æĻķ +å®Ľ å¦Ĥ +è§ ¸ +æ¸ħ æ·¡ +åļ ¼ +åģľ ç͵ +ç²ī å°ĺ +éĻįä½İ æĪIJæľ¬ +æĶ¾ æīĭ +è®°èĢħ 表示 +æĭĸ å»¶ +éª ĩ +æ®ĭ å¿į +çľģ æķĻèĤ² +çľģæķĻèĤ² åİħ +é«ĺ é¢Ŀ +éĦ Ļ +æ¥ ŀ +åĨħ ç§ij +èIJ¥ä¸ļ é¢Ŀ +åŁº çŁ³ +æµģ æ·Į +主 æĹ¨ +éĺIJ éĩĬ +建 åįİ +æĥĬ åı¹ +çī¢åĽº æłijç«ĭ +æĺ¯åIJ¦ åŃĺåľ¨ +建 åĨĽ +éĽ¾ éľ¾ +åħ¬ 认 +åħ¬è®¤ çļĦ +æ°¨ åŁº +æ°¨åŁº éħ¸ +åīį åĩłå¹´ +åι éĤ£ +æ±Ł 举 +å·¥ æ¥Ń +ä¸ĢçĤ¹ ä¹Łä¸į +ä¿® 士 +äºĨä¸Ģ éģį +åĪ ģ +æ»ļ æ»ļ +åĪĨ æł¡ +羣 çα +è¡Ģ èĦī +æĢ¥ åī§ +ä¸Ģ群 人 +ç¾ ¯ +æĪIJ é¾Ļ +ç²¾ç¥ŀ çĹħ +缸åħ³ 人åijĺ +éĿĵ 丽 +ä¸ī åŃ£åº¦ +åĪĴ å®ļ +ä¸ĸçķĮ 第ä¸Ģ +éĢļ ä¿Ĺ +åķĨä¸ļ åľ°äº§ +åĬŁèĥ½ æĢ§ +èµĦæľ¬ 主ä¹ī +详 è§ģ +æĬĵ æįķ +æĸĩ æĺĮ +å®Ŀ å®ī +è£ħéħį å¼ı +æºIJ æºIJ +æºIJæºIJ ä¸įæĸŃ +çĶŁ æĢķ +纵 åIJij +å£ ½ +çľ¼ è¢ĭ +èĤī ä½ĵ +åı¤ ä»Ĭ +èŀį åªĴä½ĵ +åģ ī +æł¼ æľĥåĵ¡ +çĥ · +åĬŁ ç͍ +æīŃ çŁ© +绿èī² éĢļéģĵ +åī§ ç»Ħ +å¼± åĬ¿ +è´¨éĩı éĹ®é¢ĺ +éĻIJ é¢Ŀ +éª Ĩ +éģµ ä¹ī +å¯Ŀ 室 +æĥ³ 念 +åł± åijĬ +ä»ħ 次 +ä»ħ次 äºİ +èŀį åĪĽ +æĭĽèģĺ ä¼ļ +åºĬ åŀ« +转åŀĭ åıijå±ķ +ä¸ŃåĽ½ çĶµä¿¡ +åIJ¬ è¯Ŀ +è«ĭ æ±Ĥ +大éĥ¨åĪĨ 人 +æ´» å¾Ĺ +åĵŃ æ³£ +è¶ Ļ +åıijçĹħ çİĩ +ä¸į 符 +åĨĽ å®ĺ +é¢Ī æ¤İ +æĸ°åĨł çĸ«æĥħ +æŁ¬ åŁĶ +æŁ¬åŁĶ 寨 +ä»»ä½ķ å½¢å¼ı +人 éĻħ +人éĻħ åħ³ç³» +æĢ» æī¿åĮħ +å¹³åĿĩ æ¯ı +æģŃ åĸľ +åĦ ĺ +åħµ 马 +è¿Ł åΰ +å·¥ 伤 +çīĪæĿĥ å½Ĵ +çīĪæĿĥå½Ĵ åİŁ +æĭ¥ æĬ¤ +ç³Ĭ æ¶Ĥ +å¹² æ¶ī +å°ij ä¸įäºĨ +æĥ³ æī¾ +è´¹ çİĩ +该 éĻ¢ +èŀį åĮĸ +è¿İ åIJĪ +è§ĨåIJ¬ èĬĤ缮 +æł¼ ç¶²ç«Ļ +çľī æ¯Ľ +欢è¿İ 大家 +å®¶åºŃ æķĻèĤ² +ä¾µ èļĢ +ç»Ļ ä½łä»¬ +è¡Ģæ¶² 循çݯ +å¯Ħ æīĺ +å°ĸ åı« +以ä¸ĭ åĩłä¸ª +è¿ĺ 以为 +åħ¶ä»ĸ çݩ家 +ç¬ij ç¬ij +æīĵ åIJ¬ +èĩªçĦ¶ ç§ijåѦ +åŁº ç«Ļ +ä¹Ŀ å·ŀ +ä¿Ŀ 驾 +ä¿Ŀ驾 æĬ¤ +ä¿Ŀ驾æĬ¤ èĪª +æĶ¾ çľ¼ +çŁ¥åIJį ä¼ģä¸ļ +ç¸ ® +ç¨ ½ +æļ ĩ +使ç͍ 網路 +é¢Ħ çķĻ +大 象 +åıijæĺİ ä¸ĵåĪ© +æĸĩ 娱 +éĢł ç¦ı +湿 润 +éĿ¢ æĿ¡ +æ¶Īè´¹ åįĩ级 +è®Ĭ å¾Ĺ +åĩł åIJį +ä» Ħ +认 æ¸ħ +è¿ľ æĻ¯ +æıĴ 座 +诸 侯 +åıĺ æĢģ +ç¦ı 彩 +è´§ æŀ¶ +失 æİ§ +ç§»åĬ¨ 端 +ä¸Ĭ åı¸ +éĢł 纸 +å¸ĥ æľĹ +çĴ ĩ +åı° åįĹ +åĮĹ京 åĨ¬å¥¥ +èĵĿ çīĻ +éķ¿ çŁŃ +æĬĺ å°Ħ +ç»ij æŀ¶ +å¯Ĵ åģĩ +转 åŁºåĽł +æĢ¥ äºİ +æŃ£ åĵģ +åħħ 滿 +大 纲 +æĬĹ ä½ĵ +è¨ĵ ç·´ +æĶ¶ ç´§ +æ¯Ķ è³½ +åħµ åĬĽ +æľ¬ æĽ¸ +äºĮ 代 +æĢ¥ è¯Ĭ +æĸĩ æ¡Ī +ç»ı åķĨ +æĻ¨ æĬ¥ +æ£ ĺ +æĢ»ä¹¦è®° åľ¨ +åıĹ éĤĢ +äºĶ åĽĽ +å²Ń åįĹ +çα åIJĥ +åŁĥ å°Ķ +å¿ĥ å¢ĥ +è¦ĨçĽĸ éĿ¢ +å®ŀåľ¨æĺ¯ 太 +æł¹ åºķ +纷纷 表示 +åĹ ħ +éļıçĿĢ æĹ¶éĹ´ +åİĨåı² æĤłä¹ħ +éħ ī +æĢ» éĺŁ +主é¢ĺ æ´»åĬ¨ +éĹ® åį· +é©¿ ç«Ļ +æı¡ ä½ı +åı¯èĥ½ 导èĩ´ +æ°ij éĸĵ +éĸĭ åķŁ +ä½Ĩ ä¸įéĻIJ +ä½Ĩä¸įéĻIJ äºİ +åįģ éĩĮ +å¨ ¥ +æįŁ èĢĹ +çĸı 导 +çݯ æ°§ +ç¥ŀ éĢļ +çα å°Ķ +çαå°Ķ åħ° +æľ´ å®ŀ +å¿« æĬ¥ +æĶ¶ åıĹ +æĪĸ 許 +èĥĮ éĿ¢ +æĸĩåĮĸ ä¼łåªĴ +ä¸ī åĢĭ +æĶ» åĬ¿ +å®ī 举 +å®ī举 å°¼ +åĿĩ å·² +顾 èĻij +éĦ Ń +è¿Ļå®¶ åħ¬åı¸ +åħ¬åijĬ ç§° +æıIJä¾Ľ ä¼ĺè´¨ +稳æŃ¥ æİ¨è¿Ľ +å¤į è¯ķ +å°Ĩ é¢Ĩ +è°Ī èµ· +å¨ Ħ +è¿ŀ 线 +æ©Ł éĹľ +åºĶç͍ åľºæĻ¯ +çĶ» åĥı +è´¢ è¿IJ +ä¿Ŀ éļª +çĹħ çIJĨ +æ¯Ľ 主å¸Ń +ä¸Ŀ 毫ä¸į +çα å¥ĩ +çαå¥ĩ èīº +ä¸ĵå®¶ ç»Ħ +åij¼ åͤ +éĭ ¼ +çģ ¸ +é¢ĨåħĪ åľ°ä½į +æıIJ æĭĶ +龸 éģĵ +å±± åĿ¡ +èĿ İ +沸 èħ¾ +该 项 +ä»Ĭ çĶŁ +ä¸Ģç¯ĩ æĸĩ竳 +æĸ¹å¼ı è¿Ľè¡Į +é»ij 客 +æĶ¹ åĬ¨ +主 é¡Į +æķ£ å¸ĥ +ä»Ģä¹Ī åľ°æĸ¹ +åĮĸ åIJĪ +åĮĸåIJĪ çī© +éĿĻ ç͵ +æĢ» æĶ¶åħ¥ +å§Ķ ç»Ħç»ĩ +å§Ķç»Ħç»ĩ éĥ¨ +éĿĻ æĢģ +èĢģ åŃĹåı· +室 åıĭ +éĥ½ä¸į æķ¢ +æŀ¶ åŃIJ +çģµ æķı +审 è§Ĩ +æĤ£ åĦ¿ +å±± 寨 +èĸª èµĦ +é©° æı´ +éĥ¨åĪĨ åĨħ容 +好 ä¼¼ +æĪIJåijĺ åĽ½ +åľ¨æĪij çľĭæĿ¥ +åħ³æ³¨ 度 +éĻĪ æŁIJ +è¿Ļç§į äºĭæĥħ +éĢī å®ļ +ç²¾ åŃIJ +å£ģ çĶ» +æ±Ł æ·® +é«ĺ æĺĤ +æł¼ åĬĽ +è¼ © +åѦ åłĤ +æĤ¨ åIJĮæĦı +ä¸ĢåĪĩ éĥ½æĺ¯ +æ½ ¤ +éĸ ĥ +å¸ĮæľĽ èĩªå·± +ä¿ ĺ +æ±Ł åİ¿ +æ³ ¾ +ç§ij æķĻ +æīĵ è¿Ľ +ä¸į æħİ +å¯Ĵ åĨ¬ +æ¸Ķ æ°ij +鼷 æĸ¯ +主 å®° +æĹħ游 度åģĩ +ç͵åŃIJ éĤ®ä»¶ +æ±Ĥ å©ļ +éļİ æ®µ +åģ¥èº« æĪ¿ +注æĺİ åĩºå¤Ħ +äºĭæķħ åıijçĶŁ +级 以ä¸Ĭ +åŃĺ æ´» +æĸ½ èĤ¥ +èľľ èľĤ +åµ © +æĮĸæİĺ æľº +æĬĹ æĭĴ +ä¼ł 导 +æĺ¯ä»Ģä¹Ī åij¢ +ä¸Ĭå¹´ åIJĮæľŁ +建 åħļ +çĶŁ æħĭ +ä¿Ŀ ä½ı +款 车åŀĭ +人 èĦī +éļIJ èͽ +失 æķĪ +éģ¿ åŃķ +ç®Ģ 便 +谢谢 ä½ł +å®Ī ä½ı +æĶ¾ æĺł +è¨Ī çķ« +çݰ代 çµģ +é¤IJ 廳 +æķħ å±ħ +大 大å°ı +大大å°ı å°ı +çī¹åĪ« 声æĺİ +éģį åıĬ +å¿ĥçIJĨ åĴ¨è¯¢ +è³ ´ +çĮ® è¡Ģ +å·²ç»ı è¾¾åΰ +æīĵ æĭĽåij¼ +åıĮ è¾¹ +ä¸Ģæĸ¹éĿ¢ æĺ¯ +å´ĩ å°ļ +éĺ¿ å¯Į +éĺ¿å¯Į æ±Ĺ +æĮģ æľī人 +è± ģ +é£İ çŃĿ +åĬ¨ èį¡ +äºĨä¸Ģ ä¼ļ +äºĨä¸Ģä¼ļ åĦ¿ +ä¸ĩ 象 +çľĭ ç͵è§Ĩ +åįģä¸ī æĿ¡ +çĮĽ çĥĪ +è¦ģ ä¸įçĦ¶ +太æŀģ æĭ³ +å¼ķ çĪĨ +ç»ıè¿ĩ å¤ļå¹´ +游æĪı éĩĮçļĦ +é¾Ļ æ³ī +æłĩ éħį +è®ĵ ä»ĸåĢij +éĢł æŀĹ +åĮºåŁŁ æĢ§ +亿 ä¸ĩ +æĪĺçķ¥ å¸ĥå±Ģ +éķĩ æĶ¿åºľ +åĶ® 票 +çĶŁäº§ å·¥èīº +éķĩ åħļå§Ķ +ä¸Ńå°ı åŀĭ +æľ¨ è̳ +æ²³ è¾¹ +èĦ¾ èĥĥ +欢è¿İ æĤ¨ +åıĺ å¼Ĥ +缤 纷 +åŀĥåľ¾ æ¡¶ +辩 è¯ģ +车 åºĵ +æ¯Ķ çİĩ +åħ´ æĹº +详ç»Ĩ äºĨè§£ +å®ī å±ħ +çħ§ æĸĻ +æĸ¹ æīį +èµ ¦ +åĨ ķ +å¥Ķ èµ´ +å®Ŀ 鸡 +åľº åĿĩ +缮åīį æŃ£åľ¨ +åIJŀ åϬ +è¿° èģĮ +æĩ µ +å¥ĩ çijŀ +ä»į å°Ĩ +èĪī 辦 +å·¥åķĨ å±Ģ +å¡ij èĥ¶ +åĬŀ å®ŀäºĭ +æĸ¹ æĸ¹éĿ¢ +æĸ¹æĸ¹éĿ¢ éĿ¢ +æĸĩåĮĸ èĬĤ +åħ¥ èģĮ +é¸ ¥ +ç©¿ éĢı +以 ä¹łè¿ijå¹³ +åį± éļª +æľ¦ èĥ§ +åİĨåı² æĢ§ +æķŀ å¼Ģ +ä¼Ļä¼´ åħ³ç³» +çŁ¿ åĮº +åĽ½éĻħ åľ¨çº¿ +ä¼łå¥ĩ éĩĮéĿ¢ +è¿ij äºĽ +è¿ijäºĽ å¹´ +åĬ£ åĬ¿ +æĶ»åĩ» åĬĽ +æĻº éĢł +ç¦ § +çİĭ åħĪçĶŁ +éĨ« çĶŁ +åĽĽ 项 +å®ŀ æĻ¯ +åĪĿ åĪĽ +å¿ĥ 裡 +æĻ¶ ä½ĵ +交 éĻħ +让 æ¶Īè´¹èĢħ +课 æĸĩ +æİĴ æ°Ķ +å¹¶ä¸į æĦıåij³ +缸 声 +第ä¸Ģ å±Ĭ +åİŁ èijĹ +éĽ ľ +没æľī 太大 +è¡¥ æ°´ +çµģ ä¼ģä¸ļ +第äºĮ æī¹ +åħ¶å®ĥ éĹ®é¢ĺ +æİĮ éŨ +责任 å¿ĥ +é¤IJ åħ· +ç¾Ĭ æ¯Ľ +没æľī å¿ħè¦ģ +ä¹IJ åĽ¢ +è¿Ľ åŁİ +ä¸ĢçĤ¹ åĦ¿ +身 å½¢ +çļ®èĤ¤ çĹħ +æĺ ± +å¢ŀ èĩ³ +èģ² æĺİ +æıIJ è´¨ +ä½ĵèĤ² åľº +çѹ 建 +é¬ Ĩ +车 çīĮ +éļĶ éŁ³ +è´Łè´£ åIJĮå¿Ĺ +丰 ç¡ķ +ä½Ľ éĻĢ +äºī åIJµ +åº ¶ +æ·¡ æ°´ +å°ı çĶ·åŃ© +ç§ģ èĩª +åĮĸ è¿Ľç¨ĭ +æĪĺ士 æĿ¥è¯´ +æ²¹ èħ» +èĦ±è´« èĩ´å¯Į +æĹ¥å¸¸ å·¥ä½ľ +交 èŀį +åĨľ è´¸ +åĨľè´¸ å¸Ĥåľº +åĵĪ çĻ» +ç͵ è´¹ +èµ ĺ +åıĮ èħ¿ +æĵĶ å¿ĥ +æĿ¥ 形容 +使åij½ æĦŁ +éĤ£ä¹Ī ç®Ģåįķ +èĬĻ èĵī +åĢŁæ¬¾ 人 +ç§Ģ 丽 +è®ĵ ä»ĸ +严åİī æīĵåĩ» +è³ ŀ +æļ « +çħ¤ æ°Ķ +çά ä¸Ĭ +æ½ĩ æ´Ĵ +太 ä¹ħ +åij½ åIJį为 +è·¯ çͱ +è·¯çͱ åύ +é© ¯ +æıIJ æĹ© +æĬĹåĩ» çĸ«æĥħ +åĩ Ľ +交 åıĭ +éĶĢåĶ® æ¸łéģĵ +毫ä¸į çĬ¹è±« +èIJ¥ åľ° +çłĶç©¶ 表æĺİ +é±¼ ç±» +æį¢ å±Ĭ +æİ¡ åıĸ +çī Ĩ +缼 å¼Ģ +æ²§ æ¡ij +åºŃ 审 +ç»ı æŁ¥ +åĬł å¼· +缸æ¯Ķ äºİ +ä¸ĵ çıŃ +ä½ĵ åŀĭ +被 害 +被害 人 +æĶ¶ 款 +åħ·æľī èī¯å¥½ +é«ĺå³° æľŁ +åģı ä½İ +åĦ Ł +åĨľä¸ļ ç§ijæĬĢ +ç®Ĭ æĥħåĨµ +å¦Ĥæŀľ çݩ家 +éķ¿ çº¦ +第åħŃ å±Ĭ +åħ¬å¼Ģ æĭĽèģĺ +åĪĩ æĸŃ +è¿« 使 +çĸĹ ç¨ĭ +第äºĮ ç§į +ä¸į åħį +å¹² èѦ +çŁ³ 榴 +åĹ £ +两 ç±» +çε 士 +åŁİ乡 å±ħæ°ij +æŃ¤ 项 +缴 è¾ĸ +缴è¾ĸ å¸Ĥ +åij¼ åºĶ +éĴ ¯ +ç¦ı å¾· +æľº 身 +æĵį åľº +æ¿Ĵ 临 +人群 ä¸Ń +èĤ¡ æ°ij +åŃ ½ +æ³ķ åħ° +é¨ İ +糯 ç±³ +æĢ» çļĦ +æĢ»çļĦ æĿ¥è¯´ +åħ¸ éĽħ +æĸ° éĻĪ +æĸ°éĻĪ ä»£è°¢ +缮 çĿ¹ +é¢Ħ è¨Ģ +è·Į çł´ +æĸ° ç¯ĩ竳 +æ¯Ĵ æĢ§ +åĸĿ èĮ¶ +æŁ¥ èİ· +亮 丽 +çĶŁäº§ åķĨ +æĶ¹ æĪIJ +为äºĨ æĽ´å¥½ +æ·± 交 +深交 æīĢ +æİ ĥ +ä¹Ļ èĤĿ +泸 å·ŀ +åħĪè¿Ľ æĬĢæľ¯ +è¾ĵ ç»Ļ +æķ£ æĪ· +æĢĿç»´ æĸ¹å¼ı +åºĹ 主 +è°ĭ æ±Ĥ +游æĪı æĬĢå·§ +ä¸Ģå¹´ 级 +çľ¼ è§Ĵ +ä¸Ńä»ĭ æľºæŀĦ +å·§ åIJĪ +éĺ² çĽĹ +导 è´Ń +æĪ Ĭ +æĽ´ éĢĤåIJĪ +åŁºæľ¬ ä¿¡æģ¯ +马 ä¸ģ +åħ»æ®ĸ åľº +åıį è¿ĩæĿ¥ +æİ¨ å´ĩ +å¯ĨåĪĩ åħ³æ³¨ +åŁºéĩij ç»ıçIJĨ +æĮī éĶ® +åĨħéĥ¨ æİ§åζ +æĪIJåijĺ åįķä½į +æľ¯ è¯Ń +åζ æľį +åĪļ éľĢ +æ£Ģ ç´¢ +大大 æıIJé«ĺ +åģ¥åº· 管çIJĨ +èĩª æŃ¤ +客æĪ· éľĢæ±Ĥ +丰 èĥ¸ +èµ· éĩį +èµ·éĩį æľº +æ¬ł 缺 +æ¡Ī åŃIJ +æĥħ人 èĬĤ +åħļ æł¡ +è¢ ľ +该 åī§ +迷失 ä¼łå¥ĩ +ç»ļ 丽 +åķ ª +æĹł ç§ģ +é̲ ä¸ĢæŃ¥ +第ä¸Ģ 竳 +åύ åħ· +åĨľ èµĦ +確 實 +åºı åĪĹ +娱ä¹IJ å¹³åı° +èŀįèµĦ ç§Łèµģ +èµĦæºIJ åħ±äº« +èģ½ åΰ +æIJŀ å¾Ĺ +ç»§ç»Ń ä¿ĿæĮģ +åIJ¯ èĴĻ +çľ º +ä¸Ŀ è·¯ +设æĸ½ 建设 +æİ¥ åľ° +æİ¥åľ° æ°Ķ +第ä¸ī åŃ£åº¦ +åŁº è°ĥ +åıij éŁ³ +社ä¼ļ èµĦæľ¬ +éĽĩ 主 +è¿ŀ èĥľ +没 åķ¥ +å» ¢ +èµ¶ èµ´ +æ¼Ķ åĮĸ +åı¤ æĢª +çİĭ çĪ· +é¢Ħ åħĪ +å¼Ģ åħ· +åĽŀ é¦ĸ +åľ°ä¸ĭ æ°´ +å°ıç¼ĸ ä¸Ģèµ· +èµİ åĽŀ +åľ° è²Į +åĪĿ ä¸ī +åı¯ ç͍äºİ +éģĹ è¿¹ +è¿Ļ æī¹ +èĸª æ°´ +å¿ħçĦ¶ ä¼ļ +æ² ½ +éį ĭ +第ä¸Ģ éĥ¨ +åĪĬ çī© +å®ŀ ä¾ĭ +æ¸ħ åĩĢ +ä¸Ĭ èµĽåŃ£ +åĽ¾ 表 +éĤ® è½® +åĵª 裡 +缸 è§ģ +æī° ä¹± +æ¯ı æ¯ı +è¿Ļ è¾ĪåŃIJ +ç¡« éħ¸ +äºī 缸 +溯 æºIJ +åĩº ä¼Ĺ +çİī çŁ³ +åħ± çĶŁ +æĹ¶éĹ´ 段 +éĩįè¦ģ æĮĩ示 +æ¶Īè´¹ éľĢæ±Ĥ +éķ¿ éķ¿ +éķ¿éķ¿ çļĦ +å®ī æĬļ +å¢ŀ é«ĺ +æľ¬ è½® +亲 çľ¼ +é£İ æ³¢ +èĢģ å¦Ī +æĶ¶è´¹ æłĩåĩĨ +åĨħ éĻĨ +æĮ¥ åıij +åįĩ åѦ +èĥ¸ åīį +åģı è¿ľ +纯 æ´ģ +æĸ½å·¥ åįķä½į +身 ä»· +è´¢ åĬĽ +çº ¶ +è£ħ çͲ +æĺ¾ç¤º åύ +毫 åįĩ +æ·± çŁ¥ +è̶ ç© +èĢ¶ç© Į +è¾ĥ éĩı +åľ¨ è¿ĩ渡 +åľ¨è¿ĩ渡 æľŁ +èĮ Ĺ +ä¸Ģ个 æĺŁæľŁ +èĬ · +è´¿ èµĤ +æ¿ ķ +æĩĤ äºĭ +ç§ § +åħħ å½ĵ +åĽ½ ç«ĭ +èĬ± çĵ£ +éĤĦ è¦ģ +åħ¬ åľĴ +触 åĬ¨ +æ³° å·ŀ +ä»Ģä¹Ī æł· +æ»ĭ åħ» +è¯Ħ åΤ +æĮ¥ æīĭ +èĦ Ī +å§¥ å§¥ +è¿IJ è´¹ +æ¯ħ åĬĽ +å¿ĥ æĻº +ä¸į æİĴéϤ +第ä¸ī 代 +éĢĢ è´§ +æĺŁ éĻħ +æ°¸ åĪ© +æĬ¤ åį« +çıŃ è½¦ +è¨Ģ è¡Į +ç¹ ª +主åĬ¨ æĢ§ +å·¥ç¨ĭ è´¨éĩı +éĥĬ åĮº +ä¸Ģ æłĭ +ä½Ĩ å®ŀéĻħä¸Ĭ +ä¸ī大 èģĮä¸ļ +åij¼ åı« +女 åħĴ +è¯ģåΏ æĬķèµĦ +èĢĥ æħ® +çĤ« èĢĢ +æ²» 好 +åĺ ¶ +èĥ ¤ +åħīä¼ı åıijç͵ +åĩł æŃ¥ +æīĢ æīĢ +æīĢæīĢ éķ¿ +çħ§ æł· +åĵ¥ 们 +è¯ Ľ +è¿Ļä¸Ģ åĪ» +çŁ¿ çī©è´¨ +ä¸įå¾Ĺ å·² +åIJĮ 缣 +ç»Ĩ å¾® +è·¯ èĻİ +çϾ èĬ± +æ·· æ²Į +ä¸Ĭæµ· è¯ģåΏ +éĢĢ ç¨İ +èµŀ åı¹ +æī®æ¼Ķ 游æĪı +åIJį åĪĹ +åIJįåĪĹ åīį +åIJįåĪĹåīį èĮħ +ç±³ å°Ķ +ä»Ģä¹Ī åİŁåĽł +å®īåħ¨ ä¿Ŀéļľ +ä¸Ģåıª æīĭ +ä¹³ ä¸ļ +ä¸į çĶĺ +æĥħ åķĨ +æĮ¡ ä½ı +åİŁåĽł ä¹ĭä¸Ģ +è¿Ļ 两天 +çĥĺ çĦĻ +è± ¬ +ä½ł 以为 +没 è§ģè¿ĩ +åĵªå®¶ 好 +åīį ä»» +è¿Ľ è´§ +éĢĢ åĽŀ +串 èģĶ +èĩ³ æĸ¼ +åĨ° æ·ĩ +åĨ°æ·ĩ æ·ĭ +æŁ¥çľĭ 详æĥħ +çı¾ 實 +æİ¨ æµĭ +æİ¥ æīĭ +éļ¶ å±ŀäºİ +åŁİå¸Ĥ 群 +æĿİ åħĪçĶŁ +çŁ¿ æ³īæ°´ +çī¹ ä»· +æĽ´å¤ļ 精彩 +ç¨ĭ å¼ı +读 æĩĤ +å±ı èͽ +奥 æŀĹ +奥æŀĹ åĮ¹ +奥æŀĹåĮ¹ åħĭ +红 èĸ¯ +å¥ ® +å®Ŀ çİī +ç¶² 絡 +è² § +欧 å¼ı +çϽ ç³ĸ +èĩªçĦ¶ çģ¾å®³ +åijĬè¯ī 她 +å» ļ +çĤ¹åĩ» æŁ¥çľĭ +é£İ 湿 +èµĦ产 éĩįç»Ħ +ä¹Łä¸į ä¾ĭå¤ĸ +åįĬ 个å°ıæĹ¶ +åIJ¸å¼ķ æĽ´å¤ļ +æĹ¶éĹ´ èĬĤçĤ¹ +æĶ¶ 纳 +åIJ¸ æ¯Ĵ +èĢģ 乡 +çIJ ħ +æľĢ çµĤ +åıį æĦŁ +ç͍ 微信 +çĶ¨å¾®ä¿¡ æī« +éĢŁ çİĩ +大 çĨĬçĮ« +åı¯ æĥ³ +åı¯æĥ³ èĢĮ +åı¯æĥ³èĢĮ çŁ¥ +åĴ § +èµ° åħ¥ +碳 éħ¸ +èĮĥ åĨ° +èĮĥåĨ° åĨ° +被 åΤ +积æŀģ æİ¨åĬ¨ +è¶³ è¶³ +ç²Ĵ åŃIJ +大 å®Ĺ +大å®Ĺ åķĨåĵģ +ç½ij绾 ç§ijæĬĢ +æĽ¼ åŁİ +å·² ä¹ħ +å·²ä¹ħ çļĦ +秦 çļĩ +秦çļĩ å²Ľ +ä»» æķĻ +å͝ ç¾İ +æ·¡ åĮĸ +æ¡Ĥ èĬ± +çŁ¥è¯Ĩ åĪĨåŃIJ +æĩĴ å¾Ĺ +主 åħ¬ +设计 çIJĨ念 +è³ º +æīĢ æıIJä¾Ľ +æīĢæıIJä¾Ľ ä¹ĭ +æĶ» åħĭ +åĤ ¾ +è¯Ń æ³ķ +åįĥ åı¤ +éĸĭ æĶ¾ +第ä¸Ģ èĬĤ +éĤĦ æ²Ĵ +éĢĥ çĶŁ +æ³ Ĺ +åİ¿ å§Ķ书记 +ä½ľèĢħ æīĢæľī +çħ ½ +ç» ħ +æł ħ +æľ´ ç´ł +çijķ çĸµ +åĮħ åĮħ +æ°ij主 åħļ +ä¸į è¿ľå¤Ħ +å¥ĩ å¼Ĥ +åĺ» åĺ» +æī ¼ +ç¿» å¼Ģ +æĢİ èĥ½ +éģ´ éĢī +è§£ éĩĭ +å¹¼ ç¨ļ +è¦ģ 好好 +è¶´ åľ¨ +ç´¢ åıĸ +ç»Ī çĶŁ +åħ¨ æµģç¨ĭ +éģ© çķ¶ +åįıè°ĥ åıijå±ķ +æĬ¥ ä»ĩ +ç§ijæĬĢ åĽŃ +ä»Ģä¹Ī éĥ½ä¸į +æľĢåIJİ ä¸Ģ次 +ç»Ļ人 ä¸Ģç§į +æł¸ å®ļ +被 åĪĹåħ¥ +æĦı æĥ³ä¸įåΰ +èĢĥ æŁ¥ +åľ¨æŃ¤ ä¹ĭåīį +æīĵ çIJĥ +è¶ĬæĿ¥è¶Ĭ å°ij +å®ļ å¾ĭ +è¡ĮæĶ¿ æľºåħ³ +ä½ıæĪ¿ åħ¬ç§¯ +å°ıå§IJ å§IJ +ä¸ī èı± +ä¿® è¡¥ +èŀĥ èŁ¹ +西 çͲ +æĢ ł +çŃī å¤ļ项 +产ä¸ļ éĽĨèģļ +ä»·æł¼ ä¸Ĭ涨 +åħ¬åħ± åľºæīĢ +è¢ĭ åŃIJ +æĨ§ æĨ¬ +çļĦæĸ¹å¼ı æĿ¥ +åΰ è´¦ +çģ ½ +å·´ èı² +å·´èı² çī¹ +æ¼Ķ ä¹ł +èŃ¦ç¤º æķĻèĤ² +çķı æĥ§ +å¼ķ æµģ +æĶ¶ æĶ¯ +å±Ĥ åĩº +å±Ĥåĩº ä¸į +å±Ĥåĩºä¸į ç©· +æijĩ æ»ļ +辦 çIJĨ +纵 è§Ĥ +æķij æµİ +å®¶ éĥ½çŁ¥éģĵ +åĮ ¯ +å°ı 鸣 +ä»» åĭĻ +计 åħ¥ +ç«ŀ éĢī +å¼ĢèįĴ æĹ¶æľŁ +åij¨ æģ© +åij¨æģ© æĿ¥ +交 ç»ĩ +çķ¢ æ¥Ń +æł¹æį® èĩªå·± +æĸ°äºº çݩ家 +åѵåĮĸ åύ +éĩĩ æļĸ +å¹³åĿĩ æ°´å¹³ +åħ¬å¼Ģ 课 +失 åĪ© +伺 æľį +çĬ ģ +忽 æĤł +主è¦ģ éĽĨä¸Ń +æ¤į æłij +æ¯Ĺ éĤ» +èĩº çģ£ +åĩºåĽ½ çķĻåѦ +æĬĹ éľĩ +æĥ© æĪĴ +å¹´åºķ åīį +åĴ¸ éĺ³ +æ°ij å±ħ +大çIJĨ çŁ³ +éĿ ³ +éķ ĸ +æ¸ħ è¿ľ +è£ħ è½½ +èĩ Ģ +å½± ä¸ļ +å¼Ł åħĦ +æĤ² è§Ĥ +çĿĢçľ¼ äºİ +æįį åį« +åī¥ å¤º +ç¯ Ĩ +å¾Ī éķ¿æĹ¶éĹ´ +è¥ Ł +第ä¸Ģ çϾ +ä¸ĢåĪĨ éĴ± +æĸ°éĹ» è®°èĢħ +éķ· æľŁ +æ³ķ æĪĺç»ĦåIJĪ +è°ģ çŁ¥éģĵ +èħ° éĥ¨ +æ±ī åł¡ +åħ¥ çĿ¡ +åįĸ æİī +æ¶Īè²» èĢħ +æĥ¯ ä¾ĭ +æĥ³ äºĨ +æĥ³äºĨ æĥ³ +èĢģæĹ§ å°ıåĮº +ä¼ł è¨Ģ +åĪĨæķ° 线 +æµģ 泪 +ç»Ħç»ĩ é¢Ĩ导 +äºļ åĨĽ +å¢ŀå̼ æľįåĬ¡ +å¾ ¹ +ä¼ ¶ +äºĽ 许 +å¸ĥ èݱ +强 æĤį +宫 å»· +绿 èĮ¶ +åĮ ¡ +å¾Ī æŃ£å¸¸ +æĺ¥ å¤ı +æ¯ Ļ +è¯Ħ æ¯Ķ +åĩ¡ äºĭ +æĬī æĭ© +åĢĴ éľī +éĩį 度 +åįıä¼ļ ä¼ļéķ¿ +å¿§ èĻij +ä¸ĭ ä¸Ģç¯ĩ +沪 æ·± +æĪ İ +æīĵ ä»Ĺ +åįĪ é¥Ń +å¹´é¾Ħ 段 +ä¸ŃåĽ½ è¶³çIJĥ +设计 æĸ¹æ¡Ī +åºĶç͍ æŁ¥çľĭ +é¢Ħ æĸĻ +åĹ ¡ +ç¥ĸ çζ +çļĦä¸Ģ åijĺ +æ´Ĺ å¹²åĩĢ +åİĨåı² æĸ° +åİĨåı²æĸ° é«ĺ +çĭ¬ åħ· +æħĭ 度 +æīĵ 交 +æīĵ交 éģĵ +é»Ħ çŁ³ +çĽ¼ æľĽ +çī§ åľº +转 弯 +åįĩ åįİ +åĨį ä¹Łæ²¡æľī +èĭ± æīį +æĽ´ åIJį为 +åĢŁ ç͍ +çºł éĶĻ +ç»Ŀ对 ä¸įä¼ļ +çİĭ çīĮ +çĽĨ åľ° +失 è°ĥ +好 象 +é³ ¥ +ä¿Ŀ ä¿® +åĽĽä¸ª èĩªä¿¡ +头 çļ® +åİŁ åīĩ +æĬ¥ æ¡Ī +奴 éļ¶ +å³ Ļ +è°ĥ æĸĻ +ä¹Ł 許 +èIJ½ åΰ +èIJ½åΰ å®ŀ +èIJ½åΰå®ŀ å¤Ħ +çĦļ çĥ§ +çĶŁæ´» çݯå¢ĥ +åºĶ åıĬæĹ¶ +è¶Ĭ è¿ĩ +æĦŁ è¬Ŀ +æĻ¯ å¾· +æĻ¯å¾· éķĩ +çĬ Ģ +身 éĤĬ +ç¨İåĬ¡ æĢ»å±Ģ +åĩĢ åľŁ +ä¾µ åįł +åĬ¨ å·¥ +å¹´ ä¹ĭ +å¹´ä¹ĭ ä¹ħ +第äºĮ èĬĤ +åĬ¨çī© åĽŃ +第ä¸Ģ 书记 +éħ ļ +çĶŁäº§ 设å¤ĩ +æŁIJç§į ç¨ĭ度 +åľ Ń +åĩŃåĢŁ çĿĢ +éĺħ è§Ī +çϽ æ²Ļ +æ²¹ çĥŁ +çªģçł´ åı£ +åıĹ å½±åĵį +åı¯ä»¥ æĽ´å¥½ +å³° å̼ +æĿĤ è´¨ +宿 è¿ģ +çĽĺ æ´» +æ¿Ģ èµ· +åĦ¿ ç§ij +åĿIJ èIJ½åľ¨ +æĮª å¨ģ +æµ· å²Ľ +绣 绣 +éĻ ¨ +ä¼ĺ äºİ +å°Ī å®¶ +ä¸Ģ éĤĬ +èIJ Ĭ +äºĨä¸Ģ åı£ +æ²ĥå°Ķ æ²ĥ +æŃ£å¸¸ 使ç͍ +æĻ®éģį åŃĺåľ¨ +丰 满 +çĶ» åį· +åºĶ æĶ¶ +åºĶæĶ¶ è´¦ +åºĶæĶ¶è´¦ 款 +å®Įæķ´ çĥŃ +å®Įæķ´çĥŃ æ¦ľ +注 è§Ĩ +çĨ Ħ +èº ¬ +éĶĢåĶ® 人åijĺ +è¶ĭ åIJij +çĦ¦ æĢ¥ +åįģå¹´ åīį +ä¼łç»Ł 产ä¸ļ +質 éĩı +åĩ¤åĩ° ç½ij +èµĦæºIJ æķ´åIJĪ +æ¶Į åħ¥ +æĸĩåĮĸ ä¼łæĴŃ +çķĮ 第ä¸Ģ +æ°´ æ³µ +宫 殿 +æİ¢ 寻 +ä¿® åīª +æĦı è¦ĭ +ç´Ĭ ä¹± +æĽ ī +çϽ è¡£ +èĻİ åį« +ç´§ æī£ +å¤Ħå¤Ħ éķ¿ +åĪĽå»º å·¥ä½ľ +红 æŀ£ +饼 å¹² +äºĨ åįĬ天 +ä¼ļå½±åĵį åΰ +çĽ¸ä¿¡ 大家 +èħ¾ é£ŀ +å°± å¦ĤåIJĮ +ä¸ĭéĿ¢ å°ıç¼ĸ +æ°ijèIJ¥ ç»ıæµİ +æĻ ¦ +è£ħ æī® +é»ij å¤ľ +常 å¾· +å·¥ä¸ļ 大åѦ +æĺİ çŁ¥ +éĺŁåijĺ 们 +åIJ¬ 课 +æ¯ı éļĶ +羣æĺ¯ 太 +åIJĪä½ľ åħ±èµ¢ +çIJĨ åıij +æīį å¹² +çľĭ èµ·ä¾Ĩ +殿 ä¸ĭ +å®ī éĺ³ +æīĢ äº§çĶŁçļĦ +éĽĩ ä½£ +æĬ¬èµ· 头 +æį® æĬ¥éģĵ +éļĨéĩį 举è¡Į +交 éĶĻ +è¶ħ é¢Ŀ +åĮĸ çĸĹ +é¡ Ĩ +纵 æ·± +çĪ±åĽ½ 主ä¹ī +éĻ¢ åī¯éĻ¢éķ¿ +è® ³ +羣æŃ£ åģļåΰ +åѤ åįķ +èĩªçĦ¶ èĢĮ +èĩªçĦ¶èĢĮ çĦ¶ +ä¿® 身 +èĬ ¹ +æģ¯ æģ¯ +æģ¯æģ¯ 缸åħ³ +驾 æł¡ +æİ© 饰 +æ³½ è¿ŀ +æ³½è¿ŀ æĸ¯åŁº +举 æŃ¢ +管çIJĨ ä½ĵåζ +åħ¶ä¸Ń ä¹ĭä¸Ģ +æĿ¾ å¼Ľ +æĭ¦ æĪª +åį« åģ¥ +åį«åģ¥ å§Ķ +ä»İ åݻ年 +åĤ ¢ +è´Ń 票 +åĽ¾ æłĩ +æ²³ 西 +æ°ijæĶ¿ å±Ģ +ç§ģ èIJ¥ +å¤ĸåĽ½ è¯Ń +å¹² è´§ +æĵ¦ æĭŃ +åľ° ä¸Ń +åľ°ä¸Ń æµ· +æµĵ æµĵ +æµĵæµĵ çļĦ +å§ĭ 建 +å§ĭ建 äºİ +ç¶ĵ æŃ· +è·¯ æ¼Ķ +æļ´ é£İ +åŁº è¾ħ +æī¶è´« å·¥ä½ľ +ä¸Ģ缴 å¤Ħäºİ +æĥħ è¶£ +äºĮ åŃ£åº¦ +åİĮ æģ¶ +顺åĪ© å®ĮæĪIJ +æŁ¥ å°ģ +é¡¶ 端 +ä¸į åŃķ +ä¸Ģ大 åłĨ +被 æ·ĺæ±° +æĺ¯ ç͍æĿ¥ +æľĢ åIJĪéĢĤ +亮 çľ¼ +å¹¶ä¸įæĺ¯ å¾Ī +ç§ijçłĶ éĻ¢ +ç§ijçłĶéĻ¢ æīĢ +ç² Ł +é¢Ī éĥ¨ +é»ĺé»ĺ åľ° +é«ĺä¸Ń çĶŁ +æĹıèĩªæ²» åİ¿ +æķĻåѦ è´¨éĩı +æĪĺ çģ« +åĿİ åĿ· +æIJŃ ä¹ĺ +è¯Ĺ æĦı +åĪij èѦ +åĩº æ±Ĺ +åįģåħŃ æĿ¡ +请 åıĬæĹ¶ +åĨľä¸ļ 大åѦ +èIJ½ åı¶ +æĢ» èĢĮè¨Ģ +æĢ»èĢĮè¨Ģ ä¹ĭ +æĿľ åħ° +æĿľåħ° çī¹ +éĻª ä½ł +åħ¬ æĬ¥ +çķĻè¨Ģ æĿ¿ +éĺħ åİĨ +ç«¶ çĪŃ +ç»Ļ åĪ«äºº +æĹ¥æĬ¥ 社 +åĿIJ èIJ½ +åĿIJèIJ½ äºİ +éĩij åŃĹ +éĩijåŃĹ å¡Ķ +åĽ ¤ +è¯Ŀ åī§ +æĮģç»Ń æİ¨è¿Ľ +æ¼ı æ°´ +詳 ç´° +æĢĢ æĬ± +åıĺ å¹» +饥 饿 +éļIJ 身 +个 èµĽåŃ£ +åĵ¡ å·¥ +æģ¢å¤į æŃ£å¸¸ +äºĨ 好å¤ļ +æĺŁ å·´ +æĺŁå·´ åħĭ +åħī çݯ +å¸ħ åĵ¥ +çϽ éĽª +ç¨į ç¨į +计 æıIJ +æĦĽ æĥħ +éİ ĸ +ä¿¡ éĺ³ +è§Ģ å¯Ł +å¦Ĥæŀľä½ł æĥ³ +缸æ¯Ķ ä¹ĭä¸ĭ +è§£ å¼Ģ +æīĵåį° æľº +身 躯 +ç²¾ç¥ŀ æĸĩæĺİ +èĤ¡ æĮĩ +å¾® åĪĽ +红 èĮ¶ +èĩ´ çĻĮ +æģ© æĸ½ +èħ¿ éĥ¨ +大åŀĭ å¤ļ人 +å®ī åĢį +è¾ħ导 åijĺ +èĪª éģĵ +å¸ĥ å°Ķ +åįĹå®ģ å¸Ĥ +ä¸ĬçıŃ æĹı +ä¾§ ç»ĵæŀĦæĢ§ +追 éļı +å½ĵåľ° æĶ¿åºľ +èµ° åĩºæĿ¥ +éĩijèŀį ä¸ļ +丼 书 +é¡¹çĽ® ç»ıçIJĨ +è¿ĩ æĪ· +骨 æŀ¶ +è¡ Ļ +ä»Ģ 麽 +èħ ĭ +è¦ģ 害 +åľ¨ åºĬä¸Ĭ +代è¨Ģ 人 +並 å°ĩ +åIJĦ个 æĸ¹éĿ¢ +è°´ è´£ +åħ± æĮ¯ +åį³å°Ĩ åΰæĿ¥ +èĤº çĻĮ +ä¾Ľ éĶĢ +丼 æŀĹ +èµ ĥ +åįģä½Ļ å¹´ +åĭĺ æİ¢ +飵 åij³ +èĭ¦ ç¬ij +æľĢ大 ç¨ĭ度 +éĩįçĤ¹ åħ³æ³¨ +ä¹ĭ 举 +满 æĢĢ +åıĹåΰ å½±åĵį +æĭĽ æĬķæłĩ +è¡¥ é½IJ +西 红 +西红 æŁ¿ +é¬ § +è£ħ åᏠ+éĤ» éĩĮ +èĤĩ äºĭ +æİĴ æ¯Ĵ +åѤ åĦ¿ +鼶 è·Ŀ离 +å®ŀ å¹² +çľĭ æŁ¥çľĭ +æĶ¶è´¹ ç«Ļ +ç» · +åħ¬çĽĬ æĢ§ +éĢĴ ç»Ļ +æĶ» æīĵ +æĺŁçº§ éħĴåºĹ +æĺİ åªļ +ç፠ç«ĭ +è¯Ŀè¯Ń æĿĥ +ä¸ĢæŃ¥ ä¸ĢæŃ¥ +书æ³ķ å®¶ +æľªç»ı æİĪæĿĥ +çŁ³ èĨı +åĩŃ ä»Ģä¹Ī +çļĦ æĹ¥ +çļĦæĹ¥ åŃIJéĩĮ +诱 人 +çϾåĪĨ çϾ +èĪĪ è¶£ +å¼ł åħĪçĶŁ +èĢģçĪ· åŃIJ +æ³¢ çī¹ +åŁºéĩij 份é¢Ŀ +æ²Ļåıij ä¸Ĭ +å¥ĭæĸŠ缮æłĩ +æ°¢ èĥ½ +æ²ĥå°Ķ çİĽ +義 åĭĻ +éŁ³ ç®± +æ²ī 浸 +æ²ī浸 åľ¨ +èĭ± åľĭ +çģ¯ çģ« +è¿Ľ 项 +两 端 +ä¹Ķ 丹 +èĦ¸ é¢Ĭ +åıijå±ķ æ½ľåĬĽ +åĭķ ä½ľ +åĵĪ ä½Ľ +å®´ ä¼ļ +æ§ į +ç«ĭ å¿Ĺ +ç¡ķ士 åѦä½į +åĭĭ 竳 +è¿Ļ åľºæ¯ĶèµĽ +æĮģ å¹³ +éķĢ éĶĮ +èĭ± çī¹ +èĭ±çī¹ å°Ķ +æķĻ èģĮå·¥ +åĬŁ åĬĽ +该 æ¡Ī +ä¸Ģ æ¢Ŀ +åĺī å¹´ +åĺīå¹´ åįİ +è¿« ä¸įåıĬ +è¿«ä¸įåıĬ å¾ħ +è¿Ļ个 æĹ¶ä»£ +精彩 æĴŃæĬ¥ +人 èĦ¸ +人èĦ¸ è¯ĨåĪ« +æ£Ģå¯Ł å®ĺ +å°ı èħ¿ +éĨĴ 缮 +åħļ æĢ» +åħļæĢ» æĶ¯ +æĪ Ł +èĮ« çĦ¶ +è±Ĩ æµĨ +主 æ²» +éĿĴæµ· çľģ +åĪijäºĭ 责任 +çł ° +ä¹ĭ æ¬ĬåĪ© +äºĶ å®ĺ +è¿· æĥij +åħ¥ åºĵ +å®¶ 纺 +å¼¹ ç°§ +åįģäºĶ æĿ¡ +ç»Ļ å®Ŀå®Ŀ +èĪªç©º èĪªå¤© +å¾Ģ å¤ĸ +å¼ķ åĬĽ +çľ¼ çļ® +æ¶ī è¶³ +æĿ¥ 宾 +åľ¨çº¿ è§Ĵèī² +çĥŃ éĶĢ +æµģ éĢĿ +泡 泡 +éĻį å¹ħ +è´ŁéĿ¢ å½±åĵį +红 楼 +红楼 梦 +éļĶ çĿĢ +ä¾¥ 幸 +许 ä¹ħ +åĴĮ çĿ¦ +èŃ ½ +使ç͍èĢħ æĪĸ +ä¹° åįķ +è¿ ´ +é£İ æīĩ +æķĻ å¸« +æ¡ĮåŃIJ ä¸Ĭ +å¾Ī æ¼Ĥ亮 +åł± å°İ +第ä¸Ģ åŃ£åº¦ +ç©© å®ļ +æĤ² åĵĢ +çĿĢåĬĽ æīĵéĢł +æĮ Ł +è·¯ æ¡¥ +åij IJ +åľ£è¯ŀ èĬĤ +çļĩ åŃIJ +ä»ĩ æģ¨ +éħĿ éħ¿ +ä¸į éĹ´ +ä¸įéĹ´ æĸŃ +æĮĩ å°ĸ +ä¸ŃåĽ½ ç½ij游 +åŀ £ +æĦıè§ģ 建议 +æ¯ħ çĦ¶ +亮 度 +èģĶ è°Ĭ +å½ķ åħ¥ +åĦ ² +å¨ĺ å®¶ +ç§ij å°Ķ +ä¹Łæ²¡ ä»Ģä¹Ī +æł¹æį® ä¸įåIJĮ +åı¶ ä¿® +å̼ å®Ī +æľ« 端 +åĪ ¨ +åĤµ åĭĻ +èģ¯ åIJĪ +å¥ĩ å¹» +èĻļ æŀĦ +é»Ħ æĺı +å¹³ åĿ¦ +æµģ æ°ĵ +æĸ° åŁºå»º +æĮ½ æķij +åįİ å°Ķ +åįİå°Ķ è¡Ĺ +æľĢ åıĹæ¬¢è¿İ +ç»Ń 约 +å¼Ĭ 端 +éŃĶ æ³ķå¸Ī +éŃĶæ³ķå¸Ī åĴĮ +åħ·ä½ĵ åĨħ容 +çIJī çĴĥ +æī© 容 +èĮ¶ åĽŃ +主ä¹ī èĢħ +ç«ĭ éĿ¢ +æİ¥åıĹ éĩĩ访 +åĩº åħ¥å¢ĥ +ç§ij åįı +éĴ ³ +çµIJ æ§ĭ +ç»ĵæŀľ æĺ¾ç¤º +åı° è´¦ +å°± æĿ¥çľĭçľĭ +èĩª æķij +åıį æĩī +åİ» åĵªåĦ¿ +è¿Ļ é¦ĸ +è¿Ļé¦ĸ æŃĮ +åIJ¬ ä¼Ĺ +å¤ĸ 壳 +ä½ĵèĤ² é¦Ĩ +實 æĸ½ +èŀº ä¸Ŀ +æĭī åįĩ +çĮĽ åľ° +åħ¨åĽ½ 人æ°ij +æĤī å°¼ +æĹı 群 +åĽ¢ åijĺ +两个 å°ıæĹ¶ +åľ¨ çݩ家 +åľ¨çݩ家 ä¸Ń +çĶľ çĶľ +æĬķ è¡Į +åįĶ æľĥ +éĻ ¡ +åĬłå·¥ åİĤ +æ¦Ĩ æŀĹ +æŃ» è§Ĵ +åĨħ å¹ķ +æīĢæľī æĥħèĬĤ +åĪ· åį¡ +æ°´ èĤ¿ +èĥĥ åı£ +å«Į å¼ĥ +æ²® 丧 +ä¸īå¹´ 级 +æ¶Ĥ å±Ĥ +å¿ĥ 仪 +å¿ĥ仪 çļĦ +å¤ Ń +é¦ĸ è½® +æĹłè®ºæĺ¯ åħ¶ +éĢı æ°Ķ +äºĮ åįģäºĶ +ç® « +åĬŁ åĬ³ +çѾ ä¸ĭ +æ²ī è¿· +æķij åij½ +éĹª éĹª +åIJĥ äºı +å±ķ åĵģ +åį³æĹ¶ åıijçĶŁ +ç¶ ľ +ç¶ľ åIJĪ +æłĩ æĺİ +çľĭ ç͵影 +åħ¬ 竳 +éĺ¿ æ£® +éĺ¿æ£® 纳 +身 åĪĽéĢł +身åĪĽéĢł çļĦ +æ¸Ľ å°ij +å̼å¾Ĺ åħ³æ³¨ +鼶åĶ® åķĨ +æįĨ ç»ij +è¸ı åħ¥ +èĽ Ł +æŁ´ 纳 +èĢģ åħµ +绿èī² çݯä¿Ŀ +é¹ Ń +麻 æľ¨ +æıŃ çīĮ +è¿Ļ款 车 +ç¾İ å¾· +ç¾İå¾· åħ¬åı¸ +æ¶ § +è°ģ çŁ¥ +æ´ĭ èij± +æ¯į æł¡ +ä¸Ģ éĹª +çĶ· 主è§Ĵ +æĹłçº¿ ç͵ +å±ł å®° +æĺ¯ éŁ©åĽ½ +æĺ¯éŁ©åĽ½ 娱 +容 è²Į +åĿĩ 使åħ¶ +太 å¿« +å¹´ çͱ +å¹´çͱ 缼 +èĭ¦ èĭ¦ +åĬĽ è¿ĺæĺ¯ +åĬĽè¿ĺæĺ¯ èĩª +æĨ © +èģ¯ çµ¡ +åĶ ¾ +åħ·æľī æĪĺ士 +追 éĹ® +åłĨ æĶ¾ +åıį 驳 +å®ŀäºĭ æ±Ĥ +å®ŀäºĭæ±Ĥ æĺ¯ +åѸ éĻ¢ +åįģ åĩłä¸ª +æķij æĬ¤ +æķijæĬ¤ 车 +ç½ij绾 ä¼łæĴŃ +åįģåħ« å±Ĭ +éĥ¨ åī¯ +éĥ¨åī¯ éĥ¨éķ¿ +çĹ´ è¿· +管çIJĨ æĿ¡ä¾ĭ +èŀį 为ä¸Ģä½ĵ +æĢ» 产å̼ +è³ ĵ +ä¸ĥ æĺŁ +çıŃ ç»Ħ +绣 é¢Ĩ +请 大家 +éĩij éϵ +èĪħ èĪħ +æµ· æ¹¾ +æĸ½ çŃĸ +享 èªī +éº ¥ +端 åįĪ +绿 åŁİ +確 ä¿Ŀ +å·´ æĭī +åĨĴ çĿĢ +æħ· æħ¨ +个人 è§ĤçĤ¹ +ä¹Ļ çĥ¯ +ç¡ħ è°· +éĸĭ å±ķ +å°ļ 书 +åĿļ 飧 +åº µ +èĢģ é¾Ħ +èĢģé¾Ħ åĮĸ +羨 çľ¼ +绿 æ°´ +绿水 éĿĴå±± +书 é¦Ļ +主åĬĽ åĨĽ +æīįæĺ¯ 羣æŃ£ +æĬ¢ åħĪ +æĪIJå°± æĦŁ +éĩį æŀĦ +éĴ¢ åİĤ +æĪIJ 份 +èĬ± 纹 +ä¹ĭ äºī +å¹² ç»Ĩèĥŀ +æĹ¢ åı¯ä»¥ +ç¹ģ çIJIJ +æĦļ èł¢ +éĿŀ常 æĺİæĺ¾ +ä½ĵ 彩 +æĬĢ æ³ķ +æĿĨ èıĮ +å¹¿æ³Ľ åħ³æ³¨ +åĮĹ å®ĭ +å§Ĭ 妹 +åįı åĬŀ +æ·® åįĹ +çĥ ı +æ´Ĺ èĦ¸ +åıĹ è®¿ +åıĹ访 èĢħ +éĩįè¦ģ åĽłç´ł +å½±è§Ĩ åī§ +综èīº èĬĤ缮 +èľķ åıĺ +äºĮ 线 +äºĮ线 åŁİå¸Ĥ +ä¼Ĭ å§ĭ +çıĬ çijļ +èĩª æŁ¥ +åħ¥ åĽŃ +åĩ¶ æīĭ +åħ¬ è¯ī +éģĩ éļ¾ +éĩĩçŁ¿ çŃī +èĩª çIJĨ +åĸ· æ¶Ĥ +æī© åħħ +éĢı è§Ĩ +é«ĺéĢŁ å¢ŀéķ¿ +åĽ¾ çĶ» +ç¾ ¹ +èĤĩ åºĨ +è¾ľ è´Ł +èµĶ ä»ĺ +è· ¡ +åģ¥åº· æĪIJéķ¿ +以ä¸Ĭ åѦåİĨ +åıĸå¾Ĺ 以åıĬ +æ²ī 积 +åįģä¹Ŀ å±Ĭ +缸éĹľ æľįåĭĻ +æī§ åĭ¤ +åī¯ åİ¿éķ¿ +å¯ ° +åģľ æ»ŀ +æ·¹ 没 +çŁ³ çģ° +çį ¸ +åĢ ¦ +ç¾İ åªĴ +æķĻ æ¡Ī +åĬł çĽĸ +åħ¬å¼Ģ èµĽ +å¥ł åŁº +æĺĨ èĻ« +çŀ ħ +磷 éħ¸ +äºī åĪĽ +çİĭ æĻĵ +ç¼ĵ åĨ² +åİļ åİļ +åİļåİļ çļĦ +æŀ£ åºĦ +ç²¾ çĽĬ +ç²¾çĽĬ æ±Ĥ +ç²¾çĽĬæ±Ĥ ç²¾ +åĪĨæĶ¯ æľºæŀĦ +å®ŀæĸ½ ç»ĨåĪĻ +æĸ° èµĽåŃ£ +總 çµ± +éĢł è¡Ģ +é¢ĩ åħ· +é»Ħ åŁĶ +è¡Ģ èĦĤ +交éĢļ å·¥åħ· +å³ ¥ +æĹıèĩªæ²» å·ŀ +寺 éĻ¢ +確 å®ļ +æ¦Ĥ念 èĤ¡ +æĦŁ å®ĺ +æŁľ åı° +åĶ Ķ +çŀŃè§£ 並 +æĢ» ä»· +åIJ¸ åħ¥ +æĢ ¼ +æĻļ éĹ´ +å±Ĭ æ¯ķä¸ļçĶŁ +çĶŁ å§ľ +éĺħ读 åħ¨æĸĩ +å¾Ĺåΰ æľīæķĪ +æIJľ æķij +åİĨ æĿ¥ +èŃī æĺİ +åĥ » +èĨ³ é£Ł +åĦĦ åħĥ +æīĵ åİĭ +宾 客 +åķ ¼ +ä¸ĢçϾ å¤ļ +æ·±åħ¥ 人å¿ĥ +æ¢ħ å·ŀ +çłĶ åѦ +åħ³ ä¹İ +è¼ Ľ +亲 åıĭ +éħį æĸĻ +æĪij çĪ±ä½ł +è´¸æĺĵ æĪĺ +æľī èī² +æľīèī² éĩijå±ŀ +æįIJ åĬ© +为 é¦ĸ +为é¦ĸ çļĦ +å¯Į åĬĽ +çĶ· ç¥ŀ +é³ ³ +æµĩ æ°´ +åIJ ± +æĺİç¡® æıIJåĩº +åı¹ äºĨ +åı¹äºĨ åı£æ°Ķ +礼 æĭľ +è¿Ļ个 åIJįåŃĹ +ä¿¡ å¾Ĵ +å¿Ĺ 强 +éĻIJ æĹ¶ +æĶ¶ è²» +åĨľå®¶ ä¹IJ +å°ıé¾Ļ èϾ +èIJ½ å¹ķ +æ§ Ł +åѦ 龸 +æĪĸ å¤ļ +æĪĸå¤ļ æĪĸ +æĪĸå¤ļæĪĸ å°ij +座è°Ī ä¼ļä¸Ĭ +æ¶ ¼ +éŃĶ çİĭ +å² ± +é¡¶ å±Ĥ +é¡¶å±Ĥ 设计 +èĦij åŃIJéĩĮ +éĻ¢ åŃIJéĩĮ +轩 è¾ķ +身å¿ĥ åģ¥åº· +èħ ij +éĹľ 注 +åıĤåĬł ä¼ļè®® +ä¸Ńåįİ æĸĩåĮĸ +追 寻 +å®ī çĦ¶ +é£Ļ åįĩ +éŁŃ èıľ +é¸ ¦ +åĤ¨ éĩı +çĶ· æĸ¹ +å¤ĩ 份 +æijĶ åĢĴ +润æ»ij æ²¹ +é̼ è¿ij +çͳ è¯ī +鸣 ç±» +çŁ³æ²¹ åĮĸå·¥ +åĿļ æŀľ +è¿Ļå®¶ ä¼Ļ +æĭĴ ä¸į +羣 çļ® +è·Ŀ éĽ¢ +è¿ĺ æĮº +éĽķ åĥı +åĪĿ æģĭ +æıIJä¾Ľ æĽ´å¤ļ +æŁ¥çľĭ åħ¨æĸĩ +æķ°åŃĹ è´§å¸ģ +åĸī åĴĻ +åı¦ä¸Ģ ä½į +åĤ¬ åĮĸ +åĤ¬åĮĸ åīĤ +ä»İæĿ¥ 没 +å¯ĨåĪĩ 缸åħ³ +éĥ¨ 主任 +产åĵģ ç»ıçIJĨ +並 åIJĮæĦı +èIJ½ åħ¥ +å±ıå¹ķ ä¸Ĭ +åħ¬åı¸ 竳ç¨ĭ +æį¢ åı¥è¯Ŀ +æį¢åı¥è¯Ŀ 说 +ä½į æĸ¼ +ä½ Ķ +åĩ» æĿĢ +缸 è¾ĥ +缸è¾ĥ äºİ +ç²½ åŃIJ +åįĹ æŀģ +宫 é¢Ī +è£ģ åijĺ +æĺİ ç»Ĩ +ä»·å̼ éĵ¾ +åĽĽä¸ª æĸ¹éĿ¢ +æĥħåĨµ æĿ¥çľĭ +æĮij åīĶ +æ® ĺ +æŀģ åĬĽ +çĸij éļ¾ +æĬµæĬĹ åĬĽ +æĢ¥ éĢŁ +æĪ Į +ä½İ ä¼° +éĹª è¿ĩ +æģ ¬ +èµŀ æī¬ +ä»ĸ å¦Ī +æĪIJ为 ä¸ĢåIJį +æ´Ĺ 礼 +é¢Ħ计 å°Ĩ +åħĪè¿Ľ åįķä½į +è¼ Ķ +éĢĥ èĦ± +çݰ åŃĺ +èĢģèĻİ æľº +åįģä¸ĥ æĿ¡ +åı¦ä¸Ģ åįĬ +温 æĥħ +åī¥ ç¦» +ä¸ĸ è´¸ +å®ĺ åı¸ +å¾Ī å·® +éĹ´ è·Ŀ +请 注æĦı +åı² è¯Ĺ +åĪ© åύ +è¿IJ ç®Ĺ +沦 为 +該 使ç͍èĢħ +èĮ ¬ +éͦ 绣 +åı² æĸĻ +çģµ æ´»æĢ§ +èģĶ ç¤¾ +æĹł åĬ© +æĬĹ æ°§åĮĸ +èıľ èĤ´ +éĢł èι +æİī èIJ½ +å¤į æŁ¥ +åĭĥ åĭĥ +åij¼ 声 +給 äºĪ +åIJĮäºĭ 们 +ç½ ° +è¯ķ æİ¢ +åħ³éĶ® åŃĹ +æįIJ çĮ® +ç»Łè®¡ æķ°æį® +åĪĽ ä½ľèĢħ +ä¸ĭ åįĬ +ä¸ĭåįĬ åľº +æī¿æĭħ 责任 +端 æŃ£ +ç©¿ è¡£ +ä¼ł çIJĥ +åĬ© éķ¿ +åĩ ± +éķ¶ åµĮ +é£ŀ ç¿Ķ +è¾ĵ åįµ +è¾ĵåįµ ç®¡ +ä¸ĩ åħ¬éĩĮ +æİ¨å¹¿ åºĶç͍ +å¿« æ¨Ĥ +ç§ ½ +èī° å·¨ +åIJ¬ å®Į +åĿļ 硬 +奥 åľ° +å¥¥åľ° åĪ© +é¢ ĵ +èĻIJ å¾ħ +ä¾Ľ æ±Ĥ +éľī ç´ł +伪 è£ħ +乡 åľŁ +åĩ¡ æľ¬ç½ij +åĩ¡æľ¬ç½ij 注 +ä¼Ĭ åĪ© +è¡¡ æ°´ +æĽ´ åĥıæĺ¯ +åĪĨéĴŁ å·¦åı³ +è¦ı 模 +äºĶ åĪĨéĴŁ +åºĹ åĬłçĽŁ +åĽ° éĽ£ +åħ³ åģľ +æĢĿ 绪 +åĴ½ åĸī +缸 符 +çĥ¦ èºģ +æĻĤ æľŁ +åijĪ çı¾ +è§£ æķ£ +诱 导 +éļĶ çĥŃ +çĮ ¶ +åįĹ å®ĭ +æ·±åħ¥ äºĨè§£ +çŃĶ çĸij +æĺ¼ å¤ľ +åįĥ ä¼ı +åĬ³åĬ¡ æ´¾éģ£ +红 è±Ĩ +åĿı äºĭ +çĤ¹ æ»´ +å°±ä¸ļ å²Ĺä½į +约 åIJĪ +åħį éϤ +éĢĨ åĬ¿ +éĩį éĩijå±ŀ +å®ĺ 宣 +ä½İ å»ī +æģ¨ ä¸įå¾Ĺ +å¾Ĺ 天 +å¾Ĺ天 çĭ¬ +å¾Ĺ天çĭ¬ åİļ +ä¸Ģå°ģ ä¿¡ +æĬ½ å¥ĸ +è¾Ĺ 转 +çķĻ å®Ī +çķĻå®Ī åĦ¿ç«¥ +çŃĶ åį· +å·¨ åŀĭ +æľĢ好 ä¸įè¦ģ +æµĻæ±Ł 大åѦ +æĨ ¨ +æı¡ æīĭ +éĴĪ ç»ĩ +æİĴ 骨 +çĤ ½ +å°ģ è£ħ +åįĢ åŁŁ +空æ°Ķ åĩĢåĮĸ +åħī å½± +åĢĴ å¡Į +å§ļ æĺİ +æ¤į 被 +åѦ åīį +åѦåīį æķĻèĤ² +èĬĿ åĬł +èĬĿåĬł åĵ¥ +缩 æ°´ +ä½ Ł +åľ¨çº¿ åĴ¨è¯¢ +èµı æŀIJ +éĿĴ èĽĻ +æĬ± ä½ı +èĮĤ åIJį +åħ¨åĬĽ æīĵéĢł +åįļ士 åѦä½į +æ²§ å·ŀ +åĻ ¢ +æĿĤ çī© +åĪ» çĶ» +æį ħ +å¾® éĩı +å¾®éĩı åħĥç´ł +ä¸Ģ åĽŀäºĭ +鸡 èĤī +åĪ©æ¶¦ çİĩ +æīį ç®Ĺ +å¾® å¦Ļ +棵 æłij +è´ª 婪 +åĩı å̼ +梦 å¢ĥ +åı¯ è§Ĩ +åı¯è§Ĩ åĮĸ +广大 å¸Ĥæ°ij +ä¸ĵä¸ļ ä»İäºĭ +ç»ı 纬 +ç´§ çĽ¯ +çŁ¥ å·± +è¤ ļ +æĸĩåĮĸ åºķèķ´ +åݦéŨ å¸Ĥ +临 港 +对åħ¶ 羣å®ŀ +岸 è¾¹ +è¦ĸ çĤº +æĬĹ çĻĮ +åĶIJ å®ĩ +ä¸įå¾Ĺ è¶ħè¿ĩ +å¨ģ æħij +æ¡Ĩæŀ¶ åįıè®® +èµ° ç§ģ +åĽ¢ å§Ķ +夸 大 +æ¬ Ħ +ç¥ŀç»ı ç³»ç»Ł +æijĦå½± ä½ľåĵģ +èĬ ¥ +å®ī åºĨ +æµ· 滨 +æŀĦ æĢĿ +çīµ æĮĤ +åı © +éĺIJ æĺİ +éģ ģ +ç²¾ æ²¹ +ç©´ ä½į +æĬ¤ 身 +æĬ¤èº« 符 +æĮĩ å°İ +åŃĺåľ¨ ä¸Ģå®ļ +å¯Ĥ éĿĻ +æµ·å¤ĸ å¸Ĥåľº +éĿ ¡ +综åIJĪ å¾ģ +ä¿ IJ +è¨Ī ç®Ĺ +æĺİ æľĹ +äºļ è¿IJ +äºļè¿IJ ä¼ļ +åīįçŀ» æĢ§ +åĮ® ä¹ı +产ä¸ļ æī¶è´« +èĦij æµ· +èĦijæµ· ä¸Ń +åħļçļĦ é¢Ĩ导 +åĪĺ éĤ¦ +æµģ æĺŁ +æĵ Ĥ +æĶĢ çĻ» +åĴ Ķ +ä¸Ģä¸ĭåŃIJ å°± +è¯Ĭ æ²» +使 åĬ² +åīµ ä½ľ +éĵŃ è®° +éĴ± è´¢ +æĹ¥æĬ¥ è®°èĢħ +çĥŁ çģ« +èĥľ è´Ł +åįļ 主 +ä¸ŃåĽ½ èģĶéĢļ +ç½ijç«Ļ é¦ĸ页 +å°± å¤Ł +å°±å¤Ł äºĨ +æīij åħĭ +å±ħ å§Ķä¼ļ +è° ¬ +å®īåħ¨ äºĭæķħ +åķĨ çĶ¨è½¦ +循çݯ ç»ıæµİ +æ· ¤ +èĢĥ è¯ģ +å®Ŀ èĹı +å®Į ç»ĵ +çłĶåıij æĬķåħ¥ +å² ij +æģŃ æķ¬ +离 éĢĢä¼ij +æ°´ 墨 +å© ¶ +è¯Ĺ åı¥ +å®ģæ³¢ å¸Ĥ +å¼± çĤ¹ +åģľ çīĮ +奶 æ²¹ +å¥ĩ纳 æ²³ +æĨ Ĥ +社ä¼ļ å®ŀè·µ +è´Ŀ 壳 +çłĤ æµĨ +èι åıª +宣 æī¬ +综åIJĪ æķ´æ²» +åĤ ij +æ°ijæĹı æĸĩåĮĸ +éĩį çݰ +积 æ·Ģ +åħ¬ çĦ¶ +çħ ī +缸 èģļ +æ± ¾ +纹 çIJĨ +çĩĥ çħ¤ +æŃ¤ ç§į +ç¾İ å¦Ĩ +åįĥ çĵ¦ +çIJ Ľ +驾驶 è¯ģ +éĺ¶ æ¢¯ +ä¸Ŀ ä¸Ŀ +å¾Īå¤ļ äºĭæĥħ +åħī éĺ´ +èijĹä½ľ æ¬Ĭ +åħ§ éĥ¨ +çĽ¸å¯¹ æĿ¥è¯´ +éĸ Ĵ +éľĩ æħij +說 話 +æĨ ij +ç«¥ è£ħ +ä½ıæĪ¿ åĴĮ +ä½ıæĪ¿åĴĮ åŁİ +å·²ç»ı è¶ħè¿ĩ +侦 å¯Ł +çŁ¿ çī© +ä¾Ľ 大家 +çī¹ éĤĢ +ç¨ĭåºı åijĺ +çķľçī§ ä¸ļ +æ° ª +çij ª +åĢĴ åľ¨ +åĢĴåľ¨ åľ° +æ¯ Ģ +梯 éĺŁ +æİ¥ èijĹ +æĬĹ èıĮ +è¤ ĩ +ç¬ Ļ +æ¯Ķ ä¸Ĭå¹´ +鸡 汤 +åŃ¦ä¹ł æĪIJ绩 +æĸij æĸĵ +åħΠ坼 +åĪĹ ä¸¾ +è°ĥæŁ¥ æĺ¾ç¤º +æ© « +ä¹Ŀ åįģ +è°¢ 飵 +è·¨è¶Ĭ å¼ı +女æĢ§ æľĭåıĭ +èIJ¥åħ» ä»·å̼ +å®ŀè·µ ç»ıéªĮ +èĭı å·ŀå¸Ĥ +çĵ¶ åŃIJ +æĸ° çļĦä¸Ģ +æĸ°çļĦä¸Ģ å¹´ +æĺİ æĻ° +å®ł çα +åŃŠ第 +æľĹ 诵 +纳 æĸ¯ +éĢĨ è¡Į +è«ĭ æĤ¨ +è«ĭæĤ¨ æıIJä¾Ľ +èĥ¸ æĢĢ +第ä¸ĥ å±Ĭ +强 壮 +代 åŃķ +æ±¶ å·Ŀ +å®¶ åĸ» +å®¶åĸ» æĪ· +å®¶åĸ»æĪ· æĻĵ +èħ ® +åIJ¯ 迪 +æĹł éļľç¢į +èĻķçIJĨ åıĬ +æĿ¥ åİĨ +å®ŀ åĬ¡ +ä¹Ł éļıä¹ĭ +æĬĢèĥ½ åŁ¹è®Ń +åѤ ç«ĭ +åī ģ +éĥ´ å·ŀ +æĶ¶ æķĽ +éł» éģĵ +èᣠ幏 +èİ« è¿ĩäºİ +æŃ¤ æĻĤ +纪å§Ķ çĽij +纪å§ĶçĽij å§Ķ +缸 éĤ» +åı¦ä¸Ģ è¾¹ +çªĴ æģ¯ +æľīå¾Īå¤ļ ç§į +æ¯ı éĢ¢ +éĹ® ä¸ĸ +ç´¯ ç´¯ +éĿĴæĺ¥ æľŁ +è·¯ åĨµ +åħĭ èݱ +è¿Ħä»Ĭ 为æŃ¢ +æĥĬ å¥ĩ +è·¨ 度 +éħ¿ éĢł +åĩ ĭ +è¿ij ä¸īå¹´ +åĨħ 马 +åĨħ马 å°Ķ +æı į +è¿Ľå±ķ æĥħåĨµ +èĮ § +æľīåºı æİ¨è¿Ľ +æĢ» åĨłåĨĽ +æĪIJ绩 åįķ +éĽ»è©± åıĬ +ç´§å¯Ĩ ç»ĵåIJĪ +åºĬ ä½į +é¹ Ĭ +æķ£åıij çĿĢ +åĭŁ èµĦ +æ°¨ éħ¸ +彩 ç¥ŀ +è®Ģ åıĸ +éĩį æ¸© +ä¸Ń åŃĺåľ¨çļĦ +ç¾İ éºĹ +ä¸įæĸŃ å¢ŀåĬł +è½® æµģ +æİ¥ åIJ¬ +å¹´ 产å̼ +åįĥ åħĭ +æĪĺåľº ä¸Ĭ +çħ§ é¡§ +å¹²éĥ¨ éĺŁä¼į +åį° ç«ł +ä¸Ģèĩ´ æĢ§ +è¿ŀ å¤ľ +åħħ è£ķ +é»ij åIJįåįķ +åĩĢ æ°´ +ä¸Ģ大 æĹ© +åĮħ 袱 +çĬ¯ è§Ħ +çIJĨ è«ĸ +æŀģ æĺĵ +éª ¸ +å¨ĺ å¨ĺ +åĽ¢ åľĨ +亿åħĥ 以ä¸Ĭ +åĪ©ç͍ æĤ¨çļĦ +带æĿ¥ æĽ´å¤ļ +ä¸Ń央 空è°ĥ +æľĪ èĸª +çĮľ æĥ³ +åĪº 客 +ä½ľ æģ¯ +åįķ è°ĥ +äºĴ åĪ© +å¦Ĥæľī ä¾µæĿĥ +å°ı å·§ +åįģ åł° +åĵĪåĵĪ åĵĪåĵĪ +è¾¹ éĻħ +æłĩ è¯Ń +åĪĩåħ¥ çĤ¹ +éĢĨ è¢Ń +è¯ķ åīĤ +绿 è±Ĩ +è® ļ +åŁºçĿ£ å¾Ĵ +å£ ¬ +åħ¨ æĺİæĺŁ +éĢī ç§Ģ +èĪĮ å°ĸ +ä¸įåIJĮ ç±»åŀĭ +çĥŁ åĽ± +çģµ æ°Ķ +åĮº 管å§Ķä¼ļ +åĨľ åī¯ +åĨľåī¯ äº§åĵģ +èĶļ æĿ¥ +沪 æĮĩ +åħ»æ®ĸ æĪ· +æĸĹ å¿Ĺ +é¦ĸ é¢Ĩ +è¡Ģ èħ¥ +åĬł ç´§ +ä¸Ģèĩ´ 好è¯Ħ +第ä¸ī èĬĤ +æī¬ å°ĺ +交éĢļ æŀ¢çº½ +鼶 ç¢İ +é»ij æ´ŀ +çľĭ ä¸įæĩĤ +å±ŀ å®ŀ +主 åŁİåĮº +å¨ Ľ +å¨Ľ æ¨Ĥ +ç¬ij æĦı +èϹ æ¡¥ +åIJĦ个 çݯèĬĤ +çķ¥ å¾® +èĢķ èĢĺ +æľ¬ åľºæ¯ĶèµĽ +æĪIJ è´¥ +éĢī èĤ¡ +èªŀ è¨Ģ +çŃĶ è¾© +èĩª ä¹ł +æ£ º +ä¸ĩ 欧åħĥ +åģľ å·¥ +对åħ¶ è¿Ľè¡Į +积æŀģ éħįåIJĪ +ä¹¾ åĿ¤ +å¦ĸ æĢª +èļĮ åŁł +èµĦ产 è¯Ħä¼° +è°ĥ çļ® +éϤ å¤ķ +åĽ´ å¢Ļ +æľį å½¹ +æ·± æ¸Ĭ +é¢Ħ åζ +ç ĥ½ +å®ī 稳 +建 æŀĦ +çĭĻ åĩ» +主åĭķ 註åĨĬ +éĥ½æľī èĩªå·± +æİĴåIJį 第ä¸Ģ +麻 è¾£ +çĢ ļ +çĥŁèĬ± çĪĨ +çĥŁèĬ±çĪĨ 竹 +èĩªçĦ¶ ä¿ĿæĬ¤ +ä»Ļ å¢ĥ +为äºĨ éģ¿åħį +åĨ· åºĵ +è§£æĶ¾ æĢĿæĥ³ +åĪĿ äºĮ +ä½ĵ è´´ +é¦ĸ å¯Į +迪 æĭľ +æļĤ ç¼ĵ +æĶ¯æĮģ åĬĽåº¦ +侦 æİ¢ +马 åĪº +åĮĹ æ±½ +ç¹ ŀ +è°İ è¨Ģ +éĢ£ çºĮ +å· ³ +ä»»ä½ķ æĹ¶åĢĻ +车 èģĶç½ij +åįķ 项 +å¸Ń åį· +建çŃij æĿIJæĸĻ +ä¸Ńç§ĭ èĬĤ +ç¡ķ士 çłĶç©¶ +ç§ģ ç«ĭ +åħļåĴĮ æĶ¿åºľ +æľ¬æ¬¡ 交æĺĵ +èººåľ¨ åºĬä¸Ĭ +ç½ijåıĭ è¯Ħ论 +å¦ Ŀ +害 ç¾ŀ +åħ¬ç«ĭ åĮ»éĻ¢ +ä¸ ŀ +çĶŁçī© è´¨ +åºĶ éĤĢ +æĬ½ åıĸ +åĩł å¼ł +æijĺ ç¼ĸ +ç»ĺ æľ¬ +详 è§£ +强 硬 +æľĢ åħĪè¿ĽçļĦ +æĭĽ èĤ¡ +æĭĽèĤ¡ 书 +åįĥ æĸ¹ +åįĥæĸ¹ çϾ +åįĥæĸ¹çϾ 计 +éħį éŁ³ +驾 çħ§ +å¾ģ æĪĺ +èªĵ è¨Ģ +æĭľ å¸Ī +æĭľå¸Ī åѦ +æĭľå¸ĪåѦ èīº +æĬ± åĽ¢ +ç±³ ç²ī +éĿŀ常 éĢĤåIJĪ +èĪª æµ· +å±¥ 约 +åįģåħ« æĿ¡ +éĶ» éĢł +éĩįè¦ģ 举æİª +åıijæĮ¥ ä½ľç͍ +æ· ļ +人 社 +人社 å±Ģ +è¯ķçĤ¹ å·¥ä½ľ +éĺľ éĺ³ +æ¡ĥ åľĴ +æ°ij ä¼ģ +æ´ģ çϽ +è´µ 宾 +åħ¬ 社 +è§ī æĤŁ +è®°å¿Ĩ åĬĽ +æľĥåĵ¡ 註åĨĬ +æŃ¤ æ¡Ī +麻 çĹ¹ +çı Ģ +æĸ© èİ· +çĶ· åŃ©åŃIJ +å±ĢéĻIJ äºİ +åĭĺ æŁ¥ +åIJĥ 饱 +èĬ¬ åħ° +æ£ķ èī² +ç¦ı ç¥ī +çͳ èĬ± +æµ· çĽĹ +èĶ ij +æĸĩ åѸ +æ´»æĢ§ çĤŃ +缴 éĢļ车 +è°¢ éĤĢ +躺 çĿĢ +åľ ĥ +æ¯ıæĹ¥ ç»ıæµİ +åħ¬åħ± æĸĩåĮĸ +讲 æķħäºĭ +å¯Ł çľĭ +æĤł éĹ² +åľ° åĿª +æ¶Į çݰåĩº +é«ĺçŃī éĻ¢æł¡ +èĮĦ åŃIJ +éĺ² åį« +ä¾ĭ è¡Į +æĺ¾ éľ² +æĸ° 常æĢģ +ç»Ŀ ä½³ +å¯Į æ°ij +以 人æ°ij +以人æ°ij 为 +éĤ¢ åı° +å±ķ æ¼Ķ +çϼ å¸ĥ +è´Ł è½½ +åģı 离 +æ°¸ éģł +éĩįè¦ģ åİŁåĽł +åįıä¼ļ ä¼ļåijĺ +éļ¾ æ°ij +çĶŁäº§ 车éĹ´ +çģµ åĬ¨ +两年 åīį +æĸ¹ åľĨ +æ´» ä¸ĭåİ» +ä¸ĸçķĮ è§Ĥ +éªĹ åıĸ +ç¾İ è²Į +èĥ½ çľĭåĩº +çϼ æı® +è§Ĥ å½± +åī ĥ +åIJĪèµĦ åħ¬åı¸ +å© § +å¹² æĹ± +åħŃ ä¸ªæľĪ +尤为 éĩįè¦ģ +èĤ ½ +秦 åĽ½ +æīĺ ç¦ı +建çŃij å¸Ī +åįĩ级 æĶ¹éĢł +å°ı é¢Ŀ +å°ıé¢Ŀ 贷款 +两个 ç»´æĬ¤ +æĭį æĭį +åı¯ çĸij +æį¢ åıĸ +æŃ¦ 士 +èµĸ 以 +èµĸ以 çĶŁåŃĺ +æĮ ļ +殿 åłĤ +èĩªçĦ¶ çķĮ +ç£ģ åľº +å¦Ĥä½ķ çľĭå¾ħ +ä»ĬæĹ¥ 头æĿ¡ +西 åŁŁ +èİ· è¯Ħ +風 æł¼ +ä¿Ħ åĽ½ +æīĵ æĭ¼ +å®£ä¼ł çīĩ +å¾Ī æĸ¹ä¾¿ +ä¾Ľç»Ļ ä¾§ +纪念 ç¢ij +毫 åħĭ +èĬ³ é¦Ļ +å·¥åķĨ éĵ¶è¡Į +请 çĤ¹åĩ» +ç¼ ª +æĹłæķ° 次 +èᝠå¸Ī +èħ ¸ +游 èīĩ +åĮ ¾ +å·¡ èĪª +æ²»çIJĨ ä½ĵç³» +èIJ¥éĢł èī¯å¥½ +æ·· æ·Ĩ +éĢļ çķħ +åĬ³ ç´¯ +ä»ĵ ä½į +å¢ŀ éķ· +éļIJ 约 +æĿĤå¿Ĺ 社 +åħ» èĤ² +åı¯èĥ½ åıijçĶŁ +èĢĥ 試 +西 ä¾§ +åĬł åĢį +主æĮģ åı¬å¼Ģ +çķ¢ ç«Ł +éĹ® 询 +æµ· æ£ł +èĹ © +注æĺİ æĿ¥æºIJ +æ£Ģ çĸ« +请 åģĩ +æĬļ æij¸ +èĵĦ çĶµæ±ł +è·Ł ä¸įä¸Ĭ +çݰ代 社ä¼ļ +çѹ èµĦ +ä½ĵèĤ² 彩票 +å»¶ 误 +è¾Ľ è¾£ +éĿ¢ 容 +åį° è®° +çģŃ äº¡ +ç´ł é£Ł +åħ´ èĩ´ +éľĢè¦ģ ç͍ +éľĢè¦ģç͍ åΰ +å®Ŀ å¦Ī +ç£ĭ åķĨ +éļ¶ å±ŀ +è´¡çĮ® åĬĽéĩı +åħ¬åħ± èµĦæºIJ +大 éĺª +åĨĽ è®Ń +æĤ¬ 念 +社ä¼ļ 稳å®ļ +å¹²äºĭ åĪĽä¸ļ +æľī æĿ¡ä»¶ +æľīæĿ¡ä»¶ çļĦ +ä¸Ģå¹´ ä¸Ģ度 +åİ ¥ +强 奸 +豪 车 +æİĮ æŁľ +æ°´åĪ© å·¥ç¨ĭ +å³ ª +积æŀģ ä½ľç͍ +æµ· æ·Ģ +æµ·æ·Ģ åĮº +çĥŃ æĴŃ +åĿļæĮģ ä¸įæĩĪ +åıĮ èĦļ +绣 æĪĺ +ä»»ä½ķ 人éĥ½ +åľ°ä¸ĭ 室 +åĨ¶ çĤ¼ +è°ħ è§£ +æ¸Ķ èι +太éĺ³ åŁİ +被 æįķ +计ç®Ĺ åύ +西 åĮ» +èĪĴ å¿ĥ +æ¡ ¦ +éģ ² +åĬ ij +è¨ Ĺ +èİ º +åĸ ¬ +çĵ ¯ +åĺ ĺ +åł ķ +æķ Ŀ +åij ¦ +èĭ ŀ +æŃ ¹ +æĵ ¬ +æ£ Ħ +èĪ µ +å¥ ª +çļ ĭ +æĶ ¸ +åľ © +ç¤ Ļ +ç¢ ĺ +éı Ī +æĦ ķ +ç¹ ³ +èĺ ¸ +è² Ĥ +æ¼ ² +æij ¹ +æĶ Ŀ +åŃ ¢ +èķ Ń +é¨ ° +æ½ ¼ +éħ ° +æĴ ¥ +è¹ ¬ +é¨ Ļ +è¸ ¹ +éģ IJ +çĺ Ģ +èĽ ¤ +æĤ ĸ +çĴ ŀ +ç£ IJ +æİ ° +è¾ Ĭ +å¾ ij +æİ ĸ +éģ ŀ +éĤ ¸ +éĽ ı +æĨ İ +æľ ½ +çį » +ç® Ķ +è¤ ¶ +æļ ¢ +æĺ µ +çı Ĥ +æĤ ¸ +åģ µ +åĻ ľ +å£ ¯ +æĴ ® +æģ į +å© ķ +ç¯ ± +éĺ Ļ +çī ł +è£ ĺ +è³ ¢ +éĩ ľ +éĵ ł +èİ ĺ +æ® Ĩ +çĻ ¸ +è´ ı +ç² ± +å« ¡ +åĨ ¢ +è¤ Ĵ +æĩ Ĭ +éľ ĵ +å¡ µ +æĭ £ +å» Ł +é£ ½ +é¢ Į +åļ İ +æ· º +èĨ ł +åİ Ń +åļ ĩ +åij ĥ +çĴ ĭ +çŃ ± +æĭ · +èį § +éĶ ° +åŃ ° +èĵ ĵ +èĨ ½ +æŀ ī +åĸ ½ +çĽ Ķ +çŃ IJ +ç¾ ļ +è ħĮ +è¾ « +æ³ ĵ +çĶ ¬ +èŁ ² +åĸ ª +å¦ ĵ +è¬ Ģ +çĤ Ĭ +æĽ ľ +æ± IJ +è´ Ī +èį Ģ +æĬ ł +ç¢ ¾ +æ« ĥ +éŀ ł +èij Ĩ +ç¥ ¯ +å½ Ŀ +é¦ į +åĮ £ +æľ Ń +åĿ Ĥ +ä¿ ij +èĵ ® +çij Ľ +æī ī +èĩ Ł +è² « +çİ ¥ +æ· ¼ +åİ ² +é³ Į +å³ Ń +åij Ľ +é § +é§ IJ +éģ · +ä¿ ª +æĢ Ĥ +è¾ į +å± į +åĭ ģ +å¥ ļ +éļ ħ +éĴ ´ +è¼ Ŀ +å® ¦ +èIJ ĥ +çĺ ĭ +æĨ ¶ +æĤ ħ +è¾ Ļ +åij ľ +çł º +éĢ ŀ +æµ ļ +éĸ £ +èĸ © +éĻ ĭ +çĤ Ļ +èª ķ +ä¸ Ł +é¹ ½ +ç± Į +è´ ° +éĭ ª +çľ © +æĴ IJ +èĨ º +éŀ ĺ +ç¾ ² +çª ® +ç´ IJ +æ® ´ +çº ¾ +èº į +ç´ ĭ +çĦ ĸ +çĶ º +çī ½ +çĤ ¯ +ç¼ Ķ +æ¯ ĵ +å¬ ° +æ¢ § +äº Ł +è¢ ħ +çį Ħ +è¿ ¥ +æ¼ ¾ +çĿ ij +ç¸ ¾ +é¦ ĭ +é¤ ħ +æ ¹Ħ +æĺ ĩ +æŀ Ń +èĸ ° +æŁ ij +æ¦ » +åĻ Ĺ +åĻ ´ +æ£ £ +åĶ § +çĨ ¹ +è¼ ¯ +å¢ Ł +é² ² +æĪ Ľ +èī ¦ +èĬ ® +åĺ Ł +å¸ ¥ +å¿ » +çĮ Ŀ +å¯ µ +è³ ¦ +èĽ ¾ +æ» ¾ +çĤ ķ +éĵ ¬ +èĴ ¿ +éĴ ¨ +çĥ Ļ +ç² ķ +æĥ ¦ +æº § +é¢ į +éħ £ +å³ ¦ +ç± ģ +çĥ ĥ +åĨ Ĺ +åı ģ +çĽ § +ç½ µ +éĴ Ĺ +å¬ ī +è° ı +ç³ § +è¾ Ń +æ· ¬ +èŁ Ĵ +è¯ © +è¦ ĥ +çĻ ĸ +é½ Ĵ +çĪ IJ +ç® į +ç¼ İ +ç£ º +è¯ « +è¤ ² +æĵ ł +èIJ ¦ +çĿ ¬ +è° į +éĦ ° +æł ¾ +é¡ ı +ç¸ ± +æ¡ ¨ +éĨ ¬ +è¥ ² +è® ª +å© º +èį Ł +åĮ Ŀ +çĨ ł +èĽ Ĭ +æ¸ ļ +å´ ½ +é² ¤ +åķ ° +åĮ ķ +ä¸ IJ +è® ¥ +åı ½ +åı ¼ +çļ ¿ +è¿ Ĥ +åIJ Ĩ +å± ¹ +èĩ ¼ +è® ¹ +é© ® +çº « +æ± ŀ +æĬ ¡ +èĭ ĩ +åIJ ł +åIJ Ń +åIJ ® +å² ĸ +ä½ ĥ +çĭ Ī +åº ĩ +åIJ Ŀ +éĹ ° +æ± ¹ +å¿ ± +æĭ Ħ +æĭ Ĺ +èĮ ī +èĭ Ľ +èĮ ģ +çŁ ¾ +èĻ ı +åij » +åĴ Ħ +å¿ ¿ +èĤ ® +çĭ ŀ +çĸ Ł +çĸ Ļ +çĸ ļ +æ³ ŀ +å¸ ļ +å± ī +è¿ ¢ +é© ¹ +ç İ· +çıĬ ó +çıĬó ł +çıĬół Ħ +çıĬółĦ ģ +æĮ İ +æĭ ´ +åŀ Ľ +èį ¤ +æ® ĥ +çĽ ¹ +åĵ Ĩ +è´ » +æ¯ ¡ +çĭ ° +çĭ ¡ +æŁ Ĵ +æģ ĥ +è¯ ¬ +è¢ Ħ +è¯ ² +èļ ¤ +èĢ Ļ +åŁ Ĥ +æį İ +æį Į +æ¢ Ĩ +é ħĮ +çł ¾ +æ® ī +åĶ ł +æĻ Į +èļ £ +èļ ª +èļ ĵ +é¸ ¯ +åĶ ģ +åĶ Ĩ +åĢ Ķ +èĪ Ģ +è± º +èĥ ° +é¸ µ +é¸ ³ +é¦ ģ +ç¾ Ķ +æ¶ £ +æ¶ ķ +æĤ ¯ +è¯ ½ +è° Ĩ +ç¥ Ł +ç» ¢ +æį º +æį ¶ +æį » +æİ Ĥ +èı ł +èIJ ¤ +éħ Ĺ +çľ ¶ +åķ Ħ +èļ ¯ +èĽ Ģ +åĶ ¬ +å¸ · +éĵ IJ +éĵ Ľ +åģ İ +å¾ Ļ +èĦ ¯ +è± ļ +çĮ ĸ +çĹ Ĭ +æ¶ ® +æĥ Ń +æĤ ´ +æĥ ĭ +è° ļ +æı © +æIJ Ģ +æIJ Ķ +æ¦ Ķ +æ¤ Ń +éĽ ³ +åĸ ³ +è· Ľ +èľ ĵ +èľ Ĵ +é¹ ĥ +éĶ Ħ +çĶ ¥ +çŃ ı +çĮ © +çĮ ¬ +çĮ ¾ +çĹ ¢ +çĹ ª +æĥ ° +çª ĺ +è° ¤ +éļ ĺ +å© ¿ +é¹ ī +çij Ļ +æĸ Ł +æ¤ ¿ +éħ ª +éĽ ¹ +åĹ ¦ +è· · +è· º +è· ¤ +èľ Ī +èľ Ĺ +å¹ Į +é¦ ı +èª Ĭ +æ¼ ĵ +è¤ Ĥ +èĶ Ĺ +èĶ ¼ +åħ ¢ +è£ ³ +èľ » +èĿ ĩ +åĺ Ģ +éĶ ¹ +ç® ķ +ç® © +çĺ © +çĺ Ł +æ¼ ± +å¯ ¥ +éª ¡ +æĴ µ +æĴ ¬ +è± Į +åĺ ¹ +èĿ ł +èĿ Į +èĿ Ĺ +èĿ Ļ +éķ IJ +ç¨ ¼ +ç¯ ĵ +èĨ Ľ +é² « +çĺ ª +é² ¨ +æĨ Ķ +ç¿ © +è¤ ¥ +ç¼ Ń +åĻ © +çĵ ¢ +éľ İ +è¸ ± +è¹ Ĥ +èŁ Ĩ +é¹ ¦ +ç¯ ¡ +çĺ ¸ +çª ¿ +ç¼ ° +èĹ IJ +è¹ ĭ +èŁ ĭ +èŁ Ģ +èµ ¡ +èĩ Ĭ +é³ Ħ +ç³ ł +æĩ ¦ +åļ £ +éķ ° +é³ į +ç° ¸ +çĻ £ +é³ ĸ +é¬ ĵ +èł ķ +éľ ¹ +èº ı +é» ¯ +çĵ ¤ +çŁ Ĺ +ä¹ Ĥ +ä¹ ľ +åħ Ģ +å¼ ĭ +åŃ ij +åŃ ĵ +å¹ º +äº ĵ +å »¿ +ä¸ ı +åį ħ +ä» ĥ +ä» ī +ä» Ĥ +åĪ Ī +çĪ » +åį ŀ +éĹ © +è® £ +å¤ ¬ +çĪ ¿ +æ¯ ĭ +éĤ Ĺ +éĤ Ľ +èī ½ +èī ¿ +åı µ +ä¸ ķ +åĮ ľ +åĬ ¢ +åį Ł +åı ± +åı » +ä» ¨ +ä» Ł +ä» ¡ +ä» « +ä» ŀ +åį ® +æ° IJ +çĬ ° +åĪ į +éĤ Ŀ +éĤ Ļ +è® ¦ +è® § +è® « +å° » +éĺ ¡ +å° ķ +å¼ ģ +èĢ Ĵ +çİ İ +çİ ij +åľ ¬ +æī ¦ +åľ ª +åľ ¹ +æī ª +åľ ® +åľ ¯ +èĬ Ĭ +èĬ į +èĬ Ħ +èĬ ¨ +èĬ ij +èĬ İ +èĬ Ĺ +äº ĺ +åİ į +å¤ ¼ +æĪ į +å° ¥ +ä¹ © +æĹ ¯ +æĽ ³ +å² Į +å± º +åĩ ¼ +åĽ ¡ +éĴ ĩ +ç¼ ¶ +æ° ĺ +æ° ĸ +çī Ŀ +ä¼ İ +ä¼ Ľ +ä¼ ¢ +ä½ ¤ +ä» µ +ä¼ ¥ +ä¼ § +ä¼ ī +ä¼ « +åĽ Ł +æ± Ĩ +åĪ ĸ +å¤ Ļ +æĹ ® +åĪ İ +çĬ · +çĬ ¸ +èĪ Ľ +åĩ « +é Ĥ¬ +é¥ § +æ± Ķ +æ± ľ +æ± Ĭ +å¿ ĸ +å¿ ı +è® ´ +è® µ +è® · +èģ ¿ +èī ® +åİ ¾ +å¦ ģ +çº ¡ +çº £ +çº ¥ +çº ¨ +çİ ķ +çİ Ļ +æĬ Ł +æĬ Ķ +åľ » +åĿ į +æĬ ĥ +ã§ IJ +èĬ « +èĬ ¾ +èĭ Ī +èĭ £ +èĭ ĭ +èĬ ¼ +èĭ Į +èĭ ģ +èĬ © +èĬ ª +èĬ ¡ +èĬ Ł +èĭ Ħ +èĭ İ +èĭ ¡ +æĿ Į +æĿ ĵ +æĿ Ī +å¿ ij +åŃ Ľ +éĤ ´ +éĤ ³ +å¥ ģ +è± ķ +å¿ Ĵ +æ¬ ¤ +è½ « +è¿ ĵ +éĤ ¶ +å¿ IJ +åį £ +éĤ º +æĹ ° +åij ĭ +åij Ĵ +åij ĵ +åij Ķ +åij ĸ +æĹ ¸ +åIJ ¡ +èĻ ¬ +åIJ ½ +åIJ £ +åIJ ² +å¸ ı +å² Ī +å² ĺ +åħ ķ +åĽ µ +åĽ « +éĴ Ĭ +éĴ ĭ +é ĴĮ +è¿ ķ +æ° Ļ +æ° ļ +çī ¤ +ä½ ŀ +ä½ ļ +ä½ Ŀ +ä½ Ĺ +å½ · +ä½ ĺ +ä½ ¥ +è± ¸ +åĿ Į +èĤ Ł +å¥ Ĥ +åĬ ¬ +çĭ ģ +é¸ ł +é¥ ¨ +é¥ © +é¥ « +é¥ ¬ +åº ij +åº ĭ +çĸ Ķ +çĸ ĸ +èĤ ĵ +éĹ ± +éĹ ³ +çĤ Ģ +æ² £ +æ² ħ +æ² Ķ +æ² ¤ +æ² ı +æ² ļ +æ± © +æ± ¨ +æ² ¨ +æ± ´ +æ² Ĩ +æ² © +æ³ IJ +æĢ ĥ +æĢ Ħ +å¿ ¡ +å¿ ¤ +å¿ ¾ +æĢ ħ +å¿ ª +æĢ Ĩ +å¿ Ń +å¿ ¸ +è¯ Ĥ +è¯ ĥ +è¯ ħ +è¯ ĭ +è¯ Į +è¯ Ĵ +éĻ Ĥ +éĻ ī +å¦ © +å¦ ª +å¦ £ +å¦ Ĺ +å¦ « +å§ Ĵ +å¦ ¤ +åĬ Ń +åĪ Ń +éĤ ° +çº Ń +çº ° +çº ´ +çİ ¡ +çİ Ń +çİ ł +çİ ¢ +çİ ¦ +çĽ Ĥ +å¿ Ŀ +åĮ ¦ +åĿ © +æĬ ¨ +æĭ ¤ +åĿ « +æĭ Ī +åŀ Ĩ +æĬ » +åĬ ¼ +æĭ ĥ +æĭ Ĭ +åĿ ¼ +åĿ » +ã§ Ł +åĿ ¨ +åĿ Ń +æĬ ¿ +åĿ ³ +èĭ · +èĭ ¤ +èĮ ı +èĭ « +èĭ ľ +èĭ ´ +èĭ Ĵ +èĭ ĺ +èĮ Į +èĭ » +èĭ ĵ +èĮ ļ +èĮ Ĩ +èĮ ij +èĮ ĵ +èĮ Ķ +èĮ ķ +è ĮĢ +èĭ ķ +æŀ ¥ +æŀ ĩ +æĿ ª +æĿ ³ +æŀ § +æĿ µ +æŀ ¨ +æŀ ŀ +æŀ ĭ +æĿ » +æĿ · +æĿ ¼ +çŁ ¸ +ç łĢ +åĪ ³ +å¥ Ħ +æ® ģ +éĥ ı +è½ Ń +éĥ ħ +é¸ ¢ +çĽ ± +æĺ Ļ +æĿ ² +æĺ ĥ +åĴ Ĥ +åij ¸ +æĺ Ģ +æĹ » +æĺ ī +çĤ ħ +çķ Ģ +èĻ ® +åĴ Ģ +åij · +é» ¾ +åij ± +åij ¤ +åĴ Ĩ +åĴ Ľ +åij ¶ +åij £ +åĴ Ŀ +å² ¢ +å² ¿ +å² ¬ +å² « +å¸ Ļ +å² £ +å³ ģ +åĪ ¿ +å² · +åī Ģ +å¸ Ķ +å³ Ħ +æ² ĵ +åĽ ¹ +ç½ Ķ +éĴ į +éĴ İ +éĴ ı +éĴ Ĵ +éĴ ķ +éĤ ¾ +è¿ ® +çī ¦ +ç« º +è¿ ¤ +ä½ ¶ +ä¾ ij +ä¾ ī +èĩ ¾ +ä¾ Ĺ +ä¾ ı +ä¾ © +ä½ » +ä½ ¾ +ä¾ ª +ä½ ¼ +ä½ ¯ +ä¾ ¬ +å¸ Ľ +ä¾ Ķ +å¾ Ĥ +åĪ ½ +éĥ Ħ +ç± ´ +çĵ ® +æĪ Ĺ +èĤ ¼ +äı Ŀ +èĤ ± +èĤ « +è¿ © +éĥ ĩ +çĭ İ +çĭ į +çĭ Ĵ +åĴ İ +é¥ ¯ +é¥ ´ +åĨ ½ +åĨ ¼ +åº ĸ +çĸ ł +çĸ Ŀ +åħ ĸ +åĬ ¾ +ð¬ ī +ð¬ī ¼ +çĤ ĺ +çĤ Ŀ +çĤ Ķ +æ³ Ķ +æ² Ń +æ³ · +æ³ ± +æ³ ħ +æ³ ł +æ³ º +æ³ ĸ +æ³ « +æ³ ® +æ² ± +æ³ ¯ +æĢ Ļ +æĢ µ +æĢ ¦ +æĢ Ľ +æĢ ı +æĢ į +ã ¤ +㤠ĺ +æĢ © +æĢ « +æĢ ¿ +å® ķ +ç© ¹ +å® ĵ +è¯ ĵ +è¯ Ķ +è¯ ĸ +è¯ ĺ +æĪ ¾ +è¯ Ļ +æĪ ½ +éĥ ĵ +è¡ © +ç¥ Ĩ +ç¥ İ +ç¥ ĩ +è¯ ľ +è¯ Ł +è¯ £ +è¯ ¤ +è¯ § +è¯ ¨ +æĪ ķ +éĻ Ķ +å¦ ² +å¦ ¯ +å§ Ĺ +å¸ ij +åŃ ¥ +é© ½ +èĻ ± +è¿ ¨ +ç» Ģ +ç» ģ +ç» Ĥ +é© · +é© ¸ +ç» ī +ç» Į +éª Ģ +çĶ ¾ +çı ı +çı IJ +çı ij +çİ ³ +é¡ ¸ +çı ī +çı Ī +æĭ ® +åŀ Ń +æĮ Ŀ +æĮ ŀ +åŀ ¤ +èµ ³ +è´ ² +åŀ ± +åŀ Į +åŀ § +åŀ ĵ +æĮ ¦ +åŀ ł +èį ļ +èį ij +è´ ³ +èį ľ +èİ Ĵ +èĮ ¼ +èĮ ´ +èĮ ± +èİ Ľ +èį ŀ +èĮ ¯ +èį ı +èį ĩ +èį ĥ +èį ł +èĮ Ń +åŀ © +èį ¥ +èį ¦ +èį ¨ +èį © +åī ĭ +èį ª +èį ¬ +èį ® +æŁ ° +æł ī +æŁ ĺ +æł Ĭ +æŁ © +æŀ ° +æł Į +æŁ Ļ +æŀ µ +æŀ ³ +æŁ ŀ +æŁ Ŀ +æł Ģ +æŁ ¢ +æł İ +æŁ Ī +æŁ ģ +æŀ · +æŁ ½ +åī Į +éħ Ĭ +éĥ ¦ +çĶ Ń +çł Ĺ +çł ĺ +çł Ĵ +æĸ « +çł Ń +çł ľ +èĢ · +èĻ º +æ® Ĥ +æ® ĩ +æ® Ħ +è½ ± +è½ ² +è½ ³ +è½ ¶ +è½ ¸ +èĻ ¿ +æ¯ ĸ +è§ ĩ +å° ľ +åĵ IJ +çľ Ħ +çľ į +ðł ³ +ðł³ IJ +éĥ ¢ +çľ ĩ +çľ Ĭ +çľ Ī +ç¦ º +åĵ Ĥ +åĴ ´ +æĽ · +æĺ ´ +åĴ ¦ +åĵ ĵ +åĵ Ķ +çķ İ +åij ² +èĥ Ħ +çķ ĭ +çķ Ī +èĻ ¼ +èĻ » +çĽ ħ +åĴ £ +åĵ ķ +åī IJ +éĥ § +åĴ » +åĽ ¿ +åĴ ¿ +åĵ Į +åĵ Ļ +åĵ ļ +åĴ © +åĴ ¤ +åĵ Ŀ +åĵ ı +åĵ ŀ +å³ £ +ç½ ĺ +å³ Ĵ +å³ ¤ +å³ ĭ +è´ ¶ +éĴ ļ +éĴ ¡ +éĴ £ +éĴ ¤ +éĴ « +æ° ¡ +çī ¯ +éĥ ľ +ç§ ķ +ç§ Ń +ç« ½ +ç¬ Ī +ä¿ ¦ +ä¿ ¨ +ä¿ ħ +åı Ł +åŀ ¡ +çī ® +ä¿ £ +ä¿ ļ +çļ Ī +ä¿ Ł +éĢ ħ +å¾ ĩ +å¾ ī +èĪ ¢ +éĥ Ĺ +ä¿ İ +éĥ ¤ +çĪ ° +éĥ Ľ +çĵ ´ +èĥ ¨ +èĥ ª +èĥ Ľ +èĥ Ĥ +èĥ Ļ +èĥ į +èĥ Ĺ +è ĥĿ +æľ IJ +èĥ « +é¸ ¨ +åĮ į +çĭ ¨ +çĭ ¯ +é£ ij +çĭ © +çĭ ² +è¨ ĩ +éĢ Ħ +æĺ Ŀ +é¥ · +é¥ ¸ +é¥ ¹ +åŃ ª +å¨ Ī +åº ¥ +çĸ ¬ +çĸ £ +çĸ ¥ +çĸ Ń +åº ł +ç« ij +é£ Ĵ +éĹ ¼ +éĹ ¾ +éĹ ¿ +éĺ Ĥ +ç¾ ij +è¿ ¸ +ç± ¼ +éħ ĭ +çĤ » +çĥ Ģ +çĤ · +æ´ ± +æ´ ¹ +æ´ § +æ´ Į +æµ ĥ +æ´ ĩ +æ´ Ħ +æ´ Ļ +æ¶ İ +æ´ İ +æ´ « +æµ į +æ´ ® +æ´ µ +æµ Ĵ +æµ Ķ +æµ ķ +æ´ ³ +æģ ¸ +æģ ĵ +æģ ¹ +æģ « +æģ » +æģ Ĥ +æģ ª +æģ ½ +å® ¥ +æī ĥ +è¡ ² +è¡ ½ +è¡ ¿ +è¢ Ĥ +ç¥ ľ +ç¥ ĵ +ç¥ ļ +è¯ ® +ç¥ Ĺ +ç¥ ¢ +è¯ ° +è¯ ³ +é¸ © +æĺ ¶ +åĴ « +å¼ Ń +çī ģ +èĥ ¥ +éĻ Ł +å§ ® +å¨ Ĩ +å§ Ŀ +å§ £ +å§ ĺ +å§ ¹ +ç¾ ¿ +çĤ ± +çŁ ľ +ç» Ķ +éª ģ +éª ħ +ç» Ĺ +ç» Ľ +éª Ī +èĢ ĸ +æĮ Ī +çı ¥ +çı Ļ +é¡ ¼ +çı ° +çı © +çı § +çı £ +çı ŀ +çIJ ¤ +çı ² +æģ ļ +åŁ ķ +åŁ ĺ +åŁ Ļ +åŁ ļ +æĮ ¹ +èĢ Ĩ +èĢ Ħ +åŁ Ĵ +æį ĭ +è´ ½ +åŀ ¸ +æį ĥ +çĽ į +èį ¸ +èİ ³ +èİ ´ +èİ ª +èİ ł +èİ ľ +èİ ħ +èį ¼ +èİ © +èį ½ +èİ ¸ +èį » +èİ ¨ +é¸ ª +èİ ¼ +æł ² +æł ³ +æ¡ ¡ +æ¡ İ +æ¡ ¢ +æ¡ ¤ +æ¢ ĥ +æł Ŀ +æ¡ ķ +æ¡ ģ +æ¡ § +æ¡ ħ +æł Ł +æ¡ ī +æł © +éĢ ij +éĢ ĭ +å½ § +é¬ ² +è± ĩ +éħ IJ +éĢ ¦ +åİ Ŀ +åŃ ¬ +çł Ŀ +çł ¹ +çł § +çł · +çł Ł +çł ¼ +çł ¥ +çł £ +åī ŀ +çł » +è½ ¼ +è½ ¾ +è¾ Ĥ +é¸ « +è¶ ¸ +é¾ Ģ +é¸ ¬ +èĻ Ķ +çľ ¬ +åĶ Ľ +çľ Ļ +åĵ § +åĵ ½ +æĻ ģ +é¸ ® +è¶ µ +è¶ ¿ +çķ Ľ +èļ ¨ +èļ ľ +èļ į +èļ ĭ +èļ ¬ +èļ Ŀ +èļ § +åĶ ¢ +åľ Ħ +åĶ £ +åĶ ı +çĽ İ +åĶ ij +å´ Ĥ +å´ ĥ +ç½ ¡ +ç½ Ł +è§ Ĭ +èµ ħ +éĴ ² +éĴ µ +éĴ ¹ +éĴ º +éĴ ½ +éĴ ¼ +éĴ ¿ +éĵ Ģ +éĵ Ħ +éĵ Ĩ +éĵ Ī +éĵ ī +éĵ Ĭ +éĵ ĭ +éĵ Į +é ĵį +ä ¥ +ä¥ ½ +éĵ İ +æ° © +æ° ¤ +æ° ¦ +æ¯ ª +èĪ IJ +ç§ £ +ç§ « +çĽ ī +ç¬ Ħ +ç¬ ķ +ç¬ Ĭ +ç¬ ı +ç¬ Ĩ +ä¿ ¸ +ä¿ µ +åģ Į +ä¿ ³ +ä¿ ¶ +åĢ ¬ +åĢ ı +æģ ģ +åĢ Ń +ä¿ ¾ +åĢ ľ +éļ ¼ +éļ ½ +åĢ Į +åĢ ¥ +èĩ ¬ +éĥ « +åĢ ¨ +è¡ Ħ +é¢ Ģ +å¾ ķ +èĪ « +è¡ ¾ +èĥ ¯ +èĥ ± +èĥ ´ +èĥ Ń +èĦ į +èĥ ¼ +èĦ Ĵ +é¸ ± +é¸ ² +çĭ · +çĮ ģ +çĭ ³ +çĮ ĥ +çĭ º +éĢ ĸ +æ¡ Ģ +é¥ ½ +åĩ ĩ +æĮ Ľ +äº ³ +çĸ ³ +çĸ ´ +çĸ ¸ +çĸ ½ +çĹ Ī +çĸ ± +çĹ Ĥ +çĹ ī +è¡ ® +é¢ ĥ +æģ £ +æĹ Ĩ +æĹ Ħ +æĹ ĥ +éĺ ĥ +éĺ Ħ +è¨ ļ +éĺ Ĩ +æģ Ļ +ç² ij +çĥ ľ +çĥ © +çĥ Ĭ +åī ¡ +éĥ ¯ +çĥ ¬ +æ¶ ij +æµ ¯ +æ¶ ŀ +æ¶ Ł +å¨ ij +æ¶ ł +æµ ŀ +æ¶ ĵ +æµ ¥ +æ¶ Ķ +æµ ľ +æµ ł +æµ £ +æĤ ļ +æ ĤŃ +æĤ Ŀ +æĤ Ĵ +æĤ Į +æĤ Ľ +çª Ī +åī ľ +è¯ ¹ +è¯ ¼ +è¢ Ĵ +è¢ ¢ +è¯ ¿ +è° Ģ +è° Ĥ +è° Ħ +è° ĩ +å± IJ +å± Ļ +éĻ ¬ +åĭ IJ +å¥ ĺ +çī Ĥ +èļ © +éĻ ² +å¨ Į +å¨ ī +å¨ ² +å¨ ´ +å¨ £ +å¨ ĵ +å© Ģ +çķ ļ +éĢ ¡ +ç» ł +éª Ĭ +ç» ¡ +éª ĭ +ç» ¦ +ç» ¨ +éª İ +éĤ ķ +é¸ ¶ +å½ Ĺ +èĢ ľ +çĦ ĺ +èĪ Ĥ +çIJ ı +çIJ ĩ +éº ¸ +æı ¶ +åŁ ´ +åŁ ¯ +æį ¯ +æİ ³ +æİ ´ +åŁ ¸ +åŁ µ +èµ § +åŁ ¤ +æį Ń +éĢ µ +åŁ Ŀ +åł ĭ +åł į +æİ ¬ +é¸ · +æį ½ +æİ Ĭ +åł ī +æİ ¸ +æį © +æİ ® +æĤ « +åŁ Ń +åŁ ½ +æİ ĩ +æİ ¼ +èģ ĥ +èIJ ģ +èı ĺ +åł ĩ +èIJ ĺ +èIJ ĭ +èı ½ +èı ĸ +è IJľ +èIJ ¸ +èIJ ij +æ£ » +èı Ķ +èı Ł +èIJ ı +èı ¹ +èı ª +èı ħ +èı Ģ +èı ° +èı ¡ +æ¢ ¿ +æ¢ ı +è§ ĭ +æ¡ ´ +æ¡ · +æ£ ģ +æ¡ « +æ£ Ĥ +åķ ¬ +éĥ ¾ +æķ ķ +è± ī +éĦ Ħ +éħ ŀ +ç¡ İ +ç¡ Ń +ç¡ ĸ +ç¡ Ĺ +ç¡ IJ +ç¡ ĩ +ç¡ Į +é¸ ¸ +çĵ ł +åĮ ı +åİ © +æ® Ĵ +æ® ĵ +æ® į +èµ ī +éĽ © +è¾ Ħ +åł ij +çľ Ń +çľ ¦ +åķ § +æĻ ¡ +æĻ ¤ +çľ µ +åľ Ĭ +åĸ ı +åķ ī +åĭ ĸ +æĻ ŀ +åĶ µ +æĻ Ĺ +åķ Ń +çķ ¦ +è¶ º +åķ ® +è· Ħ +èļ ¶ +è ĽĦ +èĽ İ +èĽ Ĩ +èļ ° +åľ ī +èļ ± +èĽ ī +èĽ ı +èļ ´ +åķ ģ +åķ ķ +åĶ ¿ +åķ IJ +åĶ ¼ +åĶ · +åķ ĸ +åķ µ +åķ ¶ +åķ · +åĶ ³ +åĶ ° +åķ ľ +å¸ » +å´ ļ +å´ ¦ +å¸ ¼ +å´ ® +å´ ¤ +å´ Ĩ +èµ ĩ +èµ Ī +èµ Ĭ +éĵ ij +éĵ Ĵ +éĵ Ĺ +éĵ Ļ +éĵ Ł +éĵ ¡ +éĵ ¢ +éĵ £ +éĵ ¤ +éĵ § +éĵ ¨ +éĵ © +éĵ ª +éĵ « +éĵ ¯ +éĵ ° +éĵ ± +éĵ ³ +éĵ µ +éĵ · +çī ¾ +é¸ ¹ +ç§ ¾ +éĢ ¶ +ç¬ º +çŃ ĩ +ç¬ ¸ +ç¬ ª +ç¬ ® +ç¬ ł +ç¬ ¥ +ç¬ ¤ +ç¬ ³ +ç¬ ¾ +ç¬ ŀ +åģ ¾ +åģ ĥ +åģ ķ +åģ Ī +åĤ Ģ +åģ ¬ +åģ » +çļ ij +çļ İ +é¸ » +å¾ ľ +èĪ ¸ +èĪ » +èĪ ´ +èĪ · +é¾ Ľ +ç¿ İ +èĦ ¬ +èĦ ĺ +èĦ ² +åĮ IJ +çĮ Ĺ +çĮ ¡ +çĮ ŀ +æĸ Ľ +çĮ ķ +é¦ Ĺ +é¦ ĥ +é¦ Ħ +é¸ ¾ +åº ¹ +åº ¾ +çĹ Ķ +çĹ į +ç¿ Ĭ +æĹ Į +æĹ İ +è¢ ¤ +éĺ ĩ +éĺ Ī +éĺ ī +éĺ Ĭ +éĺ ĭ +éĺ į +éĺ ı +ç¾ Ł +ç² Ŀ +çĦ IJ +çĦ ĵ +çĦ Ĺ +æ· ħ +æ· ŀ +æ¸ İ +æ¶ ¿ +æ· ĸ +æĮ ² +æ· ł +æ¶ ¸ +æ¸ ij +æ· ¦ +æ· Ŀ +æ¶ ª +æ· Ļ +æ¶ « +æ¸ Į +æĤ » +æĤ ± +æ ĥĿ +æĥ ĺ +æĥ Ĩ +æĥ ļ +æĥ ĩ +æĥ ® +çª ķ +è° Į +æī Ī +çļ ² +è° ij +è£ Ĩ +è¢ · +è£ ī +è° Ĵ +è° Ķ +è° ķ +è° ĸ +è° Ĺ +è° Ļ +è° Ŀ +éĢ ¯ +éĥ ¿ +éļ Ī +ç² ľ +éļ į +éļ Ĺ +å© Ĭ +å¨ ¼ +å© ¢ +å© µ +èĥ ¬ +è¢ Ī +ç¿ Į +æģ ¿ +æ¬ ¸ +ç» « +éª IJ +ç» ¯ +ç» ± +éª Ĵ +ç» ² +éª ĵ +ç» ¶ +ç» º +ç» » +ç» ¾ +éª ĸ +ç¼ ģ +èĢ ł +çIJ « +çIJ µ +çIJ ¶ +çIJ ¥ +çIJ ¨ +çIJ ° +çIJ ® +çIJ ¯ +çIJ ¬ +çIJ ļ +è¾ ĩ +é¼ ĭ +æı ³ +åł ŀ +æIJ ½ +æı ¸ +æı ł +åł Ļ +è¶ Ħ +æı ĸ +é¢ ī +å¡ Ħ +æı ¿ +èĢ ĭ +æı Ħ +èĽ © +èĽ ° +å¡ Ĩ +æij Ĵ +æı Ĩ +æİ ¾ +èģ Ĵ +èij ij +èij ļ +éĿ ° +éĿ ¸ +èij ³ +èij º +èij ¸ +èIJ ¼ +èij ¶ +è ĴĮ +èij Ń +æ¥ ® +æ £¼ +æ¤ Ł +æ£ ¹ +æ¤ ¤ +æ£ ° +èµ į +æ¤ ĭ +æ¤ ģ +æ¤ ª +æ¤ IJ +é¹ ģ +éħ ¤ +éħ ¢ +éħ ¡ +é¹ Ĥ +æ® ļ +æ® Ľ +éĽ ± +è¾ ĭ +æ¤ ł +è¾ İ +çĿ Ħ +çĿ ĩ +çĿ ĥ +æĪ ¢ +åĸ ĭ +åĹ Ĵ +åĸ ĥ +åĸ ± +åĸ ¹ +æĻ · +åĸ Ī +è· ĸ +è· Ĺ +è· ŀ +è· ļ +è· İ +è· ı +è· Ĩ +èĽ ± +èĽ ² +èĽ Ń +èĽ ³ +èĽ IJ +èĽ Ķ +èĽ ŀ +èĽ ´ +èĽ ĺ +åĸ ģ +åĸ Ł +åķ ¾ +åĹ ĸ +åĸ ij +åĹ Ł +åĹ ŀ +åĸ Ļ +åµ ĺ +åµ ĸ +å´ ´ +éģ Ħ +è© Ī +åµ İ +å µ¬ +åµ Ľ +åµ ¯ +åµ Ŀ +åµ « +å¹ Ħ +åµ ĭ +èµ ķ +éĵ » +éĵ ¼ +éĵ ¿ +éĶ ĥ +éĶ Ĩ +éĶ ĩ +éĶ ī +éĶ ı +éĶ ij +éĶ Ĵ +éĶ Ķ +éĶ ķ +æİ £ +çŁ ¬ +æ° ° +æ¯ ³ +æ¯ ½ +çĬ Ĭ +çĬ Ħ +çĬ ĭ +é ¹Ħ +çĬ į +åµ ĩ +é» į +ç¨ ĥ +ç¨ Ĥ +çŃ ļ +çŃ µ +çŃ Į +åĤ £ +åĤ Ī +èĪ Ħ +çī į +åĤ ¥ +åĤ § +éģ ij +åĤ © +å¾ ¨ +åª Ń +çķ ² +å¼ ij +ç¿ ķ +é¹ Ĩ +èħ Ī +èħ ĵ +èħ Ĩ +èħ ´ +èħ ļ +èħ ± +é± ¿ +é² Ģ +é² Ĥ +çĮ ¢ +çĮ ¹ +çĮ ¥ +é£ ĵ +è§ ŀ +è§ ļ +çĮ ± +é¢ İ +é£ § +é¦ ĩ +é¦ Ĭ +äº µ +èĦ Ķ +è£ Ĵ +çĹ £ +çĹ ¨ +çĹ ¦ +çĹ ŀ +çĹ ¤ +çĹ § +èµ ĵ +ç« ¦ +çĵ ¿ +åķ » +é¢ ı +é¹ ĩ +éĺ ij +éĺ Ĵ +éĺ ķ +ç² ŀ +éģ Ĵ +åŃ ³ +çĦ ¯ +çĦ ľ +çĦ ± +é¹ Ī +æ¸ « +æ¹ ® +æ¹ İ +æ¹ ľ +æ¹ į +æ¹ « +æº ² +æ¹ Ł +æº Ĩ +æ¹ ² +æ¹ Ķ +æ¹ ī +æ¸ ¥ +æ» ģ +æĦ ł +æĥ º +æĦ ¦ +æĥ ´ +æĦ Ģ +æĦ İ +æĦ Ķ +åĸ ¾ +å¯ IJ +è° Ł +è£ ¢ +è£ İ +è£ ¥ +ç¥ ¾ +è° ł +è° ¡ +è° ¥ +è° § +åŃ ± +å¼ ¼ +å· ½ +éª ĺ +åª ª +å· ¯ +ç¿ ļ +çļ ´ +éª Ľ +ç¼ Ĥ +ç¼ ĥ +ç¼ Ħ +å½ ĺ +ç¼ ĩ +ç¼ Ī +ç¼ Į +ç¼ ij +ç¼ Ĵ +ç¼ Ĺ +é£ ¨ +èĢ ¢ +çij ģ +çij Ĺ +çij Ħ +éģ ¨ +éª ľ +éŁ « +é« ¡ +å¡ ¬ +éĦ ¢ +è¶ Ķ +è¶ ij +æij ħ +æij ģ +èľ ĩ +æIJ ĭ +æIJ ª +æIJ IJ +æIJ Ľ +æIJ ł +æij Ī +å½ Ģ +æ¯ Ĥ +æIJ ¦ +æIJ ¡ +èĵ ģ +æĪ ¡ +è ĵį +éĦ ŀ +èĵ IJ +èĵ ¦ +é¹ ĭ +èĴ ½ +èĵ ĸ +èĵ Ĭ +èĴ ¯ +èĵ Ł +èĵ ij +èĴ º +èĵ ł +èĴ Ł +èĴ ¡ +èĴ ¹ +èĴ ´ +èĴ Ĺ +èĵ ¥ +æ¥ Ķ +æ¥ Ĥ +æ¥ Ŀ +æ¥ « +æ¥ ¸ +æ¤ ´ +æ§ Į +æ¥ ¯ +çļ Ļ +æ¦ Ī +æ§ İ +æ¦ ī +æ¥ ¦ +æ¥ £ +æ¥ ¹ +æ¤ ½ +åī ½ +éħ © +èľ ĥ +ç¢ Ľ +ç¢ ĵ +ç¡ ¼ +ç¢ ī +ç¢ ļ +ç¢ ĩ +ç¢ ľ +é¹ Į +è¾ ı +é¾ ĥ +é¾ ħ +è¨ ¾ +ç² ² +çĿ ļ +åĹ ª +éŁ ª +åĹ · +åĹ ī +çĿ ¨ +çĿ ¢ +éĽ İ +çĿ ¥ +åĹ ij +åĹ « +åĹ ¬ +åĹ Ķ +åĹ Ŀ +æĪ ¥ +åĹ Ħ +çħ ¦ +æļ Ħ +éģ ¢ +æ ļĮ +è· ¬ +è· ¶ +è ·¸ +è· IJ +è· £ +è· ¹ +èĽ ¸ +èľ Ĭ +èľ į +èľ ī +èľ £ +çķ ¹ +èĽ ¹ +åĹ ¥ +åĹ ² +åĹ ³ +åĹ Į +åĹ į +åĹ IJ +åĹ ¤ +åĹ µ +ç½ ¨ +åµ Ĭ +åµ ´ +éª ° +éĶ Ĺ +éĶ Ľ +éĶ ľ +éĶ Ŀ +éĶ ŀ +éĶ Ł +éĶ ¢ +éĶ ¨ +éĶ © +éĶ Ń +éĶ ± +éĽ ī +æ° ² +çĬ ı +æŃ ĥ +ç¨ ŀ +ç¨ Ĺ +ç¨ Ķ +çŃ ł +çŃ ¢ +çŃ ® +çŃ ² +çī Ĵ +æķ « +å¾ Ń +æĦ Ĩ +èī Ħ +è§ İ +æ¯ ¹ +è² Ĭ +è² ħ +è² ī +é¢ Ķ +èħ ł +èħ © +èħ ¼ +èħ Ń +è ħ§ +å¡ į +åª µ +é² ħ +é² Ĩ +é² ĩ +é² Ī +é² ĭ +é² IJ +èĤ Ħ +é¹ IJ +é£ ķ +è§ ¥ +éģ Ľ +é¦ IJ +é¹ ij +äº ¶ +çĺ ĥ +çĹ ± +çĹ ¼ +çĹ ¿ +çĺ IJ +çĺ ģ +çĺ Ĩ +éº Ĥ +æŃ Ĩ +æĹ Ĵ +éĺ ĸ +éĺ Ĺ +ç¾ § +è± ¢ +ç² ³ +çĮ · +çħ ³ +çħ ¨ +çħ ħ +çħ Ĭ +çħ ¸ +çħ º +æ» Ł +æº ± +æº ĺ +æ¼ Ń +æ» ¢ +æº ¥ +æº ½ +è£ Ł +æº » +æº · +æ» Ĺ +æ» « +æº ´ +æ» ı +æ» ĥ +æ» ¦ +æº ı +æ» Ĥ +æ» ĵ +æº Ł +æ» ª +æĦ « +æħ Ĭ +é² İ +éª ŀ +çª ł +çª £ +è£ ± +è£ ¨ +è£ ¾ +è£ ° +ç¦ Ĭ +è° © +è° ª +åª ¾ +å« « +åª ² +å« Ĵ +å« Ķ +åª ¸ +ç¼ Ļ +ç¼ ľ +ç¼ Ľ +è¾ Ķ +éª Ŀ +ç¼ Ł +ç¼ ¡ +ç¼ ¢ +ç¼ £ +éª Ł +èĢ ¥ +çĴ Ī +çij Ń +çį Ĵ +è§ ı +æħ Ŀ +å« ł +åı Ĩ +æij ½ +å¢ ģ +æĴ Ĥ +æij ŀ +æĴ Ħ +ç¿ ¥ +è¸ ħ +æij Ń +å¢ ī +å¢ Ĵ +æ¦ ĸ +ç¶ ¦ +èĶ « +èĶ · +éĿ º +éĿ ¼ +éŀ ħ +éĿ ¿ +çĶ į +èĶ ¸ +èĶ Ł +èĶ º +æĪ ¬ +èķ ĸ +èĶ » +èĵ ¿ +æĸ ¡ +é¹ ķ +èĵ ¼ +æ¦ Ľ +æ¦ § +æ¦ « +æ¦ Ń +æ§ Ķ +æ¦ ± +æ§ ģ +æ§ ł +æ¦ · +åĥ ° +éħ ½ +éħ ¹ +ç¢ ¡ +ç¢ ´ +ç¢ £ +ç¢ ² +èĩ § +è± ¨ +æ® ¡ +éľ ģ +èľ ļ +é¾ ĩ +é¾ Ī +ä ģ +äģ ĸ +çĿ ½ +åĺ ŀ +åĺ Ī +åĺ Į +åĺ ģ +æļ Ŀ +è¸ Į +è¸ ī +èľ ŀ +èľ ¥ +èľ ® +èĿ Ī +èľ ´ +èľ ± +èľ © +èľ · +èľ ¿ +èŀ Ĥ +èľ ¢ +åĺ ¡ +é¹ Ĺ +åĺ £ +åĺ ¤ +åĺ ļ +åĹ ¾ +åĺ § +ç½ ´ +ç½ ± +å¹ Ķ +å¶ Ĥ +å¹ Ľ +èµ Ļ +ç½ Ĥ +éª · +éª ¶ +é¹ ĺ +éĶ ² +éĶ ´ +éĶ ¶ +éĶ · +éĶ ¸ +éĶ µ +éķ Ĥ +çĬ Ĵ +ç® IJ +ç® ¦ +ç® § +ç® ¸ +ç® ¬ +ç® ħ +ç® ª +ç® ľ +ç® ¢ +ç® ĵ +åĥ ĸ +åĦ Ĩ +åĥ ³ +åĥ Ń +åĬ ģ +åĥ ® +éŃ ĥ +éŃ Ĩ +çĿ ¾ +èī ĭ +éĦ ± +èĨ Ī +èĨ ij +é² ij +é² Ķ +é² ļ +é² Ľ +é² Ł +çį IJ +è§ « +éĽ Ĵ +å¤ ¤ +é¦ ij +éĬ ® +å¡ ¾ +çĺ Į +çĺ Ĭ +çĺ ĺ +çĺ Ļ +æĹ ĸ +èĨ Ĥ +éĺ ļ +éĦ ¯ +é² ŀ +ç² ¿ +ç² ¼ +ç³ ģ +æ§ Ĭ +é¹ ļ +çĨ ĺ +çĨ ¥ +æ½ ¢ +æ¼ ķ +æ» ¹ +æ¼ ¯ +æ¼ ¶ +æ½ ĭ +æ½ ´ +æ¼ ª +æ¼ ī +æ¼ © +æ¾ ī +æħ µ +æIJ ´ +çª ¨ +å¯ ¤ +ç¶ ® +è° ® +è¤ ¡ +è¤ Ļ +è¤ ĵ +è¤ Ľ +è¤ Ĭ +è° ¯ +è° ° +è° ² +å± £ +é¹ Ľ +å« ± +å« ĸ +å« ¦ +å« ļ +å «ĺ +é¼ IJ +çŀ Ģ +é¹ ľ +éª ł +ç¼ ¥ +ç¼ ¦ +ç¼ § +ç¼ ¨ +éª ¢ +ç¼ « +èĢ ¦ +èĢ § +çĴ ľ +çĴ İ +çĴ ģ +å¥ Ń +é« ¯ +é« « +æĴ · +æĴ ħ +èµ Ń +æĴ ¸ +éĭ Ĩ +æĴ Ļ +æĴ º +å¢ Ģ +èģ © +è§ IJ +éŀ ij +èķ Ļ +éŀ Ĵ +èķ Ī +èķ ¨ +èķ ¤ +èķ ŀ +èķ º +çŀ ¢ +èķ ĥ +èķ ² +èµ ľ +æ§ ¿ +æ¨ ¯ +æ§ Ń +æ¨ Ĺ +æ¨ ĺ +æ§ ² +éĨ Į +éĨ ħ +éĿ ¥ +éŃ ĩ +é¤ į +ç£ Ķ +ç£ Ļ +éľ Ī +è¾ ĺ +é¾ ī +é¾ Ĭ +è§ ij +çŀ Į +ç ŀĭ +çŀ ij +åĺ Ń +åĻ İ +åĻ ¶ +é¢ Ļ +æļ ¹ +åĻ ĺ +è¸ Ķ +è¸ Ŀ +è¸ Ł +è¸ Ĵ +è¸ ¬ +è¸ ® +è¸ ¯ +è¸ º +è¸ ŀ +èĿ ½ +èĿ ¾ +èĿ » +èĿ ° +èĿ ® +è ŀĭ +èĿ ĵ +èĿ £ +è Ŀ¼ +åĺ ¬ +é¢ ļ +åĻ į +åĻ Ļ +åĻ Į +åĻ Ķ +é¢ Ľ +å¹ ŀ +å¹ ¡ +å¶ Ļ +å¶ Ŀ +éª º +éķ Ĭ +éķ ī +éķ Į +éķ ı +éķ Ĵ +éķ ĵ +éķ Ķ +ç¨ · +ç® ´ +ç¯ ij +ç¯ ģ +ç¯ Į +çī ĸ +åĦ ĭ +èĻ ¢ +é¹ ŀ +èĨ ĺ +é² ł +é² ¡ +é² ¢ +é² £ +é² ¥ +é² § +é² © +çį Ĺ +çį ł +è§ ¯ +é¦ ĵ +é¦ Ķ +éº ¾ +å» Ľ +çĺ Ľ +çĺ ¼ +çĺ ¢ +çĺ ł +é½ ij +ç¾ ° +𥠻 +ð¥» Ĺ +ç³ Į +ç³ į +ç³ ħ +çĨ ľ +ç Ĩµ +æ¾ į +æ¾ Į +æ½ ¸ +æ½ ¦ +æ½ ² +éĭ Ī +æ½ Ł +æ½ º +å¯ ® +çª ³ +è° ³ +è¤ ´ +è¤ Ł +è¤ « +è° µ +çĨ ¨ +å± ¦ +åĭ ° +æĪ ® +èĿ ¥ +ç¼ ¬ +ç¼ ® +ç¼ ¯ +éª £ +çķ ¿ +èĢ © +èĢ ¨ +èĢ ª +çĴ Ł +éĿ Ľ +çĴ ł +çĴ ĺ +èģ ± +èŀ ¯ +é« » +é« Ń +é« ¹ +æĵ Ģ +çĶ ı +æĵ ŀ +ç¸ ł +ç£ ¬ +é¢ ŀ +èķ » +é¢ Ł +èĸ ¤ +èĸ ¨ +æª ł +èĸ ı +èĸ ® +èĸ ľ +èĸ ħ +æ¨ ¾ +æ© Ľ +æ© ĩ +æ¨ µ +æª İ +æ© ¹ +æ¨ ½ +æ¨ ¨ +æ© ¼ +å¢ ¼ +æ© IJ +ç¿ ® +éĨ IJ +éĨ į +éĨ ļ +ç£ ² +èµ Ŀ +æ® ª +éľ ı +éĮ ¾ +è¾ ļ +éģ ½ +æ° ħ +çŀ Ł +çŀ ł +çŀ ° +åļ Ħ +åļ Ĩ +åĻ ¤ +æļ ¾ +è¹ Ģ +è¸ µ +è¸ ½ +è¹ ī +è¹ ģ +èŀ ¨ +èŀ Ī +èŀ ħ +èŀ Ń +èŀ ł +èŀ Ł +åĻ ± +åĻ « +åĻ » +åĻ ¼ +ç½ ¹ +åľ ľ +ä ¦ +ä¦ ĥ +éķ Ĺ +éķ ĺ +éķ ļ +éķ Ľ +éķ Ŀ +éķ ŀ +éķ ł +æ° ĩ +æ° Ĩ +ç© ij +ç¯ Ŀ +ç¯ ¥ +ç¯ ¦ +ç¯ ª +ç¯ Ļ +çĽ ¥ +åĬ ĵ +ç¿ ± +éŃ ī +éŃ Ī +å¾ ¼ +æŃ Ļ +èĨ ¦ +èĨ Ļ +é² ® +é² ± +é² ³ +é² ´ +é² µ +é² · +é² » +çį ´ +çį Ń +çį ¬ +éĤ Ĥ +é¹ § +å» ¨ +èµ Ł +çĺ ° +å» ª +çĺ ¿ +çĺ µ +çĺ ´ +çĻ ĥ +çĺ ³ +éº ĩ +éº Ī +å ¬´ +å£ ħ +ç³ Ĺ +çĶ ij +çĩ İ +çĩ ł +çĩ Ķ +çĩ § +æ¿ ij +æ¿ ī +æ½ ŀ +æ¾ § +æ¾ ¹ +æ¾ ¥ +æ¾ ¶ +æ¿ Ĥ +è¤ ° +çª ¸ +å¬ ĸ +çĬ Ł +éļ ° +å¬ Ĺ +é¢ ¡ +ç¼ ± +ç¼ ² +ç¼ ³ +çĴ © +çĴ ª +èŀ « +æĵ ¤ +å£ ķ +è§ ³ +ç½ Ħ +æĵ ¢ +èĸ ¹ +éŀ ¡ +éŀ ¬ +èĸ · +èĹ ĵ +èĹ ģ +æª Ħ +æª © +æĩ ĭ +éĨ ¢ +ç¿ ³ +ç¤ ħ +ç£ ´ +é¹ © +é¾ ĭ +é¾ Į +è± ³ +å£ ij +é» » +åļ ı +åļ ħ +è¹ ij +è¹ Ĵ +è¹ Ĭ +è Ł¥ +èŀ ¬ +èŀ µ +çĸ ĥ +èŀ ³ +èŁ ij +åļ ĵ +ç½ ½ +ç½ ¾ +å¶ · +é» ľ +é» Ŀ +é« ģ +é« Ģ +éķ ¡ +éķ ¢ +éķ £ +éķ ¦ +éķ § +éķ © +éķ ª +éķ « +ç½ ħ +ç° Į +ç¯ ¾ +ç¯ ¼ +ç° ĸ +ç° ĭ +é¼ ¢ +åĦ ¡ +é¹ ª +é¼ ¾ +çļ ¤ +éŃ į +é¾ ł +ç¹ ĩ +è² ĺ +éĤ Ī +è² Ķ +èĩ Į +èĨ » +èĩ Ĩ +èĩ ĥ +é² ¼ +é² ½ +é³ Ģ +é³ ĥ +é³ ħ +é³ ĩ +é³ Ĭ +èŀ ½ +çĩ ® +é¹ « +ç³ ľ +ç¸ » +çĻ į +éº ĭ +æĩ ij +æ¿ ¡ +æ¿ ® +æ¿ ŀ +æ¿ ł +æ¿ ¯ +è¹ ĩ +è¬ ĩ +éĤ ĥ +è¥ ģ +æª Ĺ +æ ĵĺ +åŃ º +éļ ³ +å¬ · +èŁ Ĭ +é¹ ¬ +éį ª +éı Ĭ +é¬ Ī +é¬ ĥ +çŀ ½ +éŀ ¯ +éŀ ¨ +éŀ « +éŀ § +éŀ £ +èĹ ľ +èĹ ł +éĨ ª +è¹ Ļ +ç¤ ĵ +çĩ ¹ +é¤ ® +çŀ ¿ +æĽ Ľ +é¢ ¢ +èº ĩ +è¹ ļ +èŁ Ľ +èŁ ª +èŁ ł +èŁ ® +é¹ ® +é» ł +é» Ł +é« ħ +é« Ĥ +éķ ¬ +éķ Ń +éķ ¯ +é¦ ¥ +ç° Ł +ç° ª +é¼ ¬ +éĽ ł +èī Ł +é³ İ +é³ ı +é³ IJ +çĻ ŀ +çĻ Ķ +ç³ ¨ +è¹ © +éİ ı +éĤ ĭ +é¬ ı +æĶ ī +éŀ ² +éŀ ´ +èĹ ¿ +èĺ § +èĺ ħ +éĨ ® +éĨ ¯ +éħ ĥ +éľ ª +éľ Ń +éľ ¨ +é» ¼ +åļ ¯ +è¹ ° +è¹ ¶ +è¹ ½ +è¹ ¼ +è¹ ´ +è¹ ¾ +è¹ ¿ +èł ĸ +èł ĵ +èŁ ¾ +èł Ĭ +é» ¢ +é« ĭ +é« Į +éķ ² +ç± Ģ +é½ ģ +éŃ ij +èī ¨ +é³ ĵ +é³ Ķ +é³ ķ +é³ Ĺ +é³ Ļ +éı ĸ +ç¾ ¸ +㸠Ĩ +çĢ £ +çĢ Ľ +è¥ ¦ +è° ¶ +è¥ ŀ +éª ¥ +ç¼ µ +çĵ Ĵ +æĶ ĺ +èĺ © +èĺ ĸ +éĨ ´ +éľ ° +éħ Ĩ +çŁ į +èº ħ +é¼ į +å· ī +é» © +é» ¥ +é» ª +éķ ³ +éķ ´ +é» § +çº Ĥ +çĴ º +é¼ ¯ +èĩ ľ +é³ ľ +é³ Ŀ +é³ Ł +çį ¾ +åŃ Ģ +éª § +ç ĵĺ +é¼ Ļ +éĨ º +ç¤ ´ +é¢ ¦ +æĽ © +é³ ¢ +éº Ŀ +å¤ Ķ +çĪ Ŀ +çģ ı +ç¦ ³ +éIJ ¾ +ç¾ ¼ +èł ¡ +èĢ ± +é¹ ³ +æ° į +é¥ ķ +èº IJ +é« ij +éķ µ +ç© ° +é¥ Ķ +é¬ » +é¬ Ł +è¶ ± +æĶ « +æĶ ¥ +é¢ § +èº ľ +é¼ ¹ +çĻ ¯ +èł ² +èł ¹ +èº ŀ +è¡ ¢ +çģ ŀ +è¥ » +çº Ľ +é¬ £ +æĶ ® +åĽ Ķ +é¦ ķ +æĪ Ĩ +çĪ ¨ +é½ ī +äº į +å° ¢ +å½ ³ +åį ¬ +æ® ³ +ðł ϶ +æ¯ Į +éĤ ĺ +æĪ ĭ +åľ ¢ +æ° ķ +ä¼ ĭ +ä» Ŀ +åĨ ® +æ° ¿ +æ± Ī +æ° ¾ +å¿ ī +å® Ħ +ð¬£ Ļ +è® ± +æī ŀ +åľ ² +åľ « +èĬ ı +èĬ ĥ +æľ ³ +æľ ¸ +ð¨ Ļ +ð¨Ļ ¸ +éĤ ¨ +åIJ Ĵ +åIJ ĸ +å± ¼ +å± ¾ +è¾ ¿ +éĴ Ĩ +ä» ³ +ä¼ £ +ä¼ Ī +çĻ ¿ +çĶ ª +éĤ ł +çĬ ´ +åĨ ± +éĤ ¡ +ð¬ĩ ķ +æ± ĭ +ä ľ +äľ £ +è® » +𬣠ŀ +åŃ ĸ +ð¬ĺ ĵ +çº © +çİ Ĵ +çİ ĵ +çİ ĺ +çİ ļ +åĪ ¬ +ð«Ń Ł +åĿ ľ +åĿ ī +æī ½ +ð«Ń ¢ +åĿ ĭ +æī º +ã§ ij +æ¯ IJ +èĬ ° +èĬ £ +èĭ Ĭ +èĭ ī +èĬ ĺ +èĬ ´ +èĬ ł +ð« ĩ +ð«ĩ Ń +èĬ ¤ +æĿ ķ +æĿ Ļ +æĿ Ħ +æĿ § +æĿ © +å° ª +å° ¨ +è½ ª +ð«IJ Ħ +åĿ Ĵ +èĬ Ī +æĹ ´ +æĹ µ +åij Ļ +ã ķ +ãķ ® +å² į +ð« µ +𫵠· +å² ł +å² ľ +åij ĩ +åĨ ı +è§ ĥ +å² Ļ +ä¼ ¾ +ãij ĩ +ä¼ Ń +ä½ ĸ +ä¼ ² +ä½ ģ +é£ ı +çĭ ĥ +éĹ ¶ +æ± § +æ± « +𣲠ĺ +ð£² Ĺ +æ² Ħ +æ² ĺ +ð¬ĩ Ļ +æ± Ń +ã³ ĩ +æ² ĩ +å¿ ® +å¿ ³ +å¿ º +𬣠¡ +ç¥ ĥ +è¯ ĩ +éĤ ² +è¯ İ +è¯ IJ +å± ĥ +ð« ¸ +𫸠© +å² Ĭ +éĺ ½ +ä¢ º +éĺ ¼ +å¦ § +å¦ ĺ +ð¨ ļ +ð¨ļ ķ +çº ® +é© ² +ð«ĺ ľ +çº » +ð¬ĺ ĺ +ð«ĺ Ŀ +çº ¼ +çİ ¤ +çİ ŀ +çİ ± +çİ Ł +éĤ ½ +éĤ ¿ +åĿ ¥ +åĿ ° +åĿ ¬ +åĿ ½ +å¼ Ĩ +èĢ µ +ä¢ ¼ +ð¦ Ń +ð¦Ń ľ +èĮ ĭ +èĭ § +èĭ ¾ +èĭ ł +æŀ ħ +ãŃ İ +æŀ ĺ +æŀ į +çŁ ¼ +çŁ » +åĮ ¼ +𬨠Ĥ +ð¬Ģ © +ð¬Ģ ª +æĹ ¿ +æĺ Ħ +æĺ Ĵ +æĺ Ī +åĴ ī +åĴ ĩ +åĴ į +å² µ +å² ½ +å² ¨ +å² ŀ +å³ Ĥ +ã Ł +ãŁ ĥ +åĽ · +𬬠© +éĴ IJ +éĴ Ķ +éĴ ĸ +çī ¥ +ä½ ´ +åŀ Ī +ä¾ ģ +ä¾ ¹ +ä½ ¸ +ä½ º +éļ ¹ +ãij Ĭ +ä¾ Ĥ +ä½ ½ +ä¾ ĺ +éĥ Ī +èĪ ł +éĥ IJ +éĥ ĥ +æĶ ½ +èĤ Ń +èĤ ¸ +èĤ · +çĭ ī +çĭ Ŀ +é¥ ³ +å¿ ŀ +çĤ Į +çĤ Ĩ +æ³ Ļ +æ² º +æ³ Ĥ +æ³ ľ +æ³ ĥ +æ³ ĩ +æĢ Ĭ +å³ ĥ +ç© ¸ +ç¥ ĭ +ç¥ Ĭ +ð«į £ +𬣠³ +𬠩½ +é¸ ¤ +å¼ ¢ +å¼ ¨ +éĻ ij +𬮠¿ +éĻ İ +ð¬¯ Ģ +åį º +ä¹ ¸ +å¦ Ń +å§ Ī +ð« ° +ð«° Ľ +è¿ ³ +åı ķ +𬳠µ +é© µ +𬳠¶ +ä Į +äĮ ¹ +é© º +ð«ł Ĭ +ç» ĭ +ç» IJ +çł ī +èĢ Ķ +ãĽ ĥ +çİ ¶ +çı ĩ +çı ħ +ð¬į Ľ +çı ĭ +çİ ¹ +çı Į +çİ ¿ +éŁ ¨ +åŀ ļ +åŀ ¯ +åŀ Ļ +åŀ ² +åŁ ı +åŀ į +èĢ ĩ +é¿ į +åŀ İ +åŀ ´ +åŀ Ł +åŀ ŀ +æĮ ĵ +åŀ µ +åŀ ı +æĭ ¶ +èį ĸ +èį ģ +èį Ļ +èį Ľ +èĮ Ī +èĮ ½ +èį Ħ +èĮ º +ð¬ľ ¬ +èį ĵ +èĮ ³ +𦠰 +𦰠¡ +èĮ Ľ +èį Ń +ãŃ ķ +æŁ · +æŁ ĥ +æŁ Ĭ +æŀ ¹ +æł IJ +æŁ ĸ +éĥ ļ +åī ħ +ä´ ĵ +è¿ º +åİ ĸ +çł Ĩ +çł ij +çł Ħ +èĢ ı +å¥ ĵ +ä ¶ +ä¶ ® +è½ µ +è½ · +è½ ¹ +è½ º +æĺ º +𪠾 +𪾠¢ +æĺ ½ +çĽ · +åĴ ¡ +åĴ º +æĺ ³ +æĺ £ +æĺ ¤ +æĺ « +æĺ ¡ +åĴ ¥ +æĺ ª +èĻ · +èĻ ¸ +åĵ ĥ +å³ ĺ +èĢ ij +å³ Ľ +𪨠° +å³ Ĺ +å³ § +å¸ ¡ +éĴ ĺ +ð«ĵ § +éĴ ľ +𬬠® +𬬠± +ð¬¬ Ń +éĴ ª +éĴ ¬ +éĴ Ń +çŁ § +ç§ ¬ +ä¿ « +èĪ ģ +ä¿ ľ +ä¿ Ļ +ä¿ į +åŀ ķ +è¡ İ +èĪ £ +å¼ ĩ +ä¾ ´ +é¸ § +äı ¡ +èĥ ł +ð¦ ϶ +èĥ Ī +èĥ © +èĥ £ +æľ ı +é£ IJ +è¨ Ħ +é¥ » +åº ¤ +çĸ ¢ +çĤ £ +çĤ Ł +ã ¶ +ã¶ ² +æ´ Ń +æ´ ĺ +æ´ ĵ +æ´ ¿ +ã³ ļ +æ³ ļ +æµ Ī +æµ ī +æ´ ¸ +æ´ ij +æ´ ¢ +æ´ Ī +æ´ ļ +æ´ º +æ´ ¨ +æµ IJ +ã³ ĺ +æ´ ´ +æ´ £ +æģ Ķ +å® ¬ +çª Ģ +æī Ĥ +è¢ Ĩ +ç¥ ı +ç¥ IJ +ç¥ ķ +åı ļ +éĻ § +éĻ ŀ +å¨ Ģ +å§ ŀ +å§ ± +å§ ¤ +å§ ¶ +å§ ½ +æŀ ² +ç» ĸ +éª ĥ +ð¬ĺ ¡ +𬳠½ +ð¬ĺ © +ð«Ħ § +å½ ĸ +éª ī +æģ Ŀ +çı ª +çı Ľ +çı ¹ +çIJ Ĭ +çİ ¼ +çı ĸ +ðª Ł +ðªŁ Ŀ +çı ½ +çı ¦ +çı « +çı Ĵ +ð¬į ¤ +çı ¢ +çı ķ +çı Ŀ +ð«Ń ¼ +åŁ Ĺ +åŀ ¾ +åŀ º +åŁ Ĩ +åŀ ¿ +åŁ Į +åŁ ĩ +èİ ° +èĮ Ŀ +ð¬ľ ¯ +éĦ Ģ +èİ ¶ +èİ Ŀ +äĵ ĸ +èİ Ļ +æł » +æ¡ ł +ð¬ Ĥ +ð¬Ĥ © +æ¡ Ħ +æ¢ ł +æł ´ +æ¢ ´ +æł Ĵ +éħ İ +éħ ı +ð«ł Ĩ +çł µ +çł ł +çł « +çł ¬ +ç¡ ģ +æģ § +ç¿ ĥ +éĥ ª +ð¨ IJ +ð¨IJ Ī +è¾ Ģ +è¾ ģ +ð¬ Į +ð¬Į Ĺ +åī ķ +èµ Ģ +åĵ ¢ +æĻ ħ +æĻ Ĭ +åĶ Ŀ +åĵ ³ +åĵ ± +åĨ Ķ +æĻ Ķ +æĻ IJ +çķ ĸ +èļ Ħ +èļ Ĩ +ð« ij +ð«ij ¡ +å¸ ± +å´ ģ +å³ ¿ +𪨠¶ +å´ Ħ +å¸ ¨ +å ´Ģ +èµ Ĩ +𬠬¸ +éĴ · +𬬠» +𬬠¹ +𬬠¿ +ð¬Ń ģ +çľ ļ +çĶ ¡ +ç¬ « +åĢ » +åĢ ´ +èĦ © +åĢ ® +åĢ ķ +åĢ ŀ +ð« ¢ +ð«¢ ¸ +åĢ ĵ +åĢ § +è¡ ĥ +èĻ Ĵ +èĪ Ń +èĪ ¯ +èĪ ¥ +çĵ ŀ +é¬ ¯ +é¸ ° +èĦ İ +æľ ĵ +èĥ ² +èĻ ĵ +é± ½ +çĭ ´ +å³ ± +çĭ » +çľ ¢ +ð«Ĺ § +åĭ į +çĹ Ħ +çĸ ° +çĹ ĥ +ç« ĺ +ç¾ ĸ +ç¾ ĵ +æ¡ Ĭ +æķ ī +çĥ ł +çĥ Ķ +çĥ ¶ +çĥ » +ð¬Ĭ Ī +æ¶ į +æµ ¡ +æµ Ń +æµ ¬ +æ¶ Ħ +æ¶ ¢ +æ¶ IJ +æµ ° +æµ Ł +æµ Ľ +æµ ¼ +æµ ² +æ¶ ĺ +æĤ Ī +æĤ ĥ +æĤ ¢ +ð¬Ĵ Ī +å® § +çª ħ +çª Ĭ +çª İ +æī ħ +æī Ĩ +è¢ ª +è¢ Ĺ +è¢ ¯ +ç¥ § +éļ º +åł ² +çĸ į +𨠺 +ð¨º Ļ +éĻ ´ +ç ĥĿ +çł ® +ãĽ ļ +åĵ ¿ +ç¿ Ģ +ç¿ Ĥ +åī Ł +𬳠¿ +ð«Ħ ¨ +ç» ¤ +éª į +ð¬ĺ « +ä Ĥ +äĤ ® +çIJ İ +çı ¸ +çı µ +çIJ Ħ +çIJ Ī +çIJ Ģ +çı º +æİ Ń +åł İ +åł IJ +åŁ ¼ +æİ İ +åŁ « +åł Į +æĻ ¢ +ð« ® +ð«® ĥ +æİ ŀ +åŁ ª +å£ ¸ +ãĻ į +èģ į +èı Ŀ +èIJ ļ +èı ¥ +èİ ¿ +äĵ « +åĭ ļ +äĵ ¬ +èIJ Ĩ +èı Ĥ +èı į +èı ¼ +èIJ £ +äĵ ¨ +èı ī +äĵ Ľ +æ¢ ¼ +æ¢ ½ +æ¡ ² +æ¢ ¾ +æ¡ ¯ +æ¢ £ +æ¢ Į +æ¡ ¹ +æķ Ķ +åİ £ +ç¡ Ķ +é¿ İ +ç¡ Ļ +ç¡ ļ +ç¡ Ĭ +ç¡ į +åĭ Ķ +ä´ ķ +é¾ ģ +éĢ ´ +åĶ ª +åķ « +ç¿ Ī +ã « +ã« ° +æĻ Ļ +çķ ¤ +𬱠ĸ +è¶ ¼ +è· Ĥ +èĽ ĥ +èļ ² +ð¬Ł ½ +èļ º +åķ ´ +äİ ĥ +å´ § +å´ Ł +å´ ŀ +å´ Ĵ +å´ Į +å´ ¡ +éĵ ı +ð«ĵ ¯ +ð«Ł ¹ +éĵ ķ +ð«Ł ¼ +éĵ ĸ +éĵ ĺ +éĵ ļ +éĵ ŀ +éĵ ¥ +éĵ ´ +çī » +çī ¿ +ç¨ Ĩ +ç¬ ± +ç¬ ¯ +åģ ° +åģ ¡ +é¸ º +åģ Ń +åģ ² +åģ ģ +ã ¿ +ã¿ ł +éĦ ħ +åģ ĵ +å¾ Ľ +è¡ Ĵ +èĪ ³ +èĪ ² +é¸ ¼ +æĤ Ĩ +éĦ ĥ +çĵ » +ä Ŀ +äĿ Ļ +èĦ ¶ +èĦ ŀ +èĦ Ł +äı ² +é± ¾ +çĮ ĩ +çĮ Ĭ +çĮ Ħ +è§ ĸ +ðł ħ +ðłħ ¤ +åº ± +åº ¼ +åº ³ +çĹ ĵ +ä´ Ķ +ç« « +åł ĥ +éĺ Į +ç¾ Ŀ +ç¾ ķ +çĦ Ĩ +çĥ º +çĦ Į +æ· ı +ð¬ĩ ¹ +æ· Ł +æ· ľ +æ· ´ +æ· ¯ +æ¹ ´ +æ¶ ´ +ð¬į ¡ +ã ¥ +㥠Ħ +æĥ Ľ +æĥ Ķ +æĤ ° +æĥ Ļ +å¯ ģ +éĢ Ń +𬤠ĩ +ð«į ¯ +è¢ ¼ +è£ Ī +ç¥ ² +𬤠Ĭ +ð«į ² +è° ŀ +èī ´ +å¼ ¸ +å¼ ¶ +ð¬¯ İ +éļ ĥ +å© ŀ +å¨ µ +å© ¼ +åª ĸ +å© ³ +å© į +å© Į +å© « +å© ¤ +å© ĺ +å© ł +ð¬ĺ ¬ +ð¬ĺ Ń +𬴠Ĥ +ð«ĺ ¦ +ç» ¹ +ð«Ł ħ +ð¬ĺ ¯ +éª ķ +ð«ĺ § +çµ ľ +çı · +çIJ ² +çIJ ¡ +çIJ Ł +çIJ Ķ +çIJ Ń +åł ¾ +åł ¼ +æı ķ +ãĻ ĺ +åł § +åĸ Ĩ +åł ¨ +å¡ ħ +åł ł +çµ · +𪠣 +𪣠» +ð¡ İ +ð¡İ ļ +è ijľ +æĥ İ +èIJ ³ +èij Ļ +éĿ ¬ +èij ´ +èĴ ĩ +èĴ Ī +éĦ ļ +èĴ ī +èĵ ĩ +èIJ © +èij ° +èij İ +éĦ ij +èĴ İ +èij ĸ +èĴ Ħ +èIJ ¹ +æ£ ¤ +æ£ ½ +æ£ « +æ¤ ĵ +æ¤ ij +ð¬ ĥ +ð¬ĥ Ĭ +é¹ Ģ +æ¤ Ĩ +æ£ ĵ +æ£ ¬ +æ£ ª +æ¤ Ģ +æ¥ Ĺ +𬠷 +𬷠ķ +çĶ ¦ +éħ ¦ +è§ Į +å¥ ¡ +çļ ķ +ç¡ ª +æ¬ ¹ +è© Ł +ð«IJ IJ +è¾ Į +æ£ IJ +é¾ Ĥ +𬠹 +𬹠¼ +é» ¹ +çī ļ +çĿ İ +æĻ « +æĻ ª +æĻ ± +ð § +ð§ ¿ +ð§¿ ¹ +èĽ ij +çķ ¯ +æĸ Ŀ +åĸ ¤ +å´ ¶ +åµ ģ +ð« ¶ +ð«¶ ĩ +å´ ¾ +åµ ħ +å´ ¿ +åµ ļ +ç¿ Ļ +ð«ĸ ® +åľ Į +åľ IJ +èµ ij +èµ Ĵ +é¿ ı +éĵ ¹ +ð¬Ń Ĭ +éĵ ½ +𨱠ĩ +ð«ĵ ¶ +éĶ Ĭ +éĶ į +éĶ İ +ð¬Ń İ +éĶ ĵ +çĬ ĩ +é¢ ĭ +ç¨ Į +çŃ Ģ +çŃ ĺ +çŃ ľ +çŃ ¥ +çŃ ħ +åĤ ĥ +åĤ ī +ç¿ Ľ +åĤ Ĵ +åĤ ķ +èĪ ¾ +çķ ¬ +ð«ĸ ¯ +èĦ ¿ +èħ ĺ +ä IJ +äIJ ĥ +èħ Ļ +èħ Ĵ +ð¬± Ł +é² ĥ +çĮ ° +ð« Ľ +ð«Ľ Ń +çĮ ¯ +ã º +㺠Ħ +é¦ ī +åĩ ĵ +éĦ Ĺ +ð« · +ð«· · +å» ĭ +å» Ĩ +éĦ Į +ç² ¢ +éģ Ĩ +æĹ IJ +𬮠± +çĦ ŀ +ð¬Ĭ ¤ +æ¬ » +𣠸 +𣸠£ +æº ļ +æº ģ +æ¹ Ŀ +æ¸ ° +æ¹ ĵ +ã ´ +ã´ Ķ +æ¸ Ł +æº ł +æ¸ ¼ +æº ĩ +æ¹ £ +æ¹ ij +æº ŀ +æĦ IJ +æĦ ĥ +æķ © +çĶ ¯ +æ£ ¨ +æī Ĭ +è£ £ +ç¥ ¼ +å© » +åª Ĩ +åª ŀ +ãĽ ¹ +åª ĵ +åª Ĥ +åª Ħ +æ¯ µ +çŁ ŀ +𬴠ĥ +ð«ĺ ¨ +ç¼ Ĭ +ç¼ IJ +éª Ļ +çij ĥ +çij ĵ +çij ħ +çij Ĩ +ä´ ĸ +çij ĸ +çij Ŀ +çij Ķ +çij Ģ +𤠧 +𤧠Ľ +çij ³ +çij Ĥ +å¶ ħ +çij ij +éģ ĺ +é« ¢ +å¡ ¥ +åł ½ +èµ ª +æij Ľ +å¡ Ŀ +æIJ Ĵ +æIJ Į +èĴ ± +èĴ ¨ +èĵ ı +èĶ Ģ +èĵ ¢ +èĵ Ĥ +èĴ » +èĵ £ +æ¤ ¹ +æ¥ ª +æ¦ ĥ +æ¦ ħ +æ¥ Ĵ +æ¥ © +æ¦ ĩ +æ¤ ¸ +æ¥ Ļ +æŃ ħ +𬠪 +𬪠© +ç¢ ĥ +ç¢ ı +ð¬Ĵ Ķ +ç¢ Ī +äĥ ħ +ç¡ ¿ +éĦ ł +è¾ Ĵ +ð¬¨ İ +ð«IJ ĵ +é¾ Ĩ +è§ ľ +ä £ +ä£ ĺ +æļ ķ +é¹ į +ð« « +ð«« ĩ +㬠Ĭ +æļ ħ +è· ± +èľ IJ +èľ İ +åµ ² +èµ Ĺ +éª ± +éĶ ĸ +ð«ĵ ¹ +éĶ ĺ +éĶ ³ +éĶ § +éĶ ª +ð¬Ń ļ +éĶ « +éĶ ¬ +ð¬Ń Ľ +ç¨ ij +ç¨ Ļ +ä ħ +äħ Ł +ð¬ ķ +ð¬ķ Ĥ +çŃ » +çŃ ¼ +çŃ ¶ +çŃ ¦ +çŃ ¤ +åĤ º +é¹ İ +åĥ ĩ +èī ħ +èī ī +è° ¼ +è² Ĩ +èħ ½ +èħ ¨ +èħ ¯ +é² ī +é² Ĭ +é² Į +ä² Ł +𬶠ĭ +𬶠į +é² ı +éĽ Ĭ +çĮ º +é£ Ķ +è§ Ł +ð¦ Ŀ¼ +é¦ Į +è£ Ľ +å» Ĵ +çĺ ħ +éĦ ĺ +é¹ Ĵ +éĦ ľ +éº Ģ +éĦ £ +éĺ ĺ +ð«Ķ ¶ +çħ ģ +çħ ĥ +çħ ´ +çħ ĭ +çħ Ł +çħ ĵ +æ» ł +æº į +æº ¹ +æ» Ĩ +æ» ī +æº ¦ +æº µ +æ¼ · +æ» § +æ» ĺ +æ» į +æĦ Ń +æħ ¥ +æħ Ĩ +å¡ ± +ð« ĮĢ +è £¼ +ç¦ ĭ +ç¦ Ķ +ç¦ ĺ +ç¦ Ĵ +è° « +é¹ Ķ +ð«ĸ ³ +æĦ į +å« Ħ +åª ± +æĪ ¤ +åĭ ł +æĪ £ +ð«ĺ ª +ð«ĺ ¬ +ç¼ ŀ +èĢ ¤ +çij § +ð« ŀ +ð«ŀ © +çij ¨ +çij ± +çij · +çij ¢ +æĸ ł +æij ı +å¢ ķ +å¢ Ī +å¢ IJ +å¢ ĺ +æij ´ +éĬ İ +ð¡ IJ +ð¡IJ ĵ +å¢ ļ +æĴ ĸ +𪠤 +ðª¤ Ĺ +éĿ ½ +éŀ ģ +èĶ Į +èĶ Ī +èĵ ° +èĶ ¹ +èĶ Ĭ +åĺ ı +æ¦ ° +æ¦ ij +æ§ ļ +ð£ Ĺ +ð£Ĺ ĭ +æ§ ľ +æ¦ į +çĸ IJ +𬸠ĺ +éħ º +éħ ¾ +éħ ² +éħ ´ +ç¢ ¶ +äĥ İ +ð¬Ĵ Ĺ +ç¢ ¨ +ð¥ Ķ +ð¥Ķ ² +ç¢ ¹ +ç¢ ¥ +åĬ Ĥ +ð«ļ ĸ +ä´ Ĺ +å¤ ¥ +çŀ į +é¹ ĸ +ã¬ İ +è· ½ +èľ ¾ +å¹ ĸ +å¶ į +åľ Ļ +𨱠ı +éĶ º +éĶ ¼ +éĶ ½ +ð¬Ń ¤ +éĶ ¾ +éĶ ¿ +éķ ĥ +éķ Ħ +éķ ħ +é¦ Ŀ +é¹ Ļ +ç® ¨ +ç® ĸ +åĬ Ħ +åĥ ¬ +åĥ ¦ +åĥ Ķ +åĥ İ +æ§ ĥ +ãĻ ¦ +é² Ĵ +é² ķ +ð«ļ ķ +é² ĸ +é² Ĺ +é² ĺ +é² Ļ +𬶠IJ +𬶠ı +ð ©½ +𩽠¾ +å¤ IJ +çį į +é£ Ĺ +𬸠ļ +åĩ ĺ +å» ij +å» Ļ +çĺ Ĺ +çĺ ¥ +çĺ ķ +é² Ŀ +éĦ « +çĨ ĩ +æ¼ ¹ +æ¼ ĸ +æ½ Ĩ +æ¼ ¤ +æ½ © +æ¼ ¼ +æ¼ ´ +ã ½ +ã½ ı +æ¼ Ī +æ¼ ĭ +æ¼ » +æħ ¬ +çª ¬ +çª Ń +ã ® +ã® ¾ +𬤠Ŀ +è¤ ķ +ç¦ Ľ +ç¦ ļ +éļ © +å« ķ +å« Ń +å« ľ +å« ª +ð¬ ĻĤ +ã » +ã» ¬ +éº ¹ +çĴ Ĩ +æ¼ ¦ +åı ĩ +å¢ £ +å¢ ¦ +å¢ ¡ +åĬ IJ +èĸ ģ +èķ ° +èĶ ĥ +é¼ Ĵ +æ§ ± +é¹ Ŀ +ç£ ı +ç£ ī +æ® £ +æħ Ń +éľ ħ +æļ µ +æļ ² +æļ ¶ +è¸ ¦ +è¸ £ +äĹ ĸ +èĿ ĺ +èĿ ² +èĿ ¤ +åĻ ĩ +å ĻĤ +åĻ Ģ +ç½ ¶ +å¶ ² +å¶ ĵ +ãł ĩ +å¶ Ł +å¶ Ĵ +éķ Ĩ +éķ Ī +éķ ĭ +éķ İ +ð¬Ń © +éķ ķ +ç¨ ¹ +åĦ ĩ +çļ ŀ +çļ Ľ +ä´ ĺ +èī İ +èī ı +é¹ Ł +𩾠ĥ +é² ¦ +é² ª +é² ¬ +æ© ¥ +è§ Ń +é¹ ł +é¹ ¡ +ç³ ĩ +ç³ Ī +ç¿ ¦ +é¹ ¢ +é¹ £ +çĨ Ľ +æ½ ĸ +æ½ µ +ã µ +ãµ IJ +æ¾ Ĥ +æ¾ Ľ +çij ¬ +æ½ ½ +æ½ ¾ +æ½ ı +æĨ Ń +æĨ ķ +𬸠£ +æĪ Ń +è¤ ¯ +ç¦ ¤ +ð«į ½ +å« ½ +éģ ¹ +𬴠Ĭ +çĴ ¥ +çĴ ² +çĴ Ĵ +æĨ Ļ +æĵ IJ +éĦ ¹ +èĸ ³ +éŀ Ķ +é» ĩ +ð¬ ŀ +ð¬ŀ Ł +èķ Ĺ +èĸ ¢ +èķ ¹ +æ© ŀ +æ© ij +æ© ¦ +éĨ ij +è§ ± +ç£ ¡ +ð¥ ķ +ð¥ķ ¢ +ç£ ľ +è± ® +ð«Ł ¦ +ð¬º Ī +ð«ł ľ +é¹ ¾ +èĻ ¤ +æļ ¿ +æĽ Į +æĽ Ī +㬠ļ +è¹ ħ +è¸ ¶ +äĹ Ľ +èŀ Ĺ +çĸ ģ +ãł ĵ +å¹ ª +𪠩 +𪩠ĺ +å¶ ¦ +ð¬Ń ¬ +𨱠ij +ð¬Ń ¯ +é¦ ŀ +ç© Ħ +ç¯ ļ +ç¯ ¯ +ç° ī +é¼ ½ +è¡ ł +çĽ ¦ +èŀ £ +ç¸ ¢ +é² Ń +é² ¯ +é² ° +é² º +é² ¹ +ð«Ĺ ´ +äº ¸ +çĻ Ģ +çĺ Ń +𬸠¦ +ç¾ ± +ç³ Ĵ +çĩ ĭ +çĨ » +çĩ Ĭ +çĩ ļ +çĩ ı +æ¿ © +æ¿ ĭ +æ¾ ª +æ¾ ½ +æ¾ ´ +æ¾ Ń +æ¾ ¼ +æĨ · +æĨ º +æĩ Ķ +é» ī +å¬ Ľ +é¹ ¨ +ç¿ ¯ +ð«Ħ · +çĴ ± +𤠩½ +çĴ ¬ +çĴ ® +é« ½ +æĵ ¿ +èĸ ¿ +èĸ ¸ +æª ij +æ« Ĩ +æª ŀ +éĨ ¨ +ç ¹Ħ +ç£ ¹ +ç£ » +çŀ « +çŀ µ +è¹ IJ +èŁ ı +ã ĺ +ãĺ İ +ð¬Ń ³ +éķ ¤ +ð¬Ń ¶ +ð«Ķ į +éķ ¥ +éķ ¨ +ð¬Ń ¸ +ð¨± Ķ +ð¬Ń ¼ +ð«Ķ İ +çŁ ° +ç© Ļ +ç© ľ +ç© Ł +ç° ķ +ç° ĥ +ç° ı +åĦ ¦ +éŃ ĭ +æĸ ¶ +èī ļ +𬸠ª +è° ¿ +ä² ł +ð¬¶ Ł +é² ¾ +𬶠ł +é² ¿ +é³ ģ +é³ Ĥ +é³ Ī +é³ ī +çį ¯ +äĹ ª +é¦ ĺ +è¥ ķ +è¥ ļ +𬶠¨ +èŀ ± +çĶ ĵ +å¬ ¬ +å¬ ¥ +ð¦ Ī +ð¦Ī ¡ +ð«Ħ ¸ +çĵ Ģ +éĩ IJ +é¬ ¶ +çĪ ĩ +éŀ ³ +éŀ ® +ð¬Ł ģ +èĹ Ł +èĹ ¦ +èĹ ¨ +é¹ ² +æª « +é» ¡ +ç¤ ŀ +ç¤ Į +ð¥ ĸ +ð¥ĸ ¨ +è¹ ¢ +è¹ ľ +èŁ « +äĹ ´ +åļ ļ +é« ĥ +éķ ® +éķ ± +éħ Ĥ +é¦ § +ç° ł +ç° Ŀ +ç° ° +é¼ « +é¼ © +çļ ¦ +èĩ ij +ä² ¢ +é³ ij +é³ Ĵ +é¹ ± +é¹ ¯ +çĻ Ĺ +ð¦ Ĵ +ð¦Ĵ į +æĹ ŀ +ç¿ · +åĨ ģ +äİ ĸ +çĢ Ķ +çĢ į +çĢ Į +è¥ ľ +ä´ Ļ +ð¬Ļ Ĭ +åļ Ń +ã ° +ã° Ģ +é¬ · +éĨ Ń +è¹ ¯ +èł ĭ +ç¿ ¾ +é³ ĺ +åĦ ³ +åĦ ´ +é¼ Ĺ +ð¬¶ Ń +𩾠Į +é³ ļ +é³ Ľ +éº ij +éº ĸ +èł ĥ +å½ Ł +å¬ ¿ +é¬ Ĵ +èĺ ĺ +æ¬ Ĥ +é Ĩµ +é¢ ¥ +çĶ Ĺ +ð¨ Ł +ð¨Ł ł +å· ĩ +éħ ħ +é« İ +çĬ ¨ +𬶠® +ð¨ Ń +ð¨Ń ī +㸠Į +çĪ Ķ +çĢ ± +çĢ ¹ +çĢ ¼ +çĢ µ +è¥ « +åŃ ħ +éª ¦ +ð¬Ļ ĭ +èĢ ° +𤠫 +𤫠ī +çĵ ĸ +é¬ ĺ +è¶ ¯ +𬺠ĵ +ç½ į +é¼ ± +é³ ł +é³ ¡ +é³ £ +çĪ Ł +çĪ ļ +çģ Ī +éŁ Ĥ +ç³ µ +èĺ ¼ +ç¤ µ +é¹ ´ +èº Ķ +çļ Ń +é¾ ¢ +é³ ¤ +äº ¹ +ç± ¥ +é¼ · +ð«ļ Ń +çİ ĥ +éĨ ¾ +é½ ĩ +è§ ¿ +èł ¼ +× § +× ¤ +× Ľ +×ķ× ª +× ¡ +×Ļ× Ŀ +× ¦ +× Ĵ +× ĺ +×ķ× ¨ +× Ŀ +×ķ× ľ +× ĸ +๠Ĥ +ï º +ðŁ į +ðŁ IJ +×Ļ× ¨ +ï » +ðŁ ij +ðĿ IJ +ðŁ ı +ðŁ Ķ +ðŁ Į +ðŁ İ +ðŁ ĵ +× Ł +ðĿ ij +×ķ× ĵ +ï ¦ +Ġ× ķ +×ķ× ij +à¸Ń à¸ĩ +ðĿ ĺ +×Ļ× ª +ðĿ ķ +à¸Ĺ ีà¹Ī +Ø§Ø ¦ +ðŁ ¤ +×ķ× Ł +ر ÙĬ +×Ļ× ľ +ร ะ +า ย +ï ¯ +ï ® +า ม +â ĩ +ðŁ ¥ +ï Ń +ðĿ Ļ +×ķ× ł +á ½ +Ġ× Ľ +ðŁ ļ +â ļ +ï § +×ij ר +×Ļ× ł +á ´ +Ġ× Ĺ +á ¼ +ðĿ Ĺ +Ġ× ¢ +×Ļ× Ķ +ãģ£ ãģŁ +ãģĵ ãģ¨ +á ¸ +ÙĬ ÙĨ +ãģª ãģĦ +ا ع +ภ¨ +à¹Ī à¸ĩ +×Ļ× ĵ +×ŀ ש +á Ī +׳ ×Ļ +×Ļ× ij +ï ¥ +ðĿ ĵ +Ġ× Ļ +× ļ +ั à¸ĩ +â ĵ +ï ¤ +ĠاÙĦ Ø£ +า à¸ģ +à¹ī à¸Ļ +à¹Ģ ร +×ķ× Ŀ +á ¹ +ภ¶ +×Ļ× § +ภĭ +à¸Ħ ร +ภĺ +ั à¸ģ +ðŁ ķ +ÙĪ ÙĨ +à¸Ń ย +â Ĭ +ðĿ Ĵ +ĠاÙĦ ع +า à¸Ļ +×Ļ× Ł +ÙĦ ÙĬ +×Ļ× © +à¸Ľ ระ +à¹Ģ à¸Ľ +Ġ× ł +×ķ× ¡ +ภł +Ùħ ÙĨ +×ķ× ¢ +×ķ× ŀ +â Į +ðŁ § +à¹ĩ à¸Ļ +ภį +ã İ +á µ +ĠاÙĦ س +×ķ× § +ห ล +ðŁ ĩ +â ı +ðŁ ¦ +Ġ×Ķ ×ŀ +ÙĪ Ø§ +Ġ× ª +ר ×IJ +à¸Ń à¸Ļ +ภ© +à¹Ī ว +×ķ× ¦ +í Ĺ +ã Ħ +ï ¨ +ï ¹ +â İ +ï ² +ðĿ ļ +ð IJ +à¸Ħ ว +ห à¸Ļ +Ġ× ¨ +ب ÙĬ +ร à¹Į +ر ا +Ø´ ر +×ķ× Ĺ +×ķ× ¤ +×ķ× © +×ķ× Ĵ +í Ŀ +â Ľ +à¸ķ ิ +à¹Ģ à¸ģ +ï ³ +ï ± +à¸Ķ à¹ī +ë ¹ +ï ¬ +á ¿ +ðŁ Ľ +ðĿ ĸ +à¹Īา à¸ĩ +ู à¹ī +Ġ×Ķ ×IJ +ĠاÙĦ ØŃ +פ ר +ÙĪ Ùħ +à¹Ģ ล +í ĸ +×Ļ× ¢ +ì Ī +í ĵ +ðŁ ħ +á ł +à¸Ħว าม +à¸Ī ะ +׳ ×Ķ +Ġ× § +à¸ Ł +à¹ī à¸ĩ +ห ม +ت Ùħ +׾ ×Ļ +ÙĬ د +à¹Ī à¸Ļ +׊ר +ש ר +à¹Ģ à¸Ĺ +×ŀ ר +ë ĸ +ع ÙĦ +×ŀ ×¢ +â ² +׾ ×Ķ +Ġ× ¤ +à¸Ń à¸ģ +س ÙĦ +×Ļ× ŀ +ÙĤ ÙĬ +í İ +ت ØŃ +×Ļ× ¡ +×Ļ× Ĺ +í Ľ +ï ° +â ½ +á ī +á Ĭ +á ¨ +Ùĩ ا +Ġ׾ ×Ķ +×ķ× IJ +Ùħ ا +à¹īà¸Ń à¸ĩ +ر ب +ĠاÙĦ ج +×ŀ ×ĵ +Ùħ ÙĦ +ت ر +à¹Ģ à¸Ķ +×§ ר +í ħ +ì ¼ +ê ¿ +ã Ī +á IJ +ðŁ Ĺ +ê ¦ +á ĭ +ðĿ Ķ +à¹Ģà¸Ľ à¹ĩà¸Ļ +à¹ĥ ห +ม า +ว à¹Īา +ม ี +ี à¹ī +à¹Ħม à¹Ī +ÙĨ ÙĬ +Ø ¤ +ร า +×ķ ×Ļ +ãĤĪ ãģĨ +ิ à¸Ķ +×Ļ× ¤ +׊׾ +ÙĤ د +à¹Ģ ส +×Ļ× ĺ +à¸ģ ล +ר ׼ +×ķ× Ľ +×Ļ× Ľ +ë Ī +ë ĥ +ðŁ ĸ +á ħ +â ¼ +ã ī +à¹Ħ à¸Ķà¹ī +ת ×Ļ +×Ļ× IJ +ĠاÙĦ Ø¥ +à¸ł า +ร ิ +ÙĤ Ø© +ØŃ د +ê » +ì ± +ת ×Ĺ +ì º +â ĭ +á Ħ +á ¾ +â µ +â ¾ +ĠÙĪ Ø§ÙĦ +׳ ×ķ +Ù Ģ +ÙĬ ا +à¸ģ à¹ĩ +×ŀ ×Ķ +ãģĦ ãĤĭ +ع د +ĠاÙĦ ÙĨ +Ġ×Ķ ×© +Ø ¦ +ั à¹īà¸ĩ +ร ัà¸ļ +ÙĪ ÙĤ +ãģ§ ãģį +à¹Ģ à¸ŀ +׼ ׾ +×ĺ ר +ั à¸Ķ +à¸Ń า +ì ¢ +à¸Ń à¸ļ +à¸ķ ร +à¹Ģ à¸Ĭ +ì Ķ +ãģĹ ãģ¾ +ë ģ +ë ķ +ðŁ Ļ +â Ĵ +á ¶ +à¹ģ ล +ÙĨ ا +à¹ĥห à¹ī +à¹Ħ à¸Ľ +× £ +ั ว +า à¸ĩ +×ĵ ר +×ij ׾ +פ ×Ļ +Ġ× ĵ +ĠاÙĦ Ùģ +à¹Ģ à¸Ĥ +ש ×Ķ +×IJ ר +ë ¬ +ãģ« ãģª +ÑĢ Ð¾ +ว ิ +Ùħ ر +×IJ ת +Ùĥ ر +س ب +ÙĨ ت +ãģĹ ãģĦ +ا ج +à¸Ń รà¹Į +Ùĥ ÙĦ +س Ùħ +ส ิ +×Ļ× ¦ +ë Ŀ +í ľ +ì ī +á Ĩ +Ùĩ Ùħ +à¸Ļ ีà¹ī +ãģĤ ãĤĭ +ãģĦ ãģ¦ +س ÙĬ +׾ ×IJ +د ر +ãģ ļ +ÙĪ Ø¬ +ĠاÙĦ Ø® +ص ر +í ı +à¹īา à¸ĩ +ุ à¸Ķ +×ķ× ĺ +×ij ×¢ +í Ĩ +à¸Ĭ า +ร ม +ש ×ŀ +×ŀ ס +ê ´ +ì ´ +ë ľ +ì ¿ +ì © +ë » +â ¤ +ðŁ Ĩ +á Į +á ķ +ذ ا +à¸Ĺ ำ +à¸ķ à¹Ī +ĠاÙĦ ÙĤ +ÙĦ Ùĥ +ู à¹Ī +à¸Ħ ุ +ÙĬ Ùħ +׳ ×Ļ×Ŀ +ืà¹Ī à¸Ń +ÙĪ Ø¹ +ãĤ ĩ +ا ÙĤ +Ġ×ij ×¢ +à¹Ģ ม +ج Ùħ +á» « +ãģĵãģ¨ ãģĮ +ب د +×ķ× Ķ +ש ׾ +Ùĩ ر +à¹Ģ à¸Ļ +ãģ ¹ +í ĭ +ì » +ì ½ +ë Ń +ì Į +í Ģ +ë Į +ë º +ã Ĭ +à¹ĥ à¸Ļ +Ġ× Ĵ +๠Ĩ +à¸Ī าà¸ģ +ว ย +à¹ĥ à¸Ĭ +à¸ĩ าà¸Ļ +ĠاÙĦ Ø´ +ا ØŃ +à¹īา à¸Ļ +ืà¹Ī à¸Ńà¸ĩ +×IJ ×Ļ +ب ÙĦ +ãģ¨ æĢĿ +׳ ס +ãģ¾ ãģĽ +Ùĥ ÙĨ +×¢ ר +ĠاÙĦ د +ש ת +í ŀ +Ùħ س +ص ÙĦ +×ķ׳ ×Ķ +ار Ø© +ÙĦ Ùħ +ส ม +Ø£ ÙĨ +ת ר +×IJ ×ŀ +ع ب +Ø® ت +ãĤ ĥ +ì ¡ +ì £ +ив а +ส ั +ึ à¸ģ +ì ¸ +ë Ĩ +алÑĮ н +ì ³ +ì į +ê ¼ +ê ½ +ì ı +ã Į +ã ı +ï © +ê ª +á İ +Ġ× ĸ +à¸ģ ัà¸Ļ +×Ļ ×ķ +à¸Ħ à¸Ļ +׳ ×ķת +à¸ľ ูà¹ī +à¹ĥ à¸Ī +ãģĦ ãģŁ +Ùģ Ø± +×ĺ ×Ļ +צ ×Ļ +ãĤĤ ãģ® +ĠاÙĦ ص +ãģ¾ãģĽ ãĤĵ +د Ø© +×ij ×Ļ +ĠاÙĦ ر +Ġ×ŀ ×IJ +ส ำ +à¹Ģ ห +ع ر +ãģª ãģı +à¸ģร ะ +×ij ×ĵ +à¹Ģ à¸Ī +×Ļ× ļ +×Ĺ ×Ļ +ÙĬ ع +ש ×ij +ÙĨ Ø© +ÙĪ Ø¶ +ÙĦ Ùģ +ÙĢ ÙĢ +פ ×¢ +í Ī +×ŀ ×§ +ภIJ +ØŃ Ø© +ا ص +Ñĭв а +à¸Ħ ม +ว ั +à¸Ľ ล +ì Ł +í ļ +ë ´ +ë ij +ë ī +ë ĩ +ì ¨ +ë ± +ë İ +â ¬ +á ¥ +á Ĺ +á Ľ +á į +Å © +à¸Ķ ี +ô i +Ġ× ¡ +׾ ×ķ +á»Ŀ i +à¸Ħุ à¸ĵ +â y +à¸Ļ า +×Ĺ ×ĵ +×ĵ ×Ļ +ห า +ج ÙĦ +à¹Ģ ว +ãĤĩ ãģĨ +Ùħ Ø© +ĠاÙĦ Ùĥ +Ġ×Ķ ×¢ +ج ر +×ĸ ר +ا Ø· +׼ ת +×ķ׳ ×Ļ×Ŀ +ØŃ Ùħ +ê ¶ +ر Ùĥ +Ġ׾ ×¢ +×ķ× ĸ +ส ร +צ ׾ +Ø ¢ +ا ست +à¹Ī ม +Ø® ر +צ ×¢ +×Ļר ×ķת +اد Ø© +Ø´ ار +×ŀ ×Ĺ +í Ĵ +à¹Ģร ีย +×Ĺ ×§ +Ø§Ø « +ร à¸ĩ +à¹Ģ à¸ķ +à¸Ī ำ +ภĿ +à¹Īา ย +à¸Ħ ล +ÙĤ ÙĪ +иÑĩеÑģ к +à¸ĵ à¹Į +ั ย +Ùħ ع +ë ¨ +ë ¿ +ë ® +ï ´ +ì ¥ +ì « +ë µ +á ¡ +â į +ð ĵ +â ° +à¸Ĥ à¸Ńà¸ĩ +Ù ĭ +à¸ģ ัà¸ļ +ãģ® ãģ§ +à¹ī ว +à¸Ńย à¹Īาà¸ĩ +ãģ Ń +á»ĩ t +à¸ķ à¹īà¸Ńà¸ĩ +×ŀ ×Ļ +à¹ģ à¸ļ +×Ĵ ר +ÙĪ Ùģ +ÙĤ ÙĦ +à¸łà¸² à¸ŀ +ר ×Ļ +ล า +ÙĬ س +Ġ× ¦ +ÙĬ Ùģ +Ġ× ĺ +à¸ľ ล +á ng +ร ว +Ġ×ŀ ש +×IJ ×ķת +×ĸ ×Ķ +ู à¸ģ +à¸Ļ ัà¸ģ +اÙĨ ÙĬ +د ا +ãģ ³ +׼ ף +ãĤī ãĤĮ +ãĤĮ ãģ° +ת ×§ +ú c +ÙĪ Ø² +×Ļר ×Ķ +Ġn gh +án h +Ġ×ķ ×IJ +á» ħ +ส ุà¸Ķ +ë į° +ا ض +اÙĦ ÙĬ +ب ار +ع Ùħ +à¸ļ า +ت ج +à¸ŀ ร +×ķר ×Ķ +ả ng +Ø® ÙĦ +ภī +ắ c +ש ×Ļ×Ŀ +í Ķ +Ùģ Ø³ +×Ļ× Ĵ +п ÑĢ +ĠاÙĦ Ø« +س Ø· +ร ูà¹ī +ีà¹Ī ย +à¸Ń à¸Ķ +ãģª ãĤĬ +×Ĵ ×ĵ +ãģĦ ãģ¾ãģĹãģŁ +ס ×§ +Ø® ص +la ÅŁ +ен но +ب ØŃ +ส à¸Ļ +ภ® +ר×IJ ש +Ùħ ÙĪ +دÙĬ د +ษ า +×ķ× ļ +ãĥ§ ãĥ³ +à¸ķ ุ +Ġê µ +ĠÑģв о +צ ×ij +à¸Ń ม +à¸Ľ ร +ت ع +×Ķ ×ª +اÙħ ÙĦ +×ŀ ׳ +ç ¶ļ +ภ¤ +í į +ë ĺ +ë ¤ +ì ij +â ´ +ã ĭ +Ġب اÙĦ +á»ģ u +ĠاÙĦ ÙĦ +à¸ķ ัว +ذ Ùĩ +ึ à¸ĩ +à¹ĥà¸Ĭ à¹ī +á»ĵ ng +à¸Ļ ั +ม าà¸ģ +ãĥ Ł +×ŀ ×ķ +à¸Ĺ ย +á»Ļ i +Ạ± +ả o +à¹Ĥ à¸Ķ +×IJ ׾ +ส าม +ÙĪ Ø¨ +à¸Ĺ ุ +ย ัà¸ĩ +×¢ ת +×ķ׳ ×ķת +à¸Ĥ ึ +à¸Ĥึ à¹īà¸Ļ +à¸ģ à¹Ī +Ạ« +á»ij c +ãģĹ ãĤĩãģĨ +á»ĭ ch +Ġ×IJ ×ķת +Ġש ×IJ +׼ ×ķ׾ +á»Ļ c +ع Ø© +à¸Ĺ ี +à¹Ģ à¸Ń +Ùĥ ت +ãģ » +Ạ» +ìĹ ħ +à¸Ń à¸Ńà¸ģ +اÙĨ ت +à¹Ħ ร +Ġ×IJ ×Ĺר +Ø· ر +ÙĨ د +ื à¹īà¸Ń +Ø· ÙĦ +×IJ ×Ķ +uy ên +í ĸī +×ij ×Ķ +à¸Ħ à¹Ī +à¸Ĭ à¹Īว +ãģĤãĤĬ ãģ¾ãģĻ +ÙĬ ب +×§ ׾ +ãĥ Ļ +Ä © +س ر +า ว +ãĤ ± +à¸ļ ริ +ר ×Ĵ +á»ĥ u +ØŃ ت +×ķ×ŀ ×Ļ +ب ÙĨ +êµ IJ +ÄŁ u +ãģª ãĤĵ +×ij ×§ +Ġפ ר +ắ n +ØŃ ÙĦ +×ij ×Ĺ +ấ u +×ij ×ķ×ĵ +ãĥ ¯ +Ġ׾ ×§ +ั à¸į +à¸ŀ ิ +×Ĺ ×Ķ +×ĸ ׼ +ãĥ¼ãĥ ł +ÑĤ елÑĮ +×ŀ ×Ļ×ĵ +ÙĬ Ø® +Ạ³ +ت ص +à¸ĺ ิ +è¾ ¼ +ì ĵ +Ùĥ Ø© +ÙĤ ب +à¸Ħ à¹Į +à¹īา ย +à¸ĵ ะ +า ะ +ë Ĵ +ê ¾ +ë · +ì ĩ +ê º +ì ģ +ë Ģ +ì ¾ +ë ½ +ë ļ +ì Ń +ì İ +á ij +ë Ĺ +ê Ĵ +à ¡ +à ¬ +ðIJ Į +ã ĩ +ðĿ Ħ +Ġ׾ ×IJ +ãģ¨ ãģĦãģĨ +Ġn hi +×Ļ ×ķת +Ġש ×Ķ +à¹ģล à¹īว +Æ°á»Ľ c +à¸Ķà¹ī วย +à¸Ĺ าà¸ĩ +׳ ת +פ ת +à¹ģ à¸ķà¹Ī +ư ng +à¸Ńย ูà¹Ī +à¹ī ำ +Ġ×IJ ׾ +Ùĥ Ùħ +ấ p +ล à¸ĩ +ãģŁ ãĤģ +×Ĵ ׾ +ห ร +ĠÑĢ Ðµ +à¹Ģà¸Ĥ à¹īา +ÙĤ ر +Ġ×Ķ ×¡ +ÙĪ ÙĬ +สาม าร +สามาร à¸ĸ +Äĥ n +à¸Ń ี +פ ×ķ +×Ļ׳ ×ķ +ว ัà¸Ļ +ặ c +íķ Ļ +×ŀ ת +ê u +Ạ¹ +Ùģ ÙĬ +×ŀ צ +à¸Ħ า +ãģĿ ãģĨ +ãĢ ħ +ا ز +ا Ùĩ +ר ×Ļ×Ŀ +ấ n +ห าร +ạ t +ÙĨ Ùĩ +à¹Ģ à¸Ħร +ج Ùĩ +׼ ×Ļ +ắ t +à¸Ħ à¹īา +ر Ø© +ãĥ ı +Ùĥ ÙĪÙĨ +ứ ng +Ġìļ ° +ย à¹Į +à¹Īว à¸Ļ +à¸ģ ำ +Ø« ر +Ñģ и +ĠاÙĦ Ø· +Ġ×Ķ ×¦ +ĠØ · +ĠاÙĦ ÙĪ +ê¹ Į +ØŃ ÙĬ +ار ات +à¹Ģ à¸ĭ +ب ا +г ÑĢ +ร ี +ืà¸Ń à¸Ļ +ع ت +ÙĤ اÙĦ +د Ùħ +Ø ¡ +Ġ×ŀ ×§ +×ĵ ×Ļ×Ŀ +×¢ ׾ +ãģ Ĵ +ëĭ ĺ +×¢ ×Ķ +Ġìĸ ´ +Ñģ ÑĮ +ÙĤ Ø· +ãĥ Ľ +èĢĥ ãģĪ +à¹ģ à¸Ļ +ÙĪ Ø§Øª +â u +ĠìĤ¬ ëŀ +ห ว +ĠاÙĦØ£ Ùħ +Ġ×Ķ ×ŀש +ب ÙĪ +à¸Ĭ à¸Ļ +ãĤĵ ãģ§ãģĻ +ว à¸Ļ +à¸ģร รม +×ŀ ×ķ×ĵ +Ùĥ اÙĨ +×ķ× £ +ол ог +ت ÙĨ +à¸ķ à¹Į +ê² ĥ +ר ×ĺ +ừ ng +×ķ×ij ×Ķ +Ùħ ØŃ +ĠÐ § +פ ×Ĵ +ส à¸ĸ +ãģĭ ãĤĬ +ını z +à¹Ģ ย +ãĥ¼ ãĥ³ +ãģĬ ãĤĬ +פ ש +ิ à¸ķ +Ø· ÙĨ +×Ļת ×Ļ +×IJ ׳ +ç ek +ì ª +×ŀ ×ij +ศ า +ãĤ¹ ãĤ¿ +à¸ļ ุ +×ĵ ×ijר +ãģĦ ãģı +ส ะ +à¹Ģ หล +ิ à¸ĩ +à¸ŀ ัà¸Ļ +ãģĦ ãģŁãģł +ãĤĤ ãĤī +à¹ī ม +ãģĵãģ¨ãģĮ ãģ§ãģį +าร à¹Į +ุ à¸ĩ +í ij +ì ¯ +ë ¼ +í Ĥ +ì · +ê ¡ +á ı +á Ĵ +ðĿ ľ +á © +ðŁ Ħ +ðIJ ¤ +Ġש ׾ +Ġ×ŀ ×Ķ +à¹ģล ะ +Ġ׼ ׾ +Ạ½ +á»Ļ ng +ذ ÙĬ +л е +× ¥ +ãģª ãģ© +ĠÙĪ Ø£ +หà¸Ļ à¹īา +ãģ¾ ãģ§ +à¸ķà¹Ī à¸Ń +à¸Ĺ ัà¹īà¸ĩ +ãģł ãģij +à¹ģà¸ļ à¸ļ +à¹Ģร า +פ ׾ +ãģŁ ãģĦ +à¹Ģล ย +ãģ£ãģ¦ ãģĦãĤĭ +ế p +ึ à¹Īà¸ĩ +ê ´Ģ +ê³ Ħ +׼ ×ķ +à¹Ģร ืà¹Īà¸Ńà¸ĩ +×§ ×Ļ +êµ Ń +פ ס +ت ÙĬ +ãĥ Ħ +Ġ×Ķ ×Ĺ +г и +ר×IJ ׾ +×ŀ ׾ +ĠØ£ ÙĬ +Ġع ÙĦÙĬ +ãģĭ ãģ£ãģŁ +ש ×Ļ +д Ñĥ +×ŀ ף +׳ ×ĺ +׳ ×Ļת +mi ÅŁ +׼ ×Ŀ +Ġ×ij ר +Ġ׾ ×ij +ĠÐ Ľ +ç e +×ķ׳ ×Ļ +ãĤĪãģĨ ãģ« +פ ×ķר +ãĥ į +Ùĥ ÙĬ +×Ĺ ×ª +Ùģ ÙĦ +Ġ×Ķ ×§ +Ġ×Ķ ×ij +Ġ×ŀ ס +à¹Īา à¸Ļ +п еÑĢ +à¹Īา ว +Ġ×ij ×IJ +ĠÙĪ Ùĩ +à¸Ļ ำ +Ġ×ij ש +׳ ×§ +ãģ© ãģĨ +ש ×ķת +×ĵ ×Ķ +à¹Ģ à¸ļ +ÙĨ س +Ġìļ° ë¦¬ +ส à¹Īวà¸Ļ +ล ัà¸ĩ +ج ز +Ġ×Ĺ ×Ļ +Ùĥ ثر +ล ะ +Ùĩ د +ĠÙĪ Ø¨ +اÙĦ Ùħ +à¹ģ ม +Æ¡ i +Ġ×ij ×Ĺ +ữ a +à¹Ģà¸Ĺ ศ +à¸ķ ัà¹īà¸ĩ +ог да +׾ ×§ +د د +สร à¹īาà¸ĩ +à¸Ĭ ี +Ùģ Ø¶ +à¹ģ ห +uy á»ĩn +ร ัà¸ģ +á»ĩ m +ส า +פ ×§ +ีย à¸ĩ +à¸ķ à¹Īาà¸ĩ +à¸Ħร ัà¹īà¸ĩ +ØŃ ÙĤ +à¹Ģ à¸Ńà¸ĩ +ائ ÙĬ +×ĺ ×¢ +اÙĦ Ø© +ิ à¹Īม +ãĤ ½ +د Ùī +Ġר ×IJ +ãģ£ ãģ¨ +ãĥĥ ãĥĹ +ÙĬر Ø© +ê± ´ +×ŀ ×IJ +×ķ ×ķ +ب ع +ãģ ² +ร าย +×ĵ ×Ŀ +ت Ùģ +à¸ķ à¸ģ +ạ ng +ãĤĴ è¦ĭ +à¸Ĭ ั +Æ°á» Ł +Æ°á»Ł ng +ج ب +×ķ×ŀ ר +ĠìĤ¬ëŀ Į +ó ng +ร ั +Ġ×Ķ ×ĸ +ר צ +Ġ×Ĺ ×ĵ +ذ ÙĦÙĥ +×ķר ×Ļ +ãģ¡ ãĤĥ +Ùģ Ø¹ +Ġ׾ צ +á i +à¹ĩ à¸ļ +ãģ İ +à¸ģ ิ +ạ c +ë© ° +ãģª ãĤĭ +×ķ׾ ×Ŀ +à¹ģ à¸Ĺ +×ķ× ¥ +м еÑĤ +ü ÅŁ +ÑĢ Ñı +ภĴ +ÑģÑĤ оÑı +ع ÙĪØ¯ +Ùħ ار +Ø· Ø© +à¸ŀ ื +к ÑĢ +à¹ģ à¸ģ +à¹Ĥ รà¸ĩ +×ij ×Ļ×ĺ +ê² ł +×ķ׾ ×Ķ +ØŃ ر +ืà¹Ī à¸Ńà¸Ļ +×ķ×ij ר +׊ש +ãĥķãĤ ¡ +×ŀ ×ĺ +ú t +Ġd ön +ắ ng +ëł ĩ +ẳ ng +ว à¸ģ +ص د +Ø® Ø· +à¸Ń ั +ãĤı ãĤĮ +سÙĦ اÙħ +à¹Ģร à¹ĩ +×Ļש ×Ļ +ج اÙĦ +ãģij ãĤĭ +à¸Ĭา à¸ķิ +ÙĪØ§ ÙĤ +à¹Ĥ à¸Ļ +ãģ¦ ãģĹãģ¾ +اع Ø© +ãĤŃ ãĥ£ +à¸į า +ÙĦا ÙĤ +ิ à¸ģ +ĠÑģ ов +ÑĢаРº +×Ļ׳ ×Ļ +ü ÄŁ +Ã¼ÄŁ ü +×§ ×ij +à¹Ī à¸Ńà¸ĩ +Ġger çek +à¸Ĺ ั +ов аниÑı +×ŀ ׼ +س Ø© +×Ļ× £ +le ÅŁ +Ùħ ؤ +ĠìĿ ĺ +à¸IJ าà¸Ļ +ĠÑģ об +Ġêµ Ń +×¢ צ +з в +ส à¸ĩ +ز ÙĦ +ãģı ãĤĮ +и ÑĢÑĥ +ت Ø£ +п олн +ìĺ Ģ +ÙĨ Ø´ +׼ ×IJ +Ùħ Ø´ +à¸Ķ à¹Į +ÙĪ ÙĬÙĦ +à¹ģ à¸Ĥ +ãģ£ãģ¦ ãģĹãģ¾ +но ÑģÑĤ +в л +Ùħ ÙĤ +را ج +å¤ ī +ë Ľ +â ¸ +ì IJ +à » +á ļ +â » +ê Ļ +â § +ð Ĵ +ðĿ ĩ +Ġ×IJ ת +ĠÙĦ ÙĦ +ĠØ£ ÙĨ +Ġ×ķ ×Ķ +ãģ« ãģ¯ +Ġ×Ļ ×© +ت Ùĩ +ÃŃ nh +ÙĬ ات +Ġ×ij ×ŀ +à¸Ļั à¹īà¸Ļ +à¸Ļ à¹īำ +Ãł o +à¸ķ าม +ãģ® ãģ¯ +d ır +Ġn ghi +ặ t +×ŀ ×Ļ×Ŀ +ãģ¦ ãģĦãĤĭ +Ġ×ij ת +หร ืà¸Ń +Ġس ÙĬ +ãģª ãĤī +à¹Ĥà¸Ķ ย +ı yor +à¸Ńี à¸ģ +á»ĩ nh +Ñĭ м +à¸Ĺุ à¸ģ +Ġ׾ ×Ĺ +Ġ×Ķ ×¨ +Ġ×Ķ ×Ļ +à¸ŀ ระ +à¹Ģว ลา +ĠØ º +ẫ n +m Ä±ÅŁ +׼ ×Ķ +á»ij n +ãģ§ ãģĹãĤĩãģĨ +ãĥ ¢ +à¸Ľ ี +ס ×Ļ +ãģĵ ãĤį +Ġ׾ פ +ร à¸ĸ +ê¸ Ī +à¸ģ วà¹Īา +ë ¬´ +á»į ng +ãĤĵ ãģ§ +ãĤĪãģĨ ãģª +á»ĵ i +ãĤ ¬ +ส à¹Īà¸ĩ +×Ļ׳ ×Ķ +à¸ĸ ูà¸ģ +à¸Ī ัà¸Ķ +Ġ×Ķ ×Ĵ +ãĥ ľ +×ŀ ×ķת +ÙĪ Ùĥ +ëĭ ¨ +ĠØ « +ãģ® ãģĮ +à¹Ģห à¹ĩà¸Ļ +ع ا +à¸Ļ ิ +Å ŀ +à¸Ń ะ +ãģĪ ãĤĭ +Ø« ÙĦ +ØŃÙħ د +à¹Ģà¸ģ ิà¸Ķ +פ שר +פ ×Ķ +ม ิ +ئ ÙĬس +à¸Ĺำ à¹ĥหà¹ī +×¢ ×ĵ +ìĭ ¤ +à¸Ĭà¹Īว ย +ĠاÙĦÙħ ÙĨ +ز ÙĬ +ع ÙĬ +Ġ׼ ×IJ +ạ nh +á» ¹ +ãĤĵ ãģª +ส ู +צ ר +Æ°á»Ľ ng +×ķ ×ķ×Ķ +à¹Ĥ ล +ĠاÙĦ Ùĩ +ว า +หล าย +Ñī е +à¸Ĥ à¹īà¸Ń +à¹īà¸Ń ย +ب Ø· +ка Ñı +ĠØ ¢ +Ġи Ñģ +ĠاÙĦ غ +à¸ģ า +à¸Ļ à¹Īา +ÙĬ ÙĪ +×ij ×ķר +á»ħ n +ว à¸ĩ +×Ļ× ĸ +ì² Ń +н им +ëŁ ° +×Ĵ ×ķר +ص ØŃ +ÙĦ ÙĪ +×Ĺ ×ķת +ส ุ +رÙĬ ÙĤ +ס ×ĺ +Ġ×ŀ ×¢ +ãĥĨ ãĤ£ +à¸Ħ ิà¸Ķ +ãĤį ãģĨ +à¹Ħ ล +à¸Ļ à¹Į +á»ı i +ÑģÑĤÑĢ Ð¾ +ส à¸Ķ +ส าร +ÙĪÙĦ Ø© +ầ m +ร à¹Īว +รà¹Īว ม +ร ุ +ĠاÙĦس ÙĬ +ìĺ ģ +Ġ×ŀ ×ij +פ ×ĺ +à¸ķิ à¸Ķ +×ĺ ×Ļ×Ŀ +Ġë ¬´ +ÙĤد Ùħ +Ġdü ÅŁ +ائ ÙĦ +м Ñĭ +ØŃ س +ÙĪ Øµ +×Ļ×§ ×Ķ +ãģ§ãģ¯ ãģªãģĦ +à¹Ģ หม +оÑĢ ÑĤ +í Ĩµ +ãģ IJ +к ÑĢа +ีย ว +ع ار +ئ Ø© +íĥ Ģ +ãģ«ãģª ãĤĬ +ج Ø© +ÙĪÙĤ ع +ÑĮ Ñı +×ķצ ×Ķ +ש ×Ŀ +ب ÙĤ +Ġ×Ļ ×Ķ +ÙĬ Ø· +ım ız +д еÑĢж +×Ļש ר×IJ׾ +غ ÙĬر +ร à¸Ńà¸ĩ +à¹Ģรีย à¸Ļ +Ġ×Ķ ×ĺ +หม าย +Ùħ Ùĩ +اÙģ Ø© +Ġо ÑĢг +ÙĪ Ùī +ãĥ© ãĤ¤ +×ŀ ׳×Ķ +ĠÄij o +Ġг оÑĢ +اÙħ Ø© +æ¥ ½ +Ø« ÙĬر +à¸ģิ à¸Ī +á»ĵ n +ÙĨ ب +ÑĢÑĥ д +ìĹ Ī +Ġ×Ĺ ×ijר +ÑĢаР¶ +ạ ch +ت ÙĪ +à¹Ĥ ม +×ij ×Ļ×ij +Ġí Ĩµ +aca ģı +جÙĦ س +à¹Ģà¸Ľ ล +ว à¸Ķ +à¸Ń ล +ãģŁ ãĤĬ +à¸Ľ ัà¸į +Ġìķ Į +عر Ùģ +à¹Ħ à¸Ł +Ø£ Ø® +å¤ļ ãģĦ +à¸Ķ ัà¸ĩ +Ø´ Ùģ +ãģ£ãģ¦ ãģĦãģ¾ãģĻ +׼ ×ł×¡ +ÑĨ е +еÑģ п +Ùħ اÙħ +à¸ŀื à¹īà¸Ļ +иÑĩеÑģ ки +Ø® د +Ùĥ ÙĪÙħ +Ġ×Ķ ×¨×IJש +ت اب +é£Ł ãģ¹ +ื à¸Ļ +оÑĢ Ð¾ +Ġb öl +×ķ×Ĺ ×ĵ +دÙĬ ر +ắ m +د ع +ãģķ ãģĽ +à¸ĺ ร +à¸ĺร รม +ãģĭ ãĤĤ +å¤ļ ãģı +r ä +س ع +×Ļ׾ ×Ķ +ض ر +ĠاÙĦ شر +×ĸ ×ķר +×¢ ×ijר +ạ m +алÑĮ но +ر ÙĨ +اÙħ ج +׼ ×ļ +d ıģ +д ен +ض ا +ÙĦÙĬ Ùħ +Ġê·¸ 룬 +تÙħ اع +ار ÙĬØ® +à¹Ĥ à¸ķ +ĠÑģ ÑĢед +Ġ׳ ×ķס +ÙĤ بÙĦ +оÑĤ ов +le ÅŁtir +Ġм еÑģÑĤ +سÙĦ Ùħ +Ġ×¢ צ +ĠاÙĦس ÙĦ +еÑĤ ÑĮ +اب Ø© +н ак +สà¸ĸ าà¸Ļ +Ġ×ij ׳ +à¸ļ ัà¸Ļ +׼ ׳ +Ġö ÄŁ +ãģ¨ è¨Ģ +uy ến +di ÄŁ +áºŃ u +ÑĢ Ð°Ñģ +ãĤ· ãĥ§ãĥ³ +n ız +×ķ×ĵ ×Ķ +ت س +Ùħ اÙĦ +à¹Ģห à¸ķุ +ย ว +à¸ŀ ัà¸ģ +ãģĦ ãģªãģĦ +Ġк аÑĩ +ล à¹Į +ר׼ ת +ÅŁt ur +×ŀ ×ķס +ãģ ¥ +б ол +عÙħ اÙĦ +×ķר ת +ÑĨи он +ศ ึà¸ģ +ภı +ÑĢ ÐµÐ½ +اس ÙĬ +ائ ر +à¹Ĥ à¸Ľà¸£ +Ġse ç +غ ÙĬ +Ñį ÑĤ +ен н +ãģª ãģ® +×Ļש ×Ķ +×Ļפ ×ķר +ãģŁãĤģ ãģ« +ز Ø© +Ġç oc +ãĤ¯ ãĥª +ÑĪ ÐµÐ½ +ãĤı ãģij +رÙĬ د +ĠÑĢ Ð°ÑģÑģ +Ùĥ ات +ส à¸Ńà¸ļ +ce ÄŁi +ãĤ¿ ãĤ¤ +à¸ļ ร +ĠاÙĦ بر +׳ ×ķ×¢ +r ün +را ض +ศา ส +à¸ķ รà¹Į +ãģį ãģŁ +×ķ׾ ×ĵ +еÑĢ Ð¸ +íĹ ĺ +ắ p +ت عÙĦ +Ùĥ د +иÑĤелÑĮ но +Ø· Ùģ +Ġав ÑĤом +Ġ×ŀ צ +ÑĪи Ñħ +ات Ùģ +ĠÑħ оÑĤ +Ùİ Ø§ +ãģı ãĤĭ +×Ķ ×¤ +à¹Ĥ à¸Ĺ +à¹ģ à¸ŀ +à¹Ī à¸Ńย +ĠاÙĦÙħ Ø´ +à¸ģาร à¸ĵà¹Į +ани з +×Ķ ×ľ +ظ Ùħ +ย ุ +li ÄŁ +à¹Ħ à¸Ĥ +à¸ĸ ืà¸Ń +ö z +ãģij ãģ¦ +à¹Ģ à¸ľ +ุ ม +ãĥĹ ãĥ¬ +Ġ×Ķ×IJ ×Ĺר +خت ÙĦÙģ +à¸ İ +ÙĦا ØŃ +Ġdü zen +צ ×Ķ +س اء +×ķר ×ļ +×ķ×ĵ ×Ļ +ÑĢа ÑĦ +ÅŁt ır +ãģ« åħ¥ +ãģĪ ãģ° +ص ÙĪÙĦ +ĠÐľ оÑģ +ا Ùĩر +ãģ£ ãģ +ĠлÑİ Ð± +×Ļ×¢ ×Ķ +Ġ×Ķ×ŀ ×§ +สิ à¸Ĺ +สิà¸Ĺ à¸ĺิ +×Ļ׳ ×Ŀ +ÙĦا Ùģ +à¸ŀัà¸Ļ à¸ĺ +×ķ×IJ ×Ķ +ม ั +à¸Ĥ à¸ĵะ +д оÑĢ +ãģ¨ ãģª +à¸ģระ à¸Ĺ +ac ı +×ķ׾ ×ķ×Ĵ +Ñĥ ÑĪ +ãĥ¥ ãĥ¼ +ãĥ ¦ +Ùħ ست +Ġa ÅŁ +ש ×§ +פ ת×Ĺ +าย à¸Ļ +í ĩ +ë ¢ +ï · +í ī +ì µ +ì ¬ +ðĿ Ľ +ì Ĵ +ë Ļ +ê § +á ĸ +â ¨ +â ± +á ĺ +ð ĸ +à ł +á Ķ +ðIJ Ń +ữ ng +Å© ng +Ġ×Ķ ×ª +ĠاÙĦ ا +Ġ×ŀ ת +à¸ĸ ึà¸ĩ +ò n +á»ĭ nh +нÑĭ м +Ġc ả +à¸Ķ ู +Ġ à¹ģà¸ķà¹Ī +Ġ×ij ×Ķ +ó i +ãģ¨ ãģĹãģ¦ +ú ng +ĠØ ° +Ġ×Ķ ×ł +Ġب ÙĨ +ÙĦ اÙĦ +à¹Ħ à¸Ĺย +á»ĩ p +t ı +ม ัà¸Ļ +ằ ng +á»ij t +к ом +à¸ĭ ึà¹Īà¸ĩ +à¸Ħร ัà¸ļ +à¸ļ à¹īาà¸Ļ +ĠاÙĦ ÙĬ +l ü +ÙĪ Ø³ +ãģł ãģ£ãģŁ +à¹Ģ à¸ĩ +Ġê³ µ +н Ñĥ +ãĤĪ ãĤĬ +м Ñĥ +à¹Ģà¸Ĥ า +ãĤ Ģ +ни е +ãģ«ãģª ãĤĭ +áºŃ y +ĠÙĪ Ø§ +ëł ¤ +ש ×ķ +á p +×ĵ ×ķ +ãģ§ ãģĹãģŁ +ع ض +Ñģк ой +æĦŁ ãģĺ +ÑİÑĤ ÑģÑı +Ġ×Ļ ×Ľ×ķ׾ +ãĤĵ ãģł +в и +à¹Ģล à¹Īà¸Ļ +ìĿ´ ëĭ¤ +ĠÙĦ Ùĩ +à¸Ħ ืà¸Ń +ت Ùĥ +Ùħ ÙĥÙĨ +a ģı +׳ ×ĵ +ë¯ ¼ +à¹Ħ ว +สำ ห +สำห รัà¸ļ +Ñģл ед +t ır +ĠÙĦ ÙĬ +ĠاÙĦع ÙħÙĦ +×ij ×ķת +×ij ×Ļ×Ŀ +à¸Ħ ำ +à¹Ģà¸Ħร ืà¹Īà¸Ńà¸ĩ +lı ģı +ืà¸Ń à¸ĩ +ج د +íŀ Ī +ìĭ ¬ +×¢ ×ķת +ส ิà¸Ļ +Ñĩ и +ر ض +à¹Ģà¸Ľ ิà¸Ķ +à¸Ħ à¹Īา +ìĦ ł +ÙĪØ± Ø© +×§ ×ĺ +ìľ ł +ع ÙħÙĦ +×IJ ×Ļ×Ŀ +׾ ×Ļ×Ŀ +à¹ĥห à¸į +à¹ĥหà¸į à¹Ī +ừ a +á»į i +ãģ ¶ +ÃŃ ch +ãĥĩ ãĤ£ +×ķר ×Ļ×Ŀ +Ñģ о +ìķ ½ +ов а +Ñĩ аÑģÑĤ +à¹Ģà¸Ī à¹īา +п ÑĢо +Ġ×ŀ ×Ĺ +ãĥ İ +×ķ×Ļ ×ķת +Ġд е +ë§ Ī +ì§ ģ +×Ļפ ×Ķ +ĠاÙĦع اÙĦÙħ +ë¥ ´ +ר×IJ ×Ķ +uy á»ĥn +×¢ ×Ļ +ม ืà¸Ń +Ø¥ ÙĨ +ร ู +ĠØ ² +×Ļ ×ķ×Ŀ +à¸ķ à¹īà¸Ļ +ãģ¦ ãģĦãģ¾ãģĻ +Ùħ اÙĨ +ĠÐ ¥ +à¸Ľà¸£à¸° à¹Ģà¸Ĺศ +á» ³ +׾ ×ij +à¹Ģà¸Ķ à¹ĩ +ãģŁ ãģ¡ +à¸Ĺี ม +à¸Ļ ะ +ìĹ ° +Ġìł Ģ +ÙĦ Ùĩ +ợ i +ĠاÙĦ ز +د ار +ãĤ³ ãĥ³ +м ин +à¹ģห à¹Īà¸ĩ +à¸Ķ ัà¸ļ +׼ ר +ж а +íĸ Ī +×ŀ ×ĸ +ợ i +à¸Ķ า +Ġع بد +à¹ģ ร +×IJת ר +×¢ ׳×Ļ +à¹Ģ à¸Ħ +×ķצ ר +ì§Ģ ë§Į +ائ Ùħ +Ø£ س +uy á»ģn +Ġ×IJ ׳ +׊׳×ķ +×ĸ ×Ļ +ร à¹īาà¸Ļ +ĠÐł оÑģ +ĠÐłÐ¾Ñģ Ñģ +رب ÙĬØ© +t ür +ãĤĭ ãģĵãģ¨ +ظ ر +б Ñĭ +à¸Ĺีà¹Ī สุà¸Ķ +Ġצ ר +èĩª åĪĨ +л аÑģ +ĠÑı в +ĠÑıв лÑı +à¸ŀร à¹īà¸Ńม +à¸Ńา à¸Ī +à¸ļริ à¸ģาร +Ġç ı +ëį ĺ +ĠاÙĦÙħ ست +ت Ø´ +ש ×ķ×ij +ãĤ ´ +Ġyap ıl +ĠاÙĦ ذ +ุ à¹Īม +à¸ĸ à¹īา +ìĦ ¤ +ì° ¨ +в аÑĢ +à¹Ģà¸ŀ ิà¹Īม +Æ°á»Ľ i +Ùĥ س +à¸Ńย าà¸ģ +ãģ¦ ãĤĤ +Ġг од +ÙĬ ار +à¸ķ à¸Ńà¸Ļ +Ġиг ÑĢ +à¹Ħà¸Ķà¹ī รัà¸ļ +ĠاÙĦÙħ ر +ÙĤ ت +Ġë ĺ +Ġëĺ IJ +ẩ n +ãģĻãĤĭ ãģĵãģ¨ +×Ĵ ×Ŀ +Ġ×ij ×ij +ت د +ÙĪ Ø§Ø± +ãĤ ® +п ол +Ġм ог +تر Ùĥ +ÙĪ Ø« +Ġç ık +ا Ø© +à¹Ģà¸Ķ ียว +มี à¸Ħวาม +Ġ×ŀ ×Ĵ +ص Ùģ +ĠТ ак +Ġ׼ ת +×Ļ×ĵ ×Ļ +ов оÑĢ +ầ y +สิ à¹Īà¸ĩ +ب ت +ür ü +ÙĨ ج +หล ัà¸ģ +×Ļ×Ķ ×Ŀ +ÙĤ ص +з Ñĭ +×Ľ×ª ×ij +ư u +m ız +ĠìĦ ¸ +л ог +Ùħ ÙĬÙĦ +ÙĬ ج +íĴ Ī +à¸ŀ à¸ļ +ห ัว +з на +ר ×§ +à¹Ĥ ร +Ġ×ij ס +ĠBaÅŁ kan +ĠëĶ ° +à¸Ń ัà¸Ļ +ีà¹Īย ว +н еÑģ +à¹Ģà¸Ķ ิà¸Ļ +ÙĬ اÙĨ +×ķ׾ ×Ļ +ا خت +צ ×ķת +ãģĵ ãģĵ +ĠاÙĦ اÙĨ +ĠпÑĢо ÑĨ +ãģ¾ ãģł +׼ ס +ĠاÙĦ Ø¢ +ÙĬ ز +ĠاÙĦد ÙĪÙĦ +Ġíķĺ ëĤĺ +ض ع +ê» ĺ +ÅĽ wi +ย ิ +ãģ¡ãĤĥ ãĤĵ +ĠÙħ Ø´ +à¸ĺ ี +ãģ¨ ãģį +׳×Ļ ×ķת +Ġë ¯ +Ġë¯ ¸ +Ġs ı +ëĭĪ ê¹Į +Ġп л +غ ÙĦ +à¹ģ รà¸ĩ +ب ÙĬر +ãģĤãĤĬ ãģ¾ãģĽãĤĵ +ê· ¼ +Ġy üz +ĠdeÄŁ er +åł´ åIJĪ +á» ¡ +м аÑĤ +รา à¸Ĭ +ÙĪØ± ÙĬ +ж ен +ãģ¾ ãĤĬ +ãģ® ä¸Ń +×Ļ×ĵ ×¢ +à¸Ń ุ +à¸ļ à¸Ńล +à¸Ľà¸±à¸į หา +ز Ùħ +ÄŁ a +à¸Ń ืà¹Ī +à¸Ńืà¹Ī à¸Ļ +п л +Ġне обÑħодим +׼ ×ij +à¹Ģ ศ +קר ×Ķ +ì² ĺ +ëł ¨ +×ŀ×§ ×ķ×Ŀ +jÄħ c +Ùĩ ÙĦ +Ġ×¢ ×ij×ķ×ĵ +à¹Ħม à¹ī +à¸ģล ัà¸ļ +×ķ׼ ׾ +×§ ×ĵ +اÙĦ ÙĬØ© +ر Ùĩ +ãģij ãĤĮãģ° +ĠÙĨ Ù쨳 +ãĤ¢ ãĥ« +ìĹ Īëĭ¤ +×§ ×ķר +н еÑĢ +ب اب +ãĤ ¶ +سب ب +ÙĦ ÙĬÙĦ +ص ÙĨ +ص در +ế m +à¸Ĭà¹Īว à¸ĩ +ØŃ ÙĨ +Ġ×ij ×Ĵ +×ŀ ×ķ×¢ +׾ ×Ĺ +大 ãģį +ت ب +н еÑĤ +×Ļ×ij ×Ķ +б л +ãĥĹ ãĥª +اص Ø© +ãģ¤ ãģij +×Ļ×ŀ ×ķש +ãģĮ ãģĤ +ëĭ ´ +ãģĭãĤĤ ãģĹ +ãģĭãĤĤãģĹ ãĤĮ +ãģ¡ ãĤī +×ij ×ĺ +Ġba ÄŁ +×Ļ×Ĺ ×¡ +×ij ×ķ×¢ +ล ี +פע ×Ļ׾ +им и +g ÅĤ +Ġим е +خد اÙħ +×IJ ×Ļר +Ġy apt +ãģ¨ ãģĦ +à¸ĩ à¹Īาย +׾×Ļ ×ķ +ØŃد Ø« +را ÙĤ +ĠÄIJ i +اد ر +ãģĵãģ¨ ãĤĤ +×ij ×Ļר +Ġв з +ض اÙģ +ת ×ķ׼ +ÑĢ Ð¾Ð¼ +ر ات +à¹Ģà¸Ĺ à¹Īา +ãģĺ ãĤĥ +ãģĿ ãģĵ +اج تÙħاع +à¹īà¸Ń à¸Ļ +ÙĤ Ùħ +ë³ ¸ +Ä ŀ +ש ×Ļ×ķ +×ij ׳×Ļ +ìľĦ ìĽIJ +à¹ģ à¸Ī +×Ĺ ×ķר +دÙĬ ÙĨØ© +ت Ø· +ằ m +ò a +ย à¸Ńà¸Ķ +Ġëĭ ¹ +สุ à¸Ĥ +×ĵר ×ļ +د ÙĨ +س ÙĬÙĨ +ÙĪÙĤ Ùģ +ÑĨ Ñĭ +г оÑĤов +еж дÑĥ +à¸ŀ วà¸ģ +اÙĤ تص +اÙĤتص اد +cz ÄĻ +ni ÄĻ +ÑĢ ÐµÐ± +ØŃ ÙĪ +à¸Ĺ à¹Į +ãĤĪ ãģŃ +д ж +à¸ģล à¹Īาว +دÙĬ Ø« +ãĤ³ ãĥŁ +ÙĤ ÙĪÙħ +Ġت ØŃ +à¹Ģ à¸ķิ +اÙģ Ø¸ +à¸Ī ุ +رÙĬ اض +×ŀש ×ļ +à¹Ĥ ย +еÑĢ Ðµ +ãģ¿ ãģŁãģĦ +ìĿ´ ëĿ¼ +ĠاÙĦÙħ ÙĪ +ĠÑģÑĤ о +à¹Ģรà¹ĩ ว +Ġд еÑĤ +ĠÑģ дел +à¹Ģà¸Ĭ ืà¹Īà¸Ń +פ ׳×Ļ +ÙĪØ¶ ÙĪØ¹ +×ij ס +à¹ģ à¸Ķ +ó c +ริ ม +ÑĢаР´ +ìĪ ł +ãĥ¼ãĤ º +ãģ« ãģĬ +и но +פ ×Ļ׾ +à¸Ĭั à¹Īà¸Ļ +×Ĺ×ĵ ש +à¹Ģà¸Ļ ืà¹Īà¸Ńà¸ĩ +׳ ×Ļס +غ رب +ãĤ¸ ãĥ£ +ส ัà¸ĩ +à¹Ģ à¸Ĺีà¹Ī +à¹Ģà¸Ĺีà¹Ī ยว +ëŁ ¼ +à¹ģ à¸Ł +ãĥ¼ãĤ · +ãĥ¼ãĤ· ãĥ§ãĥ³ +Ġвоз мож +جÙħ ÙĪØ¹ +×ijר ×Ļ×Ŀ +ãĥĪ ãĥ© +ĠкаÑĩ еÑģÑĤв +Ø· ÙĬ +ÑĤ Ñı +צ ×ķ×¢ +ÄŁ ını +ع ÙĦÙī +ا ذ +ÙĪØ§ÙĤ ع +Ùħ ÙĪØ§ +ائ ÙĬÙĦ +к ол +á»ģ m +à¸ľà¸¥ ิà¸ķ +×Ļ׳ ×ĺר +س Ùĥ +ש ×Ļר +ศึà¸ģ ษา +à¸ļ ั +Ñĩ аÑģ +×ķפ ×Ķ +×Ļפ ×ķ׾ +ĠاÙĦس اب +رÙĬ ب +ĠاÙĦ بÙĬ +ãĤ¹ ãĥĨ +Ñĩ ен +à¹ģ à¸ľ +Ġ׳ ש +ز ÙĬد +ØŃ اد +ëį Ķ +رÙĪ Ø¹ +à¸Ĺุ à¸Ļ +ส มา +c zeÅĦ +×Ļ×ĵ ×Ķ +ãģ§ ãģĤ +Ġçoc uk +Ø® ب +à¸ļ าย +à¸Ľà¸£à¸° à¸Ĭา +×ŀש ׾ +ãģª ãģĭ +à¸ģ าย +ãĥģ ãĥ£ +аÑĢ Ð¸ +ĠÑĩ а +à¸Ķ ำ +à¸Ĺั à¹Īว +Ñĥ Ñħ +Ġö z +Ġì¢ ĭ +ج رÙĬ +ائ ÙĤ +à¸ł ัย +Ø· ار +د ارة +Ä© nh +Ø« ÙĨ +zell ik +اÙĦ ت +Ġg eli +ãĥķãĤ © +ол од +رب ع +שת ×ŀש +à¸ļร ร +íĿ ¬ +Ġü rün +Ġê·¸ ëłĩ +ศาส à¸ķรà¹Į +ãģ ľ +×Ļ×ij ׾ +ĠпÑĢед ÑģÑĤав +سط ÙĬÙĨ +ãĤĴ 使 +Ġпом оÑī +×ķ×§ ר +ãĥ¯ ãĥ¼ +Ġyö net +×Ļ×§ ר +à¸Ĥ า +еÑĢи ал +ØŃ Ùģ +Ġ×Ļ ×¦ +à¸Ĺ ิ +å£ ² +à¸Ļ à¸Ńà¸ģ +×ķ׼ ר +íĻ ľ +á»§ y +ĠاÙĦÙĤ ر +×Ļ×ij ×ķת +ÅĽ ni +Ùħ شار +ượ t +ĠÙĦ دÙĬ +ÑĤ ел +ĠØ¥ ÙĦÙĬ +عÙĦ ÙĪÙħ +ìķ ĺ +в иÑĤ +à¸Ħ ะ +yr ı +ãģ¨ ãģ£ãģ¦ +à¹Ģ à¸ī +à¸ĸ าม +ÙĤ ار +عÙĦ اÙħ +ặ ng +Ùħ ÙĴ +×Ļ×ŀ ת +سب Ø© +ãĤ¯ ãĥ© +×ķס ×£ +ĠпÑĢ Ð¸Ð½ +ãģĦ ãĤį +س اس +عت بر +วิ à¸Ĺย +วิà¸Ĺย า +س Ùĥر +ãĤ· ãĥ§ +ãģ ģ +ัà¸ģ ษ +×ij ×ķ×Ķ +ห ย +ãģ¾ ãĤĮ +ĠоÑĢг аниз +каз ал +ĠÑģв Ñıз +uy ết +ĠпÑĢо из +Ġ×§ ×ĺ +à¹ģà¸ģ à¹ī +п ÑĥÑģ +Ġê·¸ ê²ĥ +ëĬ IJ +л екÑģ +ãĥ¼ãĥ Ĺ +à¸ķ ำ +ת×Ĺ ×Ļ׾ +à¸Ńà¸ĩ à¸Ħà¹Į +Ạµ +׳ צ +Ø£ Ø´ +Ø´ Ùĩ +ย ะ +à¸ģ à¸İ +ĠاÙĦØ¥ سÙĦاÙħ +ед ÑĮ +ãģ² ãģ¨ +ëıĦ ë¡Ŀ +ãģ© ãģ® +Ñĥ в +еÑĩ ение +ĠاÙĦت ج +ãģ« è¡Į +Ġп озв +ãĤı ãĤĬ +ÙĦ اث +íķĺ ìĺĢ +Ġм аÑĢ +Ġkon uÅŁ +ãĥ¬ ãĤ¹ +ãĤĴ æĮģ +ĠоÑģ нов +×Ĺ ×ij +ÙĪØ¬ ÙĪØ¯ +פ ×ķף +в оÑĢ +Ġн ик +ãģĭ ãĤĭ +ÅŁtır ma +×Ļס ×ĺ +Ø£ ÙĦ +ห à¹Į +и она +лÑĮ н +Ġг оÑģ +ĠÐľÐ¾Ñģ к +ÑĢ Ð¾Ð± +×ķ×IJ ×Ļ +ãģĬãĤĬ ãģ¾ãģĻ +ãģ£ãģ ± +к л +à¸Ļ à¸Ķà¹Į +رÙĬ Ùģ +اس ب +ĠÑĢ ÐµÑĪ +Ġд ол +ãģ¹ ãģį +×Ļ×ij ×ķר +м еÑī +Ġна ÑĪ +à¹ģ à¸Ľà¸¥ +ÑĢ Ð¸ÑĤ +кÑĥ Ñģ +и ÑĢа +аÑĤ ÑĥÑĢ +ÙĪØ§ صÙĦ +à¹Ģà¸ľ ย +à¸Ń ำ +à¹Ģà¸ģ ิà¸Ļ +غ Ùħ +ãģĻ ãģİ +lı kl +ÅĦ sk +ê² ¬ +×Ļ׼ ×Ķ +׊ש×ij +ÙĪØ± ÙĬØ© +Ġд ейÑģÑĤв +×Ĺ׾ ×ĺ +Ġ׾ ×ŀ×¢ +צ׾ ×Ļ×Ĺ +еÑĩ а +Ùģ Ø§Ø¹ +×Ĵ ×Ļ×ĵ +áºŃ m +ÄĻ b +Ø´ ع +ãģı ãĤĬ +à¸ŀ ุ +ед еÑĢ +à¸Ĥ à¸Ļ +à¸Ħ าร +ĠболÑĮ ÑĪ +ãģı ãģªãĤĬ +à¸ĵ า +×ĵ ×ķ×Ĵ +Ġм н +ä¸Ĭ ãģĮ +ç¶ļ ãģį +ฤ ษ +ภĨ +Ø® ÙĬ +à¹Ģà¸Ĺ à¸ŀ +สั ม +à¹Ģส à¸Ļ +à¹Ģสà¸Ļ à¸Ń +ãĥ ´ +Ġи ÑģÑĤ +با شر +ĠÑĥ ÑĢов +×ŀ ×ķ×ĸ +ab ı +wa ż +×ķצ ×IJ×Ķ +ÑĤ веÑĢ +à¸ŀัà¸Ļà¸ĺ à¹Į +׳ ×Ĵ×ĵ +ãĤĭ ãģĵãģ¨ãģĮãģ§ãģį +ĠÑĤÑĢ ÐµÐ± +à¸ģร ุà¸ĩ +ØŃت اج +à¹Ģ à¸Ħล +ã Ĩ +ÄĻ tr +Ġszcz eg +Ġר ש +à¸Ĺ à¸ĺ +Ġн ек +Ġнек оÑĤоÑĢ +в ÑĪ +Ð ¬ +à¹Īว ย +ล ุ +б ÑĢÑı +หม ูà¹Ī +à¹ģ à¸ķà¸ģ +ר׼ ×Ļ×Ŀ +Ġí ĸī +ã i +Ùĥر Ø© +â Ń +í IJ +ã į +á ģ +â ® +â ¥ +ì ® +à ¿ +â ¿ +á Ĥ +á ¤ +â ł +í Ł +ðIJ į +ðIJ ° +ðĿ Ĩ +ðŁ Ī +Ġ×¢ ׾ +Ġع ÙĨ +ĠÙħ ع +Ġ×ĸ ×Ķ +ĠÙħ ا +Ġm Ãł +Ġd ụ +á»ĩ c +а Ñħ +s ı +íķĺ ê³ł +Ġ×ķ ×ij +ĠÐŁ о +×ķת ר +ĠÙĦ Ùħ +Ġ×ķ ׾ +ãģĹãģ¦ ãģĦãĤĭ +Ġ×ŀ ×Ļ +Ġب ÙĬÙĨ +з а +ĠÙĥ اÙĨ +Ġ×Ķ ×Ļ×Ķ +ëħ Ħ +×IJ ×ķ +д и +ĠпеÑĢ Ðµ +d ı +Ġ׾ ש +Ġש ×ŀ +ãģĮ ãģĤãĤĭ +ãģĦ ãģĦ +ÑĢ Ðµ +×§ ×ķ +и ли +м е +ÙĬ ت +ãģ§ ãģĤãĤĭ +Ġв о +à¹ĥ หม +à¹ĥหม à¹Ī +Ġש ×ij +Ġ à¹Ĥà¸Ķย +ÙĬ Ùĩ +ãģ§ãģĻ ãģĮ +ãģ¨ ãģ¯ +ר ×ķ +Ġ à¸ĭึà¹Īà¸ĩ +ãģ§ãģį ãĤĭ +м о +à¹Ģà¸ŀ ืà¹Īà¸Ń +צ ×ķ +×ĺ ×ķ +ìķ Ī +Ġh á»į +à¹Ģà¸ĩ ิà¸Ļ +ĠاÙĦ ب +Ġ มี +ë¬ ¼ +Ñģ е +ëĵ¤ ìĿ´ +Ġë§ IJ +Ġl Ỽ +a ÅĤ +×Ĺ ×ijר +Ġd á»± +ÙĬ Ø« +Ġth á»ĭ +à¸ģà¹Ī à¸Ńà¸Ļ +Ġ×ij ׼׾ +ãģ ¸ +ã썿ĢĿ ãģĦãģ¾ãģĻ +ả nh +ย า +Ùģ Ø§ +ส ี +à¸ķ า +ë² ķ +ãĥª ãĥ¼ +รา à¸Ħา +Ġ×ķ ׾×IJ +ãģ¨ ãģĵãĤį +à¹Ģล ืà¸Ń +di ÄŁi +ÙĪ Ø§ÙĨ +Ġ׾×Ķ ×ª +รว ม +פ ×Ļ×Ŀ +à¸ľ ม +ж и +c ı +ÑĢ Ð¾Ð´ +Ġkar ÅŁÄ± +×Ĵ ×ķ +ãģ« ãģ¤ +ãģ«ãģ¤ ãģĦãģ¦ +r Ãł +×Ļ×ķת ר +ĠìĨ Į +×§ ×Ķ +ÑģÑĤв о +ãģij ãģ© +g é +à¸Ķ à¹īาà¸Ļ +çļĦ ãģ« +ĠÙĬ ÙħÙĥÙĨ +ìĨ į +ÙĬ Ùĥ +à¹Ħว à¹ī +Ñģки й +ì m +Ġ׾×IJ ×Ĺר +à¸Ńา หาร +Ġà¹Ģ à¸ŀ +รา ะ +ล ูà¸ģ +ÑģÑĤ а +Ġìľ ł +ÙĤ ÙĪÙĦ +б оÑĢ +Ñģк ого +หล ัà¸ĩ +à¸Ĥ à¹Īาว +à¹Ģม ืà¸Ńà¸ĩ +ê° ģ +t Ãł +ÙĬ ÙĬÙĨ +عر ض +ë° © +Ġëı Ļ +Ġà¹Ģ à¸Ľ +Ġà¹Ģà¸Ľ à¹ĩà¸Ļ +ç i +li ÄŁi +ìĹIJ ê²Į +ãĤ¿ ãĥ¼ +Ġ׾ ת +פ ×ķת +à¸Ĥ à¸Ń +ر س +ìł IJ +à¸ľ à¹Īาà¸Ļ +ÑĦ и +ج ÙĨ +ì¢ ħ +Ġ×Ķ ×¤ +Ġn go +á»ĭ a +Ġtá» ķ +Ġê·¸ 리 +à¹Ģม ืà¹Īà¸Ń +ذ Ùĥر +ìĸ ij +ìĹ Ń +×ĺ ׾ +k ı +Ġع ÙħÙĦ +Ġع ÙĨد +à¸ĭ ืà¹īà¸Ń +Ġê± ° +в е +r ü +à¹Ģ à¸Ńา +ส à¹Į +à¸Ī à¸Ļ +ס ת +Ġgi ả +ãĤĭ ãģ¨ +à¸ģำ ลัà¸ĩ +н ей +à¸Ī ริ +à¸Īริ à¸ĩ +Ġë į +Ġëį Ķ +à¸Ħà¹Ī ะ +ì n +Ġsü re +Ġqu y +à¸ļ าà¸ĩ +åıĸ ãĤĬ +ר ×Ĺ +×ij ת +ãģĮ ãģĤãĤĬãģ¾ãģĻ +ר ש +ìĹIJ ëĬĶ +Ġ×IJ פשר +ay ı +ãģĮ ãĤī +ØŃ ب +ан Ñģ +س ÙĪ +ĠпÑĢ Ðµ +د ÙĪ +ãģ« ãĤĪ +à¹Ģà¸ģ ม +สู à¸ĩ +m akt +makt ad +maktad ır +Ġön em +×Ļ×ŀ ×Ļ×Ŀ +б о +ÙĪ ÙĬØ© +รู à¸Ľ +à¹Ĥล à¸ģ +Ùħ ÙĬع +ÑģÑĤ Ñĥп +à¹Ĥ à¸Ń +دÙĬ ÙĨ +ì¤ ij +ãģĹãģ ı +à¹Ģส ีย +в Ñĭ +Ùħ ت +íĺ Ħ +ãĥIJ ãĥ¼ +ا Ø´ +×§ ס +Ġtá» ¥ +ล à¸Ķ +Ùģ Ø© +í ijľ +ر ج +k ÅĤad +ĠÅŁ ey +ĠØ£ Ùħ +Ġà¹Ģ ม +Ġب ÙĦ +Ñģ каÑı +ãģ¨ ãģ® +Ġìĭ ¤ +ấ m +ห à¹īà¸Ńà¸ĩ +à¸Ĭ ม +d ü +Ġç ek +Ġê³ ł +×Ĵ ×ij +à¸Ĭี วิ +à¸Ĭีวิ à¸ķ +Ù쨶 ÙĦ +ภ¯ +ç ı +Ġب Ø´ +ĠÙĩ ÙĨا +ãģį ãģ¾ãģĹãģŁ +t ü +Ġìĺ ģ +ĠTür k +к ÑĤ +פר ס +ãģ¨ãģĦãģĨ ãģĵãģ¨ +í ĶĦ +à¹ģร à¸ģ +ר ×ķף +Ġar as +×ŀצ ×IJ +Ġtá» ī +س ا +à¸ŀ à¸Ń +ĠاÙĦÙħ ØŃ +ãĥ ¤ +ĠاÙĦ است +Ùģ ÙĨ +×Ļ×ŀ ×Ķ +ر ت +ãģ¨ ãĤĤ +Ġна Ñģ +п ÑĢи +Ġ×Ĺ ×ķ +и ла +ÙĬ Ø´ +Ġgö z +Ġ×ij ׳×Ļ +ım ı +ĠÑĤ еÑħ +Ġh á»Ļ +غ ر +к он +اØŃ ت +Ġ à¸ŀ +à¸Ń à¸Ńà¸Ļ +à¸Ńà¸Ńà¸Ļ à¹Ħล +à¸Ńà¸Ńà¸Ļà¹Ħล à¸Ļà¹Į +Ñħ о +Ñı в +à¹ģ สà¸Ķ +à¹ģสà¸Ķ à¸ĩ +à¹Ģà¸ŀ ียà¸ĩ +ÑĤ ов +ا ÙĬ +Ġ×Ķ ×ĵ +Ġ×ķ ׼ +ãĤī ãģĦ +×ķפ ף +Ġë ¶Ī +ล à¸Ńà¸ĩ +Ø· اÙĦ +Ġн и +ĠÙħ ست +ế c +Ġש ׼ +ĠëķĮ 문 +วัà¸Ļ à¸Ĺีà¹Ī +×Ļ׾ ×ĵ +ØŃ ا +е ÑĨ +Ġc ứ +×ĵ ×ķר +ĠÙħ ØŃ +ר׼ ×ij +بÙĬ ع +ни и +ĠاÙĦØ£ ÙĪÙĦ +à¸Ħว ร +ã썿ĢĿ ãģĨ +ĠС о +ائ ÙĬØ© +ر اء +оÑģ об +Ġب Ø£ÙĨ +×¢ ×ķ×ĵ +ĠÑĤ е +ãģĵ ãģĨ +ÑģÑĤ ÑĢа +ай н +Ġsö z +ت ÙĨا +à¸Ń ิ +ặ p +ĠìķĦ ëĭĪ +íķ Ń +Ġר×IJ ש +Ġ à¹Ħà¸Ķà¹ī +Ġ×Ĵ ×ĵ +Ġס פר +обÑī е +ĠÙĪ Ø¥ +ada ÅŁ +ãģ¡ ãĤĩ +×§ ×ķ׾ +ÑĢ ÐµÐ· +ĠdÃ¼ÅŁ ün +Ġ×ij ×IJ×ŀ +Ġìĸ´ ëĸ +ער ×ij +н ее +ĠÑģÑĤÑĢ Ð°Ð½ +س اÙĨ +yn ı +ĠاÙĦر ئÙĬس +ãģĹãģ ª +Ġ׳ ת +ãģ«ãģª ãģ£ãģŁ +g ü +åıĹ ãģij +׾ ת +ìł Ī +ëĬĶ ëį° +Ø® ÙĬر +à¸ķà¹īà¸Ńà¸ĩ à¸ģาร +ĠÙĦ Ø£ÙĨ +Ġch á»ĭ +ÙĪ Ø© +à¹ĥ ส +ë¶Ģ íĦ° +íķĺ ë©´ +ữ u +à¹Ģหม ืà¸Ńà¸Ļ +б еÑĢ +ĠìĿ´ ìļ© +ĠÑģ еб +wiÄĻ ks +Ġ׳ ×¢ +ÑĤ ÑĥÑĢ +Ġngh Ä© +ש ×ķ×ĺ +ti ÄŁi +Ġde ÄŁi +×IJ ×ij +Ġ×ŀ ×ŀ +ãĥĹ ãĥŃ +wa ÅĤ +à¸Ī ึà¸ĩ +Ø® دÙħ +×IJ ×Ŀ +Ä±ÅŁ ı +cz Äħ +ר ×ĵ +ĠÑĢ Ñĥб +خر Ùī +ãģ® æĸ¹ +Ġд енÑĮ +×Ĺ ×Ļ×Ŀ +еÑĤ е +ëĤ ľ +×IJ ×Ĵ +×¢ ×ķר +ë³ Ħ +åIJĮ ãģĺ +ãĤ ² +ר ×ļ +×ķש ×IJ +ìľ ¡ +ا Ø® +צ ×Ļ×Ķ +á»± a +ãģĪ ãģ¦ +ש×Ķ ×ķ +ан ÑĤ +ลา à¸Ķ +ин г +ë¡ ł +اع د +ÙĪ Ø³Ø· +Ġв оп +Ġвоп ÑĢоÑģ +Ùħ ÙĬÙĨ +à¸Ħ à¸ĩ +×Ļר ×Ļ×Ŀ +c ów +ê² © +Ġê·¸ 룰 +Ġì§ Ħ +Ġש ׾×Ķ +à¹Ģร ิà¹Īม +à¸Ĭ à¸Ńà¸ļ +д еÑĤ +ÑİÑī иÑħ +à¸ļ à¸Ńà¸ģ +æĢĿ ãģĦ +ع ÙĬد +ס ×ŀ +×Ĵ ×Ļ×¢ +צ ×ĵ +ب ات +ĠëͰ ëĿ¼ +à¸Ī ัà¸ĩ +ãģłãģij ãģ§ +×¢ ×Ļר +ĠÑĩ ел +ĠÑĩел ов +ĠÑĩелов ек +ãĥĥ ãĥģ +à¹Ģà¸ģ ีà¹Īยว +à¸Ķ ิ +Ġפ ×¢ +×Ļ×ŀ ×Ļ +ë° ĺ +Ø® ار +×ij ×Ļת +×¢ ×Ļ×Ŀ +ü yor +ãĤģ ãģ¦ +к лад +Ġ à¸Īาà¸ģ +à¹Ģà¸Ħ ย +ส à¸Ńà¸ĩ +à¹ģ à¸Ħà¹Ī +ẫ u +หà¸Ļ ัà¸ĩ +ש׾ ×ķ×Ŀ +اÙĨ ÙĬØ© +åĩº ä¼ļ +åĩºä¼ļ ãģĦ +à¸ł าย +à¸ļา à¸Ĺ +à¸Ĭา ว +mu ÅŁ +Ġ׾ק ×ij׾ +ãĤ· ãĥ£ +Ġİ ÅŁ +×Ĵ×ĵ ×ķ׾ +ج عÙĦ +ë³ Ģ +ยิ à¹Īà¸ĩ +à¸Ļ าย +à¸Ļ ีà¹Ī +วิ à¸ĺี +ãĤī ãģªãģĦ +ëł Ī +Ġ문 ìłľ +Ġ à¸ģ +à¸Ĺำ à¸ĩาà¸Ļ +à¹Ģว à¹ĩà¸ļ +ÑĦ е +楽 ãģĹ +สำ à¸Ħ +สำà¸Ħ ัà¸į +ر Ùħ +ãģķãĤĮ ãģ¦ +Ġоб ла +ר×IJ ×Ļ +หม à¸Ķ +ÙĨ ÙĬØ© +ли н +Ġe ÄŁ +it im +ëł ¹ +ص اÙĦ +ÅĽ l +à¸ľ ิà¸Ķ +ãĥŀ ãĥ³ +åħ¥ ãĤĮ +à¹Ģà¸ķ à¸Ńรà¹Į +ار ÙĬ +ĠÐ ¦ +d ür +ส วย +ë¦ ½ +رÙĥ Ø© +Ġh ã +×Ļת ×Ķ +à¸Ĥ à¸Ļา +à¸Ĥà¸Ļา à¸Ķ +à¸Īำ à¸Ļ +à¸Īำà¸Ļ วà¸Ļ +ש ×ķ×§ +Ġд ом +ì± ħ +ãģĭ ãģij +פ ×ķ׾ +à¸Ĭ าย +Ñģ моÑĤÑĢ +Ñģл Ñĥж +ש ×IJ׾ +кÑĢÑĭ ÑĤ +Ġìŀ ĺ +é«ĺ ãģĦ +ĠÑĢ Ñĥк +ÙĨ ص +д ав +ưỠ¡ +ưỡ ng +ر اÙħ +×Ļ׳ ×Ļ×Ŀ +ãĥ© ãĥ¼ +ëĦ ¤ +Ġت ع +l ke +好 ãģį +æĮģ ãģ¡ +Ġë§ İ +Ġy ük +ĠÑģоÑģÑĤ ав +енÑĤ ÑĢ +pe ÅĤ +à¹Ģà¸Ľà¸¥ ีà¹Īย +à¹Ģà¸Ľà¸¥à¸µà¹Īย à¸Ļ +íı ī +ãĤĦ ãģĻ +×Ĺ ×ĸ +×ijר ×Ķ +ë£ ¨ +ìĶ Ģ +بØŃ Ø« +à¹Ģà¸ķ à¹ĩ +ów i +ب Ùĩ +ãģį ãģ¾ãģĻ +Ġ×¢ ×ŀ +×Ĵ ×ķ׾ +ез д +ÙĬÙģ Ø© +สà¸Ļ à¹ĥà¸Ī +Ġת ׾ +Ñı Ñī +Ġس ÙĨ +ĠÙĪØ§ ØŃد +ĠÑģ м +lad ı +ı ld +×Ļר ת +ีย à¸Ļ +ת×Ĺ ×ª +Ġж из +à¸ŀ ั +à¸ŀั à¸Ĵ +à¸ŀัà¸Ĵ à¸Ļา +à¸Ĭ ิ +ا Ø®ÙĦ +ãģ£ãģ¦ ãģĦãģŁ +รั à¸IJ +ãĤģ ãĤĭ +à¹Ĥ à¸ģ +ĠT á»ķ +Ġh akk +ر Ùģ +ìł Ģ +Ñģ об +ãģª ãģijãĤĮãģ° +Ùĩ ÙĪ +Ġë² ķ +ãĤ Ĩ +ĠاÙĦس عÙĪØ¯ +Ġ×IJ תר +Ø§Ø º +Ġ׾ ×ĵ +à¹ģ à¸ķ +à¹ģà¸ķ à¹Īà¸ĩ +íĮ Į +Ñĥп иÑĤÑĮ +à¸ŀืà¹īà¸Ļ à¸Ĺีà¹Ī +×ij ת×Ļ +à¹ĩ à¸ģ +ÅĤ at +Ġê°ľ ìĿ¸ +ìłķ ë³´ +ÑĤ ал +Ġgü ven +Ġİ l +Ġê° ģ +Ġب ت +×ŀ ×ķ׳×Ķ +ĠاÙĦØŃ ÙĥÙĪÙħ +ÙĤ ات +à¹ģ à¸ģà¹Ī +ห าà¸ģ +н ÑĮ +à¸Ľ รัà¸ļ +มา à¸ĵ +Ġне Ñģк +ĠØ ¶ +สม ั +สมั à¸Ħร +ãģĮ ãģĤãĤĬ +м еÑģÑĤ +Ġ×IJ צ׾ +Ġкомп ани +ס ר +ÙĬÙħ Ø© +ĠÑħ оÑĢо +ĠÑħоÑĢо ÑĪ +Ġ×Ļ ×ķ×ĵ +ü s +×Ĵ ×Ļש +à¸ļ à¸Ĺ +تÙĨ ظ +ว าà¸ĩ +ม หา +Ġ׼ ×ķ׾ +à¸Ĥ à¹īาà¸ĩ +ë° ľ +г од +д ан +ãģĭãĤĤãģĹãĤĮ ãģ¾ãģĽãĤĵ +ãģĵ ãģ¡ãĤī +ãĥIJ ãĤ¤ +ece ÄŁi +دÙĬ دة +ÙĨ Ùī +Ġëĭ¤ ìĿĮ +ว ี +غ ا +ли з +à¹Ģà¸Ķ ิ +à¹Ģà¸Ķิ ม +ĠÙĬ ست +Ġy ılı +ko ÅĦ +ãģ§ãģĹãĤĩãģĨ ãģĭ +ãģĤ ãģª +ãģĤãģª ãģŁ +ÑĨ ен +ĠÙĪ Ø² +×IJ ×Ļש +à¹Ī à¸Ń +ر ØŃ +ê´ ij +ÑĢа ÑģÑĤ +Ġ×Ķ ×ľ +ãģĹãģ¦ ãĤĤ +×ŀר ׼ +×ŀר׼ ×ĸ +éģķ ãģĦ +ãģŁ ãģı +ĠÑģ Ñĥд +в еÑģÑĤи +ĠíķĦ ìļĶ +ãĥķ ãĤ§ +ÑĤелÑĮ но +à¹Ģà¸ŀ ืà¹Īà¸Ńà¸Ļ +ÅĤu ż +à¹Ģà¸Ķิà¸Ļ à¸Ĺาà¸ĩ +ש ×ķר +Ġ×ŀ ×ĵ +×ķ×¢ ׾ +ÙĦ اÙħ +à¹Ħ à¸ĭ +л ей +кÑĥ ÑĢ +Ạ¢ +à¸Ĺ าà¸Ļ +ì§ ij +ĠгоÑĢ Ð¾Ð´ +ר ס +׾ ×ķ×Ĵ +mas ını +Ġл ÑĥÑĩ +ล à¹Īา +ìļ ¸ +ש ×ĺ +ĠÐĺ н +í Ĥ¤ +ÙĪÙĦ ا +ìķ ł +ĠØ£ÙĬ ضا +Ùĥ ار +ĠاÙĦت ع +ส ูà¹Ī +ãĤ ¼ +×ij ×Ļ×IJ +ย à¸ģ +ĠØŃ ÙĤ +ر بÙĬ +ãģĺãĤĥ ãģªãģĦ +รัà¸ģ ษา +Ñħод иÑĤ +à¸ķ à¸Ńà¸ļ +׳ ×ĺ×Ļ +ĠاÙĦÙħ ج +تÙħ ع +ов аÑĤÑĮ +ÙĦ ÙĬÙĨ +×Ļ×ŀ ×ķת +Ġm ù +n ÄĻ +Ġد ÙĬ +׼ ש×Ļ×ķ +Ġhi ç +ë ijIJ +ÙĪ Ø§Ø¡ +ÙĪ Ø· +ĠاÙĦ بÙĦ +à¹ģม à¹ī +×§ ×ķת +ÙĪØ¬ د +å§ĭ ãĤģ +ÙĬ ئة +Ġë§ ¤ +ص بØŃ +פ ×IJ +г оÑĢ +ס ×Ķ +بÙĬ ÙĤ +ย าà¸ģ +Ġн ад +ÙĬ Ùij +Ġب ÙĪ +ס ×ķר +Ùħ ÙĥاÙĨ +ר ×ij +×Ĵ ×ĸ +צ ת +b ilit +л аг +ĠN go +×IJ ×ķר +à¸ķ à¸Ļ +íĬ ¹ +à¸Ĺีà¹Ī à¸Ķี +à¸Ľà¸£à¸° à¸Īำ +ов ание +ãģĦ ãģ¤ +ãĥĥãĤ¯ ãĤ¹ +åIJĪ ãĤı +åIJĪãĤı ãģĽ +×Ļ׳ ×ķ×Ļ +ạ y +Ø« ÙĤ +ĠпÑĢ Ð¾Ð± +ĠпÑĢоб лем +ÅŁ eh +ÅŁeh ir +ع ادة +اÙĨ ÙĪÙĨ +à¸ķัว à¹Ģà¸Ńà¸ĩ +ì¶ ķ +ı lan +б ан +ãĥ³ ãĥī +à¸Ī ี +Ġ×Ķש ׳×Ļ +п оÑĤ +×ķ׾ ×Ļ×Ŀ +ล ัà¸ļ +ĠÑį ÑĤи +×ij×§ ש +ë¹Ħ ìĬ¤ +à¸Ńยà¹Īาà¸ĩ à¹Ħร +×Ļ׾ ×Ļ +à¹ĥà¸Ĭ à¹Ī +ĠاÙĦ ÙĥÙĦ +ãĥļ ãĥ¼ãĤ¸ +ص Ø© +ÑĤи ÑĢ +ãĤĵ ãģ© +зÑĭ к +wy ż +Ùĩ ÙĬ +ĠÙħ ÙĦÙĬ +Ġвид е +ظ اÙħ +دا ÙĪÙĦ +×ŀ ת×Ļ +Ġs ık +à¹Ģà¸ķิ ม +ãĤ¢ ãĤ¤ +ка Ñħ +צ ×Ļ׾ +à¹Ģà¸Ĭ à¹Īà¸Ļ +м аг +маг аз +магаз ин +à¸Ľ ั +à¸Ľà¸± à¸Ī +Ġש ×Ļר×ķת +ีย ม +ãĥĸ ãĥ« +Ġد ÙĪÙĦ +קר ×Ļ×Ŀ +Ùĩ Ùı +ов о +Ġü ret +د ÙĪÙĨ +à¹ģà¸Ļ ว +à¹Ģà¸Ļ ืà¹īà¸Ń +ĠÑĦ оÑĤ +ãĥ ĺ +ãģ¤ ãģĭ +Ñı Ñģ +ĠíķĺëĤĺ ëĭĺ +ائ ع +Ġп лаÑĤ +ìĺ Ī +Ġdost ÄĻp +ÙĪØ¬ Ùĩ +Ġ×Ķ ×Ĺ×Ļ +׳ ×Ļ×§ +д ей +í ĽĦ +ı y +بØŃ ر +à¹Ģส ริม +Ġ׾ ×Ĵ +ذÙĩ ب +ج ÙĬÙĦ +رÙĥ ز +Ġë ħ +Ġëħ ¸ +פ×Ļ׾ ×ķ +ãģ¾ ãģļ +iri ÅŁ +ĠÙĥ ÙĬÙģ +Ġ×ij צ +Ġêµ IJ +ÑĢоÑģ Ñģ +ĠØ´ ÙĬ +Ġiç er +×Ĵ ×ķ×ij×Ķ +мен но +×¢ ×ij×Ļר +×ķ×ŀ ×Ķ +ãĤī ãģĹãģĦ +ãģ ¼ +Ñī ин +è²· ãģĦ +جÙħÙĪØ¹ Ø© +Ġdön em +Ġ×ij ×IJר +в еÑģÑĤ +×ķר ×ķת +س Ùģ +à¹ģà¸Ĺ à¸Ļ +Ġд окÑĥменÑĤ +Ġا ÙĬ +ج اÙĨ +צ×ķ×¢ ×Ļ +ĠоÑģ об +ĠاÙĦÙħ س +ÑĢаР± +à¸ł ู +à¸Ķ าว +л екÑĤ +ع ÙĤ +×ķ×ĵ ×ķת +Ġol u +Ġolu ÅŁtur +ãģ¾ ãģ¾ +ед ин +à¹Ģ à¸Ńà¸ģ +ãĤµ ãĤ¤ +ëĦ Ī +Ø· ÙĨÙĬ +Ø· ÙĤØ© +ĠÐł аз +ÙĦ Ùij +Ñĩ ем +Ġ׾ ×ĺ +สั à¹Īà¸ĩ +سر ائÙĬÙĦ +Ġפר ×ĺ×Ļ +д еÑģÑĮ +Ġ׳ ׼ +اÙĨ ب +ÙĬا Ø© +Ùħ بر +Ġk ı +à¸Ľ à¸ı +à¸Ľà¸ı ิ +à¸ļั à¸ķิ +׳ ת×Ļ +ìĨ ¡ +ر اب +à¹ĥ à¸ķ +à¹ĥà¸ķ à¹ī +×Ļ׳ ת +ÙĪ ÙĬر +Ġ×Ķ×ŀ ×Ļ +ей ÑĩаÑģ +×§ ×ķ×ij +در اس +ĠÙħ ÙĤ +رÙĬ ÙĨ +Ø® اص +ãģĬ éĩij +Ġج دا +ãģĨ ãģ¡ +ëħ ¸ +ır ım +æ§ ĺ +ãģ« å¯ +ãģ«å¯ ¾ +ÑĨ ев +Ġv ard +ĠÐIJ н +e ÄŁ +ÑģÑĤв енно +Ð ¨ +س د +à¸ģ ุ +à¹ģà¸ľ à¸Ļ +รูà¹ī ส +รูà¹īส ึà¸ģ +ات ØŃاد +Ñij ÑĤ +×Ĺ ×ķ×§ +ãģĻ ãģIJ +Ø· ÙĦاÙĤ +Ġ×§ ×ķ×ĵ +à¹ĥà¸Ĭ à¹īà¸ĩ +à¹ĥà¸Ĭà¹īà¸ĩ าà¸Ļ +ãĥ¼ãĤ ¿ +Ġs ür +ÑĢ Ð¾Ðº +ë³ ij +สมา à¸Ĭ +สมาà¸Ĭ ิà¸ģ +ãĥķ ãĥ¬ +è¾¼ ãģ¿ +ãĤ» ãĥ³ +Ġê°Ģ ì§Ģ +à¸ľ à¹īา +ÑįÑĤ омÑĥ +иÑĤ ел +à¸ł ั +ภij +ãĥĸ ãĥ© +×Ľ×ª ×ķ×ij +׳ ×Ŀ +ен нÑĭе +×¢ ×¨×Ľ×ª +Ġì Ĥ +ĠìĤ ´ +à¸Ĥ à¹īา +׳ ×ķס +ãĥ¬ ãĥĵ +ÑĢ ÐµÑģ +à¹Ģล à¸Ĥ +Ø« اÙĦ +ìĹ Ĩ +ĠÑĩ аÑģÑĤ +า ศ +ãĥª ãĤ¢ +u ç +×Ļ׼ ×ķת +ล à¹īาà¸Ļ +i ë +ãĤ¸ ãĤ§ +à¸Ī à¸Ń +ÙĪ ØŃد +×Ļצ ×ķ×ij +Ġ×ij ש׾ +ок о +ض Ø© +ذ ر +ĠÑĥ д +İ L +×ķצ ×Ļ×Ŀ +×ĸ ×ŀף +à¸Ľ à¸ģ +íķĻ êµIJ +س اÙħ +à¹Ħ à¸Ķ +ละ à¹Ģà¸Ń +ละà¹Ģà¸Ń ีย +ละà¹Ģà¸Ńีย à¸Ķ +ả y +аÑĨи он +ãĤ¹ ãĤ¯ +פ ×ķס +ร à¹Īาà¸ĩ +ен нÑĭй +ع ÙĨ +عÙĦ ÙĨ +ائ Ùģ +d ÄĻ +ؤ ÙĪÙĦ +׾×ķ ×ķ +Ġ×ij ש×ij +ä»Ĭ åĽŀ +ĠاÙĦج ÙĨ +د اد +wa Äĩ +ãĥª ãĥ³ +ĠìŀIJ ìĭł +اÙĨ ÙĬا +ãĥ¡ ãĥª +ÙĦ ÙĪÙĨ +à¸Ĺ à¹Īà¸Ńà¸ĩ +à¸Ĺà¹Īà¸Ńà¸ĩ à¹Ģà¸Ĺีà¹Īยว +اÙģ ÙĬ +Ġли ÑĪ +Ùħ ÙĬØ© +оÑĤ веÑĤ +Ñĩ ин +à Ĭ +ãĥ¡ ãĥ³ +å® Ł +éļĽ ãģ« +ĠÑĢаР¹ +ãĤ¦ ãĥ³ +×Ļר ×ķש +×Ļר×ķש ׾×Ļ×Ŀ +ม ะ +Ġar a +каз аÑĤÑĮ +à¸ķ ัà¸Ķ +ÑĥÑİ ÑĤ +Ġü st +×Ĵ ×ķ×ij +×Ĵ×ķ×ij ×ķת +mal ı +ег од +егод нÑı +اÙģ ÙĤ +à¸Ĭ à¹Īà¸Ńà¸ĩ +Ġö zellik +×Ļצ ×ķר +Ġmi ÄĻd +Ġili ÅŁ +Ġна Ñħод +×¢ ×ĸר +׾ ×Ľ×ª +ÙĨت اج +ĠÑģ ем +à¸Ī à¹Īาย +à¸ķร ว +à¸ķรว à¸Ī +פר ×ķ +à¸Ĥ ัà¸ļ +ãģ ŀ +Ġп ло +к олÑĮ +×ŀ×¢ ×ĺ +íķĺ ìĭľ +jÄħ ce +ÙĨ اÙĨ +ลี à¸ģ +н ÑĥÑĤ +Ġоб ÑĢаз +Ùĥ بر +ĠاÙĦÙĪ Ø·ÙĨ +ãģķãģĽ ãģ¦ +ÙĤ اء +×ŀ×ĵ ×Ļ׳ +y ü +פ ×Ļת +׳ ×ķף +ÙħÙĨ ظ +หà¸Ļ ัà¸ģ +ìŀ Ī +ãĤ« ãĥ¼ãĥī +ع ÙĨÙĬ +п од +ض اء +à¸Ļ à¸ķà¹Į +×ŀש פ +ว à¹Į +ר ×ķ×§ +ส ืà¹Īà¸Ń +פק ×Ļ×ĵ +ãģªãĤī ãģªãģĦ +ĠìŬ 룬 +ÙĦ ج +Ñī иÑĤ +ãĥĥ ãĤ· +ÙĦÙĬ س +ĠÙĦ Ùħا +ìł ij +×ij ×Ļף +ãĥģ ãĤ§ +Ġgü ç +Ġch ứ +×ķצ ×IJ +קר ×ij +à¹Ĥ à¸ŀ +оÑĩ но +סק ×Ļ +ש׾ ×Ŀ +صر Ùģ +ĠL Ãł +×¢ ×Ļת +á» · +à¹Ĥ à¸Ńà¸ģ +à¹Ĥà¸Ńà¸ģ า +à¹Ĥà¸Ńà¸ģา ส +Ġ×Ķ ×ĵ×ijר +à¸Ļั à¹Īà¸Ļ +ز ر +нак о +íļ į +ãĤĤ ãģ¡ +ãĤĤãģ¡ ãĤį +ãĤĤãģ¡ãĤį ãĤĵ +اÙħ ت +عد اد +и нÑĭ +ÅĤy w +à¸Ħ à¸ĵะ +à¸Ĺ ะ +kt ör +×Ļ×Ĺ ×Ķ +Ġм е +Ġме ÑģÑı +׳×Ķ ×Ĵ +ĠÑģ ÑĥÑīеÑģÑĤв +à¸Ļ ัà¸Ļ +ÑĦ ÑĦ +ек ÑĤив +عÙĦÙĪÙħ ات +б Ñĥд +à¸Ļัà¸ģ à¸ĩาà¸Ļ +หà¸Ļà¹īา à¸Ĺีà¹Ī +ÙĤÙĬ ÙĤ +ãĤ· ãĥ³ +ãģ« éĸ¢ +×IJר ×Ĵ +ĠпÑĢ Ð¾ÑĤ +ĠпÑĢоÑĤ ив +ĠìŀĪ ìĸ´ +ÙĤÙĬ ÙĤØ© +ìĹ ĩ +k ür +ãģ«ãģªãĤĬ ãģ¾ãģĹãģŁ +Ġде ÑıÑĤ +ĠдеÑıÑĤ елÑĮ +פ×ķר ×ĺ +à¸Ł à¹īา +à¹Ģ à¸ł +ĠавÑĤом аÑĤ +×ĸ ×Ļ×§ +Ġold uk +ع اÙħ +ĠÑĤ оÑĢ +yrı ca +ê Ì +ãĤŃ ãĥ³ãĤ° +ãģ« ãģ¨ãģ£ãģ¦ +à¹Ģà¸ī à¸ŀ +à¹Ģà¸īà¸ŀ าะ +ãģ¯ ãģļ +×ŀ ×IJ×Ļ +สะ à¸Ķ +สะà¸Ķ วà¸ģ +ìľ¼ ë©° +à¸ģ ี +ภ¬ +Ġ×¢ ×ķש +à¸łà¸² ษา +à¸Ĺ ัà¸Ļ +ac akt +acakt ır +اع دة +ĠÑĥÑģл Ñĥг +ס ר×ĺ +×ķ×ŀ ×ķת +×Ķ ×ķר +×ŀ ×ķ×ij +×ŀ×ķ×ij ף +سÙĬ اس +اتÙģ Ø§ÙĤ +×Ķ ×¦×ľ +Ùħؤ س +Ġp ó +Ġк ни +×Ļ׼ ×ķ׾ +à¹Ģหล ืà¸Ń +׼׾ ׼ +׳ ×ĸ +ÑĪи е +r ès +ĠاÙĦØŃ ÙĤ +лÑı ÑĢ +ห à¸į +หà¸į ิà¸ĩ +ר×Ĵ ×Ļש +à¹Ģส à¹īà¸Ļ +ש×ij ×ķף +ô tel +ап ÑĢ +апÑĢ Ð¸Ð¼ÐµÑĢ +اب ÙĦ +ĠÑĢаз виÑĤ +Ġп олÑĮз +ĠС еÑĢ +×ķ×ij ×Ļ +r óż +ìĭ Ń +ãĤ¯ ãĥĪ +ãģĹ ãĤĪãģĨ +à¸ģร ม +ØŃ ÙĥÙĪÙħ +à¹Ĥ à¸ļ +à¸Ĺ à¹īาย +ĠM á +ĠÑĤ Ñĭ +à¸Ħร ัว +ÑĢÑĥ б +ạ p +Ġm ÅĤ +ĠmÅĤ od +Ġgör Ã¼ÅŁ +Ġgeli ÅŁ +ươ i +×ŀש ×§ +ÙĢÙĢ ÙĢÙĢ +รา ว +ãģĹãģ £ +ãģĹãģ£ ãģĭãĤĬ +ĠÐļ он +Ġk ê +à¹Ĥà¸Ĺ ร +èIJ½ ãģ¡ +åĩº ãģ¦ +ล ัà¸ģษ +Ġ×Ĵ ×ij×ķ×Ķ +ãĥĻ ãĥ« +ê±° ëĤĺ +ë§ IJ +×Ļ׾ ×ĵ×Ļ×Ŀ +ĠëĦ Ī +×ŀר ×Ļ +ร ส +ãĥŃ ãĥ³ +и ло +ноÑģÑĤÑĮ Ñİ +×ĸר ×Ĺ +п он +Ġ×Ķש ׾ +ê²ł ìĬµëĭĪëĭ¤ +Ġki ÅŁ +ĠÐļ и +ว ร +د اع +ÅŁ im +ÙĨ Ùij +в аÑĤ +را Ùĥ +ب اÙĦ +ид е +Ġ×Ķ×ŀ ×Ĺ +ìĸ µ +تÙģ Ø§Ø¹ +Ø£ ت +ëĬ ĺ +ש ×Ļת +ست Ùħر +ĠÑĦ ак +ĠاÙĦØ£Ùħ رÙĬ +ëŀ ¨ +اس Ùħ +Ġa ÄŁ +Ġç ev +Ùĥ ÙĪØ± +ãģķ ãģ¾ +Ġç öz +Ġر س +Äħ da +สà¸Ļ ุ +ãģĹãģ¦ ãģıãĤĮ +н Ñİ +leÅŁ me +ãĤª ãĥ³ +ãģ¨ ãģªãĤĬ +ava ÅŁ +×ĺ ×Ļ×ij +ØŃ ض +×ķצ ×IJ×ķת +ÙĨ ÙħÙĪ +ı t +ĠÑħ а +ĠÑħа ÑĢак +ĠÑħаÑĢак ÑĤеÑĢ +Ġd ÅĤ +ãĥĹ ãĥ© +à¸Ĭ ุม +à¹Ī à¸Ńà¸Ļ +×ķ×ij ׾ +Ñģ ол +×ĵ ×Ĵ +аÑĢ Ð°ÑĤ +n ivers +Ġgerçek leÅŁtir +ĠاÙĦ ÙĦÙĬ +ระ ยะ +ĠÙħ ختÙĦÙģ +Ġgö nder +Ùģ Ø§Ø± +do ÄŁ +doÄŁ an +ص ÙĦاØŃ +Ġyay ın +ãĥĨ ãĥ³ +รว à¸Ī +×Ļ×Ĺ ×Ļ×ĵ +ünk ü +ÑĨи алÑĮн +à¸ļ ู +ม ุ +h ä +Ø® Ùģ +å¢ Ĺ +å¢Ĺ ãģĪ +еÑĩ но +ĠاÙĦس ÙĨ +à¸Ĥ าว +im di +Ð « +à¸Ļà¸Ńà¸ģ à¸Īาà¸ģ +à¸ļา ล +ת ש +Ġdüzen le +мÑĭ Ñģл +ãģı ãģª +ż u +Ġwsp óÅĤ +Ġн аз +ınd aki +تر Ø© +ÅŁ ek +Ġö d +ĠÙĪ Ùĥ +Ġпозв олÑı +Ġת ×ķ׼ +ÙħÙĨ تج +ë§ ī +ĠاÙĦØ« ÙĦاث +аÑĨи Ñİ +ÙĪØ± ÙĪ +Ñĭв аеÑĤ +خص ص +ĠاÙĦÙģ ÙĦ +ĠاÙĦÙģÙĦ سطÙĬÙĨ +Ø¥ جر +إجر اء +اÙĨت Ø® +اÙĨتخ اب +ار ÙĬØ© +×ķ Ö +Ø¢ ÙĨ +×ŀ×¢ ×ķת +Ġм ал +Ġ×IJ ×Ĺ +à¸Ĺ à¹īà¸Ńà¸ĩ +ze ÅĽ +Ġë§Į ëĵ¤ +رÙĬ ع +äºĭ ãĤĴ +à¸ļริ หาร +׾ ×ŀ×Ļ×ĵ +Ġм Ñĥж +ت رÙĪ +ĠباÙĦ Ø¥ +פ ×Ļ×§ +ز ÙħØ© +ĠÃ¶ÄŁ renc +ãĥ ¶ +اÙħ عة +×§×ij ×ķצ +×ŀ ׳×ķת +رÙĬ Ùħ +Ġо каз +ãģłãģij ãģ© +Ġh ız +Ġש ×IJת +ãĤ¢ ãĥ¼ +Ġmożli wo +ìĦ ¼ +ÙĪ Ø§Ø¨ +ог ÑĢаÑĦ +Ġعبد اÙĦ +ãĤĴ è¡Į +ب ÙĬÙĦ +Ġİ ç +ย าย +ĠÑĥ ÑĩаÑģÑĤ +ÑĦ еÑģÑģ +ÑĦеÑģÑģ иона +Ạ¤ +ÙĨ ÙĬÙĨ +عد ÙĦ +สร ร +دÙĬ ÙĦ +×ij ×Ļ×§ +czy ÅĤ +ÑĢом е +Ġм ед +ìĻ Ķ +ãĥ© ãĤ¤ãĥ³ +ĠÑĤ еп +еÑĢ ÑĮ +i ÄŁi +в ели +ÑĢи ÑģÑĤ +ס ×ķפ +×ŀ׾ ×Ĺ +ĠاÙĦØ¥ ÙĨ +Ġ׾×Ķ ×© +è¶Ĭ ãģĹ +ĠÑĢ Ñĭ +×ķ×IJ ר +رÙĩ اب +פ ×ķ×IJ×Ļ +ĠгоÑģ Ñĥд +ĠгоÑģÑĥд аÑĢ +ĠгоÑģÑĥдаÑĢ ÑģÑĤв +ĠاÙĦØ£Ùħ ÙĬر +Ùħ ج +à¹Ģหม าะ +ÑĢ ÐµÐ² +à¸Ĭี à¸ŀ +ãĥķ ãĥĪ +иÑĩ но +ĠاÙĦÙħ ؤ +Ġi ht +íħ ľ +د ÙĨÙĬ +ر ص +ла ÑģÑĤ +à¹Ģหล à¹Īา +ılı r +ร à¸ĵà¹Į +×ŀש ×Ļ×ļ +Ġd á»ĭ +Ø·Ùģ Ø§ÙĦ +×ĺ ×ķף +Ġ×ij ×Ļ׳ +ãģ¾ ãģ£ãģŁ +лож ениÑı +تØŃ ر +ب اØŃ +à¹Ģส ืà¹īà¸Ń +ãģĻ ãģĶ +lt ür +à¸ĩ าม +Ġt ü +ĠпÑĢ Ð¸Ð¼ +ĠпÑĢим ен +Ġhay at +ëĥ IJ +ëĭ Į +׳×Ļ ×ķ +вед ен +ìħ ¨ +à¸Ī ัย +à¸ģà¹Ī à¸Ń +Ġв од +оÑģÑĤ оÑı +н аÑĤ +à¹ģ หล +سÙħ ÙĬ +à¸Ķำ à¹Ģà¸Ļ +à¸Ķำà¹Ģà¸Ļ ิà¸Ļ +w ód +ö yle +ãĥĢ ãĤ¤ +ÑĪи й +меÑī ен +ãģĹãģ¾ ãģĨ +ãĥī ãĥ© +ÙĪØ¶ ØŃ +à¸Ńà¸Ļ ุ +ĠاÙĦ اجتÙħاع +laÅŁ ma +à¸Ħ à¸Ńà¸Ļ +×ŀר ×Ļ×Ŀ +ÙĨ اÙħج +שר ×ķת +اÙĦ Ø£ +Ġksi Äħż +Ġа н +ÑĢаР¹ +اÙĩر Ø© +×ŀ×ĵ ×Ķ +ä¸Ģ ç· +ä¸Ģç· Ĵ +ä¸Ģç·Ĵ ãģ« +ÑĢиÑĤ оÑĢ +d ıkl +à¹ģ à¸ĸ +à¹ģà¸Ĥ à¹Īà¸ĩ +екÑĤ оÑĢ +×ŀס ×¢ +ÑĢак ÑĤи +u ÄŁu +×ķ×ij ת +สู à¸ķร +ĠçalÄ±ÅŁ m +ĠçalÄ±ÅŁm alar +Ġа на +ãĥĽ ãĥ¼ãĥł +Ġböl üm +Ġب ص +ол оÑģ +ĠìķĬ ëĬĶ +à¹Ī ะ +ÙĪ ØªØ± +ä¹ Ĺ +ست خداÙħ +פ×Ļ ×Ļס +פ×Ļ×Ļס ×ij +פ×Ļ×Ļס×ij ×ķ×§ +Ġк ÑĢаÑģ +ли к +رÙĬ ØŃ +×ŀש ׾×Ķ +à¹Ģย ีà¹Īย +à¹Ģยีà¹Īย ม +в иÑģ +ом н +ÄŁ un +ãĥŃ ãĥ¼ãĥ³ +Ø£ تÙĬ +à¸ķร ี +çͳ ãģĹ +تÙħ ر +ìĹ ĪìĬµëĭĪëĭ¤ +ĠÙĪ ØºÙĬر +red ni +ĠاÙĦص Ùģ +Ġна ÑģÑĤоÑı +ĠнаÑģÑĤоÑı Ñī +à¸ķ รา +ĠÑĥÑģл ов +ĠÑĥÑģлов иÑı +ÑĨ еп +×Ķ ×Ĺ׾×ĺ +Ø· ÙĬع +ĠB akan +ĠاÙĦ رÙĪ +илÑĮ но +Ġм еÑĤ +à¸Ķ à¸Ńà¸ģ +ãģĭãĤī ãģªãģĦ +Ġпо ÑģÑĤоÑı +ĠпоÑģÑĤоÑı н +ĠÑĩ аÑģ +ü c +wr ó +б ÑĥÑĢ +ãĥIJ ãĥĥãĤ¯ +ãĥ©ãĥ³ ãĥī +Ġо гÑĢ +สั à¸į +สัà¸į à¸įา +มั à¹Īà¸Ļ +à¸Ħ à¸Ńม +al ık +Ġн ед +üm üz +ĠÅĽ wie +é rio +×Ļ×IJ ×Ķ +دÙħ ات +ı rl +ĠоÑĤ з +ĠоÑĤз Ñĭв +ä»ĺ ãģį +Ġkaż de +мин иÑģÑĤ +ãĤ° ãĥ« +ë° ĸ +ез н +اÙĦ Ùģ +Ġש ק׾ +Ùħ ض +ãĥĿ ãĥ¼ãĥĪ +ÙħÙĨ ت +ÙĤÙĬ اÙħ +Ø´ ÙĨ +×Ļר ×ķ×¢ +ãĤŃãĥ£ ãĥ³ +доÑĢ Ð¾Ð² +×ŀ ×Ļת×Ļ +ÙĪÙĦ ÙĪØ¬ +Ùĥ اÙģ +ĠÑĢаз лиÑĩ +иÑĤ еÑĤ +н олог +ลà¸ĩ à¸Ĺุà¸Ļ +Ġyak laÅŁ +ãĥ¬ ãĤ¤ +ê²ł ëĭ¤ +æ±Ĥ ãĤģ +رÙĪ Ùģ +Ġí Ĭ +ĠíĬ ¹ +ãģ£ ãģıãĤĬ +à¸Ħวาม à¸Ħิà¸Ķ +×Ķ ×Ļס×ĺ +Ø¥ ÙĤ +ãģ¦ ãģĦ +à¹Ĥ à¸Ĭ +ĠBü yük +ĠФ едеÑĢ +ÑĨи н +ÑĢов а +ĠاÙĦ اÙĤتصاد +Ġch á +à¸ĺ าà¸Ļ +ë¥ ł +à¹Ħ à¸ķ +ÃŃ pio +Ùĭ ا +Ġоб Ñıз +Ùĩ ج +Ġì¤ij ìļĶ +ãģ® ãģ§ãģ¯ãģªãģĦ +بار اة +ãĤ¤ ãĥ« +Ġн оÑĢм +á»ī nh +m ö +mö glich +ÑĨи п +ãĤ¢ ãĤ¯ +×Ķ ×Ļ +ÑĨи алÑĮно +ĠÅĽ wi +ت ÙĤ +ĠÑģÑĤо им +بÙĬ عÙĬ +Ġ׾ ש×ŀ +г лÑı +глÑı д +ãģ¦ ãģıãĤĮ +ÄĻd zi +à¸Ĥ ั +à¸Ĥั à¹īà¸Ļ +Ø· ÙĤ +ĠìĹ Ń +ãģ£ãģ¦ãģĹãģ¾ ãģĨ +ĠdeÄŁer l +ĠdeÄŁerl endir +Ġü lk +Ġмн ог +๠ĭ +ë¿ IJ +ĠУ кÑĢа +ÄŁ ini +Ġбез оп +Ġбезоп аÑģ +à¸Ńà¸Ńà¸ģ à¹ģà¸ļà¸ļ +Ø§Ø ¸ +ØŃد اث +л еÑĢ +×Ļ× ¥ +×Ļ׳×ĺר ׳×ĺ +lar ınız +ØŃÙĬ ØŃ +ż eli +à¸Ń ัà¸ĩ +à¸Ńัà¸ĩ à¸ģ +à¸Ńัà¸ĩà¸ģ ฤษ +ĠоÑĤ лиÑĩ +ั ส +ëŀ į +ож но +ãĤ¹ ãĥĿ +ĠÑħ оÑĩ +Ġк ап +еÑĩ ен +ØŃÙĦ Ø© +ÙĬا Ùĩ +на л +×ķצ ר×Ļ×Ŀ +Ġk ald +åĥ į +ĠاÙĦØ´ خص +Ġз на +Ġwz gl +ż ycz +ê° Ŀ +à¸ŀ ลัà¸ĩ +íģ ¼ +Ġö l +Ġb ụ +Ø´ Ùĩر +Ġз ам +Ġд ев +×Ļ×ĺ ת +تعÙĦ ÙĤ +ÙĪÙħ Ø© +ãĤĴ ä½ľ +ãģį ãģ¦ +í ĥĿ +ras ında +ãĤĴ æİ¢ +ĠÙħ باشر +راج ع +Ġв озд +ÙħØŃ ا +×ķש ר +ĠиÑģÑĤ оÑĢ +ม ัà¸ģ +t ıģ +Ø« ار +تر ÙĨت +à¹ģà¸Ĥ à¹ĩ +à¹ģà¸Ĥà¹ĩ à¸ĩ +п оÑĩ +Ġ×ij ×IJ×ķת +ë¯ Ģ +ëĿ¼ ëıĦ +à¸Ĭ ัà¸Ķ +ส à¸ķà¹Į +ãĥĭ ãĥĥãĤ¯ +ид енÑĤ +Ġг ÑĢÑĥпп +ت Ø® +Ạł +ย ืà¸Ļ +ย ัà¸Ļ +ó ry +T Ãľ +ãģĹ ãĤĥ +ĠпÑĢов ед +лÑı еÑĤ +Ùħ Ø® +ย à¸Ńม +×Ľ×ł×¡ ת +ĠاÙĦÙħ ÙĨت +Ġol mad +ר׼ ×ĸ×Ļ +Ġв ÑģÑĤÑĢ +ĠиÑģ Ñģлед +ÑĤвеÑĢ Ð¶ +بد ÙĪ +еÑĢ ÑĤ +ï» · +± ħ +สัม à¸ŀัà¸Ļà¸ĺà¹Į +ิ à¹Īà¸Ļ +צ ×Ļ×ij +wiÄĻ t +Ġì° ¸ +Ġz wiÄħz +سب ÙĪØ¹ +ãĥĥ ãĤ° +à¸Ľà¸¥ à¸Ńà¸Ķ +à¸Ľà¸¥à¸Ńà¸Ķ à¸łà¸±à¸¢ +ãĤĤ ãĤĬ +ÙĤد س +Ġspr z +Ġsprz eda +Ġist edi +Ġk hu +Ġд ен +Ġko ÅĦ +Ġ×ij ×Ĺ×Ļ +à¹Ģà¸Ĺ à¹īา +×ķס ×Ļ×£ +ãĥĭ ãĥ¥ãĥ¼ +ĠпÑĢед оÑģÑĤ +ĠпÑĢедоÑģÑĤ ав +à¹Ĥ à¸Ł +é v +ĠاÙĦص ØŃ +صØŃ اب +à¹Ģà¸Ī à¹ĩà¸ļ +вл ек +วั à¸ķ +à¸ĸ ุ +ãģĵãģ¨ãģĮãģ§ãģį ãģ¾ãģĻ +ÙĤÙĬ ÙĤÙĬ +×ķ׊ר +Ñĭ ÑĪ +ĠоÑĤ но +ĠоÑĤно ÑĪ +об илÑĮ +Ùģ ØŃ +ı nt +ınt ı +Ġ׾ ×ij×ĵ +í İĺìĿ´ì§Ģ +ãĥĬ ãĥ« +ĠÙħ ساء +×Ļ×ĺ ×ij +ÑĮ еÑĢ +ëĦ · +Ñĭ ÑĤа +ĠоÑĩ еÑĢ +à¸Ķ ืà¹Ī +à¸Ķืà¹Ī ม +ĠN gh +ت عب +ÙĦاÙĤ ات +×ķ׾×ķ×Ĵ ×Ļ×Ķ +ĠìĿ´ ê²ĥ +Ġ×Ķ ×ijר +ìľ µ +à¹Ģà¸Ħล ืà¹Īà¸Ńà¸Ļ +Ùĩ Ø© +à¸Īำ à¹Ģà¸Ľà¹ĩà¸Ļ +å¤ī ãģĪ +wi ÅĽcie +ch od +chod zÄħ +в ÑĢо +×ŀ×Ĺ ×Ļר +Ġy ı +Ġyı ll +ì¡ Į +à¹Ħ หว +ãģªãģı ãģª +Ġзав иÑģ +ĠìĺĪ ìĪĺ +Ùģ Ø° +á»§ ng +à¸ŀุ à¸Ĺà¸ĺ +з н +lay an +ãĤ ¡ +à¸ģà¹ĩ à¸ķาม +ĠsaÄŁ lam +ร à¸ĵ +ĠÑģ иÑĤ +ĠÑģиÑĤ Ñĥ +ĠاÙĦت ÙĨ +×Ķ ×ĸ +ĠØ· ÙĪÙĬÙĦ +ta ÅĤ +Ġgö rd +å¤ī ãĤı +ëĥ ¥ +à¸Ħà¹Ī à¸Ńย +×IJ ×ķ×ĺ +ëħ IJ +ãĥ©ãĥ³ ãĤ¹ +วั à¸Ĵ +วัà¸Ĵ à¸Ļ +Ġol uÅŁ +פע ×ķ׾ +Ġszczeg óÅĤ +à¸Ħา สิ +à¸Ħาสิ à¹Ĥà¸Ļ +pow ied +ĠÑĤ еб +หà¸Ļ à¹Īวย +Ġм ил +ØŃ Ùĥ +à¸Ĺ à¸Ķ +ĠмаÑĤ еÑĢиал +ÅĤ ow +à¹Ģà¸ģ ีย +ĠÑģов еÑĢ +ãĤ © +à¸Ľ ริ +Ġи Ñİ +наÑĩ ен +ÑĢен д +mu ÅŁtur +ĠпÑĢод Ñĥк +з д +Ñı ÑĤи +ÑıÑĤи Ñı +à¹Ģม ีย +رات ÙĬج +Ġam acı +ש ×ķ׾ +ש×ķ׾ ×Ĺ +สะ à¸Ńา +สะà¸Ńา à¸Ķ +פ×Ĵ ×¢ +عب Ø© +d ın +íħ Ķ +Ġ×ŀש ×Ĺ×§ +Ġfi yat +Ġз аÑı +ĠзаÑı в +à¹Ĥ หล +à¹Ĥหล à¸Ķ +à¸ģรุà¸ĩ à¹Ģà¸Ĺà¸ŀ +צ×Ļ ×Ļף +ìļ ± +Ùħ ب +Ùħب اد +land ır +Ġв еÑģÑĮ +Ġh ük +ĠÐĴ оз +ÑĩиÑĤ Ñĭва +ว ล +×ķצ ×¢ +à¸Ĥà¸ĵะ à¸Ĺีà¹Ī +ĠaÅŁ aģı +׾×IJ ×ķ×ŀ×Ļ +tr zym +Ã¤ÃŁ ig +owo ÅĽci +ãģĿ ãĤĤ +Ġroz wiÄħz +ĠgÅĤ ówn +м онÑĤ +×ŀ ×ķ×ŀ +ĠÑģÑĤ ан +ÙĦا ÙĤØ© +p rowad +prowad zi +ĠÑģоÑģÑĤ оÑı +×Ļ×IJ ×ķת +r ı +g ı +ãĥij ãĥij +Ġна лиÑĩ +×Ķ ×¦×¢ +Ġ׳ ×Ķ +à¸Ħ ัà¸ļ +ع راض +и ж +Ùĩ ائÙĬ +ãĤī ãģı +ож еÑĤ +Ġоб оÑĢ +ĠобоÑĢ Ñĥд +Ø£ سÙĦ +à¹ĩ à¸Ķ +ÑĢÑĥ ÑĤ +دÙĬ ÙħÙĤ +دÙĬÙħÙĤ را +Ġjest e +×ķ×ķ ×Ļר +×ij×ĵ ×Ļ×§ +деÑĢж ива +ãģĬ ãģı +ewn ÄĻtr +ewnÄĻtr zn +à¸ŀ ฤ +Ġ×IJ ×ķ×Ķ +ת×Ĺ ×ķש +Ġz ob +д Ñĥм +ĠÑģ Ñĭ +ÙĬر ا +ĠwiÄĻ ks +à¹ģà¸ķà¸ģ à¸ķà¹Īาà¸ĩ +lar aras +lararas ı +íĺ Ģ +ëī ´ +×ķ×Ĵ ׾ +ĠоÑĤ меÑĤ +ĠÑĢ Ð°Ð½ +ت ÙĥÙĦ +иÑĤелÑĮ н +à¸Ľà¸£à¸° วั +à¸Ľà¸£à¸°à¸§à¸± à¸ķิ +ìŀ ĸ +мож но +pie czeÅĦ +pieczeÅĦ st +ëª » +ìĬ ¨ +×ŀס ×ŀ +á» ¦ +ศ ิ +ศิ ล +ศิล à¸Ľ +ĠÅļ w +ãĥĥ ãĤ·ãĥ§ãĥ³ +unit Ãł +Ġmiesz ka +Ġmieszka ÅĦ +pr zed +przed si +przedsi ÄĻb +przedsiÄĻb ior +à¸Ľà¸£à¸° สิà¸Ĺà¸ĺิ +à¸Ľà¸£à¸°à¸ªà¸´à¸Ĺà¸ĺิ à¸łà¸²à¸ŀ +ย à¹Ī +ìķ Ļ +รว à¸Ķ +รวà¸Ķ à¹Ģรà¹ĩว +å½ĵ ãģŁãĤĬ +äl le +Ñĥ еÑĤÑģÑı +ã n +ëł µ +th è +ãĤĴ åĪ©ç͍ +ì µľ +íĵ ¨ +à¸Ĺ ัà¸ļ +า à¸Ħม +ãģ ĩ +ëĤ Į +à¹Ģà¸Ľà¸¥ à¹Īา +â ¦ +ë ¾ +ê Ģ +ê ĩ +â ¡ +ðŁ Ł +ã IJ +â º +á Ń +á Ļ +á ĵ +á ² +ðĵ ı +á ¬ +â ¯ +ä ¨ +ê Ŀ +ê « +ð ij +ðĵ ĥ +ðĿ ħ +< unk + + + +Ġع ÙĦÙī +Ġm á»Ļt +Ġv Ỽi +Ġng ưá»Ŀi +ĠØ¥ ÙĦÙī +Ġnh ững +Ġth á»ĥ +Ġ×IJ ×ķ +Ġ×¢ ×Ŀ +ا Ùĭ +Ġ à¹ģละ +ĠÙĦ ا +Ġnh ư +ĠاÙĦت ÙĬ +Ġ×Ķ ×ķ×IJ +ĠÄij ến +ĠØ£ ÙĪ +Ġv á»ģ +ĠlÃł m +Ġs ẽ +Ġc Å©ng +Ġ ợ +ĠÄij ó +Ġnhi á»ģu +Ġt ại +Ġtr ên +Ġ×Ĵ ×Ŀ +Ġnh Ãł +Ġ׼ ×Ļ +Ġs á»± +ĠÄij ầu +Ġb á»ĭ +ĠÙĩ ذا +Ġnh ất +Ġph ải +Ġhi á»ĩn +Ġdụ ng +ĠÄij á»Ļng +ĠاÙĦÙĦ Ùĩ +ĠØ Į +ĠÙĥ ÙĦ +Ġvi á»ĩc +Ġn Äĥm +Ġth ì +Ġh á»įc +ĠÙĪ Øª +t é +Ġا ÙĨ +Ġt ôi +Ġ×IJ ׳×Ļ +Ġ׾ ×Ļ +Ġ×ŀ ×ķ +Ġng Ãły +Ġn Æ°á»Ľc +Ġ×Ķ ×Ļ×IJ +Ġ×IJ ×Ļ +Ġh Æ¡n +ĠÙĩ ذÙĩ +ĠÙĪ ÙĬ +ĠاÙĦ ذÙĬ +Ġ×ķ ×ŀ +Ġgi á +Ġnh ân +Ġch ÃŃnh +Ġm ình +ĠÐĿ а +Ġth ế +Ġ×Ļ ×ķתר +Ġ×IJ ×Ŀ +Ġn ên +Ġh ợ +Ġhợ p +Ġc òn +ĠÙĩ ÙĪ +Ġc Æ¡ +Ġr ất +ĠVi á»ĩt +Ġب عد +Ġש ×Ļ +Ġth á»Ŀi +Ġc ách +ĠÄij á»ĵng +Ġн о +Ġtr ưá»Ŀng +Ø Ł +ĠÄij á»ĭnh +ĠÄiji á»ģu +×Ļ ×Ļ×Ŀ +Ġth á»±c +n ın +Ġh ình +Ġn ói +Ġc ùng +Ġ×Ķ ×Ķ +ĠØ¥ ÙĨ +Ġ×IJ ×ij׾ +Ġnh ưng +Ġbi ết +Ġж е +Ġch úng +ĠÄij ang +Ġذ ÙĦÙĥ +Ġl ên +Ġkh ách +Ġn Ãło +Ġs á»Ń +Ġkh ác +Ġë° ı +Ġl ý +×Ļ ×Ļ +ĠÄij ây +Ġ׾ ×ŀ +Ġc ần +Ġtr ình +Ġph át +ãģ« ãĤĤ +п о +Ġn Äĥng +Ġb á»Ļ +Ġv ụ +ĠÄij á»Ļ +Ñĩ е +Ġnh áºŃn +Ġtr Æ°á»Ľc +Ġ×¢ ×ĵ +Ġh Ãłnh +ĠØ® ÙĦاÙĦ +Ġl ượng +Ġc ấp +Ġtá» ± +Ġv ì +Ġt ư +Ġch ất +Ġ׼ ×ŀ×ķ +Ġg ì +Ġש ׳ +Ġt ế +ת ×ķ +Ġnghi á»ĩp +Ġm ặt +ĠÙĥ Ùħا +Ġ×ij ×Ļף +Ġר ×§ +Ġth ấy +Ġmá y +ĠÙģ Ùī +Ġd ân +Ġ×IJ ×Ĺ×ĵ +Ġt âm +Ġ׼ ×ļ +Ġ׾ ×ķ +в о +Ġt ác +Ġto Ãłn +ĠÙĪ Ùħ +Ġk ết +Ġ หรืà¸Ń +ĠÙĪØ§ÙĦ Ùħ +ĠÄiji á»ĥm +Ġ×ĸ ×ķ +Ġ×ij ×ķ +׼ ×ķת +Ġh á»Ļi +Ġb ằng +ت Ùĩا +Ġ׼ ×ĵ×Ļ +Ġ×Ķ ×Ŀ +Ġxu ất +ĠÙĤ د +Ġb ảo +Ġt á»ijt +Ġt ình +ĠÙĩ ÙĬ +ĠÄij á»iji +Ġthi ết +Ġhi á»ĩu +Ġti ếp +Ġt ạo +ת ×Ķ +Ġch á»§ +o ÅĽÄĩ +Ġgi ú +Ġgiú p +Ġà ½ +Ġqu ả +Ġlo ại +Ġc ô +Ġà ´ +Ġô ng +Ġ×Ķ ×ķ +ĠاÙĦÙĬ ÙĪÙħ +ĠtÃŃ nh +г а +Ġph òng +Ġ Äĥn +Ġع اÙħ +Ġv á»ĭ +lar ını +r ÃŃa +Ġt Ỽi +ĠÄij ưá»Ŀng +Ġgi Ỽi +Ġb ản +Ġc ầu +Ġnhi ên +Ġb á»ĩnh +Ġth ưá»Ŀng +Ġ×IJ ×Ļף +ĠÄij á»ģ +Ġh á»ĩ +Ġ×Ļש ר×IJ׾ +Ġqu á +ĠÐĹ Ð° +ãģ® ãģ§ãģĻãģĮ +ĠÐŁ ÑĢи +Ġph ần +ĠÙĪ ÙĦا +ĠlỼ n +Ġtr á»ĭ +Ġcả m +Ġм о +Ġd ùng +ĠاÙĦ Ùī +ĠعÙĦÙĬ Ùĩ +ĠìŀĪ ìĬµëĭĪëĭ¤ +ÙĬ ÙĤ +ĠÙĤ بÙĦ +Ġho ặc +ĠØŃ ÙĬØ« +Ġ à¸Ĺีà¹Ī +Ġغ ÙĬر +ĠÄij ại +Ġsá»ij ng +нÑĭ ми +Ġth ức +Ġפ ×Ļ +ĠÄiji á»ĩn +ãģª ãģĭãģ£ãģŁ +Ġgi ải +Ġv ẫn +Ġи Ñħ +Ġö nce +Ġv áºŃy +Ġmu á»ijn +Ġ ảnh +à¹ĥà¸Ļ à¸ģาร +ĠQu á»ijc +Ġk ế +׳ ×IJ +Ġס ×Ļ +Ġy êu +ãģ® ãģĭ +ĠÄij ẹ +ĠÄijẹ p +Ġch ức +Ġy ıl +ĠTür kiye +d é +ĠÙĤ اÙĦ +Ġd á»ĭch +ĠolduÄŁ u +Ġch á»įn +Ġت Ùħ +หà¸Ļ ึà¹Īà¸ĩ +ãģķãĤĮ ãģŁ +Ġph áp +ìĽ Ķ +Ġti á»ģn +ãģĹ ãģ¾ãģĹãģŁ +Ġש ׾×IJ +ÙĦ Ø© +Ġ׾פ ׳×Ļ +Ġ×ij ×Ļת +ĠH Ãł +ĠØŃ ت +ĠØŃت Ùī +Ġ×¢ ×ķ×ĵ +Ġn ó +Ġth áng +à¹Ģลืà¸Ń à¸ģ +ר ×Ķ +Ġt Äĥng +Ġcá i +Ġtri á»ĥn +Ġ×IJ×ķת ×ķ +ìłģ ìĿ¸ +ĠC ông +Ġ׾×Ķ ×Ļ×ķת +Ġг ода +и Ñİ +Ġب عض +Ġ à¸ģาร +èī¯ ãģĦ +ÙĪ Øª +Ġli ên +ĠÐĿ о +ĠÐĿ е +çļĦ ãģª +ĠÙħ ت +ĠÑĤак же +ĠкоÑĤоÑĢ Ñĭе +Ġ×Ļ ×ĵ×Ļ +Ġtr á»įng +ãĤµ ãĤ¤ãĥĪ +ìłģ ìľ¼ë¡ľ +Ġt áºŃp +Ġש ׾×Ļ +íķĺ ê²Į +Ġt Ãłi +ĠÐ ¯ +Ġr á»ĵi +ا Ùĥ +Ġth ương +Ġ×Ķ ×ĸ×Ķ +ĠÙĪ ÙħÙĨ +à¸Ĺีà¹Ī มี +Ġcu á»Ļc +Ġbü yük +ãģ¨ ãģĭ +Ġ×ij ×Ļ×ķתר +Ġl ần +Ġgö re +Ġtr ợ +Ġ×ĺ ×ķ×ij +ÑĤÑĮ ÑģÑı +Ġth á»ijng +Ġ׼ ש +Ġti êu +Ġ×ŀ×IJ ×ķ×ĵ +Ø Ľ +k Äħ +Ġ à¹ĥà¸Ļ +Ġv ấn +Ġש ׾×ķ +ĠÄij á»ģu +Ùģ Øª +Ġê²ĥ ìĿ´ +Ġh óa +ĠاÙĦع اÙħ +ĠÙĬ ÙĪÙħ +к ой +Ġbi á»ĩt +ÑģÑĤ о +Ġ×Ķ ×Ļ×ķ +à¸Ĺีà¹Ī à¸Īะ +Ġ×ĵ ×Ļ +Ġ×IJ ×ļ +Ġá n +ص ÙĪØ± +Ġtr ÃŃ +ĠÐŁÑĢ Ð¾ +Ġl á»±c +ãģĹãģ¦ ãģĦãģ¾ãģĻ +Ġb Ãłi +Ġ×ĸ ×IJת +Ġb áo +à¸ļ à¸Ļ +ĠëĮĢ íķľ +Ġti ế +Ġtiế ng +Ġb ên +ãģķãĤĮ ãĤĭ +s ión +Ġt ìm +×¢ ×ķ +m é +ни Ñı +ãģ» ãģ© +Ġà¹Ģà¸ŀ ราะ +ب Ø© +Ġë¶ Ħ +Ġ×IJ ×ĸ +à¸Ĺ à¹Īาà¸Ļ +ת ×Ŀ +Ġth êm +Ġho ạt +y ı +×ĸ ×ķ +Ġgi á»Ŀ +Ġb án +à¸Ĥ าย +Ñĩ а +Ġ à¹Ĩ +ĠاÙĦÙħ ت +ĠоÑĩ енÑĮ +Ġb ất +Ġtr ẻ +ÑĤ ÑĢ +ĠØ£ ÙĨÙĩ +ĠØ« Ùħ +Ġ׼ ×ŀ×Ķ +Ġkh ó +Ġr ằng +ĠÙĪ ÙģÙĬ +ни й +Ġho Ãłn +t ó +Ġ×IJ שר +ĠìĥĿ ê°ģ +Ñģ а +Ġ׼ ×ijר +ĠÑįÑĤ ом +lar ının +Ġch ưa +з и +Ġd ẫn +ĠÐļ ак +ج ÙĪ +ĠбÑĭ ло +ĠÙĬ ت +n ı +ÅĤ am +ĠÙĪÙĩ ÙĪ +×ij ×ķ +п и +ר ת +Ġqu á»ijc +ж д +ĠÄij Æ¡n +Ùĥت ب +Ġm ắt +ระ à¸ļ +ระà¸ļ à¸ļ +ĠÙĥ اÙĨت +Ġth ân +สิà¸Ļ à¸Ħà¹īา +×Ĵ ×Ļ +Ġph ương +à¹Ħมà¹Ī à¹Ħà¸Ķà¹ī +ĠìĦ ± +ĠC ác +Ġ×Ķ×ŀ ×ķ +ĠÑĤ ем +Ġ×ĵ ×ķ +à¸Ńะ à¹Ħร +Ġv Äĥn +ãģª ãģ®ãģ§ +ĠN á»Ļi +Ġ×¢ ×ķ +ãĤīãĤĮ ãĤĭ +Ġs áng +Ġgö ster +ãģĵãģ¨ ãĤĴ +Ġtaraf ından +Ġм а +ĠпоÑģл е +Ġ׳ ×Ļת +Ġ׳×Ļת ף +Ġл еÑĤ +Ġ׾ ׳×ķ +Ñģ Ñģ +Ġ×Ļ ×ķ +п е +ĠÙĪ ÙĦÙĥ +ĠÙĪÙĦÙĥ ÙĨ +Ġngo Ãłi +ĠÄij á»ĭa +r zÄħd +dz iaÅĤ +ĠÙħ ر +иÑĤÑĮ ÑģÑı +Ġ×IJ×Ĺר ×Ļ +Ġ׾ ׼׾ +à¸Ĥ à¹īà¸Ńม +à¸Ĥà¹īà¸Ńม ูล +Ġб ол +Ġбол ее +جÙħ ع +л еÑĤ +Ġl á»ĭch +ĠÙħ Ø«ÙĦ +Ġ그리 ê³ł +Ġth ứ +ĠdeÄŁ il +ÙĪ ØŃ +Ġש׾ ×ļ +ĠÙħ ØŃÙħد +Ġn ếu +ĠÄij á»ķi +Ġv ừa +Ġm á»įi +Ġо ни +Ġl úc +ĠÙĬ ÙĥÙĪÙĨ +ì§ Ī +Ġש׾ ׳×ķ +ĠÐĶ Ð¾ +Ġש ׳×Ļ +ล ิ +×IJ פשר +Ġs ức +ê¶ Į +Ġ ứng +à¹Ħมà¹Ī มี +Ø·ÙĦ ب +ĠÑĩ ем +Ġch uyên +Ġth ÃŃch +Ġ×ķ ×Ļ +íķ © +ĠÙħ صر +д о +ĠÄij ất +Ġch ế +à¸Ĭ ืà¹Īà¸Ń +Ġìĭ ł +ĠØ¥ ذا +Ġر ئÙĬس +Ġש ×Ļש +Ġgiả m +Ñģ ка +lar ında +Ġs ợ +ĠtÃŃ ch +ĠÙĦ ÙĥÙĨ +Ġب Ùħ +×¢ ×ķ×ij +×¢×ķ×ij ×ĵ +ÅĤÄħ cz +ları na +Ġש ×Ŀ +ĠÙĦ ت +Ġש×Ķ ×ķ×IJ +t ów +Ġëĭ¤ 른 +ĠØ£ Ùĥثر +ãģ® ãģ§ãģĻ +׼ ×Ļ×Ŀ +ĠolduÄŁ unu +ãģĭ ãģª +ãĤĤ ãģĨ +ÙĬ ØŃ +Ġnh ìn +Ġngh á»ĩ +ãģ«ãģª ãģ£ãģ¦ +п а +Ġquy ết +ÙĦ ÙĤ +t á +Ġlu ôn +ĠÄij ặc +Ġ×IJ ר +Ġtu á»ķi +s ão +ìĻ ¸ +ر د +ĠبÙĩ ا +Ġ×Ķ×Ļ ×ķ×Ŀ +×ķ ×ķ×Ļ +ãģ§ãģĻ ãģŃ +ĠÑĤ ого +Ġth á»§ +ãģĹãģŁ ãģĦ +ر ÙĤ +Ġb ắt +г Ñĥ +Ġtá» Ń +ÑĪ Ð° +Ġ à¸Ľà¸µ +Ġ×Ķ×IJ ×Ŀ +íı ¬ +ż a +Ġ×IJת ×Ķ +Ġn á»Ļi +Ġph ÃŃ +ĠÅŁek ilde +Ġl á»Ŀi +d ıģı +Ġ׼×IJ ף +Ġt üm +Ġm ạnh +ĠM ỹ +ãģĿ ãĤĵãģª +Ġnh á»ı +ãģª ãģĮãĤī +Ġb ình +ı p +à¸ŀ า +ĠÄij ánh +ĠÙĪ ÙĦ +ר ×ķת +Ġ×IJ ×Ļ×ļ +Ġch uyá»ĥn +Ùĥ ا +ãĤĮ ãĤĭ +à¹ģม à¹Ī +ãĤĪ ãģı +ĠÙĪ ÙĤد +íĸ Īëĭ¤ +Ġn Æ¡i +ãģ«ãĤĪ ãģ£ãģ¦ +Ġvi ết +Ġà¹Ģà¸ŀ ืà¹Īà¸Ń +ëIJĺ ëĬĶ +اد ÙĬ +ĠÙģ Ø¥ÙĨ +ì¦ Ŀ +ĠÄij ặt +Ġh Æ°á»Ľng +Ġx ã +Ġönem li +ãģł ãģ¨ +Ġm ẹ +Ġ×ij ×Ļ +Ġ×ĵ ×ijר +Ġv áºŃt +ĠÄij ạo +Ġdá»± ng +ĠÑĤ ом +ĠÙģÙĬ Ùĩا +Ġج ÙħÙĬع +Ġthu áºŃt +st ÄĻp +Ġti ết +Ø´ ÙĬ +Ġе Ñīе +ãģĻãĤĭ ãģ¨ +ĠmÃł u +ĠÑįÑĤ ого +Ġv ô +ĠÐŃ ÑĤо +Ġth áºŃt +Ġn ữa +Ġbi ến +Ġn ữ +Ġ׾ ׼×Ŀ +×Ļ ×Ļף +Ġس ت +ĠÐŀ ÑĤ +Ġph ụ +ê¹Į ì§Ģ +Ġ׾ ×ļ +Ġk ỳ +à¹ĥ à¸Ħร +Ġg ây +ĠÙĦ ÙĦÙħ +Ġtụ c +ت ÙĬÙĨ +Ġtr ợ +Ġ׾ פ×Ļ +Ġb á»ij +ĠÐļ а +ĠÄij ình +ow Äħ +s ında +Ġkhi ến +s ız +Ġк огда +ס ׾ +ĠбÑĭ л +à¸Ļ à¹īà¸Ńย +обÑĢаР· +Ġê²ĥ ìĿ´ëĭ¤ +ëĵ¤ ìĿĢ +ãģ¸ ãģ® +Ġà¹Ģม ืà¹Īà¸Ń +Ġph ục +Ġ׊׾ק +Ġh ết +ĠÄij a +à¹Ģà¸Ķà¹ĩ à¸ģ +íĺ ķ +l ÃŃ +ê¸ ī +Ġع دد +ĠÄij á»ĵ +Ġg ần +Ġ×Ļ ×ķ×Ŀ +Ġs Ä© +ÑĢ Ñıд +Ġquy á»ģn +Ġ×IJ ׾×IJ +Ùĩ Ùħا +׳ ×Ļ×Ķ +׾ ×ķת +Ġ×Ķר ×ij×Ķ +Ġti ên +Ġal ın +Ġd á»ħ +人 ãģĮ +но Ñģ +л ÑģÑı +ĠÄij ưa +ส าว +иÑĢов ан +Ġ×ŀס פר +×Ĵ ף +Ġki ến +ĠÐ ¨ +p é +б Ñĥ +ов ой +б а +ĠØ¥ ÙĦا +×IJ ׾×Ļ +Ġx ây +Ġb ợi +Ġש ×ķ +人 ãģ® +×§ ×Ļ×Ŀ +à¹Ģà¸Ķ ืà¸Ńà¸Ļ +Ġkh á +Ġ×ķ ׾×Ķ +×ĵ ×ķת +Ġ×¢ ×ij×ķר +Ġبش ÙĥÙĦ +ĠÙĩÙĨا Ùĥ +ÑĤ ÑĢа +Ġ íķĺëĬĶ +ร à¸Ńà¸ļ +owa ÅĤ +h é +Ġdi á»ħn +Ġ×Ķ ×Ľ×ľ +ĠØ£ س +Ġch uyá»ĩn +ระ à¸Ķัà¸ļ +ĠNh ững +Ġ×IJ ×Ĺת +ĠØŃ ÙĪÙĦ +л ов +׳ ר +Ġ×ķ ׳ +Ġch Æ¡i +Ġiç inde +ÑģÑĤв Ñĥ +Ġph á»ij +ĠÑģ Ñĥ +ç§ģ ãģ¯ +Ġch ứng +Ġv á»±c +à¹ģ à¸Ń +Ġl áºŃp +Ġtừ ng +å°ij ãģĹ +ĠNg uy +ĠNguy á»ħn +ĠÙģÙĬ Ùĩ +Ġб а +×Ļ ×Ļת +Ġ×ľ×¢ ש×ķת +Ġ×ŀ ׼ +Ġnghi á»ĩm +Ġм ного +Ġе е +ëIJĺ ìĸ´ +Ġl ợi +Ġ׾ ׾×IJ +Ġ׼ ף +Ġch ÃŃ +ãģ§ ãģ® +×Ĺ ×ķ +ש ×ķ×Ŀ +Ġ×ŀ ר +ĠÐĶ Ð»Ñı +Å ģ +Ġ׼×IJ שר +ĠM á»Ļt +ĠÙĪØ§ÙĦ ت +ĠìĿ´ 룰 +ÅŁ a +Ġchi ến +Ġaras ında +Ġ×ij ×IJתר +ãģķãĤĮ ãģ¦ãģĦãĤĭ +Ø´ ÙĥÙĦ +Ġt ượng +Ġت ت +ĠC ó +Ġb á»ı +Ġtá»ī nh +Ġkh ÃŃ +ĠпÑĢ Ð¾ÑģÑĤ +ĠпÑĢоÑģÑĤ о +ĠÙĪ ÙĤاÙĦ +Ġgi áo +ĠN ếu +×IJ ×ŀר +×¢×ł×Ļ ×Ļף +íİ ¸ +Ùĩد Ùģ +ĠB á»Ļ +Ġb Ãłn +Ġng uyên +Ġgü zel +ส าย +ì² ľ +×ŀ ×ķר +Ġph ân +ס פק +×§ ×ij׾ +ĠاÙĦÙħ تØŃ +ĠاÙĦÙħتØŃ دة +ائ د +Ġ×IJ ×ŀר +Ġki ÅŁi +ì¤ Ģ +Ġtr uyá»ģn +ĠÙĦ Ùĩا +ĠÐľ а +à¸ļริ ษ +à¸ļริษ ั +à¸ļริษั à¸Ĺ +Ġש ׳×Ļ×Ŀ +Ġмен Ñı +ÅŁ e +Ġdi á»ĩn +Ġ×IJ׳ ×Ĺ׳×ķ +k ü +Ġc á»ķ +Ġm á»Ĺi +w ä +Ùħ ÙĬ +Ġhi á»ĥu +ëĭ ¬ +Ġ×Ķ ×Ĺ׾ +Ġt ên +Ġki á»ĩn +ÙĨ ÙĤÙĦ +Ġv á»ĩ +×ĵ ת +ĠÐłÐ¾ÑģÑģ ии +л Ñĥ +ĠاÙĦع ربÙĬØ© +ĠØ· رÙĬÙĤ +Ġ×Ķ×ij ×Ļת +Ñģ еÑĢ +Ġм не +ä u +Ġtri á»ĩu +ĠÄij á»§ +Ġר ×ij +ت ÙĩÙħ +à¸ĭ ี +Ġì§Ģ ê¸Ī +li ÅĽmy +د عÙħ +ãģł ãĤįãģĨ +Ñģки е +Ġh á»ıi +Ġ×§ ×ķ +ÑĢÑĥ Ñģ +ÙĨ ظر +ãģ® ãĤĤ +Ġ×Ķ ×Ľ×Ļ +ĠìĽ IJ +ÙĪ Ùĩ +ĠÙĪ Ùİ +ĠB ạn +п лаÑĤ +Ġ×ŀ ×ŀש +лÑİ Ð± +ĠнÑĥж но +Ġth ư +ãģ µ +ãģı ãĤīãģĦ +ر Ø´ +ר ×ķ×Ĺ +ĠÙĬ تÙħ +Ġצר ×Ļ×ļ +Ġph á +ม à¸Ńà¸ĩ +Ġ×ij×IJ ×ķפף +Ġcả nh +Ġíķľ ëĭ¤ +Ġ×Ķ×ŀ ת +à¸ķà¹Īาà¸ĩ à¹Ĩ +มี à¸ģาร +Ñģки Ñħ +ĠÐĴ Ñģе +Ġا ÙĪ +ج ÙĬ +ãģĵãģ¨ ãģ¯ +Ġd Ãłi +Ġh á»ĵ +èĩªåĪĨ ãģ® +à¹Ħ หà¸Ļ +ëĵ¤ ìĿĦ +ĠV Äĥn +Ġд аж +Ġдаж е +Ñĭ ми +лаÑģ ÑĮ +ÙĬ ÙĪÙĨ +ÙĨ ÙĪ +c ó +ãģĹãģ¦ ãģĦãģŁ +ãģł ãģĭãĤī +طاÙĦ ب +Ġc á»Ńa +п ÑĢоÑģ +ãģªãģ© ãģ® +รุ à¹Īà¸Ļ +Ġchi ếc +л Ñĭ +ĠÑıвлÑı еÑĤÑģÑı +Ġn á»ķi +ãģ® ãģĬ +Ġ×IJת ×Ŀ +ĠëķĮ문 ìĹIJ +à¸ģล าà¸ĩ +ĠbaÅŁ ka +ìĦ Ŀ +ĠÑĨ ел +Ùģ ÙĤ +ãģ«ãĤĪ ãĤĭ +ÙĤ ا +Ġçı kar +Ġcứ u +Ø· ا +Ġש ת +à¹Ĥ à¸Ħ +Ġ×ŀ ׾ +Ġ×Ķ ×¤×¨ +Ġг де +ĠØ® Ø· +åīį ãģ« +c jÄĻ +Ġ׊ש×ķ×ij +ר×Ĵ ×¢ +Ġkho ảng +ĠÄij á»Ŀi +ĠÐł е +Ġо на +Ġ×IJ ׳×ķ +ãģ® ãģ« +ĠاÙĦذ ÙĬÙĨ +кÑĥ п +ãĤµ ãĥ¼ãĥ +ãĤµãĥ¼ãĥ ĵ +ãĤµãĥ¼ãĥĵ ãĤ¹ +в ал +г е +Ġgi ữa +ĠKh ông +ĠâĹ ĭ +à¸ģล ุà¹Īม +ĠÙħÙĨ ذ +à¸Ń à¹Īาà¸Ļ +ĠÑģп оÑģоб +ĠÄij á»Ļi +Ġdi ÄŁer +Ġ à¸ĸà¹īา +Ùħ Ø«ÙĦ +Ġ×Ķ×IJ ×Ļ +Ġد ÙĪÙĨ +ÙĬر اÙĨ +Ñī и +بÙĨ اء +ĠØ¢ خر +ظ Ùĩر +Ġ×ij ׼ +ĠاÙĦÙħ ع +ãĥ Ĵ +Ġt ất +Ġm ục +ĠdoÄŁ ru +ãģŁ ãĤī +Ġס ×ķ +Ġx ác +ร à¸Ń +ĠcÄĥ n +Ġон л +Ġонл айн +Ġk ý +Ġch ân +Ġ à¹Ħมà¹Ī +اØŃ Ø© +r án +׳×Ļ ×Ļ×Ŀ +Ġ×ij ף +ĠÐ ĸ +à¸ķร à¸ĩ +д Ñĭ +Ġs ắc +ÙĦ ت +ãĥŃ ãĥ¼ +ĠÙĦ ÙĨ +Ġר ×ķ +Ġd Æ°á»Ľi +à¹Ģ à¸ĺ +à¹Ģà¸ĺ à¸Ń +e ÄŁi +Ġ×ķ ש +ĠÙĦ Ø£ +Ġg ặp +Ġc á»ij +ãģ¨ ãģ¦ãĤĤ +رÙĪ Ø³ +Ġ׾×Ķ ×Ļ +Ġë³ ¸ +ä¸Ĭ ãģĴ +Ġm ức +Ñħ а +Ġìŀ ¬ +à¸ī ัà¸Ļ +ÑĢÑĥ ж +Ġaç ık +ÙĪ Ø§ÙĦ +Ġ×ĸ ×ŀף +人 ãģ¯ +ع ÙĬÙĨ +Ñı Ñħ +Ġ×Ĵ×ĵ ×ķ׾ +ר ×ķ×ij +g ó +ëĿ¼ ê³ł +Ġark adaÅŁ +ÙĨ شر +Ġгод Ñĥ +ĠболÑĮ ÑĪе +ãģ¡ãĤĩ ãģ£ãģ¨ +Ġcâ u +Ġs át +íĶ ¼ +Ġti ến +íķ´ ìķ¼ +ĠÙĪ Ø£ÙĨ +à¸Ļ าà¸Ļ +Ġ×ij×IJ×ŀ צע +Ġ×ij×IJ×ŀצע ×ķת +Ġ׾ ר +Ġqu ản +ĠÙĪØ§ÙĦ Ø£ +Ġ×IJ×ķת ×Ķ +Ġìĸ´ëĸ ¤ +Ġê²ĥ ìĿĢ +ØŃس ÙĨ +Ġm ất +à¸Ħ ูà¹Ī +ãĥ¬ ãĥ¼ +ĠÐĶ Ð° +Ġol ması +Ġthu á»Ļc +׳ ×Ĺ +íĨ ł +Ġsö yle +ãģĿãģĨ ãģ§ãģĻ +Ġت ÙĥÙĪÙĨ +л ÑĥÑĩ +׾ ×Ļ×ļ +ĠØ£ ØŃد +ли ÑģÑĮ +ĠвÑģ его +Ġ×Ķר ×ij +Ġëª » +o ÄŁ +oÄŁ lu +ĠìĦ ł +Ġк аÑĢ +à¸łà¸² à¸Ħ +e ÅĦ +Ġ à¸ģà¹ĩ +Ġa ynı +Ġb Ãł +ãģªãĤĵ ãģ¦ +Ġ모 ëĵł +ÙĤر ار +ãģĹãģª ãģĦ +ĠÐĴ о +ĠÙĪÙĩ ÙĬ +ни ки +ãĤĮ ãģŁ +Ġchu ẩn +ר ×¢ +Ùģ Ø±ÙĬÙĤ +ãĤĴ åıĹãģij +ĠÄij úng +б е +׼ ×ķ×Ĺ +п Ñĥ +Ġ×ķ ×Ĵ×Ŀ +×ŀ ׳×Ļ +íĸ ¥ +צ ×Ļ×Ŀ +à¸ĭ ิ +Ùĩ ÙĨ +н ем +Ġ×ij×ij ×Ļת +ر ع +Ġ ส +ĠÄIJ Ãł +íķĺ ëĭ¤ +Ġ ấy +×Ĺ ×ķ×ĵ +×Ĺ×ķ×ĵ ש +ĠÑĩеÑĢ ÐµÐ· +Ñĥ л +ĠB ình +Ġê²ĥ ìĿĦ +Ġ×Ĵ ר +ä»ĺ ãģij +×Ĺ׾ ×§ +Ġت ÙĦÙĥ +à¹ĥส à¹Ī +sz Äħ +ÙĤ اÙħ +د ÙĪØ± +ĠÙģ ÙĤØ· +Ġh ữu +Ġмог ÑĥÑĤ +Ġg á»įi +Ġ×§ ר +à¸Īะ มี +ت ÙĤدÙħ +Ġع بر +Ġ׾×Ķ ×Ŀ +ĠÑģам о +ס ×ĵר +Ġc Ãłng +r ÃŃ +Ġìŀ ¥ +ëĵ¤ ìĿĺ +ĠÙĦ Ùĥ +п оÑĢÑĤ +Ġkh ả +ĠÑģеб Ñı +׳ ף +Ġد ÙĪØ± +Ġm ợ +Ġcâ y +Ġf ark +Ġfark lı +а ÑİÑĤ +Ġtr á»±c +wiÄĻks z +Ġthu á»ijc +Ġت ØŃت +ت ÙĦ +ов Ñĭе +ëĤ ł +Ġв ам +بÙĦ غ +Ġê°Ļ ìĿĢ +íĮ IJ +ÙĦ ب +Ġnas ıl +Ġод ин +м ан +ĠعÙĦÙĬ Ùĩا +б и +Ġפ ש×ķ×ĺ +×ijר ×Ļ +Ġש ׳×Ķ +Ġëı Ħ +ĠÄIJ ại +Ġ×IJ×ķת ×Ŀ +ĠاÙĦØŃ ر +Ġб о +à¸Ī ุà¸Ķ +Ġr õ +ĠdeÄŁi ÅŁ +Ġëĭ ¨ +ĠÑģлÑĥÑĩ а +ĠÑģлÑĥÑĩа е +Ġ×IJ׳ ש×Ļ×Ŀ +×ĵ ×£ +ש×ij ת +Ġש׾ ׼×Ŀ +Ġch ú +nik ów +Ġtan ı +Ġcá o +ĠÄij á +Ġ×IJ ×ĵ×Ŀ +Ġê° ķ +Ġnhi á»ĩm +Ġ׾ ס +Ġ×Ľ×ª ×ij +Ġ×Ķס פר +ĠÄij Äĥng +Ġë ijIJ +à¸ľ ิ +à¸ľà¸´ ว +ج ا +Ġê° IJ +ر Ø£ +ست خدÙħ +ãģ«ãģªãĤĬ ãģ¾ãģĻ +Ġtá» · +×ĺ ×ķר +г овоÑĢ +Ġв оÑģ +ĠÙħÙĨ Ùĩا +иÑĢов аÑĤÑĮ +ĠÄij ầy +׳ ×Ĵ +ĠÙħ ÙĪ +ĠÙħ ÙĪÙĤع +ר׼ ×Ļ +ت Ùı +ëª ¨ +Ġת ×ķ +ÙĬا Ùĭ +à¹ĥ à¸Ķ +ãĤĬ ãģ¾ãģĻ +à¸Ńยูà¹Ī à¹ĥà¸Ļ +ĠØ£ ÙĪÙĦ +ĠØ£ خرÙī +Ġc ư +ص ار +×ŀ׊ש×ij +б ÑĢа +ÅĦ ski +б ÑĢ +ĠÙĬ Ùı +à¸ģ ิà¸Ļ +Ġch á»ijng +Ùħ Ùı +Ġ à¸Ħืà¸Ń +Ġت ÙĨ +t ÃŃ +y Äĩ +Ġm ạng +Ùģ ÙĪ +Ġdü nya +×§ ר×IJ +Ġ×§ ׾ +ĠØŃ اÙĦ +c ÃŃa +Ġà¹Ģ รา +Ġר ×ķצ×Ķ +Ġá p +ë° ķ +ا ÙĤØ© +ни Ñİ +Ġ×IJ ׾×ķ +Ġ×ŀס ×ķ +ãģ§ãģ¯ ãģªãģı +Ġtr ả +Ġ×§ שר +mi ÅŁtir +Ġl ưu +Ġh á»Ĺ +ĠбÑĭ ли +Ġl ấy +عÙĦ Ùħ +Ġö zel +æ°Ĺ ãģĮ +Ġ×ĵ ר×ļ +Ùħ د +s ını +׳ ×ķש×IJ +r ów +Ñĩ еÑĢ +êµIJ ìľ¡ +ĠÐľ о +л ег +ĠV Ỽi +วัà¸Ļ à¸Ļีà¹ī +ÑİÑī ие +ãģĬ ãģĻ +ãģĬãģĻ ãģĻ +ãģĬãģĻãģĻ ãĤģ +ëı ħ +Ġ×Ļ×Ķ ×Ļ×Ķ +×ŀ ×ĺר +Ñı ми +Ġl á»±a +ĠÄij ấu +à¹Ģส ียà¸ĩ +Ġt ương +ëĵ ± +ĠÑģÑĤ аÑĢ +à¹ĥ à¸ļ +ว ัà¸Ķ +Ġİ stanbul +Ġ à¸Īะ +à¸ķ ลาà¸Ķ +Ġب ÙĬ +à¹ģà¸Ļ ะ +à¹ģà¸Ļะ à¸Ļำ +س اعد +Ġب Ø£ +Ġki á»ĥm +ØŃ سب +à¸Ĭั à¹īà¸Ļ +Ġ×ķ ×¢×ķ×ĵ +ов ÑĭÑħ +оÑģ нов +Ġtr Æ°á»Łng +צ ×ij×¢ +ĠÃŃ t +Ġk ỹ +cr é +Ñı м +êµ ° +ãģĮ ãģªãģĦ +ÙĬÙĦ Ø© +ãĥķ ãĤ£ +ر Ùī +ĠÙĬ جب +Ġ×IJ ×£ +Ġc á»±c +ãĤīãĤĮ ãģŁ +Ġ à¸ľà¸¹à¹ī +Ġ à¸Ń +lar ımız +Ġkad ın +Ġê·¸ ëŀĺ +Ġê·¸ëŀĺ ìĦľ +ĠëĺIJ ëĬĶ +ĠÄij ả +ĠÄijả m +Ġ×IJ ×ķ×ŀר +Ġy ếu +ci Äħ +ciÄħ g +Ġt á»ij +Ġש×IJ ׳×Ļ +Ġdz iaÅĤa +Ñī а +ĠÄij Ãłn +s ına +ãģĵãĤĮ ãģ¯ +Ġ×ij ׾×Ļ +Ġ×ij ×Ļשר×IJ׾ +л оÑģÑĮ +Ġgi ữ +ê° IJ +ÑĢ Ð¾Ð½ +تج ار +г лав +в ин +Ġh ạn +Ġyapı lan +ب س +Ġ à¸ŀรà¹īà¸Ńม +ê´Ģ 리 +mÄ±ÅŁ tır +b ü +r ück +ĠBaÅŁkan ı +ĠÙĦ ÙĬس +Ġs Æ¡ +à¸Īัà¸ĩ หว +à¸Īัà¸ĩหว ัà¸Ķ +د اء +Ġ×Ķ ×Ľ +v ÃŃ +ש ×IJר +Ġh Æ°á»Łng +Ġb óng +ĠCh ÃŃnh +Äħ c +à¹Ģà¸ģีà¹Īยว à¸ģัà¸ļ +Ġtá» © +Ġtứ c +ĠÑĨ веÑĤ +Ġt á»iji +ĠnghÄ© a +ÙĦا عب +د ÙĦ +Ġפע ×Ŀ +h ör +à¸Ĭ ุà¸Ķ +à¸ŀ ู +à¸ŀู à¸Ķ +п аÑģ +ĠÅŁ u +Ġt Æ°á»Łng +خار ج +Ġâ m +ĠинÑĤеÑĢ ÐµÑģ +ен нÑĭÑħ +×IJ ׳×Ļ +بد Ø£ +ëĿ¼ ëĬĶ +ì¹ ´ +æĸ¹ ãģĮ +ли в +Ġ à¸Ħà¸Ļ +ער ×ļ +à¸Ĥà¸Ńà¸ĩ à¸Ħุà¸ĵ +п ад +Ġc ạnh +ĠëĤ ¨ +ĠÄij âu +Ġbi á»ĥu +ãĤĤ ãģĤãĤĭ +׾ ×Ĵ +Ġ สำหรัà¸ļ +Ġxu á»ijng +ס ×ķ +Ġذ ات +ĠÐľ е +ع اÙĦÙħ +×IJ ס +ب ÙĬØ© +Ø´ ا +и ем +ĠNg ưá»Ŀi +íĺ ij +Ñģл ов +Ġп а +Ġm ẫu +ĠпÑĢоÑĨ еÑģÑģ +ĠNh Ãł +пÑĢо из +пÑĢоиз вод +à¸łà¸²à¸¢ à¹ĥà¸Ļ +Ġ à¸ļาà¸Ĺ +×ŀ ׳×ķ +ĠоÑĢг ан +רצ ×ķ +×ķ×ŀ ×Ļ×Ŀ +Ġyaz ı +Ġd ù +ãĥ¬ ãĥ³ +ÙĪÙĦ ÙĬ +ย ู +Ġtr ò +à¹Ģà¸ŀ ลà¸ĩ +Ġ×ŀ ׾×IJ +à¸ķ ล +à¸ķล à¸Ńà¸Ķ +ĠÄij ạt +Ġ×Ĺ×ĵ ש +p óÅĤ +Ġ×ŀ ×ĵ×Ļ +ujÄħ c +×ŀ׳×Ķ ×ľ +Ġש×ij ×ķ +Ġ×Ķ×ŀש פ×ĺ +Ġ×IJ ׾×Ķ +ĠÙĪ Ø°ÙĦÙĥ +à¹Ģà¸ŀ ราะ +ĠÄijo Ãłn +Ġíķ¨ ê»ĺ +Ġd ục +Ø´ ت +Ġ ula +Ġula ÅŁ +Ġqu ý +Ġ×Ķ ×Ĵ×ĵ×ķ׾ +à¸ķัà¹īà¸ĩ à¹ģà¸ķà¹Ī +Ġש ר +Ø´ Ùĩد +׳ ש×Ļ×Ŀ +à¸ŀ ล +رÙĪ Ø§ +ãĤĮ ãģ¦ +Ġн иÑħ +Ġдел а +ãģ§ãģį ãģªãģĦ +ÅĤo ż +×IJ ×Ĺר +ì ½Ķ +ãĤ¢ ãĥĥãĥĹ +د Ù쨹 +Ġti á»ĩn +Ġkh á»ı +Ġkhá»ı e +ĠاÙĦع اÙħØ© +ãģ« ãģĤãĤĭ +ĠÄij á»Ļc +ì¡ ± +Ġc ụ +й ÑĤе +Ġзак он +ĠпÑĢо екÑĤ +ìĸ ¸ +ÙĦ ØŃ +ĠçalÄ±ÅŁ ma +ãĤĴ ãģĻãĤĭ +Ñħ и +ع اد +Ġ׳ ×ŀצ×IJ +Ġר ×Ļ +à¸Ńà¸Ńà¸ģ มา +ĠT ôi +Ġth ần +ĠÙĬ ا +ล าย +Ġав ÑĤо +Ġsı ra +ĠÙĥ Ø«ÙĬر +Ùħ ÙĬز +ĠاÙĦع ÙĦÙħ +æĸ¹ ãģ¯ +×ķ×¢ ×ĵ +Ġобла ÑģÑĤи +×Ļ׾ ×Ļ×Ŀ +ãģĮ åĩº +à¸ĺ ุ +à¸ĺุ ร +à¸ĺุร à¸ģิà¸Ī +ÙĤت ÙĦ +ר×IJ ×ķ +Ġng u +Ġngu á»ĵn +Ġ มา +Ġпл ан +t ório +Ġcu á»iji +Ñģк ом +ĠاÙĦÙħ اض +ĠاÙĦÙħاض ÙĬ +Ġ×ij×¢ ׾ +Ġר ×ij×Ļ×Ŀ +Ġlu áºŃn +Ùĥ ÙĪ +à¸Ĺัà¹īà¸ĩ หมà¸Ķ +в ан +Ġtho ại +à¹Ħ à¸Ń +б иÑĢ +ĠاÙĦ ض +ت ا +ĠÑĢ Ð¾Ð´ +ĠV Ãł +×ŀ ×Ļף +ĠбÑĭ ла +к ами +ĠÐĶ Ðµ +t ık +קר ×Ļ +ĠeÄŁ itim +ĠÙĥ بÙĬر +ب Ùĥ +ĠÙĦ ÙĪ +в ой +Ġ ãģĵãģ® +ĠÑĤ ÑĢÑĥд +my ÅĽl +Ġs ư +à¸ŀ ีà¹Ī +Ġ à¹ģลà¹īว +×¢ ×§ +Ġ×Ĺ×ijר ת +ระ หว +ระหว à¹Īาà¸ĩ +×Ļ ×Ļ×Ķ +ĠاÙĦÙĨ اس +ün ü +Ġ׾ ×ŀ×Ķ +Ġch ương +ĠH á»ĵ +ار ت +ãĤĪãģĨ ãģ§ãģĻ +l á +×§×Ļ ×Ļ×Ŀ +æľ¬ å½ĵ +æľ¬å½ĵ ãģ« +ãģĵãĤĵ ãģª +Ñģ ов +Ġ×ķ ×Ĺ +à¹Ģà¸ģ à¹ĩà¸ļ +Ġк ÑĤо +à¹Ĥร à¸Ħ +ĠØ´ رÙĥØ© +ع زÙĬ +عزÙĬ ز +Ø·ÙĦ ÙĤ +п ÑĥÑģÑĤ +Ùģ ØªØŃ +ëŀ Ģ +Ġhã y +ض Ùħ +ë¦ ° +åł´åIJĪ ãģ¯ +ãĤª ãĥ¼ +Ġh ắn +Ġ×IJ ×ij×Ļ×ij +Ġש׾×Ķ ×Ŀ +Ġ×Ķ×Ļ ×Ļת×Ķ +ĠاÙĦد ÙĪÙĦØ© +ĠاÙĦ ÙĪÙĤ +ĠاÙĦÙĪÙĤ ت +ãģĤ ãģ¾ãĤĬ +Ġta ÅŁÄ± +İ N +×¢ סק +ãģ¦ ãģĦãģŁ +Ġtá»ķ ng +ĠاÙĦØ¥ ÙĨس +ĠاÙĦØ¥ÙĨس اÙĨ +ÑĢ ÐµÑĪ +Ġg ái +ĠÑĨ ен +ĠÙģ ÙĤد +Ùħ ات +ãģķãĤĵ ãģ® +Ġph ù +×ĺ ×Ķ +ĠÙĪØ§ÙĦ تÙĬ +Ġب Ùĥ +ìĿ´ ëĤĺ +к Ñģ +Ùħ ÙĬر +Ġv ùng +ĠاÙĦØ´ عب +ĠNh ưng +ãĥĢ ãĥ¼ +Ġ×Ĺ×Ļ ×Ļ×Ŀ +ĠØ´ خص +×§ ×ķ×ĵ +ê² Ģ +×¢ ש +×¢ ×ķ׾×Ŀ +צ ×ķר +ع ÙĤد +ĠiÅŁ lem +Ġ×Ķ×ij ×IJ +Ġd ưỡng +à¸Ł รี +Ġph ÃŃa +ãģ®ä¸Ń ãģ§ +Ġп и +Ġng Ãłnh +ним а +ĠÙĩ ÙĦ +Ġ×ķ ×IJת +ĠÄij áng +é quipe +ĠÑįÑĤ оÑĤ +Ġgö rev +ë§ ¤ +Ġqu ân +å¼ķ ãģį +æĻĤ ãģ« +Ġب Ùħا +×ŀ ×Ļת +Ġü lke +Ġ×ŀ×§ ×ķ×Ŀ +×ij ף +æ°Ĺ æĮģãģ¡ +Ġë§İ ìĿĢ +Ġyük sek +ÑĨ енÑĤÑĢ +ĠÙħ جÙĦس +ç§ģ ãģ® +ÙĤد ر +Ġë¶Ģ ë¶Ħ +Ġì° ¨ +خر ج +ãģĭ ãģªãĤĬ +ë³´ ëĭ¤ +Ġ×ŀ ×Ļ×ĵ×¢ +peÅĤ ni +Ġx á»Ń +ìĹIJìĦľ ëĬĶ +ĠباÙĦ Ùħ +ĠÙĪ Ùħا +ĠÑįÑĤ ой +ب ÙĬÙĨ +n ü +ØŃ ز +ØŃز ب +ĠÑĢабоÑĤ а +ĠNh áºŃt +ÙĦ اء +Ġëĵ ¤ +Ġëĵ¤ ìĸ´ +ãĤĦãģĻ ãģĦ +×Ĺ×ĸ ×§ +Ġ×Ķ×Ĺ ×ijר×Ķ +п иÑĤ +ãģĭãĤī ãģ® +Ġë§IJ ìĶĢ +Ġפ ×ķ +ÙĦ Ùİ +à¹Ģà¸ķà¹ĩ ม +ĠÐļ о +Ġm ówi +Ġt ÃŃn +ר×Ĵ ש +פר ×§ +Ġtr ạng +ĠÐŀ н +×Ĺ ×ķ×¥ +ĠعÙĨد Ùħا +Ġب ر +使 ãģĦ +Ġr á»Ļng +ëĮĢ ë¡ľ +íĪ ¬ +Ġktóry ch +в ид +ลูà¸ģ à¸Ħà¹īา +Ġmog Äħ +Ġש ×Ĺ +×ij ×Ĺר +ãĥĸ ãĥŃãĤ° +ĠTh Ãłnh +Ġ×Ķ ×¨×Ļ +ĠÑģÑĤ аÑĤÑĮ +ĠH á»Ļi +à¸ļ à¹īาà¸ĩ +çī¹ ãģ« +ĠÄIJ ức +èĢħ ãģ® +×¢ ×ŀ×ķ×ĵ +×ĺר ×Ķ +Ð ¥ +ĠÙħ Ùħا +Ġe ÅŁ +ĠнеобÑħодим о +ник ов +Ġüzer inde +a ÅĤa +Ġchá»ĭ u +ĠاÙĦ دÙĬÙĨ +أخ بار +ĠÄij au +ãģĮ å¤ļãģĦ +jÄħ cych +د Ø®ÙĦ +ları nd +larınd an +Ġs ẻ +à¸ŀิ à¹Ģศ +à¸ŀิà¹Ģศ ษ +ת ף +t ıģı +Ġlu áºŃt +ĠÅŀ e +ãĤ« ãĥ¼ +ãģ® ãģĤãĤĭ +Ġ×Ķ×IJ תר +ĠاÙĦØ¢ ÙĨ +ıld ı +Ġá o +ĠнаÑĩ ал +Ġvi á»ĩn +Ġ×ij×¢ ×ķ׾×Ŀ +з наÑĩ +×Ļ×ĺ ×Ķ +к ам +ĠÐĺ з +à¹Ģà¸Ĥ ียà¸Ļ +à¸Ļ à¹īà¸Ńà¸ĩ +ÑĤ ÑĢо +à¹Ģ à¸Ł +Ġжиз ни +Ġ สà¹Īวà¸Ļ +Ġv áºŃn +Ġê´Ģ 볨 +Ġl âu +ס ×ĺר +×§ ש +س ÙĬر +Ġ×IJ×ķת ×Ļ +Ġm ôi +ائ ب +Ġо ÑģÑĤа +Ġm ón +Ġ×ij ×ŀ×§×ķ×Ŀ +Ġد اخÙĦ +Ġ×IJ ×ķר +Ġв аÑģ +Ùĥ Ø´Ùģ +ìĺ ¨ +à¸ĸ à¹Īาย +Ġkullan ıl +Ġt ô +ãģ« ãĤĪãĤĬ +ĠëĺIJ íķľ +Ġ×¢×ij×ķ×ĵ ×Ķ +Ġri ê +Ġriê ng +Ġyak ın +ز ا +Å » +×IJ ×ķ׼׾ +شار Ùĥ +Ġб еÑģ +× ´ +Ġا بÙĨ +ĠTá»ķ ng +ÙĨ ظ +ÅĽwi ad +ãĤµ ãĥ¼ +ห าย +ĠG ün +Ġhakk ında +à¹Ģà¸Ĥà¹īา มา +ز ÙĨ +ĠÐł о +Ġbi á»ĥn +ãģ© ãģĵ +Ùģ Ø¹ÙĦ +ز ع +פר ×ĺ +Ġ×Ķ ×Ł +Ø£ ÙĩÙĦ +Ġth ất +ØŃ ÙħÙĦ +Ñĩ Ñĥ +ĠìĤ¬ ìĭ¤ +ì° ¸ +ĠìľĦ íķ´ +ÙĪ Ø¸ +ĠÐŁ од +Ġkho ản +ÑĤ ен +ĠÙģ Ø§ÙĦ +Ñģ ад +à¸Ļ à¸Ńà¸Ļ +ĠاÙĦسعÙĪØ¯ ÙĬØ© +" ØĮ +ĠاÙĦ ÙĴ +ãĤī ãģļ +Ġto án +Ġch ắc +׼ ×Ļר +m éd +méd ia +ز ÙĪ +Ġyan ı +פ ׳×Ļ×Ŀ +ØŃ ظ +Ġб еÑģп +ĠбеÑģп лаÑĤ +ĠбеÑģплаÑĤ но +ĠØ£ ÙħاÙħ +à¸Ń าย +à¸Ńาย ุ +ר שת +Ġg á»ĵ +Ġgá»ĵ m +Ġu á»ijng +ص ب +k ır +ãĥij ãĥ¼ +Ġ׾×ĵ עת +Ġк ÑĥпиÑĤÑĮ +׾ ×ķ×Ĺ +ÙĪØ¶ ع +ÙĤÙĬ Ùħ +à¸Ľ า +ж ив +à¸Ķ ิà¸Ļ +×IJ ×ķפ +à¹Ģล à¹ĩà¸ģ +ãĥĥ ãĥī +иÑĩеÑģки Ñħ +ĠCh á»§ +кÑĢ Ð°Ñģ +ÙĪ ØµÙĦ +p ÅĤat +м оÑĢ +Ġ×Ķ×IJ ×ķ +à¸Ń ิà¸Ļ +Ġíķľ êµŃ +гÑĢ Ðµ +Ġìłľ ê³µ +ì° ½ +Ġê°ľìĿ¸ ìłķë³´ +Ġngh á»ĭ +à¸ĭ า +ØŃس اب +Ġby ÅĤa +ÙħÙĦ Ùĥ +иÑĩеÑģки е +Ġb ác +ض ØŃ +ê¸ ¸ +ש ×ŀ×¢ +Ġìĸ´ëĸ » +Ġìĸ´ëĸ» ê²Į +ìĽ Į +ات Ùĩ +à¹Ĥรà¸ĩ à¹ģ +à¹Ĥรà¸ĩà¹ģ รม +خد ÙħØ© +ĠÐł а +׼×ķ׾ ×Ŀ +×ŀש ×Ĺ×§ +ĠÙĪ ÙĥاÙĨ +ס ×ķ×£ +ĠاÙĦØŃÙĥÙĪÙħ Ø© +Ġ×ij ×ĺ +Ġtr áºŃn +Ġ×Ķ×¢ ×ķ׾×Ŀ +ĠÃŃ ch +t Äħ +ש×ŀ ×ķ +Ġ×Ķר×IJש ×ķף +Ġíķĺ ê³ł +ãģķ ãĤī +ãģķãĤī ãģ« +ãģ« ãģĹãģ¦ +Ġ à¸ľà¸¡ +ãģ® ãĤĪãģĨãģª +ĠÙĪ ÙĤت +ãĥį ãĥĥãĥĪ +ÙĦ عب +ÙĪ Ø´ +ìĺ ¬ +Ġ หาà¸ģ +Ġm iaÅĤ +à¸Ĺ à¸Ńà¸ĩ +иÑĤ а +ا صر +ил ÑģÑı +з е +à¸Ľà¸£à¸° มาà¸ĵ +ãģĿãĤĮ ãģ¯ +Ġb ır +Ġbır ak +صÙĨ اع +Ð ® +Ø´ عر +Ġ׳ ×Ĵ×ĵ +Ġب سبب +ãĥĿ ãĤ¤ +ãĥĿãĤ¤ ãĥ³ãĥĪ +ĠاÙĦج ÙĪ +ĠнеÑģк олÑĮко +Ġki ếm +Ùģ Ùİ +Ġض د +×ij×Ļ×ĺ ×ķ×Ĺ +تاب ع +ÙĨ ز +ĠB ản +Ġaç ıkl +Ġaçıkl ama +Ġ à¸Ħุà¸ĵ +à¸Ĺ า +ÅĤ ów +Ø· ب +ÙĨ ØŃÙĨ +Ġ×ŀ×§ ×ķר +Ġİ s +Ġдом а +Ġ วัà¸Ļ +Ġd Ãłnh +Ñı н +ми ÑĢ +Ġm ô +ĠvÃł ng +ص اب +s ının +à¸Ħ ืà¸Ļ +Ø® بر +×ĸ׼ ×ķ +Ġ×ŀ ש×Ķ×ķ +m ü +Ġкомпани и +Ġ×Ķ×¢ ×Ļר +ĠÙĥ ÙĪ +ÙĤÙĦ ب +ĠlỼ p +и ки +׳ ×ij +à¹Ĥ à¸Ħร +à¹Ĥà¸Ħร à¸ĩ +à¹Ĥà¸Ħรà¸ĩ à¸ģาร +×ŀ×ķ×¢ ×ĵ +ÑıÑĤ ÑģÑı +หลัà¸ĩ à¸Īาà¸ģ +ени Ñİ +Ġש ×¢ +Ġb Æ°á»Ľc +ãĥ¡ ãĥ¼ãĥ« +ãĤĦ ãĤĬ +Ġ×Ļ×ķ×ĵ ×¢ +Ġê´Ģ íķľ +ĠاÙĦØ£ Ùħر +Ġböl ge +ĠÑģв ой +ÙĦ س +Ġ×ŀ×Ļ ×ķ×Ĺ×ĵ +ĠëĤ´ ìļ© +ĠØ£ جÙĦ +ĠÄIJ ông +Ġ×ŀ ×ł×ª +Ġìĭľ ê°Ħ +Ùĥ Ùİ +ãģ¨ãģĦãģĨ ãģ®ãģ¯ +Ġnale ży +تÙĨظ ÙĬÙħ +ĠÑģозд а +Ġph é +Ġphé p +ãģ§ãģį ãģ¾ãģĻ +Ġع ÙĦÙħ +大ãģį ãģª +ãĤ² ãĥ¼ãĥł +í ħĮ +Ġ׼×ķ׾ ׾ +ĠинÑĤеÑĢ Ð½ÐµÑĤ +ĠT ừ +ãģ¨ ãģªãĤĭ +ز اÙĦ +Ġktóry m +Ġnh é +ìĪ ľ +н ев +д еÑĢ +ãĤ¢ ãĥĹãĥª +i á»ĩu +×ij ×Ļ׾ +Ġت س +ĠÄIJ ây +ĠاÙĦØ® اصة +Ġà¹Ģ à¸Ĭ +Ġà¹Ģà¸Ĭ à¹Īà¸Ļ +ص اد +Ġd ạng +س عر +Ġש ×Ļ×ŀ×ķש +×Ĵ ×Ļ×Ŀ +ãģĮãģĤ ãģ£ãģŁ +п ÑĢов +пÑĢов од +Ġ×IJ ×Ļ׳×ķ +Ġ׾ ר×IJ +Ġ׾ר×IJ ×ķת +ĠØ£ Ù쨶ÙĦ +ĠØŃ ÙĦ +ĠØ£ بÙĪ +ê° ķ +Ġì§ ij +ãģ® ãĤĪãģĨãģ« +Ġפ ׳×Ļ +ס ×Ļ×Ŀ +ĠÙĪÙĩ ذا +Ġka ç +Ġé én +Ġê± ´ +ë° Ķ +Ñĥ з +à¸Ĥà¸Ńà¸ĩ à¹Ģรา +i ÅĤ +ĠÐľ Ñĭ +Ġch ết +ĠاÙĦØ« اÙĨÙĬ +×IJ ×§ +Ġ×ķ ×¢×ľ +ĠاÙĦØ· ب +×ij×ĺ ×Ĺ +Ġج دÙĬدة +Ġع دÙħ +ع ز +สิà¹Īà¸ĩ à¸Ĺีà¹Ī +ãģĻ ãĤĮãģ° +ĠÄij ô +ì£ ł +د ÙĤ +н омÑĥ +Ġk á»ĥ +ãĤ¢ ãĥ³ +å¤ļãģı ãģ® +à¸Ľà¸£à¸° à¸ģ +à¸Ľà¸£à¸°à¸ģ à¸Ńà¸ļ +פע×Ļ׾ ×ķת +ĠÑģÑĤ ол +may ı +ãģ¤ ãģĦ +Ġyılı nda +Ġ à¸Īึà¸ĩ +koÅĦ cz +ĠTh ông +Ġак ÑĤив +н ÑģÑĤ +нÑģÑĤ ÑĢÑĥ +ĠÃĸ z +Ġת ×ŀ×Ļ×ĵ +ĠÙĥ ÙĨت +Ñģ иÑģÑĤем +pr és +prés ent +Ġn â +Ġnâ ng +gÅĤ os +ĠÙĪØ² ÙĬر +ØŃ صÙĦ +Ġиме еÑĤ +ØŃ رÙĥØ© +à¸ŀ à¹Īà¸Ń +ãĤĴ ãģĬ +Ġاست خداÙħ +×IJ×Ļר ×ķ×¢ +ä»ĸ ãģ® +Ġש×Ķ ×Ŀ +ãģĹãģŁ ãĤī +ש×ŀ ×Ļ +Ñģ ла +m ı +Ġbaz ı +Ġíķĺ ì§Ģë§Į +×ĵ ׾ +Ġyapt ıģı +ãĥĬ ãĥ¼ +׾ ×Ļ׾×Ķ +ãģ¨ãģĦ ãģ£ãģŁ +änd ig +ĠÅŁ a +ĠÙģÙĬ Ùħا +иÑĤ елÑı +×ŀ ×ķש +à¸Ĥ à¸Ńà¸ļ +l ük +Ġh á»ĵi +Ġëª ħ +ĠاÙĦÙĥ Ø«ÙĬر +צ ×IJ +Ġhaz ır +طر Ùģ +ا ÙĬا +ĠÄij ôi +ен д +ÙĦ غ +×Ĺ ×ĸ×ķר +ĠвÑģ ег +ĠвÑģег да +ëIJĺ ê³ł +×ĵ ×ķ×ĵ +ан а +د ÙĪÙĦØ© +Ġho ạch +ع ÙĦا +عÙĦا ج +Ġ×ķ ×¢×ĵ +×Ķ ×Ŀ +ки й +ÙĦ ÙIJ +Ġ×¢ ׾×Ļ×ķ +ÑİÑī ий +Ġng á»§ +صÙĨ ع +ĠاÙĦع راÙĤ +à¸ķà¹Īà¸Ń à¹Ħà¸Ľ +ãģŁãģı ãģķãĤĵ +Ġph ạm +ÙĦ اÙĨ +ات Ùĩا +Ġbö yle +تÙĨ ÙģÙĬ +تÙĨÙģÙĬ ذ +Ġש×Ķ ×Ļ×IJ +Ñģ Ñĥ +ย าว +Ġש ×ķ׳×Ļ×Ŀ +Ġ×ŀ ×ķ׾ +ĠÑģ ил +Ġ×IJ×Ĺר ×Ļ×Ŀ +Ġph á»§ +ÙĤØ· ع +ĠTh á»§ +à¸Ľà¸£à¸°à¹Ģà¸Ĺศ à¹Ħà¸Ĺย +ÙĨ ÙĤ +ĠÄijo ạn +Ġب Ø¥ +п ÑĢедел +×ķת ×ķ +Ġy arı +пÑĢ Ðµ +ĠczÄĻ ÅĽci +ØŃ ÙĥÙħ +×ķ׳ ×Ļת +פע ׾ +ãĤĴ ãģĹãģ¦ +Ġktó rzy +׾ ×Ŀ +ĠÄIJi á»ģu +ĠкоÑĤоÑĢ Ð°Ñı +ĠìĿ´ ìĥģ +ãģĤ ãģ£ãģŁ +Ġ×ŀ×ĵ ×ķ×ijר +פ ×ķ×¢×ľ +d ım +éĢļ ãĤĬ +ĠбÑĥд ÑĥÑĤ +à¹Ģวà¹ĩà¸ļ à¹Ħà¸ĭ +à¹Ģวà¹ĩà¸ļà¹Ħà¸ĭ à¸ķà¹Į +ا خر +×Ĺ ×Ļ׾ +Ġ×Ļ ×ľ +Ġ×Ļ׾ ×ĵ×Ļ×Ŀ +×Ĺ ×Ļפ +×Ĺ×Ļפ ×ķש +Ġd òng +Ġש ×ĸ×Ķ +ÑĮ е +ãģĤ ãģ¨ +ìŀIJ ê°Ģ +×IJ ×ĵ +Ġü z +Ġüz ere +ظ ÙĦ +Ġ×IJ ×ķ׾×Ļ +Ġ×ij ×Ļ×ķ×Ŀ +ÙĦ ات +Ġm ê +ì¹ ¨ +تØŃ د +تØŃد Ø« +ĠØ® اصة +Ġب رÙĨ +ĠبرÙĨ اÙħج +ĠH Ãłn +×Ĺ ×¡ +ĠÙĪ ÙĦÙħ +×¢ ×Ŀ +Ġm ı +à¸Ł ัà¸ĩ +ש ×¢×Ķ +ÙĪÙģ ÙĤ +ס ×ij×Ļר +алÑĮ нÑĭй +×Ĺש ×ķ×ij +Ġn Ãłng +ë³ ¼ +ĠкоÑĤоÑĢ ÑĭÑħ +Ġ×Ĺ ×ķ×§ +t ör +ĠлÑĥÑĩ ÑĪе +ãĥij ãĥ³ +ลà¹Īา สุà¸Ķ +Ġج دÙĬد +ÙĬد Ø© +à¸Ĺ รà¸ĩ +ãĤĪãĤĬ ãĤĤ +ÙĦ ÙĦ +ãĤĤ ãģ£ãģ¨ +ש×ĺ ×Ĺ +Ġ×ķ ×IJ×Ļ +Ġgi á»ijng +Ø¥ ضاÙģ +×§ ת +ë§ Ŀ +Ġzosta ÅĤ +ÑĢ Ð¾Ð· +×Ļפ ×Ļ×Ŀ +Ġ׼׾ ׾ +ת×ķ׼ ף +dıģ ını +ÙĤ سÙħ +ĠÑģ ÑĩиÑĤ +ĠÑģÑĩиÑĤ а +×ĺ ×ķת +Ġ ưu +ĠØ¢ ÙĦ +Ġм ом +Ġмом енÑĤ +ĠاÙĦتع ÙĦÙĬÙħ +×¢×ľ ×ķת +Ġch ữa +Ġy ön +Ġtr Ãł +ĠØŃ ÙĬÙĨ +à¸ĭ ั +ĠC á +×¢ ×ĸ +ĠاÙĦØ£ ÙħÙĨ +c ÃŃ +Ġv á»ijn +Ġ à¸Ļาย +об ÑĢа +×§ ×IJ +Ġthi ếu +ãĥŀ ãĥ¼ +ส วà¸Ļ +Ġg á»Ń +Ġgá»Ń i +Ġê ¹ +Ġê¹ Ģ +Ġthi á»ĩn +ÙĤ ع +w ÄĻ +Ġн ам +ÑĤ ол +Ġs ân +ס ×ķ×Ĵ +Ġgeç ir +ÑĤ он +ев а +ĠÙĪ Ø¶Ø¹ +Ġع شر +Ñģ ло +à¸Ī ัà¸ļ +ãĤ· ãĥ¼ +ãĤĤ ãģĤãĤĬãģ¾ãģĻ +Ġv ẻ +ĠÄIJ á»ĥ +ر Ù쨹 +ĠاÙĦØ£ÙĪÙĦ Ùī +ÑĤ аÑĢ +ãģªãģı ãģ¦ +Ùħ Ùİ +qu ÃŃ +×¢×ł×Ļ ×Ļ׳ +г ен +Ġh ôm +à¸Ī า +Ġnh Ỽ +ĠاÙĦع ربÙĬ +×IJ ף +Ġl á»Ļ +Ġje ÅĽli +à¹Ģà¸Ĺà¹Īา à¸Ļัà¹īà¸Ļ +ĠØ£ÙĨ Ùĩا +Ġt uy +Ġtuy á»ĩt +Ġت ص +Ġتص ÙĨÙĬ +ĠتصÙĨÙĬ Ùģ +Ġê·¸ëŁ¬ ëĤĺ +о ÑĨен +à¸ģิà¸Ī à¸ģรรม +ãĤĦ ãģ£ãģ¦ +Ġkh á»ıi +Ġl á»ĩ +ĠاÙĦÙħج تÙħع +à¸Ńาà¸Ī à¸Īะ +à¸Īะ à¹Ģà¸Ľà¹ĩà¸Ļ +ов Ñĭй +ר ×Ŀ +ร à¹īà¸Ńà¸Ļ +ש ×ŀש +人 ãģ« +Ġüzer ine +פר ×Ļ +du ÄŁu +Ñĩ ик +Ġmù a +Ġ×ŀת ×ķ×ļ +Ġc áºŃp +Ġت ارÙĬØ® +×ij׾ ת×Ļ +Ġì¢ Ģ +ÙĦ ع +ب اÙĨ +Ġch út +Ġ×Ķ×ĸ ×ŀף +n ée +ĠLi ên +ĠÙĦÙĦ Ø£ +ØŃد ÙĪØ¯ +Ġ×¢ ׼ש×Ļ×ķ +в оз +Ġyapt ı +Ġоб о +à¹ĥหà¹ī à¸ģัà¸ļ +Ġ×ij×Ķ ×Ŀ +ãģı ãģ¦ +ر أس +ĠÑģÑĢед ÑģÑĤв +ĠB Ãłi +ãģĵãģ¨ ãģ« +ĠìĤ¬ íļĮ +Ġ모 ëijIJ +×ij ×IJ +Ġtr ắng +ĠاÙĦبÙĦ د +ĠHo Ãłng +ли бо +ĠдÑĢÑĥг иÑħ +İ R +Ñĥм а +ĠJe ÅĽli +ãĤĤ ãģĹ +Ġv òng +Ġ×IJתר ×Ļ×Ŀ +ĠÄij á»įc +Ġв оÑĤ +ãģł ãģĮ +ë° ° +à¸Ķู à¹ģล +Ġ×ŀ ׼׾ +ìĹIJ ëıĦ +г аз +Ġ׳×ķס פ×Ļ×Ŀ +ãģĵãģ¨ ãģ§ +Ġت ÙĪ +ãģ§ ãģĤãĤĬ +à¸Ļั à¹Īà¸ĩ +ĠможеÑĤ е +sz ÄĻ +ãģ® ãģł +ĠÙħÙĨ Ùĩ +Ġb á»ķ +Ġb üt +Ġbüt ün +ë³´ ê³ł +Ġch á»ĵng +à¹ģà¸Ī à¹īà¸ĩ +ĠV ì +ĠØŃ ر +Ġgi ản +ĠÙħ دÙĬÙĨØ© +تط بÙĬÙĤ +à¸Ī ิ +æĹ¥ ãģ® +б ил +à¸ģ à¸Ńà¸ĩ +ê³ ³ +ĠØ£ Ùħا +ìĨ IJ +Ġtr ái +ĠвÑģ ем +Ġس ÙĨØ© +ĠÑģай ÑĤ +Ġг оÑĤов +п Ñĭ +ĠëIJ ł +ĠاÙĦØ® Ø· +ĠاÙĦرئÙĬس ÙĬØ© +Ġíķ ©ëĭĪëĭ¤ +ĠìķĦëĭĪ ëĿ¼ +ĠìĿ´ ëłĩ +ĠìĿ´ëłĩ ê²Į +) ØĮ +h ält +ĠØ£ Ùħر +Ġع Ùħر +à¸ģà¹ĩ à¸Īะ +Ġ à¸Ĺำà¹ĥหà¹ī +Ġc ân +Ġ×ij ׾ +Ġ×ij׾ ×ij×ĵ +פ סק +ĠÙĬ ÙĤÙĪÙĦ +н ÑĥÑĤÑĮ +à¹ģ à¸Ħ +Ġ×§ צת +Ġn ằm +Ġh òa +bilit Ãł +ĠìĹĨ ëĭ¤ +Ġ׼ פ×Ļ +ÑĢ Ð¾Ð¶ +лаг а +Ġ×Ķש ×Ļ +ĠNgo Ãłi +ĠÙĪ Ø¬ +ĠÙĪØ¬ ÙĪØ¯ +ĠìľĦ íķľ +Ġus ÅĤug +Ġtu ần +d ź +×ŀ ×ķף +ĠاÙĦع دÙĬد +Ġch ẳng +สุà¸Ĥ à¸łà¸²à¸ŀ +Ġ×ij ×ĵר×ļ +ĠÑģеб е +ĠìŀĪ ìĿĦ +ĠاÙĦØŃ اÙĦ +Ġd á +Ġc ưá»Ŀi +Ġnghi ên +ie ÅĦ +ĠD ương +ï¼ ħ +Ø´ د +ãģĦãģ¤ ãĤĤ +ĠвÑĭб оÑĢ +Ġc á»Ļng +ש ×Ļ׳×ķ×Ļ +Ġch ạy +Ġ×ij×¢ ׾×Ļ +اخ بار +íķĺ ë©° +ż Äħ +ج از +Ġ׳ ר×IJ×Ķ +ศ ู +ศู à¸Ļ +ศูà¸Ļ ยà¹Į +×Ĵ ×¢ +Ġ×¢ ×ĵ×Ļ +Ġ×¢×ĵ×Ļ ×Ļף +بر ا +ÑĨи й +ĠÄIJ á»ĵng +ÙĤ اÙĨÙĪÙĨ +ĠÄij ứng +ãģĹãģŁ ãĤĬ +Ġ×Ĺ×Ļ ×Ļ +Ġë IJľ +ĠëIJľ ëĭ¤ +Ġм еждÑĥ +à¸ŀวà¸ģ à¹Ģà¸Ĥา +ĠB ắc +ล ำ +ë° ± +ĠíĻ ķ +มาà¸ģ ม +มาà¸ģม าย +бан к +à¸Ńา à¸ģาร +Ġh Ãł +Ġ׾ ׳ +à¸Ń à¸Ń +Ġë°Ķ ë¡ľ +л ом +m ática +ĠØŃ د +اب ت +à¸Ĺีà¹Ī à¸Ļีà¹Ī +Ġco ÅĽ +ÙģÙĬ دÙĬ +ÙģÙĬدÙĬ ÙĪ +ĠмеÑģÑĤ о +Ġph út +มาà¸ģ à¸ģวà¹Īา +×IJ פ +ب ÙIJ +ĠPh ú +ì± Ħ +ĠÙĪ Ø³ÙĦÙħ +à¸Īี à¸Ļ +поÑĤ ÑĢеб +Ġ×Ĺ×ĵ ש×ķת +Ø´ ÙĪ +Ġעצ ×ŀ×ķ +ĠعÙħÙĦ ÙĬØ© +à¸Ħุà¸ĵ à¸łà¸²à¸ŀ +ãģ¾ãģĻ ãģĮ +دع ÙĪ +طر ÙĤ +à¹Ħมà¹Ī à¸ķà¹īà¸Ńà¸ĩ +ë² Ķ +ìĬ ¹ +Ġk ÃŃch +ĠìĹĨ ëĬĶ +ĠÑĤ ам +ĠÙĨ ØŃÙĪ +ĠاÙĦÙĤ اÙĨÙĪÙĨ +×Ĺ ×ķ×Ŀ +Ġk ız +Ġ×ĵ ×Ļף +ĠвÑĢем ени +ãģ£ãģŁ ãĤĬ +ĠØ´ Ùĩر +ĠìĦľ ë¹ĦìĬ¤ +×¢ ש×Ķ +Ġgi ác +ĠاÙĦسÙĦ اÙħ +Ġ×IJ ש +ĠполÑĥÑĩ а +à¸Īัà¸Ķ à¸ģาร +к оÑĢ +Ġ×Ķ×ĺ ×ķ×ij +ราย à¸ģาร +주 ìĿĺ +à¹ģà¸ķà¹Ī ละ +Ġê·¸ëŁ° ëį° +à¸Ĺีà¹Ī à¹Ģà¸Ľà¹ĩà¸Ļ +Ġת ×ķ×ļ +بÙĬ اÙĨ +Ð Ļ +oÅĽci Äħ +ÑĤ ок +ĠÃ Ķ +ĠÃĶ ng +à¹Ħมà¹Ī à¹ĥà¸Ĭà¹Ī +ãģ¿ ãģ¦ +ÐŁ о +ĠЧ ÑĤо +íĻ © +×ĺ ×ij×¢ +меÑĤ ÑĢ +Ġ×ij ×ŀ×Ķ +Ġ×ij×ŀ×Ķ ×ľ +Ġ×ij×ŀ×Ķ׾ ×ļ +Ñĩ ÑĮ +×§ ש×Ķ +з нак +знак ом +uj ÄĻ +×Ļצ ר +ĠاÙĦÙħ ÙĦÙĥ +ı yla +×IJ×ŀ ת +à¸Ľ ิà¸Ķ +×IJ ×Ĺ×ĵ +ر اد +Ġm áºŃt +ëĭ¤ ëĬĶ +Ġl ạnh +ש׾ ×ķש +ØŃ دÙĬØ« +ت ز +å¹´ ãģ® +Ġк ваÑĢ +ĠкваÑĢ ÑĤиÑĢ +ä½ľ ãĤĬ +رÙĪ Ø¨ +ов ан +ĠТ е +à¸Īำ à¸ģ +à¸Īำà¸ģ ัà¸Ķ +ب اط +×Ĵ ת +Ġм аÑĪ +ĠмаÑĪ Ð¸Ð½ +×Ļצ ×Ķ +ãģ» ãģ¨ +ãģ»ãģ¨ ãĤĵãģ© +ÃŃ do +ĠÑı зÑĭк +à¸ļ ิà¸Ļ +สà¸ĸาà¸Ļ à¸Ĺีà¹Ī +ĠìĹ ´ +ãĤ¦ ãĤ§ +Ġc Ãł +п ан +åı£ ãĤ³ãĥŁ +Ġر د +اÙĤ ت +ĠÙĥ ب +ĠÙĥب ÙĬرة +ÑģÑĤ ал +ש×ŀ ×Ĺ +pos ición +ĠÙħÙĦÙĬ ÙĪÙĨ +ĠìĿ´ ìķ¼ +ĠìĿ´ìķ¼ ê¸° +Ġh út +ĠÅĽw iat +Ġë°© ë²ķ +ĠÑģв еÑĤ +Ġвиде о +ĠاÙĦÙĨ ظاÙħ +Ġtr á»Ŀi +ĠëĮĢ íķ´ìĦľ +ר ×ŀת +ت داÙĪÙĦ +×ķר ×ĵ +ת ×ŀ +ת×ŀ ×ķ׳×ķת +Ġ×ŀ ף +Ġдв а +Ġ×Ķ×§ ×ķ +æĹ¥ ãģ« +Ġ×Ķ×Ĵ ×Ļ×¢ +à¹Ģà¸ŀิà¹Īม à¹Ģà¸ķิม +Ùħار س +Ġê²ĥ ìŀħëĭĪëĭ¤ +ãģªãģĦ ãģ¨ +Ġnhi á»ĩt +ëIJ ©ëĭĪëĭ¤ +Ġ×ij׳ ×ķש×IJ +Ġê°Ģ ìŀ¥ +Ġv ợ +ĠÄij óng +צ×Ļ׾ ×ķ×Ŀ +ê´Ģ ê³Ħ +в аÑı +×IJ ×Ļ×ĸ +×IJ×Ļ×ĸ ×Ķ +ĠÙĨ ظاÙħ +ÙħØŃ اÙ쨏 +Ġt ải +기 ëıĦ +à¸Ľà¸±à¸Ī à¸Īุ +à¸Ľà¸±à¸Īà¸Īุ à¸ļัà¸Ļ +׼ ×ĵ×ķר +ĠìķĦ ìĿ´ +׼׳ ×Ļס +à¹Ģ à¸ķร +à¹Ģà¸ķร ียม +Ġngo ại +ĠدÙĪÙĦ ار +Ġr ẻ +Ġkh Äĥn +عد د +Ø´ عب +czy Äĩ +ĠاÙĦ Ùĥر +ĠÑĩеловек а +ĠÙĪ Ø¥ÙĨ +×IJ ×ĺ +Ġth Æ¡ +ĠاÙĦ رÙĬاض +оп ÑĢедел +опÑĢедел ен +×Ķ ×ŀש×ļ +ĠÐĿ ово +з Ñĭва +ĠاÙĦدÙĪÙĦ ÙĬ +ĠÄij áp +Ġк ÑĢед +ĠкÑĢед иÑĤ +ов ого +Ġm ôn +à¸Ľà¸£à¸° à¹Ĥย +à¸Ľà¸£à¸°à¹Ĥย à¸Ĭà¸Ļ +à¸Ľà¸£à¸°à¹Ĥยà¸Ĭà¸Ļ à¹Į +ÑģÑĤ е +ĠTh á»ĭ +د ÙĬØ© +×ŀצ ×ķ +Ùģ Ø§Øª +×§ ×ĵ×Ŀ +ìĿ´ëĿ¼ ê³ł +ÙĪ Ø® +Ġ×Ĺ ×ĸ +ĠÑĦоÑĤ о +׾ ×Ļת +ت Ùİ +ÙĪ Ø¨Ø± +й ÑĤи +ĠÃ¶ÄŁ ren +Ġ×Ķ×ĸ ×ķ +Ġv á»įng +ÙĤÙĪ Ø© +ĠT ây +ĠÐĿ и +Ġש ×ķ×ij +ãģ¨è¨Ģ ãĤıãĤĮ +ãģ© ãĤĵãģª +׊צ×Ļ +ï½ ľ +Ġ×ķ×Ķ ×ķ×IJ +ä¸Ģ ãģ¤ +ĠÑģÑĤо иÑĤ +ni Äħ +×ĺר ×Ļ +ĠдеÑĤ ей +нÑı ÑĤÑĮ +ĠÑģдел аÑĤÑĮ +Ġë§İ ìĿ´ +ä½ķ ãģĭ +ãģĽ ãĤĭ +à¹Ħ หม +à¸ķิà¸Ķ à¸ķà¹Īà¸Ń +Ġ×ij ת×Ĺ +Ġ×ijת×Ĺ ×ķ×Ŀ +ìĻ Ħ +ì§Ģ ëĬĶ +ÑģÑĤ аÑĤ +ÑıÑģ н +ü b +Ġth ả +Ġ×ij×IJ×ŀ ת +Ġt uyến +×ĵ ×Ļר×Ķ +Ġ×IJ ×Ļש×Ļ +×ĸ׼ ר +ãģ° ãģĭãĤĬ +Ġx ét +׼ ×Ļ×ķ +׼×Ļ×ķ ×ķף +diÄŁ ini +ĠاÙĦÙħ ÙĪØ¶ÙĪØ¹ +Ġh áºŃu +à¸Īาà¸ģ à¸ģาร +×ijס ×Ļס +Ġ×ŀ×Ĵ ×Ļ×¢ +×ij ×Ļ×¢ +ĠÙĪ Ø¬Ùĩ +à¹ģà¸Ķ à¸ĩ +à¸Ļ าà¸ĩ +ĠÅŀ a +ì ¡´ +ë¡ Ģ +à¸ķ ะ +Ġ×Ķ×Ĺ×Ļ ×Ļ×Ŀ +Ùģ ÙĬد +ãģ§ãģĻ ãģĭãĤī +ê· ľ +ź ni +ĠлÑİ Ð´ÐµÐ¹ +Ġyüz de +ıy orum +ĠاÙĦ بØŃر +e ño +п аÑĢ +ÙĬ ÙĤØ© +об ÑĢ +ר ×ķ×ļ +ت ÙĪÙĤع +ĠاÙĦØ´ ÙĬØ® +åĪĿ ãĤģãģ¦ +ĠÑĤ елеÑĦ +ĠÑĤелеÑĦ он +Ġth ôi +Ġ×Ļ׼×ķ׾ ×Ļ×Ŀ +ĠÅŁ irk +ĠÅŁirk et +Ġìļ°ë¦¬ ê°Ģ +ĠÄij ông +Ġת ×ķ×ĵ×Ķ +ÑģмоÑĤÑĢ ÐµÑĤÑĮ +ĠÙĦ ÙĩÙħ +Ġ׾ ׼ +ĠN ó +ĠØŃ اÙĦØ© +ãģĦ ãģij +קר ×ķ +az ı +ãĤ³ ãĥ¼ +ĠÙĦÙĦ ت +s ınız +ĠH ải +기 ìĪł +ยัà¸ĩ à¹Ħมà¹Ī +ëĭ¤ ê³ł +פ ×Ĺ +Ġ׾×Ĵ ×ij×Ļ +Ġع ÙĨÙĩ +Ġк аз +Ġказ ино +ب ÙĪØ± +ÑĦ еÑĢ +Ġê°Ļ ìĿ´ +تس جÙĬÙĦ +ĠاÙĦÙħ رÙĥز +ĠTh ái +д аÑĤÑĮ +×ŀ×Ļ ×Ļ׾ +Ġpay laÅŁ +ãģ¤ ãģ® +à¹Ģร ืà¸Ń +n ça +׳ ×ķ×Ĺ +Ġ×IJ פ×Ļ׾×ķ +ãģ¨ èĢĥãģĪ +ãģ¨ãģĹãģ¦ ãģ¯ +à¹Ģà¸Ī à¸Ń +×ŀ פ +Ġg iriÅŁ +л иÑĤ +ÑĤ елÑı +Ñij н +æ°Ĺ ãģ« +Ġg ó +Ġgó p +åĪĩ ãĤĬ +Ġ×Ķ ×Ĺ×ĵש +ж ал +Ġ×ĵ עת +éģķ ãģĨ +à¹Ģà¸Ĥà¹īา à¹Ħà¸Ľ +Ġס ר×ĺ +e ña +æĸ° ãģĹãģĦ +ر Ùİ +ĠÐIJ ÑĢ +Ġph ản +à¸Īะ à¹Ħà¸Ķà¹ī +Ġ×ijצ ×ķר×Ķ +Ø´ اÙĩ +شاÙĩ د +ÙĪØ± د +à¹Ģà¸Ļืà¹Īà¸Ńà¸ĩ à¸Īาà¸ģ +или ÑģÑĮ +à¹ģละ à¸ģาร +Ġ×Ķ ×ĸ׼ +Ġ×Ķ×ĸ׼ ×ķ×Ļ×ķת +ei ÃŁ +ãĥ ¨ +ìĥ Ī +ĠÃĩ a +Æ ¯ +ש ×Ĵ +ÙĬÙĨ Ø© +ร à¹īà¸Ńà¸ĩ +ãĤµ ãĥ³ +ÑĢоÑģÑģ ий +ÑĢоÑģÑģий Ñģк +a ÄŁa +ĠнаÑĩ ина +Ġص ÙĦÙī +à¸Ĺุà¸ģ à¸Ħà¸Ļ +íļĮ ìĤ¬ +Ġли ÑĨ +Ø´ ÙĬر +ĠØ´ÙĬ Ø¡ +ÙĬÙĨ ا +Ġפ ×Ĺ×ķת +Ġiçer is +Ġiçeris inde +ĠØ£ ØŃÙħد +Ġże by +ì´ Ŀ +Ġп оказ +Ġи менно +หà¸Ļัà¸ĩ ส +หà¸Ļัà¸ĩส ืà¸Ń +ĠÑĤÑĢ Ðµ +สัà¸ĩ à¸Ħม +Ø¥ ÙIJ +ãģĮ å¿ħè¦ģ +ÙĬÙij Ø© +פ צ +íĭ ° +ĠÙħ جاÙĦ +׳ פש +к ан +×Ĺ ×ķפ +×Ĺ×ķפ ש +ì²ĺ ëŁ¼ +ов аÑı +з ов +Ġh ạ +Ġdzi ÄĻki +×Ļר ×ķ +Ġ׾ ×ŀצ +Ġ׾×ŀצ ×ķ×IJ +×Ļ×ĵ ×ķ +Ġs ợ +Ġ׾×Ķ ×Ĵ×Ļ×¢ +×§ ×ij×¢ +Ġchi á»ģu +ãĥŀ ãĤ¤ +Ġd Ãłng +à¹ģà¸Ł à¸Ļ +Ġü ye +×Ļ׳ ×Ĵ +à¹Ģรีย à¸ģ +ç§ģ ãģĮ +th é +ĠÑĦ илÑĮ +ĠÑĦилÑĮ м +ĠNg Ãły +Ġж ен +Ġжен Ñīин +ج ÙĬد +n ç +à¸Ľ รา +×Ļ×ŀ ×ķ +Ġn á»ģn +×IJ ×ķ׾×Ŀ +Ġвозмож ноÑģÑĤÑĮ +Ġëĭ¤ ìĭľ +è¦ĭ ãģŁ +à¸ĸ à¸Ļ +à¸ĸà¸Ļ à¸Ļ +mız ı +ĠÙħ جÙħÙĪØ¹Ø© +c jÄħ +ĠÐł Ф +à¸ģำ หà¸Ļ +à¸ģำหà¸Ļ à¸Ķ +ĠìŬ 기 +land ı +ни ÑĨ +ÑģÑĤв е +Ġ×ĵ ×ijר×Ļ×Ŀ +Ġsk ÅĤad +ãĤĬ ãģ¾ãģĹãģŁ +ĠоÑĤ кÑĢÑĭÑĤ +нÑı ÑĤ +ĠÑģво ей +à¸Ī ิà¸ķ +ĠкаÑĩеÑģÑĤв е +Ġet tiÄŁi +ìĤ¬ íķŃ +ĠاÙĦÙĬ ÙħÙĨ +иÑĩеÑģки й +ë¸ Į +Ġ×ij×IJר ×¥ +Ġا سÙħ +Ġиз веÑģÑĤ +r ão +Ġatt ivitÃł +à¹Ģà¸Ľà¹ĩà¸Ļ à¸ģาร +ĠاÙĦد Ùĥت +ĠاÙĦدÙĥت ÙĪØ± +ĠÙĪØ§ØŃد Ø© +ĠÑģ ÑĩеÑĤ +ĠпÑĢ Ð¸Ñĩ +ĠпÑĢиÑĩ ин +ĠÙĪØ² ارة +Ġh uyá»ĩn +ĠÙĥ تاب +à¹ģà¸Ļ à¹Īà¸Ļ +à¹ģà¸Ļà¹Īà¸Ļ à¸Ńà¸Ļ +Ġgün ü +г ÑĢÑĥз +ĠاÙĦØ® اص +Ġgör ül +׾ ×ŀ×ĵ +Ġìłķ ëıĦ +×ķ×ij ×Ļ׾ +Ġ×ŀ×§ צ×ķ×¢×Ļ +ĠоÑģоб енно +à¸Ľà¸£à¸° à¸ģา +à¸Ľà¸£à¸°à¸ģา ศ +aca ģını +ë¶ ģ +à¸łà¸¹ มิ +ĠÑį лекÑĤ +ĠÑįлекÑĤ ÑĢо +Ġ×§ ש×Ķ +سÙĦ Ø· +à¸Ĭà¸Ļ ะ +×¢ ×Ļ׾ +ĠЧ е +à¹ģà¸Ļ à¹Ī +lı ÄŁ +lıģ ın +Ġ×ŀ×¢ ×¨×Ľ×ª +好ãģį ãģª +มาà¸ģ à¸Ĥึà¹īà¸Ļ +×ŀ×¢ ×ijר +ĠاÙĦÙħ غرب +ĠпеÑĢ Ð¸ +ĠпеÑĢи од +Ġnh ạc +ا ÙĪÙĬ +ĠÙĪ Ø¹ÙĦÙī +أخ ذ +ĠC ô +תר ×ij×ķת +×Ĵ ×Ķ +Ġktóre j +×IJ ×Ļת +×ij ×ķ×IJ +д елÑĮ +รี วิ +รีวิ ว +ж Ñĥ +Ġ×ij×Ĺ ×ķ +еÑĪ ÑĮ +ĠØ£ ÙĦÙģ +ĠاÙĦÙĪ Ø·ÙĨÙĬ +ĠاÙĦÙħÙĨ Ø·ÙĤØ© +nÄħ Äĩ +Ġthi ên +иÑĩеÑģк ой +ĠاÙĦÙħ ÙĦ +Ġع Ùħ +ס פר +Ġnh óm +ÙĪØµ Ùģ +ĠCh úng +Ġر ÙĤÙħ +ãģ¾ãģĹãģŁ ãģĮ +al ité +ล ม +ĠëĤ´ ê°Ģ +׾ק ×ķ×Ĺ +ĠS Æ¡n +pos ição +mi ÄĻ +Ġtr ánh +ĠÄIJ á»Ļ +׼ ×Ĺ +ãģĤ ãģ£ãģ¦ +à¸Ńย à¹Īา +Ġ×ŀ×Ĺ ×Ļר +Ġ×Ķ ×Ļת×Ķ +à¸Ľ à¹Īา +à¸Ńืà¹Īà¸Ļ à¹Ĩ +Ø´ ÙĤ +×ł×¡ ×Ļ +ë¦ ¼ +ãģ¦ãģĹãģ¾ ãģĨ +Ġ×ŀ צ×ij +ãģ« åĩº +ÙħÙĪØ§ Ø·ÙĨ +ยัà¸ĩ มี +алÑĮ нÑĭе +san ız +Ø¥ سرائÙĬÙĦ +ĠvÃł i +ì¤ Ħ +ã썿ĢĿ ãģ£ãģ¦ +×Ļ ×ķ׳×Ļ +çĶŁ ãģį +Ġs âu +Ñĩ иÑģÑĤ +Ġl á»ħ +ĠGi á +à¸Ńุ à¸Ľ +à¸Ńà¸¸à¸Ľ à¸ģร +à¸Ńà¸¸à¸Ľà¸ģร à¸ĵà¹Į +Ġnh ẹ +r ö +ס ×ĺ×Ļ +ãģķãĤĵ ãģĮ +Ġd ầu +ع Ùİ +ت را +×Ĵ×ĵ ׾ +Ġtécn ica +׼ ׳×Ļ×Ŀ +תק ש +תקש ×ķרת +Ġн его +ét ait +Ġm á»ģm +Ñģ еÑĤ +Ġnh áºŃt +Ġ×ŀ ×¢×ľ +Ġ×Ķ×¢ ×ij×ķ×ĵ +Ġ×Ķ×¢×ij×ķ×ĵ ×Ķ +Ġ×Ĵ ×Ļ׾ +ãģ¯ ãģªãģĦ +ائ ØŃ +Ġз деÑģÑĮ +×IJ ×Ļ׳×ĺר +Ùħ ÙIJ +Ġ×Ļ ×Ĺ×ĵ +ر اÙģ +ì²ĺ 리 +×ĵ ×¢×ķת +ì¹ ľ +ĠТ о +ĠTh ế +ì¶ © +Ġ׳׼ ×ķף +عÙĬ Ø´ +ни з +Ġج اÙĨب +×ŀ×§ צ×ķ×¢ +à¹Ĥ à¸ĭ +Ñģ ÑĥÑĤ +ìĸ´ ìļĶ +ãĤĴè¦ĭ ãģ¦ +ار د +Ġaç ıl +ĠاÙĦØŃ ÙĬاة +à¸ģà¹ĩ à¹Ħà¸Ķà¹ī +ãģĿãĤĮ ãĤĴ +عض ÙĪ +Ġг ÑĢаж +ĠгÑĢаж дан +à¸Īะ à¸ķà¹īà¸Ńà¸ĩ +ĠìĿ´ 룬 +ĠìĿ´ë٬ íķľ +Ġtr ách +ÙĨ Ùİ +Ġkı sa +Ã Ķ +ÑĪ ÐºÐ° +ãģ® äºº +ĠÐŁ оÑģ +ĠÐŁÐ¾Ñģ ле +Ñĥ лÑĮ +ÙĪØ§ جÙĩ +ÙĤ رب +à¸Ľà¸ıิ à¸ļัà¸ķิ +ê° Ļ +Ġ×ŀ ׳ +ĠÑģво и +بر اÙħج +Ġر ÙĪ +пÑĢ Ð¾Ð´ +пÑĢод аж +Ġby ÅĤy +วั ย +Ġgör ün +ĠÃ Ī +ÑİÑī им +ĠÑĤак ой +Ùģ ÙĪØ± +ĠÙģ Ø¹ÙĦ +Ġб ел +ëIJ ł +er ÃŃa +ĠÑģво Ñİ +Ġl ã +Ġlã nh +à¹Ģà¸ŀืà¹Īà¸Ń à¹ĥหà¹ī +ÙĤ ÙĨ +تط ÙĪÙĬر +Ġsay ı +ĠÑģ ейÑĩаÑģ +Ġ×IJ×Ĺר ת +×§ ×ķפ×Ķ +×§×ķר ס +Ġس Ùħ +Ġ×ĺ ×Ļפ×ķ׾ +ìĿ´ëĿ¼ ëĬĶ +دراس Ø© +èµ· ãģĵ +×Ĺ ×Ļ׳ +×Ĺ×Ļ׳ ×ķ×ļ +×ĵ ×§ +Ġë§ ŀ +Ġком анд +ĠÐij о +Ġиг ÑĢÑĭ +à¸ļ ี +ĠØ£ Ùİ +в ен +ĠاÙĦج دÙĬد +ĠÙĦ Ø¥ +Ġ×ķ×IJ ׳×Ļ +Ġ×Ķס ×Ļ +иÑĩеÑģк ого +رÙĪ ØŃ +à¸ģาร ศึà¸ģษา +ĠTr ưá»Ŀng +иг ÑĢа +ıl ması +Ġм аÑģÑģ +ãģ¨ãģį ãģ« +à¸Ĺีà¹Ī à¸ľà¹Īาà¸Ļ +à¸Ĺีà¹Īà¸ľà¹Īาà¸Ļ มา +ĠاÙĦساب ÙĤ +Ġ×ŀ×¢ ×ĺ +в аÑĤÑĮ +m Ã¼ÅŁ +Ġ׾ ׼×ļ +Ġt á»ĭch +Ùģ ÙĩÙħ +تد رÙĬب +Ø´ Ùĥ +Ġ×ij ×ŀ×Ļ +Ġ×ij×ŀ×Ļ ×ķ×Ĺ×ĵ +ÙĤØ· اع +ãģª ãģĹ +×ķצ ×Ļ×IJ +ĠÙĪ Ø³ÙĬ +з Ñĥ +Ġy at +Ġyat ırım +ë§ İ +Ġth ắng +ãģĬ 客 +ãģĬ客 æ§ĺ +ĠThi ên +ãģ«å¯¾ ãģĹãģ¦ +ÑĢ Ð¸Ñģ +ÙĨت ائ +ÙĨتائ ج +Ġ×ŀ שר +Ġ×ŀשר ×ĵ +Ġتع اÙĦ +ĠتعاÙĦ Ùī +ש ׳×Ļ +Ùĩ اÙħ +×IJ׳ ש×Ļ×Ŀ +Ġżyc ia +ĠÑĢÑĥб лей +ÙĬ ض +Ġkat ıl +ĠÙħ ÙĪØ¶ÙĪØ¹ +Ġvard ır +ĠÙħÙĨ Ø·ÙĤØ© +ĠTr ần +Ġв еÑģ +ü p +Ùħ ÙĪÙĨ +ÑĪ Ð»Ð¸ +Ġn óng +Ø® ÙĦÙģ +ĠС ÑĤа +Ġд оÑĢ +ĠдоÑĢ Ð¾Ð³ +ĠwÅĤa ÅĽnie +eÄŁ in +Ġhi á»ĥm +ĠС ам +ê»ĺ ìĦľ +ĠÑĦ а +ãģ» ãģĨ +ãģ»ãģĨ ãģĮ +×ķפ ×Ļ×¢ +ê° Ī +د ÙĪÙĦ +Ġthu ê +Ġch á»Ĺ +Ġëĭ¹ ìĭł +ãģij ãĤĮ +ãģijãĤĮ ãģ© +ë³´ íĺ¸ +ãģķãĤĮ ãģ¦ãģĦãģ¾ãģĻ +Ġнад о +ĠìĤ¬ëŀĮ ëĵ¤ +à¹Ģà¸Ĥ à¸ķ +สม ัย +z ÅĤ +ت ÙĪØ± +Ġש ת×Ļ +v ê +Ġ×ijת ×ķ×ļ +à¸Ĭ ัย +ãģĦ ãģ£ãģŁ +ìĿ ij +Ġt ầ +Ġtầ ng +ש ׼ר +Ġê¸ Ģ +Ġ×Ķש ׳×Ķ +Ġا ÙĨÙĩ +ç«ĭ ãģ¡ +r és +füh ren +ر ØŃÙħ +ê· ¹ +ĠâĢ « +Ġsu ất +à¸Ł ิ +ÙĬ Ùĩا +ĠاÙĦ اتØŃاد +Ġt uyá»ĥn +ãģ¾ ãĤĭ +Ġm ại +Ġng ân +ãĤ° ãĥ© +欲 ãģĹãģĦ +س ار +ãĤĤãģ® ãģ§ãģĻ +ки е +Ġseç im +åħ¥ ãĤĬ +ãģªãģ© ãĤĴ +ÑĤ ÑĢи +ĠÑģп еÑĨ +ĠØ£ د +Ġод но +ÑĪ ÐµÐ» +ãĥĩ ãĥ¼ãĤ¿ +ãĤ· ãĤ¹ãĥĨ +ãĤ·ãĤ¹ãĥĨ ãĥł +è¡Į ãģį +ã썿ĢĿ ãģ£ãģŁ +à¹Ģà¸ģิà¸Ķ à¸Ĥึà¹īà¸Ļ +ĠÑĤ ож +ĠÑĤож е +Ġs ạch +ĠÑģ ÑĢок +Ġкли енÑĤ +ĠÙħØ´ رÙĪØ¹ +Ġalt ında +Ġì ·¨ +ä¸Ń ãģ® +ãģķãģĽ ãĤĭ +ãģĻ ãģ¹ +ãģĻãģ¹ ãģ¦ +ê°ľ ë°ľ +ĠÄij êm +ãģªãģĦ ãģ®ãģ§ +ì² ł +×¢ ×ij×ĵ +Ġd ấu +à¸Ħà¸Ļ à¸Ĺีà¹Ī +ĠC ách +تع ÙĦÙĬÙħ +Ġh ại +ãĤ» ãĥķãĥ¬ +ĠÙĨÙ쨳 Ùĩ +ĠíĨµ íķ´ +ÑĪ Ð»Ð¾ +Ġнап ÑĢав +ĠнапÑĢав лен +ÑĢÑĥ Ñĩ +íĶ Į +Ġ×ijר ×Ļ×IJ +ãģ® ãģ¿ +ãģ«ãģĬ ãģĦãģ¦ +×ij ׳ק +ãĤ¨ ãĥ³ +Ø«ÙĦ اث +Ġm ỹ +ĠÑģай ÑĤе +Ġе мÑĥ +ت غÙĬ +تغÙĬ ÙĬر +خص ÙĪØµ +ÑĤе ли +Ġ×ķ׾ ׼ף +פע ×Ŀ +Ġпо ÑįÑĤомÑĥ +ر اÙĨ +иÑĤел ей +пиÑģ ан +×¢ ×¥ +ĠìĤ¬ ìĹħ +Ùħ ز +جÙħ ÙĬع +ë©´ ìĦľ +à¸ľà¸¥à¸´à¸ķ à¸łà¸± +à¸ľà¸¥à¸´à¸ķà¸łà¸± à¸ĵ +à¸ľà¸¥à¸´à¸ķà¸łà¸±à¸ĵ à¸ij +à¸ľà¸¥à¸´à¸ķà¸łà¸±à¸ĵà¸ij à¹Į +ĠпÑĢ Ð¸Ð¼ÐµÑĢ +ãĤŃ ãĥ¼ +l â +Ġch Äĥm +缮 ãģ® +ãģĦ ãģĭ +ãģ¨è¨Ģ ãģĨ +×ĸ ×ķ×Ĵ +Ġ×ij ×ĵ×Ļ +Ġ×ij×ĵ×Ļ ×ķ×§ +ãģĬ åºĹ +à¸ķà¸Ńà¸Ļ à¸Ļีà¹ī +Ġph á»iji +п ÑĤ +สà¸Ļ าม +Ø· ÙĪ +ص اØŃ +صاØŃ ب +ĠD ü +ĠDü nya +Ġп ока +п ал +ĠÄij ảo +ĠاÙĦÙģ ÙĪØ± +ĠاÙĦÙģÙĪØ± Ùĥس +Ġmá u +кÑĢ ÐµÐ¿ +ĠاÙĦس اعة +ĠгоÑĢ Ð¾Ð´Ð° +Ùģ ØµÙĦ +ай ÑĤе +Ġд ог +Ġдог овоÑĢ +ĠØ¥ ذ +Ġ×ij׼׾ ׾ +ÙĬ تÙĩ +×Ĵ ×ijר +Ġbir ç +Ġbirç ok +문 íĻĶ +ãģĿãģĨ ãģª +را ØŃ +ĠÙħ رة +ĠденÑĮ ги +f ä +à¸Ĥà¹īา ว +ĠÑģов ÑĢем +ĠÑģовÑĢем енн +׾×Ĺ ×¥ +èī¯ ãģı +ĠÙģ Ø£ +Ġ×ķ ×ĸ×Ķ +Ġз ани +Ġзани ма +Ġê°Ģì§Ģ ê³ł +Ġh Æ¡i +ãģªãģ® ãģĭ +ãĥĨ ãĥ¬ãĥĵ +Ġר ×ij×ķת +à¸ķ ี +Ġ×ijש ×ł×ª +ĠT ại +Ġthu áºŃn +Ñģ ел +Ñij м +dzi Äĩ +ĠÑģ ка +ĠÑģка Ñĩ +ĠÑģкаÑĩ аÑĤÑĮ +×ķ×ŀ ×ķ +г ла +Ġмин ÑĥÑĤ +åĩº ãģĻ +Ġ×Ĺ×Ļ ×Ļ×ij +Ġת ×Ĵ×ķ×ij×Ķ +à¸£à¸¹à¸Ľ à¹ģà¸ļà¸ļ +ни ÑĨа +Ġİ n +ĠØ£ ع +Ġض ÙħÙĨ +Ùħ ثاÙĦ +ĠyaÅŁ an +ĠìŰ 구 +ĠL ê +ש׾ ×Ĺ +ãģı ãģªãĤĭ +ìĹĨ ìĿ´ +ĠÑĤ ÑĢи +ĠÑĩаÑģÑĤ о +Ġоб ÑĢаÑĤ +п ло +د Ø® +دخ ÙĪÙĦ +س Ùĩ +à¸Ń าà¸ģ +à¸Ńาà¸ģ าศ +Ġ׼ ×ĸ×Ķ +Ġ×Ķ×¢ סק +ĠاÙĦØ£ ÙĨ +å¹´ ãģ« +×¢ ש×ķ +Ġש ×¢×ķת +Ġm Ãłn +×IJר ×Ļ +sı yla +Ù쨱 ÙĤ +ни Ñħ +Ġت ست +è¦ĭ ãģ¦ +ØŃا ÙĪÙĦ +×IJ ×Ļ׼×ķת +ĠbaÅŁ ladı +st Äħ +stÄħ pi +à¸Ĺีà¹Ī à¹Ģรา +ÙĤر ر +ج اب +Ġ×ijר ×ķר +à¹Ģà¸Ĥà¹īา à¹ĥà¸Ī +×ŀ׊קר +al ım +Ġס ×Ļפ×ķר +ãģ§ãģĤ ãĤĮãģ° +Ġש×ŀ ×ķר×ķת +Ġ×ķ ×ŀ×Ķ +ãģĵ ãģĿ +id ée +ä¸ĭ ãģķãģĦ +تÙĨا ÙĪÙĦ +Ġ ลà¹īาà¸Ļ +Ġìļ°ë¦¬ ëĬĶ +اÙĨ ا +ÑģÑĤ ой +б оÑĤ +ĠyaÅŁ am +kö y +Ø¥ ÙĦ +ÑĢ Ñĭв +기 ìĹħ +Ġ×Ķ×ŀ ×ĵ +Ġ×Ķ×ŀ×ĵ ×Ļ׳×Ķ +د ب +×¢ ×Ļ׳×Ļ +×ŀ ת×Ĺ +Ġפ ר×Ļ +ãĥĭ ãĥ¼ +اÙħ ÙĬ +Ġnh ằm +ãĤĮ ãģªãģĦ +ت عرÙģ +Ġë§Ī ìĿĮ +ìĵ ° +Ġh ấp +ר×Ĵ ×Ļ׾ +ب Ùİ +Ġr Äĥng +gl Äħd +ĠÑģиÑģÑĤем Ñĭ +Ġkh óa +ãģ§ãģĻ ãĤĪãģŃ +大ãģį ãģı +기 를 +Ġké o +ÙĪ Ø¡ +ج اÙħ +جاÙħ ع +Ġ×¢ ×Ļצ×ķ×ij +t éri +Ġת ש +Ġ×IJ ×ij×Ļ +ĠCh ương +à¸ļริ à¹Ģว +à¸ļริà¹Ģว à¸ĵ +ãģ¤ ãģı +Ġ×Ĺ ×ķ׾ +עת ×Ļ×ĵ +ש ×Ļ×ŀ×Ķ +ëĤ ¨ +Ġש×IJ ×Ļף +ĠÙĪØ§ÙĦ Ø¥ +ÑĦ а +Ġkh ám +Ġ×ĺ ×ķ×ij×Ķ +ĠвÑĭ Ñģ +ĠвÑĭÑģ око +ĠاÙĦØŃ دÙĬØ« +人 ãĤĤ +d Ã¼ÄŁÃ¼ +×Ļ×Ĺ ×ķ×ĵ +تع ÙĦÙĬ +تعÙĦÙĬ ÙĤ +l ö +تØŃ دÙĬد +н его +ĠÑĥд об +Ġ׾ ×ŀ×Ļ +Ġר ×ķצ×Ļ×Ŀ +Ġج اء +Ġ×ij ×ĸ×ŀף +à¸Ľà¸ģ à¸ķิ +é«ĺ ãģı +à¸Ľà¸¥ า +Ġart ık +Ġbug ün +×§ ׳×Ļ +Ġkho á +ĠÙħ رÙĥز +ĠìŀIJ 기 +در جة +×ŀש ר×ĵ +Ġgi ấy +Ġch óng +×§ פ +ÙĬب Ø© +ĠczÄĻ sto +в али +Ùĥ ب +ìŁ ģ +ส à¸ļาย +à¸Ľà¸£à¸°à¸Ĭา à¸Ĭà¸Ļ +×Ĵ ×ķ×£ +ëŁ ī +ãģ® ãģĵãģ¨ +ล à¸Ń +Ġngh á»ī +åŃIJ ãģ© +åŃIJãģ© ãĤĤ +à¹Ħà¸Ķ à¹īà¸Ńย +à¹Ħà¸Ķà¹īà¸Ńย à¹Īาà¸ĩ +×ĵ ×¢ +ĠاÙĦت Ùī +ĠÑģов еÑĤ +Ġqual itÃł +åĩº ãģĹ +ĠÑĢÑĥк ов +ĠÑĢÑĥков од +ราย ละà¹Ģà¸Ńียà¸Ķ +ãģªãģĭ ãģªãģĭ +기 ê´Ģ +Ġ×Ĺ ×ķש +Ġ×Ĺ×ķש ×ij +л оÑĤ +à¸Ļะ à¸Ħรัà¸ļ +×§×ij ×ķצ×Ķ +Ġth ái +Ġש ×ij×Ķ +ĠÑĪ ÐºÐ¾Ð» +ĠÙĦ ÙĥÙĦ +à¹ĥà¸Ļ à¸Ĭà¹Īวà¸ĩ +ĠÙħ ÙĥاÙĨ +ë ķĮ +Ġc ải +ĠCh ÃŃ +ÑĥÑĩ а +ìĿ µ +Ġx ảy +à¸Ĭà¸Ļ ิà¸Ķ +Ġc áºŃu +к ÑĢов +ss é +ĠÙĨ ÙĪØ¹ +ĠТ а +Ø® Ùħس +פ×ķס ×ĺ +Ġm ắc +ĠÄij em +à¸ģาร à¹ĥà¸Ĭà¹ī +ר ×ķס +ĠÐĽ е +Ġth á»Ń +รà¹Īาà¸ĩ à¸ģาย +üz ü +æĹ¥æľ¬ ãģ® +ê³¼ ìłķ +ש ×Ļ×IJ +ĠìŀĪ ê³ł +×ij ×ķ׾ +ìķ ħ +ĠÙĪØ§ÙĦ ا +ĠÐĽ и +ĠвÑģ Ñij +Ġużytk ow +×Ĺ ×ķ׾ +ر Ù쨶 +Ġson uç +ãģĦ ãģ¾ãģĽãĤĵ +ìĤ¬ ìĹħ +ëĪ Ħ +ÑĤ ек +Ġud ziaÅĤ +л ез +Ġ×Ķ×Ļ ×Ļת×Ļ +ãĤīãĤĮ ãģ¦ +Ùħس ؤÙĪÙĦ +ر ار +ÑĤ ан +ĠÄij Ãło +Ġר ×ķ×ij +Ġ×ijש×ij ×Ļ׾ +ä»ĬåĽŀ ãģ¯ +ãĤ¸ ãĥ¥ +Ġ×¢ ×ijר +ãģĽ ãģ¦ +п олÑĮ +ak lı +Ġk ÃŃnh +د ت +лож ение +ĠاÙĦÙħ ص +ĠاÙĦÙħص رÙĬ +à¸Īริà¸ĩ à¹Ĩ +ĠاÙĦشر ÙĥØ© +ĠÄij á»ı +ãĥĽ ãĥĨ +ãĥĽãĥĨ ãĥ« +Ñį кон +Ñįкон ом +ĠÙĪ Ø¹ÙĨ +Ġת ׳ +Ġ×ª×ł ×IJ×Ļ +ĠاÙĦدÙĪÙĦ ÙĬØ© +Ġì§Ģ ìĹŃ +ãģ§ãģĻ ãģĭ +Ġв аÑĢи +ĠваÑĢи анÑĤ +ĠاÙĦع رب +ел а +Ġt Æ°á»Ľng +sk Äħ +Ġm ặc +ส ัà¸ģ +ãĥĵ ãĥ¼ +Ġ×ij ×Ĵ׾ +Ġ×ij×Ĵ׾ ׾ +ãĥķãĤ¡ ãĥ³ +×ij ×Ļצ +×ij×Ļצ ×ķ×¢ +ли ÑģÑĤ +à¸Ł ุ +à¸Łà¸¸ à¸ķ +à¸Łà¸¸à¸ķ à¸ļà¸Ńล +à¸Ŀ à¹Īาย +ìŀIJ ìĿĺ +Ġس ÙĪÙģ +Ġש ×Ķת +Ġê± ¸ +×¢ ×ij×ķ×ĵ +ãģĻãĤĭ ãģĵãģ¨ãģĮ +ĠÑĩа ÑģÑĤÑĮ +ãĤ¢ ãĥ¡ãĥª +ãĤ¢ãĥ¡ãĥª ãĤ« +Ġtak ım +Ġs Ỽ +ĠsỼ m +שר ×Ķ +è¨Ģ ãģĨ +л ан +ì» ¤ +׼ ׳×Ķ +ÙĪÙģ ÙĬ +íĹ Ī +lu ÄŁu +ĠëĮĢ íķ´ +Ġ׾×ij ×Ļת +Ġ×Ķר×IJש ×ķ׳×Ķ +ص Ùħ +Ġsö yled +Ġsöyled i +à¸Ľ าà¸ģ +Ġard ından +ãģĪ ãģŁ +à¸Ĺัà¹Īว à¹Ħà¸Ľ +Ġ׳×ķס ×£ +б олÑĮ +ãĤĵãģ§ãģĻ ãģijãģ© +ĠлиÑĪ ÑĮ +Ġ×ij ×IJ×Ļ +ĠбÑĭ ÑģÑĤÑĢо +ส ัà¸Ļ +Ġ×ij פ׳×Ļ +л еÑĩ +ĠاÙĦØ® بر +Ġsó c +Ġth ú +Ġп ÑıÑĤ +ãģĬ é¡ĺ +ãģĬé¡ĺ ãģĦ +ÑĤ ин +ãģ«ãģ¤ãģĦãģ¦ ãģ¯ +פ ף +Ġдв ÑĥÑħ +à¸į ีà¹Ī +à¸įีà¹Ī à¸Ľ +à¸įีà¹Īà¸Ľ ุ +à¸įีà¹Īà¸Ľà¸¸ à¹Īà¸Ļ +оп еÑĢ +ĠاÙĦب شر +ĠاÙĦÙħ اÙĦ +ıyor uz +تØŃ ÙħÙĬÙĦ +à¸ģ ะ +éĸĵ ãģ« +×Ĺ ×ķש +ĠNg uyên +ãģĦãģ¦ ãģĦãĤĭ +дÑĥ ÑĪ +ש פע +ÑĪ Ñĥ +å®Ł éļĽãģ« +ĠÑĢай он +ĠCh á»ī +ÙĨ صر +Ġìļ ´ +Ġìļ´ ìĺģ +Ġ×Ķ×ĵ ×Ļף +ØŃد د +ر ز +ĠاÙĦد Ùħ +ĠPh áp +ÑĤ ÑģÑı +è¦ĭ ãģĪ +Ġti á»ĥu +Ġs á»Ńa +а ÑİÑĤÑģÑı +ĠB á +Ġ×ķ ׼׾ +Ð ĸ +ÑĪ Ð¸Ð¼ +ìĿ´ ëĬĶ +л ев +d ık +Ġprés ente +Ġara ç +صد ÙĤ +Ġпом ог +ĠاÙĦشر ÙĤ +ĠÙĪØ§ÙĦ ذÙĬ +رÙĬ ا +×ij ׳×ķת +Ġng á»ĵi +ר ×ķפ +ר×ķפ ×IJ +Ġth ấp +ãĤĦ ãģ¯ +ãĤĦãģ¯ ãĤĬ +ĠاÙĦج دÙĬدة +éĿŀ常 ãģ« +ÙĬÙĦ ÙĬ +ìª ½ +تع اÙħÙĦ +ãģł ã썿ĢĿãģĦãģ¾ãģĻ +Ùħ Ùħ +иÑĤе ли +ãĤµãĤ¤ ãĤº +اد ات +ĠاÙĦÙħ اÙĦÙĬØ© +Ùĥات ب +к ли +веÑĢ Ñħ +ни Ñĩ +Ġ×ľ×¢ ×ij×ķ×ĵ +׾ ×Ļ×Ķ +ØŃ Ùİ +ãĤ¤ ãĥĻ +ãĤ¤ãĥĻ ãĥ³ãĥĪ +Ġת ×Ĵ×ķ×ij×ķת +ÑĦ он +ĠдÑĢÑĥг ие +×IJ ×ĸ×ķר +Ġper ò +ìķ ŀ +åĢŁ ãĤĬ +ר צ×Ļ +×IJ ×ĸ +алÑĮ нÑĭÑħ +Ġê²ĥ ìľ¼ë¡ľ +ĠпÑĢав о +ĠاÙĦØ£ رض +à¹Ģà¸Ĺ à¸Ħ +à¹Ģà¸Ĺà¸Ħ à¹Ĥà¸Ļ +à¹Ģà¸Ĺà¸Ħà¹Ĥà¸Ļ à¹Ĥล +à¹Ģà¸Ĺà¸Ħà¹Ĥà¸Ļà¹Ĥล ย +à¹Ģà¸Ĺà¸Ħà¹Ĥà¸Ļà¹Ĥลย ี +צ ר×Ļ +ĠÐļ Ñĥ +ıl ma +決 ãĤģ +ا ÙĪ +Ġ×ĵ ×§×ķת +à¸Ħร ู +ĠÙħست ÙĪÙī +à¸Ľ à¹īà¸Ńà¸ĩ +à¸Ľà¹īà¸Ńà¸ĩ à¸ģัà¸Ļ +×ĵ ×ķ×ŀ×Ķ +ĠÑģ егоднÑı +س ÙĪÙĤ +ר×Ĺ ×ķ×ij +ĠØ¥ دارة +Ñħ ож +éģİ ãģİ +à¸Ħ à¸Ń +нÑĥ л +×ķ׼ ×Ķ +ÙĪ Ø§ÙģÙĤ +׼׾ ׾ +Ġ×Ķ ×ĵ×ķ +Ġl Ä©nh +Ġkh ảo +×IJ×ŀ צע +ë¨ ¸ +Ġ׼ ×Ļצ +Ġ׼×Ļצ ×ĵ +Ġдолж нÑĭ +หว ัà¸ĩ +ãĥĩ ãĤ¶ +ãĥĩãĤ¶ ãĤ¤ãĥ³ +Ġng á»Ŀ +ä¸Ń ãģ« +à¸ģลัà¸ļ มา +جÙħ اÙĦ +à¸Ķัà¸ĩ à¸ģลà¹Īาว +س ÙĥÙĨ +س ÙĨ +Ġözellik le +з еÑĢ +rz ÄĻ +×ŀ ×ķר×Ķ +Ġl ạ +×ŀ ×Ļ׳×Ļ +ר ×Ļת +ãģĿãĤĮ ãģĮ +ãģĭ ãĤĮ +ĠÙĬÙħÙĥÙĨ Ùĥ +öff entlich +г ан +ĠاÙĦØŃ ÙĦ +ĠmiÄĻd zy +ĠÑĩа ÑģÑĤи +ujÄħ cy +ĠbaÄŁ lı +ĠiliÅŁ ki +Ùģ Ø§Ø¡ +ãĥª ãĥ³ãĤ° +Ġhã ng +ĠконÑĤ ÑĢ +ĠконÑĤÑĢ Ð¾Ð» +к оп +ש ×Ļ×¢ +ש×Ļ×¢ ×ķר +ĠÐĴ аÑĪ +Ġ×Ķ ×ª×§ +ÙħÙĨ ع +ĠpolÃŃt ico +Ġг олов +ĠØ¥ ÙĬ +Ø¥ ÙĨتاج +à¸ļ ิ +Ġг овоÑĢ +ĠговоÑĢ Ð¸ÑĤ +Ġph á»ķ +ĠÑģем ÑĮ +ãģ¯ ãģĤãĤĬãģ¾ãģĽãĤĵ +ĠÙĪ Ø§Ø³Øª +×ŀש פ×ĺ +з ем +×ŀ×ĵ ×ijר +Ġíģ ° +ĠìĿ´ ë²Ī +ê°Ģ ëĬĶ +Ġì§Ģ ìĽIJ +Ġca ÅĤy +Ġgeli ÅŁtir +Ñģк ое +pos é +Ġkh ô +à¸ķิà¸Ķ à¸ķาม +miss ão +Ġ׾ ×ŀר +Ġ׾×ŀר ×ķת +Ġb ó +à¸ķรวà¸Ī สà¸Ńà¸ļ +Ġngh á»ģ +Ġб из +Ġбиз неÑģ +ÑģÑĤ еÑĢ +ÙĪ Ùİ +楽 ãģĹãģ +楽ãģĹãģ ¿ +ãģĵãĤĮ ãģĭãĤī +wiÄħ zan +ส à¸Ńà¸Ļ +Ùħ ÙĪØ± +׳×ĵ ׾ +Ġ×Ķ×IJ ×ĵ×Ŀ +Ġм олод +ØŃ Ùħا +ØŃÙħا ÙĬØ© +ÑģÑĤ ÑĢан +Ġbu á»ķi +ת×Ļ ×Ļ×Ŀ +abile ceÄŁi +L İ +à¹Ģย à¸Ńะ +à¸Ī ร +س ÙĥاÙĨ +à¸Ļ ัà¸Ķ +Ġm ấy +ĠÐij а +s ÅĤaw +ĠÙģ ÙĦا +ĠкоÑĤоÑĢ Ð¾Ð¹ +Ġпло Ñī +ĠплоÑī ад +ãĤĤ ãģĤãĤĬ +sz czÄĻ +×Ļפ ×ķ +ש×ŀ ת +owa ÅĤa +Ġn ông +צ×ij ×IJ +ĠìŀĪ ìĹĪ +ãģ¾ ãģ¨ +ãģ¾ãģ¨ ãĤģ +ÙĤÙĪ Ø§Øª +ãģ¿ ãĤĵãģª +Ġ׼ ×ŀ×¢×ĺ +Ġx úc +ï¼ Ĩ +r ÄĻ +rÄĻ cz +×ĵ ×ŀ×Ļ +Ġt áºŃn +à¸Ķ วà¸ĩ +ê²½ ìłľ +п ÑĥÑĤ +Ø£ ربع +Ġ×ŀ שת×ŀש +ãĤ¿ãĤ¤ ãĥĹ +Ġìłľ ê°Ģ +Ġ׾ ׼ף +ĠобÑĢаз ом +ÙĬÙĥ ا +w ÅĤ +wÅĤ asn +ĠاÙĦÙĪØ·ÙĨ ÙĬØ© +بÙĬ ب +×ŀ ׾×Ļ +к ÑĢаÑĤ +기 ìĹIJ +ÙĤ اد +ĠÙĦ دÙī +à¸Ħวาม รูà¹ī +×ŀ×ĵ×Ļ׳ ×Ļ×ķת +ê² ¨ +Ġíĺ Ħìŀ¬ +ש ת×Ļ +м ол +Ġmá i +à¸ŀิ ม +à¸ŀิม à¸ŀ +à¸ŀิมà¸ŀ à¹Į +หล วà¸ĩ +Ġx uyên +×Ĺ ×¡×¨ +رÙĪ ÙĨ +ãģĿãģĨ ãģĦãģĨ +ãģĿãĤĮ ãģŀ +ãģĿãĤĮãģŀ ãĤĮ +Ġ׼ ש×Ķ +ÐŁ ÑĢав +×ŀ×ij צע +ع رب +Ġbü yü +פ×Ļת ×ķ×Ĺ +à¸Ī à¸ļ +ĠØ£ Ùĥبر +שר ת +×ŀ׼ ש×Ļר +ĠÙĪ Ùħع +ãģ® ãģŁãĤģãģ« +à¸Ļ ัà¸ļ +ì° ° +ãĥª ãĥķãĤ© +ãĥªãĥķãĤ© ãĥ¼ãĥł +Ġc ưá»Ŀng +ĠìłĢ íĿ¬ +ÙħÙĨظ ÙħØ© +Ġhiç bir +ãģ§ãģ¯ ãģĤãĤĬãģ¾ãģĽãĤĵ +ร à¸Ńย +ëIJľ ëĭ¤ +ãģĻãģIJ ãģ« +к ла +Ġürün ler +Ġki á»ĥu +ĠëĤĺ ëĬĶ +ÑĤ ки +Ñģ им +Ġchá»ī nh +ãĤĤ ãģªãģĦ +ศ รี +æĽ¿ ãģĪ +ta ÅŁ +Ġب ÙĥÙĦ +Ġ×ķ ×Ļש +vis ão +ä¼ Ŀ +ä¼Ŀ ãģĪ +ÙĦ د +׾ ×Ļ×ŀ +׾×Ļ×ŀ ×ķ×ĵ +t ória +د Ùij +اÙħ ر +Ġê·¸ëłĩ ê²Į +Ġmateria ÅĤ +à¸Ĺ รา +à¸Ĺรา à¸ļ +ã쮿ĸ¹ ãģĮ +ãģ¦ ãģįãģŁ +ض غ +ضغ Ø· +ĠÙĬ عÙĨÙĬ +ел о +×IJ×Ķ ×ij×Ķ +×¢ ×ŀ +ÅŁ ık +ìŀIJ ëĬĶ +ãĤ¿ ãĥ³ +Ġb áºŃt +×ŀשפ ×Ĺ×Ķ +к ÑĢи +б ли +สั à¸ķ +สัà¸ķ วà¹Į +ĠسÙĨ ÙĪØ§Øª +ĠPh ương +ãģ¦ãģĹãģ¾ ãģ£ãģŁ +ãģª ãģľ +Ġ×ij×IJ ×ķ +Ġc án +س جÙĦ +Ġl ẽ +ãĤ± ãĥ¼ãĤ¹ +Ġ×§ ×Ļ×ij׾ +à¸ļà¸Ĺ à¸Ħวาม +Ġ×ķ ׼ף +ĠпÑĢедÑģÑĤав лен +Ġn á»iji +Ġcoment ário +ени ем +Ġtá» ı +l Ãł +Ġש×Ķ ×Ļ×Ķ +Ñģл ав +ĠاÙĦ ÙĪÙĦا +ĠاÙĦÙĪÙĦا ÙĬات +ÙĦج ÙĨØ© +×§×ķר ×IJ +бÑĭ ÑĤ +Ġì ¦ +Ġì¦ ī +ãģ§ãģĻ ãģĹ +หรืà¸Ń à¹Ħมà¹Ī +за ÑīиÑĤ +ÙģÙĦ سطÙĬÙĨ +Ġmi á»ħn +à¹Ģย à¹ĩà¸Ļ +ĠçalÄ±ÅŁ an +×Ļ×Ĵ ×Ķ +ĠE ÄŁ +ĠEÄŁ itim +ãĥĥãĤ· ãĥ¥ +Ġоп Ñĭ +ĠопÑĭ ÑĤ +ر غ +رغ ب +ĠÑģво иÑħ +à¸Ľà¸£à¸° à¸ķ +à¸Ľà¸£à¸°à¸ķ ู +Ġ×ŀ×IJ ×ĵ +׼ ×ķ׳×Ļ×Ŀ +à¸Ļ ี +ĠвÑĭ Ñħод +ãģ®ä¸Ń ãģ« +פ ׾×IJ +ĠÙĪ ÙĦÙĬس +פ×ķר ס +פ×ķרס ×Ŀ +Ùħ سÙĦÙħ +Ġng ôi +×ĵ ×ŀ×ķת +ãĤĴ使 ãģ£ãģ¦ +ĠпомоÑī ÑĮÑİ +Ø£ سر +бл ок +ÙĤ Ùĩ +ãģĹãģ¾ ãģĦ +ãģ¨ ãģĹãģŁ +Ġп еÑģ +ãĥī ãĥ« +×Ĺ ×Ŀ +ãģĹãģª ãģĮãĤī +ĠÐŁ ÑĢед +ãĥģãĤ§ ãĥĥãĤ¯ +å¼· ãģĦ +ש ×Ļר×ķת +д аеÑĤ +×Ļ×ij ×ķ +Ġgen ç +ил аÑģ +илаÑģ ÑĮ +ĠبÙĦ د +æĤ ª +æĤª ãģĦ +Ġ×ŀ שת +æ§ĺ ãĢħ +æ§ĺãĢħ ãģª +à¸ĺรรม à¸Ĭาà¸ķิ +ĠÙĥ اÙħÙĦ +ĠاÙĦس Ùħ +×ij×ĺ ×Ļ×Ĺ +c á +g ência +ãĤ¹ãĤ¿ ãĥ¼ +à¸Ĺำ à¸ģาร +×Ļ׾ ת +Ġ×Ļ ×ķצ×IJ +w ój +à¸ļุ à¸Ħ +à¸ļุà¸Ħ à¸Ħล +ع تÙħ +عتÙħ د +ãģĿãĤĮ ãģ« +ĠاÙĦت ارÙĬØ® +ÙĤر اء +Ġyönet im +×§ שר +ĠÑģп оÑĢÑĤ +Ġר×IJש ×ķף +Ġseñ al +Ġch ắn +çĦ¡ ãģĦ +ĠдоÑģÑĤ аÑĤ +ĠдоÑģÑĤаÑĤ оÑĩно +Ġá gua +à¸ģร à¸ĵ +à¸ģรà¸ĵ ี +Ġ×ŀש ×ķ +Ġtr ải +ë² Į +ujÄħ cych +Ù쨱 د +à¹ĥ à¸ģล +à¹ĥà¸ģล à¹ī +ãĤĭ ãģ®ãģ¯ +ר×ķ ×ķ×Ĺ +ÙĨ Ùĥ +ĠاÙĦÙĨ ÙĤ +ãģ®ãģ§ ãģĹãĤĩãģĨ +ãģ®ãģ§ãģĹãĤĩãģĨ ãģĭ +Ùħ عرÙģ +ÙħعرÙģ Ø© +ÑĥÑī е +Ġ×ij×¢ ×Ļקר +ت صÙĦ +Ġ×Ķ×IJ ר +Ġ×Ķ×IJר ×¥ +ĠÅŀ i +à¸Ĥา à¸Ķ +íŀ ĺ +ãģªãĤĵ ãģ¨ +ĠìĤ¬ëŀ ij +l Ã¼ÄŁÃ¼ +ب اء +ĠاÙĦØ¢ خر +Ġfam ÃŃlia +ĠTh áng +Ñī ениÑı +ãĤ¯ ãĥŃ +ĠTh ứ +æĽ¸ ãģį +ен ной +ìŀ ¡ +бл аг +благ о +п ов +à¹ģ ว +à¸ĩ à¸Ħà¹Į +à¸Ńัà¸Ļ à¸Ķัà¸ļ +ãģĤ ãģĴ +ร à¹īาย +ün ün +Ġ×Ļ׼×ķ׾ ×Ķ +з он +ĠÐľ и +маÑĤ еÑĢиал +Ġë³´ ë©´ +ØŃÙģ Ø¸ +ê Ìģ +ãģ« ãģĻãĤĭ +Ġת ×IJ +Ġ×Ķס ×ķ +ĠÑģÑĤ оÑĢ +ĠÑģÑĤоÑĢ Ð¾Ð½ +ãĥĪ ãĥĥãĥĹ +ÅĤo ÅĽÄĩ +ëħ ¼ +ëĵ Ŀ +ĠÙĪØ§ÙĦ ع +ì¶ Ķ +Ġ×Ļצ ×IJ +ĠÑĢаз дел +алÑĮ наÑı +×IJ׳ ש×Ļ +spo ÅĤ +spoÅĤ ec +spoÅĤec zn +Ø¥ عÙĦ +إعÙĦ اÙĨ +ÙĤÙĪ Ùī +íķĺë©´ ìĦľ +تط ÙĪØ± +Ġsi êu +Ỽ t +д ви +дви ж +Ġqu ần +k ıl +ĠпÑĢи зна +ĠH ã +ĠHã y +ĠباÙĦ ت +man ın +ãĤ« ãĥ« +Ġk á»· +×§ ׾×Ļ +ëIJĺ ì§Ģ +تعÙĦ Ùħ +ìĭľ ìĦ¤ +ìĭ ¶ +íĺ ¼ +Ùĥ ÙĬÙģ +売 ãĤĬ +วิ à¸Ĭา +б ал +ĠØ£ ØŃ +Ġдолж ен +รา à¸ĩ +ราà¸ĩ วั +ราà¸ĩวั ล +Ùħ اء +ج ار +Å ļ +Ġ×ŀ×IJ ×ĸ +ר ×ŀ×Ķ +ãģĭãĤĤãģĹãĤĮ ãģªãģĦ +ét ude +czÄħ c +Ġg ór +×ł×¡ ×Ķ +Ùħ ÙĬد +ĠÐŁ еÑĢе +Ø£ خر +ãģĿãģ® å¾Į +à¹Ģà¸Ķียว à¸ģัà¸Ļ +×ŀ ×Ĵ×ķ +×ŀ×Ĵ×ķ ×ķף +д ов +mas ına +×¢ ׳×Ķ +ãĤ± ãĥĥãĥĪ +ס ×¢ +סע ×Ļ×£ +ĠT ư +Ġt óc +íĻľ ëıĻ +ĠÐŀ д +ĠÐŀд нако +Ġdol ayı +ؤ Ùĥد +ê³Ħ íļį +׾ ר +в еÑĩ +Ġkh ợi +Ġth á»§y +×ĵ ף +ร à¸ģ +à¸ļั à¸ķร +à¹Ģà¸ģ à¹Īา +ĠاÙĦØ« اÙĦ +ĠاÙĦثاÙĦ Ø« +Ġpod rá +ער ×Ļ +ÙĨج اØŃ +Ġkh ắc +ì¸ ¡ +İ M +ãĤ» ãĥĥãĥĪ +ż enia +Ġ׾×Ĺ ×ijר +er Ãł +ì ´Ī +Ġkü ç +Ġküç ük +ات ÙĩÙħ +à¸ĭ à¹Į +Ùħشار ÙĥØ© +ĠاÙĦ بط +Ġd ây +ен нÑĭм +à¸Ĺีà¹Ī à¹Ħมà¹Ī +ÙĤ Ùİ +Ġv ượt +Ġtr ì +Ġwp ÅĤyw +A Åŀ +з о +ĠاÙĦس ÙĬد +à¸Ĺะ à¹Ģล +ĠÑģодеÑĢж а +ع Ø·ÙĬ +ĠاÙĦع ÙĨ +èĢħ ãģĮ +à¹Ģ หà¸Ļ +à¹Ģหà¸Ļ ืà¸Ń +Ġb ÃŃ +Ġüzer inden +ĠV Å© +Ġnu ôi +ÙĨ Ùħ +алÑĮ ного +×¢ ×Ļף +ØŃ ضر +ĠоÑĤ дел +ëª ĩ +ìķ ¡ +ĠÙĦدÙĬ Ùĩ +ìĻ ľ +Ġse ktör +Ġвозмож но +ĠÐĶ Ð¶ +Ġh ô +äºĭ ãģĮ +иÑĢов ание +алÑĮ ной +Ġ미 êµŃ +ر ØŃÙĦ +ĠÑįк Ñģ +пÑĢав лÑı +Ġnh á»Ŀ +ĠÄij ẩ +ĠÄijẩ y +Ùģ Ùĥر +ĠÙĪØ£ ضاÙģ +ãĥIJ ãĤ¹ +ת×ķ׼ ׳×Ļת +ÑĤел ей +ĠØ¥ÙĦÙĬ Ùĩ +ãģ¨è¨Ģ ãģ£ãģ¦ +Ġдв е +Ġch ấp +ĠL ö +à¸Ħล ิ +à¸Ħลิ à¸Ľ +Ġس ÙĪØ± +ĠسÙĪØ± ÙĬا +×ŀ×Ĺ ×ķ +st ä +д об +Ġni á»ĩm +ãģ® å¤§ +פר×ķ ×Ļ×§ +פר×ķ×Ļ×§ ×ĺ +ĠCh âu +Ġ×ŀ×Ķ ×Ŀ +Ñģк им +ĠполÑĥÑĩ иÑĤÑĮ +ÙĬ ÙĪÙħ +Ø« ÙĪØ± +פ×ķ׾ ×Ļ×ĺ +פ×ķ׾×Ļ×ĺ ×Ļ +ĠмеÑģÑı ÑĨ +åħ¨ ãģ¦ +ĠاÙĦÙħ جÙĦس +ĠاÙĦت اÙĦÙĬ +Ġ׊ר +åIJij ãģij +׼ ×ŀ×Ķ +б ед +Ø£ عض +أعض اء +ÙĪÙĦ د +วà¹Īา à¸Īะ +Ġb ánh +à¸Ļิ ย +à¸Ļิย ม +à¸Ľà¸£à¸° à¸ģัà¸Ļ +ÑģÑĤав иÑĤÑĮ +à¸ŀ à¸Ļัà¸Ļ +ĠÑį ÑĦÑĦ +ĠÑįÑĦÑĦ екÑĤив +Ġав ÑĤоÑĢ +ĠÄIJ Äĥng +Ġth Æ°á»Łng +ãĤĴ æĦŁãģĺ +à¸ģัà¸ļ à¸ģาร +å¾Į ãģ« +Ġya ÄŁ +ست اÙĨ +Ġli á»ģn +ãģĦ ãģ¾ +i êu +à¹Ĥà¸Ķ à¸Ļ +ĠÙĦ ذÙĦÙĥ +à¹Ĥรà¸ĩ à¹Ģรียà¸Ļ +צ ×Ļ×Ĵ +ĠاÙĦÙħ عÙĦÙĪÙħات +ç§ģ ãģŁãģ¡ +à¸Ĺีà¹Ī à¸Ħุà¸ĵ +ãģ«ãģª ãģ£ãģ¦ãģĦãĤĭ +×ŀ×ĵ ×Ļ׳×Ķ +ס ׼×Ŀ +Ġв не +à¸ŀ à¸Ļัà¸ģà¸ĩาà¸Ļ +ÑĢ ÐµÐ¹ +à¹Ģà¸Īà¹īา หà¸Ļà¹īาà¸Ĺีà¹Ī +ĠHi á»ĩn +Ġméd ico +ĠتØŃ ÙĤÙĬÙĤ +ÑĮ ÑĤе +miÅŁ ti +ÙĤÙĬ ادة +ãĤı ãģĭãĤĬ +มา à¸Īาà¸ģ +ëħ Ģ +ãģ«éĸ¢ ãģĻãĤĭ +×IJר×Ĵ ×ķף +m ètre +Ġעצ ×ŀ×Ļ +ĠCh úa +รูà¹ī à¸Ī +รูà¹īà¸Ī ัà¸ģ +ì£ Ħ +ëĭ µ +à¹ģà¸Ĺ à¹ī +Ġgeç en +Ġlan ça +ĠاÙĦ بØŃØ« +×ĵ ×ŀ×ķ +ãģ¯ ãģĺ +ãģ¯ãģĺ ãĤģ +Ġdön Ã¼ÅŁ +è¿ij ãģı +à¹Ģส ม +à¹Ģสม à¸Ń +ëĿ ½ +Ġü ç +á» ŀ +ÑĪ Ð°Ñı +à¸Ĺ ร +ØŃ ÙĤÙĬÙĤØ© +à¸Ĥà¸Ńà¸ĩ à¸ģาร +Ġ무 ìĹĩ +Ġ×Ķ ×Ľ×¨ +ĠاÙĦص ÙĬÙĨ +ĠлÑİ Ð´Ð¸ +à¸ķ าย +ب ÙĪÙĦ +Ġvi êm +Ġthi á»ĩu +à¸ģ à¸Ķ +Ġ׾ ×ĵ×ijר +פ ׳×Ķ +×IJר ×ij×¢ +س Ùī +ĠاÙĦسÙĬ اس +ĠاÙĦسÙĬاس ÙĬØ© +yd ı +ÙĪØŃØ¯ Ø© +ĠдеÑıÑĤелÑĮ ноÑģÑĤи +Ġ×ķ×Ķ ×ŀ +п еÑĩ +пеÑĩ аÑĤ +иÑĢов аниÑı +ĠÑģ ог +ĠÑģог лаÑģ +Ġ׼ ×ĵ +Ġ׼×ĵ ×IJ×Ļ +ĠиÑģполÑĮзов аÑĤÑĮ +ס פ×ķר×ĺ +Ġil çe +exp érience +ĠTh á»Ŀi +İ K +à¹Ħà¸Ł à¸Łà¹īา +ëĵ¤ ìĹIJê²Į +à¸Ľà¸£à¸° à¹Ģà¸ł +à¸Ľà¸£à¸°à¹Ģà¸ł à¸Ĺ +Ġmü mk +Ġmümk ün +Ġ×IJ×ķת ׳×ķ +ìĦ± ìĿĦ +ĠìĿ´ ìľł +زÙĬ ارة +Ġolduk ça +r ób +ĠØ£ ÙĨا +Ġ×Ķ ×ij×Ļ +Ñģ ен +×¢ ×Ļקר +×Ļ×ĵ ×ķ×¢ +d zÄħ +Ùħ عÙĦÙĪÙħات +Ø´ اب +Ġpar ça +à¸Ļะ à¸Ħะ +ب اس +ĠÑĤоÑĢ Ð³ +ĠÑĤоÑĢг ов +Ġ×Ĺ ×ĵר +׼ ר×ĺ +׼ר×ĺ ×Ļס +ĠA yrıca +ÃªÌ £ +ìľ ¨ +ĠÑĤак ие +Ġ×ŀצ ×ķ×Ļ +ãĥ©ãĥ³ ãĤŃãĥ³ãĤ° +ש×Ļ×ķ ×ķ×§ +åīį ãģ® +ĠB ảo +Ñī Ñĥ +æĹ© ãģı +ĠPh òng +à¸ŀระ ราà¸Ĭ +פ ×Ĺ×ķת +Ġг л +Ġгл аз +à¸Ĺ à¹Īา +Ġd ạy +ÑĢ Ð¾ÑģÑĤ +à¹Ĥà¸Ķย à¹Ģà¸īà¸ŀาะ +Ġqu áºŃn +Ġ×Ĺ×ijר ×ķת +m ême +mÄ±ÅŁ tı +ĠاÙĦت داÙĪÙĦ +Ġn ạn +Ġ×Ķ ×ĵ×Ļ +ĠاÙĦØ· رÙĬÙĤ +×Ĵ ×ķת +Ġ×Ķ ×ĵר×ļ +ujÄħ ce +Ġch ữ +ãĤĤãģ® ãģ® +ë° Ľ +ãģķãĤĵ ãģ¯ +Ġyard ım +ĠاÙĦع Ùħ +Ġì§Ħ íĸī +Ġ×Ļ ×Ĺ +Ġ×Ļ×Ĺ ×¡×Ļ +ĠاÙĦÙħ دÙĬÙĨØ© +Ġc ú +à¸ģี ฬ +à¸ģีฬ า +Ġni ên +mis ión +׳×Ļס ×Ļ +׳×Ļס×Ļ ×ķף +Ġвоз ÑĢаÑģÑĤ +Ġ×¢×ķש ×Ķ +ĠÙħ دÙĬر +Ñı ÑģÑĮ +ØŃ جÙħ +íĻĺ ê²½ +ĠاÙĦØ£ خرÙī +u ÃŁer +ĠاÙĦعاÙĦÙħ ÙĬØ© +ĠNg á»įc +êµIJ íļĮ +ä¸Ĭ ãģ§ +×Ļ×Ķ ×ķ×ĵ +×Ļ×Ķ×ķ×ĵ ×Ļ×Ŀ +Ùħس اعدة +Ġжиз нÑĮ +ĠпоÑĤ омÑĥ +ĠاÙĦÙħ ÙħÙĦ +ĠاÙĦÙħÙħÙĦ ÙĥØ© +ĠG ör +ر ÙIJ +×ŀ×§ ×ķ×ŀ×ķת +åĩºæĿ¥ ãĤĭ +ÑĦ ÑĤ +ĠìĿ´ ìłľ +ĠÑĢ ÐµÐ¼ +ĠÑĢем онÑĤ +ת ×ķ×ļ +æĻĤ ãģ¯ +ãĤīãĤĮ ãģªãģĦ +alt ı +å®¶ ãģ® +ĠاÙĦØ¥ عÙĦاÙħ +리 ëĬĶ +ãģĭãĤī ãģ¯ +ĠH ạ +ãģĤ ãģ® +×ĵ×Ļ ×ķף +رÙĬ س +Ġsoci etÃł +ĠاÙĦÙĥ بÙĬر +Ġ×ij ×ŀס +Ġ×ij×ŀס ×Ĵר +Ġ×ij×ŀס×Ĵר ת +ĠìŀĪ ìľ¼ë©° +Ġn ặng +Ùĩ Ùī +ĠB Ãł +×ŀר ×ķ +Ġj ÄĻ +ĠjÄĻ zy +ĠjÄĻzy k +Ġ׼ ×ŀ×ķ×ijף +×¢ ׾×Ķ +à¸Ĺีà¹Ī à¹Ħà¸Ķà¹ī +ãģ¾ ãģĹãĤĩãģĨ +×ŀס פר +Т Ðŀ +سÙĬاس Ø© +Ġкажд Ñĭй +ë² ł +t ım +y á»ĩn +ร ีà¹Ī +ĠдеÑĤ Ñģк +วิà¸ĺี à¸ģาร +m ówi +×ĺ×¢ ×Ŀ +×Ķצ׾ ×Ĺ×Ķ +ض ÙĬÙģ +ĠÑħоÑĤ Ñı +ãĤĵãģ§ ãģĦãĤĭ +à¸Ħา à¸Ķ +à¸Ħร à¸ļ +Ġк ÑĥÑĢÑģ +ĠbaÅŁ arı +×ijר ×ķ +ÙĬع Ø© +ĠÐĿ Ñĥ +à¸Ħวาม à¹Ģà¸Ľà¹ĩà¸Ļ +Ġ׾ ×ŀש׾ +Ġì¢ĭ ìĿĢ +Ùħؤس س +Ùħؤسس ات +Ġpréc is +Ġth ảo +à¸ģà¹ĩ à¸Ħืà¸Ń +Ġש ׼׾ +führ ung +ãģĦ ãģ§ +à¹ģละ มี +à¸ģà¹ĩ มี +Ġש ש +м ел +Ġкни г +ĠباÙĦ ÙĨ +ĠباÙĦÙĨ سبة +Ġald ı +ÑĤ ай +Ġ×Ĺ×ĵ ש×Ļ×Ŀ +å®Ł ãģ¯ +ع ÙĪØ§ +ĠìĿĺ 미 +из м +ÑĢабоÑĤ аÑĤÑĮ +Ùģ Øµ +Ġ×ij׳ ×ķסף +ãģ¨ãģĹãģ¦ ãĤĤ +à¹Ģà¸Ľà¹ĩà¸Ļ à¸Ĺีà¹Ī +ĠÑģлед ÑĥеÑĤ +èĢĥãģĪ ãģ¦ +Ġ׼ ×Ļ×ķ×Ŀ +ÑģÑĤ Ñĭ +׼׾׼ ׾×Ļ +æµģ ãĤĮ +ãĤĴ ãģ¤ãģij +Ñĩ аÑĤ +×Ļ׼ ×ķף +×Ļר ×Ļ +ları yla +ãĤ¤ ãĥ¡ +ãĤ¤ãĥ¡ ãĥ¼ãĤ¸ +׳×ĸ ×§ +Ġci ò +Ġs ın +Ġsın ır +à¸Ļ à¸Ħร +к аÑĤ +Ġl á»Ĺi +ëŀ Į +تÙģ Ø§Øµ +تÙģØ§Øµ ÙĬÙĦ +ëĨ ĵ +ĠÙħ ض +il miÅŁ +بار Ùĥ +ÐĿ Ðĺ +Ġth ẩm +Ġ×IJ×ķת ×ļ +ĠпÑĢин им +ĠпÑĢиним а +Ġyö nt +Ġyönt em +Ġ×ŀ×§ ×ij׾ +Ġktó rego +ê· Ģ +شر Ùģ +د اÙħ +ãģĦãĤį ãģĦãĤį +ĠAl ém +Ġgör ü +Ġgörü nt +Ġgörünt ü +د س +ÑĪ ÐºÐ¸ +г ÑĢад +Ġl ạc +Ġs ữa +ãĤīãĤĮ ãģ¾ãģĻ +o Ãłi +Ñī ен +ãģĭ ãģªãģĦ +Ġп оп +Ġпоп Ñĥ +ĠпопÑĥ лÑıÑĢ +ĠاÙĦÙħ ÙĪÙĤع +rä g +ï¼ ¡ +íķ Ħ +ãĤĴè¦ĭ ãĤĭ +اÙħ ا +ĠاÙĦØŃ رب +ĠÐŁ а +Ġ׾ ×IJתר +Ġt á»ijc +×ij ׾×Ķ +ر ئÙĬس +в Ñĥ +ÙĬ دÙĬ +каз ан +Ġ׊ש×ij×ķף +h ôtel +×¢ ×ķ׳×Ķ +ب ÙĨÙĬ +×ŀ ×ķ׾ +Ġд нÑı +éĽ£ ãģĹãģĦ +вед ениÑı +Ġ×ķ ×ŀת +н апÑĢимеÑĢ +ÙĤ ابÙĦ +Ġrésult at +ĠÑĢазвиÑĤ иÑı +ر Ùij +ìłĦ 문 +ĠاÙĦÙħ زÙĬد +ĠìľĦ íķ´ìĦľ +ëĨ į +íĻ ķ +ĠThi ết +íĮ ¨ +malı dır +Ġcz ÅĤ +ĠczÅĤ owie +ĠczÅĤowie k +ĠÙĦ بÙĨ +ĠÙĦبÙĨ اÙĨ +üs ü +ãģªãĤĵ ãģł +Ġżyc ie +ĠÑħоÑĢоÑĪ Ð¾ +æĸ¹ ãģ« +ëĭ¤ ë©´ +иÑĩеÑģ каÑı +ער ×Ļ׼ +ער×Ļ׼ ת +ãģ¾ãģĽãĤĵ ãģ§ãģĹãģŁ +ĠÑģоб ой +Ġg á»Ĺ +Ġдел аÑĤÑĮ +da Äĩ +аÑĢ Ð° +róż ni +à¹Ģล ีà¹ī +à¹Ģลีà¹ī ย +à¹Ģลีà¹īย à¸ĩ +à¸Ŀ าà¸ģ +Ġت ÙĤ +ĠتÙĤ دÙĬ +ĠتÙĤدÙĬ Ùħ +หà¸Ļ ุà¹Īม +Ġmü cade +Ġmücade le +ì§Ģ 를 +ãĤ¤ ãĤ¹ +ĠØ£ ساس +jÄħce go +ĠÅŁ eh +н ÑĤеÑĢ +ÑĨи Ñİ +ï» » +ÑİÑī его +à¹Ĥà¸Ľà¸£ à¹ģ +à¹Ĥà¸Ľà¸£à¹ģ à¸ģรม +Ġmie Äĩ +ØŃÙĥÙĪÙħ Ø© +ãģ§ãģĹãģŁ ãģĮ +×Ļס ×Ķ +ãĤĤãģ® ãĤĴ +Ġ×ŀ ×IJת +สุà¸Ķ à¸Ĺà¹īาย +Ġc Å© +ÙĨ سب +ĠпÑĢ Ð¾Ñĩ +Ġд ней +ĠÑįÑĤи Ñħ +׾ ×ŀת +нÑı Ñı +Ñį к +Ġì§Ģ ëĤľ +มหา วิà¸Ĺยา +มหาวิà¸Ĺยา ล +มหาวิà¸Ĺยาล ัย +d ão +ĠMá y +ĠêµŃ ê°Ģ +à¸ļุ รี +×Ĵ ×Ļ׾ +ĠÑĤÑĭ ÑģÑı +ĠÑĤÑĭÑģÑı Ñĩ +Ùģ Ùĥ +ĠÐĺ Ñģ +è¡Į ãĤıãĤĮ +פר ×ĵ +ãģ¤ ãģį +à¸Ħร à¸Ńà¸ļ +à¸Ħรà¸Ńà¸ļ à¸Ħรัว +à¸Ĥึà¹īà¸Ļ มา +ä»ĬæĹ¥ ãģ¯ +ĠìĤ¬ëŀĮ ìĿ´ +עצ ×ŀ×Ķ +п оÑĢ +ĠK ỳ +Ġ Æ¡n +Ġth Äĥm +Ùģ Ø§ÙĤ +ãģļ ãģ« +Ġ׾ קר +Ġ׾קר ×ķ×IJ +اÙģ ÙĬØ© +Ùħ ÙİØ§ +г аÑĢ +ص ÙĦا +صÙĦا Ø© +Ġ×ŀ ×ĸ×Ķ +lı ģını +Ġ×IJ ×Ļ׳×Ķ +к ÑĢо +Ġng ươi +Ġв ним +Ġвним ание +jÄħ cy +ÙĢÙĢÙĢÙĢ ÙĢ +Ñģ Ñħод +ãģªãĤĵ ãģĭ +×ŀ ×Ļ׾ +Ġ×Ķ×IJ ×Ĺ +ãĤı ãģªãģĦ +ع سÙĥر +ĠìĦ¸ ê³Ħ +ĠÑĩ его +ĠÑģÑĢед ÑģÑĤва +ĠÐł аÑģ +ãģª ãģģ +ÙĨ Ù쨳 +ר×Ļ ×ķף +Ñģ Ñĥд +ĠìĿ¸ ê°Ħ +ĠاÙĦÙħ ÙĤبÙĦ +ÙĨ عÙħ +تÙĪ Ù쨱 +ש ×ij×¢ +ı lm +ılm Ä±ÅŁ +Ġ×ľ×ª ת +تص Ùģ +×Ķפ ×ķ×ļ +à¹ĥà¸Ļ à¸Ľà¸µ +ìĿ´ ê³ł +Ùģ ÙĪØ² +à¸ľà¸¥ à¸ĩาà¸Ļ +ĠGi áo +à¸ļà¸Ńà¸ģ วà¹Īา +Ġd Ä±ÅŁ +ĠdÄ±ÅŁ ında +ì£ ½ +Ġdzie ÅĦ +к ÑĨии +и ÑĨе +ãģ® ä¸Ģ +ع Ø´ +пÑĢ ÐµÑģÑģ +หà¸Ļ à¹Īà¸Ńย +ลัà¸ģษ à¸ĵะ +Ġpossibilit Ãł +à¹Ħà¸Ķà¹īรัà¸ļ à¸ģาร +หย ุà¸Ķ +Ġphi ên +çĶŁ ãģ¾ãĤĮ +Ø· ÙĪÙĦ +ÑĦ ин +f ür +ØŃ ÙĬاة +íĸ ĪìĬµëĭĪëĭ¤ +׼ ׳×ķת +à¸Ľà¸£à¸° ส +à¸Ľà¸£à¸°à¸ª à¸ļ +à¸Ľà¸£à¸°à¸ªà¸ļ à¸ģารà¸ĵà¹Į +ëIJĺ ìĹĪ +Ġkaż dy +Ġl uyá»ĩn +ĠоÑĢганиз аÑĨии +å°ij ãģªãģı +ÑģÑĤÑĢо ен +Ġtécn ico +×§ ×Ķ׾ +Ġ×ķ×IJ ×Ĺ +ĠعÙĦÙĬ Ùĥ +Ñī ение +Ġ×Ķ ×Ļ׾×ĵ×Ļ×Ŀ +ÙĪØ³ ائÙĦ +Ġ×ķ ×Ķת +تÙħ ÙĬز +ĠÑģ казал +Ġпол и +Ġ×Ķ×ŀ ס +ÙĦÙij Ùİ +Ùħؤس سة +Ġ×ŀ ×Ļ×ĵ +ãģ£ ãģ¡ +ĠëĦĪ ë¬´ +à¸ŀ ี +Ġt ặng +Ġt ấn +ר ש×Ŀ +Ġméd ica +Ġ×¢ ×ķ×ŀ +Ġ×¢×ķ×ŀ ×ĵ +ÑĦ оÑĢ +Ùħر Ø© +Ġvat anda +Ġvatanda ÅŁ +Ġдел о +à¸Ļ ม +ãģ¨ åIJĮãģĺ +Ùģ Ùī +Ñģ оÑĢ +Ġ×Ķס ר×ĺ +Ġép oca +ìłķ ì±ħ +ĠÑģвÑıз ан +ض رب +ĠÙĦ ÙĨا +Ġuży wa +ĠاÙĦج ÙĬØ´ +Ñİ ÑĢ +×ijס ×ķ×£ +Ġм Ñĥ +ĠмÑĥ зÑĭк +bilit é +Ġma ç +س Ùİ +ت ÙĦÙĥ +ãģ ¬ +ÙĬ ÙĦا +ÑĪ Ð»Ð° +ÙĢÙĢ ÙĢ +Ġод ной +зв ан +ĠÑģ ÑĢаз +ĠÑģÑĢаз Ñĥ +ÙĨ ظÙħ +را Ùĩ +ĠÙĦÙĩ ذا +׼ ×ķר +Ġ×Ķש ×ij×ķ×¢ +Ġ×Ķש ת +ĠQu ảng +ãĥ« ãĥ¼ +ãģĪ ãģªãģĦ +×ĺ ×IJ +Ġmi á»ģn +ĠPh áºŃt +ĠاÙĦس ÙĪÙĤ +Ä Ĥ +ĠاÙĦج Ùħع +ĠاÙĦجÙħع Ø© +ÑİÑī ей +a ÅĤem +عت ÙĤد +Ø£ ÙĦÙħ +Ñģ ке +ĠìĿ´ íķ´ +ÙĨس Ø® +è¨Ģ ãģĦ +д обав +سب ÙĤ +×¢×ķר ר +ÑĤи п +ãģĿãģĵ ãģ§ +vis ión +عÙĪØ¯ Ø© +ë¨ ¹ +×ŀ ×ĸר×Ĺ +ĠØ¥ ØŃ +Ġ׾×ij ×Ļף +Ġ׾צ ×IJת +Ġyard ı +Ġyardı mc +Ġyardımc ı +İ Z +×§ פ×Ķ +tr é +liÄŁ ini +клÑİÑĩ а +Ġüret im +Ġa yrı +ĠkiÅŁ iler +à¸Ħ à¹īà¸Ļ +à¸Ħà¹īà¸Ļ หา +ĠS á»± +Ġ׼ ס +Ġ×Ľ×¡ ×£ +ĠÑĤак иÑħ +ĠXu ân +Ġл ег +Ġлег ко +Ø«ÙĤ اÙ쨩 +ÐĿ Ðŀ +ãĤ¹ãĤ¿ ãĥĥ +ãĤ¹ãĤ¿ãĥĥ ãĥķ +åIJĪ ãģĦ +Ġ×Ķש ×Ļ×ŀ×ķש +man ız +ĠÐĴ аÑģ +g ün +ìľĦìĽIJ íļĮ +Ġwsp óln +ĠÑģв ое +í ĥģ +à¹Ģà¸Ļ ีย +ÙĪØ¨ Ø© +в Ñıз +ı dır +ëIJĺ ìĹĪëĭ¤ +ĠdeÄŁi ÅŁtir +ãĤĭ ãģĵãģ¨ãģĮ +Ġ×Ĺ×ĵ ש×Ķ +ãĤīãĤĮ ãģ¦ãģĦãĤĭ +×Ĺ×Ļ ×Ļ×ij +ĠÐļ аÑĢ +׳×Ļת ×ķ×Ĺ +Ġ×§×ĺ ף +ר ×ĸ +ÙĪ Øº +èªŃ ãģ¿ +Ġت ÙĤÙĪÙħ +ĠÙĥ اÙĦ +à¸Ŀ ึà¸ģ +Ġë°ľ ìĥĿ +ológ ico +ر اع +à¹ģà¸ģà¹ī à¹Ħà¸Ĥ +ĠÑĢабоÑĤ Ñĥ +ÙĨÙij Ùİ +à¸Ńยูà¹Ī à¸Ĺีà¹Ī +ĠاÙĦØ« اÙĨÙĬØ© +ĠNh ân +Ñħ ваÑĤ +ö ne +Ġع دة +à¹ģ สà¸ĩ +ÑĤ оп +пÑĥÑģ ка +شر اء +ĠÐļ ом +Ġפע ×ķ׾×Ķ +ìĤ¬ ìĿ´ +ìĤ¬ìĿ´ íĬ¸ +è¡Į ãģ£ãģ¦ +Ġ×Ķ ×Ķת +ĠÑģÑĤ оÑĢо +ĠÑģÑĤоÑĢо нÑĭ +در س +à¸ĭ ู +à¸ķà¹Ī ำ +ĠØ£ بÙĬ +под об +ãģ« ãģ¦ +ار تÙģØ§Ø¹ +ĠÙħ ؤ +ик ов +ge führt +มืà¸Ń à¸ĸืà¸Ń +ĠÙĦ ÙĤد +ĠØ£ÙĨ Ùij +سÙĬ طر +ãģ¾ãģļ ãģ¯ +ס ×ĵ +Ñģк олÑĮко +ãģ¿ãģŁãģĦ ãģª +×ĵר ×Ĵ +×¢ ×Ļ×ĵ +à¹ĥหà¹ī à¸ļริà¸ģาร +ĠÐĶ Ð¸ +×ij×¢ ×Ļ×ķת +Ġ×Ķ×Ĺ ×ķ +пиÑģ ÑĮ +ĠاÙĦØ® ÙĦ +б ав +Ġİ lk +ĠاÙĦØ® Ùħ +ĠاÙĦØ®Ùħ ÙĬس +ĠÙĬ ÙĤÙĪÙħ +æĻĤ ãģ® +ĠsÅĤ ow +ĠØ£ ÙĩÙħ +Ø®ÙĦ ÙĤ +ĠØ£ صبØŃ +Ġchứ a +Ġth ác +Ùģ Ø§ÙĦ +Ġch á»Ŀ +ĠاÙĦØ® ار +ĠاÙĦخار ج +ĠاÙĦخارج ÙĬØ© +Ø· ائر +Ġt Ãł +ĠtÃł u +à¸ģล à¹īà¸Ńà¸ĩ +ĠاÙĦÙħر Ø£ +ĠاÙĦÙħرأ Ø© +åħ¨ ãģı +ĠÃĸ n +çļĦ ãģ«ãģ¯ +Ġpiè ce +×Ĵ ×Ļ×ij +ĠاÙĦ ÙĪØ§ÙĤع +ä»Ĭ ãģ® +ĠاÙĦÙħ ÙĤ +cz nÄħ +Ù쨹 اÙĦ +ен ного +ĠÑĦак ÑĤ +ìĭł ì²Ń +ĠÐŀ ни +ĠاÙĦبÙĦ اد +ов иÑĩ +ëı Į +ÑĦ ÑĥнкÑĨи +Ġìĸ´ ëĬIJ +ãĥķãĤ© ãĥ¼ +d ÃŃ +ил оÑģÑĮ +Ùħ Ùī +ĠاÙĦØ£ÙħرÙĬ Ùĥ +ĠاÙĦØ£ÙħرÙĬÙĥ ÙĬØ© +×ĺ ×Ļפ×ķ׾ +íĶĦ ë¡ľê·¸ +íĶĦë¡ľê·¸ ëŀ¨ +Ġש ×ķ׳×ķת +Ø´ ÙħÙĦ +ĠпаÑĢ Ð° +Ġ×Ķ×Ĺ ×ķ×§ +ÙĪØ² ارة +ãģ¨ ãģĻãĤĭ +Ġqu ảng +ĠaÄŁ ır +ĠاÙĦÙĦ ج +ĠاÙĦÙĦج ÙĨØ© +ê¸ ´ +ĠT ân +ج ÙħÙĦ +д ол +à¹ģà¸ŀ à¸Ĺย +à¹ģà¸ŀà¸Ĺย à¹Į +Ġר×IJ ש×Ļ +Ñī ей +Ġçev re +Ġкомп лекÑģ +Ġ×ij ×ŀש×ļ +Ġalt ın +ĠØ£ عÙħاÙĦ +ĠÑģво его +ãĤĪ ãģĦ +×Ĺ׾ ×Ļ×ĺ +×ŀ׳ ×¢ +Ġר ×ij×Ķ +ĠØ£ÙĬضا Ùĭ +×ĸ ׾ +ĠاÙĦسÙĬ اسÙĬ +æĢĿ ãģĨ +קר ×§ +קרק ×¢ +ĠاÙĦÙģ Ø±ÙĬÙĤ +б иÑĤ +×§ ׳×Ķ +ĠØ¥ ÙĨÙĩ +ĠÐĴ ам +Ðł Ðŀ +ãĥĪ ãĥª +å¿ħè¦ģ ãģª +Ġch âu +ç¶ļ ãģij +Ġçöz üm +gÅĤ ow +ع ÙĤÙĦ +売 ãĤĭ +i ết +à¸Ĭิ à¹īà¸Ļ +ĠØŃÙĤ ÙĪÙĤ +Ø·ÙĦ ع +ĠÄij en +ĠÙĥ اÙ쨩 +ãģ® ãģĶ +Ġë ¬ +Ġë¬ ¼ +Ġ물 ë¡ł +Ġرس ÙĪÙĦ +з ам +зам ен +Ġkullan ıcı +×¢ ×ķ׾ +èī² ãĢħ +ÑĪи ÑĢ +Ġ׊ש +Ġwy gl +Ġwygl Äħda +ש ×Ļ×ŀ×ķש +å¿ĺ ãĤĮ +×¢ ×Ļצ×ķ×ij +ĠاÙĦس ÙĪØ±ÙĬ +å°ij ãģªãģĦ +Ġпо иÑģк +สำ à¸Ļัà¸ģà¸ĩาà¸Ļ +Ġ×ŀצ ×ĵ +Ġmü ÅŁ +ĠmÃ¼ÅŁ ter +ĠmÃ¼ÅŁter i +ĠÙħÙĨ ÙĩÙħ +à¸ķำ à¹ģ +à¸ķำà¹ģ หà¸Ļ +à¸ķำà¹ģหà¸Ļ à¹Īà¸ĩ +ÅĽ mie +Ġש ×ł×ª +Ġ×Ķ ×¤×Ļ +פר ש +×¢×ijר ×Ļת +สà¸Ļ ัà¸ļ +สà¸Ļัà¸ļ สà¸Ļุ +สà¸Ļัà¸ļสà¸Ļุ à¸Ļ +è¨Ģ ãģ£ãģ¦ +à¸ģาร à¸Īัà¸Ķ +ĠMo że +из аÑĨии +ứ t +ĠÙĪØ¨ عد +ĠdeÄŁ ild +ĠdeÄŁild ir +Ġת ×ŀ +Ġ×ŀ×ŀ ׳×ķ +話 ãĤĴ +ĠÑĨ ена +Ġth úc +×Ļ×ŀ ×ķף +ĠB áo +ãĤĴ åıĸãĤĬ +å®ī ãģĦ +Ġ×¢×ķש ×Ļ×Ŀ +èĩªåĪĨ ãģĮ +l ée +ãĤĭ ãģ®ãģ§ +иÑĢÑĥ еÑĤ +ãģ¦ ãĤĭ +ست ر +ĠاÙĦØŃ ÙĬ +×Ļ׾ ×ķת +Ġ×Ĺ ×ij +ÙĤر Ø£ +تÙħ ÙĥÙĨ +س ائÙĦ +prü f +ãģĭ ãģijãģ¦ +ĠÑģоб ÑģÑĤвенно +ĠìľĦ íķĺìŬ +׾ ×Ļ×ĺ +ãģĮ å¤ļãģı +ÙĬت Ùĩا +ç«ĭ ãģ¦ +ม à¸Ńà¸ļ +ìĭľ ìŀ¥ +оÑĢ Ð° +Ġs avaÅŁ +×ĺ×Ļ×ij ×Ļ +×ij ׳×ķ +Ùħا ذا +기 ê°Ħ +ãģªãģ© ãģ§ +Ġ×ŀ ת×Ĺ×Ļ׾ +Ġnhi á»ħ +Ġnhiá»ħ m +ка ÑĢ +каÑĢ ÑĤ +Ġ׾×Ķ ×©×ª×ŀש +׳ ×Ļ×Ĺ +اد ÙĬØ© +ราย à¸ĩาà¸Ļ +Ġprzy kÅĤad +Ñī ий +ØŃض ÙĪØ± +Ġh ôn +à Ŀ +ת ×ķצ×IJ×ķת +راب Ø· +Ġb ếp +ĠполÑĥÑĩ и +åĩºä¼ļãģĦ ç³» +à¸Ľà¸¥ à¹Īà¸Ńย +ĠاÙĦØ´ باب +اÙĩ ÙĦ +ä»Ĭ ãģ¾ãģ§ +رج ع +ãĤ¶ ãĥ¼ +ÙĤ Ùģ +ĠGro ÃŁ +ĠíļĮ ìĽIJ +اج ر +Ġ×ij×ŀ קר×Ķ +Ġseg urança +fü hl +ãģ¦ ãģĦãģı +หม à¸Ń +ĠкоÑĤоÑĢ Ð¾Ð¼ +ĠN Äĥm +ĠdÅĤ ugo +ÙħÙĨ ØŃ +ש×ķ ×ķ×Ļ +ĠØ£ÙĬ اÙħ +ส à¸łà¸²à¸ŀ +r zÄħ +شر Ùĥات +ãĤĴ èĢĥãģĪ +д аÑĢ +à¸Ľà¸£à¸° à¸Ĭุม +Ġ×ķ×IJ ×ĸ +i á»ĩn +Ġt ươi +ש ×Ļ×Ĺ +à¸Ń à¹Īà¸Ńà¸Ļ +æĽ¸ ãģĦãģ¦ +Ġng ữ +×ij×Ļ×ĺ ×Ĺ +×ij×Ļ×ĺ×Ĺ ×ķף +Ġs ẵ +Ġsẵ n +ì§Ģ ëıĦ +ĠпÑĢ ÐµÐ¿ +ĠпÑĢеп аÑĢаÑĤ +Ġна ÑĥÑĩ +ĠÃľ nivers +ĠÃľnivers ites +ĠÃľniversites i +Ġ×Ĵ×ĵ ×ķ׾×Ķ +Ġ×Ķ ×ł×ª +Ġ×Ķ×ł×ª ×ij×¢ +ãģ§ãģĤ ãģ£ãģŁ +Ġmies iÄħ +ĠmiesiÄħ c +г ÑĢам +гÑĢам м +Ġبش Ø£ÙĨ +ĠÑħ ÑĢ +×§ ×Ļ×ĵ +×§×Ļ×ĵ ×ķ×Ŀ +Ø´ Ùĥر +Ġ á»ķ +Ġá»ķ n +ãģĮãģĤ ãģ£ãģ¦ +ãģķãĤĮ ãģ¾ãģĻ +Ġ×Ĺ ×ķ×ĵ +Ġ×Ĺ×ķ×ĵ ש×Ļ×Ŀ +ÙħÙĪØ§ جÙĩ +ÙħÙĪØ§Ø¬Ùĩ Ø© +أش خاص +ب غ +à¹Ģรียà¸Ļ รูà¹ī +ãģĹãģ¦ ãģĦãģı +Ġs ạn +å¿ħ ãģļ +׳ ×Ļ×Ĵ +׳×Ļ×Ĵ ×ķ×ĵ +باÙĦ غ +׊ש×ŀ +×Ĺש×ŀ ׾ +Ġnap raw +Ġnapraw dÄĻ +Ø´Ùĩ اد +×IJ ×ķ×Ķ +×IJ×ķ×Ķ ×ij +и ÑĨÑĭ +Ġ×Ķ ×¨×Ľ×ij +ëŀ ij +Ġת ×¢ +Ġ×Ķ ×Ļש +Ġ×Ķ×Ļש ר×IJ +Ġ×Ķ×Ļשר×IJ ׾×Ļ +Ø£ ÙħÙĨ +ÑİÑī аÑı +sk ór +LER İ +Ġ×Ķ×IJ×Ĺר ×ķף +×¢ ׳ק +ĠÙĪ ÙĥÙĦ +ãģĵãģĵ ãģ§ +Ġqu án +liÄŁ in +à¸ģà¸İ หมาย +Ø· Ùħ +Ø£ جÙĩ +أجÙĩ زة +ĠEr doÄŁan +ãģ§ ãģĬ +Ġв ÑĢа +ĠвÑĢа Ñĩ +ĠPh ó +à¸Ĭั à¹Īว +à¸Ĭัà¹Īว à¹Ĥม +à¸Ĭัà¹Īวà¹Ĥม à¸ĩ +Ġph úc +×Ļפ ×ķת +×¢×Ļ ×ķף +Ġduż o +ãĥģ ãĥ¼ãĥł +ĠÙĬ Ùİ +Ġзад аÑĩ +Ġ×Ĵ×ij×ķ×Ķ ×Ķ +Ġ׼ ׼׾ +лож ен +ét at +Ġng Äĥn +èµ· ãģį +ĠTi ến +ص عب +Ġexperi ência +Ø® Ùħ +à¸ģาร à¸Ĺำà¸ĩาà¸Ļ +س ÙĬد +ĠD á»± +ĠкоÑĤоÑĢ Ð¾Ð³Ð¾ +lad ıģı +Ġkh á»ķ +Ġê³Ħ ìĨį +Ñī ик +สà¹Īวà¸Ļ à¸ķัว +з оÑĢ +ÙĨ Ùı +Ġ à¸Ķัà¸ĩ +Ġà¸Ķัà¸ĩ à¸Ļัà¹īà¸Ļ +Ġc ấu +ĠÄij á»ijc +о ÑĦ +ĠاÙĦØ£ عÙħاÙĦ +ãģªãģı ãģ¦ãĤĤ +×ķ׼ ×Ļ×Ŀ +à¹ģ à¸Ľ +ĠB ên +ãĥ¯ ãĥ³ +Ġgi ám +ĠÅŀ u +Ġd áng +ع ÙĦÙĬ +à¹Ģà¸ģ ษ +à¹Ģà¸ģษ à¸ķร +ÙĪØ¬ ب +н нÑĭе +ÙĤ ضاء +à¸Ħว à¸ļ +à¸Ħวà¸ļ à¸Ħุ +à¸Ħวà¸ļà¸Ħุ ม +ãģ¤ ãģ¤ +ĠVi á»ĩc +×ŀ×ij ×ĺ +ש×Ļת ×ķ×£ +Ġв едÑĮ +k aza +kaza ÅĤ +à¸ķำ รวà¸Ī +ãĤ¿ ãĥ« +Ġпов Ñĭ +ĠповÑĭ ÑĪен +ĠS ợ +ĠìĦ¤ ëªħ +ĠÃĩ ünkü +ìĥĿ íĻľ +Ö ¾ +ãĤĮ ãģ¦ãģĦãĤĭ +Ġ×ij ר×IJש +ר ×ķ×Ĵ +Ġо ÑĦи +ĠоÑĦи ÑĨиалÑĮн +ĠÑĥ ÑģÑĤанов +ĠÑĥÑģÑĤанов лен +ĠاÙĦÙħ صر +ĠاÙĦÙħصر ÙĬØ© +ĠÐŁÐ¾ ÑįÑĤомÑĥ +ÙĨ صÙģ +ĠÙĪØ§ÙĦ ÙĨ +Ġh Ãłi +à¸Ħ ิ +ĠApr ès +ì³ IJ +à¹Ģà¸ĭ ีย +×ĵ ×ŀ×Ķ +activ ité +à¸Ħิà¸Ķ วà¹Īา +ÑĤ ÑĢен +à¹Ģ ฮ +ãĥı ãĤ¤ +ãģĮ å¢ĹãģĪ +ен наÑı +Ġìĺ¤ ëĬĺ +ãĥ¢ ãĥ³ +Ġкон еÑĩно +ĠÙħÙĤ ابÙĦ +cl é +Ġh ü +Ġth ẳng +ìłģ ìĿ´ +ĠÐIJ лекÑģ +ĠÐIJлекÑģ ан +ĠÐIJлекÑģан дÑĢ +ãĥŀãĥ³ ãĤ·ãĥ§ãĥ³ +ãģ²ãģ¨ ãģ¤ +ãģª ãģĬ +à¹Ģà¸Īà¹īา à¸Ĥà¸Ńà¸ĩ +ëĵľ 리 +Ø´ اء +ĠsaÄŁ lık +ĠÅŁ imdi +×Ļ×IJ ׾ +تأ Ø«ÙĬر +Ø£ سب +أسب اب +ĠвÑĭполн ен +л ок +ש ×Ļ×ij×Ķ +Ġl ắm +ĠTr Æ°á»Ľc +Ġ×Ķ×¢ ׾ +리 를 +ĠÑĢ ÐµÐ¶ +ĠÑĢеж им +int é +inté gr +×Ĵ ׳×Ļ +ĠاÙĦØ´ عر +Ġmil hões +Ġpeque ño +ãĤ³ ãĥ¼ãĤ¹ +×ķ׼ ×Ĺ +à¹Ģà¸Ĭ à¹īา +شر ÙĤ +Ġh ương +รัà¸IJ à¸ļาล +à¸ģล าย +à¸ģลาย à¹Ģà¸Ľà¹ĩà¸Ļ +Ġпод Ñħод +תש ×ķ×ij×Ķ +ãģıãģª ãģ£ãģ¦ +ĠاÙĦØ£Ùħ Ùħ +ĠH á»įc +ĠwspóÅĤ pr +ĠwspóÅĤpr ac +Ñĩ Ñĥв +ÑĩÑĥв ÑģÑĤв +ÃŃst ico +à¹Ģà¸ģ าะ +ìĽ Ģ +Ġназ ад +ãĤĭ ãĤĪãģĨãģ« +ĠС Ш +ĠСШ ÐIJ +м он +ĠAs ÃŃ +×ķר ×Ĵ +полн ен +×ŀס ׾ +×ŀ×¡×ľ ×ķ׾ +à¹Ģลืà¸Ń à¸Ķ +à¹Ģริà¹Īม à¸ķà¹īà¸Ļ +ĠاÙĦØ¥ Ùħ +ĠاÙĦØ¥Ùħ ارات +צ×Ķ ×¨ +ãĥ¡ãĥª ãĥĥãĥĪ +ĠпоÑĤ ом +в из +ĠÙģ ØªØ±Ø© +å¾Į ãģ® +ÐĿ ÐIJ +×ŀס ר +ÙĬر ÙĬ +pr é +Ġte ÅŁek +ĠteÅŁek kür +Ġöd eme +د اÙĨ +ãģ¾ ãģĹãģ¦ +缮 ãģ« +ĠÑĤ еÑĩение +l ard +lard ır +à¹Ģรา à¸Īะ +ס פ×Ļ +ĠÙĪÙĥ ذÙĦÙĥ +Ġh át +Ġt á»Ļc +à¸Ħุ ย +Ġb ức +ØŃ ÙĬÙĨ +èģŀ ãģĦãģ¦ +Ùħؤ شر +ĠNh ư +Ġмен ее +ละ à¸Ħร +Ñģ ин +ĠÑĢ ÐµÐº +ĠÑĢек л +ĠÑĢекл ам +ĠÙģ ÙĩÙĪ +Ġ׾ ×ĸ +×Ļ׳ ×ķת +ĠÅŁ art +ÑģÑĤав ка +Ġíı¬ íķ¨ +ãģ«è¡Į ãģı +ï¼ Ŀ +ĠпозволÑı еÑĤ +Ġת×ķ׼ ׾×ķ +ов ал +صÙĦ Ø© +Ġ׾ש ׳×ķת +ĠÐĺ гÑĢ +ÙħÙĨتج ات +Ġsat Ä±ÅŁ +Ñģ ко +ĠاÙĦØ«ÙĦاث اء +Ġ×Ķ×ĵ×ijר ×Ļ×Ŀ +ãģĹãģ¾ ãģĹãĤĩãģĨ +بÙĤ Ùī +åĬĽ ãĤĴ +ĠÃĩ ok +ãĥģ ãĥ¥ +à¹Ģà¸Ĭ ืà¹īà¸Ń +ยุ à¸Ħ +ศา ล +Ġ×§×ķ×ĵ ×Ŀ +×ĸר ×Ļ×Ŀ +ãģ® åł´åIJĪ +ĠìķĬ ìķĺ +ãģĤãĤĬãģ¾ãģĻ ãģĮ +×IJ שר +è¡Į ãģı +ãģ» ãģĭ +æ°Ĺ ãģ«ãģªãĤĭ +й деÑĤ +íķĺìĺĢ ëĭ¤ +ستÙħر ار +ĠÐŁÑĢ Ðµ +ĠÑģ боÑĢ +ĠìķĦ 무 +ç§ģ ãĤĤ +ع ص +Ġн иÑĩ +ĠниÑĩ его +ĠпÑĢи ем +×§ ×ķ×ŀ +ĠìĪĺ ëıĦ +Ġì ¡´ +Ġì¡´ ìŀ¬ +ĠØ£ Ø«ÙĨ +ĠأثÙĨ اء +ĠÙĪØ§ÙĦ ØŃ +ãģĮ ãģ§ãģįãĤĭ +Ġת ×Ķ +Ġת×Ķ ×Ļ×Ķ +ר ף +ĠÑģвÑıз и +×Ĵ שת +Ñģп екÑĤ +ס ×ij×Ļ×ij +ס×ij×Ļ×ij ×Ķ +ĠíķĦìļĶ íķľ +ت خصص +Ġж ив +Ġжив оÑĤ +ĠMay ıs +تع ا +تعا ÙĪÙĨ +ĠعÙĨ Ùĩا +ów ki +ĠاÙĦÙģÙĦسطÙĬÙĨ ÙĬ +ãģłãģijãģ§ ãģªãģı +ìĿ¸ ì§Ģ +ĠاÙĦس ÙĪØ¯ +ĠاÙĦسÙĪØ¯ اÙĨ +إجراء ات +Ġkö tü +Ġ×Ļ ×ª×¨ +×Ĵ ×Ļש×Ķ +Ġצ ×ķר×ļ +รà¸ĸ ย +รà¸ĸย à¸Ļà¸ķà¹Į +Ñħ оÑĤ +Ðł ÐIJ +ÙĪ Ø·ÙĨ +Ġsay ısı +ס ×Ĺר +Ùħ ÙĪÙĦ +ãĤĴæĮģ ãģ£ãģ¦ +ع اÙĨ +Ġt á»Ļi +ĠвÑĭ ÑĪе +Ġt ầm +ãĥĪ ãĥ¬ +×Ļצ ×ķ +ม ุม +س ÙĪØ¯ +ìłĦ ìŀIJ +ãĤµ ãĥŃãĥ³ +ìĤ° ìĹħ +ĠоÑģнов ан +Ø® Ù쨶 +רצ ×Ķ +بÙĬ ض +×ķÖ ¹ +ס×Ļ ×Ļ×¢ +Ġש ×IJ×Ļ +ĠاÙĦÙĤر Ø¢ÙĨ +ĠТак же +×ŀש ×ŀ×¢×ķת +س ÙĩÙĦ +Ġ×Ķ ×ł×Ķ +ãĤĴ ãģĹãģ¦ãģĦãĤĭ +×Ļ ×Ļס +×Ķ ×ķ×IJ +ĠB ÃŃ +Ġмал о +ĠëͰëĿ¼ ìĦľ +Ġר ×Ĺ×ij +ãģĮ é«ĺãģĦ +ÙĪ Ø§Ø³ +ìĤ ¼ +׳ ×¢ +ãģ£ ãģ¡ãĤĥ +ĠT üm +à¸Ńีà¸ģ à¸Ķà¹īวย +ãģĹãģ¦ ãģıãģłãģķãģĦ +ÙĨØ´ اط +ãĥĹ ãĥ©ãĥ³ +али ÑģÑĮ +×ĵ ×ľ×ª +Ġwc zeÅĽ +ĠwczeÅĽ niej +ĠÑįÑĤ им +Ġthá»ĭ t +à¸ļ ัà¸į +à¸ļัà¸į à¸Ĭี +ãģļ ãģ£ãģ¨ +ÑĢ Ð¸Ð½ +Ġswo jÄħ +íķĺëĬĶ ëį° +Ġë§Įëĵ¤ ìĸ´ +تش Ùĥ +تشÙĥ ÙĬÙĦ +ائ Ùĩ +Ġ׾פ ×Ĺ×ķת +ãĥĭ ãĥ¥ +ãĥĭãĥ¥ ãĥ¼ãĤ¹ +׼×IJ ף +ãģ§ãģį ãģŁ +зв он +Ġsta ÅĤ +×Ĺ×ijר ת×Ļ +ĠØ£ عÙĦÙĨ +à¹ģà¸ļà¸ļ à¸Ļีà¹ī +بد Ø¡ +ãĤģ ãģŁ +Ġ×ŀש ×ŀ×¢×ķת +Ġ×ŀש×ŀ×¢×ķת ×Ļ +ör ü +Ġh ạnh +z ähl +ĠL ý +Ġ×ij ×Ķת +Ġ×ij×Ķת ×IJ×Ŀ +б аÑĢ +ì¦ Ī +ä»ĬåĽŀ ãģ® +Ġy ü +Ġyü ks +Ġyüks el +ãĤ½ ãĥ¼ +ãģĤ ãĤĮ +ת ׾×ŀ×Ļ×ĵ +ãģ¤ ãģª +×ij ׳×Ļ×Ŀ +Ġx ếp +ĠмÑĥж Ñĩин +ĠاÙĦÙĥ تاب +׼ ×ŀ×ķת +Ġç e +Ġçe ÅŁ +ĠçeÅŁ it +ĠçeÅŁit li +×ĵ ×Ļר×ķת +à¸ļุ à¸į +ĠاÙĦØ¥ ÙĦÙĥ +ĠاÙĦØ¥ÙĦÙĥ ترÙĪ +ĠاÙĦØ¥ÙĦÙĥترÙĪ ÙĨÙĬ +ĠباÙĦØ¥ ض +ĠباÙĦإض اÙ쨩 +Ġyö nel +Ġyönel ik +mys ÅĤ +à¸Ķà¹īวย à¸ģาร +à¸ģาร à¸Ĺำ +ов Ñĭм +Ø£ زÙħØ© +æİ¢ ãģĹ +íļ ¨ +Ġ×ķ×IJ ×Ŀ +Ġnghi êm +ÑĪ Ð¸Ð½ +ка л +Ġcrian ças +èĩªåĪĨ ãģ§ +Ġн ай +Ġнай ÑĤи +ĠS á»ij +ĠÃ¶ÄŁrenc iler +ãĥ¶ æľĪ +Ñģ ан +ĠJ á +ĠkonuÅŁ ma +شر Ø· +ëĪ Ī +ar rière +ضر ÙĪØ±Ø© +ãĥĶ ãĥ³ +×¢ שר +аÑĢ ÑĮ +جÙħ اع +Ġdé co +Ġ×Ļ×Ķ ×ķ×ĵ×Ļ +à¸ŀ ลาà¸Ķ +ĠÙĬ ÙĥÙĨ +Ġج اÙħعة +Ø· بÙĤ +Ġbo ÅŁ +×ķ ×ķ×IJ +×ŀ×ĵ ×¢ +×§×ij×ķצ ת +פ ×Ļר +jÄħc ym +ÙħØ´ ا +Ùħشا ÙĥÙĦ +צ פ×ķף +Ø¥ ست +×ŀ׼ ר +سÙħ ع +Ġкак ой +ÑĤ воÑĢ +ØŃ ج +Ù쨱 ض +пÑĢав лен +Ġник ак +Ġmi á»ĩ +Ġmiá»ĩ ng +ü ÃŁ +иÑĢов ал +׾ ×ŀ×ķת +次 ãģ® +ÙĦ Ø· +à¸ķ ัà¸Ļ +×Ķ ×ª×Ĺ×Ļ׾ +Ġfoto ÄŁ +ĠfotoÄŁ raf +طر ØŃ +à¸Ńà¸Ńà¸ģ à¹Ħà¸Ľ +Ġy ên +Ġп ок +Ġпок Ñĥп +ĠпокÑĥп а +ÑĨ Ñĥ +Ġкомп ÑĮÑİ +ĠкомпÑĮÑİ ÑĤеÑĢ +ĠاÙĦÙĥ رÙĬÙħ +تص Ùħ +تصÙħ ÙĬÙħ +Ġоказ а +Ġzar ówn +Ġzarówn o +ëĮĢ ì¶ľ +ãĤ»ãĥ³ ãĤ¿ãĥ¼ +Ġjako ÅĽci +æĤ © +æĤ© ãģ¿ +Ø£ÙĨ ÙĪ +Ø£ÙĨÙĪ Ø§Ø¹ +ë¹ ł +Ġìłķ ë§IJ +Ġk ẻ +ĠÑģай ÑĤа +Ġ×Ķ ×¢×¨×ij +Ùĩ ز +pres ión +ĠÑģÑĤ ен +ãģ£ãģ¦ ãĤĭ +Ġhız lı +Ðļ ÐIJ +×ŀשפ ×Ĺת +ĠÙĨ Ùĩا +ĠÙĨÙĩا ÙĬØ© +ãģ¾ ãģĦ +о ÑħÑĢан +ร à¹īà¸Ńย +ล ึà¸ģ +ĠÙĪØ¨ اÙĦ +ãĤĤãģ® ãģĮ +ר׼ ×Ļ×ij +ãĤ¤ ãĥ¤ +س ؤ +سؤ اÙĦ +ĠÙĦØ£ÙĨ Ùĩ +ĠkonuÅŁ tu +Ðļ ÑĥпиÑĤÑĮ +Ġש×IJת ×Ķ +ĠÙĪØ§ÙĦ س +Ġmożliwo ÅĽci +Ġpró b +ëĶ ° +ãģ© ãĤĮ +ĠÐľ ин +ĠоÑĢганиз м +ãģ«å¯¾ ãģĻãĤĭ +ĠPr é +Ġpriv é +ch è +ãģĦãģŁãģł ãģį +สà¸Ļุ à¸ģ +ajÄħ ce +ĠD zi +ĠDzi ÄĻki +ÅĤat w +r än +rän k +æĿ¥ ãģŁ +Ġ×Ķ×Ļ×Ķ ×ķ×ĵ×Ļ +ãĤ¬ ãĥ¼ +ĠÑĢаР´ +ĠÑĢад и +к ÑĤив +Ø£ Ùĩد +Ø£Ùĩد اÙģ +ש ×IJ×Ļר +ãģ¦ ãģĦãģªãģĦ +Ġfr üh +Ġок ол +Ġокол о +Ġreg ião +ĠÑĩиÑģ ле +Ġpon iew +Ġponiew aż +ìĦ¼ íĦ° +Ġb ầu +Ġê · +Ġê· ľ +Ġê·ľ ìłķ +ĠH òa +ĠÑĤ оÑĤ +ãĤĤ å¤ļãģĦ +ĠاÙĦإسÙĦاÙħ ÙĬØ© +ãģĭ ãģĦ +Ñį н +ĠÑĥказ ан +ĠÑĤак ое +ï¼ ³ +ëĮĢ íķĻ +Ġgen iÅŁ +ĠاÙĦØ® ÙĬ +ĠاÙĦØ®ÙĬ ارات +ãĤĴè¡Į ãģĨ +ש ×ŀ×Ķ +ĠLÃł m +ÙĪÙĨ ÙĬ +Ġ×IJ ׾×Ļ×ķ +Ä ĺ +à¹Ħมà¹Ī สามารà¸ĸ +人 ãģ¨ +بر ز +×Ļס ×ķ×ĵ +×Ĵ ׾×Ļ +ĠÙĬ ÙĨا +ĠÙĬÙĨا ÙĬر +ĠкаÑĢÑĤ ин +Ġt ôn +à¹Ģ à¸ģร +à¸Ħ à¸Ķี +Ġ׾×IJ ×ķר×ļ +ãĤĤãĤī ãģĨ +ãģĭ ãģĭãĤĭ +ани и +Ġara ÅŁtırma +ÙĦاØŃ ظ +ãģĦ ãĤĦ +ĠT Ãłi +Ġ à¸Ļà¸Ńà¸ģà¸Īาà¸ģ +Ġà¸Ļà¸Ńà¸ģà¸Īาà¸ģ à¸Ļีà¹ī +ĠÄIJ ảng +ãģ£ãģ¦ ãģįãģŁ +Ġà¸ĭึà¹Īà¸ĩ à¹Ģà¸Ľà¹ĩà¸Ļ +Ġt ả +Ġmożliwo ÅĽÄĩ +ĠS ản +Ġİ ki +Ġc ắt +س Ø£ÙĦ +Ġbak ım +Ø´ ب +à¸ķ ีà¹ī +à¸ŀ ยาย +à¸ŀยาย าม +สั à¸Ľ +à¸ªà¸±à¸Ľ à¸Ķา +à¸ªà¸±à¸Ľà¸Ķา หà¹Į +ë° Ģ +еÑĢ Ñĭ +Ġc ánh +Ġthu ế +ت بع +ãģ«åħ¥ ãĤĮ +Ñİ ÑģÑĮ +íļĮ ìĿĺ +ç°¡ åį +ç°¡åį ĺ +ç°¡åįĺ ãģ« +Ġtr úc +ĠاÙĦÙĥ ÙĪÙĬ +ĠاÙĦÙĥÙĪÙĬ ت +ãĤıãģij ãģ§ãģĻ +ĠÑģв об +ĠÑģвоб од +ĠÑĥÑĩаÑģÑĤ ник +สิ à¹īà¸Ļ +ĠпÑĢо ÑĦеÑģÑģиона +ĠпÑĢоÑĦеÑģÑģиона лÑĮн +Ñģп оÑĢ +×Ĺ ×ķ×ij×Ķ +Ùħع ÙĨÙī +ĠاÙĦÙģ ØªØ±Ø© +สูà¸ĩ สุà¸Ķ +ãĤı ãģļ +ĠÄij è +ĠÄijè n +æ¯Ķ ãģ¹ +า à¸ĺิ +Ġmoż emy +à¹ģ à¸ĭ +à¸Īะ à¹Ħมà¹Ī +Ġs ắp +Ðļ Ðŀ +Ġprá ctica +ÙĪÙĥ اÙĦØ© +è¾¼ ãĤĵãģ§ +ológ ica +Ġе Ñī +ĠеÑī Ñij +تع دÙĬÙĦ +ĠØ£ Ùĥد +Ġצר ×Ļ׼ +Ġצר×Ļ׼ ×Ļ×Ŀ +Ø« Ùħ +Ġк ÑĢÑĥ +ĠкÑĢÑĥ п +×ij×Ļ×§ ×ķרת +Ġì¡° ê¸Ī +ãģ¨ãģį ãģ¯ +Ġb ạc +ĠÑĢаÑģ пол +ĠÑĢаÑģпол ож +ĠÑĢаÑģполож ен +ز ÙĬÙĨ +ĠÐļ ÑĢоме +ĠاÙĦÙĨ ظر +×Ķ ×ķ×ĵ +ĠاÙĦس بت +ã썿ĢĿ ãģĦ +Ġpa ÅĦst +ĠpaÅĦst w +ĠÙĦÙĬ ست +ĠбÑĥд Ñĥ +à¸Ĺัà¸Ļ à¸Ĺี +ร าม +ØŃ صÙĪÙĦ +ãģĹãģ¦ãģıãĤĮ ãĤĭ +ĠاÙĦØ¥ سرائÙĬÙĦ +ĠاÙĦإسرائÙĬÙĦ ÙĬ +ãģĵãĤĮ ãģ¾ãģ§ +ìĤ¬ 를 +Ġs ürü +à¹Ģว à¸Ńรà¹Į +à¹Ģà¸ĭ à¸Ńรà¹Į +Ġutilis é +ĠÑģиÑģÑĤем а +Ġdw ó +Ġdwó ch +Ġpróp rio +Ġëĵ± ìĿĦ +arr êt +ĠЧ а +×IJ×ŀ ׳×ķת +عار ض +à¹Ģà¸ģม สà¹Į +Ġ׾×Ķ ×ij×Ļף +Ġ׾ ×ij×Ĺ +Ġ׾×ij×Ĺ ×ķר +สา à¸Ĥา +ĠÐľÐ¾Ñģк ве +ب عد +ĠاÙĦÙĤر ار +ĠÄIJ á»ĭa +Ġ×Ĺ ×Ĵ +Ùģ ØªØ± +ÙĪÙĨ Ø© +Ġ×Ķ×ĸ ×IJת +å¸Ĥ ãģ® +ãģ» ãģĹãģĦ +Ġ×ij×¢ ×Ļר +ĠÑĤеп еÑĢÑĮ +ìĬµ ëĭĪê¹Į +à¹Ħม à¹Īว +à¹Ħมà¹Īว à¹Īา +à¹Ħมà¹Īวà¹Īา à¸Īะ +×ŀ ×IJ×Ķ +æĥħ åł± +æĥħåł± ãĤĴ +غ ÙĨ +Ġпо Ñı +ĠпоÑı ви +éģİ ãģĶ +تش غ +تشغ ÙĬÙĦ +в ел +Ġ×Ĺ ×ŀ +ãģ¨ãģªãĤĬ ãģ¾ãģĻ +Ġra ÄŁ +ĠraÄŁ men +ãģĭ ãģ©ãģĨ +ãģĭãģ©ãģĨ ãģĭ +ен ко +ì§Ģ ê³ł +Ġ×IJ׾ ×Ļ×Ķ +ĠØ£ ÙĦ +à¸Īำ หà¸Ļ +à¸Īำหà¸Ļ à¹Īาย +nız ı +Ġ׾ק ×Ĺת +Ø£ ÙĩÙħ +Ø£ÙĩÙħ ÙĬØ© +ت غÙĬر +ש ×Ĺר +ס×ķפ ר +×ĵ ×Ļר +èī¯ ãģĭãģ£ãģŁ +×ŀ׾×Ĺ ×ŀ×Ķ +ÑģÑĤв ие +ÑĤ ÑĢаÑĤ +ĠاÙĦØ£ Ø® +ĠاÙĦأخ ÙĬرة +ĠاÙĦØŃ صÙĪÙĦ +Ġcréd ito +צ ×Ļ×¢ +ãĥ¬ ãĥĻãĥ« +بر ÙĬ +ëIJ IJ +ãģł ãģ£ãģ¦ +Ġreal tÃł +س Ù쨱 +×ķ׳ ×ķ +×Ĵ ×ķ×ĵ +×Ĵ×ķ×ĵ ׾ +ฮ า +ãģĹãģ¦ ãģĬãĤĬãģ¾ãģĻ +Ġg Ãł +Ġ׾×ij צע +å¼ķ è¶ĬãģĹ +Ġ×ŀ ×Ļ׾×Ļ +Ġ×ŀ×Ļ׾×Ļ ×ķף +Ùħ در +Ùħدر سة +פ ×ķ×ĺ +à¸Ļà¹īำ มัà¸Ļ +ëģ Ŀ +ع Ùĥس +ĠÙĤ ض +ĠÑĢÑĭ б +خط Ø· +×ŀ×ķס ×ĵ +Ġ׼׾ ׾×Ļ +ĠкоÑĤоÑĢ Ð¾Ðµ +צ×Ļ ×ķף +ĠмеÑģÑĤ а +ãģĭ ãģ¤ +г ÑĢÑĥпп +׾ ×Ļ׾ +ת ×ķ×IJר +ë³µ ì§Ģ +à¹ģà¸ľ à¹Īà¸Ļ +Ġ×ij×¢ ת +æĻĤéĸĵ ãĤĴ +ï¼ £ +ãģ¨ãģĦãģĨãģĵãģ¨ ãģ§ +Ġ׾×Ķ ×§ +Ġ׾ ×ĸ×Ķ +ĠìłĢ ëĬĶ +ĠاÙĦØ¥ رÙĩاب +ĠìŀĪëĬĶ ëį° +ĠÑĤ огда +Ġ×Ķ ×¦×Ļ +×ķ׾ ×ĺ +Ġר פ×ķ×IJ×Ļ +ãģĵãģ¨ ãģ§ãģĻ +ĠÄij ÃŃch +ØŃ ÙĬا +Ġ×Ķ×ŀש ×Ĺ×§ +ãģľ ãģ² +Ġ×ŀ×IJ פשר +ãģ¿ ãģ¾ãģĹãģŁ +ĠاÙĦØ£ÙħÙĬر ÙĥÙĬ +Ùħج تÙħع +Ġس اب +Ġساب ÙĤ +׼ ×Ļ׾ +Ạ¾ +ãĥª ãĤ¹ãĥĪ +Ġì ĥ +Ġìĥ Ī +ĠìĥĪ ë¡ľ +ĠìĥĪë¡ľ ìļ´ +ĠD á»ĭch +à¹Ģหมาะ สม +ĠاÙĦÙĨ بÙĬ +׾ ׾ +ÙĨ ع +Ðĵ лав +Ðĵлав наÑı +Ùħر ض +Ġ×ķ ×ĵ +ت ÙĤÙĬ +تÙĤÙĬ ÙĬÙħ +Ġb ảng +ĠÙģ ÙĤاÙĦ +×¢ ×ŀ×Ļ +д ÑĢа +Ġsu á»ijt +سر عة +Ġc á»Ń +Ġ×Ķ ×Ļ×Ĺ×Ļ×ĵ +سع ÙĬد +à¸Ńา à¸Ĭีà¸ŀ +Ġس ÙĪØ§Ø¡ +ãĤ½ ãĥķãĥĪ +Ġл иÑĩно +ĠÐļ оÑĢ +اÙĩ تÙħ +اÙĩتÙħ اÙħ +à¸Ń à¸Ķี +à¸Ńà¸Ķี à¸ķ +ãģIJ ãĤīãģĦ +Ġiht iya +Ġihtiya ç +ãģ¾ãģ§ ãģ® +ìĭľ ìĬ¤ +ìĭľìĬ¤ íħľ +ÑĢÑĥ ÑĪ +ãĤĦ ãģ£ãģ± +ãĤĦãģ£ãģ± ãĤĬ +к еÑĢ +Ġ ży +Ġży w +кл он +Ġl ượt +à ¾ +да Ñĩи +tür k +غ ÙĪ +ĠигÑĢ Ð¾Ðº +Ġph ê +Ġש ×¢×ľ +ĠاÙĦÙħ دÙĨÙĬ +ĠìŬ룬 ë¶Ħ +ער ×Ļ×Ŀ +Ñħод ÑıÑĤ +Ġx ứ +ÐĹ Ð° +ĠÙģ Ø±Øµ +à¸Īะ à¸Ĺำà¹ĥหà¹ī +íģ ´ +×¢ ×ij×ķר +à¹Ģหลà¹Īา à¸Ļีà¹ī +èĢĥãģĪ ãĤĭ +ÑĢ ÐµÑģÑĤ +н нÑĭй +Ġc ầm +دا Ø®ÙĦ +ĠÙħÙĦÙĬ ار +ĠÐIJ л +ĠвÑĢем ен +à¸Ĭà¹Īวย à¹ĥหà¹ī +ר×Ļ ×ķת +ëĵ ¯ +飲 ãģ¿ +׳ ׾ +שת ×£ +ĠاÙĦسعÙĪØ¯ ÙĬ +u ÃŁ +ìĿ¸ ëį° +ĠìĿ¼ ë°ĺ +ÅĤ ÄĻ +Ġm á»iji +×ŀ ×Ļ׳ +ĠاÙĦØ£ Ø·Ù쨧ÙĦ +Ġçı kan +é cole +×§ ×Ļש +×§×Ļש ×ķר +ĠоÑģ ÑĥÑīеÑģÑĤв +ĠоÑģÑĥÑīеÑģÑĤв лÑı +×ij ×IJר +à¹Ħà¸Ľ à¸Ķà¹īวย +Ġ×¢ ×ķ׾×Ķ +à¸ģà¹ĩ à¹Ħมà¹Ī +ãĥ¢ ãĥĩ +ãĥ¢ãĥĩ ãĥ« +تØŃ ÙĪÙĦ +Ġод ного +ת×Ĺ×Ļ׾ ת +Ġت Ø® +Ġch cia +Ġchcia ÅĤ +ãĥIJ ãĥ³ +èĢħ ãģ¯ +ĠÙħ ØŃÙĦ +Ñģл ож +Ñģлож н +Ġt ÄĻ +Ġçı kt +Ġçıkt ı +ĠC Æ¡ +à¹Ħà¸Ķà¹ī à¹Ģลย +ır ken +à¹Ģà¸Ĥà¹īา สูà¹Ī +ÙħØŃ Ùĥ +ÙħØŃÙĥ ÙħØ© +à¸Ħุ à¹īม +à¸Ļà¹Īา à¸Īะ +лÑİ Ð´ +де ÑģÑı +деÑģÑı ÑĤ +ĠлÑİб ой +تØŃر ÙĬر +צע ×ĵ +Ġе Ñij +ĠاÙĦØŃ ÙĥÙħ +Ġص باØŃ +à¹Ģà¸ļ à¸Ńรà¹Į +Ġróż nych +ги б +ĠÑģ оÑĤ +ĠÑģоÑĤ ÑĢÑĥд +ĠÑģоÑĤÑĢÑĥд ник +ĠобÑĬ ем +פ ×ĺר +ãģĻãģĶ ãģı +ãģ«éĸ¢ ãģĹãģ¦ +в ол +Ø« ÙħاÙĨ +Ġd ần +æĬ ľ +æĬľ ãģij +Ġ×¢ ש +Ġעש ×ķ×Ļ +ס ×ķף +ãģªãģ® ãģ§ãģĻ +ãģ¯ ãģ©ãģĨ +×ŀ×¢ ר×ij +ï¼ ° +Ùħ صر +ÙħÙĨ اسب +ÙħÙĨاسب Ø© +ä¸Ĭ ãģ® +×IJ×Ļש ×ķר +ĠìĦ¤ ì¹ĺ +×ŀ×ĵ×Ļ׳ ×ķת +×ŀר ת +ãĤĭ ãģ®ãģĮ +د Ùİ +ĠاÙĦشر Ùĥات +ìĭľ ê°Ħ +ĠÑĢеÑĪ ÐµÐ½Ð¸Ðµ +ãģĻãĤĭ ãģ®ãģ¯ +ĠìŀIJìĭł ìĿĺ +׾ ×ŀ×ķ +ãģ¨ãģĵãĤį ãģ§ +Ġ×§ צר +Ġmã i +Ġkü ltür +ãĥ©ãĤ¤ ãĥĸ +à¸ľà¸¹à¹ī หà¸įิà¸ĩ +æĻĤéĸĵ ãģĮ +клÑİÑĩ и +diÄŁ iniz +มาà¸ģ à¹Ĩ +تØŃ ÙħÙĦ +Ġh ạt +ãĤ¦ ãĤ£ +п ле +×ŀ ׾×IJ +ÅĤ ó +Ġg á»ijc +Ġ×IJ ×ķ×ĵ×ķת +หว าà¸Ļ +ĠاÙĦ ÙĪØ² +ĠاÙĦÙĪØ² راء +ëĵ¤ ê³¼ +Ġص ØŃ +ĠصØŃ ÙĬÙ쨩 +Ġм м +تد Ø®ÙĦ +Ġpersön lich +Ġز ÙĬ +ĠزÙĬ ادة +ãĤ· ãĤ¢ +Ġng ắn +à¸Ħล ิà¸ģ +Ġs ông +Ġtü ket +Ñį ÑĦÑĦ +ÑįÑĦÑĦ екÑĤ +ש ×Ļ×ij +Ġا عت +ت ض +تض ÙħÙĨ +ĠاÙĦÙħØ´ رÙĪØ¹ +Ġprodu ção +ĠпÑĢимен Ñı +ни ÑĨÑĭ +주 ëĬĶ +ر Ùı +Ġm Æ¡ +Ġhayat ı +ëŁ ½ +Ġü cret +Ġyan ında +Ġpr ática +×ij×Ļ×§ ×ķר +Ãľ N +Ñģ оÑĤ +ãĤıãģij ãģ§ +Ġдол го +ת ׼×ķ +ĠìķĦ ëĭĮ +ë į°ìĿ´ +Ġç iz +Ġcho Äĩ +Ġ×Ķ ×Ļת +Ġ×Ķ×Ļת ר +Ġso át +׼ ×ij×ĵ +à¹Ģล à¹Īา +Ġд еÑĢ +ĠдеÑĢ ÐµÐ² +ãĤĴ åħ¥ãĤĮ +×Ĺ ×ķס +×Ĺ×ķס ר +ج ÙĬÙĨ +t ón +onn é +Ġпол ноÑģÑĤÑĮÑİ +人 ãģŁãģ¡ +Ġpr êt +ëł ¸ +Ġdéc embre +cı lar +Ġת ת +Ġê²½ìļ° ìĹIJëĬĶ +ÙĪ Ø¹Ø¯ +è¦ĭ ãĤĭ +วิ à¸Īัย +ë ¶Ī +ز ÙĪØ§ +زÙĪØ§ ج +d ì +ãģ§ãģĻ ãĤĪ +Ġвод о +ĠÙĬ ÙĪØ¬Ø¯ +Ñģ оÑģÑĤоÑı +Ðŀ С +ĠÄIJ ó +׊פש +Ġצ ×Ļ×ij×ķר +ĠاÙĦÙĤ Ø· +ĠاÙĦÙĤØ· اع +Ġиме ÑİÑĤ +Ġph áºŃn +×Ľ×¡ פ×Ļ +полн иÑĤелÑĮ +éĻIJ ãĤĬ +ĠÑģ ÑĢав +ĠÑģÑĢав н +ÙħاÙĦ Ùĥ +×ĵר ×ķ×Ŀ +çļĨ ãģķãĤĵ +ØŃÙĤ ÙĤ +à¹ģหล à¹Īà¸ĩ +ĠاÙĦر سÙħÙĬ +оÑĩ ки +×ĺ ×ij×Ĺ +Ġcan lı +Ġ׾ ׾ +Ġ׾׾ ×ŀ×ķ×ĵ +×ŀ×ij ×ķ +ת ׼ +×ª×Ľ ׳×Ļת +ĠاÙĦÙħ شار +ĠاÙĦÙħشار ÙĥØ© +İ Åŀ +ĠسÙĬ اسÙĬ +в олÑĮ +ĠÑģ пÑĢав +æĿ¥ ãģ¦ +פ×ķר ×ķ×Ŀ +สำ à¹Ģรà¹ĩ +สำà¹Ģรà¹ĩ à¸Ī +ĠÅŁ öyle +Ġzosta ÅĤa +ĠH ü +ר ×ķש +د ÙĦÙĬÙĦ +ÑĢи д +ש ף +×ŀ×§ ×ķר +ĠÑĥ Ñĩ +ĠÑĥÑĩ еб +ĠÑį ÑĤа +ков а +à¸ķà¸Ļ à¹Ģà¸Ńà¸ĩ +ÙĨ ÙIJ +à¸Ńีà¸ģ à¸Ħรัà¹īà¸ĩ +ระ à¸ļุ +Ġd ữ +ĠاÙĦØŃ اÙĦÙĬ +׼ ×ķ׼ +׼×ķ׼ ×ij +Ġ×ŀ×IJ שר +Ġtr ụ +ÑĤел ем +Ġв ли +Ġвли Ñı +Ġש×IJת ×Ŀ +Ġuw ag +Ġuwag ÄĻ +×ĺ ×Ļת +×IJ ×ĵ×Ŀ +à¸Ķ ุ +Ġ×Ķ×IJ ׾×Ķ +Ġkar Ä±ÅŁ +ĠÄIJ á»iji +да ÑİÑĤ +ãģªãģ® ãģ« +Äħ cych +à¹Ģà¸Ļ à¹īà¸Ļ +ãģĹãģ¦ ãģĹãģ¾ãģĨ +int érieur +ĠfÃŃs ica +ĠÐŁ ол +ãģĹãģ ķ +à¸Ĺำ à¹Ħม +ĠL âm +ĠاÙĦÙħ سÙĦÙħ +ĠاÙĦÙħسÙĦÙħ ÙĬÙĨ +ص ØŃØ© +ìĹ Ħ +à¹Ģà¸Ķà¹ĩ à¸Ķ +ĠÑĥ ÑĩеÑĤ +â Ìģ +Ġب ÙĦا +ĠاÙĦاجتÙħاع ÙĬ +פרס ×Ŀ +ãĥķ ãĥ© +ĠÐļ огда +mie ÅĽci +ĠبÙĬÙĨ Ùħا +Ġ×ŀ×IJ ×ŀר×Ļ×Ŀ +Ġ×ij×IJ ×ĸ×ķר +×ķש ×Ļ×Ŀ +ĠÑģдел а +entr ée +à¹Ģ à¸Ħà¹īา +Ñĥг л +ĠاÙĦÙģ ÙĨÙĬ +ĠÐĴ оÑĤ +à¸Ĺีà¹Ī มา +×ķצ ×Ĵ +ÙĤد رة +Ġëª © +Ġ목 ìłģ +íıī ê°Ģ +ĠاÙĦØ£ ربع +ĠاÙĦأربع اء +פס ×Ļ×§ +ĠÑıвлÑı ÑİÑĤÑģÑı +ب ÙĪÙĨ +ì° ¾ +×ŀ×¢ ר׼ +×ŀ×¢×¨×Ľ ×ķת +ãĤ· ãĤ§ +ĠباÙĦ Ø£ +íĸĪ ëįĺ +ĠاÙĦبر ÙĨاÙħج +ĠاÙĦØ£ ØŃد +Ġm Å© +ĠmÅ© i +п аÑĤ +ب Ø« +ĠÑĨ енÑĭ +Ġ×ijת ׾ +è¨Ģ ãĤıãĤĮ +ĠاÙĦÙħ جاÙĦ +ĠìĦ¸ ìĥģ +Ġ×Ĵ ×ķפ +ĠнаÑĪ ÐµÐ¹ +Ġкомп аниÑı +б ин +öl ü +×Ļ ×Ļ×ĺ +Ġ×ŀס פ×Ļ×§ +ยัà¸ĩ à¸Ħà¸ĩ +ĠЧ и +Ġан ÑĤи +ĠÑģÑĢед и +สà¹Īวà¸Ļ à¹ĥหà¸įà¹Ī +оÑĩ ка +íĬ¹ ë³Ħ +ว à¹Īาà¸ĩ +гоÑĢ Ð¾Ð´ +با Ùĥ +à¹Ģส ีà¹Īย +à¹Ģสีà¹Īย à¸ĩ +ãĤĤãĤī ãģĦ +×§ ×ķ×Ŀ +ãģĽ ãģļ +ĠاÙĦÙĤ اÙĩرة +Ġ×ij ׼×ļ +Ùħشار ÙĬع +باØŃ Ø« +Ġпо Ñĩ +ĠпоÑĩ ÑĤи +ĠÑĦоÑĢм а +S İ +Ġ×ŀצ ×Ļ×¢ +ล ื +ลื ม +ĠÑĤ еÑĢ +ĠÑĤеÑĢ ÑĢиÑĤоÑĢ +ĠÑĤеÑĢÑĢиÑĤоÑĢ Ð¸Ð¸ +Ġв меÑģÑĤ +ĠвмеÑģÑĤ е +dıkl arı +op ération +à¹Ĥ ห +ص دÙĬ +صدÙĬ ÙĤ +íĸī ìłķ +تج ا +تجا ÙĪØ² +Ġsu ç +Ġar ty +Ġarty ku +Ġartyku ÅĤ +ãĤ·ãĥ§ ãĥĥãĥĹ +ש פ +שפ ×Ļ×¢ +Ġ×Ķש ×Ļר×ķת +à¹ģà¸ĸ ม +ë¸ Ķ +Ġuk ÅĤad +Ġ×ķ ׼×Ļ +หล าà¸ģ +หลาà¸ģ หลาย +æĸ¹ ãĤĤ +Ġpodr óż +ĠE ÄŁer +Ġком наÑĤ +ĠÑģам ÑĭÑħ +Ġв кÑĥÑģ +б еж +Ġ×ij ×§×ķ +æİĽ ãģij +ãģ¿ ãĤĭãģ¨ +ĠiliÅŁ kin +ĠÙĬ عÙħÙĦ +Ġпод аÑĢ +Ġyaz ılı +ãĤĴ å¾Ĺ +Ġwyst ÄĻp +à¸Ĺีà¹Ī à¹ĥà¸Ĭà¹ī +ØŃاد Ø« +ÙĪ ÙĬد +кÑĥ лÑĮÑĤ +кÑĥлÑĮÑĤ ÑĥÑĢ +à¸ģาร à¹ģà¸Ĥà¹Īà¸ĩ +à¸ģารà¹ģà¸Ĥà¹Īà¸ĩ à¸Ĥ +à¸ģารà¹ģà¸Ĥà¹Īà¸ĩà¸Ĥ ัà¸Ļ +ÙħÙĪ Ø¸ +ÙħÙĪØ¸ Ùģ +ÙĬÙħ ÙĬ +ãĤĵãģ§ãģĻ ãģĮ +diÄŁ im +diÄŁim iz +ĠÐŁ еÑĢ +ĠÐŁÐµÑĢ Ð² +Ġm ão +ĠÑģ ез +ĠÑģез он +Ġ×Ķ×ŀ ×¢ +Ùħ جÙħÙĪØ¹Ø© +ĠинÑĦоÑĢм аÑĨии +i ếc +ã ng +ĠÄij ấy +ãģĶ ç´ +ãģĶç´ ¹ +ãģĶç´¹ ä»ĭ +Ġad ım +à¹Ħ หล +Ġп ÑĢакÑĤи +ĠпÑĢакÑĤи Ñĩ +ĠпÑĢакÑĤиÑĩ еÑģ +ĠпÑĢакÑĤиÑĩеÑģ ки +ĠاÙĦÙĨ Ù쨳 +ĠÑĢабоÑĤ е +ÙĦÙĬ Ùģ +ĠاÙĦجÙĨ ÙĪØ¨ +Ġвод Ñĭ +ì¹ Ļ +Ġм иÑĢа +ĠÄij ừng +ĠпÑĢоÑĤив о +ĠÑģÑĤÑĢан Ñĭ +ล ู +ìĤ ¶ +kre ÅĽl +Ġbul und +Ġbulund uÄŁu +à¹ģ สà¸Ļ +ãĤ± ãĤ¢ +ת×Ĺ ×ķ×ŀ×Ļ +ר׼ ×Ķ +Ġ׾ק ×ķ×Ĺ +Ġ׾ק×ķ×Ĺ ×ķת +Ġ×Ľ×ª ×ķ×ijת +ĠÙĦ ÙĥÙħ +ب شر +Ġr Ãłng +Ġ×ŀ×Ķ ×ŀ +Ġ×IJ×Ĺר ×ķת +Ġб он +Ġбон ÑĥÑģ +ï½ Ĺ +à¹ģ ยà¸ģ +ãģĤãģªãģŁ ãģ® +ĠÑĥÑĩаÑģÑĤ ие +ĠE yl +ĠEyl ül +ĠçalÄ±ÅŁmalar ı +Ø® طر +ìĿ ½ +à¸ģาร à¹ĥà¸Ĭà¹īà¸ĩาà¸Ļ +Ġана лиз +תק ×ij׾ +ни ем +Ġİ ns +Ġİns an +ĠبÙĪ Ø§Ø³ +ĠبÙĪØ§Ø³ طة +Ġ׳ ×Ľ×ł×¡ +Ġ×Ķ×ŀ ×Ļ×ĵ×¢ +Ġç o +Ġço ÄŁu +á» ĺ +ĠêµŃ 민 +ãĤĤ ãģĦãģĦ +Ġ׼ ׾×Ļ +ĠÑģÑĢед не +g ÅĤo +gÅĤo ÅĽ +Ġneg ó +Ġnegó cio +ĠÑĢ ÐµÐ³Ð¸ÑģÑĤ +ĠÑĢегиÑģÑĤ ÑĢа +ĠÑĢегиÑģÑĤÑĢа ÑĨии +Ġtr á»ĵng +ĠпÑĢ Ñı +ĠпÑĢÑı мо +ëłĪ ìĿ´ +Ġk ém +к ле +à¸Ļำ มา +ĠÑĦ ин +ĠÑĦин анÑģ +ĠÑĦинанÑģ ов +Ġki á»ĩm +ยัà¸ĩ à¹Ħ +ยัà¸ĩà¹Ħ à¸ĩ +ย ิà¸ĩ +à¹Ĥ à¸Ľ +ĠполÑĥÑĩ ил +×Ļ×ĸ ×Ŀ +à¹ģละ à¸Ħวาม +Ġво обÑīе +ص ÙĬر +ãĥı ãĥ³ +ĠاÙĦÙĤ اد +ĠاÙĦÙĤاد Ùħ +Ġب دÙĪÙĨ +ع ظÙħ +ת ׳×ķ×¢ +×ª×ł×ķ×¢ ×Ķ +Ø£ ÙħÙĦ +ãģķ ãģĪ +ÑĤ ем +ÑĤем пеÑĢ +ÑĤемпеÑĢ Ð°ÑĤÑĥÑĢ +Ġ׾ ×Ļצ×ķר +Ġr ÄĻk +ر سÙĦ +ìŀIJ 를 +Ġ×Ļצ ×Ļרת +ÙĨ بÙĬ +Ñĩ наÑı +تØŃ ÙĦÙĬÙĦ +Ġм ик +Ġмик ÑĢо +ĠS öz +Ġfor ça +Ñģ он +ĠاÙĦع را +ĠاÙĦعرا ÙĤÙĬ +ĠH á»ĵng +ãģĻãĤĭ ãģŁãĤģãģ« +à¸Ĺีà¹Ī à¸Ńยูà¹Ī +Ġ×ķ×IJ ×£ +ص ÙĬد +ĠìķĬ ê³ł +ร ัà¸ĩ +ĠاÙĦت ÙĪØ§ØµÙĦ +à¹Ģม à¸ķร +Ñĥ ÑģÑĤÑĢой +ÑĥÑģÑĤÑĢой ÑģÑĤв +m ıyor +Ġبا سÙħ +Ġ×ķ ׼×ķ +ĠG ül +á» IJ +Ãī tat +غ اÙĦ +Ø¥ ÙĨØ´ +Ø¥ÙĨØ´ اء +T İ +à¸Ĥà¹īา ม +Ġtro ch +Ġtroch ÄĻ +Ø¥ ص +إص ابة +ĠØ« اÙĨÙĬ +ĠاÙĦص ØŃØ© +Ġ×ĸ×Ķ ×ķ +jÄħ cej +ãĥĢ ãĥ³ +ìĿ¸ ìĿ´ +Ġв олоÑģ +ëIJĺ ë©´ +Ġzak ÅĤad +ãģĻ ãģĵãģ¨ +以ä¸Ĭ ãģ® +Ġ×Ķ×ŀ×§ ×ķ×Ŀ +ÙħØ´ اÙĩ +ÙħشاÙĩ دة +Ñĩ ив +ب Ø´ +ย à¹īาย +Ġsür dür +ĠN ẵ +ĠNẵ ng +ĠигÑĢ Ð°ÑĤÑĮ +Ġê·¸ëŁ¬ ë©´ +ãĥķ ãĥ« +ล à¹Īะ +Ġtend rá +Ġb Ãły +à¹Ģà¸Ľà¹ĩà¸Ļ à¸ľà¸¹à¹ī +Ġok o +Ġoko ÅĤo +w ÅĤa +wÅĤa ÅĽci +wÅĤaÅĽci w +æĢĿ ãĤı +ĠYa ÅŁ +ĠB á»ĩnh +íı Ń +بÙĬ د +קר ף +à¹Ģศ ร +à¹Ģศร ษ +à¹Ģศรษ à¸IJ +à¹Ģศรษà¸IJ à¸ģิà¸Ī +ĠاÙĦØ£ ÙĪØ±ÙĪ +ĠاÙĦØ£ÙĪØ±ÙĪ Ø¨ÙĬ +fl äche +ä¹Ĺ ãĤĬ +Ġb á»ģn +Ùĩ ب +æľĢ ãĤĤ +Ġsa ç +à¸Ńำ à¹Ģà¸ł +à¸Ńำà¹Ģà¸ł à¸Ń +ĠØ£ ج +ĠاÙĦد اخÙĦ +ĠاÙĦداخÙĦ ÙĬØ© +×ĺ ×ķ×ij +ãĤĤ ãģªãģı +Ġли ÑĨа +à¹ģลà¹īว à¸ģà¹ĩ +×ĸ׼ ×Ļר +Ġqu Ãł +ĠÙĥ ذÙĦÙĥ +صØŃ Ùģ +ĠÃĤ u +ÙĪØ¨ ا +à¹Ģà¸Ľà¸¥à¸µà¹Īยà¸Ļ à¹ģà¸Ľà¸¥ +à¹Ģà¸Ľà¸¥à¸µà¹Īยà¸Ļà¹ģà¸Ľà¸¥ à¸ĩ +à¸ķัว à¸Ńยà¹Īาà¸ĩ +Ġráp ida +Ġtas ar +Ġtasar ım +ĠعÙĦÙĬ ÙĩÙħ +ס ×ķ׾ +c ılı +cılı k +Ġر غÙħ +ìĭľ íĤ¤ +Ġ×IJ׾ ×§ +Ġ×IJ׾ק ×ĺר +Ġ×IJ׾ק×ĺר ×ķ׳×Ļ +à¹ģà¸ļ à¹Īà¸ĩ +Ġh ạng +ãģ£ãģ¦ ãģıãĤĮ +ĠÙĨ تÙĬ +ĠÙĨتÙĬ جة +ıkl ı +غ اÙĨ +à¸Ĥà¹īà¸Ń à¸Ħวาม +à¸Ľà¸¥ าย +ĠØ£ Ùħس +à¸Ĺีà¹Ī à¹Ģà¸ģีà¹Īยว +à¸Ĺีà¹Īà¹Ģà¸ģีà¹Īยว à¸Ĥ +à¸Ĺีà¹Īà¹Ģà¸ģีà¹Īยวà¸Ĥ à¹īà¸Ńà¸ĩ +Ġdé fin +Ġdéfin i +ÙģÙĨ اد +ÙģÙĨاد ÙĤ +à¹Ħà¸Ķà¹ī วà¹Īา +ãģªãģĦ ãĤĪãģĨãģ« +Ġpróp ria +ĠPh át +ãĤĦãģĻ ãģı +สวย à¸ĩาม +ê³ł ìļĶ +Ñı еÑĤ +ãģĭãĤĤãģĹãĤĮãģ¾ãģĽãĤĵ ãģĮ +تر جÙħ +ĠкÑĢаÑģ ив +Ġ×ŀ ר×IJש +д еж +ĠÙĬ ÙĪÙĨ +ĠÙĬÙĪÙĨ ÙĬÙĪ +Ñģк оÑĢ +ĠKas ım +ê³Ħ ìķ½ +к оÑģ +Ġна ÑĢÑĥ +ĠнаÑĢÑĥ ÑĪен +Ġdu że +acc ès +Ġh á»ĵng +Ġv Å© +ãģĦãģŁ ãģĹãģ¾ãģĻ +Ġ×ĺ ×Ļ +Ġ×ĺ×Ļ ×ķ׾ +lıkl arı +Ġqu ê +ëħ¸ ëıĻ +ìķ Ķ +CI ÃĵN +Ġt ắc +press ão +ĠìŀĪ ìľ¼ +สิà¸Ĺà¸ĺิ à¹Į +íĥ Ħ +Ġ×Ķ×ŀ ×ŀש׾×Ķ +å¬ī ãģĹãģĦ +ĠÄIJ ặc +ÙĨ زÙĦ +ĠдÑĢÑĥг ой +д ÑĥÑĤ +ìĪ Ļ +Ġth ụ +à¹Ģส ร +à¹Ģสร à¹ĩ +à¹Ģสรà¹ĩ à¸Ī +Ġto plant +Ġtoplant ı +×IJ×ŀ ף +×ķ׾ ת +п омн +Ġyo ÄŁun +ÅĦsk iego +ì° © +ĠØ« ÙĦاث +ĠØ«ÙĦاث Ø© +Ġl ắng +ë¦ ´ +ราà¸Ĭ à¸ģาร +ĠÑģлов а +á» Ĩ +à¸Ķี à¸ģวà¹Īา +ãģĶãģĸ ãģĦãģ¾ãģĻ +Ġд из +Ġдиз айн +fé rence +lıkl ar +ãģªãĤĵ ãģ§ãģĻ +ajÄħ cy +Ġëĭ¤ ìĸij +Ġëĭ¤ìĸij íķľ +×§ ×Ļר +ØŃ ار +ส ูà¹ī +Ġz ro +Ġzro bi +Ġzrobi Äĩ +×ŀ ×Ļ׼×Ķ +à¸Ĭà¹Īวย à¹Ģหลืà¸Ń +ĠÑįÑĤ Ñĥ +ë´ ī +楽 ãģĹãģĦ +س ÙĪØ± +íķĺ ê±°ëĤĺ +Ùħؤ تÙħر +Ġpoc zÄħ +ĠpoczÄħ tk +ĠpoczÄħtk u +Ġع ربÙĬ +اÙĦØ£ ر +اÙĦأر دÙĨ +à¸Ķ ร +Åĵ uvre +ĠÙĪÙĥ اÙĨت +ĠÅĽ redni +Ø® ضر +Ġch uyến +н ÑĤ +ĠìķĮ ê³ł +Ġv á»Ŀi +Ġ×ij ×Ļ×ĵ×Ļ +×ŀ×ĵ ×ķ×ijר +ÙĪ Ù쨱 +ÙĬ Ø¡ +׳ ×Ľ×¡ +ĠÐĽ а +л он +Ġx ấu +Ùģ ÙĬÙĨ +Ġfé vrier +ĠÐŀ на +ĠV á»ģ +ĠÅŁey ler +ĠполÑĥÑĩ ен +з ад +Ġn ét +à¹Ħà¸Ľ ยัà¸ĩ +×Ĺש×ij ×ķ +à¸ļัà¸Ļ à¸Ĺ +à¸ļัà¸Ļà¸Ĺ ึà¸ģ +Ġgerçek leÅŁ +иÑĩеÑģк ое +ìĪĺ ê°Ģ +Ø« بت +ãģ¤ ãģ¾ãĤĬ +ĠÑĥÑģловиÑı Ñħ +ëĭ¤ ê°Ģ +ราย à¹Ħà¸Ķà¹ī +׼×IJ ×ij +à¹Ĥà¸Ľà¸£ à¹Ĥม +à¹Ĥà¸Ľà¸£à¹Ĥม à¸Ĭัà¹Īà¸Ļ +j ähr +jähr ige +×§ ׳×Ļ×Ŀ +×ŀ ×ķ×§ +×ŀ×ķ×§ ×ĵ +ãģ«è¡Į ãģ£ãģ¦ +Ø¢ ÙĦ +вед ение +Ġ׾ ×Ľ×ª×ķ×ij +جÙħ Ùĩ +جÙħÙĩ ÙĪØ±ÙĬØ© +à¸ī à¸ļ +à¸īà¸ļ ัà¸ļ +ĠC òn +à¸ľ สม +ãģªãģ© ãģĮ +×IJ×Ķ ×ij +ĠдейÑģÑĤв иÑı +y ız +à¹Ħมà¹Ī à¹Ģà¸Ħย +ج ÙĪØ² +×Ķ×Ĺ׾×ĺ ×Ķ +f ällt +ãĥĵ ãĤ¸ +ãĥĵãĤ¸ ãĥį +ãĥĵãĤ¸ãĥį ãĤ¹ +Ġ×IJ ×Ļ׳×Ŀ +ĠнаÑħод иÑĤÑģÑı +Ġdzi ÅĽ +ست Ø·ÙĬع +׾ ×Ļף +Ø® ÙĦاÙģ +Ùĩ ÙIJ +Ġatr ás +íĺ ģ +ãĤĴ ãģĶ +Ġ×Ķ×ŀ ×ķצר +ĠBakan lıģı +ÑİÑī ее +ÙħÙĨ اط +ÙħÙĨاط ÙĤ +Ùģ Ø¯ +à¸Ļำ à¹Ħà¸Ľ +Ġв аж +Ġваж но +Ġm ạch +׼ ׳×ķ +بع Ø« +lan ması +Ġa yr +Ġayr ıl +ìĤ¬ íļĮ +d ÃŃa +p ÅĤyw +اÙħ ÙĬØ© +íĺ ľ +×IJ׳ ×Ĵ׾ +×IJ׳×Ĵ׾ ×Ļת +ĠìŀĪëĭ¤ ëĬĶ +Ġس اعة +ĠëĤĺ íĥĢ +b ö +à¸Ħ ัà¸Ļ +ĠdziaÅĤ ania +Ø© Ùĭ +Ġng Å© +׳צ ×Ĺ +ãģ¯ ãģĤãĤĭ +ĠyaÅŁ ında +st ück +car acter +caracter ÃŃsticas +Ġr á»Ńa +ĠÙħختÙĦÙģ Ø© +ãģ«ãģĬ ãģijãĤĭ +à¹ģà¸ŀ à¸ĩ +วิ à¹Īà¸ĩ +ת פ×ķ +سا ÙĩÙħ +使 ãģĨ +Ùĥ رÙĬ +×IJ פ×Ļ +........ ....... +ĠÑĤак им +×Ļ׼ ×ķ×Ļ +Ø´ بÙĩ +ج ÙĬر +ãģĿãģ® ãģ¾ãģ¾ +ac jÄĻ +ĠاÙĦت رÙĥ +ĠاÙĦترÙĥ ÙĬ +ĠпÑĢав илÑĮно +Ġت عÙħÙĦ +à¸ģล à¹īา +Ġbi ên +Ġ×ij׳×Ļ ×Ļת +Ġкл Ñĥб +Ġ×ŀ ש×Ķ +в ÑĪий +ãģĵãģ¨ãģĮãģ§ãģį ãĤĭ +à¸ŀัà¸Ļà¸ĺ ุ +à¸ŀัà¸Ļà¸ĺุ à¹Į +ר ×ķ×Ŀ +ĠاÙĦÙģ Ø±ÙĨ +ĠاÙĦÙ쨱ÙĨ سÙĬ +à¹Ģà¸Ľà¹ĩà¸Ļ à¸Ħà¸Ļ +ãģĹãģ¦ ãģĬãĤĬ +Ġth ầy +ãĤĵ ãģłãģijãģ© +ìĶ ¨ +Ùħ دÙĨ +ت ÙĪÙĨ +ĠмеÑĤ ал +ĠмеÑĤал л +Ġin ÃŃcio +à¸Ńà¸Ńà¸ģ à¸Īาà¸ģ +ëĴ ¤ +Ġcu á»ijn +Ġbu á»Ļc +ÙĨ سÙĬ +ä cht +×ŀ ×Ļ׳×Ļ×Ŀ +ãģķ ãģ¦ +ãģĮ ãģ§ãģį +ÑĬ ем +Ġtá i +ĠЧ ÑĤ +ĠЧÑĤ обÑĭ +à¸Ľà¸¥ ูà¸ģ +à¸Ĭุม à¸Ĭà¸Ļ +н Ñģкий +Ġv ững +Ġ×Ķ ×ľ×ij +ë le +Ġש ×¢×ijר +в аÑĤÑĮÑģÑı +б ой +ع ÙĪÙĨ +à¹ģà¸Ķ à¸Ļ +Ġספר ×Ļ×Ŀ +Ġt uyên +Ġnhi êu +ĠQu ý +Ġh uyết +ãĤı ãģĭãĤīãģªãģĦ +Ġ×ŀ ׼ף +Ġ×Ķ ×§×ľ +Ġ׾×IJ ×ķר +ĠÄIJi á»ĩn +Ø´ ؤ +شؤ ÙĪÙĨ +Ġ×ŀ׊פש +ĠпоÑģÑĤоÑıн но +×ŀ ×Ļר +ìħ Ķ +Ðŀ Ñģ +ÐŀÑģ нов +×ĸ ×Ļת +ĠH á +ĠÑĩаÑģ ов +×IJ ×ķ׾×Ļ +Ġm át +Ø® رÙĪ +خرÙĪ Ø¬ +ÙĤ ضا +ÙĤضا ÙĬا +à¹Ģà¸Ľ à¸Ńรà¹Į +ĠÙĬ ÙĪÙĦ +ĠÙĬÙĪÙĦ ÙĬÙĪ +à¹Ĥà¸Ĺ ษ +׳ פ׾ +ת ×ķש +ת×ķש ×ij×Ļ +Ġv ários +×ŀ ר×IJ×Ķ +ëĿ¼ ìĿ´ +ÙĨ غ +×ij צע +г он +ĠÄIJ ược +ع Ùı +пÑĥÑģ к +ĠÙĪØ§ÙĦ Ùģ +üc ü +×Ļ×§ ×Ļ×Ŀ +Ġس بÙĬÙĦ +׾×ij ף +ĠاÙĦÙĤ رÙĨ +ס ×ķת +ĠQu áºŃn +ãģĵãĤĮ ãģĮ +ãĥĸ ãĥ©ãĥ³ãĥī +×Ĵ ×ŀר +Ġwarto ÅĽci +ĠÙĪØ¨ ÙĬÙĨ +Ġd ạ +ÐIJ в +ÐIJв ÑĤо +Ġol acaktır +à¸Ļ à¸Ĺà¹Į +Ùħ طار +Ġ×¢ ×§×ij +Ġת פ +ãģĹãģ¦ ãģĦãģ¦ +צ ×ŀ×Ĺ +à¸Ī à¸Ńà¸ĩ +Ġö de +ìį ¨ +ÙĨ اس +調 ãģ¹ +ĠогÑĢ Ð¾Ð¼Ð½ +ë³´ íĹĺ +×ĺ ×§ +×ĺ×§ ס×ĺ +ĠbaÅŁ v +ĠbaÅŁv uru +Ġpom ys +Ġpomys ÅĤ +ãģ« ä¹Ĺ +Ġש ׼ף +ĠاÙĦÙħس ؤÙĪÙĦ +Ġз ан +Ġзан ÑıÑĤ +Ġd ương +ãĥĹãĥ¬ ãĤ¤ +ล à¸ļ +ÑĤи ка +ĠAr alık +Ġнед о +Ġm á»Ļ +Ġor an +Ġoran ı +Ġktó r +Ġktór Äħ +Ġ×Ķ×IJ×Ĺר ×ķ׳×ķת +ائ ÙĨ +ÅĦ s +ÅĦs ka +åĽ½ ãģ® +×ŀ ×ĺ×Ļ +ĠвопÑĢоÑģ Ñĭ +à¸Ńà¸ĩà¸Ħà¹Į à¸ģร +×ŀ ×ķצ×IJ +Ġpó ź +Ġpóź niej +ש×ŀ ×IJ׾ +Ġk aps +Ġkaps am +Ġkapsam ında +Ġmá quina +ĠÅĽwie cie +Ġho Ãłng +Ġöz gü +×Ĵ×ķר ×Ŀ +ãģĤ ãģŁãĤĬ +à¸ķัà¸Ķ สิà¸Ļ +à¸ķัà¸Ķสิà¸Ļ à¹ĥà¸Ī +б ÑĢи +ãģ«ãģªãĤĭ ãģ¨ +ت ÙĥÙĪÙĨ +Ġ×ķ×Ķ ×Ļ×IJ +Ġchi ếu +ÑģÑĤан ав +ÑģÑĤанав ли +ÑģÑĤанавли ва +×ŀ ×ķ×Ĵ +c ité +ĠK örper +Ġש ×Ĵ×Ŀ +ع ظ +عظ ÙĬÙħ +Ġ×Ķ×IJ ×Ļש×Ļ +Ġmat ière +ĠÙģ ÙĪÙĤ +Ġk to +Ġkto ÅĽ +à¸Ļ à¹Ĥย +à¸Ļà¹Ĥย à¸ļาย +å¾ħ ãģ¡ +à¹Ģม à¸Ļ +à¹Ģมà¸Ļ ู +A ÃĩÃĥO +Ġt ù +Ġtù y +ãĥĪ ãĥ³ +ĠоÑĤ каз +Ġ×ŀ ×ķצר +ül ü +ãģķãĤĵ ãģ« +Ġ×Ĺ ×ķ×ij +קר ×Ļ×IJ×Ķ +ĠاÙĦØ® دÙħات +ĠÙĦÙħ دة +ر ؤ +رؤ ÙĬØ© +ãĤĴè¦ĭ ãģ¤ãģij +à¸Ł า +Ġréuss i +à¸Ļัà¸ģ à¹Ģรียà¸Ļ +ĠÑĩиÑģ л +à¸ģาร à¹Ģลà¹Īà¸Ļ +Ġhaz ırl +Ġhazırl an +ĠпеÑĢв Ñĭй +ли м +ĠоÑĤзÑĭв Ñĭ +Ġwy jÄħ +ĠwyjÄħ tk +ĠØ£ ÙĤÙĦ +ס ×ļ +Ġê²° ìłķ +Ġ׾×ŀ×¢ ש×Ķ +Ġl ắp +à¹ģà¸ļ ร +à¹ģà¸ļร à¸Ļà¸Ķà¹Į +วà¹Īา à¹Ģà¸Ľà¹ĩà¸Ļ +Ġب دا +Ġبدا ÙĬØ© +ãģ¨ãģĦãģĨ ãģ®ãģĮ +иÑĩеÑģк им +à¸ģาร à¸ŀัà¸Ĵà¸Ļา +Ġb Ãło +Ġmia ÅĤa +y waÄĩ +ĠMär z +ĠÙĨ سبة +Ġéconom ique +×ĸ ×ŀ +×ĸ×ŀ ׳×Ļ×Ŀ +æŃ¢ ãĤģ +Ġt á»§ +íķĺ ìĭł +Ġkażde go +stra ÃŁe +à¸Ĭ ีà¹ī +à¹Ģ à¸ļา +ÑĢеÑģ ÑĥÑĢÑģ +ев ой +Ø´ باب +à¸ķà¹Īาà¸ĩ à¸Ľà¸£à¸°à¹Ģà¸Ĺศ +Ġ×IJ ×Ļש +Ġ×IJ×Ļש ×Ļת +×Ļ ×ķפ +×Ļ×ķפ ×Ļ +ĠìļĶ êµ¬ +ì¡° ìĤ¬ +ãģ£ãģŁ ãĤī +׾ ×Ļ×§ +миниÑģÑĤ ÑĢ +ãĤĤãģ® ãģ¯ +Ġl ương +Ġна и +Ġнаи бол +Ġнаибол ее +íİ ĺ +à¹ģà¸ŀ à¹ī +ãĤŃ ãĥ¥ +ĠкоÑĤоÑĢ Ñĭм +à¹ģà¸Ĺ à¸ĩ +à¹ģà¸Ĺà¸ĩ à¸ļà¸Ńล +Ġ׳ ×Ļ×Ķ +Ġ׳×Ļ×Ķ ×ķ׾ +âĤ ª +ĠGi ải +ĠиÑģполÑĮзов а +ëł¥ ìĿĦ +ãģĹãģĭ ãĤĤ +à¸ģà¹ĩ à¸ķà¹īà¸Ńà¸ĩ +ĠÑĢ ÐµÐ± +ĠÑĢеб ен +ĠÑĢебен ка +ت ÙĪØ§ØµÙĦ +ãĤ°ãĥ« ãĥ¼ãĥĹ +ãĤĦ ãĤī +à¹Ģà¸Ľà¸´à¸Ķ à¸ķัว +б ÑĢо +ë°ĸ ìĹIJ +ÙĨ ÙİØ§ +×Ķ ×Ĵ +×Ķ×Ĵ ׳×Ķ +à¸Ĺ รั +à¸Ĺรั à¸ŀ +à¸Ĺรัà¸ŀ ยà¹Į +Ġkh á»iji +עצ ×ŀ×ķ +бол езн +Ġë°Ľ ìķĦ +ม à¸Ļ +มà¸Ļ ุ +มà¸Ļุ ษ +มà¸Ļุษ ยà¹Į +âĹ Ĩ +×ŀ צ׾×Ļ×Ĺ +Ñıв ление +Ùħ Ø·ÙĦ +ÙħØ·ÙĦ ÙĪØ¨ +Ø® اÙĦÙģ +ت ÙĪÙĤÙģ +ãģ§ãģį ãģ¾ãģĽãĤĵ +оÑģÑĤ ей +м еÑĩа +기 ëĬĶ +תש ×¢ +ص ÙĬب +Ġ×ij×¢ ×ķ×ĵ +à¸Ĥà¸Ńà¸ĩ à¹Ģà¸Ĥา +ÑĤÑı ж +ĠÑĥ пÑĢав +ĠÑĥпÑĢав лениÑı +Ġgén ér +Ġth ÃŃ +פ ×ļ +Ġر Ùħض +ĠرÙħض اÙĨ +Ġtr uyá»ĩn +Ø¥ عداد +ãĤµ ãĥĿãĥ¼ãĥĪ +Ġпол но +Ø® اÙħ +ÐŁ еÑĤ +ÐŁÐµÑĤ еÑĢ +ÐŁÐµÑĤеÑĢ Ð±ÑĥÑĢ +ÐŁÐµÑĤеÑĢбÑĥÑĢ Ð³ +ÙħÙĨت دÙī +ãģķãĤĮ ãģ¾ãģĹãģŁ +ĠëĮĢ íķĺìŬ +à¸ľà¸¹à¹ī à¸Ĺีà¹Ī +Ġ×ŀ×IJ ×ķ +׾ ׳×ĵ +оÑĩ нÑĭе +ĠнаÑĩ ала +Ġ׾ ×Ļ׾×ĵ×Ļ×Ŀ +ов ое +ãģĻãĤĭãģĵãģ¨ ãģ§ +ĠاÙĦÙĨ Ùģ +ĠاÙĦÙĨÙģ Ø· +ìŀĪ ëĬĶ +غ ÙĨÙĬ +פ ×ĵ +ãĤ ¾ +ĠCr é +ãģ© ãģ¡ãĤī +Ø« اÙĨ +ÑĢаб аÑĤ +ÑĢабаÑĤ Ñĭва +Ġê°Ļ ëĭ¤ +à¸Ī ั +à¸Īั à¸ģร +Ġch ụ +Ġchụ p +Ġм аÑģÑĤ +ĠмаÑģÑĤ еÑĢ +Ġn ắm +ĠÑģÑĤ али +Ġ×Ķ×IJ ×Ļר×ķ×¢ +ãĤ½ ãĥ³ +åĪĨ ãģĭãĤĬ +Ø· بع +بد ا +gr áfico +г еÑĢ +à¸Ķำà¹Ģà¸Ļิà¸Ļ à¸ģาร +Ġsal dır +Ġsaldır ı +в ÑĪиÑħ +ãģĭãģ£ãģŁ ãģ§ãģĻ +Ġyapı yor +ĠاÙĦÙģ Øª +צר פת +з доÑĢов +×ij×¢ ׾ +Ġ×IJ ×ŀ×Ļת×Ļ +Ġоб Ñĭ +ĠобÑĭ Ñĩ +ĠобÑĭÑĩ но +Ġ׾ ×ķ×ŀר +ت ÙĥÙĨ +تÙĥÙĨ ÙĪÙĦÙĪØ¬ +تÙĥÙĨÙĪÙĦÙĪØ¬ ÙĬا +Ġhakk ı +ĠÑĢаР² +ĠÑĢав но +رÙĬ Ùĥ +Ġ×ij ×ŀ×Ļ×ĵ +Ġ×ij×ŀ×Ļ×ĵ ×Ķ +à¹ģà¸ģ à¹īว +Ġìĸ ĺ +Ġìĸĺ 기 +ãģĹãģ¦ ãģĦãģ¾ãģĹãģŁ +Ġkı sm +Ġkısm ı +ê± ¸ +åĨħ ãģ® +ì§ ķ +à¹Ģหมืà¸Ńà¸Ļ à¸ģัà¸Ļ +ĠÙģ ÙIJ +ĠÙģÙIJ ÙĬ +ÙĤ اعدة +Ġmoż esz +Ùħ صاÙĦ +ÙħصاÙĦ ØŃ +ãģ¾ãģŁ ãģ¯ +б ег +Ġs ıc +Ġsıc ak +Ñĩ иÑģ +ÑĩиÑģ лен +Ġн ог +ãĥģãĥ£ ãĥ³ +ãĥ« ãĥī +Ġgi ó +Ġs ını +Ġsını f +ив аÑĤÑĮ +Ġqu ên +Ġì łģ +Ġìłģ ìļ© +ĠJo ão +Ùģ Ø§Ø¯ +ĠGl ück +à¸Ĺ à¸Ńà¸Ķ +Ġg ói +ï¼ Ĭ +Ġdé tail +ĠدÙĬ سÙħ +ĠدÙĬسÙħ بر +ë¡ľ ìĦľ +×ŀ ×ķ×Ĺ +à¹Ħ ฮ +ĠоÑĤ д +ĠоÑĤд ÑĭÑħ +Ġkh uyến +à¸Ħ à¸Ńย +Ġج ÙĨÙĬ +ĠجÙĨÙĬ Ùĩ +ĠاÙĦد ÙģØ§Ø¹ +à¸Ļà¹īำ หà¸Ļัà¸ģ +ĠìĤ¬ëŀĮ ëĵ¤ìĿ´ +Ġth ừa +ĠÃ¶ÄŁrenc i +ĠпомоÑī и +ĠczÄĻ ÅĽÄĩ +ש ×ĺר +ĠN hi +ĠNhi á»ģu +׳ צ×Ļ +ĠнаÑĪ ÐµÐ¼ +ĠkarÅŁÄ± laÅŁ +Ġ×Ķש ׳×Ļ×Ŀ +ĠÄIJ ưá»Ŀng +Ġtr ú +ĠÑĢазлиÑĩ нÑĭÑħ +ĠاÙĦØ´ Ùĩر +Ġ×ľ×¢ ×ķ׾×Ŀ +ØŃ جر +ĠÄij á»ķ +ĠìĿĺ íķ´ +à¸ļ à¹Īà¸Ńย +Ġ×Ķ ×Ļ׾×ĵ +ãģ¨ãģª ãģ£ãģŁ +Ġ×Ĺ×ķ ×ķת +Ġש×Ļר×ķת ×Ļ +Äħ cy +س رÙĬ +K İ +פ ׳×ķ +ÑģÑĤÑĢÑĥк ÑĤÑĥÑĢ +ÑĤ ÑĢÑĥд +Ġ×Ķ ×§×¨ +Ġ×Ķקר ×ķ×ij +Ġth áºŃm +èģŀ ãģį +ÙĤÙĪ ÙĬ +клÑİÑĩ ен +ÑĤе Ñħ +ÑĤеÑħ нолог +è¡Į ãģ£ãģŁ +Ġ×ķ×IJ ×Ļף +ĠÅŁek lin +ĠÅŁeklin de +r ô +ÑĢ Ð¾Ð³ +Ġнов Ñĭе +Ġס ×ij×Ļ×ij +Ġtecn ologÃŃa +ס ׼ +×¡×Ľ ×ķ×Ŀ +ĠÅŀ ub +ĠÅŀub at +Ġ×Ķ×ŀ ׾×IJ +Ġwy pos +Ġwypos aż +ãģ¯ ä½ķ +ãĤ¬ ãĥ³ +ê° ĸ +Ġкак ие +Ġçocuk lar +Ġ׾צ ×ĵ +Ġkay ıt +ĠмеÑģÑĤ е +Ùħ دÙĬÙĨØ© +Ġ׼ ×Ĵ +Ġ׼×Ĵ ×ķף +ãģĹãģ¦ ãĤĭ +ĠÙħا ÙĬÙĪ +ãģ£ãģ¦ãģĹãģ¾ ãģ£ãģŁ +ĠпÑĢогÑĢамм Ñĭ +à¹ģล à¸Ļà¸Ķà¹Į +ãĥ¯ ãĤ¤ +ער ×ķ×¥ +Ñģ ид +ĠB öyle +Ġì²ĺ ìĿĮ +Ġת פק×Ļ×ĵ +ĠTr ên +íĥ Ī +ĠÐłÐ¾ÑģÑģ ий +ĠÐłÐ¾ÑģÑģий Ñģкой +Ġs Ãłn +Ġrè gle +ĠyaklaÅŁ ık +à¹Ģล ิà¸ģ +Ġد ائÙħ +Ġ×ķ ×Ĵ +اب ر +Ġb è +ĠاÙĦ ÙĤدÙħ +ĠÑĢеÑĪ ÐµÐ½Ð¸Ñı +hi ên +ÑĤи к +Ä Ħ +à¸ļรร ยาà¸ģ +à¸ļรรยาà¸ģ าศ +רצ ×ķף +åĭķ ãģį +ĠGä ste +Ġ기 본 +ĠÙĬ عرÙģ +ĠS á»Ń +gÅĤ ÄĻb +à¹Ģà¸Ń ส +×IJ×ŀ ×Ļף +Ġп Ñĥнк +ĠпÑĥнк ÑĤ +Ġ×Ļ×ķ×ĵ ×¢×Ļ×Ŀ +ãĤ« ãĥ©ãĥ¼ +Ġ×ijס ×ĵר +Ġbu á»ĵn +й ÑĤ +йÑĤ еÑģÑĮ +ãĤĴ æ±ĤãĤģ +Ġ×IJת ׼×Ŀ +Ġ모 르 +ظ رÙĪÙģ +Ñĩ еÑģÑĤво +ìĸ´ ìĦľ +Ġод на +Ġkap ı +Ġëħ¸ ëł¥ +ĠKü che +ĠاÙĦت Ø´ +Ø· ÙĬب +ĠíĬ¹ íŀĪ +ĠвÑĭп ÑĥÑģ +ĠвÑĭпÑĥÑģ к +×ĵ ת×Ļ +Ġu ÄŁ +ĠuÄŁ ra +ائ Ùĩا +Ġtho át +ãģª ãĤĤãģ® +Ñij ÑĢ +기 ê°Ģ +ĠgeliÅŁ me +تØŃ ÙĤ +تØŃÙĤ ÙĤ +Ġоп аÑģ +б ÑĢоÑģ +ห ุ +หุ à¹īà¸Ļ +ì¼ Ģ +ãĤ¹ ãĥŀ +ãĤ¹ãĥŀ ãĥĽ +Ø£ Ù쨱 +Ø£Ù쨱 اد +ĠTh á»±c +Ġth ắ +ãĥªãĥ³ ãĤ¯ +Ġni á»ģm +ĠHö he +عÙħ ار +ÙĥÙĪØ± ÙĪÙĨ +ÙĥÙĪØ±ÙĪÙĨ ا +ĠÄIJ ến +ĠÑģам ом +ĠÑĤ еле +ĠÄijo án +à¸Ħวามà¸Ħิà¸Ķ à¹Ģหà¹ĩà¸Ļ +Ġд иÑģк +Ø£ Ø·Ù쨧ÙĦ +ม ารà¹Į +à¸Ĺ หาร +à¸Ĺ à¸Ļ +Ġب عÙĬد +ĠاÙĦÙĩ ÙĨد +åĩº ãģĹãģ¦ +Ġkar de +Ġkarde ÅŁ +×Ķ×Ļס×ĺ ×ķר +×Ķ×Ļס×ĺ×ķר ×Ļ×Ķ +éģ¸ ãģ³ +ع اÙħÙĦ +à¸Ĥ ยาย +Ġtü rl +Ġtürl ü +ĠìĿ¼ ìĿ´ +Ġmaté ria +Ġ׼׾ ×ķ×ŀר +ãĥģãĥ£ ãĥ¼ +جÙħ اعة +ĠÑģво им +Ø¥ÙĤ اÙħØ© +ä¾ĭ ãģĪãģ° +س اب +Ø¢ خر +ÙĤ دÙĬر +×IJ×ŀ ×Ļ +ìĸ » +Ġ׳×ķס פת +ĠÐĴ лад +ĠÐĴлад им +ĠÐĴладим иÑĢ +Ġest ará +ãģĵãģĨ ãģĦãģĨ +ãĤĴ 使ç͍ +มา à¸ķร +มาà¸ķร à¸IJาà¸Ļ +ãģ£ãģ ½ +Ġn ú +Ġnú i +ย าà¸ĩ +ĠاÙĦج ÙĨس +Ġüst ün +ëľ » +ãĤ» ãĥ« +ãģ¦ãģĦ ãģįãģ¾ãģĻ +Ġ×Ĺ ×ķ×ĸ +Ġ×Ĺ×ķ×ĸ ר +ĠÐĵ лав +à¹Ĥà¸Ĭ à¸Ħ +íı IJ +ÙĨت ظر +Ġ×Ĵ ×ij×Ļ +ع ÙĤب +int ér +intér êt +×ŀ פ×Ĵ +×ŀפ×Ĵ ש +Ġth ù +اÙģ Øª +Ġ×ŀש פ +Ġ×ŀשפ ×ĺ×Ļ +ĠÙħ ÙĪØ§ÙĤع +è¦ ļ +è¦ļ ãģĪ +×ĵ ×Ļף +à¹Ģรืà¹Īà¸Ńà¸ĩ ราว +ãģ¾ ãģĤ +Ġgh ế +иÑĢÑĥ ÑİÑĤ +à¸ģ ว +à¸ģว à¹īาà¸ĩ +Ġпов еÑĢ +ĠповеÑĢ Ñħ +ĠповеÑĢÑħ ноÑģÑĤ +׳ ×ĵר +Ġкон ÑĨе +Ġдолж на +Ġ×Ļש ×Ļר +acaģı z +ìĹ Ķ +Ġn ÃŃvel +Ġö r +Ġör nek +Ùĥ Ùģ +ĠФедеÑĢ Ð°ÑĨии +Ġ구 ìĦ± +หัว à¹ĥà¸Ī +ĠV áºŃy +м ед +мед и +меди ÑĨин +медиÑĨин Ñģк +از ÙĬ +×Ĵ×ij ×ķ׾ +ÑĦ ÑĢ +Ġzus ätzlich +à¸ģ à¸ģ +ĠاÙĦاÙĤتصاد ÙĬØ© +Ġh è +lu ÄŁun +ج Ùİ +à¹Ħà¸Ł ลà¹Į +ÄIJ T +ãģĿãģ® ä»ĸ +à¸Ĺิ à¹īà¸ĩ +ĠاÙĦØ£ ÙĪ +ر سÙħ +æ°Ĺ ãģ¥ +ìĿ´ ë©° +ÑĮ ев +ص Ø· +ĠاÙĦاست Ø« +ĠاÙĦاستث Ùħار +à¸Ńา à¸Ħาร +ĠÑĤоÑĩ но +ĠV ân +à¸Ń ร +à¸Ńร à¹Īà¸Ńย +ĠاÙĦس ÙĨØ© +Ġc Æ°á»Ľi +×Ļ×Ķ ×Ł +íį ¼ +話 ãģĹ +âĹ ĭ +ĠìķĬ ìĿĢ +ãĥ¡ ãĥ¼ãĤ +ãĥ¡ãĥ¼ãĤ « +ãĥ¡ãĥ¼ãĤ« ãĥ¼ +ĠÑĤеп ло +å½¼ ãĤī +Ġİ z +Ġİz mir +íĻ į +Ġr ượ +Ġrượ u +æĢĿãģĦ åĩº +ĠPh ạm +Ġchá u +צ×Ļ ×ķת +ĠìĿ¼ 본 +ìĤ¬ ëĬĶ +ĠÑģозд ан +Ġar acı +Ġ×¢ ר +Ġער ×Ļ׼×Ķ +ĠíķĺëĤĺëĭĺ ìĿĺ +dzi ÅĤ +à¸Ľà¸£à¸° à¸ĺาà¸Ļ +Ġser ÃŃa +ĠìŀĪ ëıĦë¡Ŀ +در ج +íķľëĭ¤ ëĬĶ +à¸Ńา à¸Ĺ +à¸Ńาà¸Ĺ ิà¸ķ +à¸Ńาà¸Ĺิà¸ķ ยà¹Į +ÑĤелÑĮ нÑĭй +ĠØ® دÙħات +×ŀ׳ ×ĺ +Ġl ược +ĠS Ãłi +ĠÙĪ Ø§Ø¶ +ĠÙĪØ§Ø¶ ØŃ +غ از +ĠdoÄŁ al +Ġ×ijש ×Ŀ +Ġд лин +ĠØ¥ طار +Ġ×ijס פר +ãĤĴ ä¸İ +ãĤĴä¸İ ãģĪ +Ġë²ķ ë¥ł +ĠÑĥ вели +ĠÑĥвели Ñĩи +ส à¹Ħà¸ķ +สà¹Ħà¸ķ ลà¹Į +à¹Ħ à¸ģล +×ij׊ף +ĠìĿ´ íĽĦ +Ġm unic +Ġmunic ÃŃpio +تÙħ Ø«ÙĦ +ĠÄij áo +H ôtel +Ġl á»Ńa +ĠÄij ẳng +Ñĩ ки +Ø´ رÙĪ +شرÙĪ Ø· +ĠìĿ´ 를 +ÙĬ Ùĭا +×ŀ׾ ×ļ +×ŀ×Ķ ×Ļר×ķת +ĠобÑıз аÑĤелÑĮ +ĠобÑıзаÑĤелÑĮ но +é nergie +Ġmud ança +Ġm ụ +Ġmụ n +Ġn º +ĠاÙĦت عا +ĠاÙĦتعا ÙĪÙĨ +ĠاÙĦاجتÙħاع ÙĬØ© +Ġп лаÑģÑĤ +Ġëĵ± ìĿĺ +ãĥIJãĤ¤ ãĤ¯ +Ùĩج ÙĪÙħ +ĠSa úde +Ġì¤ijìļĶ íķľ +Ġ×Ķצ ×Ļ×ij×ķר +תק ף +ĠاÙĦعاÙĦÙħ ÙĬ +ĠболÑĮÑĪ Ð¾Ð¹ +ĠÙĥ ÙĦÙħ +ĠÙĥÙĦÙħ Ø© +ãģ®ãģ§ãģ¯ãģªãģĦ ãģ§ãģĹãĤĩãģĨãģĭ +ĠÙħ باراة +Ġש×IJ ׳ +Ġש×IJ׳ ×Ĺ׳×ķ +ãĤ¹ãĤ¿ ãĤ¤ãĥ« +ĠSa ÄŁ +ĠSaÄŁ lık +Ġh ư +׳ ×Ĺ×Ķ +Ġ×ij קר×ij +Ø· عÙħ +ห ิà¸Ļ +à¸Ĺุà¸ģ วัà¸Ļ +à¸Ħรัà¹īà¸ĩ à¸Ĺีà¹Ī +ĠlÃł nh +Ġdonn é +ãģĽ ãģĦ +جز ÙĬرة +доÑĢ Ð¾Ð¶ +ì¼ ľ +تÙĨظ ÙĬÙģ +ãĥģ ãĥ§ +Ġald ıģı +ج اج +ĠÑĤ омÑĥ +à¸Ľ ิ +Ġ×ijר שת +ãģıãģªãĤĬ ãģ¾ãģĻ +ĠпÑĢин ÑĨип +Ġ׊׾×ķ +ëı ¼ +×ķ×Ĵ ש +س س +à¸Ľ ู +Ġh ầu +æĦŁãģĺ ãĤĭ +ï¼ ´ +د ÙĪØ§ +ĠÑģм ог +scri ção +Ġth áºŃn +Ġר ×ķ×IJ×Ķ +обÑĢаж ен +ĠاÙĦتج ارÙĬØ© +Ø· بÙĬع +jÄħc Äħ +íĸī ìľĦ +Ġнов Ñĭй +Ġ×ŀ ×Ĺ×ĵש +æĮ¯ ãĤĬ +gu é +Ġ×IJ ×Ļר×ķ×¢ +Ġ×IJ×Ļר×ķ×¢ ×Ļ×Ŀ +ĠاÙĦ ذÙĩب +×ĵ ×IJ +ت اÙĨ +ãģł ãģĹ +à¸Ńั à¸ķรา +à¹Ĥ à¸Ī +بÙĦ اد +×Ķ×Ļ ×Ļ׳×ķ +ĠÑģп е +ĠÑģпе ÑĨиалÑĮно +ĠÅĽwi ata +ãĤĵãģ§ãģĻ ãĤĪ +شر ÙĥØ© +ĠpÅĤ yt +Ġsitu é +Ġ׼×IJ ׾×Ķ +ס ×ijר +Ġkaż d +Ġkażd ym +ãĤĴæĮģ ãģ¤ +׾×Ķ ×ľ +׾×Ķ׾ ף +ĠwÅĤ as +ĠwÅĤas ne +ĠsaÄŁ lan +×ŀ×¢ ׾×Ķ +ĠاÙĦا ÙĪÙĦ +ìĹIJìĦľ ëıĦ +×IJ×Ļר ×ķפ×Ķ +تÙĤ ÙĨÙĬØ© +Ùħ ائ +Ùħائ Ø© +Ġcompañ ÃŃa +Ġsü rek +Ġsürek li +ĠиÑģ кÑĥÑģ +ĠиÑģкÑĥÑģ ÑģÑĤв +ĠB ürger +ת ×Ĺר +ת×Ĺר ×ķת +à¸ŀรà¹īà¸Ńม à¸ģัà¸ļ +Ø´ Ùħ +à¸ĸืà¸Ń วà¹Īา +è¾¼ ãĤĢ +ä¼ij ãģ¿ +ĠاÙĦØ£ ب +ĠÑģÑĤоим оÑģÑĤÑĮ +ĠпÑĢав а +may ın +ห วย +ĠاÙĦØ· بÙĬعÙĬ +à¸Ĺีà¹Ī à¸ŀัà¸ģ +ĠEst á +Ñĭва ÑİÑĤ +ب سÙĬ +بسÙĬ Ø· +Ġ×ij×¢ ×ijר +åı¯èĥ½ ãģ§ãģĻ +Ġ×ĵ ×ķ׾ +Ġ×ĵ×ķ׾ ר +Ùĩ ÙİØ§ +воÑĢ Ð¾ÑĤ +ãģ¦ ãģĦãģ¾ãģĹãģŁ +à¹Ĥà¸Ĺร ศ +à¹Ĥà¸Ĺรศ ั +à¹Ĥà¸Ĺรศั à¸ŀ +à¹Ĥà¸Ĺรศัà¸ŀ à¸Ĺà¹Į +Ġ×§ ׳ +ĠاÙĦØ« ÙĨ +ĠاÙĦØ«ÙĨ ائÙĬØ© +Ġco ût +à¸ķิà¸Ķ à¸ķัà¹īà¸ĩ +Ġö rg +Ġörg üt +ĠاÙĦØ® ÙĦÙĬ +ĠاÙĦØ®ÙĦÙĬ ج +Ġb á»įn +×ķ׾×ķ×Ĵ ×Ļ +ëŀ ľ +ĠÐij олÑĮ +ĠÐijолÑĮ ÑĪ +×Ĵ ×ijר×Ļ×Ŀ +ÙĤ ÙĬد +×ij×Ļ×ĺ ×ķ×Ļ +æīĵ ãģ¡ +Ġol muÅŁ +f äh +fäh ig +ล าà¸Ļ +ĠÙĤ طر +ש פ×Ķ +èªŃ ãĤĵãģ§ +à¸Ĥ วา +Ġchi ếm +ãĤ¤ãĥ³ ãĤ¿ +ãĤ¤ãĥ³ãĤ¿ ãĥ¼ãĥ +ãĤ¤ãĥ³ãĤ¿ãĥ¼ãĥ į +ãĤ¤ãĥ³ãĤ¿ãĥ¼ãĥį ãĥĥãĥĪ +Ġ׾ש×ŀ ×ķר +Ġت رÙĥ +ĠترÙĥ ÙĬا +ר ×ķ×ĺ +ã썿ĢĿ ãģĦãģ¾ãģĹãģŁ +ĠاÙĦت ÙĤ +Ġd ư +ãģ¦ãģıãĤĮ ãĤĭ +ãģĹãģŁ ãģĵãģ¨ +Ġróż ne +ĠاÙĦØ· ÙģÙĦ +ĠPost é +Ġ×ŀש ×ķ×Ŀ +Ñį ÑĢ +ĠÑĢабоÑĤ аеÑĤ +ãĤ· ãĥª +ãĤ·ãĥª ãĥ¼ãĤº +Ġ×ij×Ķ ×Ĺ׾×ĺ +×§×Ķ ×Ļ׾×Ķ +ãĤ« ãĥ¡ +ãĤ«ãĥ¡ ãĥ© +ï¼ ¯ +ĠìĤ¬ ìĿ´ +Ġk ì +Ġth Æ°á»Ľc +ض بط +ÙĤب ÙĪÙĦ +åĪ¥ ãģ® +Ġparticul ière +ĠÑģво ем +Ġ×¢ סק +Ġעסק ×Ļ×Ŀ +×ij×Ĺ ×Ļר×ķת +×ij ×Ļ׳×ķ +à¸ĭ à¸Ń +Ġ×¢ ×ķ×ijר +ãģłãģ£ãģŁ ãģ®ãģ§ +ıld ıģı +Ùħ دار +Ùħدار س +주 ìĭľ +à¸Ńา ศ +à¸Ńาศ ัย +Ġt ấm +à¸ŀิ à¸Ī +à¸ŀิà¸Ī าร +à¸ŀิà¸Īาร à¸ĵา +ÑĤелÑĮ нÑĭе +Ñģк ÑĥÑİ +Ðľ Ðĺ +à¹Ģà¸ģ า +à¹Ģà¸ģา หล +à¹Ģà¸ģาหล ี +×ĵ ×Ĺ +à¹Ģà¸Ĭ ิà¸ĩ +Ġد ÙĤÙĬÙĤØ© +íķĻ ìĥĿ +Ġש×IJ ׾×Ķ +Ġcontr ôle +Ġsit uação +à¸Ĥà¸Ńà¸ĩ à¸ľà¸¹à¹ī +ÙĨ Ø·ÙĤ +ê³¼ íķĻ +หลาย à¸Ħà¸Ļ +Ġn ắng +ÙĤ Ùı +ì¡° ê±´ +Ñ ķ +ãĥĥ ãģ¨ +×ŀ ×Ļ׾×Ķ +Gr ün +×Ļ ×Ļ×¢ +×Ļ×Ļ×¢ ×ķ×¥ +×ŀ׳ ׼ +ë ŃIJ +×ŀ×¢ ×ŀ×ĵ +สำ à¸Ļัà¸ģ +ج دد +à¸Ħ ัà¸Ķ +Ġ×Ķ×ŀש פ +Ġ×Ķ×ŀשפ ×Ĺ×Ķ +×ŀש ק׾ +ÙĦ Ùı +Ġty tu +Ġtytu ÅĤ +ÑĪ ÐµÐ¹ +ĠìĿ¼ ë¶Ģ +ÑĪ ÐµÐ½Ð¸Ðµ +Ġph óng +ĠìĹŃ ìĤ¬ +ãĤ« ãĥ³ +Ġtú i +ĠÙĨ ÙĪÙģ +ĠÙĨÙĪÙģ Ùħبر +gr ün +ĠاÙĦØ´ ÙħاÙĦ +ÅĽwi adc +ÅĽwiadc zenie +ער ×Ķ +Ġ×¢ ×ķ×ij +Ġ×¢×ķ×ij ×ĵ×Ļ×Ŀ +×ĵ×ķ×Ĵ ×ŀ×IJ +ä»Ĭ ãģ¯ +Ġv ão +ĠТ ем +Ñģ илÑĮ +Ġch ợ +Ùħ را +Ùħرا ÙĤب +à¹Ħมà¹Ī รูà¹ī +Ġر ائع +×IJ׳ ×Ĺ׳×ķ +สà¹Īà¸ĩ à¹Ģสริม +צ ×Ĺ +ĠìŀĪìĸ´ ìĦľ +Ġkur ulu +Ġkurulu ÅŁ +ĠÃĸ zellik +ĠÃĸzellik le +Ġת ×Ļ×§ +Ġgh é +Ġspr zÄĻ +ĠsprzÄĻ t +ער ×ķת +را ØŃØ© +ãģ£ ãģį +ãģ£ãģį ãĤĬ +ĠìķĦ ëŀĺ +stit uição +Ġдолж но +×Ķ ×¨×© +×Ķרש ×ŀ×Ķ +×Ķ׾ ×ļ +ãģ¡ ãģª +ãģ¡ãģª ãģ¿ +ãģ¡ãģªãģ¿ ãģ« +פ ×Ĺ×ĵ +ĠاÙĦج ÙħÙĬع +×ij×¢ ׾×Ļ +Ġtr ùng +Ġפ ת×Ĺ +×ŀ׾×Ĺ ×ŀת +ãĥĨ ãĥ¼ãĥ +ãĥĨãĥ¼ãĥ ŀ +Ùħ تاب +Ùħتاب عة +Ġ모 ìĬµ +ÙĬ ص +åIJĪ ãģĨ +ĠY ap +ĠYap ı +ĠÑģ казаÑĤÑĮ +ëª ° +à¸Ĺีà¹Ī สำà¸Ħัà¸į +ĠìĹĨ ìĬµëĭĪëĭ¤ +Ġnh ắc +Ġülk eler +Ġмног ие +íķĺ ìħ¨ +มาà¸ģ à¸Ĺีà¹Īสุà¸Ķ +à¸ģ à¹īา +à¸ģà¹īา ว +Ġİ yi +л еж +леж а +ãĤ¸ ãĥ§ +à¸Ĺั à¸ŀ +ا ÙĪØ± +Ġ×Ĺ×ijר ×Ļ +Ġ׾ ש×Ŀ +ì² « +ĠT á»Ń +×ŀ ×ķ׳×Ļ +ÙĤ ÙĪØ¯ +à¸ģระ à¹Ģà¸Ľ +à¸ģระà¹Ģà¸Ľ à¹ĭ +à¸ģระà¹Ģà¸Ľà¹ĭ า +ĠпÑĢоблем Ñĭ +Ġaç ıs +Ġaçıs ından +Ġ×Ķ×ŀ ׼ +ĠÙħع ظÙħ +ÙĤÙĬ اس +ĠпÑĢод олж +ĠпÑĢодолж а +Ġver diÄŁi +ĠпÑĢед меÑĤ +ãģĦãģ¾ãģĻ ãģĮ +ĠëͰ 른 +ĠاÙĦ ÙĤÙĬاÙħ +ĠØ¥ÙĦÙĬ Ùĩا +Т ÐIJ +п оз +ãĤ· ãĥ¥ +ä¸ĬãģĮ ãĤĬ +à¹Ģà¸Ķิม à¸ŀัà¸Ļ +à¸ģุ ล +ØŃر ÙĬØ© +×§×ij×ķצ ×ķת +ë¯ ¿ +ĠاÙĦÙħ ÙĨا +ĠاÙĦÙħÙĨا Ø·ÙĤ +ĠвÑĭп ол +ĠвÑĭпол нÑı +ãĥĭ ãĤ¢ +Ġê²° êµŃ +×Ĺ ×ķ×ŀ +×Ĺ×ķ×ŀ ר×Ļ×Ŀ +ĠУкÑĢа инÑĭ +ห à¸Ńม +ר ×Ļס +ĠÑħоÑĤ ел +ĠобÑĢаз ованиÑı +Ġkh ẳng +Ġm ưa +Ġgör me +Ġgüç lü +سع Ùī +มัà¹Īà¸Ļ à¹ĥà¸Ī +íķĺ ê²łìĬµëĭĪëĭ¤ +Ġпол Ñĥ +Ġfün f +ã썿ĢĿ ãģ£ãģ¦ãģĦãģ¾ãģĻ +Ġê·¸ê²ĥ ìĿĢ +ĠdÃ¼ÅŁÃ¼n ce +ìŀ ł +ĠH Æ°á»Ľng +ĠTi á»ĥu +Ġç ift +ãģij ãģ° +à¸Īà¸Ļ à¸ĸึà¸ĩ +à¸Ĺำ à¹Ħà¸Ķà¹ī +ĠìŀIJ ì²´ +Ġd õ +Ġdõ i +à¸Ī ัà¸Ļ +à¸Īัà¸Ļ à¸Ĺ +à¸Īัà¸Ļà¸Ĺ รà¹Į +ece ÄŁini +׳×ķ×¢ ר +غ ار +ĠاÙĦØ£ÙħرÙĬ ÙĥÙĬ +داع Ø´ +ĠбезопаÑģ ноÑģÑĤи +Ġб Ñİ +ĠбÑİ Ð´Ð¶ +ĠбÑİдж еÑĤ +ãĥĬ ãĤ¤ +à¸ŀà¸ļ วà¹Īา +da ÄŁ +×IJ ×ķפף +íĹ Į +ãĥĢãĤ¤ ãĤ¨ +ãĥĢãĤ¤ãĤ¨ ãĥĥãĥĪ +ĠëĮĢ íĨµ +ĠëĮĢíĨµ ëł¹ +D İ +Ø£ ØŃداث +ĠA ÄŁ +ĠAÄŁ ust +ĠAÄŁust os +ØŃÙĦ ÙĪÙĦ +Ġw ÅĽ +ĠwÅĽ ród +ĠÑģо оÑĤвеÑĤ +ĠÑģооÑĤвеÑĤ ÑģÑĤв +ĠÑģооÑĤвеÑĤÑģÑĤв ии +ĠLu áºŃt +Ġ׼׾ פ×Ļ +Ġв еÑī +ĠвеÑī еÑģÑĤв +×§ ×Ļ×¥ +ĠبÙĩ ذا +عا Ø´ +à¹Ģà¸Ľà¹ĩà¸Ļ à¹Ģรืà¹Īà¸Ńà¸ĩ +Т Ðķ +Ġ×ij×IJ ×Ļ׳×ĺר׳×ĺ +س عد +Ġ×Ķ×ĺ ×Ļפ×ķ׾ +פ ×Ļס +à¸ĩà¹Īาย à¹Ĩ +ĠGer ät +׾ ×Ļ×ĵ×Ķ +ĠÑĢ Ð¸Ñģк +׾ק ×Ĺ +н наÑı +ר ×Ļ×ĵ +п ÑĢакÑĤи +пÑĢакÑĤи к +à¸Ĥัà¹īà¸Ļ à¸ķà¸Ńà¸Ļ +à¸Ļà¹Īา รัà¸ģ +larınız ı +à¸Ńà¸Ļุ à¸įา +à¸Ńà¸Ļุà¸įา à¸ķ +ĠzdjÄĻ cia +Ġb ây +Ñģ ÑĢ +ÑģÑĢ Ð¾Ñĩ +ãĥĭ ãĥ³ãĤ° +Ġö ner +Ġöner i +Ġнов ÑĭÑħ +دع ÙĪØ© +Ġg ắn +ĠاÙĦÙĦ بÙĨ +ĠاÙĦÙĦبÙĨ اÙĨÙĬ +ãĥĨãĤ£ ãĥ¼ +Ġص ØŃÙĬØŃ +ем ÑĭÑħ +çĸ² ãĤĮ +ĠпÑĢо иÑģ +ĠпÑĢоиÑģ ÑħодиÑĤ +ส à¸ķิ +ĠT ết +Ġ×Ķ׾ ׾×ķ +à¹Ģรืà¹Īà¸Ńà¸ĩ à¸Ļีà¹ī +×ŀ×ij ׳×Ķ +Ġconte údo +Ġا خت +Ġاخت ÙĬار +Ùħ سÙĦ +ÙħسÙĦ سÙĦ +ëı Ī +Ġ׾ ×Ļ×ĵ +à¸ŀิ à¸ĺี +ĠÑģов Ñģ +ĠÑģовÑģ ем +ãģĮãģĤãĤĬ ãģ¾ãģĹãģŁ +Ġsó ng +Ø¥ صÙĦاØŃ +ë§ ģ +Ùģ ÙĬر +ĠJe żeli +ìłľ ëıĦ +d ÅĤug +ìĥģ ìĿĦ +Ġc áºŃn +Ġhá»į p +Ø£ ست +أست اذ +Ġ×ŀ ×Ļש×Ķ +Ġ×ŀ×Ļש×Ķ ×ķ +Ġd Ãły +Ġch Ãłng +ãģ¡ãĤĥãĤĵ ãģ¨ +ĠÄij ám +Ġsw ój +Ġpoder á +ĠоÑĤлиÑĩ а +Ġpéri ode +ünd ig +×ĺ×¢ ף +ÑģÑĤÑĢо иÑĤелÑĮ +ר ת×Ļ +Ġ×Ļ×Ķ ×Ļ×ķ +׾ ס +ĠاÙĦÙħÙĨ زÙĦ +à¸Ļิ à¹īว +иÑĦ ика +иÑĦика ÑĨи +ðŁĺ ī +Ġad ına +ãĢĤãĢĤ ãĢĤ +×IJ ×Ļף +ס ×Ļר +ĠÙĬ عد +çŃĶ ãģĪ +اÙĦ جز +اÙĦجز ائر +енÑĮ к +ร ห +รห ัส +ĠTürk çe +ê¾ ¸ +Ġ×Ļ ×ķ׼׾ +Ġש ×ķ׳×Ķ +Ġ×ij×ŀ צ×ij +ĠдейÑģÑĤв иÑĤелÑĮно +ĠبأÙĨ Ùĩ +×ŀ×§ ×ĵ +Ġ×Ķש ×§ +Ø®ÙĬ ارات +Ġf ı +Ġfı rs +Ġfırs at +ëij ĺ +ĠìĦľ ìļ¸ +Ġ×Ķ×Ĵ ×ķ×£ +ر عا +رعا ÙĬØ© +ĠK ết +к Ñģи +ĠÑĥÑģлÑĥг и +ноÑģÑĤ ей +ìļ´ ëıĻ +ĠобÑĬ Ñı +ĠобÑĬÑı вл +н еж +×Ķפ ×ļ +Ġ×ij×¢ ×Ļ׳×Ļ +ëĨ Ĵ +ĠпÑĢоÑĨ ед +ĠпÑĢоÑĨед ÑĥÑĢ +Ġiht iy +Ġihtiy acı +Ġë°Ķ ëŀį +Ġë°Ķëŀį ëĭĪëĭ¤ +à¸ģล ัว +ĠÑģл ожно +×§×Ļ ×Ļ×ŀת +ĠÄIJ ình +ĠÙħ ÙĦÙģ +Ġà¹Ĥà¸Ķย มี +Ġkat kı +تØŃ ÙĪÙĬÙĦ +à¹Ħ à¸ŀ +ĠH á»į +ñ e +Ġдо Ñħод +Ġtho ải +íķĺìŬ ìķ¼ +ãĤ¹ãĥĿ ãĥ¼ãĥ +ãĤ¹ãĥĿãĥ¼ãĥ Ħ +ĠG òn +Ġk è +Ġkè m +é̲ ãĤģ +ãĤ¹ ãĥ¼ãĥ +ãĤ¹ãĥ¼ãĥ ij +ãĤ¹ãĥ¼ãĥij ãĥ¼ +ĠgiÃł u +ĠØ¥ عادة +Ġ׾ ×ķ×§ +Ġ׾×ķ×§ ×Ĺ +ĠÑħоÑĩ еÑĤ +×ĺ ׾×ķ×ķ +×ĺ׾×ķ×ķ ×Ļ×ĸ +×ĺ׾×ķ×ķ×Ļ×ĸ ×Ļ×Ķ +Ġth uyết +ãģĿãĤĮ ãģ§ +Ġvard ı +à¹Ħร à¹ī +ع بد +ĠRep ública +ãĥ¼ãĤ¿ ãĥ¼ +Ġ×ŀ×IJ ×ķת +à¹Ħà¸Ľ à¹ģลà¹īว +Ġyapıl acak +ãĤ¹ãĤ¿ ãĥ¼ãĥĪ +ãģ» ãģ¼ +Ġko ÅŁ +ĠмаÑĤ еÑĢи +Ġsiè cle +ĠاÙĦÙħ ختÙĦÙģ +ĠاÙĦÙħختÙĦÙģ Ø© +Ġ׾ק ר×IJ +Ġ׾קר×IJ ת +Ġ×Ķפ ×ķ×¢×ľ +Ġt òa +Ġr Æ¡i +åij¨ ãĤĬ +à¸Ŀ à¸Ļ +j ÅĽÄĩ +ĠìķĬ ìĿĦ +اÙĨت ÙĤاÙĦ +ëĸ ł +ив аеÑĤ +ãĥĪ ãĥ« +ĠاÙĦÙģÙĦسطÙĬÙĨ ÙĬØ© +à¸ģลà¹Īาว วà¹Īา +ا Ùĥت +ĠÃĸ l +ĠÑĢе ÑĪи +ĠÑĢеÑĪи л +Ġ׳×ķס פ×ķת +Ġìłķ ì¹ĺ +вл еÑĩен +Ùħر ØŃÙĦØ© +Ġcome ça +Ġy ık +ìĤ ´ +à¸ĺ à¸Ļา +à¸ĺà¸Ļา à¸Ħาร +à¸Ńà¸Ļ า +à¸Ńà¸Ļา à¸Ħ +à¸Ńà¸Ļาà¸Ħ à¸ķ +Ġpeque ña +ä»ķ äºĭãĤĴ +Ġب ذÙĦÙĥ +Ġнов ого +ãģĹãģ¦ ãģĦãģªãģĦ +ĠاÙĦÙħ ÙĬاÙĩ +à¸ģà¹ĩ à¹Ģà¸Ľà¹ĩà¸Ļ +Ġж ÑĥÑĢ +ĠжÑĥÑĢ Ð½Ð°Ð» +в еÑģ +خت ار +Ġ매 ìļ° +ĠM ã +ĠавÑĤомаÑĤ Ñĭ +ضع Ùģ +ĠاÙĦÙģ Ùĥر +ãģ§ãģĻ ãģ®ãģ§ +ãĥ¡ãĥ³ ãĥIJãĥ¼ +Ġк ÑĢÑĥг +ĠاÙĦسÙĦ طة +à¸Ħรัà¹īà¸ĩ à¹ģรà¸ģ +à¸ģระà¸Ĺ รว +à¸ģระà¸Ĺรว à¸ĩ +ÑĨ ов +éķ· ãģĦ +大ãģį ãģĦ +Ġgeç miÅŁ +ìĦ± ìĿ´ +Ġצר ×Ļ׼×Ķ +Ġм оÑī +ĠмоÑī н +Ġ×§ ×Ļש +Ġ×§×Ļש ×ķר×Ļ×Ŀ +ĠNas ıl +г ÑĢан +Ġ×ŀ ×ķצר×Ļ×Ŀ +Ġ×ŀס ×ķ×Ĵ +Ġy ür +Ġyür üt +Ġ׾׊צ×ķ +×ķÖ ¼ +ĠìŀĪ ìĹĪëĭ¤ +Ġter ör +ĠTh ương +ĠÙĪ ÙĬÙħ +ĠÙĪÙĬÙħ ÙĥÙĨ +ج ÙĪÙĨ +ĠÙĪØºÙĬر Ùĩا +×ŀ פ×ķ +×Ĵ×ķר ×ŀ×Ļ×Ŀ +׼×ij ×Ļש +ĠاÙĦÙĦ غ +ĠاÙĦÙĦغ Ø© +شر Ùĥ +ĠاÙĦر اب +ĠاÙĦراب ع +ĠпÑĢ ÐµÐº +ĠпÑĢек ÑĢаÑģ +ĠпÑĢекÑĢаÑģ н +Ġenerg ÃŃa +×§×ĵ ×ŀ×Ļ +ãģıãģª ãģ£ãģŁ +ĠÄij ứ +ĠÄijứ a +Serv i +Servi ço +Ġkald ır +åĥį ãģį +Ġод еж +Ġодеж д +물 ìĿĦ +ãģĿãģĨ ãģ§ +ãģĮãģĤ ãĤĮãģ° +ìĻ ķ +צ×ĵ ×§ +Ġart ır +Ġile ti +Ġileti ÅŁim +ãĤĪãģĨ ãģ§ +ãĥĪ ãĥ¼ +ãĤ¢ ãĥĭ +ãĤ¢ãĥĭ ãĥ¡ +×ĺ×Ļ ×Ļ׾ +ãĥķ ãĥªãĥ¼ +ãĥĿ ãĥ³ +ÐŁÑĢ Ð¾ +Ġع اÙĦÙĬØ© +ĠÃ¶ÄŁ ret +ĠÃ¶ÄŁret men +ĠкаÑĩеÑģÑĤв а +Ġ×Ķ×ĺ ×ij×¢ +Ġзна Ñİ +ãģ¦ ãģıãĤĭ +Ġm ừng +ÙħÙĪ Øª +ש ×ķ×ŀר +×Ĺ׾ ×ij +Ġwzgl ÄĻ +ĠwzglÄĻ du +ë²Ī 째 +Ġtá» ĵ +Ġtá»ĵ n +ãĥ¯ãĥ¼ ãĤ¯ +Ġpo życz +Ġpożycz k +×Ļ ×ķצר×Ļ×Ŀ +Ùĥر Ùħ +Ġг аÑĢ +ĠгаÑĢ Ð°Ð½ +ĠгаÑĢан ÑĤи +ล à¹īาà¸ĩ +Ġìĺģ íĻĶ +×ĺ ×Ļס +Ġth ẻ +ĠìŀĪëĭ¤ ê³ł +اÙĦت ز +اÙĦتز اÙħ +Ġна ÑĪи +is ée +ãģĵãĤĮ ãĤĴ +Ġm ẽ +ض ÙĦ +بÙĪ Øª +Ġ׼ ׼×Ķ +h ợ +ĠاÙĦس ÙĪØ±ÙĬØ© +Ġ×ľ×¢ ×ķ×ŀ +Ġ×ľ×¢×ķ×ŀ ת +ĠbaÅŁ ar +ĠbaÅŁar ılı +е ÑģÑĤÑĮ +à¸Ħร ี +à¸Ħรี ม +ĠìłĦ ì²´ +ĠسÙĬ ÙĥÙĪÙĨ +Ġ×ŀ×ĵ ×ķ×¢ +ĠëķĮ문 ìĿ´ëĭ¤ +Ġc ứng +ger ät +Ġм иÑĢ +ĠмиÑĢ Ðµ +ĠÙĥÙĬÙģ ÙĬØ© +Ġפר ×ĺ×Ļ×Ŀ +Ġgo ÅĽci +иÑĤ еÑģÑĮ +ÑĥÑĪ ÐºÐ¸ +ؤ ÙħÙĨ +Ġ×IJ ׼ף +ĠاÙĦر جÙĦ +Ġl á»įc +à¹Ģรีย à¸ģวà¹Īา +ãģĵãģ® ãĤĪãģĨãģª +ë§Į íģ¼ +Ġп еÑĩ +ÙĪÙĦ ات +ĠÃľ ye +liÄŁ inde +à¸Ħะ à¹ģà¸Ļ +à¸Ħะà¹ģà¸Ļ à¸Ļ +ãĤĭãģĵãģ¨ ãģ¯ +วิ à¹Ģà¸Ħร +วิà¹Ģà¸Ħร าะ +วิà¹Ģà¸Ħราะ หà¹Į +Ġвозмож ноÑģÑĤи +ĠاÙĦÙĨ ساء +ãĥīãĥ© ãĥŀ +Ġgü c +Ġgüc ü +Ġt ưá»Ŀng +Ġacomp aña +ãĤ¤ ãĥ© +×§ צ×ij +ĠY ö +ĠYö net +ĠYönet im +สัม à¸ľ +à¸ªà¸±à¸¡à¸ľ ัส +à¸Ļ าม +ĠÄij ợi +à¹ģหà¹Īà¸ĩ à¸Ĭาà¸ķิ +ãģĿãĤĮ ãģ§ãĤĤ +ät ig +ת ×ķ×Ŀ +ĠbaÅŁ lat +ĠвÑģ ей +ת ×Ļ×§ +ת×Ļ×§ ×ķף +ĠNg ô +ĠGesch ä +ĠGeschä fts +Ø£ Ùħ +Ø£Ùħ راض +à¹Ģà¸Ĺ à¸Ħà¸Ļ +à¹Ģà¸Ĺà¸Ħà¸Ļ ิ +à¹Ģà¸Ĺà¸Ħà¸Ļิ à¸Ħ +Ġм енÑĮ +ĠменÑĮ ÑĪе +Ġöl ç +Ġölç ü +ĠÙĬ جعÙĦ +ĠÄij ỡ +ש ×Ļ׾ +ש×Ļ׾ ×ķ×ij +ĠGr Ã¶ÃŁe +ĠÙĩ اتÙģ +รà¹īาà¸Ļ à¸Ńาหาร +×Ķ׾ ×Ļ׼ +×Ķ׾×Ļ׼ ×Ļ +иÑĢÑĥ ÑİÑī +èĭ¥ ãģĦ +ĠÃĸ zel +ãģĦãģŁ ãĤī +à¸Ħำ à¸ĸาม +Ġzosta ÅĤy +Ġ×Ķס ×Ļפ×ķר +×Ķ ×ķ׾ +×Ķ×ķ׾ ×ļ +à¹Ģà¸Ĭà¹Īà¸Ļ à¸ģัà¸Ļ +à¹Ĥ à¸Ĩ +à¹Ĥà¸Ĩ ษ +à¹Ĥà¸Ĩษ à¸ĵา +×IJר צ×ķת +×Ĵר פ×Ļ +Ġao ût +ĠÙĬ رÙĬد +ت ÙĪØ¬ +تÙĪØ¬ ÙĬÙĩ +ĠÑįÑĤ ап +ãĤ¹ãĤ¿ ãĥ³ +Ġkr ó +Ġkró tk +ãĤĴ使 ãģĨ +ì ·¨ +éĸ¢ ãĤı +à¸Ķà¹īวย à¸Ħวาม +à¸Ļำ à¹Ģสà¸Ļà¸Ń +Ġa yrıca +à¸Ī à¹īาà¸ĩ +ĠÑĦоÑĤ огÑĢаÑĦ +Ġв еÑĩ +ĠвеÑĩ еÑĢ +åĩº ãģĹãģŁ +ĠÐ¥ о +Ġ×ŀ ר×Ĵ×Ļש +à¹ĥหà¹ī à¹Ģà¸Ľà¹ĩà¸Ļ +ãĤĴ 缮 +ãĤĴ缮 æĮĩ +׾ ×ŀ×Ļ×Ŀ +nÄħ ÅĤ +ĠÑģÑĤ анд +ĠÑģÑĤанд аÑĢÑĤ +ĠSü d +ĠT âm +اخت بار +à¹Ģà¸ģ à¸Ńรà¹Į +Ùħس رØŃ +Ġbi á»ĩn +ب Ùı +Ġص اÙĦ +ĠصاÙĦ ØŃ +ĠPh ụ +íľ ´ +ãĥ¬ãĥĵ ãĥ¥ãĥ¼ +Ġbụ ng +Ġrég ime +ĠØ£ Ø´Ùĩر +ĠÑĢабоÑĤ ник +à¸Ŀ ัà¸Ļ +اع تÙħ +اعتÙħ اد +Ġзам еÑĤ +ãģ¾ ãģ£ãģ¦ +Ġch ặt +æĿ¥ ãĤĭ +ĠاÙĦÙĤ ÙĪØ§Øª +ãģ«åħ¥ ãģ£ãģ¦ +تØŃ اÙĦÙģ +Ùħ زÙĬد +ĠÙĬ صÙĦ +ìĹ ¼ +à¹Ģà¸Ĭ à¹ĩ +à¹Ģà¸Ĭà¹ĩ à¸Ħ +Ġk á»ĭ +Ġká»ĭ p +ĠìķĦ ì§ģ +×IJ׳ ×Ĵ +Ġобла ÑģÑĤÑĮ +Ġpomoc Äħ +Ġ×ķ ש׾ +ëĵł ì§Ģ +ĠGi ám +ĠSt ück +Ġchá y +ĠëĤĺ ìĺ¤ +ש ×Ļ×ĺת +×ŀ×ĵ ר +×ŀ×ĵר ×Ļ×ļ +Ġsüre ç +к ва +×ij׾ ×Ļ×Ŀ +×Ķ ×ª×Ļ +×Ķת×Ļ ×Ļ×Ĺס +ÙĤب اÙĦ +Ġס ×ķ×Ĵ +Ġס×ķ×Ĵ ×Ļ +ÑģÑĤ олÑĮ +ä½ķ ãĤĤ +×ĸ׼ ×ķר +è²· ãģĨ +å®ī ãģı +à¸Ħรัà¹īà¸ĩ à¸Ļีà¹ī +kö p +ĠÑģеÑĢ Ð²Ð¸Ñģ +оÑĩ нÑĭÑħ +ê±° ëŀĺ +تأ Ùĥ +تأÙĥ ÙĬد +×ĵ ׾ק +Ġпо Ñĩем +ĠпоÑĩем Ñĥ +пиÑģ аÑĤÑĮ +×ij שר +ĠH Ãłng +ĠT ìm +Ġtr ừ +ãĤ» ãĥĥãĤ¯ãĤ¹ +×ķ׳ ×Ĵ +mız da +п Ñģи +ĠìŀĪ ê¸° +Ġr út +ز اÙĨ +تÙĨ ÙĪØ¹ +ÙħÙĤ ا +ÙħÙĤا ÙĪÙħØ© +Ġ׾צ ×ķר×ļ +Ġ×ij ×Ļר×ķש׾×Ļ×Ŀ +ãĥ´ ãĤ£ +eb ile +ebile ceÄŁi +ãĥ¦ ãĥ¼ãĤ +ãĥ¦ãĥ¼ãĤ ¶ +ãĥ¦ãĥ¼ãĤ¶ ãĥ¼ +ãĤĴä½ľ ãĤĭ +Ñģ меÑĢ +ÑģмеÑĢ ÑĤ +Ġì§ ģ +Ġì§ģ ìłij +ĠÐŁ аÑĢ +ØŃ اض +ØŃاض ر +Ùħ ÙĥاÙģ +ÙħÙĥاÙģ ØŃØ© +ล ิà¸Ļ +ãģ¦ ãģįãģ¦ +ÑĢоÑģ л +ĠÄ°ÅŁ te +ÙĤص ÙĬر +Ġ×ij×Ĵ ×Ļ׾ +Ġ×ŀת ×IJ×Ļ×Ŀ +Ġ×Ķ ×Ĺ×ĵ +Ġ×Ķ×Ĺ×ĵ ש×Ķ +ר ×ķ×¢ +Ġprodukt ów +ĠÙħ صدر +не ÑĨ +ĠاÙĦعÙħÙĦ ات +Ġçık ma +Ġد بÙĬ +×§ ×Ļף +ת ×IJר +ת×IJר ×Ļ×ļ +׳×Ļ ×Ļ×ĵ +صر اع +l ève +צ ×Ļר +à¸Ķ ัà¸Ļ +à¹ĥหà¹ī à¹Ħà¸Ķà¹ī +ãĤ¿ãĤ¤ ãĥł +Ġgi ảng +С ÐŁ +ĠاÙĦÙħ ØŃÙĦ +ĠاÙĦÙħØŃÙĦ ÙĬØ© +ĠT ất +׾ ×ķ×ĺ +h á»ķ +Ġam éric +Ġaméric ain +Ġ×ijש׾ ×ij +Ġ׾×IJ ×ķ×ŀ×Ļ +Ġpe ça +ĠÑĢаз нÑĭÑħ +ãģĦãĤĭ ãģ¨ +ãĥĩ ãĥ³ +ס קר +Ġ×Ķ×ŀ×Ĺ ×Ļר +ãģ¨ãģĦãģĨ ãĤĤãģ® +رت بط +ĠиÑģÑĤ оÑĩ +ĠиÑģÑĤоÑĩ ник +สมัà¸Ħร สมาà¸Ĭิà¸ģ +Ġ à¸Ĺัà¹īà¸ĩ +Ġà¸Ĺัà¹īà¸ĩ à¸Ļีà¹ī +ĠT áºŃp +ãģ£ãģ¦ ãģĦãģĨ +ĠاÙĦÙĪ ØµÙĪÙĦ +Ġdéc ada +Ġо ÑĦоÑĢм +ĠоÑĦоÑĢм лен +สำหรัà¸ļ à¸ģาร +Ġog óln +ãģĨãģ¡ ãģ« +Ġvá rias +ãģĻãģİ ãĤĭ +ÙĪ Ùĩا +à¹Ĥà¸Ľà¸£ à¸Ķ +ĠÐłÐ¾ÑģÑģ иÑı +人 ãĢħ +ãģĹãģ¦ ãģįãģŁ +Ġsı rasında +Ġng ôn +س ÙĨØ© +تÙħ تع +×ŀ׼ ×ij×Ļ +Ġnh ấn +×¢ ×ŀ×Ļ×ĵ +á» ¨ +ж иÑĤÑĮ +ãĤī ãģĽ +gr áf +gráf ica +ĠÙĤ ÙĪÙĦ +ĠÙĤÙĪÙĦ Ùĩ +ëĭ¨ ì²´ +ห à¹īา +หà¹īา ม +使 ãģ£ãģ¦ +ת ×Ļ×ij +ת×Ļ×ij ת +i á»ĥu +à¹ģ à¸Ĭม +à¹ģà¸Ĭม à¸Ľ +à¹ģà¸Ĭà¸¡à¸Ľ à¹Į +Ạ¬ +ĠëĤĺ ëĿ¼ +ĠÙħباشر Ø© +Ġtr Äĥm +سÙĥ ÙĪ +ĠاÙĦذ Ùī +Ġbi ç +Ġbiç im +ت راجع +Ġоб еÑģп +ĠобеÑģп еÑĩ +ĠобеÑģпеÑĩ ива +Ġвозд ÑĥÑħ +Ñĭв аÑĤÑĮ +ÙĦ ØŃÙĤ +ĠMü dü +ĠMüdü rl +ĠMüdürl Ã¼ÄŁÃ¼ +Ġyapt ır +Ġפר ס +Ġפרס ×ķ×Ŀ +Ø· ÙĪØ± +ÑģÑĤв оваÑĤÑĮ +ìŀ¥ ìĿĦ +à¸Ĺีà¹Īà¸Ķี à¸Ĺีà¹Īสุà¸Ķ +à¸Ńั ล +ÑĢ Ñİ +Ùħست ÙĤبÙĦ +Ñģл ÑĥÑĪ +ÑģлÑĥÑĪ Ð° +èªį ãĤģ +Ġ׾ ×Ļ×ŀ +Ġ׾×Ļ×ŀ ×ķ×ĵ×Ļ +ת ש×ķ×ij +תש×ķ×ij ×ķת +ĠgerçekleÅŁtir il +ĠاÙĦ اتÙ쨧ÙĤ +ĠÑĥÑĢов не +ĠÑĤ ÑĢав +Ġ×Ķ×ŀ ×ķף +ØŃÙģ Ø§Ø¸ +ĠÙħ ÙIJ +ĠÙħÙIJ ÙĨ +ĠÙħÙIJÙĨ ÙĴ +Ġdem ás +×ŀ×ķ×ĸ ×Ļ×§×Ķ +ש ×Ļ×Ĺ×Ķ +Ġb ú +алÑĮ нÑĭм +ãĤı ãģŁ +ãĤıãģŁ ãģĹ +ĠاÙĦÙħÙĪ Ø§Ø¯ +ת ׼׳ +×ª×Ľ×ł ×ķף +ãĥŃ ãĥĥãĤ¯ +hi ếu +ĠÑĥ ме +ÙħØŃا ÙĪÙĦØ© +×IJ ×ķשר +Ġкон кÑĥÑĢ +ĠконкÑĥÑĢ Ñģ +Ġ×ŀ ×ij×Ĺ +Ġ×ŀ×ij×Ĺ ×Ļ×ł×ª +Ġan lam +Ġanlam ı +Ġli á»ĩt +Ġв Ñħод +ĠH ình +ĠÙĨ ÙĬ +ĠÙĨÙĬ ÙĪØ² +ãĤ¸ãĥ£ ãĥ¼ +×ij ×Ļ×¥ +ÑĤелÑĮ нÑĭÑħ +à¸Ĺุà¸ģ à¸Ńยà¹Īาà¸ĩ +ĠkiÅŁ inin +Ø£ Ùĥثر +ĠиÑģÑĤоÑĢ Ð¸Ð¸ +Ġë³Ģ íĻĶ +פ׾ ס×ĺ +×¤×ľ×¡×ĺ ×Ļ׳×Ļ +ĠÑģ еÑĤ +ĠÑģеÑĤ и +dıģ ımız +íķĺ ëıĦë¡Ŀ +×Ķ ×¨ +×Ķר ×ij×Ķ +ãģĻãĤĭãģĵãģ¨ ãģ¯ +Ġphi ếu +تØŃ سÙĬÙĨ +ĠÅĽ rod +ĠÅĽrod ow +ĠÅĽrodow isk +ĠÑĢаÑģ Ñħод +بر ÙĬد +Ġر ÙĬ +ĠرÙĬ اÙĦ +Ġ×ķ ׼×ļ +ì§Ģ ìļĶ +׼ ×ŀ×ķ +Ġ×¢×ľ ×Ļ×Ķ×Ŀ +f ÃŃcio +Ġkar arı +tıģ ını +ĠС ов +ĠСов еÑĤ +ãģĬéĩij ãĤĴ +м еждÑĥ +междÑĥ на +междÑĥна ÑĢод +междÑĥнаÑĢод н +Ġm á»Ŀi +ĠاÙĦØ¥ ÙĬر +ĠاÙĦØ¥ÙĬر اÙĨÙĬ +ĠاÙĦرÙĪ Ø³ÙĬ +ص ÙĨد +صÙĨد ÙĪÙĤ +ĠاÙĦØ¥ÙĨ ترÙĨت +Ġt ắm +ĠÑĤак ого +Ġ×ij ׾×ķ×Ĵ +Ġü crets +Ġücrets iz +×Ĺ×ĸ ×Ļר +ìĸ´ ìķ¼ +ĠPh ần +ï¼ ľ +Ġ×ĺ ×ij×¢ +Ġ×ĺ×ij×¢ ×Ļ +×IJ×ŀ ×IJ +اÙĤ ÙĦ +Ġcondi ções +ÙĤات ÙĦ +ĠÑĢезÑĥлÑĮÑĤаÑĤ е +ĠÑģво ими +צ×ij ×Ļ×¢ +gé ni +Ġz es +Ġzes po +Ġzespo ÅĤ +ÑĪ Ð¸Ð² +Ġפר×ĺ×Ļ ×ķת +Ùħست Ø´Ùģ +ÙħستشÙģ Ùī +شر ع +Ġko ÅĽci +Ġ×Ķ×IJ ×Ļ׳×ĺר׳×ĺ +ĠЧ еÑĢ +поÑĩ ÑĤ +Ġactiv ités +çŁ¥ ãģ£ãģ¦ +Ġ×ij ×ĸ×Ķ +Ġyüz den +ãģªãĤĬ ãģ¾ãģĽãĤĵ +Ġíĺ ¹ +Ġíĺ¹ ìĿĢ +Ġ×ŀש ׳×Ķ +ĠÐĴ еÑĢ +Ġ×ij×IJ×ķת ×ķ +éĿ¢ çϽ +éĿ¢çϽ ãģĦ +شر ØŃ +gr ünde +Ùģ Ø´ +Ù쨴 ÙĦ +Ġsé jour +ë´ IJ +Ġr ôle +Ø´ عار +ем Ñĭе +ĠاÙĦج سÙħ +алÑĮ ное +Ġìĥģ íĥľ +ï¼ ¤ +ë¯Ģ ë¡ľ +ĠÙĨ ÙĤØ· +ĠÙĨÙĤØ· Ø© +ãģĿãģĨ ãģł +ãģĻãĤĭ ãģ®ãģĮ +ห ู +Ġnh á»ĭ +Ġeconóm ica +ס×ĺ ×ķ×ĵ +ס×ĺ×ķ×ĵ ׳×ĺ +มี à¹Ĥà¸Ńà¸ģาส +Ġgest ão +รูà¹ī วà¹Īา +Ġlo ạt +ĠاÙĦÙħ Ùı +ĠاÙĦØŃ ÙħÙĦ +ĠاÙĦعÙħÙĦ ÙĬØ© +Ġê²ĥ ëıĦ +ĠÐľÐ¾Ñģк ва +×§×ĺ ×ķר +Ġпод ÑĢоб +ĠподÑĢоб н +Ġl ưng +ت Ù쨳 +تÙ쨳 ÙĬر +ĠاÙĦ بع +ĠاÙĦبع ض +ئ ت +Ðķ ÐĿ +ìŰ 구 +à¹ĥหà¹ī à¸Ħุà¸ĵ +ãģĤãĤĬ ãģ¾ãģĹãģŁ +Ġbir ka +Ġbirka ç +Ġİ sl +Ġİsl am +çĹĽ ãģ¿ +Ġh ảo +Ġм аÑı +ĠiÅŁ çi +ש × +×©× ģ +à¸ģาร à¹Ģมืà¸Ńà¸ĩ +×ķ×Ķ ×¨ +Ġch ó +ëĨ Ģ +Ġyan lı +Ġyanlı ÅŁ +幸 ãģĽ +×IJר×Ĵ ×ķ׳×Ļ +à¸Ńาà¸Ī าร +à¸Ńาà¸Īาร ยà¹Į +ĠинÑĦоÑĢм аÑĨиÑİ +Ðĵ Ðŀ +׳ ×Ĺש +ĠìķĮ ìķĦ +ĠÑħаÑĢакÑĤеÑĢ Ð¸ÑģÑĤ +ĠÑħаÑĢакÑĤеÑĢиÑģÑĤ ик +à¸Ħุà¸ĵ สามารà¸ĸ +è¦ĭ ãģĪãĤĭ +à¸Ĭัà¸Ķ à¹Ģà¸Ī +à¸Ĭัà¸Ķà¹Ģà¸Ī à¸Ļ +ĠdziaÅĤ al +ĠdziaÅĤal noÅĽci +à¹Ĥà¸ŀ สà¸ķà¹Į +ĠÐļ ол +ĠÙģ ÙĩÙĬ +Ġ×ŀ פ׳×Ļ +Ġ×Ķ×§ שר +Ùħر Ùĥ +ÙħرÙĥ ز +Ġho á +Ġа пп +Ġапп аÑĢаÑĤ +Ġp ami +Ġpami ÄĻ +ĠpamiÄĻ ta +Ġç ünkü +×ĵ ×ķף +ãģ¯ ãģĵãģ¡ãĤī +ĠM Ãł +ĠÙĬ ÙĤدÙħ +ĠпÑĢ ÐµÐ· +ĠпÑĢез иденÑĤ +à¸Ńุ à¸ķ +à¸Ńุà¸ķ สา +à¸Ńุà¸ķสา ห +à¸Ńุà¸ķสาห à¸ģรรม +ì§Ģ ìĽIJ +Ġ×IJפשר ×ķת +sch üt +schüt z +ĠTi ên +Ġsay ılı +ĠгÑĢÑĥпп Ñĭ +оÑĩ нÑĭй +Ġ×ľ×¢ ×ŀ×ķ×ĵ +Ġwr zeÅĽ +ĠwrzeÅĽ nia +ĠÄIJ ầu +à¹Ģà¸Ĥà¹īา รà¹Īวม +nız da +Ø®ÙĬ ص +Ġgü nc +Ġgünc el +ĠÙĦÙĩ ذÙĩ +ĠÙĬ عتبر +lé gi +ãĤı ãģĭãĤĭ +Ġr ừng +ظ Ùĩ +ظÙĩ ÙĪØ± +Ġ×ŀ×ij ×Ļף +Ġ기 íĥĢ +åĪĩ ãĤĮ +lan mÄ±ÅŁ +à¸Ĺีà¹Ī มีà¸Ħวาม +Ġh á»ģ +ت ÙĪØ¬Ùĩ +ĠاÙĦØ¥ دارة +Ġú til +ס פ×ķ +à¸Ħวาม รัà¸ģ +à¹Ĥ ฮ +Ġпол иÑĤ +ĠполиÑĤ ик +Ġsat ın +ĠÅŀ imdi +×ŀ ×ķר×Ļ×Ŀ +ìķĺ ëĭ¤ +×Ĺ ×ķ×ķ +×Ĺ×ķ×ķ ×Ļ×Ķ +à¸Ħà¸Ńม à¸ŀิ +à¸Ħà¸Ńมà¸ŀิ ว +à¸Ħà¸Ńมà¸ŀิว à¹Ģà¸ķà¸Ńรà¹Į +Ġا ذا +تخ اذ +ãĤ¨ ãĥ« +Ġpossibilit é +ยืà¸Ļ ยัà¸Ļ +Ġü nivers +Ġünivers ite +ĠاÙĦد ÙĪØ±ÙĬ +ĠìķĬëĬĶ ëĭ¤ +ĠìĦľ ë¡ľ +ØŃ اÙĦ +Ġë ¨ +Ġë¨ ¼ +Ġ먼 ìłĢ +à¸Ĺีà¹Ī à¸ĸูà¸ģ +ì§ ľ +Ġsk óry +лÑĮ ÑĨ +à¹ĥà¸Ĭà¹ī à¹Ģวลา +×ij×§ שת +Ġذ ÙĪ +æĹ¥ ãĢħ +ĠкоÑĤоÑĢ ÑĥÑİ +ĠÑĥÑĢов енÑĮ +ê¹ ¨ +à¹Ħ à¸Ĺ +ãĤµ ãĥĹãĥª +ãĤ¸ ãĥ§ãĥ³ +ãģĻ ãģ¹ãģį +ĠG ór +ãĥĪ ãĤ¤ +ãĥĪãĤ¤ ãĥ¬ +ĠyaÅŁ ama +Ġdá»ĭ p +Ġb ữa +à¸ĭ ุ +Ġöl üm +ãģ£ãģ¦ ãģıãĤĭ +à¸ģาร à¸Ħà¹īา +ש ער +ĠÑĤип а +Ġг еÑĢ +ĠгеÑĢ Ð¾ +רק ×¢ +Ġu waż +Ġuważ a +ש×ŀ ף +Ġhast alık +ãĤıãĤĮ ãĤĭ +ba ÅŁÄ± +Ñĩ ÑĤо +Ġ×ij ×ŀר׼×ĸ +Ġìļ°ë¦¬ ìĿĺ +ĠÙĥاÙĨ ÙĪØ§ +ĠØ£ بر +Ġأبر ÙĬÙĦ +ì¸ µ +à¹Ħà¸Ĥ à¹Ī +ĠÙĪ ÙĦÙĪ +à¸Ĺ ัว +à¸Ĺัว รà¹Į +ĠÙĪØ£ Ùĥد +à¸Ĭ วà¸Ļ +׾ ×ķ×§ +æį ¨ +æį¨ ãģ¦ +Ġİç in +p éri +Ġy al +Ġyal nız +ÑĮÑı н +Ġg ắng +à¸ģà¹ĩ ยัà¸ĩ +ĠУкÑĢа ин +ĠÑģ ами +ĠпÑĢовед ен +à¸ķà¸ģ à¹ģà¸ķà¹Īà¸ĩ +ĠQu ân +é paration +ĠbaÅŁ ında +Ġzn ale +Ġznale ź +Ġznaleź Äĩ +ãĤ± ãĥ¼ +ãĥİ ãĥ¼ +à¸ĸูà¸ģ à¸ķà¹īà¸Ńà¸ĩ +ëª ¸ +Ġëı Į +ĠëıĮ ìķĦ +ĠSch üler +Ġпод гоÑĤов +ĠподгоÑĤов к +ع رÙĪ +عرÙĪ Ø¶ +la ÅŁtır +ĠÑģоÑģÑĤав лÑıеÑĤ +ĠпÑĢоиз вод +ĠпÑĢоизвод ÑģÑĤва +ĠоÑģнов е +ĠØ´ ÙħاÙĦ +à¸ģร ี +ĠgörÃ¼ÅŁ me +оÑĩ ек +Ġ×Ĺ×ijר ×Ļ×Ŀ +ÙħØ® اط +Ùħخاط ر +ï¼ Ń +ר פ×IJ +ĠM ẹ +ยà¸Ńม รัà¸ļ +Ġv ết +Ø® ذ +ĠاÙĦت Ø· +ĠاÙĦتط بÙĬÙĤ +à¸Ļ ึà¸ģ +Ġ×Ķ ×Ľ×ł×¡×ª +ĠогÑĢ Ð°Ð½Ð¸ +ĠогÑĢани Ñĩен +ĠÃĩ alÄ±ÅŁ +ĠاÙĦÙħÙĨت دÙī +à¸Īำà¸Ļวà¸Ļ มาà¸ģ +ĠÑĤоÑĢ ÑĢ +ĠÑĤоÑĢÑĢ ÐµÐ½ÑĤ +ĠìĤ´ ìķĦ +à¸ŀลัà¸ĩ à¸ĩาà¸Ļ +à¸Ĭ ัà¸Ļ +ĠÐIJн дÑĢ +Ġréalis é +×ŀש ×IJ +à¹ģ à¸Ĭ +à¹ģà¸Ĭ รà¹Į +Ġб ог +มา à¹ģลà¹īว +ĠاÙĦÙĨ ار +Ġolmad ıģı +×ĵ ×¢×Ķ +ĠÑĥ веÑĢ +ĠÑĥвеÑĢ ÐµÐ½ +ãĤĭ ãĤĤãģ® +Ø£ د +أد ÙĪØ§Øª +Ġ×Ķ×ĸ ×ķ×Ĵ +Ø¥ عÙĦاÙħ +h á»ı +ĠNä he +ĠÑĤ еÑģÑĤ +Ġ×ŀ ×ķ׼ר +Ġë¬¸ìłľ ê°Ģ +ת ×ķצ×IJ×Ķ +m ó +mó vel +ĠاÙĦتج ارة +Ġмног иÑħ +обÑī а +Ġ×¢ סק×Ļ +ĠEdu cação +×§ ש×Ļ×Ŀ +é tabl +établ issement +Ġд еле +иÑĢÑĥ еÑĤÑģÑı +Ø¢ ثار +Ġ×Ķ×ŀ ר׼×ĸ×Ļ +ãĥIJ ãĥ« +ĠвÑģÑĤÑĢ ÐµÑĩ +ãģĴ ãĤĭ +Ġci Äħ +ĠciÄħ gu +ÙĬ ست +à¸łà¸² ว +à¸łà¸²à¸§ ะ +Ø£ Ùħر +Ġо жи +Ġожи да +Ġ á»§y +ãĥŀ ãĥ« +ر اس +оÑĩ ной +ת ×Ĵ×ķ×ij×ķת +تع رÙĬÙģ +ĠÑģо ÑĨиалÑĮно +ãĤĴ éĸĭ +ĠиÑģÑģлед ова +Ġd ú +Ġdú vida +Ġsk ÅĤ +ĠskÅĤ ada +Ġhä ufig +ĠвÑĭб ÑĢ +ĠвÑĭбÑĢ Ð°ÑĤÑĮ +ãģ®ãģ§ãģ¯ãģªãģĦ ãģĭ +ĠÑģ илÑĮно +ÑĤвеÑĢж ден +ר פ +רפ ×ķ×IJ×Ķ +æĢĿ ãģĦãģ¾ãģĻ +ØŃر ص +ש×ķת ×£ +Ùħس جد +à¹Ĥà¸Ĭ วà¹Į +ем ÑģÑı +в ÑĪие +Ġм л +Ġмл н +Ġ׾×Ķ ×ij×Ļ×IJ +ĠÙĬ تعÙĦÙĤ +à¸ķ ูà¹ī +Ġп ÑĢаз +ĠпÑĢаз д +ĠпÑĢазд ник +Ġн ем +Ġнем ного +Ġs Ãłng +تÙĨ سÙĬ +تÙĨسÙĬ ÙĤ +Ġtá» Ŀ +Ġмед и +ãģ« æĪ +ã쫿Π» +à¸Ħว à¹īา +ãģĭ ãģijãĤĭ +×ij׾ ×ķת +ĠÑįк Ñģп +ĠÑįкÑģп еÑĢÑĤ +Ġдев ÑĥÑĪ +ĠдевÑĥÑĪ Ðº +ĠØŃ ص +ÙĨØ´ Ø£ +ãģĮãģĤãĤĭ ãģ®ãģ§ +Ġت راÙħ +ĠتراÙħ ب +أس ÙĪØ§ÙĤ +Ġ׾פ ׳×ķת +Ġا ï»· +ãģ« ãģı +ãģ«ãģı ãģĦ +ĠØ£ عÙĦÙī +Ġ׾×Ķ ×ŀש×Ļ×ļ +rä u +ש×ŀ ×Ļ×Ŀ +åĪĨ ãģij +ãģĻ ãģ§ +ãģĻãģ§ ãģ« +×Ķ׾ ׼×Ķ +×Ĺ׾ ×Ļ×£ +Ġì ±ħ +Ġì±ħ ìŀĦ +à¹Ģà¸Ī ริ +à¹Ģà¸Īริ à¸į +éģĬ ãģ³ +ج سد +สา à¸ĺ +สาà¸ĺ าร +สาà¸ĺาร à¸ĵ +Ġbas ın +ÑĢаР³ +г ад +Ġho ÅŁ +íķ µ +×ij×Ĺ ×Ļר×Ķ +×ŀס ×ļ +Ġìłľ íĴĪ +تÙħ ÙĪÙĬÙĦ +ĠL ưu +ë¡ľ ë¶ĢíĦ° +Ġп об +Ġпоб ед +ÙħÙĨ ذ +常 ãģ« +ÙĤ س +ĠاÙĦÙħ صدر +ĠÙĪØ§ÙĦ است +Ġkh ắp +ĠاÙĦج اÙĨب +Ġng uyá»ĩn +éĸĵ éģķãģĦ +ĠÑģÑĤ ÑĢа +ĠÑģÑĤÑĢа Ñħ +ĠÑģÑĤÑĢаÑħ ов +รี à¸ļ +Ġx ương +Ġì° ¾ +Ġì°¾ ìķĦ +Ġng ại +г ал +à¸ĭ ีà¹Ī +Ġ×ij פ×Ļ×Ļס×ij×ķ×§ +Ц енÑĤÑĢ +Ġaval iação +Ġeconóm ico +×ĸ ף +ĠÐľ ак +Ġinter és +à¸ģล ิà¹Īà¸Ļ +ÑģÑĤÑĮ Ñİ +ĠÄij ương +å¼· ãģı +ĠKh ách +à¹Ģà¸Ļืà¹īà¸Ń หา +ĠYaz ı +è²· ãģ£ãģ¦ +Ðł Ðķ +à¹Ģà¸ŀิà¹Īม à¸Ĥึà¹īà¸Ļ +สม à¸ļู +สมà¸ļู รà¸ĵà¹Į +Ġм иÑĢов +×Ĵ ׳×Ļ×Ŀ +ĠÄij ức +à¸Ń ารà¹Į +ص اص +ãģĬ ãĤĪ +ãģĬãĤĪ ãģ³ +ÃªÌ ī +ĠاÙĦÙħؤ تÙħر +ĠاÙĦÙħر ØŃÙĦØ© +สà¸Ńà¸ļ à¸ĸาม +Ġà¸Īาà¸ģ à¸Ļัà¹īà¸Ļ +Ġت عد +ãģĿãģ® ãģŁãĤģ +Ġkh áng +à¸Ļ ิà¸Ķ +ãĥĬ ãĥ³ +ëĦ¤ ìļĶ +ĠاÙĦ اØŃت +ĠاÙĦاØŃت ÙĦاÙĦ +ìļ ķ +Ġмод ели +ĠпÑĢоÑĨ енÑĤ +à¸ŀวà¸ģ à¹Ģรา +Ġ×Ķצ ×ĵ +Ġ×Ķצ×ĵ ×ĵ×Ļ×Ŀ +ständ e +׳ ×Ĵר +Ġdot yc +Ġdotyc zÄħ +ĠdotyczÄħ ce +ĠÅĽ wiÄĻt +×ŀר ×Ķ +ãģĻãģĶ ãģĦ +ãĥĩãĤ£ ãĥ³ãĤ° +à¸ģาร สรà¹īาà¸ĩ +ë Ĥ¬ +Ġì°¸ ìŬ +Ñģ Ñħ +ÑģÑħ ем +ÙħÙĪ Ø³ +Ġn ấu +Ġ׾×ŀ×¢ ׾×Ķ +à¹Ģà¸Ľ à¹īา +à¹Ģà¸Ľà¹īา หมาย +Ġmù i +ائ ز +íĽ Ī +×Ĺ×ij ×ķר×Ķ +à¸ľà¸¹à¹ī à¹ĥà¸Ĭà¹ī +Ġpa ź +Ġpaź dzi +Ġpaździ ern +Ġpaździern ika +ลà¸ĩ à¹Ħà¸Ľ +ÙĤ اع +Ġch áºŃm +Ġözellik leri +ĠÄIJ o +ĠÄIJo Ãłn +ж ение +Ġh ẳ +Ġhẳ n +ĠaÅŁ k +ï½ į +ãĥij ãĤ¹ +×Ķ×ķר ×IJ×ķת +ĠÅ » +ĠÅ» y +×ŀ×ĸ ׾ +ĠÑĥ кÑĢа +ĠÑĥкÑĢа ин +à¹Ģà¸Ĭ ิ +à¹Ģà¸Ĭิ à¸į +Ðł Ðĺ +ĠzwiÄħz ku +×Ķ×Ĺ׾×ĺ ת +ãĤĵãģ§ãģĻ ãĤĪãģŃ +ãģ¦ ãģĬãĤĬ +лож иÑĤÑĮ +×ŀ ×ķ׳×Ļ×Ŀ +ฮ ิ +ì° ¬ +ĠاÙĦÙħØ´ ترÙĥ +ĠdÃ¼ÅŁ ük +аг енÑĤ +ĠاÙĦØ£ سبÙĪØ¹ +ĠÙĤ رÙĬب +ин д +инд ив +индив ид +индивид Ñĥ +индивидÑĥ алÑĮн +för der +Ġseç en +Ġseçen ek +Ġét ant +ĠлÑİб им +каз ÑĭваеÑĤ +ว ิà¸Ļ +Ġ×Ķ×ij ×IJ×Ļ×Ŀ +Ġд ов +Ġдов олÑĮ +ĠдоволÑĮ но +×¢×ĵ ×Ļ×£ +Ġok re +Ġokre ÅĽ +ĠokreÅĽ lon +Ġت رÙĬد +à¹Ģมืà¹Īà¸Ń วัà¸Ļà¸Ĺีà¹Ī +ãĤĪ ãģĭãģ£ãģŁ +Cum h +Cumh ur +Cumhur ba +Cumhurba ÅŁ +CumhurbaÅŁ kan +CumhurbaÅŁkan ı +Ġn ợ +à¸ľà¸¹à¹ī à¹Ģลà¹Īà¸Ļ +Ġcompl ète +à¹Ģà¸ŀ ศ +د ÙIJ +Ġdü z +Ġdüz ey +ãģ§ãģĤãĤĭ ãģĵãģ¨ +ext érieur +× ³ +Ġinform ação +ãĤ¯ãĥª ãĥĭãĥĥãĤ¯ +ĠPub li +ĠPubli é +ר ×ķ×ĵ +à¸Ħวาม à¸Ľà¸¥à¸Ńà¸Ķà¸łà¸±à¸¢ +ĠØ£ÙĬ ض +ĠØ£ÙĬض Ùĭا +ت سبب +ãģ¤ ãĤĤãĤĬ +из ма +à¸Ĥึà¹īà¸Ļ à¹Ħà¸Ľ +Ùĥ ÙIJ +ÙĦ ÙĪÙħ +Ġש צר +Ġשצר ×Ļ×ļ +ãģ¯ ãĤĤãģ¡ãĤįãĤĵ +Ġк ан +Ġкан ал +ãģ«ãģª ãģ£ãģ¦ãģĦãģ¾ãģĻ +ĠاÙĦØ£ Ùĥثر +ت اØŃ +ÙĨت Ùĩ +ÙĨتÙĩ اء +ا ÙĪÙĬØ© +ĠBug ün +н Ñģкого +à¸Ķ à¹Īวà¸Ļ +é volution +ãģ£ãģ¦ ãģĦãģ¾ãģĹãģŁ +ãĤ ħ +ĠV ương +à¸łà¸²à¸ŀ ย +à¸łà¸²à¸ŀย à¸Ļ +à¸łà¸²à¸ŀยà¸Ļ à¸ķรà¹Į +Ġ×Ķ ×¦×ľ×Ļ×Ĺ +ĠاÙĦإسÙĦاÙħ ÙĬ +ÙĦÙĬ ب +Ġed ição +ÑģÑĤÑĢ ÐµÐ» +Ġkh úc +ÙĨÙħÙĪ Ø° +ÙĨÙħÙĪØ° ج +׾ צ×Ķ +ÑģÑĤав ил +à¸ĸ า +สรà¹īาà¸ĩ à¸Ħวาม +ãģĦ ãģ£ãģ± +ãģĦãģ£ãģ± ãģĦ +ÑģÑĤав лен +ĠاÙĦ ÙĤدس +Ġng ược +ب Ø® +ส หร +สหร ั +สหรั à¸IJ +ĠØ£ غ +Ġأغ سط +Ġأغسط س +ãģĨ ãģ¾ +ãģĨãģ¾ ãģı +ĠêµŃ ìłľ +ØŃض ار +Ġd ừng +æĬ¼ ãģĹ +ت ÙĪØ§ +تÙĪØ§ جد +ש×ŀ ×Ĺ×Ķ +ãģı ãĤĵ +Ġ×ij×¢ צ +Ġ×ijעצ ×Ŀ +×ŀ ׳×Ļ×ķת +×ķ ×Ļ×ĵ +×ķ×Ļ×ĵ ×IJ×ķ +à¸Ĭ ิà¸ĩ +Ġprac ÄĻ +Ġз аÑĤ +ĠзаÑĤ ем +ĠìŀIJ ìľł +Ġì¤ Ģ +Ġì¤Ģ ë¹Ħ +Ġb áºŃ +ĠbáºŃ c +Ġ×Ķ×ŀ צ×ij +ĠÙĤ ÙĬÙħØ© +à¹Ģà¸Ń à¹Ģà¸Ĭ +à¹Ģà¸Ńà¹Ģà¸Ĭ ีย +Ġperch è +ĠاÙĦع سÙĥر +ĠاÙĦعسÙĥر ÙĬØ© +ج ÙĬب +ëŀ µ +Ùħ Ùĩر +ÙħÙĩر جاÙĨ +Ùħ راÙĥ +ÙħراÙĥ ز +Ġод нако +à¸Ķี à¹Ĩ +Ġצ פ×ķ +Ġkullan ılan +Ġк ино +ãĥĨãĤ£ ãĥ³ãĤ° +ĠGi Ỽi +ت ÙĪØ² +تÙĪØ² ÙĬع +ย ิà¸Ļ +ยิà¸Ļ à¸Ķี +Ġc Åĵur +ĠiÅŁ aret +Ġ×ij×¢ ×ĸר +Ġ×ij×¢×ĸר ת +Ġп аÑĨи +ĠпаÑĨи енÑĤ +ãģ¿ãģŁãģĦ ãģ§ãģĻ +в ез +ли на +од е +Ġ×IJ×ķת ף +dıģ ınız +ĠÐIJ в +ĠÐIJв ÑĤоÑĢ +ï¼ ® +ĠC ần +ĠاÙĦا Ø® +ĠاÙĦاخ بار +Ġê±° ìĿĺ +Ġat enção +Ġgeld iÄŁi +ãĤª ãĤ¹ +ãĤªãĤ¹ ãĤ¹ +ãĤªãĤ¹ãĤ¹ ãĥ¡ +ев Ñĭе +кÑĢÑĭ л +à¹Ģà¸Ĭ ียà¸ĩ +à¹Ģà¸Ĭียà¸ĩ à¹ĥหมà¹Ī +Ġmar ço +ĠاÙĦÙħ ادة +Ġг ол +Ġsprzeda ży +Ġíķ´ ê²° +ĠÐķ го +ê¹ Ģ +Ġ׾ק×ij׾ ת +ĠاÙĦÙģ ÙĨاÙĨ +Ġcomunic ación +à¹Ģสà¹īà¸Ļ à¸Ĺาà¸ĩ +íĺ ¹ +à¸Ĭ ำ +à¸Ĭำ ระ +Ġ׼ ×IJ×ŀ +Ġ׼×IJ×ŀ ×ķר +à¸Ĭ à¹Īาà¸ĩ +ز Ùĩر +Ġklient ów +ива ÑİÑĤ +ан г +׳ ×ļ +Ġg á»įn +Ãľ R +ìĺģ ìĥģ +Ġغ زة +ìĿĮ ìĿĦ +Ġbez po +Ġbezpo ÅĽ +ĠbezpoÅĽ redni +ĠاÙĦÙħ ÙĪØ§ +ĠاÙĦÙħÙĪØ§ Ø·ÙĨ +ĠاÙĦÙħÙĪØ§Ø·ÙĨ ÙĬÙĨ +ãĤĮ ãģ¾ãģĻ +ĠмаÑĤ Ñĩ +×IJ ×ķף +Ġر سÙħÙĬ +ĠÑįк он +ĠÑįкон ом +ĠÑįконом иÑĩеÑģк +ãĥľ ãĥ¼ +Ġд иÑĢ +ĠдиÑĢ ÐµÐºÑĤоÑĢ +ĠÑģк оÑĢо +à¸ļ ำ +à¸ļำ ร +à¸ļำร ุà¸ĩ +ĠÑĦ ÑĥÑĤ +ĠÑĦÑĥÑĤ бол +Ġ×IJ ×Ļ׾ +Ġì¤ij êµŃ +ìľ ¤ +eÄŁ e +à¹Ħ à¸ģà¹Ī +tra î +traî n +ĠÑĤ ÑĢÑĥб +à¹Ģà¸ļ ื +à¹Ģà¸ļื à¹īà¸Ńà¸ĩ +à¹ģม à¸Ļ +ĠتØŃ دÙĬØ« +Ġ׼ עת +ØŃ اسب +lı ÄŁa +×§×Ļ ×Ļ×ŀ×Ļ×Ŀ +оÑģÑĤ ÑĮÑİ +à¸Ŀ ั +à¸Ŀั à¹Īà¸ĩ +Ø´ غÙĦ +ìĽ ¹ +Ġкажд ого +Ġbölüm ü +หà¸Ļ ี +Ġistedi ÄŁi +Ġtr ưng +ãĥ Į +ฮ à¸Ń +Ø£ÙĨ Ø´ +Ø£ÙĨØ´ طة +ĠاÙĦÙħ سÙĬ +ĠاÙĦÙħسÙĬ ØŃ +ลัà¸ģษ à¸ĵà¹Į +Ġn á»Ńa +à¸Ĺีà¹Ī à¸ķà¹īà¸Ńà¸ĩà¸ģาร +ÑĪ ÐµÐº +л Ñij +Ġש ×Ļ×Ķ +Ġש×Ļ×Ķ ×Ļ×Ķ +Ġkhu ôn +ĠÑĤÑĢеб ованиÑı +Ġ×ľ×¢ ×ĸ×ķר +ĠاÙĦع Ùħر +ราà¸Ħา à¸ĸูà¸ģ +ÙĩÙı ÙħÙĴ +ü st +üst ü +Ġден ег +Ġn ạ +à¸Ĥà¸Ļ ม +Ġбл аг +Ġблаг од +Ġблагод аÑĢ +ĠблагодаÑĢ Ñı +Ø¥ سÙĦاÙħ +à¸Ļิ ว +çŁ¥ ãĤīãģªãģĦ +Ø« ÙĤØ© +Ġг олоÑģ +×IJ×ķר ×Ĺ +Ġtr ứng +Ġод ном +ĠkoÅĦ cu +Ġ×ķ רק +Wi ÄĻ +WiÄĻ cej +Ġ×IJ ×Ļ׼×ķת +Ġ×IJ×Ļ׼×ķת ×Ļ +Ñģ оÑģ +Ġje żeli +以ä¸ĭ ãģ® +å°ı ãģķ +å°ıãģķ ãģª +олог ии +Ġоб ÑģлÑĥж +ĠобÑģлÑĥж ива +Ùĥت ابة +Ġê´Ģ ìĭ¬ +×¢ ש×Ļר +Ġaras ındaki +ĠÑĢай она +ÙĪØ§ جب +Ġ×ij×Ĺ×Ļ ×Ļ +íķ´ ì£¼ +Ġg óc +ай л +ĠT ình +æļ® ãĤī +æļ®ãĤī ãģĹ +æĻĤ ãģ«ãģ¯ +ĠгоÑĢод е +Ġ׼×IJ ×Ļ׾ +Ġ׼×IJ×Ļ׾ ×ķ +ĠC á»Ļng +ãģ©ãģĨ ãģĹãģ¦ãĤĤ +×Ĺ ×ķ×£ +تØŃ رÙĥ +ĠÑģлов ам +à¸Īะ à¸Ĭà¹Īวย +ĠاÙĦÙħست ÙĤبÙĦ +ÙĤ ض +ÙĤض ÙĬ +×ijס ×ķפ +×ijס×ķפ ×ķ +iÄĻ Äĩ +ĠY ıl +Ø´ ÙĬØ® +à¸Ħุà¸ĵ à¸Īะ +ש×ŀ ×ķת +Ġت عرض +Ġanál ise +ĠÑģоб иÑĢа +à¹Ģà¸ŀ à¸Ĭ +à¹Ģà¸ŀà¸Ĭ ร +Ġв ели +Ġвели к +สั à¹īà¸Ļ +Ġpop ulação +รà¹Īวม à¸ģัà¸Ļ +×Ĺ ×ŀ +×Ĺ×ŀ ×Ļש×Ļ +ס ×Ļס +åĨħ ãģ§ +Ġsob Äħ +ĠY ay +ĠYay ın +ãĥ¡ ãĥĭãĥ¥ãĥ¼ +ĠпÑĢедоÑģÑĤав лÑı +ãģł ã썿ĢĿãģĨ +Ġê³ł ê°Ŀ +Ġод ним +à¹ĥà¸Ļ à¹Ģรืà¹Īà¸Ńà¸ĩ +Ġs á»ķ +ĠÐĹ Ð´ÐµÑģÑĮ +Ġизмен ениÑı +ĠìĿ¼ ìĿĦ +ãģªãģ® ãģł +клад Ñĭва +ÑĢ Ð¼Ð° +Ġ×ķ×ij ׼׾ +تأ ÙħÙĬÙĨ +ĠпÑĢи ÑıÑĤ +ĠпÑĢиÑıÑĤ н +Ùħ Ùħار +ÙħÙħار سة +ãģ¨ãģª ãģ£ãģ¦ +Ġج ÙħÙĬÙĦ +Ġì§ Ī +Ġì§Ī 문 +Ġquest ão +i é +ié ndo +หà¹īà¸Ńà¸ĩ à¸ŀัà¸ģ +ãĥij ãĥ¼ãĥĪ +ÑĤвеÑĢж да +н Ñģкой +з ал +มุ à¹Īà¸ĩ +á» Ĭ +Ġ×Ķ×IJ×Ĺר ×ķ׳×Ķ +ĠTh ư +주 민 +ĠاÙĦع ب +év én +évén ement +ÙĤÙĪ Ø§Ø¹Ø¯ +د Ùı +ĠìķĬ ìĬµëĭĪëĭ¤ +Ġë³´ 기 +Ġyapıl ması +à¹Ģร าà¸ģ +à¹Ģราà¸ģ à¹ĩ +ØŃ ذر +ÙĤ صر +ãģ¦ãģĹãģ¾ ãģĦãģ¾ãģĹãģŁ +Ġà¹Ģà¸Ľà¹ĩà¸Ļ à¸ķà¹īà¸Ļ +ãģ¨ ãģ« +ãģ¨ãģ« ãģĭ +ãģ¨ãģ«ãģĭ ãģı +н ÑĨе +зв Ñĥк +ãģĹãĤĪãģĨ ãģ¨ +ĠاÙĦصØŃ ÙĬØ© +Ġש×Ķ ×Ļ×ķ +ĠDi ÄŁer +ÙĤÙĦ ÙĤ +ãĤ¸ãĥ£ ãĥ³ +Ġr á»Ŀi +Ġл еÑĩ +ĠлеÑĩ ениÑı +تب اد +تباد ÙĦ +צ פ×Ķ +à¸Ħวาม à¹Ģหà¹ĩà¸Ļ +ĠØ´ ب +Ġشب ÙĥØ© +ר ×Ļ×§ +Ùħ عد +Ùħعد ات +dıģ ında +Ġ×ijש ׳×Ļ×Ŀ +Ġ×Ķ ×Ļשר×IJ׾ +Ġ×Ķ×Ļשר×IJ׾ ×Ļת +Ġsı nav +׳צ ×Ļ×Ĵ +วัà¸ķ à¸ĸุ +ĠاÙĦبر ÙĦÙħ +ĠاÙĦبرÙĦÙħ اÙĨ +t ivitÃł +ãĤĵãģł ãĤįãģĨ +×§×Ļ ×Ļ×ŀ +ÙĦÙĬ Ùĥ +ĠÄij ò +ĠÄijò i +ĠÐĺн ÑĤеÑĢ +ĠÐĺнÑĤеÑĢ Ð½ÐµÑĤ +ãģ«ãģ¨ãģ£ãģ¦ ãģ¯ +ãģ£ ãģĵ +×§ ×ķס +ست ØŃÙĤ +æķĻ ãģĪãģ¦ +ãĥĢ ãĥ¡ +ĠÙħÙĨ زÙĦ +à¹Ģà¸ĭ à¹ĩà¸Ļ +使 ãģĪãĤĭ +è¦ĭ ç©į +è¦ĭç©į ãĤĤãĤĬ +Ø£ Ùģ +Ø£Ùģ Ùĥار +Ġиг ÑĢов +ĠигÑĢов Ñĭе +Ġm ÄĻż +ĠmÄĻż czy +ĠmÄĻżczy zn +ĠاÙĦØŃ ÙĤÙĬÙĤÙĬ +ع بر +׼×ķ׾ ׳×ķ +íĿ ¥ +×ŀ×IJ ×ķ×Ĺר +خت ص +ãĥŀ ãĥŀ +Ġ×IJ×Ĺ ×ķ×ĸ +í ĮĢ +Ġr á»iji +Ġв ÑĤоÑĢ +ĠвÑĤоÑĢ Ð¾Ð¹ +Ġl ẫn +пÑĢ Ð¾Ð¼ +пÑĢом ÑĭÑĪ +пÑĢомÑĭÑĪ Ð»ÐµÐ½ +пÑĢомÑĭÑĪлен н +ĠоÑĤноÑĪ ÐµÐ½Ð¸Ñı +Ġs ứ +Ġм обилÑĮ +ĠмобилÑĮ н +ĠÑįÑĤ омÑĥ +Ġt ạp +ĠìĤ¬ ê±´ +ĠìķĮ 볤 +Ùĥ Ùı +ÙĥÙı ÙħÙĴ +Ġ×§ ×ķר×Ķ +ĠÑĦ иÑĢ +ĠÑĦиÑĢ Ð¼ +Ġsık ıntı +׳ ׼ +׳׼ ×ķף +ÙĪÙĦÙĪØ¬ ÙĬ +ØŃ اÙĨ +Ġlo ạn +Ġ×IJ׾ ×£ +Ġm ắn +abh äng +abhäng ig +ĠÑĥÑĢов нÑı +Ġ׾×ij×ĵ ×ķ×§ +ÙĬ ÙħÙĨ +lay ın +Ġh ải +Ġзав од +ĠìķĦ 주 +สà¸ĸ า +สà¸ĸา à¸ļัà¸Ļ +Ġgüven lik +à¹Ģà¸Ķ à¹Īà¸Ļ +×ij×ĵ ×§ +Ġë Ī +ĠëĪ Ħ +ĠëĪĦ 구 +éĩįè¦ģ ãģª +รà¸Ńà¸ĩ รัà¸ļ +sch lie +schlie ÃŁen +Ġìĸ ¼ +Ġìĸ¼ ë§Ī +Ġìĸ¼ë§Ī ëĤĺ +ÑĤи ки +íķľëĭ¤ ê³ł +ãģłãģ£ãģŁ ãĤī +Ġ×Ķ ×Ļ×ĺ×ij +ãģªãģijãĤĮãģ° ãģªãĤīãģªãģĦ +â Ì +Ã¢Ì £ +Ġph ạt +ak Ä±ÅŁ +ãģ¦ãģĹãģ¾ ãģĦãģ¾ãģĻ +à¹Ģà¸ĭ à¹ĩ +ĠС егоднÑı +Ġinsan ların +Ġdévelop pe +ת פר +תפר ×Ļ×ĺ +اÙĨت شار +ê° ij +Fran çois +Ø£ÙĦ ع +Ø£ÙĦع اب +ãĤĴ è¶ħ +ãĤĴè¶ħ ãģĪ +Ġê°Ļ ìĬµëĭĪëĭ¤ +ãĤ³ ãĥ¬ +ĠмеÑģÑı ÑĨев +íĮ ħ +ĠاÙĦج اÙħعة +ìĿ¸ íĦ° +ìĿ¸íĦ° ëĦ· +×ĵר ×ķש +ĠÙĪØ£ شار +ĠпÑĢав ила +ãģĿãģĵ ãģ« +×Ĺ ×ŀ×ĵ +à¹Ģหà¸ķุ à¸ģารà¸ĵà¹Į +Ġê²½ íĹĺ +ãģ¶ ãĤĬ +׾ ש +׾ש ×ķף +à¹Ģ à¸ĸ +ĠDo ÄŁu +ĠиÑģполÑĮзов ание +Ġçoc uÄŁu +магазин е +ĠÄiji á»ĥn +Ġas lı +Ġaslı nda +Ġdoen ça +Ġس اع +Ġساع ات +ĠиÑģполÑĮзов аниÑı +ר ×ķצ×Ļ×Ŀ +ĠзнаÑĩ иÑĤ +ĠÑĢаР¼ +ĠÑĢам каÑħ +ê±° 리 +Ġп ÑĭÑĤа +ãĥģ ãĥ³ +Ġпо Ñģк +ĠпоÑģк олÑĮ +ĠпоÑģколÑĮ кÑĥ +Ø¥ بر +إبر اÙĩ +إبراÙĩ ÙĬÙħ +ĠÑĤÑĢ ÐµÑħ +ĠGen ç +س ÙĪÙģ +Ġve ÃŃculo +ĠNg ân +ĠоÑĩеÑĢ ÐµÐ´ÑĮ +à¸Ħร ึà¹Īà¸ĩ +×IJ ×ij×Ļ +à¸ķ à¹īม +ãĤĴè¡Į ãģĦ +ĠاÙĦساب ÙĤØ© +на ÑĨи +наÑĨи она +наÑĨиона лÑĮн +Ġgest ión +ت ÙĤد +ĠاÙĦبÙĬ اÙĨ +ĠاÙĦبÙĬاÙĨ ات +ĠاÙĦ اÙĨتخاب +ĠاÙĦاÙĨتخاب ات +à¹Ģà¸Ĭ à¹Īา +×ĵ ×IJ×Ĵ +Ġ׾×Ĵ ×ŀר×Ļ +Ġت ØŃتاج +Ġth ôn +à¸ķ à¹īà¸Ńà¸Ļ +à¸ķà¹īà¸Ńà¸Ļ รัà¸ļ +女 ãģ® +女ãģ® åŃIJ +Ġth ợ +Ø· ØŃÙĨ +ารà¹Į à¸Ķ +ת ×ŀ×Ļ×ĵ +ĠÑģам Ñĭм +Ġìĭľ íĸī +Ø¥ صد +إصد ار +ĠNgh á»ĩ +ìķ ķ +س ئ +سئ ÙĦ +à¸Ń าร +à¸Ńาร ม +à¸Ńารม à¸ĵà¹Į +à¹ģ ฮ +׳×ĺ ׾ +Ġì¢ĭ ìķĦ +×ķ׾ ׾ +Ġ×ij ×Ľ×ª×ij +ãĤ« ãĥ© +צע ×Ļר×Ļ×Ŀ +تعب ÙĬر +Ġ×ŀ קר×Ķ +ĠÑĦак ÑĤоÑĢ +Ġت ÙħاÙħ +ĠتÙħاÙħ ا +ëį ķ +Ġv ưá»Ŀ +Ġvưá»Ŀ n +Ġd Ä±ÅŁÄ± +ãģĦ ãģ¡ +Ġ׾ק ׳×ķת +ĠاÙĦع ÙĦاÙĤات +п Ñĥб +пÑĥб ли +Ø¥ ÙĬÙħ +Ø¥ÙĬÙħ اÙĨ +à¸Ńำ à¸Ļา +à¸Ńำà¸Ļา à¸Ī +åIJ« ãģ¾ãĤĮ +ãĤĭ ãģŁãĤģãģ« +ס ×Ĵ +ס×Ĵ ׳×ķף +تØŃ دÙĬ +Ġaup rès +ĠاÙĦج Ùĩا +ĠاÙĦجÙĩا ز +Ġ×ŀ ת×Ĺת +ен нÑĥÑİ +Ġз им +à¸ģา à¹ģà¸Ł +Ġ×ijת ×ķר +Ġngh è +Ġnghè o +ĠÐĽ Ñİ +ĠÐĽÑİ Ð± +תק צ×Ļ×ij +×ŀ×¢ ש×Ķ +ĠاÙĦبÙĬ ت +צ ×Ļפ +ĠобÑıз ан +ĠM á»Ĺi +ĠТ ÑĥÑĢ +ĠÙĪØ¨ اÙĦت +ĠÙĪØ¨Ø§ÙĦت اÙĦÙĬ +Ġdéc ision +Ġب د +Ġبد أت +Ġc ục +Ġb ask +Ġbask ı +Ġhat ırl +Ġhatırl a +å°ı ãģķãģĦ +Ġgerçek ten +à¸ľ ัà¸ģ +åı¯èĥ½ ãģª +×ŀ×IJ ס +Ġcr ÃŃtica +ĠìĿĺ ìĽIJ +عÙĤ ÙĪØ¯ +×ĺ ׼׳ +×ĺ׼׳ ×ķ׾×ķ×Ĵ×Ļ×Ķ +è¨Ģ ãģĪãģ° +ĠÙĤ ÙĨا +ĠÙĤÙĨا Ø© +ĠìĿ´ê²ĥ ìĿĢ +ت صر +à¸Ł ัà¸Ļ +ĠÑĢе ÑĨеп +ĠÑĢеÑĨеп ÑĤ +ĠبÙĨ Ù쨳 +ÑĢо ÑĪ +ĠмаÑĢ ÑĤа +Ġson ras +Ġsonras ı +×ķ×ij ש +ãĥª ãĤ¹ãĤ¯ +ĠFranç ais +á» ļ +ê° Ķ +Ġ×Ķ×ijר ×Ļת +פ ×Ļצ +פ×Ļצ ×ķ×Ļ +ĠÙĦÙħا ذا +ĠÐļи ев +ĠÑģ мÑĭÑģл +ê¸Ī ìľµ +ãĤ·ãĥ£ ãĥ« +ãĥ© ãĤ¤ãĥĪ +ìĽ ĥ +×ŀ ×Ĺר +ãĨ į +Ġkullan ım +Ġ×IJצ׾ ׳×ķ +Ġt Ãłn +ãĥı ãĥ¼ +ãģ¨ ãģ¨ãĤĤ +ãģ¨ãģ¨ãĤĤ ãģ« +ÑĢ ÐµÐ³ +ÑĢег и +ÑĢеги он +ãģªãģı ãģªãĤĭ +Ġch ảy +Ġج ÙĩØ© +ÅĦsk iej +à¸Ńี à¹Ģม +à¸Ńีà¹Ģม ล +ãģį ãģ£ãģ¨ +ĠìĺĪ ìĤ° +Ġkit abı +Ġedu cação +Ġbul uÅŁ +олог иÑı +Ġкон кÑĢ +ĠконкÑĢ ÐµÑĤ +×Ĵ ×Ļר +ĠпÑĢед лаг +ĠпÑĢедлаг аеÑĤ +ĠY ên +Ġíķľ ë²Ī +Ġ×ŀ ר׼×ĸ×Ļ +à¹Ģà¸Ľà¸´à¸Ķ à¹Ģà¸ľà¸¢ +ÑĤвеÑĢ Ð´ +ĠH á»ĩ +ĠÐĵ ÑĢ +à¸Ŀ à¹īา +×Ķ ×©×§ +×Ķשק ×¢×Ķ +Ġна Ñĥк +ìłIJ ìĿĦ +Ġн елÑĮ +ĠнелÑĮ з +ĠнелÑĮз Ñı +г ин +ĠB öl +ĠBöl ge +Ġв ла +Ġвла ÑģÑĤи +à¹Ģà¸Ļ à¹ĩ +à¹Ģà¸Ļà¹ĩ à¸ķ +ê³ ¨ +Ġö ld +Ġöld ür +׼׳ ×¢ +ĠاÙĦÙĩ ÙĬئة +ت ارÙĬØ® +ĠÐij ÑĢ +ĠÑģ мож +ĠÑģмож еÑĤе +ĠL úc +à¹Ħà¸Ľ à¸ĸึà¸ĩ +ĠBakan ı +Ġerklä rt +ĠÐIJ на +Ġsc ène +åķı ãģĦ +åķıãģĦ åIJĪãĤıãģĽ +ÙħÙĩ ÙĨد +ÙħÙĩÙĨد س +Ġн азвание +ив аниÑı +ãĤĴ å¤īãģĪ +ä»ĺãģį åIJĪ +ãĥij ãĤ½ +ãĥijãĤ½ ãĤ³ãĥ³ +æĺİ ãĤī +æĺİãĤī ãģĭ +à¹Ģà¸Ńà¸ģ สาร +à¹Ģà¸ģิà¸Ļ à¹Ħà¸Ľ +л еп +ãģĹãģŁ ãĤĤãģ® +ĠC âm +ĠCâm ara +×§×ķ׾ ׳×ķ×¢ +Ġ×ij×Ĵ ×Ļף +Ġoc zy +Ġoczy wiÅĽcie +att ivitÃł +ãĥĵ ãĥ¥ãĥ¼ +Ġeduc ación +İ YE +ê¹Į ìļĶ +ãĤ¨ ãĥªãĤ¢ +н еÑģÑĤи +Ġm óg +Ġmóg ÅĤ +Ġ×§×ĺ ׳×Ļ×Ŀ +ĠPr ä +Ġ×ľ×¢ ×ij×ķר +بÙĨ Ùī +з ол +зол оÑĤ +Ġwn ÄĻtr +ĠwnÄĻtr z +Ġconstr ução +รัà¸ļ รà¸Ńà¸ĩ +س جÙĨ +Ġ×§ ×ķ׳ +ס ×Ļפ×ķר +ĠÙħ دÙī +رض Ùī +п лав +ï¼ ¥ +Ġil a +Ġila ç +ãĤĭ ãģ¹ãģį +ĠÙħ ÙĪÙĤÙģ +à¸ģร ุ +à¸ģรุ à¸ĵา +chodzÄħ c +ĠÑĤÑĭ Ñģ +Ðķ вÑĢо +ĠÙĬ ØŃدث +ãĥ¡ ãĤ¤ãĥ³ +ĠاÙĦص ØŃÙĬ +ĠÐĶ Ð°Ð½ +دع اء +ãĤ´ ãĥ¼ãĥ« +ש ×ł×ª×Ļ +×©×ł×ª×Ļ ×Ļ×Ŀ +à¸Ķà¹īวย à¸ģัà¸Ļ +Ġol acaģı +Ġ×ij ×ŀ×Ĺ×Ļר +×Ķ ×§ +×Ķ×§ ×ŀת +ãĥ¢ ãĥİ +ĠçalÄ±ÅŁ tı +Ġjó venes +ãģĦãģı ãĤī +ĠÙħ عدÙĦ +ĠC Å©ng +ĠSeg ún +Ġdönem de +Ġ׾ ×Ļ×ĵ×Ļ +ãģį ãģ¡ +ãģįãģ¡ ãĤĵ +ãģįãģ¡ãĤĵ ãģ¨ +Ù쨱 ÙĨس +Ù쨱ÙĨس ا +åIJij ãģį +Ġcamp aña +ĠÑģам оÑģÑĤоÑı +ĠÑģамоÑģÑĤоÑı ÑĤелÑĮно +á» Ģ +ÙĤ ÙĪØ§ +س ÙĦاØŃ +à¸ģระ à¹ģ +à¸ģระà¹ģ ส +ĠполÑĮз Ñĥ +n qu +nqu ête +รà¹Īวม à¸ģัà¸ļ +ëĬIJ ëĥIJ +à¸Ĺีม à¸Ĭาà¸ķิ +Ġyıll ık +ìĬ ¬ +ĠØ£ صØŃاب +ill é +Ġdó la +Ġdóla res +Ġк ож +Ġкож и +ล à¹īà¸Ń +à¹Ģรีย à¸ļร +à¹Ģรียà¸ļร à¹īà¸Ńย +à¹Ģà¸ŀ ิ +à¹Ģà¸ŀิ à¹Īà¸ĩ +ÑĢиÑĤоÑĢ Ð¸ +Ġí ijľ +Ġíijľ íĺĦ +ĠпеÑĢ ÐµÐ² +ĠпеÑĢев од +פ×Ĵ ×Ļ×¢×Ķ +ĠdeÄŁerlendir me +Ùģ Ø§Ø¦ +ĠвÑĭ год +ınız ı +×ķ׼ ×Ļ×Ĺ +ĠдоÑģÑĤ иг +Ġng Ãłn +æĢĿ ãģ£ãģŁ +ĠÐķ ÑģÑĤÑĮ +ĠاÙĦر غÙħ +ĠzwiÄħz ane +رب Ø· +à¸Ļ ึà¸ĩ +Ġ׾×Ĺ ×ķ×§ +Ġszczeg óln +Ġszczególn ie +Ġبا ستخداÙħ +ĠfÃŃs ico +×¢ ס +עס ×ķ×§ +سÙĦ ÙĪÙĥ +Ġا ØŃد +Ñĩ ÑijÑĤ +×ĸ׼ ×Ķ +Ġl á»ĩnh +ĠÙĪ ØŃت +ĠÙĪØŃØª Ùī +à¸Ħวาม สามารà¸ĸ +à¸Ńยูà¹Ī à¹ģลà¹īว +à¸ģาร à¹Ģà¸Ķิà¸Ļà¸Ĺาà¸ĩ +تخ ذ +צ×Ļ ×ķ×ĵ +ĠاÙĦØ£ س +ĠاÙĦأس ÙĩÙħ +Ġt á»ĩ +ãģ£ãģ¦ ãģĦãģ¦ +สร ุ +สรุ à¸Ľ +Ġком ÑĦ +ĠкомÑĦ оÑĢÑĤ +ìĺ¤ ëĬĶ +ĠÑĢаз в +ĠÑĢазв ива +л анд +h änge +ĠبÙĨ سبة +à¹Ģà¸Ĥ ียว +עצ ×Ŀ +Ġ׾ ×ľ×Ľ×ª +Ñģо ÑĨиалÑĮн +Ġëĭ¤ìĿĮ ê³¼ +Ġרש ×ķ×ŀ +×ŀר ×Ĺ×ij +س ÙĤØ· +Ġalan ı +ĠÄij á»ĩ +é£Łãģ¹ ãĤĭ +à¸Ķ ึà¸ĩ +Ġgegen über +ĠبÙĩ ذÙĩ +à¸ĸืà¸Ń à¹Ģà¸Ľà¹ĩà¸Ļ +ëķ ħ +à¸Ħà¸Ļ à¹Ħà¸Ĺย +ãĤ¢ ãĤ¦ +ãĤ¢ãĤ¦ ãĥĪ +ศ ัà¸ģ +ศัà¸ģ à¸Ķิ +ศัà¸ģà¸Ķิ à¹Į +ÙĤÙĪ Ø§ÙĨ +ÙĤÙĪØ§ÙĨ ÙĬÙĨ +Ġhá»Ļ p +ãģªãģıãģª ãģ£ãģ¦ +Ġ×IJ ×ŀ׳ +Ġ×IJ×ŀ׳ ×Ŀ +à¹Ģà¸ķ ืà¸Ńà¸Ļ +ĠзавиÑģ им +ĠзавиÑģим оÑģÑĤи +ת ×Ļ×IJ +ת×Ļ×IJ ×ķר +å§ĭãĤģ ãģŁ +Ġng á»į +Ġngá»į t +íĴ į +ê³¼ ìŀ¥ +Ġb ại +ãģ§ãģį ãģ¦ +Ġcomeç ar +à¸Ľà¸£ าà¸ģ +à¸Ľà¸£à¸²à¸ģ à¸ı +Ġгод Ñĭ +м еÑģ +ĠاÙĦÙħست ÙĪÙī +ĠÑģам Ñĭе +л леÑĢ +ãģ£ãģ¦ãģĹãģ¾ ãģĦãģ¾ãģĻ +ãģ¨ãģ® ãģĵãģ¨ +bi ó +à¸ģล à¹Īà¸Ńà¸ĩ +ĠاÙĦز ÙĪØ¬ +ãģ«è¡Į ãģ£ãģŁ +à¸Ħà¹Ī à¸Ńà¸Ļ +à¸Ħà¹Īà¸Ńà¸Ļ à¸Ĥà¹īาà¸ĩ +ĠbaÄŁ l +ĠbaÄŁl ant +ĠbaÄŁlant ı +確 ãģĭ +確ãģĭ ãģ« +ãĥľ ãĥ¼ãĥ« +çµĤ ãĤıãĤĬ +ש ×ŀר +à¸Ĺีà¹Ī สามารà¸ĸ +ÙĦ زÙħ +д аеÑĤÑģÑı +รัà¸ļ à¸Ľà¸£à¸° +รัà¸ļà¸Ľà¸£à¸° à¸Ĺาà¸Ļ +å¤ī ãĤıãĤĬ +ï¼ ¢ +ĠìĺĪìĪĺ ëĭĺ +ãĤĪãģĨ ãģ¨ +มัà¸ģ à¸Īะ +ĠH ương +ÙĨ Ù쨰 +×ŀ×ĵ ×ĵ +ĠìĿ¸ ìłķ +Ñħод иÑĤÑĮ +ĠзавиÑģ иÑĤ +×ķ×ĵ ×Ļ×¢ +ãģĵãģ¨ãģĮ ãģĤãĤĬãģ¾ãģĻ +ع راÙĤ +سط ØŃ +à¸ģำ à¹Ħร +ëĵ¤ ëıĦ +×Ļצ ×Ļר×Ķ +ãģĨ ãģĵãģ¨ +ÙĦا ØŃÙĤ +ãģĦ ãĤĮãģ° +ĠиÑģполÑĮз ÑĥÑİÑĤ +ĠB ợi +Ġשק׾ ×Ļ×Ŀ +ÑĨи кл +ÐIJ Ðŀ +Ġ×ijש ׳×Ķ +ÙĨØ´ Ø· +Ġש ×Ļ׳×ķ×Ļ +Ġש×Ļ׳×ķ×Ļ ×Ļ×Ŀ +Ġpobl ación +ĠH ưng +ระ ว +ระว ัà¸ĩ +رÙĬاض Ø© +ر صد +تÙĤ ÙĦÙĬ +تÙĤÙĦÙĬ د +Ġülk em +Ġülkem iz +à¸Ĭ ะ +ãĤ¯ãĥª ãĥ¼ãĥł +èģŀ ãģĦãģŁ +Ġwa ż +Ġważ ne +ê±° ëĵł +ê±°ëĵł ìļĶ +×ŀ×IJ ×ij×§ +×Ĺ×ĵ ש×ķת +ĠW roc +ĠWroc ÅĤaw +ĠKü ltür +s ist +sist ência +×¢×ĸר ×Ķ +Ġg ương +รà¹īาà¸Ļ à¸Ħà¹īา +ĠÙĪØ£ ÙĪØ¶ØŃ +ánd ose +ãĤ· ãĥ¼ãĥ³ +×IJ׳ ר×Ĵ +×IJ׳ר×Ĵ ×Ļ×Ķ +ãģªãģĦ ãģ§ãģĻ +Ġkh á»§ng +Ġ문 ìĦľ +Ġ×ij ×ĵ×ijר +×ĵ ×Ļ×ķ +×ĵ×Ļ×ķ ×ķ×Ĺ +Ġré gl +ÙħÙĪ Ø§Ø¯ +об оÑĢ +обоÑĢ Ð¾ÑĤ +Ġ×Ķ ×ij׾ +Ġ×Ķ×ij׾ ×ķ×Ĵ +ØŃ اÙħ +ĠاÙĦع اص +ĠاÙĦعاص ÙħØ© +пеÑĢ Ð°ÑĤоÑĢ +ت Ø®ÙĦ +تخÙĦ ص +ãģŁãģł ãģĹ +ت سÙħ +à¹Ĥรà¸ĩ à¸ŀ +à¹Ĥรà¸ĩà¸ŀ ยา +à¹Ĥรà¸ĩà¸ŀยา à¸ļาล +ĠY ük +ĠYük sek +Ġש ׳×Ļת +Ġש׳×Ļת ף +liÄŁ e +Ġפ ת +Ġפת ×ķ×Ĺ +Ġbe ÄŁ +ĠbeÄŁ en +Ġ×ŀ ×ķר +Ġ×ŀ×ķר ׼×ij +Ġرس اÙĦØ© +íĨµ ìĭł +Ġaval ia +Ġavalia ções +Ġman h +Ġmanh ã +Ġìķ ŀ +Ġìķŀ ìľ¼ë¡ľ +ÙĤ تر +ÙĤتر ØŃ +à¹Ģà¸ģ ืà¸Ń +à¹Ģà¸ģืà¸Ń à¸ļ +Ġpropos é +Ø£ Ùħا +Ø£Ùħا ÙĥÙĨ +ĠÐŀ Ðŀ +ĠÐŀÐŀ Ðŀ +ÙħÙĤ ار +ÙħÙĤار ÙĨØ© +ëĦ IJ +ãģĦãģŁãģł ãģı +ÙĤ ÙĬÙĦ +Ġна ÑĪиÑħ +ãĤ« ãĥĥãĥĹ +×Ĺ׾ ת +Ġëĭ¤ ë§Į +à¸Ĺัà¹Īว à¹Ĥลà¸ģ +ãĥį ãĤ¿ +ØŃس اس +ãģ«ãģª ãĤĮ +ج ائ +جائ زة +é change +é conom +économ ie +Т Ðĺ +סת ׼׾ +à¸Ĺัà¹īà¸ĩ สà¸Ńà¸ĩ +ĠاÙĦØ® اÙħ +ĠاÙĦخاÙħ س +×§ ×ĺ×¢ +au waż +à¸ľà¸¹à¹ī à¸Ĭาย +à¹ģà¸Ľà¸¥ à¸ģ +åIJĮæĻĤ ãģ« +зн аниÑı +ãģĦãģŁãģł ãģįãģ¾ãģĹãģŁ +Ġ×ŀ×ij ׾×Ļ +à¸Ĥà¸Ń à¹ĥหà¹ī +ĠاÙĦت ربÙĬØ© +Ġdécou vert +Ġżyc iu +apr ès +Ġy ab +Ġyab anc +Ġyabanc ı +ĠbaÅŁ layan +ìĹĪ ëįĺ +Ġhes abı +Ġë§Į ìķ½ +ë§ Īëĭ¤ +ĠTh ánh +ãĥ´ ãĤ¡ +à¸Ľà¸£à¸±à¸ļ à¸Ľà¸£ +à¸Ľà¸£à¸±à¸ļà¸Ľà¸£ ุà¸ĩ +ĠM ặc +à¹Ģหà¸ķุ à¸ľà¸¥ +ĠÐij ез +Ġcapac itÃł +ÅĤe ÅĽ +ĠпÑĢе им +ĠпÑĢеим ÑĥÑīеÑģÑĤв +ĠÅļ wiÄĻt +Ġpubli é +×ŀ×¢ צ×ij +Ùħشار Ùĥات +à¸łà¸² ษ +à¸łà¸²à¸© ี +Ġdeux ième +ĠÙħØŃ اÙ쨏 +ĠÙħØŃاÙ쨏 Ø© +ĠSch ön +ï½ ¤ +Ġ×Ķ ×ij×¢ +Ġ×Ķ×ij×¢ ×Ļ×Ķ +ĠÙĪØ§ÙĦ ÙĦÙĩ +è¨Ģ ãģ£ãģŁ +à¸ķ à¹īาà¸Ļ +วร รà¸ĵ +à¸Ĺิ ศ +ĠbaÅŁ ına +Ġmog ÄĻ +ש ×Ļפ×ķר +ĠÙĪ Ø¹Ø¯ +ĠÙĪØ¹Ø¯ Ùħ +Ġhistó rico +Ġk ısı +ĠìĿ´ ê²Į +ĠPol ÃŃtica +ĠÑģиÑĤÑĥ аÑĨии +ĠkoÅĦ ca +×ij×ĵ ×Ļ×§×Ķ +ĠاÙĦسÙĬ ارات +ãģªãĤī ãģ° +ãĤµ ãĥ© +ãĤĭãģĵãģ¨ãģĮãģ§ãģį ãĤĭ +Ġdecis ão +×ķ ×ķ×ĵ +lä ss +läss ig +Ġ׾ ×Ļשר×IJ׾ +ĠÙĬ أتÙĬ +ר ×ķ×ĸ +ö ÄŁ +Ã¶ÄŁ ret +Ã¶ÄŁret im +Ġд ек +Ġдек аб +Ġдекаб ÑĢÑı +Ġש ×Ĺ×ķר +ãģ¦ãģıãĤĮ ãģŁ +عب ارة +Ġélect rique +ĠاÙĦتÙĨ ÙħÙĬØ© +جر Ùī +ĠìĪĺ íĸī +à¸Ĺ ู +ĠÑĢе алÑĮно +Ñģп оÑģоб +à¸Ħล à¹īาย +Ġس عÙĪØ¯ +ön ü +ĠÙģ ÙħÙĨ +تÙĥ ÙĪ +تÙĥÙĪ ÙĬÙĨ +ĠкаÑĩ еÑģÑĤво +ĠконÑĤ ак +ĠконÑĤак ÑĤ +Ġsöz leÅŁme +à¸Ń à¹īาà¸ĩ +Ġت ÙĪÙģ +ĠتÙĪÙģ ÙĬر +×Ķ×ĸ ×ĵ +×Ķ×ĸ×ĵ ×ŀ׳×ķת +ĠØ·ÙĪÙĬÙĦ Ø© +Ġtér mino +Ġ×IJ ×Ļפ×Ķ +ãĥĵ ãĥ« +ส à¹Ĥม +สà¹Ĥม สร +ĠاÙĦ اث +ĠاÙĦاث ÙĨÙĬÙĨ +ев иÑĩ +Ġopin ión +à¸Ľ วà¸Ķ +åı¤ ãģĦ +ร à¹Īา +ĠB iaÅĤ +ĠÑģÑĤ ал +ĠÑģÑĤал о +ó logo +ĠìķĦ ëĭĪëĭ¤ +Ġ×IJ ×Ļת +Ġ×IJ×Ļת ×ķ +à¹Ģหà¹ĩà¸Ļ วà¹Īา +à¸ļ ารà¹Į +çĦ ¼ +çĦ¼ ãģį +ĠìĿ´ìļ© ìŀIJ +ĠнекоÑĤоÑĢ Ñĭе +ks z +ksz taÅĤ +ksztaÅĤ c +ãĤŃãĥ£ ãĥĥãĤ· +ãĤŃãĥ£ãĥĥãĤ· ãĥ³ãĤ° +Ġro ÅĽ +ĠroÅĽ lin +ÑĢаж а +×ij׳×Ļ ×Ļ×Ķ +à¸Ľà¸£ สิ +à¸Ľà¸£à¸ªà¸´ à¸ķ +Ġgörd ü +×ŀ׳×Ķ ×Ļ×Ĵ +å¤īãĤı ãģ£ãģ¦ +Ġ×IJ ×Ķ +Ġ×IJ×Ķ ×ijת×Ļ +à¹Ģร à¹Īà¸ĩ +Ġön ünde +Ġê·¸ ëĥ¥ +пол иÑĤ +полиÑĤ иÑĩеÑģк +ãĥ¡ ãĥĩãĤ£ +ãĥ¡ãĥĩãĤ£ ãĤ¢ +ĠDet ay +ĠDetay lı +ĠاÙĦصÙģ ØŃØ© +à¸ģาร à¹Ģà¸ĩิà¸Ļ +Ġìµľ ê·¼ +׼ ש׾ +ï¼ © +вÑĪ ÐµÐ³Ð¾ +íķĺ ìĭ¤ +ĠÐŃ ÑĤ +ĠÐŃÑĤ оÑĤ +ส ื +สื à¸ļ +Ġng ừng +ĠдокÑĥменÑĤ ов +дав аÑĤÑĮ +ĠاÙĦشخص ÙĬØ© +Ġצ ×¢×Ļר +در Ùĥ +س ØŃب +à¹Ħมà¹Ī à¸Ħà¹Īà¸Ńย +Ġ×Ķ×ŀ×§ ×ķ×ŀ×Ļ +สัà¹Īà¸ĩ à¸ĭืà¹īà¸Ń +Ġê·¸ê²ĥ ìĿĦ +ãģĤãĤĭ ãģĦ +ãģĤãĤĭãģĦ ãģ¯ +×IJ×ķ×ĺ ×ķ×ij +×IJ×ķ×ĺ×ķ×ij ×ķס +к ÑĨион +ĠÐľ ожно +ãģı ãģł +ãģıãģł ãģķ +ĠинÑĦоÑĢм аÑĨиÑı +ï» Ł +Ġìŀij ìĹħ +Ġ×Ļ ×ķסף +Ø¥ دارة +ĠاÙĦØŃ اج +×ł×¡ ×Ļ×¢×Ķ +из аÑĨиÑı +×IJ׾ ×ij +×IJ׾×ij ×ķ×Ŀ +п ед +Ġ×§×ĺ ׳×Ķ +ĠÙĨÙ쨳 Ùĩا +ĠMinist ério +Ġп ен +Ġпен Ñģи +ãĥIJ ãĥ©ãĥ³ãĤ¹ +Ġ×Ķת ×ķר×Ķ +Ġt ạm +ĠìĹŃ ìĭľ +ï½ ¡ +Ġth á»± +Ġ ısı +ì» ¨ +ãģĹãģ£ãģĭãĤĬ ãģ¨ +Ġx ưa +Ġc ặp +×Ĺ ×Ļ×ij×ķר +วัà¸Ĵà¸Ļ à¸ĺรรม +st är +stär ke +ĠÑģам Ñĭй +p isa +pisa Äĩ +ĠoluÅŁ an +ĠاÙĦØ¥ ÙħاÙħ +ĠcÄĥ ng +Ġgü nl +Ġgünl ük +Ġ׳ש ×IJר +Ġkhi á»ĥn +ç¶ļ ãģijãĤĭ +stit ución +Ġcapac ité +Ġj aki +Ġjaki ÅĽ +вÑĪ Ð¸Ñģ +вÑĪиÑģ ÑĮ +פע×ķ׾ ×ķת +ĠØŃ ÙĬات +ĠØŃÙĬات Ùĩ +Ġник огда +ÐĽ Ь +Ġ×Ķ×¢ ×ķ×ij +Ġ×Ķ×¢×ķ×ij ×ĵ×Ķ +Ġch Ãło +หลาย à¹Ĩ +ĠÑı н +ĠÑıн ваÑĢ +ĠÑıнваÑĢ Ñı +à¸Īำà¹Ģà¸Ľà¹ĩà¸Ļ à¸ķà¹īà¸Ńà¸ĩ +Ġhö her +ãģķãĤĮãģ¦ ãģĦãģŁ +สà¸ĩ สั +สà¸ĩสั ย +ĠاÙĦ اس +ĠاÙĦاس ÙĦاÙħ +ĠاÙĦØ´ Ùħس +สà¸ĸาà¸Ļ ี +ãĤ¯ãĥ© ãĤ¹ +à¸ŀร ร +à¸ŀรร à¸Ħ +p õ +põ e +Ġpor ém +à¸Ľà¸£à¸° สà¸ĩ +à¸Ľà¸£à¸°à¸ªà¸ĩ à¸Ħà¹Į +powied zie +powiedzie Äĩ +Ġмог Ñĥ +Ġж ел +Ġжел ез +ĠاÙĦØ« ÙĤ +ĠاÙĦØ«ÙĤ اÙģÙĬ +ĠпÑĢав ило +Ġgdy ż +פש ×ķ×ĺ +ÑĢабоÑĤ ка +ĠÙĥ رة +Ø´ دد +Ùħار Ùĥ +Ùħ ÙĥØ© +Ġпод пиÑģ +×ĺ×ķ ×ķ×Ĺ +ĠÅĽ c +ĠÅĽc ian +Ġر جاÙĦ +Ġ×ª×ľ ×ķ×Ļ +и ÑĪ +иÑĪ ÑĮ +Ġmé dec +Ġmédec in +ëįĶ ëĿ¼ëıĦ +ĠÑĤеб Ñı +Ġ׾×Ķ ×ķס×Ļ×£ +ãģĬ 話 +Ġà¹ģà¸ķà¹Ī à¸ģà¹ĩ +د اÙģ +داÙģ Ø¹ +ĠC ùng +ãĥ»ãĥ» ãĥ»ãĥ» +ê¶ ģ +Ġdeber ÃŃa +หà¸Ļà¹Īวย à¸ĩาà¸Ļ +Ġva ÌĢ +Ġעצ ×ŀ +Ġעצ×ŀ ×Ŀ +à¹Ģà¸Ĭืà¹Īà¸Ń วà¹Īา +שק ×¢ +Ġ×Ķ ×Ľ×ķ׾ +Ġ×Ķ׼×ķ׾ ׾ +ни бÑĥд +нибÑĥд ÑĮ +ĠëĦĪ íĿ¬ +Ġоб ÑĢаÑī +ĠобÑĢаÑī а +Ġ×¢×ij×ķ×ĵ ת +ĠاÙĦÙħÙĨت خب +ıy ord +ıyord u +ÙĪ Ø° +×Ĺש ×Ļ×ij×ķת +Ġ×Ķ×¢ ×Ļ×§ +Ġ×Ķ×¢×Ļ×§ ר×Ļ +ì¢ Į +ยุ à¹Ĥร +ยุà¹Ĥร à¸Ľ +Ġа пÑĢ +ĠапÑĢ ÐµÐ»Ñı +sz ed +szed ÅĤ +д он +à¹Ģà¸ķิ à¸ļ +à¹Ģà¸ķิà¸ļ à¹Ĥà¸ķ +кол о +Ġkażde j +å¸ ° +帰 ãĤĬ +Ġмил ли +Ġмилли он +ç¾İåij³ ãģĹãģĦ +ت ÙĤار +تÙĤار ÙĬر +ĠìĿ´ 루 +ĠìĿ´ë£¨ ìĸ´ +Ġsprzeda ż +×Ķ ×ķצ×IJ×ķת +ãĤ¢ãĤ¯ ãĤ» +ãĤ¢ãĤ¯ãĤ» ãĤ¹ +ר ×ķ×¥ +ĠгоÑģÑĥдаÑĢÑģÑĤв енн +Ø£ ØŃÙĥ +Ø£ØŃÙĥ اÙħ +ĠoluÅŁ u +ĠA ç +ĠAç ık +ãĤ¸ ãĥ¼ +ç´ł æĻ´ +ç´łæĻ´ ãĤīãģĹãģĦ +Ġ×ijש×ij ×ķ×¢ +ب ذ +بذ ÙĦ +สา à¹Ģหà¸ķุ +Ġpoz osta +Ġpozosta ÅĤ +ØŃر Ùħ +Ġimport ância +leÅŁtir me +Ġд ÑĢев +Ġmó vil +ĠA ynı +Ġна лог +Ġналог ов +Ġ×Ĺ ×Ļפ×Ķ +ĠÑĦоÑĢм Ñĥ +à¸Ĺà¸Ķ สà¸Ńà¸ļ +ĠksiÄħż ki +Ġma ÅĤe +Ùħس Ø£ÙĦ +ÙħسأÙĦ Ø© +ï¼¾ ï¼¾ +ç ãeste +év iter +Ġкон ÑģÑĤÑĢÑĥк +ĠконÑģÑĤÑĢÑĥк ÑĨи +ï¾ ŀ +Ġת×ķ׼ ׳ +ãĤ¹ãĥĪ ãĥ¬ãĤ¹ +ĠاÙĦاÙĤتصاد ÙĬ +×ŀ×ĵ ×Ļ +Ġw ÅĤad +ĠwÅĤad z +Ø® ÙĪÙģ +ĠмаÑĤеÑĢиал ов +ãģ¨ãģ£ãģ¦ ãĤĤ +Ġznaj du +Ġznajdu jÄħ +Ùģ Ø¦Ø© +ãģ©ãģ® ãĤĪãģĨãģª +æĬij ãģĪ +׳ ×Ĺ׾ +Ġdü ny +Ġdüny an +Ġdünyan ın +гÑĢ Ð°Ð½Ð¸ +гÑĢани Ñĩ +Ġ×Ķש׾ ×Ļש×Ļ +Ġ×Ķ×IJ ש +åıĬ ãģ³ +ìĭŃ ìĭľ +ìĭŃìĭľ ìĺ¤ +Ġдол л +Ġдолл аÑĢ +Ġпов ÑĤоÑĢ +Ġ×Ĺ ×Ļ׳×Ŀ +ת פת×Ĺ +Ñĥв ели +Ñĥвели Ñĩен +ãĤ« ãĥª +raw id +rawid ÅĤow +×ķ ×ķ׾ +ãĥŁ ãĥ¥ +ì½ ĺ +ĠBy ÅĤ +Ðľ ÐIJ +ع ÙIJ +ĠÑģовеÑĢ ÑĪ +ĠÑģовеÑĢÑĪ ÐµÐ½Ð½Ð¾ +Ġм ой +Ġ×ķ׾×IJ ×Ĺר +æħ £ +æħ£ ãĤĮ +ØŃ اÙ쨏 +Ġ무 ë£Į +à¸Ħà¸ĵะ à¸ģรรม +à¸Ħà¸ĵะà¸ģรรม à¸ģาร +Ġìĸ´ ëĶĶ +Ġdif eren +Ġdiferen ça +ĠاÙĦØ£ ساس +ĠاÙĦأساس ÙĬØ© +Ġ׾×IJ×Ĺר ×ķ׳×Ķ +ê· ł +Ġ×Ķש׳×Ļ ×Ļ×Ķ +ìľĦìĽIJ ìŀ¥ +ลุ à¸ģ +ç iler +Ġ×Ķ×IJ ׾×ķ +èģŀ ãģı +Ġ×ķ×IJ פ×Ļ׾×ķ +ĠÑĢе ализ +ĠÑĢеализ аÑĨи +ระยะ à¹Ģวลา +Ġجدا Ùĭ +تب اع +Ġveh ÃŃculo +Ġдол г +à¸Ľà¸£à¸´ มาà¸ĵ +ì¦ IJ +Ġ׾ ×ŀ×§×ķ×Ŀ +ĠìĤ¬ ì§Ħ +à¸Ĭ à¹īา +Ġ×ŀ×¢ ×ķ׾×Ķ +Ġgö rm +Ġgörm ek +ĠÙĪÙĩ ذÙĩ +пеÑĢ Ð² +пеÑĢв ÑĭÑħ +ê·¸ ëŀĺ +ĠاÙĦبر ÙĬØ· +ĠاÙĦبرÙĬØ· اÙĨÙĬ +ĠиÑİ Ð½Ñı +ĠÐĵ оÑĢ +Ġ׾ ש׾×Ŀ +ÐIJ ÐĿ +Ġназ наÑĩен +о оÑĢ +ооÑĢ Ñĥж +Ġöz elli +Ġözelli ÄŁi +Ġни же +ç¶ļ ãģijãģ¦ +Ġа ÑĢенд +Ġkat ılı +Ġkatılı m +ĠØ¥ Ø·ÙĦاÙĤ +ĠÙĪØ¥ ذا +Ġок ÑĤÑı +ĠокÑĤÑı бÑĢÑı +à¹Ĥà¸ķ ๠+à¹Ĥà¸ķ๠Ĭ +à¹Ĥà¸ķà¹Ĭ ะ +Ġolduk ları +Ùħ ÙĪÙĤع +ëĤ © +ã썿ĢĿ ãģ£ãģ¦ãģĦãĤĭ +Ġש ×Ļ׼×ķ׾ +วา à¸Ķ +س ÙĬÙĦ +à¸Ĥ วั +à¸Ĥวั à¸į +تØŃ ÙĥÙħ +ì ĤŃ +Ġconna ît +׳ פת×Ĺ +Ġch ặ +Ġchặ n +ĠÙħ ØŃÙħ +ĠÙħØŃÙħ ÙĪØ¯ +ãģ ´ +ĠпÑĢодÑĥк ÑĨии +зд ÑĢав +ãģĶ è¦ +ãģĶè¦ § +×IJ×ij ×IJ +Ġvé ritable +ĠØ· ÙģÙĦ +ãĥĪãĥ© ãĥĸãĥ« +ê³ ¡ +Ġת ×ŀ×ķ׳×Ķ +Ġki ên +ĠÙĤ ادر +Ø¥ÙĤ ÙĦÙĬÙħ +ĠпÑĢед пÑĢи +ĠпÑĢедпÑĢи ÑıÑĤиÑı +Ġb Äĥng +Ġay ında +Ġg ấp +еÑħ ал +Ġgi Ãłnh +Ġд ав +Ġдав но +ìĺĢ ëĭ¤ +à¸Ļัà¸ģ à¹Ģà¸ķ +à¸Ļัà¸ģà¹Ģà¸ķ ะ +Ùħست شار +ست راتÙĬج +ستراتÙĬج ÙĬ +رÙħ ز +Ġt Ä©nh +ë¡ Ń +ĠÑĩ еÑĤ +ĠÑĩеÑĤ Ñĭ +ĠÑĩеÑĤÑĭ ÑĢе +ĠEnt ão +Ġص غ +Ġصغ ÙĬرة +×ij×Ļ×ĺ ×ķ׾ +خط ÙĪØ· +ĠÑĢазвиÑĤ ие +Ġamacı yla +à¸Ĺี วี +Ġо ÑģÑĤ +ĠоÑģÑĤ алÑĮн +ש×ķ׾׊ף +Ġ׼ ׳×Ļס +Ġ׼׳×Ļס ×Ķ +Ġd áºŃy +ĠyaÅŁ ayan +Ġ×ŀ×Ķ ×ķ×ķ×Ķ +ĠÑĥ Ñģи +ĠÑĥÑģи ли +×ŀ פ×Ļ +ĠпÑĢовед ениÑı +Ġر ب +Ġرب Ùħا +ĠاÙĦØ£ ÙĪØ³Ø· +Ġìľł ì§Ģ +Ġprac ownik +Ġpracownik ów +×ŀס ×ķרת +ÙĤار ب +à¸Ħวาม รูà¹īสึà¸ģ +à¹ģหล ะ +ĠاÙĦÙĨ ÙĤد +Ġ×IJ׾ פ×Ļ +Ùħس ئ +Ùħسئ ÙĪÙĦ +ев ÑĭÑħ +клÑİÑĩ ениÑı +×ij ×Ļ׳ +×ij×Ļ׳ ×Ļ×Ķ×Ŀ +ש ×ķ×IJ×Ķ +ĠÅŁ ark +ĠÅŁark ı +Ġsü rec +Ġsürec in +à¹Ģà¸Ħร à¸Ķ +à¹Ģà¸Ħรà¸Ķ ิà¸ķ +ãĥIJ ãĥ¬ +ĠØ´ Ø£ÙĨ +à¹Ģà¸Ńา à¹Ħวà¹ī +niÄĻ cie +רצ ×Ĺ +ĠaÅŁ ama +׳ פ×Ĵ×¢ +Ġth á»Ŀ +Ġkhu ẩn +diÄŁ inde +ÑıÑī иÑħ +ãĥĺ ãĥ« +Ġüber h +Ġüberh aupt +ĠÑĤÑĢеб ова +ĠdÅĤ ugi +×ĺ ×Ļף +à¸Ĥà¸Ļาà¸Ķ à¹ĥหà¸įà¹Ī +ĠاÙĦØ£ Ùĩ +ĠاÙĦØ£Ùĩ ÙĦÙĬ +ĠMü d +ĠMüd ürü +Ġ×Ļ×Ķ ×ķ×ĵ×Ķ +Ñĭв аеÑĤÑģÑı +س اط +×Ķת ׳×Ķ×Ĵ +×Ķ×ª×ł×Ķ×Ĵ ×ķת +à¸ģาร à¸ľà¸¥à¸´à¸ķ +íĴ Ģ +สà¸ĸาà¸Ļ à¸ģารà¸ĵà¹Į +Ġо ÑĦ +ĠоÑĦ иÑģ +ĠÙĦ عبة +Ġstron ÄĻ +Ġר×IJ ×ķ×Ļ +×Ĺ ×ij׾ +ĠÑĢÑĭ н +ĠÑĢÑĭн ке +Ġ׾×ŀ×¢ ף +اس ÙĦ +ห ัà¸Ļ +Ġ×IJ ×Ĺ×Ļ +ĠпÑĢод ол +ê°Ģ ìŀħ +Ġ×ijר ×Ĺ +Ġ×ijר×Ĺ ×ij×Ļ +дж еÑĢ +Ġ׾ ×Ĺ׾ +Ġ׾×Ĺ׾ ×ķ×ĺ +Ġ׾×Ĺ׾×ķ×ĺ ×Ļף +ศาส à¸Ļา +ãĤ¢ãĤ¤ ãĥĨ +ãĤ¢ãĤ¤ãĥĨ ãĥł +Ġפר ×ķפ +جز اء +ล à¸Ńย +Ġc iaÅĤa +Ġgi ết +ĠзнаÑĩ иÑĤелÑĮно +Ġolmad ıģ +Ġolmadıģ ını +н д +нд екÑģ +تأ Ùĥد +Ġìĸ ¸ +Ġìĸ¸ ìłľ +ay dın +ãĥī ãĥ¬ãĤ¹ +Ġs ắt +Ġíĺ¸ íħĶ +Ġë¶ ģ +Ġë¶ģ íķľ +ãĥij ãĤ¤ +Ġ×ŀש×Ĺ×§ ×Ļ +à¸Ħà¸Ļ à¸Ńืà¹Īà¸Ļ +Ġиз гоÑĤов +ĠизгоÑĤов лен +à¹Ģà¸ģีย ร +à¹Ģà¸ģียร à¸ķิ +תק שר +ĠÑĢаÑģ ÑĩеÑĤ +ส à¹Ģà¸ķ +Ġl änger +ĠiÅŁ let +ĠiÅŁlet me +Ġع ÙĦÙĬÙĨ +ĠعÙĦÙĬÙĨ ا +é lection +ĠاÙĦغ ربÙĬØ© +íĭ Ģ +ãĤĤãĤī ãģĪ +Ġкни ги +Ø£ سÙħ +أسÙħ اء +Ġth á»ı +Ġthá»ı a +หà¸Ļ ู +Ġ×ł×¢ ש×Ķ +à¸łà¸²à¸¢ à¹ĥà¸ķà¹ī +à¸ŀื à¸Ĭ +رÙĬ Ø· +Ùģ ÙĪØ¶ +ãģĤãĤĬãģĮãģ¨ãģĨãģĶãģĸ ãģĦãģ¾ãģĹãģŁ +ש ×ĵ×Ķ +Ġng á»±c +ĠÑģеÑĢ ÑĮ +ĠÑģеÑĢÑĮ езн +T ôi +Ġfiyat ları +ĠвÑģ Ñİ +ĠC ódigo +Ġ×Ķש ×IJ +Ġ×Ķש×IJ ׾×Ķ +ĠP ública +Ø¥ Ø® +إخ ÙĪØ§ÙĨ +ĠзаÑıв ил +ãĥ¦ ãĥ¼ +ר×IJ ×Ļת +vol ución +Ġsz ko +Ġszko ÅĤy +جرÙĬ دة +Ġpens é +ìī ¬ +ĠBüyük ÅŁehir +ĠØ£Ùħ رÙĬ +ĠØ£ÙħرÙĬ ÙĥÙĬ +à¸Ļัà¸ģ ศึà¸ģษา +Ġtod av +Ġtodav ÃŃa +ĠС ан +ĠСан кÑĤ +íķĺ ìŀIJ +ØŃÙĪ Ø§ÙĦ +׼ ×ķשר +à¹Ģลย à¸Ħรัà¸ļ +Ġal gu +Ġalgu ém +Ùģ Ø² +Ġçek il +Ġ×ĵ ר׼×Ļ×Ŀ +ãĥIJ ãĥ© +à¸ģà¹ĩ สามารà¸ĸ +สà¹Īวà¸Ļ ลà¸Ķ +íı ° +ĠP úb +ĠPúb lico +à¹ģà¸Ļว à¸Ĺาà¸ĩ +×IJת ×Ĵר +Ø´ اش +شاش Ø© +ci ÅĽni +ĠÃľ rün +ÙĦÙĪ ØŃ +ĠاÙĦ بÙĨ +ĠاÙĦبÙĨ Ùĥ +ì¡° ì¹ĺ +Ġorganiz ación +ãģĤãĤĬãģĮãģ¨ãģĨãģĶãģĸ ãģĦãģ¾ãģĻ +s ätze +ĠÑģем ей +ÙĤ صد +ÑģÑĤв еннÑĭе +Ġpréc éd +Ġprécéd ent +à¸ģรุà¸ĩà¹Ģà¸Ĺà¸ŀ ฯ +ãģ¨è¨Ģ ãģĦ +×ij׳×Ļ ×Ļף +ĠØŃ ÙĪ +ĠØŃÙĪ Ø§ÙĦÙĬ +סק ס +ĠsaÄŁlam ak +Ġ׾ צ×Ļ×Ļף +×§×ĵ ש +Ġ×Ķ×ŀ ×¢×¨×Ľ×ª +Ġ׾×Ķ ×¢×ij×Ļר +Ġg ünd +Ġgünd em +ĠнаÑĪ ÐµÐ³Ð¾ +à¹ĥà¸Ļ à¸ŀืà¹īà¸Ļà¸Ĺีà¹Ī +à¹Ģà¸Ħร ืà¸Ń +à¹Ģà¸Ħรืà¸Ń à¸Ĥ +à¹Ģà¸Ħรืà¸Ńà¸Ĥ à¹Īาย +ظ اÙĩرة +ÙħÙĨ ظÙħ +ÙħÙĨظÙħ ات +Ùħت از +追 ãģĦ +dı kt +dıkt an +ĠëįĶ ìļ± +ĠÐĿ апÑĢимеÑĢ +tw ór +×ŀ×ķ×¢ צ×Ķ +Ùĥ ÙĪÙĥ +Ð © +×ŀ×ĺ פ׾ +ó lica +訪 ãĤĮ +ĠëĮĢ ë¶Ģ +ĠëĮĢë¶Ģ ë¶Ħ +ãĤ¯ãĥª ãĥĥãĤ¯ +ãĤĴ éģ¸ +ãĤĴéģ¸ ãģ¶ +Ġpow sta +Ġpowsta ÅĤ +Ġraz ón +×ij ×ķ×Ĺר +ĠÑģообÑī ил +Ġ×§ ×ij×ķ×¢ +r êt +à¸Ķี à¸Ĥึà¹īà¸Ļ +×ŀס ×¢×ĵ +×ŀסע×ĵ ×ķת +ĠÃĸ sterreich +Ġ׳ ×Ĺש×ij +Ùħباد رة +ì´ ī +×Ĵ ׳×ĺ×Ļ +ä¿¡ ãģĺ +du ÄŁ +duÄŁ unu +Ġph ú +ĠاÙĦØ£ Ø®ÙĬر +Ġت عتبر +landır ıl +ãģ¨ãģ¯ ãģĦ +ãģ¨ãģ¯ãģĦ ãģĪ +ĠاÙĦ Ø·ÙĦ +ĠاÙĦØ·ÙĦ اب +ĠN º +éģ¿ ãģij +اÙĦ Ùħع +اÙĦÙħع رÙĪÙģ +ส à¸łà¸² +éĽ¢ ãĤĮ +ĠпомоÑī ÑĮ +Ġзна еÑĤ +ãĥĹãĥ¬ ãĤ¼ +ãĥĹãĥ¬ãĤ¼ ãĥ³ãĥĪ +Ġsup érieur +Ġש׾ ×Ļש×Ļ +ĠاÙĦÙĨ ÙĪØ¹ +ãĤĵãģ§ãģĻ ãģŃ +à¸Ńà¸ļ รม +Ġgi á»įng +Ġwzgl ÄĻd +ĠاÙĦÙģ ÙĤر +è rent +Ġ×ŀ×IJ ×Ĺ +Ġ×ŀ×IJ×Ĺ ×ķר×Ļ +×Ĵ ×Ĵ +×Ļ ×Ļ×ij +ÙħÙĦ اب +ÙħÙĦاب س +Ġhük ü +Ġhükü met +Ġ×ŀ×Ĵ ×Ļ×ij +ĠÐŀ Ñĩ +ĠÐŀÑĩ енÑĮ +æĹ© ãģĦ +Ġconstr ucción +Ġth ượng +ï¼ ĭ +Ġcor ação +à¹Ģหล à¹ĩà¸ģ +ĠBaÅŁ b +ĠBaÅŁb akan +éĢ£ ãĤĮ +ãģĻãĤĭ ãģĵãģ¨ãģĮãģ§ãģįãģ¾ãģĻ +ĠÙĤ اÙħت +Ġا Ùĥثر +ÙģØ§Ø¹ ÙĦ +ĠÑĦ оÑĢ +ĠÑĦоÑĢ Ñĥм +غ ذÙĬ +ĠiÅŁ le +ĠiÅŁle ml +ĠiÅŁleml eri +ĠìĤ¬ëŀĮ ìĿĢ +Ġìŀij ìĦ± +Ġë§Ī 볨 +Ùħ جÙĦس +หม ู +д в +дв иг +двиг а +à¹Ģสีย à¸Ĭีวิà¸ķ +×Ķת פת×Ĺ +×Ķתפת×Ĺ ×ķת +ĠмеÑĤ ÑĢо +ĠÑģ енÑĤ +ĠÑģенÑĤ Ñı +ĠÑģенÑĤÑı бÑĢÑı +ê³ § +Ġ׾ פע +Ġ×ľ×¤×¢ ×ŀ×Ļ×Ŀ +à¹Ģà¸ļ ีย +詳 ãģĹãģı +çķ° ãģªãĤĭ +Ġİl çe +ĠAt at +ĠAtat ür +ĠAtatür k +รุ à¹Īà¸ĩ +Ġkald ı +Ġ주 ìŀ¥ +Ġprés ence +Ġн аб +Ġнаб лÑİ +ĠнаблÑİ Ð´Ð° +ĠÑģам ого +×Ĵ ×ķש +×ŀ×ĺ ×ķפ +×ŀ×ĺ×ķפ ׾ +ĠвÑĭб иÑĢа +ĠìŀIJ 리 +åĪĨ ãģĭãĤīãģªãģĦ +Ġз Ñĥб +Ġש׼ ×ijר +Ġد ائ +Ġدائ Ùħا +ĠпаÑĢ ÑĤи +ï¼ ² +ĠاÙĬ ضا +ĠÑħ оз +ĠÑħоз Ñı +ĠÑħозÑı й +ĠÑħозÑıй ÑģÑĤв +ĠاÙĦØ£ ج +ĠاÙĦأج ÙĨب +ĠاÙĦأجÙĨب ÙĬØ© +ĠÐĹ Ð½Ð° +ĠAp ós +ĠÑį неÑĢ +ĠÑįнеÑĢ Ð³Ð¸ +Ġy ans +Ġyans ı +ĠJust i +ĠJusti ça +Ġpré vu +ม วล +ìŀ¥ ëĭĺ +à¸ģระ à¸ļ +à¸ģระà¸ļ วà¸Ļ +à¸ģระà¸ļวà¸Ļ à¸ģาร +×ŀ ×ŀ +×ŀ×ŀ ×ķצע +Ġh ẹ +Ġhẹ n +зд ание +Ġak ÅŁ +ĠakÅŁ am +×ĺ ×ķפ +Ġgere kt +Ġgerekt i +Ġgerekti ÄŁini +Ġnar z +Ġnarz ÄĻdzi +é po +épo que +ĠTh ần +Ġwys oko +Ġwysoko ÅĽci +à¸ľà¸¹à¹ī à¸Ľ +à¸ľà¸¹à¹īà¸Ľ à¹Īวย +ĠÙĬ بدÙĪ +ÑĤелÑĮ ного +Ġвз глÑıд +Ġjed nÄħ +ĠìĿĺ 견 +Ġ à¸Ĥà¸ĵะà¸Ĺีà¹Ī +פ ×Ļ×ĵ +ìĥģ ëĭ´ +Ġm ỡ +×Ķ ×ŀ׾ +×Ķ×ŀ׾ צ×ķת +ĠÑģоÑģÑĤ о +ĠÑģоÑģÑĤо иÑĤ +Ġав и +Ġави а +ĠL änder +تص ÙĪÙĬر +×ŀ×ĵ ×Ļ×Ķ +ìłĪ ì°¨ +ãģ¨ ãĤĬ +ãģ¨ãĤĬ ãģĤ +ãģ¨ãĤĬãģĤ ãģĪ +ãģ¨ãĤĬãģĤãģĪ ãģļ +ĠÑĢ Ñıд +ĠÑĢÑıд ом +ĠNh ất +ĠاÙĦÙĥ اÙħÙĦ +×Ĺ׾ ׾ +ĠGi ấy +צ ×ĺר +צ×ĺר ×£ +Ġ׾×ij ×ĺ׾ +Ġим еÑĤÑĮ +ס×ŀ ×ķ×ļ +Ġparticip ação +íķľëĭ¤ ë©´ +ÙħÙĨت دÙĬ +ÙħÙĨتدÙĬ ات +ĠeÄŁ len +g änge +رب ØŃ +ãĤ® ãĥ£ +ĠاÙĦر ÙĤÙħ +à¸ĭ à¹īำ +ĠH óa +×ŀר ×Ĺ×§ +ØŃÙħ اÙħ +بÙĪ Ùĥ +ĠArt ÃŃculo +ãĥĦ ãĤ¢ãĥ¼ +×Ķפ ׼×Ķ +×Ĺ׾ ×ķף +ĠпеÑĢе Ñħод +len miÅŁ +زر اعة +Ġseñ or +ãģ£ãģ¦ ãģįãģ¦ +Ø¥ Ø´ +إش ارة +Ġpod ÃŃa +ĠÃľ lke +н ÑģкаÑı +Ġadapt é +Ġdüzen len +Ġdüzenlen en +ĠÑģÑĤ ала +ĠÙĬ ØŃتاج +Ġn ier +Ġnier uch +Ġnieruch omo +Ġnieruchomo ÅĽci +ãģĵãģ¨ãģĮ ãģĤãĤĭ +ยà¸Ńà¸Ķ à¹Ģยีà¹Īยม +ĠÙħ ج +ĠÙħج اÙĨÙĬ +Ġз аб +Ġзаб ол +Ġзабол ев +Ġзаболев аниÑı +ĠÅĽ ro +ĠÅĽro dk +ĠÅĽrodk ów +Ġ×Ķ ×ľ×IJ×ķ×ŀ×Ļ +Ġdok ÅĤad +ĠdokÅĤad nie +ãģŁãģı ãģªãģĦ +ãģ¯ãģļ ãģ§ãģĻ +ã썿ĢĿ ãģ£ãģ¦ãģĦãģŁ +é cran +ìĹħ ì²´ +trzym aÅĤ +ÑģÑĤв еннÑĭй +ĠNot ÃŃc +ĠNotÃŃc ias +Ùħ رÙĬ +ÙħرÙĬ ض +æ°Ĺ è» +æ°Ĺè» ½ +æ°Ĺ軽 ãģ« +ëĵ £ +Ġ×ĵ ×ķ×IJר +Ġ׾ ×ŀ׳ +Ġ׾×ŀ׳ ×ķ×¢ +ĠçalÄ±ÅŁ ıyor +ĠÅŁ idd +ĠÅŁidd et +ĠM ặt +Ġate ÅŁ +ĠполÑĥÑĩ ениÑı +à¹Ģà¸Ħรืà¹Īà¸Ńà¸ĩ มืà¸Ń +Ġgrö ÃŁer +د ائ +دائ رة +Ġbul un +Ġbulun maktadır +à¹Ģห ร +à¹Ģหร ีย +à¹Ģหรีย à¸į +à¸Ļัà¸ģ à¸Ĺà¹Īà¸Ńà¸ĩà¹Ģà¸Ĺีà¹Īยว +Ġalan ında +ĠÑĥ зна +Ġл еÑĩение +売 ãĤĮ +Ġçev ir +Ġdeste ÄŁi +ĠheiÃŁ t +âĸ ² +ØŃ Ø· +à¸Ħำ à¸ķà¸Ńà¸ļ +ãĤªãĥ³ ãĥ©ãĤ¤ãĥ³ +Ġ×ij×Ĺ×Ļ ×Ļ×Ŀ +ãĥ¦ ãĥĭ +Ġdüzenle me +Ġmodal itÃł +سر Ø· +سرط اÙĨ +×ŀ׼ ×ķף +ĠданнÑĭ й +تر ت +ترت ÙĬب +à¸ļาà¸ĩ à¸Ħà¸Ļ +ĠÄIJ á»ĭnh +ม ูล +มูล à¸Ħà¹Īา +ÙĨ ÙĤص +à¸ģาร รัà¸ģษา +ĠÑĦ он +ĠÑĦон д +ãĤĪãģĨ ãģ«ãģªãģ£ãģŁ +Ùħع اÙĦ +ÙħعاÙĦ جة +ĠOs man +ĠOsman lı +иÑĩеÑģк ом +à¸Ńยาà¸ģ à¸Īะ +ãģķãģ¾ ãģĸ +ãģķãģ¾ãģĸ ãģ¾ +ãģķãģ¾ãģĸãģ¾ ãģª +Ġת ×ķ׼׾ +×¢ צ×ij +ĠاÙĦع سÙĥ +ĠاÙĦعسÙĥ رÙĬ +Ġvé hic +Ġvéhic ule +Ġ×Ļצ ×Ĺ×§ +ĠاÙĦÙĪ ØŃ +ĠاÙĦÙĪØŃ ÙĬد +ĠاÙĦع دÙĪ +ĠQu ản +Ġê³µ ëıĻ +بد ÙĦ +ĠÄij ảng +Ġm á»ĩnh +Ġnie zb +Ġniezb ÄĻ +ĠniezbÄĻ dn +Ġyayın lan +обÑī и +Ġgö tür +צ פ +צפ ×ķ×Ļ +ĠÙĦÙĬ بÙĬ +ĠÙĦÙĬبÙĬ ا +ØŃ ÙĪØ§ +Ġд об +Ġдоб ÑĢо +иÑĢÑĥ ем +ĠاÙĦØŃÙĥÙĪÙħ ÙĬØ© +m Ã¤ÃŁig +Ġed ición +влек аÑĤелÑĮ +влекаÑĤелÑĮ н +Ġת ש׾×ķ×Ŀ +Ġ×Ķש ×ķ׳×Ļ×Ŀ +มิ à¸ĸุ +มิà¸ĸุ à¸Ļ +มิà¸ĸุà¸Ļ ายà¸Ļ +é£Łãģ¹ ãģ¦ +ĠìĪĺ ì§ij +ס ×ij×Ļ +ĠиÑİ Ð»Ñı +Ġà¹Ħà¸Ķà¹ī à¹ģà¸ģà¹Ī +׾×Ĺ ×Ŀ +tr ä +trä gt +ãģĿãĤĤ ãģĿãĤĤ +ÐĿ Ðķ +Ġв нÑĥÑĤ +ĠвнÑĥÑĤ ÑĢи +ãģ¨ ä¸Ģç·Ĵãģ« +ãĤ« ãĥķãĤ§ +Ġ×ij×Ĺ ×ĵר +×Ĺ ×ŀש +ãĤ¨ ãĥį +ãĤ¨ãĥį ãĥ« +ãĤ¨ãĥįãĥ« ãĤ® +ãĤ¨ãĥįãĥ«ãĤ® ãĥ¼ +à¸Ĥà¸Ńà¸ĩ à¸ķัวà¹Ģà¸Ńà¸ĩ +بÙĤ اء +פס ×Ļ׼ +פס×Ļ׼ ×ķ׾×ķ×Ĵ +ãĥ¡ ãĥĥ +ãĥ¡ãĥĥ ãĤ» +ãĥ¡ãĥĥãĤ» ãĥ¼ãĤ¸ +ÙĦ ÙĤب +A Äŀ +שק ×Ļ×¢ +ÙĤ ساÙħ +×ĵ×ķ×Ĵ ×ŀ×Ķ +æ·± ãģĦ +íĸĪ ëĬĶëį° +ĠrozwiÄħz anie +à¸Ļัà¹Īà¸Ļ à¹Ģà¸Ńà¸ĩ +×Ļצ ×ij +Ġtr ông +à¹ĥà¸Ĭà¹ī à¸ļริà¸ģาร +ĠاÙĦÙħÙĪ Ø³Ùħ +ĠдеÑĤ и +ãģĹãģĭ ãģªãģĦ +ס ×Ļף +Ġréfé rence +à¹ģห à¹īà¸ĩ +ãĤĤãĤī ãģ£ãģŁ +Ġ׾ ר׼ +Ġ׾ר׼ ×ķש +شع ÙĪØ± +ĠÐij ог +Ġlaz ım +Ġ×Ļש ׳×Ŀ +Ġп аÑĢÑĤ +ĠпаÑĢÑĤ неÑĢ +ĠÑĥ ника +ĠÑĥника лÑĮн +Ġmaté riel +×ŀר ×§ +Ġph ưá»Ŀng +Ġз ай +Ġзай м +Ùģ ÙĤد +Univers itÃł +×¢ ר׼×Ļ×Ŀ +Ġba ño +Ġн оÑı +ĠноÑı бÑĢÑı +à¸Ľ à¹īาย +Ġt ats +Ġtats äch +Ġtatsäch lich +ĠÑĤÑĢ ÐµÑĤÑĮ +Ñį м +ãĥĻ ãĥ¼ãĤ¹ +Ġnh á»±a +ìĬ¤ íģ¬ +ĠعبداÙĦ ÙĦÙĩ +Ġת ×ķר×Ķ +أش ÙĬ +أشÙĬ اء +ĠÙĦÙĦ غا +ĠÙĦÙĦغا ÙĬØ© +Ùħ ÙĪØ§ÙĤ +ÙħÙĪØ§ÙĤ Ùģ +ĠgÅĤówn a +Ġart Ä±ÅŁ +Ġ×ŀ×§ ×ķ×ŀ×Ļ +ãĤ¯ãĥ© ãĥĸ +Ġس ÙĪÙī +ĠìŬ ìĦ± +اس ر +اسر ائÙĬÙĦ +Ġ׳ ×Ľ×ª×ij +ย à¹īà¸Ńà¸Ļ +Ġdeber á +Ġph ẫu +ÑİÑī ем +ĠÙĦدÙĬ ÙĨا +×ŀ×ĺ ×Ķ +Ġ׳ ×ķ׾×ĵ +ĠвÑģÑĤÑĢ ÐµÑĩа +ãĤīãĤĮ ãģ¦ãģĦãģ¾ãģĻ +ĠcaÅĤ ej +ย ึ +ยึ à¸Ķ +поÑĤ ен +поÑĤен ÑĨи +Ġл иÑĤ +ĠлиÑĤ еÑĢ +ĠлиÑĤеÑĢ Ð°ÑĤÑĥÑĢ +Ġкажд ом +ĠíĮ IJ +ĠíĮIJ ëĭ¨ +à¸Ī ู +Ġpres ença +ãģªãĤĵ ãģ§ +Ùħ ÙĬاÙĩ +ин ÑĦоÑĢм +инÑĦоÑĢм аÑĨион +инÑĦоÑĢмаÑĨион н +ĠìŀIJ ìŰ +ר׼ ש +Ġöd ül +ç¶ļ ãģı +Ġп Ñģ +ĠпÑģ иÑħ +ĠпÑģиÑħ олог +ت ذÙĥر +Ġìŀħ ìŀ¥ +ล à¸Ķà¹Į +ìĦł ê±° +ãģ£ãģ¦ ãģĬãĤĬãģ¾ãģĻ +Ġ×Ļ ×¢ +Ġ×Ļ×¢ ×§×ij +ĠاÙĦØ· عاÙħ +ãĥĨ ãĤ¹ãĥĪ +ĠTu ấn +Ġparticip ación +×ŀ×ķ×ŀ ×Ĺ×Ķ +×Ĵר ס×Ķ +ĠاÙĦتÙĨ ÙģÙĬ +ĠاÙĦتÙĨÙģÙĬ ذÙĬ +ĠбезопаÑģ н +ge f +gef ähr +Ø´ ÙĪØ± +Ġmy ÅĽli +ÙĪØ§ Ø´ÙĨ +ÙĪØ§Ø´ÙĨ Ø·ÙĨ +׳×ķס ×¢ +Ùĥ Ùĩ +ÙĥÙĩ رب +ÙĥÙĩرب اء +Ġmus iaÅĤ +ìĭ ¸ +ãĥĸãĥ© ãĥĥãĤ¯ +Ġcré é +ÙĨÙĩ ار +owo ÅĽÄĩ +ÙħØŃا ÙĥÙħ +ĠwÅĤa ÅĽ +ĠwÅĤaÅĽ c +ĠwÅĤaÅĽc iciel +ĠÙĬ ؤ +ĠÙĬؤ دÙĬ +×ŀ×¢ ×ķ׳ +×IJ ×ij׾ +خط Ø£ +ĠÑħ олод +×ĸ ×ķ׾ +ãģĵãĤĮ ãĤī +ãģĵãĤĮãĤī ãģ® +Ġbás ica +ฤ à¸Ķ +ฤà¸Ķ ูà¸ģ +ฤà¸Ķูà¸ģ า +ฤà¸Ķูà¸ģา ล +èIJ½ãģ¡ çĿĢ +ãģªãģĦ ãģĵãģ¨ +ص ÙĪÙħ +ÙĨج ØŃ +׳ק ×ķ×ĵ +׳ק×ķ×ĵ ת +кл аÑģÑģ +íķĺìĭľ ëĬĶ +ëĦ ĺ +Ġש×IJ ×Ļ׳×ķ +ĠС ейÑĩаÑģ +may acaģı +Ġyap ılır +Ġcategor ÃŃa +عب اد +ĠТ еп +ĠТеп еÑĢÑĮ +×Ķ×Ļס×ĺ ×ķר×Ļ +h ế +ãĤ³ ãĥ¼ãĥī +Ġcabe ça +ج Ùħا +جÙħا Ùĩ +جÙħاÙĩ ÙĬر +ä½İ ãģĦ +ĠÑĤоваÑĢ Ð¾Ð² +à¸Ĭาว à¸ļà¹īาà¸Ļ +ĠÑģÑĤан ов +ĠÑģÑĤанов иÑĤÑģÑı +ĠавÑĤом обилÑĮ +ĠÑģлÑĥÑĩ ай +à¸Ńั à¸ŀ +ĠG iriÅŁ +ĠìĿ¼ ëĭ¨ +ĠпÑĢ Ð¾Ñģ +ĠпÑĢоÑģ моÑĤÑĢ +ãģªãģıãģª ãģ£ãģŁ +มี à¸Ľà¸±à¸įหา +ïº İ +éc oute +ĠÙħ ÙĪØ¬ÙĪØ¯ +Ġس رÙĬع +ĠÙĪÙĩ ÙĨا +ĠÙĪÙĩÙĨا Ùĥ +à¸Ħุà¸ĵ สม +à¸Ħุà¸ĵสม à¸ļัà¸ķิ +Ġìļ° ìĦł +à¸ŀระ à¸ŀุà¸Ĺà¸ĺ +好 ãģ¿ +ظ ÙĦÙħ +Ġм акÑģ +ĠмакÑģ ималÑĮ +ĠмакÑģималÑĮ но +ãĥª ãĤ¢ãĥ« +à¹ģมà¹ī วà¹Īา +ĠاÙĦØŃ ÙĪØ§Ø± +ãĥĹãĥ© ãĤ¹ +Ġع ÙĦاÙĤØ© +Ġíĸī ëıĻ +Ġgönder il +Ġl ãi +ĠsaÄŁ lıkl +ĠsaÄŁlıkl ı +ĠÑĪ Ð°Ð³ +Ġ×ij×IJר ×Ķ +prowadzi Äĩ +ãģĦãģı ãģ¤ãģĭ +Ġبت ارÙĬØ® +Ġ×ij×IJ×ķת ×Ķ +Ġmó c +ĠÐľ не +ãĥĹãĥ¬ ãĥ¼ +×IJ ×ĸר×Ĺ +åł´åIJĪ ãģ«ãģ¯ +使 ãģĪ +à¹Ģร ืà¸Ńà¸Ļ +ĠÐŁ еÑĤ +ĠÐŁÐµÑĤ ÑĢ +ãģ«åħ¥ ãĤĭ +Ùħ ادة +à¹Ģà¸ĩ ืà¹Īà¸Ńà¸Ļ +à¹Ģà¸ĩืà¹Īà¸Ńà¸Ļ à¹Ħà¸Ĥ +ĠÑģоÑģÑĤоÑı ние +ôn ica +ĠÑĦ ев +ĠÑĦев ÑĢа +ĠÑĦевÑĢа лÑı +Ġ×ķ ×ĸ +Ġ×ķ×ĸ ×IJת +à¸Ħร ิ +à¸Ħริ ส +ĠÐķ Ñīе +ãģ£ãģ¦ãģĹãģ¾ ãģĦãģ¾ãģĹãģŁ +ĠпÑĢав иÑĤелÑĮ +ĠпÑĢавиÑĤелÑĮ ÑģÑĤв +Ġtä glich +Ġëĭ¹ ìĭľ +×ŀ×ķ×¢ ×ŀ×ĵ +Ġдв оÑĢ +æī ķ +æīķ ãģĦ +ĠÑģÑĤан еÑĤ +Ġвозд ейÑģÑĤв +ĠвоздейÑģÑĤв и +Ġf ête +à¹Ģส า +תק ×ķ×ķ×Ķ +Ġu yar +Ġuyar ı +à¸ģลัà¸ļ à¹Ħà¸Ľ +Ġgi ưá»Ŀng +Ġв а +Ġва ÑĪи +ĠÄij áºŃu +ĠSpa ÃŁ +ĠìķĦ ë§Ī +à¹Ħà¸Ķà¹ī à¸ĩà¹Īาย +Ġ×Ķ×ŀ ×ijקש +æĸ° ãģŁ +æĸ°ãģŁ ãģª +ılı yor +пл ан +Ġ×Ķ×ijר ×Ļ×IJ×ķת +ĠaÄŁ rı +Ġsay gı +建 ãģ¦ +Ġnaj wyż +Ġnajwyż sz +سÙĬاس ات +ãģĬ å¾Ĺ +ĠاÙĦع ÙĦÙĬ +ĠاÙĦعÙĦÙĬ ا +Ġcoraz ón +ì¹ĺ ë£Į +หัว à¸Ĥà¹īà¸Ń +Ġب ØŃÙĬ +ĠبØŃÙĬ Ø« +зв езд +بÙĪ Ø§Ø¨Ø© +ÐĽ Ðĺ +ÙĦا زÙħ +Ġroz p +Ġrozp oc +Ġrozpoc zÄĻ +触 ãĤĮ +ĠاÙĦج ÙħÙĩ +ĠاÙĦجÙħÙĩ ÙĪØ± +Ġsp ÄĻd +ĠspÄĻd z +วิà¸Ĺยา ศาสà¸ķรà¹Į +ив аеÑĤÑģÑı +Ġдан ной +Ġreprés ente +ĠÄij á»ĭch +Ġ×¢×ŀ ×ķ×§ +à¸Ńัà¸Ļ à¸ķร +à¸Ńัà¸Ļà¸ķร าย +Ġestr atég +Ġestratég ia +pad ÅĤ +Ġв полн +Ġвполн е +ĠпÑĢедоÑģÑĤав лен +×Ĺ׾ ×ķ×§ +×Ĺ׾×ķ×§ ת +ãĤ¢ ãĥĬ +ĠاÙĦغ ذ +ĠاÙĦغذ ائÙĬ +ĠÑĥ зн +ĠÑĥзн аÑĤÑĮ +à¸ĭ à¹īาย +å½ĵ ãģ¦ +ØŃÙĬ اء +Ġbás ico +×§×ķ×ij ×¢ +ĠاÙĦÙħ باراة +ĠاÙĦÙĩ اتÙģ +Ġ׼ ׳×Ĵ×ĵ +à¸Ľà¸£à¸° หย +à¸Ľà¸£à¸°à¸«à¸¢ ัà¸Ķ +Ðļ ак +à¸Ĺีà¹Ī à¸Ļà¹Īา +à¸Ĺีà¹Īà¸Ļà¹Īา สà¸Ļà¹ĥà¸Ī +ãģ¾ ãģģ +ï½ ¢ +Ñģк оп +Ġson rasında +Ġur zÄħd +ĠurzÄħd zenia +׼×ķ ×ķ׳ +׼×ķ×ķ׳ ת +Ġ׾×Ķת ×ŀ×ķ×ĵ +Ġ׾×Ķת×ŀ×ķ×ĵ ×ĵ +ĠÑģ ли +ĠÑģли ÑĪ +ĠÑģлиÑĪ ÐºÐ¾Ð¼ +ĠÑģÑĤ Ñĥд +ĠÑģÑĤÑĥд енÑĤ +Ġ×Ķ ×ķ×ĵ +Ġ×Ķ×ķ×ĵ ×¢×Ķ +ë¹Ħ ìļ© +à¸Ńยาà¸ģ à¹ĥหà¹ī +Ġb á»ģ +ยุ à¸Ĺà¸ĺ +Ðĺ ÐĿ +س ائر +Ø£ صÙĪÙĦ +ĠاÙĦغ رÙģ +ãģĵãģ¨ãĤĤ ãģĤãĤĬãģ¾ãģĻ +è¾¼ ãģ¾ãĤĮ +ĠاÙĦساب ع +Ġc á»§ +ãģĦãģŁãģł ãģĦãģŁ +ì§ ĵ +ìĤ¬ 무 +powied ź +تÙģ Ùĥ +تÙģÙĥ ÙĬر +иÑĢов ки +ĠíĨµ íķ´ìĦľ +ãĤ¨ ãĤ¹ãĥĨ +ĠдеÑıÑĤелÑĮ ноÑģÑĤÑĮ +ĠданнÑĭ м +Ġ×¢ ×ķר +Ġ×¢×ķר ׼×Ļ +×ķ×ĵ עת +Ġhayat ını +Ġb Äħd +ĠbÄħd ź +obs ÅĤug +à¹Ģà¸ŀียà¸ĩ à¹ģà¸Ħà¹Ī +à¸ĭ à¹Īา +è²ł ãģij +ĠÑģÑĤÑĢ ÐµÐ¼ +ĠÄij á»īnh +ĠÐł ÑĥÑģ +ĠN ữ +Ġ׾×Ķש ×Ļ×Ĵ +Ġjed noc +Ġjednoc ze +Ġjednocze ÅĽnie +Ġ×Ķ×Ĵ ×ij×ķ×Ķ +أخ ÙĦاÙĤ +ĠнаÑģ ел +ĠнаÑģел ениÑı +ĠÙĬ ÙĨب +ĠÙĬÙĨب غÙĬ +ãģĮ ãģĭ +ãģĮãģĭ ãģĭ +×Ĵ עת +Ðŀ Ðł +ĠналиÑĩ ии +Ġë§Ī ì§Ģ +Ġë§Īì§Ģ ë§ī +Ġíĸī ìĤ¬ +Ġtre ÅĽci +Ġê°Ģ ì¹ĺ +ì¦ ĺ +Ġана лог +×Ķצע ת +в лад +влад е +ĠÑģдел ал +Ġ׳ ×Ĵ×Ļש +Ġ׳×Ĵ×Ļש ×ķת +полн ение +à¸Ĩ à¹Īา +ĠD ön +׼׾׼ ׾×Ķ +×ŀ×ĸ ×Ĵ +Ùħ Ùģ +ÙħÙģ Ùĩ +ÙħÙģÙĩ ÙĪÙħ +×Ķ ×ĵ +×Ķ×ĵ פס +×Ķ×ĵפס ×Ķ +ãģĻãģİ ãģ¦ +Ġг ÑĢ +ĠгÑĢ Ð½ +×ŀ×ĺ ×ķס +Ġ기 ìĸµ +ï¾ Ł +ĠpÅĤ yn +ĠGr ünde +ĠBü cher +Ġwed ÅĤug +ãģ¾ãģł ãģ¾ãģł +Ġ׳×Ķ ×ĵר +ĠÙĬست Ø·ÙĬع +ĠHi á»ĩp +ãĤŃãĥ£ãĥ³ ãĥļ +ãĤŃãĥ£ãĥ³ãĥļ ãĥ¼ãĥ³ +Ġth á»ķ +Ġeuropé enne +à¸ļ ัà¸ĩ +à¸ļัà¸ĩ à¸Ħัà¸ļ +ĠszczegóÅĤ owo +׳ שק +ãĥķ ãĥ©ãĥ³ãĤ¹ +×ŀ×ķ×ŀ ×Ĺ×Ļ +Ġcom ún +Ġç arp +ØŃت ÙĬا +ØŃتÙĬا ج +ØŃتÙĬاج ات +ëĭ´ ëĭ¹ +ä½ķ 度 +ä½ķ度 ãĤĤ +×ĵ ×ij×§ +ãģį ãĤĮ +ãģįãĤĮ ãģĦ +Ġк ам +Ġкам еÑĢ +ĠespecÃŃf ico +Ġtel éfono +à¸ķัà¹īà¸ĩ à¸Ńยูà¹Ī +I Åŀ +ãģ© ãĤĵãģ© +ãģ©ãĤĵãģ© ãĤĵ +עצ ×ŀ×IJ×Ļ +à¸Ķัà¸ĩ à¸Ļีà¹ī +ĠÑĦоÑĢм иÑĢов +ĠÑĦоÑĢмиÑĢов а +×ķ×ŀ ×ij +Ġkullan ımı +Ðľ Ðŀ +×¢ ש×Ļ +עש×Ļ ×Ļ×Ķ +Ġön lem +à¹Ģà¸Ń à¹ĩ +à¹Ģà¸Ńà¹ĩ ม +×ŀשק ×Ļ×¢ +ר ×Ļ×Ĺ +à¸Ĥ ัà¸Ķ +ĠíĻ ľ +ĠíĻľ ìļ© +à¸ĭ ะ +ãĤĪãģĨ ãģ«ãģªãĤĬãģ¾ãģĹãģŁ +ĠÑĢаÑģ пÑĢ +ĠÑĢаÑģпÑĢ Ð¾ÑģÑĤ +ĠÑĢаÑģпÑĢоÑģÑĤ ÑĢан +ĠÑĢаÑģпÑĢоÑģÑĤÑĢан ен +׼×Ļ ×ķף +ÙĤب ض +تص رÙĬØŃ +تصرÙĬØŃ ات +Ġо ÑĢи +ĠоÑĢи г +ĠоÑĢиг ина +ĠоÑĢигина л +ĠاÙĦع اÙĦÙĬ +à¹ģหà¹Īà¸ĩ à¸Ļีà¹ī +ãĥķãĤ¡ ãĥ¼ +ãģ¦ãģĦ ãģį +ãģ¦ãģĦãģį ãģŁãģĦ +פ תר +פתר ×ķ׳×ķת +Ġ×ij ×Ļ×Ĺ +Ġ×ij×Ļ×Ĺ ×ĵ +Ġod by +Ġodby ÅĤ +ĠоÑĩеÑĢ ÐµÐ´ +Ġtr ương +ãĤŃ ãĥ³ +×ŀ ×ķפ +×ŀ×ķפ ×¢ +ëĵľ 립 +ëĵľë¦½ ëĭĪëĭ¤ +à¸ŀืà¹īà¸Ļ à¸IJาà¸Ļ +ìŀIJ 격 +ĠVi á»ĩn +ĠDes pués +Ġ×IJ׾ ×Ļ׳×ķ +Ġdur ée +íĩ ´ +Ġmü zik +i ếu +ĠÑĢаз меÑīен +Ġк Ñĥд +ĠкÑĥд а +غ ض +غض ب +ĠTamb ém +à¸Īัà¸Ķ สà¹Īà¸ĩ +à¸ģาร à¹ģสà¸Ķà¸ĩ +onom ÃŃa +Ġан г +Ġанг ли +Ġангли й +Ġанглий Ñģк +Ġzn al +Ġznal az +Ġznalaz ÅĤ +תר ×Ĵ +תר×Ĵ ×ķ×Ŀ +ĠÑģ нов +ĠÑģнов а +ĠÑĩаÑģ а +Ġcommun auté +ĠespecÃŃf ica +ĠL á»ĭch +Ġli é +Ùģ Ø¬Ø± +à¹Ģà¸ģ à¹Īà¸ĩ +ع اÙĦ +عاÙĦ ج +Ø£ÙĨ ظ +Ø£ÙĨظ ÙħØ© +ES İ +ĠاÙĦØŃ دÙĬد +à¸ŀระ à¸Ńà¸ĩà¸Ħà¹Į +Ġפר שת +Ġдв иж +Ġдвиж ениÑı +ĠاÙĦج ارÙĬ +à¸ĺาà¸Ļ ี +неÑģ ен +ĠاÙĦÙĨ ÙĩائÙĬ +Ġб еÑĢ +ĠбеÑĢ ÐµÐ¼ +ĠбеÑĢем енн +Ġdépart ement +à¹Ģà¸Ĺ ีย +à¹Ģà¸Ĺีย à¸ļ +ĠÐľ аÑĢи +ĠнекоÑĤоÑĢ ÑĭÑħ +об еÑģп +обеÑģп еÑĩен +×Ĺ ×ķ×ĸ +×Ĺ×ķ×ĸ ×Ķ +ÙĨت ج +à¸Īะ à¹Ħà¸Ķà¹īรัà¸ļ +á» ° +Ġél éments +ع Ø· +عط اء +Ġt ắt +i á»ĩm +ÑİÑīиÑħ ÑģÑı +ãģĹãģ ° +ãģĹãģ° ãĤīãģı +Ġпом ожеÑĤ +à¸Ĥà¸ĵะ à¸Ļีà¹ī +Ġ×¢ שר×ķת +éģķ ãģ£ãģ¦ +ĠпÑĢ Ð¾Ð³ +ĠпÑĢог н +ĠпÑĢогн оз +Ġt ÅĤ +ĠtÅĤ um +ĠtÅĤum acz +T ür +Tür kiye +ãģį ãģ£ +ãģįãģ£ ãģĭãģij +Ġ×Ķ׳ ×ķ׼ +Ġ×Ķ׳×ķ׼ ×Ĺ×Ļ +ĠìĥĿ ìĤ° +ĠÑĦоÑĢм Ñĭ +ç¾İ ãģĹãģĦ +à¸Ľà¸£ ึà¸ģ +à¸Ľà¸£à¸¶à¸ģ ษา +Ġlum ière +ãĤª ãĥ¼ãĥĹ +ãĤªãĥ¼ãĥĹ ãĥ³ +à¸Ľ ืà¸Ļ +วั สà¸Ķ +วัสà¸Ķ ุ +еÑĢÑĤ в +ÙĥÙĦ Ùģ +ï½ £ +à¸ĺรรม à¸Ķา +׳ ×ĺר +ĠпÑĢедÑģÑĤав лÑıеÑĤ +Ġanál isis +Ġb ãi +با ÙĤÙĬ +à¸Ľà¸£à¸° à¹Ģà¸Ķ +à¸Ľà¸£à¸°à¹Ģà¸Ķ à¹ĩà¸Ļ +ĠÑģлÑĥÑĩ аÑı +ĠÑģлÑĥÑĩаÑı Ñħ +ÐĽ ÐIJ +สัà¸ĩ à¹Ģà¸ģ +สัà¸ĩà¹Ģà¸ģ à¸ķ +Ġprz ec +Ġprzec ież +Ùħ صÙĦ +ÙħصÙĦ ØŃØ© +ש×ķ×§ ×ķ׾×ĵ +ĠобоÑĢÑĥд ованиÑı +Ġtr waÅĤ +رÙĪ Ùħ +ìķĪ ëĤ´ +ĠNgh á»ĭ +Ø® Ø´ +à¸ļา à¸Ħาร +à¸ļาà¸Ħาร à¹Īา +Ġоп ÑĨион +ĠÑģозд аниÑı +ãĤ³ ãĤ¹ãĥĪ +Ġ×Ķ×¢ ׾×Ļ +Ġ×Ķ×¢×ľ×Ļ ×ķף +lä uft +ãĥĻ ãĤ¹ãĥĪ +Ġr ê +Ġrê ve +×IJ ×ij×Ļ×ij +×Ļ ×Ļ×ļ +ë¶ Ļ +ãĤ¤ãĥ³ ãĥī +ÅĤo ży +ÅĤoży Äĩ +ع ائÙĦ +عائÙĦ Ø© +Ø£ ÙĪØ± +Ø£ÙĪØ± اÙĤ +à¸Ĺà¹īà¸Ńà¸ĩ à¸ĸ +à¸Ĺà¹īà¸Ńà¸ĩà¸ĸ ิà¹Īà¸Ļ +Ġä hn +Ġähn lich +ãĥŁ ãĥĭ +à¸ľ ู +à¸ľà¸¹ à¹īà¸Ļ +à¸ľà¸¹à¹īà¸Ļ ำ +ĠмаÑĤеÑĢиал Ñĭ +Ġкап иÑĤ +ĠкапиÑĤ ал +ï¼ ¦ +Ġseç il +Ġh ứng +Ġintéress ant +ãģ£ãģ¦ ãģĦãģı +Ġe ÄŁer +ëIJĺ ìĹĪìĬµëĭĪëĭ¤ +Ġan laÅŁma +ãģĶ åĪ©ç͍ +Ġ×ij ×ĸ׼ +Ġ×ij×ĸ׼ ×ķת +ëĿ¼ ë©´ +ĠÙĬ ÙĪØ³ +ĠÙĬÙĪØ³ Ùģ +أسÙĦ ØŃØ© +ĠGef ühl +ĠноÑĢм алÑĮн +ãĥĻ ãĥ³ +ãģķãĤĮ ãĤĭãģĵãģ¨ +ĠÐij еÑģ +ãģ¨ãģĦ ãģĪãģ° +ĠÙħ ÙĩÙħ +ĠÙħÙĩÙħ Ø© +ãģ§ãģĹãĤĩãģĨ ãģŃ +ĠêµŃ ëĤ´ +à¹Ģม à¹ĩà¸Ķ +×ŀ×ij קר +ĠاÙĦد ÙĨÙĬ +ĠاÙĦدÙĨÙĬ ا +à¸Ĭ ู +к ÑĢÑĥÑĤ +Ġtho áng +Ġ׳ ×ĵר +Ġ׳×ĵר ש +ĠÑĢаÑģÑģ казал +ĠAu ÃŁerdem +פ ×IJר +פ×IJר ×§ +Ġ×ŀש×Ĺ×§ ×Ļ×Ŀ +צ ר׼×Ļ×Ŀ +×ŀ×ĵ ×ķ +×ŀ×ĵ×ķ ×Ļ×§ +èĭ¦ ãģĹ +ĠÑģ иг +ĠÑģиг нал +ĠM á»įi +Ġtr ữ +Ġnast ÄĻp +ĠnastÄĻp nie +Ġì¶Ķ ì§Ħ +ĠاÙĦÙģ ÙĨد +ĠاÙĦÙģÙĨد ÙĤ +koÅĦ czyÅĤ +ส ีà¹Ī +×§ ×Ļ×ij +×§×Ļ×ij ×ķ×¥ +ĠнÑĥж нÑĭ +大 åĪĩ +大åĪĩ ãģª +æıĽ ãģĪ +ת ×ķס +ת×ķס פת +ãģ£ãģ¦ ãģĦãģªãģĦ +Ġм Ñı +ĠмÑı г +ĠмÑıг к +Ġjak ie +Ġjakie ÅĽ +à¸ķำ à¸ļ +à¸ķำà¸ļ ล +ĠìŀĪ ì§Ģ +×ij×ĺ ×IJ +ĠоÑĤлиÑĩ но +ÙĤ ÙIJ +ĠавÑĤом об +ĠавÑĤомоб и +ĠавÑĤомоби лÑı +دÙĬÙħÙĤرا Ø·ÙĬ +ĠاÙĦ ÙĪØ§ +ĠاÙĦÙĪØ§ ØŃد +Ġس ÙĪØ±ÙĬØ© +Ø£ غÙĦ +أغÙĦ ب +ĠÑįк ÑĢан +ãĥĹ ãĥ©ãĤ¤ +Ġjeste ÅĽ +ãĥIJ ãĥª +Ġ×Ķ×IJ ×ķ×ķ×Ļר +ائ Ùĥ +à¸Ńยà¹Īาà¸ĩ ยิà¹Īà¸ĩ +ÑĢ ÐµÐºÑĤ +Ġum o +Ġumo ż +Ġumoż li +Ġumożli w +Ġumożliw ia +Ġnäch ste +ĠìŀĪ ì§Ģë§Į +ĠпÑĢед н +ĠпÑĢедн аз +ĠпÑĢедназ наÑĩен +Ġma çı +Ġp omi +Ġpomi ÄĻd +ĠpomiÄĻd zy +ĠاÙĦÙĦ ÙĤاء +à¹Ģà¸Ķ à¸Ńะ +Ġнов оÑģÑĤи +×ŀ׊׾×Ķ +رÙĬاض ÙĬ +à¸Ķ à¸Ļ +à¸Ķà¸Ļ à¸ķรี +ب صر +ìĬ¤ íĥĢ +scri pción +Ġnap isa +Ġnapisa ÅĤ +Ġ׳ש ×ŀ×¢ +ĠاÙĦÙħØŃ ÙĦÙĬ +Ġhi á»ĥn +×IJ ×Ĺ +×IJ׊ר×IJ×Ļ +Ġг ÑĢаниÑĨ +æīĭ ç¶ļãģį +Ùĥ سب +Ġà¹ģà¸ķà¹Ī à¸ĸà¹īา +à¸Ķาว à¸Ļà¹Į +à¸Ķาวà¸Ļà¹Į à¹Ĥหลà¸Ķ +ãĤĭãģĵãģ¨ãģĮãģ§ãģį ãģ¾ãģĻ +åŁºæľ¬ çļĦãģ« +ÙĪÙĦ اد +rä ume +د ÙģØ§Ø¹ +×Ļצ ×¢ +ĠO czy +ĠOczy wiÅĽcie +ĠÅ ģ +ĠÅģ a +اÙĦÙĬ اب +اÙĦÙĬاب اÙĨ +áºł I +ĠBir liÄŁi +×Ķ ×ķצ +×Ķ×ķצ ×IJת +ĠÄij ua +Ġê·¸ëŁ¬ ëĭĪê¹Į +Ġréal ité +ع ÙĦاÙĤات +J este +Jeste ÅĽ +Ġмн ож +Ġмнож еÑģÑĤво +ï¼ « +ãĥĹãĥŃ ãĤ¸ãĤ§ +ãĥĹãĥŃãĤ¸ãĤ§ ãĤ¯ãĥĪ +ĠÑĦ л +ظ ÙĨ +×Ĵ׾ ×Ĵ׾ +ĠmÅĤod zie +ĠmÅĤodzie ż +à¸Ļà¹īำ à¸ķา +à¸Ļà¹īำà¸ķา ล +ÐĽ Ðķ +×ij ×ķ×ĺ +Ġ׾×Ķ ×Ĵ×Ļ×ĵ +ãģĵãģ¨ãĤĤ ãģĤãĤĭ +ز اد +×ŀ×Ļ×ĵ ×¢ +ĠgÅĤówn ie +ãĥı ãĤ¦ +ãĥıãĤ¦ ãĤ¹ +б ел +Ġét ape +ðŁĺ Ģ +Ġмод елÑĮ +a ģını +ש ×Ĺ×§ +ש×Ĺ×§ ף +Ġni ño +à¸Ĭ à¹īาà¸ĩ +à¹Ģล ีย +ĠÑĦоÑĢм е +ĠاÙĦØ´ رÙĬÙģ +ĠÑĥд аÑĢ +arr iv +arriv ée +Ġmies iÄĻ +ĠmiesiÄĻ cy +ØŃ رÙĥ +ØŃرÙĥ ات +ĠDi á»ħn +ÐĿ Ы +ãģ¾ãģ£ãģŁ ãģı +Ġ×Ļ ×¨×ķ×§ +еÑģÑĤ еÑģÑĤв +еÑģÑĤеÑģÑĤв енн +Ġê·¸ ëŁ¼ +ĠاÙĦÙħ تÙĪ +ĠاÙĦÙħتÙĪ Ø³Ø· +Ġbéné fic +Ġbénéfic ie +Ġwy bra +Ġwybra Äĩ +ĠاÙĦز ÙħÙĨ +ĠпÑĢин Ñı +ĠпÑĢинÑı л +Ù쨱 ØŃ +Ġk sz +Ġksz taÅĤ +ĠksztaÅĤ t +ק׾ ×ĺ +×ij×ĵ×Ļ×§ ת +Ġgi ấ +Ġgiấ c +Ġpropriet Ãł +деÑĢж ан +ĠKö ln +ĠGü zel +×Ļפ ×ķ×Ļ +ĠCu á»Ļc +ÑįÑĤ аж +تر ÙĥÙĬ +ترÙĥÙĬ ز +лож ений +Ġп Ñĥ +ĠпÑĥ ÑĤи +اخت ÙĦاÙģ +åĩºãģ¦ ãģıãĤĭ +à¸ļุ à¸ģ +âĿ ¤ +ÑĦ ан +פש ×ĺ +à¸ļัà¸Ļ à¹Ģà¸Ĺ +à¸ļัà¸Ļà¹Ģà¸Ĺ ิà¸ĩ +ĠاÙĦس اد +ĠاÙĦساد س +ĠاÙĦÙĤ ÙĪÙħ +ĠاÙĦÙĤÙĪÙħ ÙĬ +Ġyönet ici +Ùĩ ÙĪØ§Øª +ÙĩÙĪØ§Øª Ùģ +Ġrespons ável +Ġпод деÑĢжива +ĠاÙĦسÙĦ Ø· +ĠاÙĦسÙĦØ· ات +ãģĹãģ¦ ãģĬãģı +ãĥļ ãĥĥãĥĪ +à¸Ľ ุà¹Īม +Ġogl Äħda +ÙĨا ÙĤ +ÙĨاÙĤ Ø´ +à¸Ħà¸Ńà¸Ļ à¹Ĥà¸Ķ +ĠMü sl +ĠMüsl ü +ĠMüslü man +ĠMo ż +ĠMoż na +Ġnum érique +Ġv á»ı +ĠسÙĬ تÙħ +Ġyer leÅŁ +монÑĤ аж +Ġgo ût +ãģ¦ ãģĬãĤĬãģ¾ãģĻ +ĠKh ánh +Ġе дин +Ġедин ÑģÑĤв +اÙĨ Ø®Ùģ +اÙĨØ®Ùģ Ø§Ø¶ +ìĭľ íĹĺ +Ġl ặng +ĠÑĢ Ð¾Ð»ÑĮ +à¸ķัว à¹ģà¸Ĺà¸Ļ +à¸Ħà¹Īา à¹ĥà¸Ĭà¹ī +à¸Ħà¹Īาà¹ĥà¸Ĭà¹ī à¸Īà¹Īาย +Ġver füg +Ġverfüg bar +ìĻĶ ëĭ¤ +ãģĦ ãģļ +ãģĦãģļ ãĤĮ +ĠиÑģÑģлед ованиÑı +меÑī а +×Ķ ×Ĺ +×Ķ×Ĺ ×ĸר +à¹ģà¸Ł à¸Ĭัà¹Īà¸Ļ +ت صرÙģ +Ø¥ رÙĩاب +Ġexerc ÃŃcio +Ġé lev +Ġélev é +สัà¸įà¸įา à¸ĵ +Ãĸ Z +ãĥĹ ãĥŃãĤ° +ãĥĹãĥŃãĤ° ãĥ© +ãĥĹãĥŃãĤ°ãĥ© ãĥł +Ġw ewnÄĻtrzn +Ġhen üz +é£Ľ ãģ³ +à¹Ģà¸Ķ à¸Ńรà¹Į +Ñģ Ñĥж +ÑģÑĥж ден +شع ÙĪØ¨ +ãģ²ãģ¨ ãĤĬ +Ġwy ÅĤÄħ +ĠwyÅĤÄħ cznie +Ġпло Ñħо +ÐĶ Ðķ +Ạ¦ +Ù쨹 اÙĦÙĬ +ÙģØ¹Ø§ÙĦÙĬ ات +ĠاÙĦع شر +ÑģÑĤÑĥп ил +Ġy arg +Ġyarg ı +нÑİ Ñİ +×ķ×IJ ×ij +Ġu ç +Ġuç ak +ë² ½ +تÙĪ ÙĤÙĬ +تÙĪÙĤÙĬ ع +Ġì¤ij ìĭ¬ +׳×Ļ×ķ ×ķ×ĺ +Ø£ ÙĥÙĦ +ç½® ãģĦãģ¦ +éłĤ ãģį +Ġ×Ķת ×ij +Ġ×Ķת×ij ×Ļ×¢×Ķ +Ġdür fen +Ùħ ÙĤاÙĦ +ÙħÙĤاÙĦ ات +Ġز ÙħÙĨ +à¸ŀฤ ศ +à¸ŀฤศ à¸Ī +à¸ŀฤศà¸Ī ิà¸ģ +à¸ŀฤศà¸Īิà¸ģ ายà¸Ļ +ĠнеÑģк олÑĮ +ĠнеÑģколÑĮ ки +ĠнеÑģколÑĮки Ñħ +Ġcrian ça +มิ à¸ķร +×ŀ׼ ×Ļר×ķת +à¸ģาร à¸ļริหาร +Ġtélé charg +Ġ×IJ×ķ×Ķ ×ijת +ĠBü ro +ä½ľ ãģ£ãģŁ +ĠKi ÅŁi +ç¾İåij³ ãģĹ +à¹Ģลย à¸Ħà¹Īะ +à¸ŀà¸ļ à¸ģัà¸ļ +à¸Ī à¹īา +Ġç er +Ġçer ç +Ġçerç eve +ãĤĴä½ľ ãģ£ãģ¦ +ĠпеÑĢв ÑĥÑİ +×ŀצ ר×Ļ×Ŀ +×IJ׾ ×ķ×Ķ +×IJ׾×ķ×Ķ ×Ļ×Ŀ +Ġagr é +Ġagré able +Ġay ır +İL İ +ãĤ ¥ +Ġíĺ Ħ +ĠíĺĦ ìĭ¤ +ثاÙĦ Ø« +ת ×ĸ +ת×ĸ ×ķ׳×Ķ +ãģ¨ãģĦ ãģ£ãģ¦ +ãģ¨ãģĦãģ£ãģ¦ ãĤĤ +Ġا بÙĪ +ĠÑģоб ак +é£Łãģ¹ ãģŁ +Ġдан ном +à¹Ģล ิ +à¹Ģลิ ศ +Ġí ļ +Ġíļ ¨ +Ġíļ¨ ê³¼ +ãĤĤãĤī ãģĪãĤĭ +׳ צ׾ +ÑĦ ик +ÑĦик Ñģ +Ġjeste ÅĽmy +ת×Ĺ×ķש ×Ķ +à¹Ħมà¹Ī à¸Ħวร +ĠØŃ سÙĬÙĨ +à¸ģาร ลà¸ĩà¸Ĺุà¸Ļ +ë´ ¤ +ĠÐĺ менно +à¸ļ à¸Ńรà¹Į +à¸ļà¸Ńรà¹Į à¸Ķ +ĠC ảnh +ìĦľ ë¹ĦìĬ¤ +Ġпол ов +Ġполов ин +Ġзам еÑĩа +ãģĦãĤį ãĤĵãģª +Ġ×ij ×Ļ×§ +Ġ×ij×Ļ×§ ש +л ÑĥÑĪ +ãĤĴ è¿İ +ãĤĴè¿İ ãģĪ +جرÙĬ ÙħØ© +Ġt ây +ĠاÙĦÙĨ ÙĪ +ĠاÙĦÙĨÙĪ ÙĪÙĬ +ÃĤ N +ì¿ ł +หà¸Ļ าว +Ġ×ij׊ש×ij×ķף +ز ار +à¸Ķ าร +à¸Ķาร า +ĠÅĽ l +ĠÅĽl ub +มีà¸Ħวาม สุà¸Ĥ +Ġn hu +Ġnhu áºŃn +ÙħØŃ طة +à¹Ģสืà¹īà¸Ń à¸ľà¹īา +ĠТ олÑĮко +ĠÙĥ س +ĠÙĥس ارة +ÙħØ´ رÙĪØ¹ +niÄĻ cia +×¢ ׼ש×Ļ×ķ +ت ÙĦÙģ +تÙĦÙģ Ø²ÙĬ +تÙĦÙ쨲ÙĬ ÙĪÙĨ +Ġl Æ°á»Ľi +ĠÐľÐ¾Ñģк вÑĭ +Ġré serve +Ġan laÅŁ +ĠanlaÅŁ ıl +Ġed eceÄŁi +รà¸Ńà¸ĩ à¹Ģà¸Ĺà¹īา +Ġب Ø· +Ġبط رÙĬ +ĠبطرÙĬ ÙĤØ© +ãģ¦ãģĹãģ¾ ãģ£ãģ¦ +ãĤĤãĤī ãģ£ãģ¦ +بر ج +æ± ļ +æ±ļ ãĤĮ +Ġch oc +Ġchoc ia +Ġchocia ż +Ġzob ac +Ġzobac zyÄĩ +пÑĢ Ñı +пÑĢÑı жен +ĠÑĨ иÑĦ +ĠÑĨиÑĦ ÑĢ +Ġм ам +Ġвз ÑıÑĤÑĮ +Ġch ạm +ج سÙħ +ØŃÙħ اس +à¹Ģล à¹Īม +à¸ŀิ ษ +×Ķפ ׼×ķ +à¸Ĭà¹Īà¸Ńà¸ĩ à¸Ĺาà¸ĩ +Ġв ек +Ġвек а +Æ¡ Ìģ +Æ¡Ìģ i +ĠTi á»ģn +Ġtr ầm +мÑĭ ÑĪ +мÑĭÑĪ Ð» +ĠÑĤ Ñĥ +ĠÑĤÑĥ ÑĢиÑģÑĤ +Ġch c +Ġchc Äħ +Ġав г +Ġавг ÑĥÑģÑĤ +ĠавгÑĥÑģÑĤ а +ס ×IJ×ķת +Ġר ×Ĵ׾ +à¸ľà¸¥ à¸ģระà¸Ĺ +à¸ľà¸¥à¸ģระà¸Ĺ à¸ļ +å¤īãĤı ãĤĭ +Ġ×Ķ×IJ×Ĺר ×ķ׳×Ļ×Ŀ +سÙģ ÙĬر +ĠÑĩа Ñīе +ãģĦ ãĤī +ãģĦãĤī ãģ£ +ãģĦãĤīãģ£ ãģĹãĤĥ +×ķ×ŀ ׳×Ļ×Ŀ +Ġart tır +ĠCh á»ĭ +Ġì¡° ì§ģ +ĠÑĥÑģп еÑħ +Ġ×¢ ×ķס +Ġ×¢×ķס ×§ +ĠìĥĿ ëªħ +ÑĨ иÑĤ +Ġreg ión +Ðŀ ÐĿ +ĠdoÄŁ um +ĠyaÅŁ ad +ĠyaÅŁad ıģı +à¸Ĺà¸Ķ ลà¸Ńà¸ĩ +Ġgöz ü +ש ×Ļר×Ķ +дÑĥм ал +Ġda ģı +Ġdaģı t +à¸Ĺีม à¸ĩาà¸Ļ +Ġti á»ģm +ĠاÙĦÙĥ بر +ĠاÙĦÙĥبر Ùī +ì¹ Ń +ĠGü nc +ĠGünc elle +ĠGüncelle me +ê¹ Ĭ +ĠобоÑĢÑĥд ование +ĠÑĢеÑĪ Ð° +á» ¤ +Ġп иÑĤ +ĠпиÑĤ аниÑı +à¹Ģรีย à¸ļ +×Ľ×ª ×Ļ×ij×Ķ +Ġп он +Ġпон ÑĢав +ĠпонÑĢав и +Ġ×Ķ ×ķ׾×ĵ +Ġ×Ķ×ķ׾×ĵ ת +Ġê² ģ +Ġê²ģ ëĭĪëĭ¤ +ĠпеÑĢв ой +ãĥ©ãĤ¤ ãĥķ +ĠÅŁi ir +kr ÄĻ +krÄĻ c +Ġthi á»ĥu +à¹Ģลย à¸Ĺี +à¹Ģลยà¸Ĺี à¹Ģà¸Ķียว +×ĺ×¢ ׳×ķת +ائ ÙĩÙħ +Ġ×IJ ס×ķר +ĠплаÑĤ еж +تر دد +Ġmożli we +Ġkh Ỽ +ĠkhỼ p +تÙģØ§Ø¹ ÙĦ +ĠÑĪ ÐºÐ¾Ð»ÑĮ +ĠÑĪколÑĮ н +ĠÙĤ صة +Ġmét ier +nÄĻ ÅĤa +หล à¹Īà¸Ń +Ġ á»§ng +Ġprz egl +Ġprzegl Äħd +ĠاÙĦÙħ تعÙĦ +ĠاÙĦÙħتعÙĦ ÙĤØ© +ĠÑģÑĭ н +Ġв олн +ãĥĩ ãĥ¼ãĥĪ +ĠÐŃ ÑĤи +Ġк ÑĢоме +à¸Ħ ารà¹Į +׳ק ×ķ×ĵ×Ķ +Ġ׾ש×ŀ ×ķ×¢ +Ġ×ĸ ×ķ׼ר +ï¼ § +ÙĬ ÙİØ§ +Ġgi á»ıi +åĥį ãģı +ĠÑģ ни +ĠÑģни жен +à¹ģà¸Ķ à¸Ķ +รุ à¸Ļ +รุà¸Ļ à¹ģรà¸ĩ +Ġhi á»ĩp +ograf ÃŃa +à¹Ģà¸Ī à¸Ńรà¹Į +Ġдв иг +Ġдвиг аÑĤ +ĠдвигаÑĤ ел +Ġü y +Ġüy eler +Ġüyeler i +Ġб Ñĥк +ĠбÑĥк в +ãĤĤ å¤ļãģı +Ġthi á»ĩt +ĠPa ÃŃs +ĠØ· بÙĬعÙĬ +à¹ģà¸Ī à¸ģ +ĠاÙĦص ØŃÙĬØŃ +Ġapp ré +Ġappré ci +Ġdecis ión +Ġë°ĺ ëĵľ +Ġë°ĺëĵľ ìĭľ +ĠÑĤеб е +ãĤ· ãĥ¼ãĤº +ãĤ·ãĥ¼ãĤº ãĥ³ +Ġд алÑĮн +ĠìĬ ¤ +ĠìĬ¤ ìĬ¤ +ĠìĬ¤ìĬ¤ ë¡ľ +ĠTh á»ĥ +Ġkar ÅŁ +ĠkarÅŁ ıs +ĠkarÅŁÄ±s ında +ĠK ön +ĠKön ig +ив ание +×ij ×ķצע +г лаÑģ +Ġtw ó +Ġtwó rc +à¸Ľà¸ģ à¸Ħร +à¸Ľà¸ģà¸Ħร à¸Ńà¸ĩ +ĠG ÅĤ +ĠGÅĤ ówn +ĠUnter stüt +ĠUnterstüt zung +Ġд ÑĥÑħ +ĠдÑĥÑħ ов +Ø£ ÙħاÙĨ +×Ĺש ש +ت ظ +تظ اÙĩر +ĠлÑİб ом +à¸ķ าร +à¸ķาร าà¸ĩ +Ġkr ól +Ø£ ØŃدث +ì¡Į ëĭ¤ +Ðļ ÑĥÑĢÑģ +ãĥĥ ãĥĦ +×ŀ×§ ×ķ×ij׾ +ĠÑģимв ол +Ġdés orm +Ġdésorm ais +w üns +wüns che +Ñĥ ни +Ñĥни ÑĨип +ÑĥниÑĨип алÑĮн +หลัà¸ģ สูà¸ķร +ÙĨت شر +Ġа л +Ġал к +Ġалк ог +Ġалког ол +ĠÑĥ ÑĩиÑĤÑĭва +à¸ģำ à¸ģัà¸ļ +Ġ׾ פע×ķ׾ +ĠìŰ ê²° +s Äħd +ĠاÙĦØ£ ÙĬ +ĠاÙĦØ£ÙĬ اÙħ +غÙĬ اب +Ġна ÑĢ +ĠнаÑĢ ÐºÐ¾ +×ŀ×ķ×ĵ ×¢ +ĠÑģеÑĢ Ð¸Ð¸ +пиÑģ Ñĭва +สิ ว +ç¶ļ ãģĦãģ¦ +çͳãģĹ è¾¼ãģ¿ +Ġ׾ ×Ĵר +Ġ׾×Ĵר ×ķ×Ŀ +Ġд ем +Ġдем о +Ġë³´ ëĤ´ +تÙĩ دÙĬد +ĠÙħØ´ ÙĬرا +Ġdu y +Ġduy á»ĩt +ĠwiÄĻks ze +Ùħع اÙĬ +ÙħعاÙĬ ÙĬر +ĠG da +ĠGda ÅĦsk +Ġr ah +Ġrah ats +Ġrahats ız +ר ×ķצ×Ķ +l ös +lös ung +ĠТак им +ÑĪ ÐµÐ´ +ÑĪед ÑĪ +ع زÙĦ +Ġרש ×Ļ×ŀת +Ġ׾×Ķ ×Ļ׼ +Ġ׾×Ķ×Ļ׼ ×ł×¡ +Ġп ÑĥÑĤ +ĠпÑĥÑĤ еÑĪ +ĠпÑĥÑĤеÑĪ ÐµÑģÑĤв +Ġnot ÃŃcia +Ġal Ä±ÅŁ +ĠalÄ±ÅŁ ver +ĠalÄ±ÅŁver iÅŁ +ĠwÅĤ os +ĠwÅĤos ów +Ġب غ +Ġبغ داد +Ġver öffent +Ġveröffent licht +ĠKh á +Ġt án +ëIJĺ 기 +Ġë°© 문 +Ùģ ÙĬÙĦ +à¹Ģà¸ģิà¸Ķ à¸Īาà¸ģ +åı¯ æĦĽ +åı¯æĦĽ ãģĦ +à¸ĸ ุà¸ĩ +Ġz ewnÄĻtrzn +à¸łà¸²à¸©à¸² à¸Ńัà¸ĩà¸ģฤษ +Ġmá xima +Ġul us +Ġulus lararası +Ġ׳×Ķ ×ł +à¸Ĥà¹Īาว สาร +ĠìĿĺ ìĤ¬ +à¹Ģหล ืà¸Ńà¸ĩ +Ġد ÙĤ +ĠدÙĤ ائÙĤ +สืà¹Īà¸Ń สาร +ë¨ ¼ +ĠÑģоÑģÑĤоÑı нии +สมา à¸Ħม +á» Ĥ +ĠÐľÐ¾Ñģ ков +ĠÐľÐ¾Ñģков Ñģк +×ŀס ×ķ×Ĵ׾ +ãģĭ ãģĭãĤĬ +ĠTr uyá»ģn +à¹ģà¸Ĥà¹ĩà¸ĩ à¹ģรà¸ĩ +×ŀ×Ĺ ×ĸ×Ļ×§ +à¹Ĥà¸ģ à¹ī +ÙĬس ر +ìĶ © +×IJ ×ķ×§ +×IJ×ķ×§ ×ĺ +×IJ×ķ×§×ĺ ×ķ×ijר +Ġprox imité +ÙħÙĨ Ùĩج +ĠاÙĦج ز +ĠاÙĦجز ائ +ĠاÙĦجزائ رÙĬ +ĠÄIJi á»ĥm +Ġден еж +Ġденеж н +ÙģØŃ ص +Ùģ Ø¦ +ĠÐij Ñĥд +×Ĵ×Ļ×ĵ ×ķ׾ +ĠÐĴ едÑĮ +عÙĦ اÙħØ© +Ġ×IJ×Ĺר ×ķ׳×ķת +ãģĦãģŁãģł ãģĦãģ¦ +سÙĦ ØŃ +ØŃ ÙĦÙħ +ز ÙĪØ§Ø± +Ùĥ سر +×ĺ קס +Ġб ан +Ġбан ков +ĠпÑĢ Ð¾Ð¶ +ĠпÑĢож ива +li wo +liwo ÅĽci +ĠTi ếp +ĠاÙĦÙħÙĨ اسب +ĠاÙĦØ® ÙĬار +ãģĬ ãģĭ +ãģĬãģĭ ãģĴ +à¸Ķà¸Ńà¸ģ à¹Ħมà¹ī +ä mp +ämp fe +à¸ķัà¹īà¸ĩ à¹ĥà¸Ī +Ġза ÑīиÑĤ +ĠзаÑīиÑĤ Ñĭ +ĠTh ưá»Ŀng +Ġص Ùģ +ĠصÙģ ØŃØ© +×Ĺ×ķר ×£ +ãĥIJ ãĥĥãĤ° +Ġ×ĵ ×Ļ×Ĵ +Ġ×ĵ×Ļ×Ĵ ×Ļ×ĺ +Ġ×ĵ×Ļ×Ĵ×Ļ×ĺ ׾×Ļ +Ġ×Ķ×Ĺ ×ķ׾×Ļ×Ŀ +в еÑī +веÑī а +Ġк ÑĥлÑĮÑĤ +ĠкÑĥлÑĮÑĤ Ñĥ +ĠкÑĥлÑĮÑĤÑĥ ÑĢÑĭ +ĠاÙĦاÙĨ ترÙĨت +Ġhö ch +Ġhöch st +Ġíĺ ķ +Ġíĺķ íĥľ +Ġв ой +Ġвой нÑĭ +ÐĽ Ðŀ +ìĭł ìļ© +Ġ×ŀ×ij ×ķס +Ġ×ŀ×ij×ķס ס +×ŀ׳ ×Ļ×¢ +Ġfiyat ı +ĠÑģл Ñĥж +ĠÑģлÑĥж бÑĭ +à¸Ĺั ศ +à¸Ĺัศ à¸Ļ +ãģĵãģ¨ãģĮ å¤ļãģĦ +Ġ×Ķ×ŀש ת +Ġ×Ķ×ŀשת ×ŀש +å¯Ħ ãģĽ +×ŀש׾ ×ķ×Ĺ +æĻĤ çĤ¹ +æĻĤçĤ¹ ãģ§ +à¸ŀร ี +à¸ŀรี à¹Ģมีย +à¸ŀรีà¹Ģมีย รà¹Į +à¸ŀรีà¹Ģมียรà¹Į ลีà¸ģ +Ġdiffic olt +Ġdifficolt Ãł +ãĥ¬ ãĤ¹ãĥĪ +ãĥ¬ãĤ¹ãĥĪ ãĥ©ãĥ³ +สม à¹Ģà¸Ķà¹ĩ +สมà¹Ģà¸Ķà¹ĩ à¸Ī +Ġж ид +Ġжид к +Ġzu peÅĤ +ĠzupeÅĤ nie +ĠÙħ جر +ĠÙħجر د +ãģĮ å§ĭ +ãģĮå§ĭ ãģ¾ +ãĤŃãĥ£ ãĥ© +Ġ×IJ ×ķ×ķ×Ļר +ãģĬ äºĴ +ãģĬäºĴ ãģĦ +Ġpot rÃł +ĠPa ÅĦst +ĠPaÅĦst wo +Ġب ÙĬاÙĨ +ĠبÙĬاÙĨ ات +Ġин огда +ĠÑĢ Ð° +ĠÑĢа ÑģÑĤв +ĠÑĢаÑģÑĤв оÑĢ +Ġ×ĸ ×ŀ׳ +ยิ à¹īม +Ä Ĩ +ãģ¾ ãģķ +ãģ¾ãģķ ãģ« +ãĥķãĤ¡ ãĤ¤ãĥ« +Ġgörd Ã¼ÄŁÃ¼ +สà¸ĩ à¸Ħร +สà¸ĩà¸Ħร าม +ĠArk adaÅŁ +ĠrozwiÄħz ania +×ŀ ×ķ×ĺ +pi ÄĻ +piÄĻ t +ص غر +ส ย +สย าม +ãĤĨ ãģ£ãģıãĤĬ +Ġtr ần +Ġeconom ÃŃa +Ġgeh ören +ãĤ·ãĥ§ ãĥ¼ +ĠsÅĤ ucha +à¸ŀà¸Ń à¹ĥà¸Ī +ĠоÑĤмеÑĤ ил +ÙĨت ÙĤÙĦ +Ġprop ósito +ĠваÑĪ ÐµÐ³Ð¾ +Ġnh ắn +à¹ģà¸ĸ ว +Ġком иÑģ +ĠкомиÑģ Ñģи +waż nie +Ġy avaÅŁ +×ŀ ×Ļ×§ +×ŀ×Ļ×§ ×ķ×Ŀ +ש×IJ׾ ת +Ġyıll arda +ĠÐ ® +ĠЮ ÑĢ +×ł×¡ ×Ļ×ij×ķת +ת צ +תצ ×ķ×Ĵ +Ġод нÑĥ +Ġ à¸Ńยà¹Īาà¸ĩà¹Ħร +Ġà¸Ńยà¹Īาà¸ĩà¹Ħร à¸ģà¹ĩà¸ķาม +ëģ ¼ +à¹Ħล à¹Ī +تس ÙĦÙĬÙħ +بÙĦ اغ +Ġì ī +Ġìī ½ +Ġìī½ ê²Į +ãĥļ ãĥ³ +зв ÑĥÑĩ +ĠW äh +ĠWäh rend +Ġ×Ļ ×Ļת +Ġ×Ļ×Ļת ׼ף +Ġkh uyên +Ġv ẽ +Ġа меÑĢ +ĠамеÑĢ Ð¸Ðº +ĠамеÑĢик ан +ĠамеÑĢикан Ñģк +ع جب +ãĥĽãĥ¼ãĥł ãĥļãĥ¼ãĤ¸ +Ġник ÑĤо +ĠÙĤ Ùİ +ĠÙĤÙİ Ø§ÙĦ +ĠÙĤÙİØ§ÙĦ Ùİ +ÐIJ ÐĹ +Ùħ جÙħÙĪØ¹ +ÙħجÙħÙĪØ¹ ات +Ġnecess itÃł +Ġpob li +Ġpobli żu +Ġph ấn +ĠСо обÑī +ÙħÙĤ اط +ÙħÙĤاط ع +Ġ×Ķצ ×ķר×ļ +la ÅŁtırma +ว ิà¸Ķ +วิà¸Ķ ี +วิà¸Ķี à¹Ĥà¸Ń +Ġ그리 ìĬ¤ +Ġ그리ìĬ¤ ëıĦ +ãĤ¿ãĤ¤ ãĥŁ +ãĤ¿ãĤ¤ãĥŁ ãĥ³ãĤ° +×§×ĺ ×Ĵ×ķר +×§×ĺ×Ĵ×ķר ×Ļ×Ķ +Ġ×Ĺ ×ķפ +Ġ×Ĺ×ķפ ש×Ļ +Ø£ جر +Ġим ени +ĠÑĢан ее +à¹Ģà¸ŀืà¹Īà¸Ńà¸Ļ à¹Ĩ +ĠJes ús +Ñģо един +Ñģоедин ен +Ġר ×Ĺ×ķ×§ +à¹Ĥà¸ļ รา +à¹Ĥà¸ļรา à¸ĵ +ĠH Æ¡n +Ġth áºŃp +تع ÙĬÙĬÙĨ +Ġtart Ä±ÅŁ +ĠtartÄ±ÅŁ ma +ĠGes pr +ĠGespr äch +תר ×ķפ +תר×ķפ ×ķת +Ġcat égorie +Ġоказ Ñĭва +ĠналиÑĩ ие +Ġprésent é +Ġk ull +Ġkull and +Ġkulland ı +Ġü nl +Ġünl ü +ĠÙģ Ùĥرة +из аÑĤоÑĢ +×IJ ×ķ׳ +×IJ×ķ׳ ×Ļ×ij +×IJ×ķ׳×Ļ×ij רס +×IJ×ķ׳×Ļ×ijרס ×Ļ×ĺת +ĠÑĢаÑģÑģ маÑĤ +ĠÑĢаÑģÑģмаÑĤ ÑĢ +ĠÑĢаÑģÑģмаÑĤÑĢ Ð¸Ð²Ð° +تÙĥÙĦ Ùħ +Ùĥت رÙĪ +ÙĥترÙĪ ÙĨÙĬ +ĠÑģо ÑĩеÑĤ +ĠÑģоÑĩеÑĤ а +ãĤĴè¦ĭ ãģĽ +Ġng ừa +ĠÐł еÑģп +ĠÐłÐµÑģп Ñĥб +ĠÐłÐµÑģпÑĥб лик +ãĤ¦ ãĤ© +ãĤ¦ãĤ© ãĥ¼ +ĠÐľ еждÑĥ +ĠìŀĪ ê²Į +Ġm â +ĠìļĶ ì²Ń +ض ار +ลุ à¹īà¸Ļ +ëĮĢ íķĻêµIJ +×ĸ ×Ļ׼ +×ĸ×Ļ׼ ר×ķף +ãĤ¹ ãĥļ +ãĤ¹ãĥļ ãĥ¼ãĤ¹ +ĠкÑĢаÑģ оÑĤ +ï¼ ¨ +ê¼ Ń +ãĤĴ éĽĨ +ãĤĴéĽĨ ãĤģ +ë° Ŀ +Ġ×Ķ׳ ×IJ +Ġ×Ķ׳×IJ ש×Ŀ +Ġê°Ģ ìļ´ +Ġê°Ģìļ´ ëį° +تÙĥÙĦ Ù쨩 +ĠØŃ ÙĤÙĬÙĤÙĬ +Ġh alk +Ġhalk ın +ÑİÑī ÑĥÑİ +ĠÑģп ин +סר×ĺ ף +ĠпеÑĢв ого +Ġпол ож +Ġполож иÑĤелÑĮн +Ġд л +Ġдл иÑĤелÑĮн +ĠV Ä©nh +ê´ ´ +ĠÑģÑĭ ÑĢ +ĠíĨµ íķĺìŬ +ë³ij ìĽIJ +à¹Ĥรà¸ĩ à¸ĩาà¸Ļ +รัà¸ļ à¸ľà¸´à¸Ķ +รัà¸ļà¸ľà¸´à¸Ķ à¸Ĭà¸Ńà¸ļ +تج ÙĨب +s ÅĤ +sÅĤ uch +ãĤ¢ãĥ« ãĥIJ +ãĤ¢ãĥ«ãĥIJ ãĥł +ëī´ ìĬ¤ +Ġpat ië +Ġpatië nt +Ġìĺ ¤í +Ġìĺ¤í ŀ +Ġìĺ¤íŀ Ī +Ġìĺ¤íŀĪ ëł¤ +ĠDer ne +ĠDerne ÄŁi +wró ci +wróci Äĩ +Ġоб Ñī +ĠобÑī еÑģÑĤв +ĠобÑīеÑģÑĤв енно +ĠêµIJ ìĪĺ +tıģ ımız +Ġ×Ķ×ŀש ×Ļ×ij +k örper +Ġпозв ол +Ġпозвол иÑĤ +ĠChi ến +أخ ÙĪ +ĠAy dın +à¸Ķà¹īาà¸Ļ ล +à¸Ķà¹īาà¸Ļล à¹Īาà¸ĩ +Ġdr u +Ġdru ż +Ġdruż yn +Ġë°ľ íijľ +ĠTh ảo +جÙĩ اد +à¸ģระà¸Ĺ ูà¹ī +Ġк ÑĢов +ĠкÑĢов и +Ġiçer ik +Ġnad zie +Ġnadzie jÄĻ +ĠС моÑĤÑĢ +Ġph ức +ج تÙħاع +جتÙħاع ÙĬØ© +ком пон +компон енÑĤ +Ġб ил +Ġбил еÑĤ +ãĥIJ ãĥ³ãĥī +ĠPol ÃŃcia +اÙĦ تÙĩ +اÙĦتÙĩ اب +ØŃر Ùģ +ت خط +تخط ÙĬØ· +ãĤ³ ãĥ¼ãĥ +ãĤ³ãĥ¼ãĥ Ĵ +ãĤ³ãĥ¼ãĥĴ ãĥ¼ +・・ ï½¥ +à¸ĭ à¸Ńย +Ġcréd it +è²· ãģ£ãģŁ +ĠпоÑĢ Ñıд +ĠпоÑĢÑıд ке +Ġph ó +Ġw ida +Ġwida Äĩ +جر ائÙħ +à¸ľ ี +ĠbÄĻd ÄĻ +Ġ×ŀ פת×Ĺ +ãĥij ãĥ¼ãĥ +ãĥijãĥ¼ãĥ Ĩ +ãĥijãĥ¼ãĥĨ ãĤ£ +ãĥijãĥ¼ãĥĨãĤ£ ãĥ¼ +ĠKa ż +ĠKaż dy +ĠнеобÑħодим оÑģÑĤи +à¸Ł à¸Ńรà¹Į +à¸Łà¸Ńรà¹Į ม +Ġмал ÑĭÑĪ +Ġпл оÑĤ +ĠÑĥ ÑģÑĤÑĢой +ĠÑĥÑģÑĤÑĢой ÑģÑĤва +à¸ĸ à¸Ńà¸Ļ +ĠoluÅŁtur ul +ĠÅĽwi ad +ĠÅĽwiad om +Ùħع Ùĩد +ĠпÑĢоиз веден +Æ ł +ר ×Ļש +Ùħست Ø« +Ùħستث Ùħر +׳×Ļ ×Ļר +pa ñ +Ġ; -) +Ġë°ľ 견 +Ġgör üyor +Ùħؤ ÙĦÙģ +ĠÄIJ á»ģ +ĠاÙĦÙĨ ÙĪØ§Ø¨ +×Ĺ×§ ×Ļר×Ķ +Ġm á»ıi +è¿° ãģ¹ +ÐĿ ик +ìŀĸ ìķĦ +ìŀĸìķĦ ìļĶ +prowadzi ÅĤ +l óg +lóg ica +פס ×ĺ +פס×ĺ ×Ļ×ij׾ +Ġ×ŀ ×ĵ×Ķ +Ġ×ŀ×ĵ×Ķ ×Ļ×Ŀ +ãģĵãģĵ ãģ¾ãģ§ +×Ķ ×ª×Ĺ +×Ķת׊׾×Ķ +Ġפ ×ķס +Ġפ×ķס ×ĺ×Ļ×Ŀ +Ġн ев +Ġнев оз +Ġневоз можно +ĠdostÄĻp ny +Ġغ اÙĦ +ĠغاÙĦ ب +Ġbez pieczeÅĦst +ĠbezpieczeÅĦst wa +åĪĨ ãģĭãĤĭ +ĠF ührung +à¸ģ ีà¹ī +gem Ã¤ÃŁ +à¸Ĭà¹Īวà¸ĩ à¹Ģวลา +Ġìļ°ë¦¬ ëĤĺ +Ġìļ°ë¦¬ëĤĺ ëĿ¼ +ãģ¥ ãģıãĤĬ +ĠاÙĦÙħ سÙĦ +ĠاÙĦÙħسÙĦ ØŃØ© +Ġlibert é +клÑİÑĩ ение +Ġzam ów +Ġzamów ienia +รà¸ĸ à¹Ħà¸Ł +Ø£ ÙģÙĦ +Ø£ÙģÙĦ اÙħ +Ùħ راج +Ùħراج عة +Ġë¹Ħ êµIJ +ĠاÙĦت اب +ĠاÙĦتاب عة +Ġë§Į ëĤĺ +Ġб Ñĥм +ĠбÑĥм аг +Ġgé nero +Ġìŀĺ 못 +×ŀ פ×ķר×ĺ +è²·ãģĦ çī© +ĠÙĦدÙĬ Ùĥ +Ġ×ľ×¢ ×Ļת +Ġ×ľ×¢×Ļת ×Ļ×Ŀ +ĠsÅĤ ab +ĠпÑĢедÑģÑĤав лÑı +ãĤ¿ ãĤ¤ãĥĪ +ãĤ¿ãĤ¤ãĥĪ ãĥ« +Ùħ ص +Ùħص Ø·Ùģ +ÙħصطÙģ Ùī +Ġdifficult é +ãĥĨãĤ£ ãĥĸ +Ġpew noÅĽci +ĠpewnoÅĽci Äħ +Ġ무 ìĬ¨ +Ø¥ رس +إرس اÙĦ +Ġд алÑĮ +ĠдалÑĮ ÑĪе +Ġ׾ ×ł×¡ +Ġ×ľ×ł×¡ ×ķת +หมูà¹Ī à¸ļà¹īาà¸Ļ +×ŀס×ŀ ׼×Ļ +أسÙĦ ÙĪØ¨ +Ġzw ÅĤ +ĠzwÅĤ as +ĠzwÅĤas zc +ĠzwÅĤaszc za +ĠпÑĢ ÐµÐ¶ +ĠпÑĢеж де +ĠоÑĢганиз аÑĨиÑı +Ġdön emin +Ġdönemin de +Ġ Ủ +ĠỦ y +ä¸ĭ ãģĴ +ĠпоÑģлед ние +Ġgü ne +Ġgüne ÅŁ +Ġ×IJ ×ĸר +Ġ×IJ×ĸר ×Ĺ×Ļ +ãģ§ãģĤ ãĤįãģĨ +ĠÙĨ ÙĤ +ĠÙĨÙĤ اط +æŃ£ ãģĹãģĦ +ĠÑĢ ÐµÐ³ +ĠÑĢег иона +ĠFör der +ê²½ ìĺģ +dıkl ar +dıklar ını +trzym aÄĩ +أش Ùĥ +أشÙĥ اÙĦ +×Ķת ×IJ +×Ķת×IJ ×ŀ×Ķ +à¸Ĺำà¹ĥหà¹ī à¹Ģà¸ģิà¸Ķ +ĠGeb ä +ĠGebä ude +ĠСеÑĢ Ð³ +ĠСеÑĢг ей +Ġз доÑĢов +ĠздоÑĢов ÑĮÑı +Ġr ãi +ĠпÑĢед ÑĥÑģ +ĠпÑĢедÑĥÑģ моÑĤÑĢ +ĠпÑĢедÑĥÑģмоÑĤÑĢ ÐµÐ½ +Ġ×Ķצ ×Ļ×ij +Ġ×Ķצ×Ļ×ij ×ķר×Ļ +Ġdés ir +Ġн оÑĩ +ĠноÑĩ ÑĮ +möglich keiten +Ġ×IJ×Ĺר ×ķ׳×Ļ×Ŀ +Ġsoir ée +ĠNh áºŃn +Ù ª +à¸Ľà¸£à¸°à¸§à¸±à¸ķิ ศาสà¸ķรà¹Į +êµIJ íĨµ +ĠØ£ Ø®ÙĬ +Ġdé cid +Ġdécid é +Ġwy ja +Ġwyja ÅĽni +Ġ สิ +Ġสิ à¸ĩ +Ġสิà¸ĩ หา +Ġสิà¸ĩหา à¸Ħม +à¹ģ à¸Ńรà¹Į +หà¸Ļà¹īา à¸Īà¸Ń +ס תר +Ġê ¶ +Ġê¶ Į +Ġê¶Į 리 +pl ätze +ب Ø·ÙĦ +ê±´ ìĦ¤ +Ġ×IJ ×Ļ×ŀ×Ļ +Ġ×IJ×Ļ×ŀ×Ļ ×Ļ׾ +ãģ ½ +تر اث +×IJ׾ ×Ļ×ŀ×ķת +Ġdispon ÃŃveis +Ġz ale +Ġzale ży +à¸Ľà¸£à¸°à¸Ĭา สัมà¸ŀัà¸Ļà¸ĺà¹Į +ĠÅļw iat +Ġpor ówn +Ġporówn a +Ġ׾×ĺ ×ķ×ijת +×Ķ×ĸ ×ŀ׳×Ķ +Ġ×Ľ×ª ×ķצ×IJ×Ķ +Ġ×ij ק׾ +Ġ×ijק׾ ×ķת +ĠоÑĤ кÑĢ +ĠоÑĤкÑĢ Ñĭва +ãĥij ãĥ¯ãĥ¼ +ë¿IJ ë§Į +Ġв ÑģÑı +ĠвÑģÑı к +ãģ¨ãģª ãģ£ãģ¦ãģĦãĤĭ +Ġgi áºŃn +Ġок ÑĢÑĥ +ĠокÑĢÑĥ жа +ĠокÑĢÑĥжа ÑİÑī +ĠUnivers ität +ĠÑĢ Ð¾Ð¶ +ĠÑĢож д +ĠÑĢожд ениÑı +Ø® ÙĬÙĦ +Ġкомпани й +ĠÑĢазлиÑĩ нÑĭе +ĠЦ ена +׳×Ļ ×ķ×ĸ +׳×Ļ×ķ×ĸ ׾ +׳×Ļ×ķ×ĸ׾ ×ĺר +Ġê³µ ê°Ħ +Ġê°ľ ëħIJ +landır ma +ĠÑĥдал ен +à¸ŀัà¸ģ à¸ľ +à¸ŀัà¸ģà¸ľ à¹Īà¸Ńà¸Ļ +Ġprote cción +Ġb ÅĤ +ĠbÅĤ ÄĻd +Ã Ī +Ġíĸī ë³µ +ĠÅŁ ü +ĠÅŁÃ¼ phe +Ġí Ķ +ĠíĶ ¼ +Ġíͼ íķ´ +Ġëĭ¤ 르 +à¹Ħมà¹Ī à¹Ģà¸ģิà¸Ļ +ãģ¿ ãģª +ãģ¿ãģª ãģķãĤĵ +ĠпоÑĤ ÑĢеб +ĠпоÑĤÑĢеб иÑĤел +ĠاÙĦÙĥÙĦ اÙħ +ìķĦ ë²Ħ +ìķĦë²Ħ ì§Ģ +ãĤĴ使 ãģ£ãģŁ +Ġbụ i +ĠпоÑĤ еÑĢ +ĠпоÑĤеÑĢ Ñı +ĠØ¢ ÙĦاÙģ +ĠнаÑģÑĤоÑıÑī ее +ãģıãģªãĤĬ ãģ¾ãģĹãģŁ +clus ão +ãĤ³ ãĥĶãĥ¼ +צ פ×Ļ +צפ×Ļ ×Ļ×Ķ +Ø® ÙĦا +Ø®ÙĦا ص +ล à¹īำ +ãĥ¯ ãĤ¤ãĥ³ +Ġมี à¸Ļา +Ġมีà¸Ļา à¸Ħม +Ø´ خص +شخص ÙĬات +Ġ×ĸ ×§ +Ġ×ĸ×§ ×ķ×§ +×Ļ ×Ļצ +×Ļ×Ļצ ×Ĵ +èĢĥãģĪ æĸ¹ +Ġürün ü +ĠиÑģп ол +ĠиÑģпол ни +Ġcompañ ero +×§ צ×Ķ +×ŀ×¢ ׳×Ļ×§ +Ùħ ØŃÙħد +Ġc ámara +Ġп ед +Ġпед аг +Ġпедаг ог +м аÑĢ +маÑĢ Ðº +×Ķת ׳×Ĵ×ĵ +ĠìĨĮ ê°ľ +Ġcom unitÃł +ê³ ¤ +ĠNg Ãłi +สà¸ĩ à¸ļ +ĠmieszkaÅĦ ców +ĠÙĨ ÙĩائÙĬ +iv ité +Ġи де +Ġиде алÑĮн +ĠØ£ سبÙĪØ¹ +Ġ×Ļ ×¢×ľ +Ġ׾ ר×IJש +Ġ׾ר×IJש ×ķ׳×Ķ +ĠзапиÑģ и +ĠкоÑĢ Ð¿ÑĥÑģ +วà¸ĩ ศ +วà¸ĩศ à¹Į +ĠÐĶ Ð¼ +ĠÐĶм иÑĤ +ĠÐĶмиÑĤ ÑĢ +Ġkön nt +Ġböl ges +Ġbölges inde +׼ ×Ļ׼ +׼×Ļ׼ ר +ĠاÙĦØ¥ Ø«ÙĨ +ĠاÙĦإثÙĨ ÙĬÙĨ +Ġng á»Ļ +ì¹ ł +د راج +Ġu da +Ġuda ÅĤo +ìº IJ +بر ÙĨاÙħج +ĠÑģÑĥд еб +ĠÑģÑĥдеб н +Ġzun ächst +ĠEduc ación +ãģ¨ãģª ãģ£ãģ¦ãģĦãģ¾ãģĻ +Ġ×Ķ×IJ ×ŀ×Ļת×Ļ +Ġİ nt +Ġİnt ernet +ĠcaÅĤ ego +ãĥĹãĥª ãĥ³ +Ø¥ بد +إبد اع +ĠпоÑĢ ÑĤал +à¹Ĥà¸ķ à¹ī +Ġ×Ķ×§ ש×ķר +пл од +ĠÙħ د +ĠÙħد رÙĬد +×ŀסע ×ĵ×Ķ +ĠØ´ÙĬ ئ +ĠØ´ÙĬئ ا +à¸ģà¹Īà¸Ń สรà¹īาà¸ĩ +Ġì°¸ ê³ł +à¹Ģà¸Ĺ ร +à¹Ģà¸Ĺร à¸Ķ +Ġ×ij×ŀ קר×Ļ×Ŀ +Ġb ât +Ġbât iment +åij¼ ãģ³ +ç´ł æķµ +ç´łæķµ ãģª +przedsiÄĻbior st +przedsiÄĻbiorst w +Ġ×ł×ª ×ķ׳×Ļ×Ŀ +×Ĺ׾ ×ķ×Ŀ +ร วย +Ùħ ÙĪØ¶ÙĪØ¹ +ĠÑģоб ÑĢан +вед ÑĥÑī +ĠÑĤе аÑĤ +ĠÑĤеаÑĤ ÑĢ +m eye +meye ceÄŁi +Ġpien iÄħ +ĠpieniÄħ d +ĠpieniÄħd ze +ÑĢез иденÑĤ +ØŃ صر +ìĺ ¥ +à¹Ģย ืà¸Ńà¸Ļ +ĠÑĥ ни +ĠÑĥни веÑĢ +ĠÑĥнивеÑĢ Ñģ +ĠÑĥнивеÑĢÑģ иÑĤеÑĤ +ĠاÙĦر ØŃ +ĠاÙĦرØŃ ÙħÙĨ +ĠÑĤеÑħ нолог +ĠÑĤеÑħнолог ии +ìĹIJ ëĦĪ +ìĹIJëĦĪ ì§Ģ +Ġíķ Ń +ĠíķŃ ìĥģ +à¸ĺ า +à¸ĺา à¸ķุ +ĠEspañ ol +×ĵ×Ĵ ש +Ġêµ ī +Ġêµī ìŀ¥ +Ġêµīìŀ¥ íŀĪ +ĠÅĤ at +ĠÅĤat wo +Ġk á»ĭch +Ø¥ ز +إز اÙĦØ© +ĠдейÑģÑĤв ие +ĠsaÄŁ layan +สุà¸Ķ ยà¸Ńà¸Ķ +Ġzosta Äĩ +Ġdispon ÃŃvel +ïº į +ver ständ +verständ lich +tw or +twor zyÄĩ +ع جز +à¹Ģà¸Ĥ à¹īม +ยà¹Ī à¸Ńม +Ġstrat ég +Ġstratég ie +à¸ľà¸¥ à¹Ħมà¹ī +Ġê°ģ ì¢ħ +ĠÙħ ÙĪØ§ +ĠÙħÙĪØ§ ض +ĠÙħÙĪØ§Ø¶ ÙĬع +اØŃ تج +اØŃتج اج +Ġ Ấ +ĠẤ n +×ŀ ×ŀש׾×Ķ +ĠÅŁek il +×ŀ ×Ĺ׾ +×ŀ×Ĺ׾ ×ķת +Ġ à¸ĺ +Ġà¸ĺ ัà¸Ļ +Ġà¸ĺัà¸Ļ วา +Ġà¸ĺัà¸Ļวา à¸Ħม +Ġìĭ¤ ìłľ +Ġìĭ¤ìłľ ë¡ľ +ì¤ij ìķĻ +ëįĶ ëĿ¼ +ĠÑĪ Ð¸ÑĢ +ĠÑĪиÑĢ Ð¾ÐºÐ¾ +Ġsol ución +วาà¸ĩ à¹ģà¸ľà¸Ļ +×IJ×ķ×ĺ ×ķ×ŀ +×IJ×ķ×ĺ×ķ×ŀ ×ĺ×Ļ +ĠÑĢ ÐµÑģÑĤ +ĠÑĢеÑģÑĤ оÑĢ +ĠÑĢеÑģÑĤоÑĢ Ð°Ð½ +ëį ¸ +ÑĤ ÑĢад +ÑĤÑĢад и +ÑĤÑĢади ÑĨион +ÑĤÑĢадиÑĨион н +มะ à¹Ģรà¹ĩ +มะà¹Ģรà¹ĩ à¸ĩ +à¹Ĥ ส +Ġol masını +×ŀ×ķס ר +ĠоÑĤноÑĪ ÐµÐ½Ð¸Ð¸ +Ġê°ĢëĬ¥ ìĦ± +Ġy uk +Ġyuk arı +ìĨ Ķ +ĠÑģ ÑĦ +ĠÑģÑĦ еÑĢе +Ġ×§ ×ķפ +ãĤ± ãĥ¼ãĤ +ãĤ±ãĥ¼ãĤ Ń +âĢķ âĢķ +ĠاÙĦØ£ ÙĦÙħ +ĠاÙĦØ£ÙĦÙħ اÙĨÙĬ +Ả N +ת×ķ׼ ׳×Ļ×ķת +ĠÑģÑĥÑīеÑģÑĤв ÑĥеÑĤ +æĪij ãĢħ +ĠاÙĦص ادر +ĠTr á»įng +Ġа д +Ġад миниÑģÑĤ +ĠадминиÑģÑĤ ÑĢа +ĠадминиÑģÑĤÑĢа ÑĨи +ĠдÑĢÑĥг ими +Ñģп еÑĪ +عÙĦاÙħ ات +Ġа б +Ġаб Ñģол +ĠабÑģол ÑİÑĤ +ĠабÑģолÑİÑĤ но +ฤ à¸Ķู +é tr +étr anger +нÑı ÑĤи +нÑıÑĤи е +×¢ ×ķ׳ +×¢×ķ׳ ש +ĠÙĤ ائ +ĠÙĤائ ÙĦا +Ġм аÑģ +ĠмаÑģ ло +ãĥī ãĤ¤ +ãĥīãĤ¤ ãĥĦ +å¿ħè¦ģ ãģĮãģĤãĤĬãģ¾ãģĻ +×ŀ×ķ×ĸ ×Ļ×IJ +×ŀ×ķ×ĸ×Ļ×IJ ×ķף +ĠNgo ại +Ġkê nh +à¸ģาร à¸Ńà¸Ńà¸ģà¹ģà¸ļà¸ļ +×ŀ פק +×ŀפק ×ĵ +ÙħÙĨ از +ÙħÙĨاز ÙĦ +ë· ° +íĹ ¤ +ÙħÙĩ ارات +Ġpropri été +פ×Ĵ ×Ļש×Ķ +Ñĩ ÑĢ +ÑĩÑĢ ÐµÐ¶ +ÑĩÑĢеж ден +×Ķ ×ķצ×IJ×Ķ +ØŃÙĥ ÙĬÙħ +ĠíĻ Ī +ĠíĻĪ íİĺìĿ´ì§Ģ +åİ ³ +åݳ ãģĹãģĦ +×¢ ×ŀ×ĵ×Ķ +ĠAu ÃŁen +سÙĪ Ø¡ +ë¹ Ī +ĠÙĪ Ø® +ĠÙĪØ® اصة +ин ÑĤеÑĢ +инÑĤеÑĢ ÐµÑģ +èĩ´ ãģĹãģ¾ãģĻ +Ġhük üm +à¹Ħà¸Ĥ มัà¸Ļ +Ġdav ran +Ġdavran Ä±ÅŁ +à¹Ģà¸ķ ียà¸ĩ +в ÑĢем +вÑĢем енно +à¹Ģà¸Ĺศ à¸ģา +à¹Ģà¸Ĺศà¸ģา ล +å¼ķ ãģ£ +å¼ķãģ£ è¶ĬãģĹ +×IJר ×ķ×Ĺ +×IJר×ķ×Ĺ ×ª +à¹Ģ วิ +à¹Ģวิ รà¹Į +à¸Ńยà¹Īาà¸ĩ รวà¸Ķà¹Ģรà¹ĩว +ĠìŬ íĸī +ĠÑĢан ÑĮ +ĠÑĢанÑĮ ÑĪе +Ġzob ow +Ġzobow iÄħ +ĠzobowiÄħ z +Ġ×ķ׼ ×ŀ×ķ×ijף +ĠاÙĦÙħ Ùĩ +ĠاÙĦÙħÙĩ ÙĨÙĬ +ãĤ¢ ãĤ¸ +ãĤ¢ãĤ¸ ãĤ¢ +ë°© ìĨ¡ +à¸Ńà¸Ńà¸ģ à¸ģำลัà¸ĩ +à¸Ńà¸Ńà¸ģà¸ģำลัà¸ĩ à¸ģาย +am éli +améli orer +å½ĵãģŁãĤĬ åīį +Ġreg elm +Ġregelm Ã¤ÃŁig +ãģĬ åĭ +ãģĬåĭ § +ãģĬåĭ§ ãĤģ +Ġm ưá»Ŀi +بر Ùħج +ĠNat ürlich +ĠD Å©ng +ĠاÙĦر جاÙĦ +Ġthé p +Ġol muÅŁtur +×ŀ×ķס ×Ļ×§×Ķ +f älle +주 íĥĿ +ĠاÙĦÙģ Ø±Øµ +Ġnaj wiÄĻks +ĠnajwiÄĻks zy +Ġça ÄŁ +ĠçaÄŁ rı +ì¸ ł +ĠvÃŃ ct +ĠvÃŃct ima +ĠÑģовеÑĢ ÑĪен +×Ķ×Ļ ×Ļת×Ļ +à¹Ģà¸Ķ ี +à¹Ģà¸Ķี à¹ĭ +à¹Ģà¸Ķีà¹ĭ ยว +ü yü +Ġд оп +Ġдоп олн +Ġдополн иÑĤелÑĮно +à¹ģà¸ķà¸ģà¸ķà¹Īาà¸ĩ à¸ģัà¸Ļ +Ġá l +Ġál bum +à¸Ľà¸£à¸°à¸Īำ à¸Ľà¸µ +ĠÑĦ едеÑĢ +ĠÑĦедеÑĢ Ð°Ð»ÑĮн +Ġobs ÅĤ +ĠobsÅĤ ugi +à¹Ģร ืà¹Ī +à¹Ģรืà¹Ī à¸Ńย +à¹Ģรืà¹Īà¸Ńย à¹Ĩ +ëģ Į +Ġngh ìn +ĠBaÅŁkan lıģı +تأ سÙĬ +تأسÙĬ س +Ġ×ij×ij ×ķקר +Ġ×¢×ij×ķ×ĵ ×ķת +Ġبص ÙĪØ±Ø© +ãĤıãģij ãģ§ãģ¯ãģªãģĦ +führ er +ãĤ¹ ãĤŃ +ãĤ¹ãĤŃ ãĥ« +ĠاÙĦÙĤ ض +ĠاÙĦÙĤض ÙĬØ© +Ġдолж ноÑģÑĤ +ÙģØ§Ø± ÙĤ +Ġcomeç ou +Ġorganis é +Ġxu ân +ĠÑģообÑī аеÑĤ +ĠпÑĢи д +ĠпÑĢид еÑĤÑģÑı +TÃľ RK +ãĥ¬ ãĥ¼ãĤ·ãĥ§ãĥ³ +Kh ông +است Ùģ +استÙģ Ø§Ø¯Ø© +ä¸ĬãģĮ ãģ£ãģ¦ +Ġum ie +Ġumie jÄĻ +ĠumiejÄĻ tn +ĠumiejÄĻtn oÅĽci +ëĤ ¸ +à¹Ģà¸Ļ à¸Ńรà¹Į +×ĵ×ķ ×ķ×Ĺ +ÃŃs imo +I ÃĬ +IÃĬ N +Ġalcan ç +Ġ à¸ķุ +Ġà¸ķุ ลา +Ġà¸ķุลา à¸Ħม +ש׾ ×ĺ×ķף +Ġél è +Ġélè ves +ĠÄij u +ĠÄiju á»ķi +ĠØ£ Ùģ +ĠØ£Ùģ Ø±ÙĬ +ĠØ£Ù쨱ÙĬ ÙĤÙĬ +ĠØ£Ù쨱ÙĬÙĤÙĬ ا +ãĤĴæİ¢ ãģĻ +ĠпÑĢед ложениÑı +ج اد +ĠÑħоÑĤ ÑĮ +Ñģ ал +Ñģал он +à¸Ľà¸£à¸° à¹Ģม +à¸Ľà¸£à¸°à¹Ģม ิà¸Ļ +ãĤŃ ãĥĥãĥģ +ãĤŃãĥĥãĥģ ãĥ³ +×ij×ĵ×Ļ×§ ×ķת +Ġch ù +Ġchù a +ÐĴ иде +ÐĴиде о +иÑĢов ка +ĠÑħоÑĤ иÑĤе +Ġspéc ifique +รส à¸Ĭาà¸ķิ +è¾¼ ãĤĵãģł +伸 ãģ³ +×Ķצ׾ ×Ĺת +ãģ©ãģ® ãĤĪãģĨãģ« +سع ادة +Ġл ид +Ġлид еÑĢ +ม à¸ĩ +มà¸ĩ à¸Ħล +ØŃ اÙħÙĦ +หล ุà¸Ķ +à¸Ńยà¹Īาà¸ĩ à¸ķà¹Īà¸Ń +à¸Ńยà¹Īาà¸ĩà¸ķà¹Īà¸Ń à¹Ģà¸Ļืà¹Īà¸Ńà¸ĩ +ãģķãģĽãģ¦ éłĤ +تس ÙĪÙĬ +تسÙĪÙĬ ÙĤ +ĠaÅŁaģı d +ĠaÅŁaģıd aki +ĠÑĨ елÑĮ +ĠÑĨелÑĮ Ñİ +ĠAra ÅŁtırma +à¸Ĥัà¸ļ รà¸ĸ +Ùĩ ذÙĩ +ลà¸ĩ à¸Ĺะ +ลà¸ĩà¸Ĺะ à¹Ģà¸ļ +ลà¸ĩà¸Ĺะà¹Ģà¸ļ ียà¸Ļ +تÙĥ اÙħÙĦ +Ġc io +Ġcio è +ãģ¦ ãģĬãģı +ĠاÙĦصØŃ ÙģÙĬ +ĠíĬ¹ ìłķ +полн иÑĤÑĮ +ãĤĵ ãģĺãĤĥãģªãģĦ +ãĤĵãģĺãĤĥãģªãģĦ ãģĭ +ĠاÙĦج Ùĩ +ĠاÙĦجÙĩ ات +ĠÑĥÑģпеÑĪ Ð½Ð¾ +Ġв ок +Ġвок ÑĢÑĥг +ĠÑģиÑĤÑĥ аÑĨиÑı +Ġ×Ķ×IJ ×ŀר +Ġ×Ķ×IJ×ŀר ×Ļ×§ +Ġ×Ķ×IJ×ŀר×Ļ×§ ×IJ×Ļ +×ŀ ×Ĵ×ĸ +×ŀ×Ĵ×ĸ ×Ļף +Ġак ÑĤÑĥ +ĠакÑĤÑĥ алÑĮн +é ta +éta is +Ġmog ÅĤa +ĠÑĤоÑĩ ки +Ġ×ŀ×Ķ ×ŀ×¢ +Ġ×ŀ×Ķ×ŀ×¢ ×¨×Ľ×ª +มี à¸Ľà¸£à¸°à¸ªà¸´à¸Ĺà¸ĺà¸´à¸łà¸²à¸ŀ +×Ļר ×Ļ×ĵ×Ķ +×Ĵר ×ŀ׳ +×Ĵר×ŀ׳ ×Ļ×Ķ +Ġг лав +Ġглав ное +Ġ미 ëŀĺ +Ġ׳׼ ×ķ׳×Ķ +ĠÙĪ Ø·ÙĨÙĬ +op port +opport unitÃł +Ġh á»§y +ĠÙĦ تØŃ +ĠÙĦتØŃ ÙĤÙĬÙĤ +Ġó rg +Ġórg ão +ãĤ¹ ãĥĶ +ãĤ¹ãĥĶ ãĥ¼ãĥī +Ġön ü +Ġönü ne +Ùħع اÙħÙĦ +ש×ŀ ×Ļר×Ķ +ĠвеÑģÑĮ ма +ĠwiÄĻks zo +ĠwiÄĻkszo ÅĽÄĩ +Ġاست راتÙĬج +ĠاستراتÙĬج ÙĬØ© +ĠÙģ Ø¥ +ĠÙ쨥 ذا +à¹Ģà¸Ĭืà¹Īà¸Ń ม +à¹Ģà¸Ĭืà¹Īà¸Ńม à¸ķà¹Īà¸Ń +Ġ׾ פר +Ġ׾פר ×ĺ×Ļ×Ŀ +Ùħض ÙĬ +ĠGer çek +Ġçocuk ların +ÙĪØ« ائÙĤ +ĠÙħساء Ùĭ +Ġunterstüt zt +Ġpré st +Ġprést amo +ĠÐłÐ°Ð· меÑĢ +ĠÅŁ eker +Ġsé culo +×ij×Ķ ×Ļר +Ø´Ùĩ ÙĪØ± +Ġ à¸Ńีà¸ģ +Ġà¸Ńีà¸ģ à¸Ĺัà¹īà¸ĩ +Ġlleg ó +à¸¨à¸´à¸¥à¸Ľ ะ +æĪij ãģĮ +æĪijãģĮ å®¶ +ع ÙĤÙĪ +عÙĤÙĪ Ø¨Ø§Øª +ĠF älle +Ġs ÅĤuż +ĠsÅĤuż b +ĠاÙĦØŃÙĤ ÙĪÙĤ +Ġпл иÑĤ +Ġи ноÑģÑĤ +ĠиноÑģÑĤ ÑĢан +ĠиноÑģÑĤÑĢан н +à¹ĥà¸Ļ à¸Ĥà¸ĵะà¸Ĺีà¹Ī +ãĤ« ãĥĨ +ãĤ«ãĥĨ ãĤ´ +ãĤ«ãĥĨãĤ´ ãĥª +à¸Ńิ ส +à¸Ńิส ระ +à¹Ģà¸ľà¸¢ à¹ģ +à¹Ģà¸ľà¸¢à¹ģ à¸ŀร +à¹Ģà¸ľà¸¢à¹ģà¸ŀร à¹Ī +ãģĬ ãģĦ +ãģĬãģĦ ãģĹãģĦ +است ÙĤÙĦ +استÙĤÙĦ اÙĦ +تØŃ ض +تØŃض ÙĬر +åĬ© ãģij +Ùħر اÙģÙĤ +Ġ×ĵ ×ķר +Ġ×ĵ×ķר ש +×ŀת×Ļ ×Ļ×Ĺס +ס ×Ļ׼ +ס×Ļ׼ ×ķ×Ŀ +íĮĮ íĬ¸ +Ġwy ÅĽ +ĠwyÅĽ w +ĠwyÅĽw iet +ĠwyÅĽwiet l +ĠاÙĦاÙĨ ساÙĨ +ĠStra ÃŁen +ï¼ ¬ +ãģ« åŁº +ãģ«åŁº ãģ¥ +Ġcap ÃŃtulo +ลุ ย +Ġ×Ķ×ŀ×§ צ×ķ×¢×Ļ +ãģĤãĤĭ ç¨ĭ度 +á» ¢ +ĠاÙĦ ÙĦا +ĠاÙĦÙĦا زÙħØ© +æķĻ ãģĪ +Ġרש ×IJ×Ļ +з ав +зав иÑģ +завиÑģ им +à¸Ľà¸±à¸Ī à¸Īัย +à¹Ģà¸ĭ ล +à¹Ģà¸ĭล ลà¹Į +Ġdiffé rence +ĠAlt ın +Ġк ÑĢай +ĠкÑĢай не +Ġз ло +Ġgün ümüz +Ġн аÑĤÑĥÑĢ +ĠнаÑĤÑĥÑĢ Ð°Ð»ÑĮн +×Ĵ×ķ׾ ש×Ļ×Ŀ +Ġк аÑĤегоÑĢ +ĠкаÑĤегоÑĢ Ð¸Ð¸ +Ġз нак +à¸ģà¹Īà¸Ńà¸Ļ หà¸Ļà¹īา +à¸ģà¹Īà¸Ńà¸Ļหà¸Ļà¹īา à¸Ļีà¹ī +ĠÙħÙĨ ت +ĠÙħÙĨت خب +ãĥĽ ãĥ¼ãĥ« +Ġе вÑĢо +ส ว +สว ม +ĠìľĦ ìĽIJ +ĠìľĦìĽIJ ëĭĺ +ĠاÙĦØŃ ÙĪØ« +ĠاÙĦØŃÙĪØ« ÙĬ +ĠÑģодеÑĢж иÑĤ +ãĥķãĤ¡ ãĥĥãĤ·ãĥ§ãĥ³ +Ġ à¸ģัà¸Ļ +Ġà¸ģัà¸Ļ ย +Ġà¸ģัà¸Ļย ายà¸Ļ +ãĤª ãĥª +ãĤªãĥª ãĤ¸ +ãĤªãĥªãĤ¸ ãĥĬãĥ« +Ġб ÑĢенд +ãĤĴæĮģ ãģ£ãģ¦ãģĦãĤĭ +Ġinvers ión +Ġê° ĸ +Ġê°ĸ ê³ł +Ġnov itÃł +ê´Ģ ê´ij +Ġà¸ŀ ฤษ +Ġà¸ŀฤษ à¸łà¸² +Ġà¸ŀà¸¤à¸©à¸łà¸² à¸Ħม +×ķר ×Ĺ×Ļ×Ŀ +׼׾ ×ķ׾ +Ġng ạc +×Ļ ×Ļש +×Ļ×Ļש ×ķ×ij +f äll +fäll ig +ĠÑĤÑĢеб ÑĥеÑĤÑģÑı +Ġcar á +Ġcará cter +Ġprinc ÃŃpio +ĠÅĤ az +ĠÅĤaz ien +ĠÅĤazien k +Ġgi ãn +ÑģÑĤÑĢа ива +Ùħس اب +Ùħساب ÙĤØ© +à¹Ģà¸Ħรืà¹Īà¸Ńà¸ĩ à¸Ķืà¹Īม +ترÙĥ ÙĬب +vol ução +ĠÐŁ оÑĩ +ĠÐŁÐ¾Ñĩ ем +ĠÐŁÐ¾Ñĩем Ñĥ +казал оÑģÑĮ +ĠпÑĢимен ениÑı +à¹Ģà¸Ĺ ียม +íĮ Ķ +à¸Ĥà¹īà¸Ń à¹Ģสà¸Ļà¸Ń +à¸Ľà¸±à¸į à¸įา +Ġоб ÑĥÑĩ +ĠобÑĥÑĩ ениÑı +ĠÑģеÑĢ Ð¸ +ĠÑģеÑĢи ал +Ġingl és +ĠÙĦ Ùĥرة +Ġ×ĺ ׾ +Ġ×ĺ׾ פ×ķף +Ġìł ij +Ġìłij ê·¼ +×IJ ×ķ×Ĵ +×IJ×ķ×Ĵ ×ķס +×IJ×ķ×Ĵ×ķס ×ĺ +ĠболÑĮÑĪ Ð¾Ðµ +ĠÐļон еÑĩно +×¢×Ļת ×ķ׳ +×¢×Ļת×ķ׳ ×IJ×Ļ +Ġкноп к +Ġз н +Ġзн аÑĤÑĮ +ĠÄij á»± +ĠÄijá»± ng +вл аж +влаж н +×ŀ ×Ļ×ĺ×ij +ãĤ¬ ãĤ¤ +ãĤ¬ãĤ¤ ãĥī +........ .. +Ġà¸ģ ุม +Ġà¸ģุม à¸łà¸²à¸ŀ +Ġà¸ģà¸¸à¸¡à¸łà¸²à¸ŀ ัà¸Ļ +Ġà¸ģà¸¸à¸¡à¸łà¸²à¸ŀัà¸Ļ à¸ĺ +Ġà¸ģà¸¸à¸¡à¸łà¸²à¸ŀัà¸Ļà¸ĺ à¹Į +be z +bez pieczeÅĦst +bezpieczeÅĦst w +ãĥijãĥij æ´» +ع اط +عاط Ùģ +ĠÄij áºŃm +Ġз ÑĢ +ĠзÑĢ ÐµÐ½Ð¸Ñı +Ġbor ç +Ġнед ел +Ġнедел Ñİ +Ġh á»ı +Ġhá»ı ng +ìŀ¥ ìķł +ìŀ¥ìķł ìĿ¸ +ĠاÙĦع ÙĦاÙĤØ© +Ġíģ ¬ +Ġíģ¬ ê²Į +à¹Ħร à¹Ī +à¸ļา à¸Ķ +à¸ļาà¸Ķ à¹Ģà¸Īà¹ĩà¸ļ +à¸Ŀ รั +à¸Ŀรั à¹Īà¸ĩ +à¸Ŀรัà¹Īà¸ĩ à¹Ģศ +à¸Ŀรัà¹Īà¸ĩà¹Ģศ ส +ר ×¢×Ļ +רע×Ļ ×ķ׳×ķת +Ġë Į +ĠëĮ ĵ +ĠëĮĵ ê¸Ģ +Ġnaj b +Ġnajb li +Ġnajbli ż +Ġnajbliż sz +ĠиÑģполÑĮз ÑĥеÑĤÑģÑı +Ġcient ÃŃf +ĠcientÃŃf ico +×¢ ×ŀ×§ +Ġg ợi +Ø´ ØŃÙĨ +ĠÅĽ m +ĠÅĽm ier +ĠÅĽmier ci +à¸Ħาสิà¹Ĥà¸Ļ à¸Ńà¸Ńà¸Ļà¹Ħลà¸Ļà¹Į +×Ĺש×ij ת×Ļ +Ġn ingu +Ġningu ém +è¾¼ ãĤģ +ãģ · +ĠÑĥ г +ĠÑĥг ол +ï½ ° +פת ×Ļ×Ĺ +פת×Ļ×Ĺ ×ª +Ġ×Ķר×IJש ×ķ׳×Ļ×Ŀ +p ósito +ãĤŃ ãĥ¬ãĤ¤ +ãģ© ãģĵãĤį +à¹Ģà¸Ĺà¹Īา à¹Ħ +à¹Ģà¸Ĺà¹Īาà¹Ħ หร +à¹Ģà¸Ĺà¹Īาà¹Ħหร à¹Ī +ĠинÑĤеÑĢ ÑĮеÑĢ +ĠØŃ اج +ĠØŃاج Ø© +สี à¸Ĥาว +ìĸ ¼ +Ġn á»Ļ +Ġná»Ļ p +ĠÃŃ nd +ĠÃŃnd ice +สำ รวà¸Ī +Ġкажд ой +Ġhot éis +Ġnast ÄĻ +ĠnastÄĻ pn +Ġ×Ķ×§ ×ķ×ĵ +Ġ×Ķ×§×ķ×ĵ ×Ŀ +פ ×ķפ +פ×ķפ ×ķ׾ +פ×ķפ×ķ׾ ר×Ļ +вÑĪ ÐµÐ¹ +ãĤ·ãĥ³ ãĥĹ +ãĤ·ãĥ³ãĥĹ ãĥ« +ĠzdjÄĻ Äĩ +ĠгÑĢÑĥпп а +Ġпом еÑī +ĠпомеÑī ениÑı +ãģ©ãģĨ ãģĦãģĨ +ĠиÑģп ÑĭÑĤа +Ġog ÅĤ +ĠogÅĤ os +ĠogÅĤos zen +ĠogÅĤoszen i +สรà¹īาà¸ĩ สรร +สรà¹īาà¸ĩสรร à¸Ħà¹Į +à¸ŀร รà¸ĵ +Ġçık Ä±ÅŁ +ĠÑĩаÑģÑĤ ноÑģÑĤи +Ġ×ķ ×Ļ×ķתר +ç¶ļãģį ãĤĴ +ç¶ļãģįãĤĴ èªŃ +ç¶ļãģįãĤĴèªŃ ãĤĢ +à¸ģร ั +à¸ģรั ม +г ÑĢаÑĦ +Ġв лад +Ġвлад елÑĮ +ĠвладелÑĮ ÑĨ +Ġistedi ÄŁ +ĠistediÄŁ iniz +×ij׾ ×¢ +×ij×ľ×¢ ×ĵ×Ļ +ÙħÙĪ Ø§Ùģ +ÙħÙĪØ§Ùģ ÙĤØ© +Ġ×Ļ ×ķר +Ġ×Ļ×ķר ×§ +ãĤ«ãĥ¼ãĥī ãĥŃãĥ¼ãĥ³ +ĠاÙĦÙħØ´ ÙĥÙĦ +ĠاÙĦÙħØ´ÙĥÙĦ Ø© +ĠêµŃ íļĮ +ס פ×ĺ +ספ×ĺ ×ŀ +ספ×ĺ×ŀ ×ijר +Ġìĸ´ ëłµ +Ùĥ اÙħ +ÙĥاÙħ ÙĬرا +sch lü +schlü sse +ĠØ« ÙĨ +ĠØ«ÙĨ ائÙĬ +ìī ½ +ĠÐŀ Ñģоб +ĠÐŀÑģоб енно +Ġин веÑģÑĤи +ĠинвеÑģÑĤи ÑĨи +اØŃ تÙħ +اØŃتÙħ اÙĦ +E Äŀ +EÄŀ İ +íķĺ ê²łëĭ¤ +Ġ×IJ ×ijר×Ķ +Ġ×IJ×ijר×Ķ ×Ŀ +Ġ×ij×Ĺ ×Ļ׳×Ŀ +Ø£ ÙĪØ¶ +Ø£ÙĪØ¶ اع +Ġdé l +Ġdél ai +Ġ×IJ×ķ×Ķ ×ij×Ļ×Ŀ +ĠÑģо Ñħ +ĠÑģоÑħ ÑĢ +ĠÑģоÑħÑĢ Ð°Ð½Ð¸ +ĠдоÑģÑĤ иж +ĠдоÑģÑĤиж ени +สิà¹Īà¸ĩ à¹ģ +สิà¹Īà¸ĩà¹ģ วà¸Ķ +สิà¹Īà¸ĩà¹ģวà¸Ķ ล +สิà¹Īà¸ĩà¹ģวà¸Ķล à¹īà¸Ńม +ĠاÙĦÙħ باشر +ĠÑĦ иг +ĠÑĦиг ÑĥÑĢ +мож ем +׾×ŀ×Ļ×ĵ ×Ķ +Ġcin é +Ġciné ma +Ġb ada +Ġbada ÅĦ +جب ÙĩØ© +Ġд еп +Ġдеп ÑĥÑĤ +ĠдепÑĥÑĤ аÑĤ +Ġdist ância +ĠاÙĦÙħ عار +ĠاÙĦÙħعار ضة +thè se +ü nc +ünc ü +Ġдан ного +ĠBel gi +ĠBelgi ë +Ġ×ij ×ij×§ +Ġ×ij×ij×§ ש×Ķ +ย à¹Īาà¸Ļ +Ġsol ução +Ġ×Ķצ ×ĺר +Ġ×Ķצ×ĺר פ×ķ +ĠØ£ÙĨ ØŃ +ĠØ£ÙĨØŃ اء +Ġد ÙħØ´ +ĠدÙħØ´ ÙĤ +มั à¹ī +มัà¹ī ย +Ùħ غرب +است عÙħاÙĦ +ĠS ÅĤow +ĠëıĻ ìĭľ +ĠëıĻìĭľ ìĹIJ +ĠÑģ оÑģ +ĠÑģоÑģ ед +ì²Ń ìĨĮ +ì²ŃìĨĮ ëħĦ +Ġг ÑĢаÑĦ +ĠгÑĢаÑĦ ик +Ġìŀij ìĿĢ +Ġyet i +Ġyeti ÅŁtir +ĠìĿ´ê²ĥ ìĿ´ +ห à¹Īาà¸ĩ +Ø¥ ÙħÙĥاÙĨ +Ø¥ÙħÙĥاÙĨ ÙĬØ© +است عراض +ÙħØ® در +ĠÑĩ ÑĥÑĤÑĮ +Ùħ دÙĬر +ÙħدÙĬر ÙĬØ© +Ġà¹Ģม ษ +Ġà¹Ģมษ ายà¸Ļ +Ġм еÑħ +ĠмеÑħ аниз +ĠмеÑħаниз м +ĠÑģ Ñĥм +ĠÑģÑĥм мÑĥ +Ġv ö +Ġvö ll +Ġvöll ig +Ġд ÑĢÑĥз +ĠдÑĢÑĥз ÑĮÑı +ãĤĴåĪ©ç͍ ãģĹãģ¦ +à¸ļรร à¸Īุ +po życz +×ŀש ׼ +×ŀש׼ ×ł×ª +×ŀ×©×Ľ×ł×ª ×IJ +Ġeuropé en +Ġpropri é +Ġproprié taire +Ġkh ấu +ãģĦãģŁãģł ãģijãĤĭ +Ġtec rü +Ġtecrü be +×Ķ ×ij +×Ķ×ij ׳×Ķ +Ġcu Ì +ĠcuÌ ī +ĠcuÌī a +×IJ ×ķ×ķ +×IJ×ķ×ķ ×Ļר×Ķ +Ġ׼×ķ׾ ×ķ +U lus +Ulus lararası +Ġ׳ ×ķת +Ġ׳×ķת ף +ãģ« åIJij +ãģ«åIJij ãģijãģ¦ +ë¹ Ľ +à¸Ĺ ัà¸ģษ +à¸Ĺัà¸ģษ ะ +س ÙĤÙĪ +سÙĤÙĪ Ø· +Ġв н +Ġвн еÑĪ +ĠвнеÑĪ Ð½Ðµ +Ġur z +Ġurz ÄĻd +Ġá mb +Ġámb ito +à¸Ń à¸ĺิ +à¸Ńà¸ĺิ à¸ļาย +Ġ ÅĤad +ĠÅĤad n +ê±´ ì¶ķ +wód zt +wództ w +Ġquest ões +Ġש ×§ +Ġשק ×Ļ×ij׾ +Ġmiejsc owoÅĽci +Ġв ал +Ġвал ÑİÑĤ +hä user +หà¸Ļ à¸Ńà¸ĩ +ãģ¨ åħ± +ãģ¨åħ± ãģ« +ãĥı ãĥ¼ãĥī +Ġê°ľ ìµľ +ĠоÑģнов ном +Ġм ÑıÑģ +اع ت +اعت ÙĤاÙĦ +สà¸ĸ ิ +สà¸ĸิ à¸ķิ +N gu +Ngu á»ĵn +ĠÙħ جÙĦ +ĠÙħجÙĦ Ø© +à¹ģà¸Ĥ à¸Ļ +ĠاÙĦÙĦÙĬ بÙĬ +פע×Ļ׾ ×ķ×Ļ×ķת +Ġ×Ķר פ×ķ×IJ×Ļ +פר ×ķפ +פר×ķפ ×Ļ׾ +×§ ׾×IJ +ק׾×IJ ס×Ļ +Ùĥت Ø´Ùģ +ãģ«ãģª ãģ£ãģ¦ãģĹãģ¾ãģĨ +à¹Ģà¸Ħล à¹ĩà¸Ķ +à¹Ģà¸Ħลà¹ĩà¸Ķ ลัà¸ļ +Ġì» ´ +Ġì»´ íĵ¨ +Ġì»´íĵ¨ íĦ° +Ġ×Ĺ×Ļ ×ķ×ij×Ļ +Ġnä m +Ġnäm lich +åij¼ ãģ° +åij¼ãģ° ãĤĮ +ĠÑĢ Ð¾Ð» +ĠÑĢол и +Ġspécial isé +à¸Ļ วัà¸ķ +à¸Ļวัà¸ķ à¸ģรรม +ÙĨص ÙĪØµ +пеÑĢ ÐµÐ´ +пеÑĢед аÑĩ +thè que +Ġר×IJ ×Ļת×Ļ +ãĥĢ ãĤ¦ãĥ³ +ãĤı ãģĭ +ãĤıãģĭ ãģ£ãģ¦ +беÑĢ ÐµÐ¶ +ĠÑģ ек +ĠÑģек ÑĢ +ĠÑģекÑĢ ÐµÑĤ +ĠпоÑģÑĤоÑıн н +à¸Ĥà¸Ļ สà¹Īà¸ĩ +Ġm ük +Ġmük em +Ġmükem mel +еÑĤ еÑģÑĮ +ĠاÙĦسÙĨ ÙĪØ§Øª +ĠìłĦ íĺĢ +Ġ×Ķ×ŀ×§ ×ķר×Ļ +Ġmü d +Ġmüd ah +Ġmüdah ale +Ġwy b +Ġwyb ór +Ġtend ência +Ø¥ دار +إدار ÙĬØ© +Ġunterstüt zen +ת ×ijר +ת×ijר ר +Ġdi á +Ġdiá logo +ĠÃĸ nce +ĠÃĸnce ki +ãĤ¹ãĥĿ ãĥĥãĥĪ +ëĦ £ +ĠG eli +ĠGeli ÅŁ +ãĤĴ éĢļ +ãĤĴéĢļ ãģĹãģ¦ +ĠFuÃŁ ball +Ġsal ari +Ġsalari é +ĠпÑĢодÑĥк ÑĤов +صÙģ ÙĤØ© +รว à¸ļ +รวà¸ļ รวม +à¹ĥà¸Ļ à¸IJาà¸Ļ +à¹ĥà¸Ļà¸IJาà¸Ļ ะ +Ġkay na +Ġkayna ģı +Ġìŀij íĴĪ +ĠвÑĭ ÑĢаж +ĠвÑĭÑĢаж ен +ĠÑģÑĤ еп +ĠÑģÑĤеп ени +ĠاÙĦÙħ ÙĪØ¬ÙĪØ¯ +ĠاÙĦÙħÙĪØ¬ÙĪØ¯ Ø© +ล à¹īม +Ġnaj czÄĻ +ĠnajczÄĻ ÅĽcie +ĠnajczÄĻÅĽcie j +Ġz wy +Ġzwy k +Ġzwyk ÅĤ +Ġê·¸ëłĩ ì§Ģ +à¸ģระ à¸Ī +à¸ģระà¸Ī าย +Ġëĭ µ +Ġëĭµ ë³Ģ +ĠÑĢе ак +ĠÑĢеак ÑĨи +ĠÅĽwie ż +ĠÑģÑĤоим оÑģÑĤи +ÙħÙĨ اÙĤ +ÙħÙĨاÙĤ Ø´ +ÙħÙĨاÙĤØ´ Ø© +ĠÑħоÑĩ Ñĥ +ãĥľ ãĥ¼ãĥī +Ġróż nic +Ġк ÑĢÑĭ +ĠкÑĢÑĭ ÑĪ +âľ ĵ +ãĤ³ãĥ³ ãĥĨãĥ³ +ãĤ³ãĥ³ãĥĨãĥ³ ãĥĦ +ĠпÑĢед поÑĩ +×ŀר ×ij×Ļת +ĠØ´ Ùĥ +ĠØ´Ùĥ را +Ġд ал +Ġдал ек +Ġдалек о +بر ÙĬØ· +برÙĬØ· اÙĨÙĬا +ع ÙĨا +عÙĨا ÙĬØ© +ĠÑĢаÑģÑģ каз +ĠÑĢаÑģÑģказ Ñĭва +Ø£ ÙĦÙĪ +Ø£ÙĦÙĪ Ø§ÙĨ +æĮģ ãģ£ãģ¦ +æĮģãģ£ãģ¦ ãģĦ +Ùħباد ئ +×Ķ ×¢×ijר +×Ķ×¢×ijר ת +Ġyay ı +Ġyayı ml +Ġyayıml a +m át +mát icos +à¸ģ ัà¸ĩ +à¸ģัà¸ĩ วล +Ġ׾ פת +Ġ×ľ×¤×ª ×ķ×Ĺ +à¸ŀฤ à¸ķิ +à¸ŀฤà¸ķิ à¸ģรรม +í Ĥ¬ +Ġок ÑĢÑĥг +Ġ×ŀצ ×ķ×ķ×Ķ +ÐĽ ени +ÐĽÐµÐ½Ð¸ н +ĠTri á»ģu +ãĤ³ãĥŁ ãĥ¥ +ãĤ³ãĥŁãĥ¥ ãĥĭ +ãĤ³ãĥŁãĥ¥ãĥĭ ãĤ± +ãĤ³ãĥŁãĥ¥ãĥĭãĤ± ãĥ¼ãĤ·ãĥ§ãĥ³ +Ùĥ ÙĨÙĬ +ÙĥÙĨÙĬ سة +ãĤĴ ä¸Ńå¿ĥ +ãĤĴä¸Ńå¿ĥ ãģ« +ĠmiÄĻd z +ĠmiÄĻdz yn +ĠmiÄĻdzyn ar +ĠmiÄĻdzynar od +ĠmiÄĻdzynarod ow +ÙĦ ÙĨ +ÙĦÙĨ دا +بر Ø´ +برش ÙĦÙĪÙĨ +برشÙĦÙĪÙĨ Ø© +à¸ģระ à¸ķุ +à¸ģระà¸ķุ à¹īà¸Ļ +Ġg ı +Ġgı da +à¸Ľà¸£à¸° à¸Ĺัà¸ļ +à¸Ľà¸£à¸°à¸Ĺัà¸ļ à¹ĥà¸Ī +Ġë¶Ī 구 +Ġë¶Ī구 íķĺê³ł +ĠÙĨ Ø· +ĠÙĨØ· اÙĤ +ĠÐľ ожеÑĤ +Pr äs +Präs ident +ĠÑģк оÑĢ +ĠÑģкоÑĢ Ð¾ÑģÑĤÑĮ +Ġ×Ķ×ij ×ķקר +еÑħ аÑĤÑĮ +Ġg ạo +Ġש×IJ ×Ļ׳×Ŀ +Ġ×ij׳ ×ķ×Ĵ +Ġ×ij׳×ķ×Ĵ ×¢ +Ġо пиÑģание +Ġucz ni +Ġuczni ów +à¹Ģà¸Ń à¹ĩà¸Ļ +Ġت Ø´ +Ġتش رÙĬÙĨ +Ġnh ãn +ë¹ ¨ +Ġcaract ère +×¢ ׾×Ļ +×¢×ľ×Ļ ×Ļ×Ķ +楽ãģĹ ãĤģãĤĭ +ĠÑģ аÑħ +ĠÑģаÑħ аÑĢ +дÑĥм аÑĤÑĮ +ĠÐĴоз можно +ص ÙĬاÙĨ +صÙĬاÙĨ Ø© +öm ür +ส ล +สล à¹ĩ +สลà¹ĩ à¸Ń +สลà¹ĩà¸Ń à¸ķ +ë¡ ¯ +Ġth ói +gr Ã¶ÃŁe +Ġksi ÄĻ +ĠksiÄĻ g +ĠÑĢ Ð¾Ð¼ +ĠÑĢом ан +ÙĤ اسÙħ +×ŀ×ij ×ķ×Ĵ +×ŀ×ij×ķ×Ĵ ר×Ļ×Ŀ +bes ch +besch äft +beschäft ig +×Ķצע ×Ķ +ĠÃģ rea +ĠзаÑıв к +Ä ¹ +ĠлÑİб ого +Ġ ม +Ġม à¸ģร +Ġมà¸ģร าà¸Ħม +ÑĦ из +ÑĦиз иÑĩеÑģк +ин ÑĦ +инÑĦ ек +инÑĦек ÑĨи +اÙĦ Ø· +اÙĦØ· ائÙģ +Ġкол л +Ġколл екÑĤив +ез жа +Ġس بØŃ +ĠسبØŃ اÙĨ +ĠسبØŃاÙĨ Ùĩ +sch lä +schlä ge +Ġд и +Ġди аг +Ġдиаг ноÑģÑĤ +ĠоÑĤмеÑĤ иÑĤÑĮ +Т Ь +ĠاÙĦ در +ĠاÙĦدر اسÙĬ +עצ ×ŀ +עצ×ŀ ×IJ×ķת +Ġdém arch +Ġdémarch e +Ġ×ĺ ×ķ×¢ +Ġ×ĺ×ķ×¢ ף +Ġfuncion ários +á» µ +׾ ׼×IJ +׾׼×IJ ×ķר×Ķ +à¸ĭ à¹Ī +à¸ĭà¹Ī à¸Ńม +ĠÑĩ Ñĥв +ĠÑĩÑĥв ÑģÑĤво +âĸ ¼ +п ÑĥÑī +пÑĥÑī ен +Ġм еÑĢ +ĠмеÑĢ Ð¾Ð¿ +ĠмеÑĢоп ÑĢи +ĠмеÑĢопÑĢи ÑıÑĤиÑı +Ġu çu +Ġuçu ÅŁ +ãĤĴåĪ©ç͍ ãģĻãĤĭ +a ÄŁ +aÄŁ lı +ìĺĪ ìĪł +à¹ģ ยà¹Ī +ĠاÙĦÙĥ Ùħ +ĠاÙĦÙĥÙħ بÙĬ +ĠاÙĦÙĥÙħبÙĬ ÙĪØªØ± +ت ÙĪÙĬ +تÙĪÙĬ تر +à¹Ģà¸Ĭ ีà¹Īยว +à¹Ģà¸Ĭีà¹Īยว à¸Ĭา +à¹Ģà¸Ĭีà¹Īยวà¸Ĭา à¸į +á» Ķ +Ġhi ếm +ذا Ùĥرة +Ġ×Ķ×ŀ×Ļ ×ķ×Ĺ×ĵ +ĠìĪ ľ +ĠìĪľ ê°Ħ +ĠK ı +ĠKı sa +Ġgele ceÄŁi +пÑĢо ÑĦеÑģÑģиона +пÑĢоÑĦеÑģÑģиона л +Ġog ó +Ġogó le +ĠgÅĤ ów +ĠgÅĤów ne +ĠÑģÑĤ илÑĮ +×IJ פ׾ +×IJפ׾ ×Ļ×§ +×IJפ׾×Ļ×§ צ×Ļ×Ķ +สม ารà¹Į +สมารà¹Į à¸Ĺ +สมารà¹Įà¸Ĺ à¹Ĥà¸Ł +สมารà¹Įà¸Ĺà¹Ĥà¸Ł à¸Ļ +Ġth ánh +ÐŁ од +ÐŁÐ¾Ð´ ÑĢоб +ÐŁÐ¾Ð´ÑĢоб нее +ĠاÙĦت ÙĪÙĨ +ĠاÙĦتÙĪÙĨ سÙĬ +Ġbah çe +à¹ģà¸ģà¹ī à¸Ľà¸±à¸įหา +é ducation +eu rop +europ ä +europä ische +ĠK si +ĠKsi ÄĻ +ĠëĦ ĺ +ĠëĦĺ ìĸ´ +Ġv üc +Ġvüc ud +Ġyay g +Ġyayg ın +Ġnie kt +Ġniekt óry +Ġniektóry ch +ãģŃ ãģĩ +Ġк аж +Ġкаж еÑĤÑģÑı +к аж +каж еÑĤ +ĠاÙĦ دÙĬÙħÙĤرا +ĠاÙĦدÙĬÙħÙĤرا Ø· +ĠاÙĦدÙĬÙħÙĤراط ÙĬØ© +æŃ © +æŃ© ãģĦãģ¦ +Ġv az +Ġvaz ge +Ġvazge ç +Ġмин ималÑĮ +ĠминималÑĮ н +ãĥij ãĤ¿ +ãĥijãĤ¿ ãĥ¼ãĥ³ +Ġë Ĭ +ĠëĬ IJ +ĠëĬIJ ëĤĮ +ãģ¡ ãĤĩãģĨ +ãģ¡ãĤĩãģĨ ãģ© +Ġ à¸ģร +Ġà¸ģร à¸ģà¸İ +Ġà¸ģรà¸ģà¸İ าà¸Ħม +تج دÙĬد +ĠØ´ اÙħÙĦ +หลัà¸ģ à¸IJาà¸Ļ +ĠмаÑĢ ÑĪ +ĠмаÑĢÑĪ ÑĢÑĥÑĤ +Ġv ÃŃt +ĠvÃŃt ima +Ġquiz á +ay gı +×ĵ×ijר ×Ļ×ķ +Ġиз д +Ġизд ели +Ġиздели Ñı +п ла +пла Ñĩ +плаÑĩ ива +ä»» ãģĽ +Ġéquip é +ä¹ħ ãģĹãģ +ä¹ħãģĹãģ ¶ +ä¹ħãģĹãģ¶ ãĤĬ +Ġк аÑĤ +ĠкаÑĤ ал +ĠкаÑĤал ог +ส à¹īม +ĠÑĢ ÐµÐ¹ +ĠÑĢей ÑĤ +ĠÑĢейÑĤ инг +Ġth uyá»ģn +ĠاÙĦÙħ ÙĤدس +esp ère +ãģ«åħ¥ ãģ£ãģŁ +หมาย à¹Ģลà¸Ĥ +ת×Ĺ×ķש ת +à¸Ļ à¹Īะ +Ġpe ÅĤ +ĠpeÅĤ ne +Ġpé rd +Ġpérd ida +หม วà¸Ķ +หมวà¸Ķ หมูà¹Ī +иÑĩеÑģк ÑĥÑİ +çµĤ ãĤı +çµĤãĤı ãģ£ãģŁ +Ġ×Ĵ ×ķ×Ĵ׾ +à¸Ĺำ à¸Ħวาม +à¸Ĺำà¸Ħวาม สะà¸Ńาà¸Ķ +Hot éis +Ġз аÑĢ +ĠзаÑĢ ÐµÐ³Ð¸ÑģÑĤ +ĠзаÑĢегиÑģÑĤ ÑĢи +ĠзаÑĢегиÑģÑĤÑĢи ÑĢова +ĠÑģ обÑĭÑĤи +ĠÑģобÑĭÑĤи Ñı +Ġ×ĸ ׼×IJ +ÙħÙĨظ ÙĪÙħØ© +Ġ×Ķ×ŀ צ +Ġ×Ķ×ŀצ ×Ļ×IJ×ķת +Ùħ ÙĥÙĪÙĨ +ÙħÙĥÙĪÙĨ ات +ä¸ĬãģĮ ãĤĭ +Ġm ÄĻ +ĠmÄĻ sk +หรืà¸Ń à¹Ģà¸Ľà¸¥à¹Īา +ëĤ ® +Ġnok tas +Ġnoktas ı +ĠболÑĮÑĪ Ð¸Ð¼ +ĠлÑĥÑĩ ÑĪиÑħ +Ø´Ùĩ ÙĬد +à¸Ńำ à¸Ļ +à¸Ńำà¸Ļ วย +à¸Ńำà¸Ļวย à¸Ħวาม +à¸Ńำà¸Ļวยà¸Ħวาม สะà¸Ķวà¸ģ +Ġе в +Ġев ÑĢ +ĠевÑĢ Ð¾Ð¿ +ĠевÑĢоп ей +à¸ī าย +ìĦ Ń +Ùħ Ù쨧 +ÙħÙ쨧 ÙĪØ¶ +ÙħÙ쨧ÙĪØ¶ ات +ë¹ Į +赤 ãģ¡ãĤĥãĤĵ +ĠÑĥдал оÑģÑĮ +ĠÐ¥ оÑĤ +ĠХоÑĤ Ñı +przedsiÄĻbior c +ĠH ôm +íķĺìĺĢ ìĬµëĭĪëĭ¤ +Ġн аг +Ġнаг ÑĢÑĥз +ĠнагÑĢÑĥз к +Ġ×ij×Ļ׳ ׾×IJ×ķ×ŀ×Ļ +Ġê°ĢëĬ¥ íķľ +ĠH ữu +à¸Ń ุà¸Ķ +à¸Ńุà¸Ķ ม +ת ×ķפ +ת×ķפ ×¢×Ķ +Ġmi ÅĤo +ĠmiÅĤo ÅĽci +ksi Äħż +ksiÄħż ka +ĠاÙĦÙĦ عبة +à¸ī าà¸ģ +สะ สม +×ŀ תר +×ŀתר ×Ĺש +Ġlég ère +Ġ׾צ פ +Ġ׾צפ ×Ļ×Ķ +ĠиÑģÑĤоÑĢ Ð¸Ñı +Ġ ãĥĪãĥ© +ĠãĥĪãĥ© ãĥĥãĤ¯ +ĠãĥĪãĥ©ãĥĥãĤ¯ ãĥIJãĥĥãĤ¯ +Ġк а +Ġка ÑĦе +×ŀס×ŀ ×ļ +Ġc üm +Ġcüm le +à¹Ģà¸Ħลืà¹Īà¸Ńà¸Ļ à¹Ħหว +ãģĬ ãģĿ +ãģĬãģĿ ãĤīãģı +ìŀIJ ëıĻ +ìŀIJëıĻ ì°¨ +à¸Ńั à¸ķ +à¸Ńัà¸ķ à¹Ĥà¸Ļ +à¸Ńัà¸ķà¹Ĥà¸Ļ มั +à¸Ńัà¸ķà¹Ĥà¸Ļมั à¸ķิ +ĠÅŁ ik +ĠÅŁik ay +ĠÅŁikay et +extr ême +kr ä +krä fte +ëĤ Ļ +íķ ij +ì² Ļ +íĺ Ī +ì° į +âĻ ¡ +ìŀ Ķ +ë¢ ° +íĿ Ķ +íĿ IJ +âĩ Ĵ +ë§ Ľ +ìĬ Ī +á» Ĵ +ìĺ µ +âĹ İ +í Ĥ¨ +ê¿ Ī +ìĪ ¨ +ìĽ ¨ +ë§ ¥ +ï½ Ģ +ï¼ ª +Ạ¨ +ãħ İ +Ñ Ĺ +ìĦ ¬ +ì¹ ¼ +ï¼ ¶ +ìĽ ł +ëŁ ´ +Å ĥ +ëĤ ¼ +ëĭ IJ +âĢ ¹ +ë¦ Ń +ì§ IJ +âĢ ¤ +à ħ +ëľ ¨ +íĦ ¸ +íľ ĺ +ê² ģ +ë´ ħ +à ĺ +ëŃ Ķ +ëĺ ij +âĹ ĩ +ìĹ ĺ +ï» ´ +ë§ ¹ +ï¾ Ŀ +ìĬ · +íĥ ķ +ï¼ ł +ì» ´ +ëł Į +ì½ ľ +ï» ¹ +ãħ ł +ì¡ ¸ +ëħ ¹ +âĤ º +âĸ ¶ +íĥ IJ +êµ ´ +íij ¸ +Ñ Ķ +íĶ ½ +Ð ħ +ë° ¤ +Ô ģ +ì² ¨ +ì¶ ĺ +ë² Ĺ +ë© ¸ +ï¼ » +ï¼ ½ +ï¼ · +ì° Į +à Ĵ +íı ´ +ìĵ ¸ +ì´ Į +ëģ Ķ +ëĶ © +ëĩ Į +ë© Ģ +ë² ¨ +ï¼ µ +ë§ ¡ +ëĭ « +ภ¿ +ãģ ± +ìĩ ¼ +ìº ł +ë® ¤ +ê± ± +ì» ¬ +âĦ ĥ +ëĶ ± +ëĥ Ī +ìĭ ± +íĻ Ī +ëŀ IJ +ìħ Ģ +ìł ł +Ð Ĩ +ëł ī +ï½ ħ +ï½ ı +íĻ Ģ +ëĽ ° +á» ® +í Ĥ¹ +ê½ ĥ +ï» ¤ +ïº Ķ +êº ¼ +ìķ ī +âĻ ¦ +ï½ ģ +ìĵ ´ +ãĢ ī +ì° ® +ì¤ ĺ +á» ª +ëģ Ħ +ëIJ ¨ +ìķ Į +íĿ ĺ +íħ IJ +ãĢ Ī +ê² ª +ëĭ ¥ +ê² ¼ +á» Į +ë§ ¨ +ëģ Ĭ +ë² ¤ +ëij Ķ +íĿ ¡ +á» ¬ +ë¬ ĺ +ãģ ī +ëŀ « +íĶ Ī +í ħį +ìŀ ĥ +ï½ ī +ìģ ľ +âĸ ½ +ë¬ » +âĸ ³ +ï¼ ¸ +ìģ ĺ +ì¶ ° +ìĬ ´ +ìķ ± +ìĩ Ħ +Ạ® +ï´ ¿ +ï´ ¾ +âĤ ½ +ëĦ ĵ +ë£ © +ì³ ¤ +ê´ ľ +Ã Ļ +á» ľ +ï¿ £ +ëĵ Ń +ë© ĺ +ê» ´ +ëł ´ +Ð ĥ +ë¬ µ +ì§ Ŀ +ãģ º +ðŁĺ Ĥ +ëŀ ¬ +ìł Ĭ +ê´ Ħ +ìŀ Ĭ +íŀ Į +ìĦ ¯ +âĪ Ģ +âĸ ¡ +ëĢ Į +ëŀ Ļ +ï½ ĥ +Ạ¶ +ï¾ Ħ +ïº ĺ +ë¹ ¼ +à Į +âĸ · +ê¸ į +ë© ĭ +ãģ ĥ +ìĺ Ĩ +ìĺ ® +ëª ¬ +ë¡ ¤ +ëł ¬ +ëĬ ¦ +âĸ ª +ì¼ ĵ +ìľ Ī +ì§ § +ï½ ½ +ëĥ ī +ï¾ Į +ëĺ IJ +ï¼ ĥ +á» Ħ +ì´ ¬ +ì¶ ¤ +ï¼ ¹ +ï» Ń +âĤ « +ï½ ĩ +ìĺ · +ëĸ ¨ +âī « +ë¦ ¿ +âľ ¨ +Ù ± +ì¯ ¤ +ê¹ Ķ +ðŁĺ Ĭ +ìĪ « +ê³ ± +êµ ³ +ï½ ĭ +ภĮ +Ä ł +ëĶ ¸ +ë° ij +ìħ ĭ +íİ ´ +âľ ħ +íĥ ij +ëĪ ĩ +íı ¼ +ðŁĺ į +ìĺ Ľ +ï» £ +Ñ ĺ +ì© Į +ë¦ ħ +ìĿ į +ï½ ¸ +ëį ľ +ãģ ħ +íİ ¼ +ëĭ Ŀ +ë¿ Į +ì¼ ° +ìĭ « +ë° ¥ +íĽ Į +ì¨ Į +ë¹ Ļ +ï½ İ +ë´ Ħ +ìĦ ¹ +ï½ ² +ìĮ ĵ +Ò ij +ë° į +ëł Ģ +íĨ ¤ +ï½ ¯ +ë¤ Ħ +ê½ ¤ +ï½ Ĵ +ìķ ¨ +ï½ ¼ +ê¹ IJ +íģ IJ +âĦ ĸ +ë§ º +ïº ® +ëħ ģ +ê² ¸ +ï» ł +íĬ ľ +Å ¹ +ë¥ Ń +ëĪ ī +ï½ Ķ +íĮ ¬ +ìŀ ĩ +ï ¬ģ +ï» ¨ +ëij ¥ +ëŀ Ħ +Ù ¬ +íĭ ´ +ìŀ ī +Ú ¾ +ìĽ ħ +ï» ® +ëĭ ī +âī ª +âĹ Ħ +ëĪ Į +íĽ ¼ +ì¤ į +Å ¸ +ì¤ ¬ +ì¾ Į +ï½ ĵ +ï¾ Ĭ +ðŁı » +ï¾ ī +Ð ģ +íĺ IJ +ï¾ Ļ +ê¼ ¬ +íŀ IJ +âĢ ¥ +ëŁ Ń +ë§ ŀ +ìĥ ¤ +ïº Ĵ +íĭ ± +ë½ ij +à ķ +âĪ ļ +ëĤ Ħ +ê¹ Ŀ +ëĨ Ī +Ạº +ìħ Ī +ìĮ į +âĢ ¡ +ï¼ ± +ìģ ¨ +âĺ º +ëĴ · +ìĺ ³ +ðŁij į +ëª ½ +ëĤ Ń +ïº Ń +ë© Ī +á» Ī +íķ Ģ +ëĭ Ļ +ë¦ ĩ +ìķ ¤ +ìį ¼ +ãĥ µ +Ñ £ +ìľ Ĺ +â ŃIJ +ï¾ ĺ +íĹ ¬ +ê¾ ¼ +ìķ Ĺ +ï» Į +ê± · +ëħ ķ +ë¡ ± +ìķ Ĭ +ï¾ Ģ +ìĩ ł +íĮ © +ïº ª +ë§ Ļ +ï¼ ¿ +ê¿ Ķ +íİ ľ +ë£ ¸ +íĶ Ķ +ï» ³ +ëı ķ +ìĭ ¼ +á» İ +ë§ ĺ +ì¢ ĭ +íĨ ¡ +ï½ ± +íĿ ij +á» ¸ +ì¦ Į +ì¹ ¸ +ëŃ ĺ +ï¾ Ĺ +ï» ĭ +íĬ Ģ +ë¥ Ļ +ì½ © +ëģ Ĺ +ëį ´ +ìħ ľ + ¸ +ë» IJ +ìĥ µ +ê² IJ +ëĵ ¬ +ë£ ° +ãħ ĭ +ìĹ ī +á» ĸ +ëĦ Į +ï½ ¶ +ë´ ĩ +ëĤ ³ +ãĤ ľ +ëĸ » +íİ Ģ +ëį © +íķ ¸ +à · +ê¼ ¼ +ëĶ ľ +ë° ´ +ë© į +âĹ ¯ +ìĹ ij +ìĻ ¼ +ïº ij +ë¶ ķ +ë¡ ¬ +ï½ Į +íĨ ¨ +ïº ´ +ëł ĺ +ê° ¤ +ìĪ ² +Ñ ĵ +ìħ ī +ï» ĵ +ëĪ Ķ +ëį § +âĢ ¼ +ï» ² +ê° ± +ê¿ Ģ +ëĭ · +Ạ¸ +Ạª +Æ Ĵ +ëį ¤ +ìĪ Ń +ï½ Ĥ +ï½ Ī +Å ł +ë£ ¬ +Ñ µ +ëĸ ¡ +ëĥ Ħ +ìĦ ° +ëĵ Ī +ï¾ ĥ +ëĩ ¨ +ï½ IJ +êµ ½ +ìĹ ½ +ëĤ Ģ +ë¬ ¶ +ï½ · +ìı Ł +íĺ Ķ +ê¼ Ī +ëģ Ī +ì¥ IJ +ïº Ĺ +Ä Į +ëĪ ł +ëĸ ¼ +íĢ ´ +âī ¥ +ëĭ Ń +ì± Ļ +ê» ı +ë© ¤ +ìĥ ĺ +ëį ® +ë£ ¡ +ìĤ ½ +ãĪ ľ +Ä ¨ +âĢ § +ï½ º +Ä £ +ì¦ ī +ï¼ ¼ +Û © +âĪ Ļ +ë° ı +ë¹ ħ +ðŁĺ Ľ +íĪ ´ +ðŁĴ ķ +ãĢ Ĵ +ìŀ ĺ +ïº ¤ +ï½ ĸ +ë© ľ +ë² ¼ +ëĿ Ħ +ëļ ľ +ï» ĺ +ìĥ Į +ï½ Ħ +ì© Ķ +ï½ Ļ +ïº © +Û ŀ +âĺ İ +ìł ¤ +ëIJ © +Å Ŀ +âŀ ¡ +ï» § +Ð ı +ì« ĵ +ê³ ½ +É ij +ãĥ ² +ëĤ « +ë¦ ī +ì¢ ģ +ë° Ń +ðŁĺ ģ +ë¹ µ +ì² © +ì» µ +ðŁĺ ĺ +ë± ħ +âī Ī +ë¹ ļ +ï» ľ +ðŁĻ ı +íģ ° +ìĦ ŀ +ï¾ ļ +ìĺ ¹ +ë¼ Ī +ëĤ ¯ +ëŀ © +íļ ¡ +ï½ ķ +íĥ ĵ +ëĿ ł +ê³ ģ +ëĵ Ģ +ìĹ ł +ï¼ º +ë§ ij +ëĭ ¿ +ì¿ ¨ +ãİ ¡ +Ð Ĭ +íĦ ± +Å ¨ +ïº ³ +ï¾ ı +âĭ ħ +ê¼ ´ +âī ¤ +íĮ ģ +Î © +ê¶ ¤ +ìĪ į +âľ ¿ +ì½ ¤ +ëĪ ħ +íĨ ± +ãħ ľ +áIJ ħ +Å Ĵ +ðŁij ī +ï» ¦ +Ð ª +ë¥ ľ +íķ « +ï¾ ĭ +âĻ « +ê¹ ľ +ë° ¸ +ëĶ ĺ +íĿ ī +ï¾ ģ +ï¾ Ľ +ëł Ľ +ê² ¹ +ì¿ ¼ +ï» ¬ +âŀ ¤ +ðŁĻ ģ +ïº ł +ëĨ ¨ +ë¯ ¹ +ê¸ ĭ +ë» Ķ +ê¹ ĥ +ëij ij +íĭ ¸ +íİ Ļ +âŀ ĸ +ãĥ ½ +ì§ ļ +ï½ ¬ +ï» ¥ +íĮ ½ +âĢ Ĵ +ì ĮĢ +ìŃ ī +ëļ ± +ãĤ ŀ +íĭ Ī +ãĤ IJ +ëī ĺ +Î £ +ê³ ° +ë¹ Ĺ +ï¾ İ +ðŁĺ Ń +íĿ ł +ìĹ ¿ +ê° ļ +ì¤ Į +ë§ µ +ï½ ³ +ãģ ¢ +ï» Ĺ +âī ¦ +Ú ¤ +ë łģ +ê¼ ½ +ï» « +âī § +ì´ Ľ +ìł Ŀ +Ạ° +âĻ £ +ìº ĺ +âĪ ĩ +ê² ī +ë° Ł +ï» Ķ +íĸ ĩ +âĸ Ĵ +ðŁij ı +à ŀ +ðŁĺ Ĩ +ïº ¼ +âĿ Ĺ +ìº Ķ +ì¹ © +ëĸ ¤ +ëĥ ħ +âĶ ľ +ï½ » +Î Ķ +áĥ ¦ +ìŀ İ +âĺ Ģ +âĪ ¼ +ðŁĶ ¥ +ë° Į +ìł ĸ +íĹ Ľ +Î ķ +ïº ĥ +ë¶ ī +âĪ ŀ +íĥ Ń +à ĭ +âģ Ħ +ãħ ĩ +ëĦ ¥ +ëĭ ® +ëł · +íĮ Ŀ +ìº ¡ +ë· Ķ +ì© į +íĤ ´ +ëļ « +âĵ Ĵ +íķ į +âĻ Ĥ +ï¾ Ĩ +âĨ © +ìį © +ïº ķ +íĿ Ļ +Ñ ľ +íĤ · +íĿ ° +íĥ ± +ëķ IJ +ï¾ Ĵ +× ĥ +ëĮ Ħ +ìĺ ´ +ìķ µ +ê¹ ¥ +ëŀ Ń +ìª ¼ +ãİ Ŀ +ðŁĺ ħ +ëı ĭ +ëª « +ïº ¸ +ë® ¬ +ë² ħ +ëij ł +ìħ ° +ì» · +ëĶ ª +ëħ Ķ +ãħ ¡ +ìĶ » +íķ ı +ëį ± +ïº ¨ +ï¾ į +ï½ µ +ì¢ Ģ +íİ Į +ï» ° +ïº £ +Æ £ +ðŁ¤ £ +ï· º +ëĤ ļ +âĭ Ĩ +ë³ į +ðŁĺ Ħ +ìĸ Ģ +ìĻ ł +ëĨ Ķ +íĹ ¨ +ï» Ľ +ï» Ŀ +á» ¶ +ìĸ ĺ +ìİ Ħ +Ú Ĩ +ï» ŀ +ëĢ IJ +ê² Ķ +ï» µ +âĹ ¦ +íļ Ł +ê¹ ģ +ê° ĵ +ëĶ ´ +ìı ĺ +ëļ Ŀ +á» ł +ëŀ ´ +ëĦ ī +âĺ ŀ +ï½ ĺ +Å ½ +ë¦ İ +âĸ ¬ +ëŃ ī +âĩ Ľ +ìį ¬ +ïº Ł +Ë ľ +ë¶ ĵ +ìĽ ° +Å ľ +ëŃ ĩ +á» ² +Ë ļ +ëķ Ģ +âĺ ij +ðŁı ¼ +ìĸ ½ +âĮ Ĵ +Ð İ +É ¾ +íĮ ¡ +ï¾ ħ +ìŀ Ń +ï½ ¨ +ì¹ « +ìľ Į +Ò Ľ +êµ ¿ +ëĭ ¦ +âĶ Ķ +ï¾ ij +ì§ ĸ +ìº Ħ +ãĢ ĥ +Ê ¼ +ê² Ł +ï½ § +Ä ¢ +íİ ł +ë§ · +ê° ĩ +ìĭ ¹ +ðŁĴ ¦ +ï¾ ľ +ëĬ Ļ +ë² ¡ +Å ¿ +ðŁĺ ĭ +ðŁĴ ª +ì¿ Ħ +ë© ķ +ìŃ ¤ +ëĬ Ħ +ðŁĮ ¸ +ãĤ Ŀ +Ç İ +ï½ ļ +Ä Ĺ +ëģ ĵ +ê¶ IJ +áµ ī +ãĥ Ĥ +ê» į +ðŁĺ ¦ +ãĢ Ŀ +ðŁ¤ Ĺ +Ñ Ł +ìĹ İ +âľ Į +ìī IJ +à Ĩ +íĹ IJ +ðŁİ ī +Î ij +ï½ Ń +ðŁĴ Ļ +ìĽ ¬ +íĢ ĺ +ï» ¢ +ðŁĺ İ +íij ¼ +íĿ © +ï» Ħ +íħ Ģ +ëł IJ +ì¥ ¬ +Ð ĭ +ìĥ · +ëľ ¬ +ðŁĺ ĥ +ëĦ ¬ +ë¥ ¨ +ìĽ į +ï½ Ĩ +ï½ ´ +ãĥ ħ +à ı +ï» ª +âĻ ł +ëĬ ¬ +ë± Ģ +ë° ĭ +ìĥ Ģ +ï½ ¾ +ëĤ ± +ì» ¸ +ðŁĴ ĸ +ðŁij Į +Ñ ŀ +ì§ ± +Ë Ĩ +ðŁĵ ļ +âŃ ķ +ï¬ Ĥ +ï» ¡ +ëij ¬ +íĪ ¼ +âĸ ¸ +ê° ¯ +ê¹ ħ +ï½ ® +ëĺ ¥ +Ä ¡ +íĮ Ł +Ð Į +ìĨ Ł +ïº ĵ +ï» ¼ +à Ľ +ãĥ ¾ +ëĮ ĵ +íĴ ĭ +ìķ ĵ +ï½ ¹ +ëĤ ¡ +ðŁij ĩ +Ạ¼ +ãĢ Ł +ðŁĮ Ł +íĥ ł +ãĢ Ĩ +âĢ Ł +ë¸ IJ +ðŁĮ ¹ +ìł ¼ +ðŁĵ Į +ìĶ ¬ +âĹ Ģ +ðŁĴ ĵ +ê¹ İ +ìĤ IJ +ìĶ Į +Ñ Ľ +âĶ Ī +ë² ³ +ãİ ŀ +Õ ¡ +íĤ µ +ðŁ¤ Ķ +ëĢ Ķ +ìĬ IJ +íĻ ī +âľ ¦ +ëľ ¯ +ìł ¯ +ëĶ § +Î ¦ +Ë Ī +ìī ¼ +âĹ Ĭ +ëľ © +ëľ ° +ï¾ IJ +ë¿ Ķ +ìĹ ® +ì· Į +ïº § +Î Ĵ +ëµ Ļ +ï» Ĭ +ì° Ķ +íİ Ħ +ðŁĴ Ĺ +Ạ´ +ì° ¢ +íľ ¼ +ê½ Ĥ +ì± Ķ +ìī ´ +âĸ ¾ +íĪ ° +ëĭ Ľ +âĿ £ +ï½ ª +ðŁĴ ľ +Ë ĺ +ãħ ¤ +âĨ Ĺ +íĸ Ħ +âĻ ¬ +ìķ ° +ïº ľ +âī ¡ +ãĢ ĵ +ìij ¥ +íĮ į +íī ģ +ë» Ĺ +íľ ł +íľ © +âľ Ī +íĢ Ħ +ìĸ ĩ +ì¢ ĩ +íŀ Ļ +ëª ¹ +ãĤ Ľ +ðŁĺ ± +ëį Ł +๠ħ +êµ ¶ +Ù « +ìĶ ģ +âľ ª +ï¾ Ī +ðŁĻ Į +âļ ¡ +Î ļ +ì¼ Ī +ï¾ Ķ +ï¾ Ĥ +êµ ī +ïº » +ðŁĴ ĭ +á¹ £ +Ó Ļ +ìĨ ľ +ìĹ £ +âľ © +ìľ Ļ +ïº ° +Ạ² +ìŀ £ +âĿ Į +âĺ ģ +ìķ İ +Ä ½ +Û ģ +ãĦ ± +ëŁ ¿ +íĮ ¸ +ê½ ī +ìı ł +ðŁį Ģ +âĨ Ķ +ëŃ ¡ +ï» ģ +ï¼ Ħ +ðŁĴ ¥ +âĺ Ľ +íĹ · +ëij ¡ +Î ł +Î ¤ +âĦ ĵ +ïº · +Î Ļ +ëı Ķ +ì§ ¤ +âĶ ĥ +ãĦ · +Ç Ĵ +ðŁ¥ ° +ëĶ ķ +ìļ ¥ +ì¸ Ħ +íĽ Ķ +ïº ĩ +ïº ¬ +ðŁĺ ¢ +ë¹ ¡ +ìĶ ¹ +Å ³ +Ë Ŀ +íİ ij +ï¾ ĵ +ðŁĴ ļ +ëĬ ij +êº ¾ +íĨ ° +à ¿ +Ð Ħ +ëĮ IJ +ë½ Ģ +ì· Ħ +ðŁ ĵį +ðŁĻ Ī +âĹ Ī +ê¿ ĩ +ì¼ Ħ +íİ « +ðŁĩ · +âĶ ĭ +âļ ł +ë± ī +ì į° +ìĻ Ī +É ª +ïº ĭ +ðŁĺ ľ +Î Ł +ðŁ ĻĤ +âļ ½ +Å Ī +ë¹ Ķ +íĮ ľ +๠ı +ìĸ ¹ +íĪ Ń +ðŁ¥ ĩ +ãĦ ´ +ëĶ ¥ +ìŃ Ī +âĪ Ĩ +ëĸ ³ +ë± ĥ +ìŀ ¦ +ï» IJ +Î ľ +âľ § +Ï į +ìł ĵ +âĹ ķ +ëĴ Ģ +ï» Ģ +ðŁĶ ´ +ê½ ģ +ëĮ Ī +ëİ Į +ãĤ İ +⦠ģ +ì½ § +ï¯ ¾ +âĿ ¯ +ภħ +ðŁĻ Ħ +âĿ Ģ +ðŁĶ ¹ +âĩ IJ +êµ µ +âĩ Ķ +ë¶ IJ +ðŁĴ Ľ +Î ¾ +íĥ ¬ +âĿ Ħ +Ò £ +ãĢ ° +âĪ ij +âĺ ¼ +âī ł +Ò ¯ +ïº ¯ +ê¿ ¨ +âľ ĸ +Ê ĸ +íĢ Ģ +ê¾ Ģ +íĹ Ŀ +âĶ £ +ãİ ľ +ëĶ Ľ +ëľ ¸ +ï º« +ê¿ ° +ðŁĩ ¹ +Ç IJ +Û Ĵ +ë£ » +ïº ĸ +Ñ ļ +ëĬ ł +Û ķ +ê¹ ¡ +ë¿ ľ +ì² ¼ +ï¨ ij +ë¥ µ +ìį ¸ +íħ ħ +íij ¹ +Ö Ģ +ï³ Į +ãħ £ +ìij ¤ +ì½ ķ +ëķ ł +ðŁĮ ¿ +íĥ Ķ +ìĽ ģ +Î ¶ +âŀ ľ +ìĬ ĺ +íĽ Ĺ +ë© § +ìī ĺ +Õ ¶ +á¹ ĩ +ðŁİ ģ +ï½ ¿ +ï¼ Ĥ +á¼ IJ +âľ ķ +âŀ ¢ +ëĦ ¨ +ì» « +ì¯ Ķ +ì° ľ +ðŁĴ ° +íħ Ŀ +ãİ ı +ë³ ¶ +Ò ĵ +âĨ ³ +ìĥ ´ +íģ ĺ +âĸ Ģ +ë² Ļ +ภĥ +á½ ¶ +Ä ķ +⬠ĩ +ë¤ ĺ +ðŁİ µ +âľ ļ +ïº ı +Î ¡ +âĹ ī +ðŁĴ « +Ð Ī +ìĸ Ħ +ì§ Ļ +ï» ĥ +ðĿij Ĵ +ëŃ Ħ +âĿ ¥ +âĿ ĸ +âĺ Ŀ +Ê ¹ +Ḡ¥ +âĢ ¿ +ãħ ħ +ê¸ ģ +ëķ ¡ +ëį ¥ +âĪ © +ê» Ħ +ë® Į +Ò ± +âĪ Ĺ +ëł Ļ +ïº Į +Ë IJ +ðŁĺ ³ +ðŁij © +ðŁİ ¶ +ì¿ µ +ðŁ¤ © +ê· ¤ +ëĮ Ķ +ïº IJ +Ï İ +ì¶ ¥ +ï½ Ĭ +á¹ Ń +ë¤ ¼ +âĸ « +ì§ ł +á¼ Ģ +ê» ij +ëĮ ģ +íĢ ¸ +âĻ Ľ +ðŁĴ ŀ +âĸ ° +ðĿij ĸ +ëĿ ¤ +ठ¦ +ì´ ĺ +ðŁĺ ĩ +ëĶ ¤ +Î Ĺ +ðŁĻ ĩ +Ë Ľ +ì© ¡ +âĪ § +Õ ¥ +Ñ Ļ +ëIJ ¬ +ëĸ Ħ +ðŁĮ · +ìĹ Į +ðŁĺ ¥ +ëĪ ´ +ï» ļ +É Ľ +ïº Ħ +ï» ı +Å Į +ë² ļ +ìĭ £ +ïº Ģ +Î ĵ +ðŁĺ Į +Ë Ļ +ëŀ ı +ðŁĶ ¸ +ðŁĵ · +ëģ ½ +íģ ½ +ðŁĴ ¡ +ðŁĮ ± +ëº ı +ìģ ł +ìĥ IJ +ëı Ĺ +ì¸ ° +ëĪ ķ +Î Ŀ +âģ ī +ðŁĮ ¼ +íĮ ł +âĭ ¯ +áĥ ĺ +âľ ¤ +ê± Ķ +íĮ İ +ðŁĴ ¯ +ìı Ļ +íĹ ī +Ù Ń +ì½ ° +ïº ¿ +ï» ± +ì± Į +âĺ ķ +ðŁİ Ģ +Ä Ŀ +ë° § +ìĤ ¿ +áij ķ +ðŁį ĥ +âĩ ¨ +Î Ľ +ë§ ´ +ë³ ķ +á ijIJ +âĸ ĵ +ðĿ ijľ +âĻ » +íĤ ¥ +Õ ¸ +ãĪ ± +ëº Ģ +ì² ¸ +ïº Ľ +ðŁı Ĩ +ðŁĩ ª +âĿ ĵ +Ä Ģ +ì½ ¥ +ðŁĩ § +á½ · +âľ Ĥ +ìŀ ¼ +ï§ ¡ +ðŁĵ ¸ +âĻ ¯ +É Ķ +á½ ¸ +âĮ ª +ï» ĸ +ï¥ § +âļ « +âĶ Ĺ +ðŁĮ Ī +ï» © +ðŁĵ ² +Ï Ī +ðŁĺ ¡ +ðĿij İ +ìľ ½ +ì§ ¬ +ì§ Ĭ +á½ ³ +ìĮ ¤ +ëĤ į +âī Ĵ +ðŁij ¨ +âĺ ĺ +Ó © +âĤ ĵ +âĪ Ĥ +ï¹ ģ +ðŁĴ IJ +íħ ĥ +ðŁı ½ +ê· Ħ +ðŁĺ ı +ðŁĮ º +ðŁĺ Ķ +ï½ « +âľ İ +ëµ Ī +ðŁĩ ¸ +âĢ £ +âŀ Ķ +ëĺ ĺ +ìĥ ¬ +Ê ĥ +⬠ħ +ì© IJ +ðŁĻ Ĩ +ðŁİ Ħ +Ä ¾ +⣠¶ +áĥ IJ +âĺ » +ì± ķ +ìģ © +ë½ ķ +ìº £ +ðŁij Ī +ðŁĻ ĭ +ï¾ ĸ +Ò ļ +Õ « +ìĮ Ī +ë² § +ðŁĩ ® +ï½ Ŀ +ðŁį ģ +ìĹ ¥ +Ä ³ +ë½ IJ +íį ½ +íĽ ij +âĤ ¹ +ãħ ģ +ìĶ ½ +ðŁĶ ģ +ठ¯ +ê¾ ¹ +ëī ľ +âĹ ¡ +íķ Į +Î ĺ +ë£ ¹ +ìĻ ĵ +ðŁĩ ¦ +ðŁij Ģ +âĶ Į +á¿ ¦ +ëĦ Ľ +ìĦ £ +ìŃ Ļ +ï± ł +Î ŀ +Ê » +á¿ ¶ +âĿ Ŀ +ê± Ģ +ëĸ ´ +ãĦ ¹ +ðŁĴ İ +Ï ¹ +⼠ħ +ï» ķ +ãĥ ± +ï½ Ľ +ëĮ ķ +ë¹ ½ +ì¥ Ķ +ì¿ ¤ +ðŁĸ ¤ +Ñ Ĵ +ê¹ į +ëİ Ģ +ìĭ ¯ +ë» ¤ +ðŁĵ ŀ +ðŁĵ £ +ðŁĺ Ŀ +ìį ¹ +ìĹ ¡ +ì° IJ +á½ IJ +ï» Ī +âľ į +Ä ı +ðŁĮ ŀ +âĦ ¦ +ê½ Ŀ +ë» ĺ +ìĪ ± +âĶ ĺ +ðŁĮ » +âĤ ´ +âŀ ¨ +íIJ ģ +ê ¶Ī +âĺ ¢ +ðŁĺ Ī +ï½ © +âĦ Ĺ +ê° Ń +ê° ¸ +ë» ij +ì¥ ´ +ì» ¥ +ï¤ Ĭ +ï» Ĵ +ðŁĺ ķ +âĺ Ķ +ìĺ IJ +ðŁļ Ĺ +ëĹ Ħ +ë§ ı +Õ ½ +âĸ » +⣠µ +ìī ° +ï» ij +âĻ © +Î ¥ +ðŁĺ £ +âĬ Ĥ +ãħ Ĥ +ìħ ¸ +íı Ħ +âľ ½ +ì¦ Ļ +âĸ £ +ê± į +ê¿ ĭ +ì« Ħ +ìº ĩ +ðŁĩ µ +ðŁij ij +âľ ĺ +ðĿij Ľ +ìį ½ +ìº ī +ï¬ µ +ðŁĶ º +âĦ ® +íĥ ¤ +ðŁĩ º +ðŁĴ µ +íħ ¨ +ï½ ij +Î ¨ +ìĥ ¹ +ìĸ ķ +ì¹ µ +ðŁĵ ± +ठµ +ðŁij Ĭ +ðŁĴ Ħ +ðŁĴ Ŀ +ãĮ Ķ +ìĻ ģ +Ð ĩ +à® IJ +âĸ ¹ +á´ Ľ +âĹ ĺ +ëº ¨ +íĥ ī +ìĸ Į +ðŁIJ ¶ +ãĤ ij +Ë ĩ +Å ı +á½ ¹ +ìħ § +ï¹ ° +ðĿij ¡ +ðŁĶ Ŀ +ðŁĺ » +ðŁĴ ĥ +ðŁ¤ ¦ +ðŁį Ĵ +íĢ µ +âľ Ĩ +ë¹ ´ +ï§ ¤ +ï» Ļ +á´ Ĺ +ðŁĮ ´ +Í ¾ +ëĮ ij +ì¨ ĭ +ìµ ¸ +ðŁİ Ī +ðŁı ł +á½ ± +Û Ĩ +á¿ ĸ +âĢ Ľ +ì° ¼ +íķ ¥ +íĹ ´ +ðŁĩ ¬ +ì° Ŀ +âĪ ł +ï¼ ĩ +âĬ Ļ +âĿ ij +ëĦ ĭ +ëŀ Ĺ +ë° ī +ìĹ Ĭ +ì¢ Ĩ +íĮ ¥ +ï° ² +ðŁĵ ĸ +ðŁĺ ® +âļ ª +ðŁĺ ļ +âĿ ŀ +ðĿij Ł +ðŁİ Ĥ +Å ķ +áIJ Ī +êº ½ +ì± ł +ïº Ŀ +ê¿ ī +áĥ ł +ðŁı ĥ +ðŁĴ ¸ +âĿ ģ +âĹ ¾ +Ú ª +á¹ ĥ +íĬ ¬ +ðŁĩ ± +íİ Ń +ðŁĺ ŀ +ë¾ ° +á¹ Ľ +ëĽ ¸ +âĿ Ĥ +êĴ ³ +âĶ IJ +íĵ ° +âŀ ł +ê´ ĺ +ëħ ĺ +ë» ¥ +ì¾ ħ +ðŁĺ IJ +âĪ ª +ðŁij ģ +âĪ ´ +âĹ ģ +ëº IJ +ìŀ ¤ +ì± Ĺ +ðŁı ¾ +Î § +á½ » +âŀ ¥ +ìŁ Ī +ï» ī +âĸ Į +ãĥ ® +ðŁ¤ ¤ +âĩ ĵ +ì¼ ł +á´ ı +ë§ ¬ +ë» £ +ðŁĴ ¬ +ðŁį ĵ +Ä ¸ +Ù ¹ +Ê ¿ +á½ ° +ëķ ľ +ì° ¡ +ì° » +íİ į +ðŁİ ¯ +ðŁį Ĥ +ðŁij § +âĻ ¢ +áĨ ŀ +âĻ § +âļ ľ +âľ ī +ëĵ ¦ +ëŃ £ +ìĪ ı +ìĵ ± +Å Ń +Ê Ĭ +âĴ ¸ +âĩ © +ðŁĴ Ķ +Õ µ +Ð ī +Ò » +ë§ £ +ìĽ ľ +ì¿ ¡ +íĽ ħ +íĽ ¤ +ïº ¢ +âľ ĭ +âĪ Ī +ðŁĮ į +Ê ľ +ëĬ ª +ëĴ ¹ +ïº ² +âĸ Ħ +ãħ Ī +ëļ ¤ +íİ © +âĪ ¨ +ðŁ¤ ª +áĥ ļ +ê³ ¶ +íĬ ķ +ðŁĺ ¬ +âĪ « +ðŁij ĭ +Ò IJ +íĬ ¿ +ðŁĶ µ +ðŁĴ ¨ +ðŁĮ Ļ +ëĩ © +âľ ³ +ë¨ ģ +ëº Ħ +ìĻ ij +ìº ħ +íı Ī +ðĿij Ļ +ðŁĴ ĺ +ãİ ¥ +âĿ ı +âľ ° +ï¯ ¿ +ëµ IJ +ì¼ IJ +ïº ± +Õ ´ +ï¬ Ģ +âľ ´ +ðŁ¤ Ń +ðŁij Ĩ +âĽ Ķ +ê· ĵ +ìĮ Į +ðŁ¤ · +Û Ķ +ðŁ§ ¡ +ðŁĺ ĵ +Î ĸ +âı ° +ê² ľ +ëĭ ³ +ëİ ħ +ë° Ī +ï® IJ +ðŁı ¡ +âĨ ª +âĵ Ķ +âľ Ĭ +Ï ² +Ü IJ +ðŁĩ ³ +Ö Ĥ +âľ ı +ìĸ Ĺ +ì« Ļ +ðŁĺ ² +Ä Ń +âĻ Ń +âĶ ı +âĹ Į +ðŁĺ ¯ +áµ Ĵ +íĬ ł +Ä · +Ê ģ +à¤ Ł +á¹ ģ +á¼ ° +á¿ Ĩ +â « +â« ¸ +ëį « +ì³ ĩ +ì¼ ¤ +íĽ ¨ +ðŁĴ Ł +Ê Ģ +Ê ³ +ëĵ IJ +âķ ° +âĿ ĩ +Ç Ģ +Ç Ķ +É ´ +âĺ ļ +âĺ ľ +ê¶ Ĥ +ì« Ĵ +ì± Ī +ðŁĩ ¨ +ðŁİ ¥ +ðŁĵ Ŀ +Ä § +ðĿ ijIJ +Û Ī +ठ¬ +ì¬ IJ +íĹ ¥ +âĻ ¨ +ðŁį ´ +ï¹ ı +Ë ĭ +ðŁ¥ º +âĸ ¨ +íĻ ĭ +âĪ ħ +ëģ Ļ +ëŀ ł +ìĨ ¥ +âĢ ĸ +ðŁ¤ ĺ +ðŁIJ » +áµ ķ +Ç Ŀ +âĺ ı +ïº ļ +ï» Ĥ +ðŁļ © +ìĪ Ł +Ë Ĭ +⤠µ +ðŁĴ § +ã ħį +ë© © +Æ ¬ +Î ĩ +âĩ § +âĵ ļ +ìĤ ¯ +ìĪ ¯ +ëĨ ĭ +âľ ¯ +ðŁļ Ģ +Ú ĺ +Ú ¨ +âľ Ń +ê² ħ +íĮ ° +íľ Ļ +ðŁĮ Ĭ +ðŁİ ĵ +ðŁĺ Ļ +Ë ĥ +ðŁĴ ģ +ðŁij İ +âĺ ¹ +ðŁĺ « +ðŁĴ » +ëĤ µ +ìĿ Ĭ +íĮ » +Ò ³ +á½ ² +âŀ ŀ +ëĤ ij +ëĿ Ī +ì£ ¤ +ï» ¯ +ðŁĩ © +ðŁ¥ ³ +âĴ ¼ +ðŁ¦ ĭ +âĺ Ĥ +ðŁĺ ° +ðŁĻ ĥ +ðŁĺ Ĵ +Û İ +Ï ķ +Ḡ¤ +ë£ ½ +ìĬ ¥ +ðĿij ī +É IJ +ðŁį İ +âķ ¯ +âķ ¹ +ຠ² +ï¾ ł +ë¹ ķ +ïº Ĩ +Ê º +Ó § +âĨ ł +ëĥ ĩ +ìİ Ī +ìŁ ¤ +ï± ¢ +âķ ¬ +âĺ ł +ðŁİ Ĭ +ãį į +ãİ İ +âĺ ° +âľ ĥ +ãħ ī +ë¯ Ī +ë¹ ¤ +ìı Ń +ðĿij ¢ +ðŁIJ ¾ +Å ĭ +ðŁij ¶ +âĶ Ľ +ï¿ ¢ +áĥ ¡ +Ä ¼ +Å Ĩ +Ñ IJ +ìĥ Ľ +ìĺ Į +ì± ¤ +íħ ģ +íļ ĥ +ï³ Ĭ +ðĿij Ķ +ðŁĩ « +âĭ ° +ðŁĺ ¨ +âĤ © +Õ ¬ +Ḡį +á» ´ +âĨ ĺ +âĺ ¯ +ãħ ı +ìł ¬ +âĻ Ķ +ðŁĶ Ķ +ðŁĺ ł +ðŁĻ Ĭ +à® ľ +á¹ ħ +âĹ IJ +âĿ Ī +âŀ ½ +ìĥ ħ +ðĿij ł +Æ ¢ +âĭ Ļ +ê° Ľ +ëĿ µ +ë£ Ł +ìı ľ +ïº ģ +ðŁĴ Ń +âĬ ĥ +ðŁIJ ° +ãħ Į +Ü ĵ +âŀ ķ +á½ ģ +ìķ ³ +ðĿij Ŀ +ðŁİ ¬ +É ¡ +à¤ Ĺ +áIJ ī +ì© ľ +ì¶ § +ï³ ī +ï» ħ +ðĿIJ ŀ +ठ¶ +ðŁĵ ¢ +ðŁį ĭ +ðŁĴ ħ +ï¾ ķ +⬠Ĩ +âĪ µ +ðŁ¤ ij +áĥ £ +Æ Ħ +Ñ ¹ +á¼ Ķ +ê° ł +ê´ Į +ê· IJ +ëĽ ´ +ì± ĺ +ï® Ń +ïº ¹ +ïº ¾ +âľ Ĺ +âĿ ¦ +ðŁij ¦ +áĥ Ĺ +Ù ² +á½ ´ +âĪ ı +âľ ® +ê¹ ° +ë² µ +ìĦ Ģ +ì© Ŀ +ïº ŀ +ïº ½ +ðŁĩ Ń +Ë Ĥ +ðŁį ij +ðŁį Į +ðŁĶ » +ê¹ ¬ +ìĬ Ń +ìľ · +ðŁĽ ij +Ç § +ë¼ Ľ +ïº ¡ +ïº º +ðĿij ļ +ðŁĵ ¦ +ðŁĶ İ +ðŁĹ ĵ +áĥ Ķ +âľ Ĵ +âľ ¡ +ðŁĮ µ +âĶ ķ +ëĢ Ŀ +ðŁį Ĭ +âĺ ĥ +ìĺ ħ +ঠ¬ +ðŁ¦ ģ +âİ ¯ +ðŁIJ ķ +Ñ ¿ +ॠ¤ +༠ĭ +ê· Ī +ì« Į +ðŁĩ ° +âĿ ī +ì« Ģ +íĿ Ħ +ðĿIJ ¢ +ðŁļ ¨ +âĻ ¤ +ðŁĺ © +ðŁį į +ðŁĺ ij +ðŁļ ļ +Ö Ħ +ë « +ë« ¼ +ठı +á¿ · +âĮ © +âĺ IJ +âŀ £ +ê¸ ± +ê¼ ¿ +ëĦ Ŀ +ìı ´ +ìļ ¤ +ì¿ ± +íİ IJ +ðŁĴ ¢ +ì´ IJ +âĩ ij +âĶ ĵ +âģ ¾ +Ü Ŀ +ðŁ į° +â´ ° +Æ ı +Ï Ł +Ú º +Û ĥ +áĦ Ĵ +âĪ Ł +âĿ į +ãĦ ² +ìľ ħ +ì¤ ı +ðŁĩ ² +êº Ħ +ðŁİ ¤ +âľ £ +⸠Ŀ +ï¸ µ +ຠ§ +áĢ Ļ +âķ ł +Õ ¯ +âı © +ðĿij £ +ðŁĴ £ +Å ĺ +ॠIJ +âģ ĥ +âĮ ĺ +ê» Į +ìĮ Ķ +ðĿij ĺ +ðŁ¤ ĵ +Õ ¿ +à¤ Ń +âĮ ļ +âľ Ŀ +ðŁIJ ¼ +Ë Į +âķ ļ +ï¦ Ĺ +âĿ ķ +âķ £ +ðŁIJ ± +à® ¤ +Ñ ¾ +ठļ +ठľ +ìĪ Ħ +ìļ ľ +ðŁİ ® +É Ĵ +Ú · +ຠį +âĨ µ +â Īĺ +âĿ Ĭ +ë¿ į +ìIJ Ī +ìļ ĺ +ì¯ § +íĥ ¯ +ìĸ ı +ï¸ ° +ðŁĩ ¯ +ðŁ§ ļ +ðŁĺ µ +ðŁĺ · +ðŁĮ ³ +ຠ¥ +Ä ī +Ä ¥ +âľ ¶ +á¿ ¾ +âĬ ± +âĺ ¾ +ê° ī +ê¼ ° +ëº ij +ðŁĶ Ĭ +ðŁĸ IJ +Å ¤ +Ò « +à® ® +âĮ Ī +âĹ Ĺ +ëĦ µ +ëħ ľ +ëľ ¹ +ðĿij ¥ +ðŁĴ ¿ +ðŁĽ Ĵ +Ê Ĵ +áŀ ĵ +ðŁIJ Ŀ +ðŁ¦ Ħ +ðŁį · +âĺ Ł +ï¸ ¶ +ðŁ¤ Ł +Ô ± +âĨ ² +âĪ İ +âľ « +ëĩ ½ +ëı IJ +ëķ Ħ +ï¦ ³ +ï§ Ŀ +ïº Ļ +ðŁij » +ðŁĵ º +êµ ¼ +ìĮ © +ðŁĮ ² +È ± +íĶ ķ +ðŁĺ ¤ +ãĮ ¢ +Ê Ķ +ठ¡ +á¼ Ī +ëİ ĥ +ë© ± +ë® Ī +ðĿIJ « +âĬ ķ +ëĥ ł +ë» ¬ +íĭ Ķ +Õ ¤ +á¼ ± +âľ ¥ +âĺ Ħ +âĪ ¥ +âļ ķ +ðŁij Ħ +ðŁİ ħ +àº Ļ +âĶ ¬ +á½ µ +Õ ¾ +Ö ģ +âĹ Ķ +ê¿ į +ëĸ µ +ë© İ +ë® ´ +ìķ ´ +áĥ ľ +á¼ ¡ +âĶ Ĭ +âķ ® +âĹ ¼ +ðŁį ¾ +ðŁĽ į +ðŁij Ĺ +ðŁ¤ ŀ +âľ Ħ +Õ Ģ +ঠ² +Ë ī +⣠¨ +Ä ¯ +Ï Ĭ +á´ ľ +ë¹ ³ +ï³ ĭ +ï¿ ł +Ä ª +âĤ ¸ +âľ ± +ê» IJ +ëĭ » +ë§ ¸ +ìŀ ¿ +ì© ¨ +ì ŃIJ +ì° ¿ +íħ Ł +ðĿIJ § +ðĿij ij +ðŁĮ İ +ðŁĵ ® +ðŁķ Ķ +âĹ Ļ +âĹ » +âŀ § +ìŁ Ŀ +âľ ¬ +ãĥ ° +âģ Ī +â ĵĺ +ðŁ ĴĮ +ï¬ ĥ +àº Ķ +ìĶ ° +ðŁĺ ª +× Ģ +ìĥ ¨ +ïŃ ĭ +ðŁį ķ +ðŁĺ ´ +Ï ³ +á¼ Ħ +á½ ħ +âĩ ¢ +âķ Ń +ìĺ » +íĬ ¤ +Ü ĺ +⤠´ +âĹ į +áŀ Ł +ðŁį º +áŀ ļ +ðŁı Ĭ +ðŁIJ · +Ê Į +á½ º +âģ » +ê½ Į +ëĪ Ĺ +ë Ĺı +ì¿ ° +íĢ ¼ +íį ħ +ï· ² +ðŁĮ ı +ðŁį « +ðŁį ³ +ðŁİ ° +ðŁij ° +ðŁĴ ² +á¥ Ļ +ðŁIJ Ł +ï¿ ¡ +ðŁĹ £ +ðŁį ľ +âľ ² +ãİ ¢ +ðŁĶ ° +á¼ ¸ +á½ ij +Ä İ +áĦ Ģ +âĻ ķ +ëł Ŀ +ìĪ ´ +ïŃ Ń +Ó ľ +Ô Ģ +ëĢ ľ +ëĥ Ķ +ìĬ Ľ +ì« ij +ìº ¥ +ìº ¬ +ðĿij ¦ +ðŁĶ ¶ +ì¾ ¨ +ðĿIJ ļ +ðŁį » +ðŁĴ į +ðŁ¤ ¡ +ðŁķ Ĭ +â½ ĩ +âĵ IJ +ðŁį Ń +ðŁį ª +ðŁĶ Ĩ +Ò ¡ +á´ ĩ +É Ĺ +Ü Ķ +âĦ İ +âĿ ĥ +ëĹ Ģ +ï² Ķ +ïº Ī +ðĿIJ » +ðŁĴ Ĭ +ðŁļ « +Ñ ° +Ñ ³ +ठ· +âĹ ł +ðŁij ¤ +ï¾ ĩ +âĺ ĵ +ðŁį µ +ðŁ¤ ¨ +âĸ Ń +à® ´ +Ü ¢ +Ü ¬ +à´ ® +ðŁķ º +Ô ¹ +Õ £ +à´ ¯ +á ´Ģ +âĮ ī +âľ IJ +âŀ ¦ +ê¹ ½ +ëĮ ľ +ðŁı ¥ +ðŁĵ © +Ò ¹ +Ó ĺ +ठħ +âĿ § +Æ Ĺ +âĹ ½ +ðŁij « +ðŁİ § +ðŁij £ +âľ » +ðŁĻ ħ +ðŁĺ ĸ +ðŁĴ ® +ຠ° +ðŁĶ ľ +ðŁį Ħ +ðŁ¤ Ŀ +á ĥĿ +áŀ Ģ +âĩ ¦ +Ê ¾ +Ò ® +Õ ¼ +ठĨ +âĹ ħ +âļ ĵ +âļ ĸ +ê¿ © +ë¯ Ħ +ìIJ IJ +ìŀ ° +ì§ Ń +íĭ ĭ +íİ ¨ +íĻ § +ï² ij +ðŁİ Ĺ +Ù ³ +ðŁij ¸ +ঠ® +ðŁij ķ +Ú µ +âĢ ¾ +âŀ ° +ðŁij ¯ +ðŁİ ¼ +ðŁı ģ +Ä º +Ê ı +Ú ³ +âı ± +ê½ Ī +ëĿ Į +ìĮ ī +ìĹ · +ìŀ ´ +íĹ ¹ +íľ ¨ +ðĿĹ ² +ðŁĮ IJ +ðŁİ Ļ +ðŁı µ +íĽ Ļ +ðĿij ħ +ðŁĺ ¶ +âĵ ħ +âķ ¥ +ðŁį ı +ï¦ İ +Õ © +ðĿIJ Ħ +Ó £ +Ú ¿ +âĻ ļ +ðŁĶ Ĺ +Ḡ« +âĭ ® +âĸ ¦ +⼠½ +âľ µ +ãħ Ĩ +ãħ Ĭ +ëĦ Ļ +ëĿ ¨ +ë¥ Ħ +ìĦ ¦ +ì§ ° +ì§ ¹ +íī Ī +ï§ ij +ï» ĩ +ðŁĮ ¾ +ðŁı ĸ +ðŁIJ ij +ðŁĴ ³ +ðŁĵ Ĩ +Û ĩ +Ü ķ +á½ ½ +ëĦ ľ +à´ ² +à´ ³ +àº Ń +áĥ Ľ +âĿ Ķ +âij ħ +áĥ ¥ +ðŁĵ ħ +âŀ ³ +á´ µ +ï¹ ¡ +ï¹ ¶ +Î Ĩ +ठ¥ +áī µ +âĿ Ļ +âĿ ± +ëī ł +ëİ ł +ëı Ľ +ë¿ ħ +ìĶ ¸ +íij ¯ +íŀ ī +íŀ Ľ +ï§ Ħ +ïŃ ĺ +ïº ¦ +ï» ¸ +ðĿij Ĥ +ðĿij ı +Ï ij +Ú ł +áĢ Ķ +áŀ Ķ +á¹ ¢ +ëĦ ¸ +ðĿIJ ¨ +ðŁĩ ´ +Õ ° +ðŁij ł +ðŁį Ĩ +ðŁı Ģ +ðŁ ijIJ +ðŁį ĩ +ðŁIJ £ +áĪ Ń +Ü ª +ðŁ ĮĢ +áŀ ĺ +âĩ Ħ +ðĿIJ Ģ +Ê Ļ +âĶ ¼ +ðŁı ¿ +Æ · +È ł +Ñ ½ +âĤ ¨ +ê´ Ń +ê¹ » +ëĶ ¨ +ìĪ Ģ +ì¾ ° +íĨ Ī +ï® § +ï¯ ½ +ðŁĶ ħ +ðŁĶ ® +Å ¢ +Ê ° +Ñ ¸ +ठ£ +âĬ Ĺ +ëª Ħ +ï¹ · +ïº ħ +ðĿIJ µ +ðŁĮ ¶ +ðŁĵ ° +ðŁĶ · +ðŁĸ Ĵ +ðŁ¤ ² +ëī © +ðŁİ Ĩ +ðŁ§ IJ +ðŁį ® +âĨ º +âĿ ¢ +ðŁij ª +ðŁij ± +âĨ ¡ +áŀ ı +Ú ķ +ðŁį ¹ +ðŁĴ Ģ +Ë ® +Ó ¨ +Ö ħ +ठĩ +âĤ ¡ +âĪ ķ +âĺ ī +ê¹ ¼ +ê¼ IJ +ì½ ¸ +ðĿIJ ¬ +ðŁı ħ +ðŁij Ļ +ðŁĴ ī +ðŁ¤ Ļ +È ĺ +É ³ +É ¹ +Ù º +áĢ Ħ +á¿ ³ +âļ ĺ +âĿ Ĩ +ëĨ ī +ìĸ į +ìĺ ĩ +ì¥ ĺ +íĸ ħ +íĻ ij +ï® Ĭ +ï¿ Ń +ðĿĴ IJ +ðĿĹ ¢ +ðŁĶ ĸ +ðŁĶ ¨ +ðŁļ ij +ðŁļ ² +Æ ¸ +âĹ ¥ +ðĿIJ Ń +ðŁį ½ +âĹ ij +âĵ ĩ +ðŁĶ ± +âľ ¼ +ï¹ ĥ +âķ ± +ãĢ Ĺ +ðŁı ĭ +ðŁļ ´ +ðĿIJ ® +Ä ļ +Õ ı +Ä ¶ +áĥ ij +á¹ ¬ +Ä Ī +Ä Ĵ +Ò ° +Ó ķ +â IJ +âIJ £ +âĹ ¢ +âļ Ļ +ãħ Ĺ +ê° ¬ +ê³ ª +ê» Ģ +ëĦ ´ +ëİ ģ +ëĿ Ķ +ë¬ ½ +ëŃ į +ìĩ ³ +ì° ¹ +íĮ ¹ +íŀ Ŀ +ï® ĭ +ï ¶Ī +ðĿĴ Ĥ +ðŁ¥ Ģ +ðŁ¦ ħ +Ê ĺ +á¼ ij +âģ İ +ðŁį ŀ +âĨ ĸ +âĨ Ļ +ðŁİ ĥ +âĦ ¡ +âĭ ± +ðŁĶ į +ಠ¨ +áµ ĥ +âĶ « +⦠¿ +ðŁĩ » +Æ ¤ +Ò ı +Ò · +Û ī +à® ķ +Ḡ³ +ï¬ ± +ðŁĨ Ķ +Ú Ń +Û ¦ +áħ ¡ +âĦ ¹ +ê¿ İ +ëķ Ķ +ë¼ ī +ìļ § +ì² µ +ì´ ¨ +íĬ Ī +íĸ IJ +ðĿĹ ĺ +ðŁĩ ¿ +ðŁİ ĸ +ðŁij ħ +ðŁ ĵĺ +ðŁļ Ļ +ðŁĽ µ +à¶ ½ +⼠µ +ðĿIJ ³ +ðĿIJ ¸ +âļ Ķ +ðŁij Ń +Ó ij +âĶ ¯ +ðŁħ ¿ +ðŁĺ ¹ +ï¿ « +â¼ ¤ +ðŁĴ ĩ +ðŁĵ İ +ðŁĸ ĭ +ঠ¸ +ðĿIJ į +Ä ² +Ï ĭ +Ñ ¬ +Ú ¬ +Ü Ĵ +á´ ¬ +ï¨ Ħ +É £ +Ë ij +Ï µ +Ò Ŀ +Û ¥ +Ü ł +๠Ľ +áĥ ķ +áĬ ķ +á¾ ¶ +âĤ · +âĩ ¾ +âķ © +âĸ IJ +âĺ ª +âĺ ® +âĿ ļ +âĿ Ń +âŀ ± +âµ İ +ãı Ĭ +ë© ĵ +ìĹ ¾ +ìª Ħ +íĵ Į +íķ ¼ +ïŃ ¬ +ðĿij Ĩ +ðĿij ŀ +ðĿĸ Ĭ +ðŁİ ¸ +ðŁı Ħ +ðŁij µ +ðŁĴ ł +ðŁĶ ĺ +ðŁ¥ Ĥ +Å ª +à· ĥ +á´ ¼ +âĬ ° +ë³ ı +ë´ £ +ï¥ ľ +ðŁĵ Ī +ðŁķ ¯ +ðŁ§ Ģ +âĻ IJ +ðŁĨ Ĺ +ðŁĵ ķ +ðŁ§ ģ +Ü « +âĿ IJ +Õ ķ +འķ +âŀ Ŀ +ঠķ +ðĿIJ ¶ +É ¢ +Î Ħ +áĨ ¢ +âĤ ± +Õ į +à¡ ķ +á´ ° +Ḡ© +⼠· +âĿ ® +ê¡ ĵ +ëı ¤ +ëĹ IJ +ëµ Į +ìij Ī +íı ¿ +íĹ µ +ðĿIJ İ +ðŁĨ ĺ +ðŁı Ł +É ¥ +Õ » +à¡ Ķ +ठĸ +á´ ¸ +âİ Ļ +âİ ¥ +âı ³ +ëģ ķ +ëĬ ī +ì¡ į +ì¹ ¡ +ï¦ ¶ +ï¬ Ł +ï® « +ï® ¯ +ï± ĥ +ï ·» +ïº µ +ðĿĹ Ķ +ðĿĹ ¡ +ðŁİ ¨ +ðŁĶ Ĵ +Ú Ľ +ठ§ +âŀ ¹ +áĢ Ģ +ðŁį ħ +âĹ ¤ +ठł +ðŁIJ ¥ +áĥ Ĵ +ðŁı Ŀ +ðŁį ¼ +ãĮ § +âĿ Ľ +ðŁIJ Ī +ঠ¯ +áĢ ŀ +ãĢ ĸ +áŀ Ļ +ঠª +Õ Ĩ +âĬ Ĩ +âľ ¾ +ðŁIJ Ĺ +ï¹ ¿ +Ä ¦ +Ü Ł +ಠł +ಠ¥ +áŀ ī +á´ ¥ +á´ © +á½ Ģ +á½ ¡ +âĨ ķ +âŀ ¯ +ê¡ ij +ëij £ +ë± Į +ìĪ ij +ìľ Ķ +ìŀ ½ +ì¨ į +ðĿij Ģ +ðŁĮ Į +ðŁį ¦ +ðŁį © +ðŁIJ ļ +ðŁĵ Ĵ +ðŁĵ ¹ +ðŁ¥ ij +Ä ĭ +Ë Ĺ +Ñ « +Õ ¢ +Ú ° +â ĮĢ +âĹ Ĥ +âĹ £ +âľ Ľ +âĿ Ĵ +âĿ ĺ +âŀ Ļ +âŀ ² +ãİ į +ê¡ IJ +ëŀ ĸ +ìĬ Ŀ +ìĽ ¤ +ì¡ ĭ +ì¨ ° +íĹ Ļ +ï¥ ¸ +ï³ į +ï» İ +ðĿij ĵ +ðŁĵ Ĭ +ðŁļ ¼ +ï¦ ģ +ðĿķ Ĵ +ðŁ ijľ +ðŁij ¿ +ðŁĩ ½ +à· Ħ +âĸ ´ +ãį ī +âĬ ĩ +ðŁ§ ¸ +Ú ¡ +â¾ ĥ +ðŁĹ » +âĵ ij +ðŁ¤ ¸ +ðŁ¤ ¯ +êĴ ° +ðĿIJ ĵ +âĶ ´ +êĴ ± +áĢ ĺ +â ĽĦ +ï¹ ¹ +Ó Ķ +áĥ ± +Ü ¡ +ß ŀ +âĻ ı +âľ ¸ +ìij ¨ +ðĿIJ Ŀ +ðĿIJ ¥ +ðŁį ī +ðŁij ¼ +ðŁ¥ Ŀ +Æ Ķ +Ý ¬ +ठ« +ຠļ +á´ ´ +á½ ĸ +âĤ ¶ +âİ ¢ +âĿ ħ +⣠« +ãİ Ľ +ë® ¨ +ëº Į +ë¼ ĺ +ìĨ Ŀ +ìľ ³ +ìŀ Į +ì£ Ĺ +ìª ĺ +ì» ¹ +ï· ¼ +ïº Ĥ +ðĿIJ ´ +ðĿIJ ¼ +ðŁĮ ļ +ðŁı « +ðŁĴ ¤ +ðŁĴ ¶ +ðŁĴ ¼ +Ê ķ +Ê ½ +â² Ł +ãī ł +ê¡ Ĵ +ëľ Ģ +ìĥ ¾ +ì¸ ¤ +ï¥ ģ +ðĿļ Ĭ +ðŁļ ĥ +âŀ Ľ +ìħ ´ +áĦ ĭ +âĩ Ĺ +ï§ · +âĺ ĸ +ðŁIJ ¦ +⸠ľ +ðŁĴ ´ +ðŁ¤ ļ +ãĬ Ĺ +âĮ Ľ +áĪ Ľ +༠º +â½ ī +ðŁı ¢ +âĵ ŀ +âĺ ½ +ãĢ Ļ +ðŁ¤ ® +Å IJ +áĥ ¬ +ðĿĹ » +ðŁį ĸ +Æ Ĭ +Ê Ł +ß ĭ +ठĭ +áµ Ķ +á¿ ĥ +âĦ ī +âĮ ĭ +âı ² +âĵ Ī +âĵ ¢ +âķ Ķ +âļ ij +âĿ ĭ +âĿ İ +â µľ +âµ £ +ëĴ Ī +ëľ ģ +ë¶ ĩ +ìį » +ìĺ Ń +ì§ ¢ +íĹ Ģ +ï§ Ĭ +ï ¬¸ +ï± ¡ +ðĿIJ º +ðĿij § +ðĿĺ ¦ +ðŁĵ ¥ +ðŁĺ Ł +ðŁ¥ IJ +Ä ĸ +É ¨ +áĢ IJ +áĥ ĵ +Ạĵ +á¼ ¶ +á½ Ħ +âĤ ¤ +âĮ ľ +âĮ Ł +âİ ł +⼠¸ +âµ į +âµ ı +âµ ĵ +ãĢ ĺ +ë ·¸ +íħ ¼ +ï¦ Į +ïŃ Ħ +ïŃ İ +ðĿĻ ļ +ðĿļ ĺ +༠ĵ +ëŃ ħ +áIJ Ľ +ãİ ¾ +ï¨ Ģ +ðŁĹ ½ +âĻ ŀ +Ë ĸ +âĹ ŀ +ðŁ¤ « +ðŁĺ Ĺ +ï½ ¦ +ðŁ¤ ¢ +âģ ĩ +ãĢ µ +ðŁį Ķ +áĬ ł +ðŁĺ ¼ +ðĿĹ ® +ðŁIJ ³ +ðĿIJ ĭ +ðŁĨ ļ +ðŁĶ Ľ +Ñ » +Ü ¨ +à® ² +âľ ŀ +âµ Ļ +êµ £ +ì¸ ¨ +ðĿ IJľ +ðĿĺ ° +ðŁĶ ½ +Ç » +Ç ¿ +Ê ĩ +Î IJ +Ð Ģ +Ñ ¡ +Ñ ² +Ò Ĵ +Ù ¶ +ß ķ +à¶ ± +áIJ ģ +âģ ŀ +âĸ § +âĽ Ī +âľ ľ +âľ ¹ +⣠¹ +⤠ĩ +ê² Ĭ +ê¾ ľ +ë¯ IJ +ë³ IJ +ìħ © +ìIJ ¬ +ìij ¹ +ï¤ Ķ +ï¦ ļ +ï¬ ł +ïŃ Ķ +ïº ¶ +ðĿĴ ı +ðĿĸ Ĩ +ðĿĹ ¶ +ðŁı Ĥ +ðŁIJ ½ +ðŁĴ © +ðŁĵ ½ +ðŁĹ ¨ +ðŁĹ º +ðŁĺ ¸ +ðŁ¥ § +Å Ĺ +Ê İ +Ò Ļ +× ² +à¤ Ī +á¼ ´ +á¿ ij +âµ ī +ãħ ĵ +ì½ ´ +ðĿĸ ĵ +ðŁĵ Ĺ +ðŁĶ ª +ðŁĸ į +Ï Ĵ +ðŁij ¬ +áĥ Ļ +âĨ ¬ +âĶ ¤ +⼠¹ +âĻ Ł +ðŁļ ¶ +ðŁij ¾ +âĪ ĭ +ðŁIJ ¯ +à¼ İ +âľ · +ï¨ Ļ +âĶ » +ðŁij ¹ +áĦ ī +ຠª +â¾ ı +â½ ħ +ãİ ĸ +Ñ ´ +Õ ® +Ú ¼ +áĢ ķ +áĨ ¼ +ëŃ ı +ðŁIJ ¸ +ðŁļ £ +Æ Ŀ +Ô » +áĥ ¢ +ðŁį ¯ +É ¦ +Õ ¦ +âĻ ĭ +ï¬ « +ðĿĹ ¦ +Ç ļ +É ± +ठī +á´ Ħ +âĻ ĵ +⼠° +⣠ª +ëĥ ĺ +ë¢ ¸ +ìĤ ij +ï® Ķ +ðĿķ ĸ +ðĿĹ § +ðŁĩ ¼ +ðŁĵ ĭ +ðŁļ ľ +ðŁ¥ ¤ +Ä ® +Å · +ß Ĭ +ॠ¥ +à® ª +áŀ Ħ +áµ Ģ +Ḡħ +á¼ ¢ +âĪ Ŀ +âĬ ¹ +âĴ ¶ +âķ ´ +⼠± +⼠³ +⼠º +âŀ Ł +ãı Ħ +ê¸ Ķ +ê¹ Ł +ëĩ ° +ë¹ » +ìĤ ¥ +ìĽ » +ì° Ł +íĥ ° +íĨ º +íļ ½ +ï¤ ´ +ï¥ ¾ +ï³ Ŀ +ðĿIJ ¦ +ðĿĴ ľ +ðĿĴ Ł +ðĿļ Ĺ +ðŁİ Ń +ðŁı ĵ +ðŁı ³ +ðŁı º +ðŁIJ į +ðŁij ĥ +ðŁĴ ı +ðŁ¤ ĸ +ðŁ¤ µ +Õ ² +âµ Ķ +ëĺ ¬ +ï¦ £ +Ê Ĥ +áĨ « +áŀ ij +ðĿĸ İ +ðĿĹ ĸ +áĦ ĥ +âĩ ł +áĢ ¡ +འĦ +âŀ ¸ +ï¦ Ļ +âĩ ļ +ðŁIJ ¬ +ðŁIJ ¢ +â¾ Ĵ +ðŁIJ ¤ +ðŁĶ « +ãĢ ŀ +ï¸ º +ðŁĺ º +â½ ´ +ðŁĨ ķ +âģ ¿ +ðŁį ¨ +ಠķ +ðŁļ ĺ +áŀ ħ +ঠħ +áŀ ¢ +ਠľ +â ļĮ +ãĢ ½ +à· ´ +âĵ Ľ +áĢ ľ +ìĨ ¨ +Ë © +Ü Ĺ +âĭ ¼ +ðŁĻ ī +Å Ĭ +É ĵ +Ê ² +Î ° +Ñ ¼ +Ô ¿ +à¡ IJ +༠ľ +འ¦ +á¶ ľ +âĤ ² +âĨ ¨ +âĬ ¥ +âķ § +âĻ ľ +ãĭ ¡ +ë´ ¬ +ë¶ ij +ìī ¿ +ìİ ħ +ìł ± +ì° § +ï² ¡ +ðĿĴ Ľ +ðĿķ £ +ðĿĹ ľ +ðŁį ² +ðŁİ © +ðŁIJ IJ +ðŁIJ ł +ðŁij ½ +ðŁĴ ij +ðŁĵ ľ +ðŁķ µ +ðŁ ļĮ +ðŁĽ £ +Ê ĭ +Ó ¯ +Ù ¸ +ß Ķ +ß Ļ +à¡ ĵ +á´ į +Ḡ¿ +âı º +âĸ ¥ +ë¤ ½ +íľ ij +ðĿIJ ¹ +ðĿĸ Ķ +ðĿļ İ +ðŁĵ Ħ +ðŁ¦ · +Æ ĥ +à¦ Ł +âĮ Ĥ +âĺ Ń +â² ļ +ëĿ ķ +ðŁİ £ +à® ĩ +འĨ +áħ µ +áĹ ľ +âĢ ½ +âĮ £ +âģ ½ +ðŁĵ ¬ +ðŁ¤ § +âĩ ª +â½ £ +âĹ Ł +ï¨ Ĺ +êĴ ª +ðŁĽ Ģ +Ç Ĥ +ðŁ¥ ¶ +ðŁİ į +ï¿ © +ðŁij Ĵ +áµ Ī +ï¸ ¿ +áħ © +â¾ ¦ +à° ¤ +á´ ĸ +ਠ¬ +àº Ĺ +༠» +Ñ º +ਠª +á´ ³ +ðĿIJ Ī +à» Ģ +á´ ¿ +âĤ į +âĩ ¡ +⼠ª +ðĿIJ Ĥ +ðĿĴ ķ +ðŁ IJľ +Ê į +Ñ ± +འĥ +ë® IJ +ìĽ ¡ +ìľ ģ +ðĿIJ ¿ +ðĿķ ł +ðŁij Ľ +Æ ª +Ï º +Ó ¬ +Ù ¿ +Ý £ +ઠī +à® ¹ +འij +áĨ ¯ +áµ ĩ +âĩ ¥ +âı ª +âĻ ° +âļ Ń +âļ ¾ +ãħ Ħ +êĢ ° +ê° Ĺ +ê² ĭ +ê² » +ê¶ ľ +ê¼ ĩ +ê½ ¹ +ëĤ Ł +ëħ Ī +ëĭ ¢ +ë§ Ł +ëª Ĩ +ëµ Ģ +ì½ ± +íĩ ĺ +íľ ľ +ï§ ¾ +ï± µ +ï² ¢ +ï² ¤ +ðĿĴ Ĭ +ðĿĺ ¯ +ðŁį Ĺ +ðŁı į +ðŁIJ ĺ +ðŁĵ ¡ +ðŁĶ ŀ +ðŁ¤ ³ +ðŁ¥ ģ +ðŁ¥ Ĺ +ðŁ¦ Ĭ +Ä µ +Æ ¦ +Ç µ +É ¯ +Î ı +Õ Ħ +Ü ¥ +འģ +ᨠł +âķ « +ãİ ī +ë· ´ +ìĨ İ +ìİ Į +ì£ µ +íĽ ł +ï§ ª +ï³ ı +ï» º +ðĿij ģ +ðĿij ĩ +ðĿĴ Ĩ +ðŁİ ł +ðŁIJ Ķ +ðŁij Ł +Å ĸ +ठĮ +á¾ ½ +ê¦ Ĵ +à® Ł +á´ ± +ðŁı ° +ðŁIJ ŀ +à½ Ģ +áĢ ħ +âĬ ¿ +ðŁIJ § +ἠģ +â¼ Ī +âĶ ¿ +ðŁ¥ ´ +â¼ ¿ +ðŁ§ ľ +ãħ ¿ +âĦ « +ãĢ ³ +ãĬ Ļ +â¼ Ģ +ï ¦¬ +ðŁı ¬ +ðŁĵ » +áĬ Ľ +áĦ ħ +ຠĬ +ຠĽ +áħ ³ +ðŁij ® +à® ± +âĺ ĩ +ðĿIJ ı +à´ µ +à» ģ +འı +འ¢ +ᥠ± +âĤ £ +ï¥ ¦ +ïŃ Ļ +ï´ © +ï¹ Ĥ +ðŁį £ +ðŁķ ¹ +Ï ĸ +à¶ ¸ +ຠ¢ +áĭ Ń +âİ Ŀ +âĹ Ŀ +âĻ Ī +âĻ İ +ê½ ¥ +ì³ Ķ +ì¼ ij +ï± ° +ðĿij ĥ +ðŁĮ ª +ðŁį ¡ +Å İ +Ê ¦ +Ñ § +Ó İ +Ô ´ +Ú Ī +ß ĵ +ß § +à¤ Ķ +áĪ « +áĪ µ +áĹ © +á´ ł +á¼ ł +âĢ Ĺ +âģ ij +âĦ ı +âĸ ĩ +â² £ +ãĦ ³ +ãī ® +ê³ Ĺ +ëĦ Ĵ +ëĸ « +ë¡ Ħ +ë¹ ° +ë½ ģ +ìĦ ģ +ìĮ ĺ +ìŁ Į +ì³ ī +ì¼ ķ +ï¬ » +ï³ İ +ï¹ ¸ +ï¹ ¾ +ðĿIJ Ĩ +ðĿij · +ðĿĽ ¼ +ðŁİ ı +ðŁİ ŀ +ðŁIJ Ļ +ðŁij Ĥ +ðŁĵ ģ +ðŁĸ ± +ðŁļ į +ðŁļ § +ðŁĽ ¡ +ðŁ¤ Ĵ +ðŁ¥ ŀ +ðŁ¥ © +ðŁ¦ Ģ +ðŁ¦ ĸ +Ë ¢ +Ü ļ +à® µ +áĢ ģ +áī ° +âı Ń +âĻ ¿ +ê³ ĺ +ëı Ŀ +ëķ ĥ +ìħ Į +ìĴ ¸ +ìĽ Ł +íħ Ħ +íľ « +ï§ ĺ +ï¿ ¬ +ðŁı · +ðŁĶ § +ðŁ¥ Ī +Æ ĸ +áŀ ĩ +áŀ ĸ +âģ º +âĹ ľ +âŀ © +ê¦ Ń +ëĻ ¤ +ïŃ ¼ +ðĿĻ ĸ +ðĿĻ £ +ðĿĻ ¤ +ðŁĮ Ŀ +ðŁĶ ij +ðŁĽ ł +ຠĩ +âĺ £ +ãĦ ¨ +ðĿĸ Ĺ +Ó ĵ +âĨ £ +ðŁ¥ ī +ðŁĮ ł +ðŁĺ ½ +ãİ ł +Å § +ðŁIJ Ĵ +ï§ IJ +ðŁĺ ¿ +âĪ ¬ +ðŁIJ ® +⣠± +ಠ¡ +â¾ ¼ +à° ² +Ë ¶ +âĸ ¿ +Õ Ī +áŀ İ +áħ ¥ +áŀ Ĺ +Õ § +ðŁ¤ IJ +ðŁį ł +ঠ¤ +à¶ º +âĻ į +ìĺ Ļ +íĺ ĵ +ï¹ º +ðŁĽ ³ +Å ī +á´ İ +âı ľ +âĶ ³ +ê¸ · +ì¡ Ķ +ðĿĴ Ī +ðĿĴ į +ðĿĴ ¹ +ðĿĵ ĩ +ðĿķ Ł +ðĿĹ ¹ +ðŁĮ ħ +ðŁı ´ +Ä Ķ +Ä ¤ +Å µ +Ç ¾ +Ï ŀ +Ï ¶ +Ô ³ +Ü Ĩ +ß © +à¡ Ĵ +ठĺ +à¶ ļ +འĸ +áģ Ĭ +áĥ ŀ +áĦ Ĥ +áĭ « +á´ º +Ḡ£ +Ḡª +á¹ Ĥ +á¼ · +á¿ ĩ +âĩ Į +âı ¬ +âĻ Į +â® Ł +â´ » +âµ Ł +ê¦ ķ +ê¦ ª +ê¦ ® +ê² Ħ +ê¾ IJ +ëĥ ij +ëķ ĭ +ë¡ ¸ +ë¬ Ģ +ìĩ ¤ +ìĪ © +ìľ ķ +ìŃ ĺ +ì· ° +ì ·¸ +íľ Ģ +ï¤ £ +ï§ į +ï± Ħ +ï³ ij +ðĿIJ ¤ +ðĿĴ ĵ +ðĿĴ ¶ +ðĿĹ ¼ +ðĿĻ Ĭ +ðŁĩ ¾ +ðŁĮ Ľ +ðŁĮ ® +ðŁİ ĩ +ðŁİ ² +ðŁı Ľ +ðŁij ¥ +ðŁij ´ +ðŁĴ Ĩ +ðŁĵ Ĥ +ðŁĵ § +ðŁķ IJ +ðŁĸ ķ +ðŁĺ § +ðŁĻ Ģ +ðŁļ Ĵ +ðŁĽ « +ðŁ¤ ł +ðŁ¥ ļ +ðŁ¥ Ľ +ðŁ¥ £ +Ç ¯ +È § +Î Ĭ +Ò ² +× ° +Û ij +áĥ © +áĦ Į +áĪ į +áī ¥ +áı Ĥ +âģ ± +âĬ ¢ +âĹ ĵ +âĿ ° +ë¿ ¡ +ìĽ © +íģ Ń +íĨ ³ +íĬ Ħ +íĵ ¸ +ï¥ £ +ï¥ ´ +ï± IJ +ï± ¯ +ï³ ļ +ðĿĸ ĺ +ðĿĺ Ģ +ðŁIJ Ĭ +ðŁIJ Į +ðŁij ļ +ðŁĵ ĥ +ðŁļ Ľ +ðŁļ ª +ðŁ¤ ° +Ä ´ +áĥ ® +áĹ ¨ +âĻ ® +â² ŀ +ãĪ Ķ +ì ħį +ãħ ĥ +ï¥ ¡ +ຠ¡ +Õ İ +Õ º +⬠Ľ +â½ ¤ +ðĿIJ ² +âŀ µ +áĢ Ľ +âĶ ħ +âĨ Ł +â¼ Ĭ +ðŁĮ ½ +ðŁļ ¿ +ï¦ Ĭ +ãĦ £ +⼠© +ï© Ľ +ðŁį ± +â¾ ¨ +à´ ¤ +áŀ ģ +ຠŀ +Ê ļ +ðĿIJ Ĵ +à´ ± +áŀ ľ +à® © +à° Ĺ +à´ ļ +âĩ £ +ï¦ ķ +Õ ħ +Æ ĺ +âĤ ¦ +âĶ Ħ +ï¦ Ł +ï¦ « +ðĿIJ ģ +ðĿIJ ĥ +ðŁį ¸ +ðŁIJ ² +Å ¶ +É ĸ +ß ĺ +ภ¦ +à½ Ķ +áĨ · +âģ ķ +âĵ Ĥ +âĿ ľ +ï¥ ¥ +ï¬ ® +ðĿĹ Ŀ +ðĿĹ ¿ +ðŁİ ¾ +ðŁĹ Ŀ +ðŁ¦ Į +Æ ħ +Ç ª +Ò Ĺ +Ü Ľ +ß ł +à¡ ij +áī £ +áĬ Ń +á¹ ¡ +âŀ ¼ +âŀ ¾ +â´ ± +ãī ¡ +ê³ ¯ +ë½ Ī +ìĤ ĺ +ìī ij +ì «ĺ +íĮ ĥ +íĻ ° +ï¤ Ĺ +ðŁĮ ¬ +ðŁĮ ° +ðŁį ¤ +Ä » +Å ĩ +Æ ¨ +É ķ +Ò ¢ +Ò º +Ö į +× ± +Ú ± +Ú ½ +Û IJ +ठĽ +à· Ģ +๠ļ +ຠ« +á´ ¹ +á ½Ķ +á¾ ³ +âĤ Ĵ +âĨ ´ +âĩ Ŀ +âī ħ +â Į¨ +âĵ ĵ +âĸ ¢ +âļ ¬ +âŀ Ń +â² Ĵ +ãİ ¿ +ê¿ ´ +ëĪ ± +ëį ¬ +ëİ IJ +ëIJ « +ëĶ « +ë± ģ +ìĥ ¥ +íĮ ¼ +ïŃ ĵ +ï® ¥ +ï² ° +ðĿIJ ĩ +ðĿIJ ij +ðĿij Į +ðĿĵ ª +ðĿķ ļ +ðĿĺ ª +ðĿĺ ¼ +ðĿļ Ľ +ðŁĩ ¶ +ðŁĮ Ħ +ðŁĮ ķ +ðŁĮ ¤ +ðŁĮ § +ðŁį ¬ +ðŁİ ĭ +ðŁİ » +ðŁı ¨ +ðŁIJ ĩ +ðŁij ĵ +ðŁĵ IJ +ðŁĵ Ļ +ðŁĶ ¼ +ðŁķ Ĵ +ðŁĸ ı +ðŁĸ ¥ +ðŁ¤ ¬ +ðŁ¥ Ĭ +ðŁ¥ Ĵ +ß Į +ຠĦ +á¼ µ +âķ ¡ +â² ¤ +â´ ¼ +âµ ¢ +ãĪ ¯ +ëĵ ¸ +ëŁ ĩ +ëº į +ðĿĻ § +ðŁį Ī +ðŁĶ ¬ +ðŁĸ Ĭ +ðŁ¤ ¾ +Ë ¡ +Ü © +âĮ ¡ +âŃ ij +â² ¦ +ë© ī +ì¼ Ń +ï¿ ¤ +ðĿĴ İ +ðĿĹ ¥ +ðŁIJ µ +ðŁķ ¶ +ðŁķ ¸ +ðŁ¤ ľ +Õ ª +áĪ ĭ +ðŁ¥ µ +ï° ģ +áµ IJ +âķ ĵ +áĢ ĸ +âĭ Ī +É ŀ +âŀ ® +ॠ° +ãĨ ģ +ðŁĴ ± +ðŁı Ń +áĨ ¨ +ðŁį ļ +ðŁ¦ IJ +á´ » +âĺ Į +à´ ķ +Õ ± +áħ ® +ðĿIJ Į +Å ¦ +ຠķ +âľ Ļ +Ë ³ +Ô µ +âķ Ĵ +ðĿĹ Ĺ +ðĿĹ ł +Ú ļ +ঠ§ +âĨ Ŀ +âĻ ī +ãĮ » +ì¹ Ĭ +ðĿĹ º +ðŁ§ ĺ +ì³ £ +ï¬ Ŀ +ðŁij º +Ç Ł +Î Ī +Î « +Ñ ¥ +Ô ² +Õ ¨ +Ü ¦ +ঠĨ +ঠ¥ +áIJ ¢ +á¼ ģ +á¼ ĺ +á¼ ¦ +âĵ Ŀ +ãĪ ° +ãİ Ĺ +ê² ¡ +ë¨ Ģ +ì£ Ķ +ì´ ¤ +ìµ Ŀ +ï§ ´ +ïŃ Ĭ +ï² Ł +ðĿIJ · +ðĿij ĭ +ðĿĵ ī +ðĿĺ µ +ðŁĴ · +ðŁĽ © +ðŁ§ ¹ +Å Ķ +Ê ŀ +Ë ¥ +Î Į +Ñ © +Ó IJ +Ó ł +Ú ij +Ú Ĵ +ß ¨ +àª Ī +áIJ ĥ +á¹ ¯ +âĤ ĭ +âĤ µ +âĦ ħ +âĦ ł +âĪ £ +âī º +âī » +âĬ Ľ +âĮ IJ +âİ ĵ +âĺ ¸ +âĻ Ĵ +âļ Ĵ +âľ ĩ +âľ ł +â´ · +âµ ĸ +ãĦ ¸ +ãī ¢ +ãī ° +êĩ ´ +ê´ ¸ +êº ł +ëĤ ı +ëĤ ¢ +ëIJ Ģ +ëº ´ +ìĥ ľ +ìį ħ +ì¤ « +ì± ¦ +ìº ij +ì¼ ģ +ì¿ ³ +íĤ ģ +íħ ¡ +íĴ Ĥ +íĴ ī +íľ Ħ +ïŃ ª +ï® ¬ +ï¯ ¦ +ï± ª +ï² ı +ï ´Ģ +ï» Ĩ +ï¿ ¦ +ðĿij Ĺ +ðĿĸ Ļ +ðŁĮ ¡ +ðŁį Ŀ +ðŁį § +ðŁİ « +ðŁı ĺ +ðŁı ª +ðŁIJ ĭ +ðŁIJ Ľ +ðŁIJ º +ðŁij ĸ +ðŁij ŀ +ðŁij · +ðŁĵ Ģ +ðŁ ĶĦ +ðŁĶ Į +ðŁķ Ļ +ðŁĻ į +ðŁĻ İ +ðŁ¦ į +Ç ° +É Ł +Ê Ĩ +Ô ¼ +Ú ľ +ঠ¡ +ঠ¶ +áĴ ĥ +á¼ © +âĵ ķ +â² Ī +ê° ° +ê¹ ł +êº ħ +ëĦ ¹ +ë¯ ĵ +íIJ Ī +ï§ ¶ +ï® ij +ï² ¨ +ðĿĴ ī +ðĿĴ Ķ +ðĿĹ ¨ +ðĿĻ ŀ +ðĿļ Ĵ +ðĿļ ķ +ðŁIJ İ +ðŁ¤ ķ +ðŁ§ Ķ +Ï ° +Ô Ŀ +âĮ Ĭ +âĴ ¾ +ãī £ +ïŃ © +ðĿļ ŀ +Ê ij +ঠ¦ +áĦ ĩ +âī ĥ +â² Ģ +ìŁ İ +ðĿij ¶ +ðĿĵ ² +ðŁ İ· +ðŁļ ¹ +ຠģ +áł ł +ãĦ ļ +ðŁIJ ¿ +ἠļ +âķ ³ +ðŁIJ Ń +âĴ ¹ +ðĿĸ ļ +âĻ ĸ +ãĪ ² +âĨ ¾ +áĦ Ĩ +âķ Ľ +ðŁ¤ į +â½ ¥ +ðŁ Į¨ +âĪ ® +ãĮ ĺ +ãį ij +ï¹ Ģ +âĵ Ĺ +âĬ Ħ +ðŁı ¹ +Ë Ĵ +ðŁ¤ ± +ãı ľ +ðŁİ Į +ï¥ Ń +ঠ£ +ðŁİ ¹ +ãĬ Ł +à´ ° +ðĿIJ Ķ +à´ ¨ +འļ +âľ º +Õ · +ðŁij ³ +ঠľ +âĺ ĭ +âĻ Ĭ +ãĢ Ľ +È ĭ +à® ° +áĥ ¨ +âĦ ķ +íij Ģ +ðĿĵ ĥ +ðŁ¦ Ķ +Ä ¿ +Å Ģ +Æ ³ +É ļ +Ö ĥ +Ü £ +ß Ł +à¦ Ń +à§ ¡ +à¶ » +ຠ£ +འĩ +Ḡ¨ +á½ Ī +â½ ¬ +ê¡ Ķ +ì³ Ħ +ï¨ ī +ðĿIJ ¡ +ðĿĺ ¢ +ðŁį ¿ +ðŁİ Ł +ðŁı ī +ðŁĶ IJ +ðŁļ ħ +ðŁ¤ ½ +Æ į +Ç « +Ç ½ +È ļ +Î ī +Ó ¤ +Ó ª +Õ Ĭ +Ù ¼ +Ú ´ +ß Ŀ +à¶ ľ +á¼ ķ +á¿ ¥ +âİ ŀ +ãĢ ļ +ãī ¤ +ê³ ¸ +ê· ģ +ëĵ Ħ +ëĵ ķ +ì¨ Ķ +ì± ¨ +ðĿIJ ¾ +ðĿij » +ðĿĶ ¼ +ðĿķ Ŀ +ðĿĺ Ń +ðŁĨ Ļ +ðŁĵ ¤ +ðŁĶ Ł +ðŁĹ ¼ +Ä ľ +Æ ģ +Æ ¿ +Ç ³ +Ç · +É ĥ +É ł +Ê ī +Ê § +Ë ² +Ï ´ +Õ ģ +Õ ŀ +Ö ĩ +Û Ĥ +Û ĵ +ß Ĺ +ß ¦ +ঠ¹ +à® ³ +à´ ¸ +à» Ĥ +áĪ Ŀ +áĪ ª +áĭ µ +áIJ Ĭ +áĴ ª +áļ ĸ +áŀ Ľ +á´ ¢ +áµ ı +áµ Ń +á¶ « +Ḡı +ẠĴ +á¼ ¥ +á½ ķ +á½ ¼ +âĤ Ĭ +âĦ Ĥ +âĦ © +âĩ ī +âī £ +âĮ ł +âİ Ł +âı ® +âķ ĺ +âĹ ĸ +âĺ © +âĻ ij +âĻ ² +âļ Ľ +ãĦ Ł +ãī ± +ãİ ļ +ê¡ ķ +êª ĸ +ê° ¹ +ê² Ĩ +êµ Ħ +ëĩ ¬ +ëĭ ¯ +ëı ł +ëĴ ¬ +ëĸ Ī +ëĸ ½ +ëĺ Ķ +ëŀ ¸ +ë¸ ħ +ë» ł +ë¿ Ł +ìĤ µ +ìĬ ī +ìľ ° +ìł ĭ +ìł Ķ +ì¥ ¡ +ìŃ Ŀ +ì¼ ¬ +íĪ ĩ +íī ľ +íį Ħ +íĽ ¾ +íĿ £ +ï¤ © +ï¤ ¯ +ï¦ ľ +ï¦ § +ï§ ľ +ï¨ Ī +ï¬ ª +ï ¬´ +ïŃ ½ +ï® ī +ï¯ ŀ +ï° Ĵ +ï± ĩ +ï¿ Ħ +ðĿIJ ħ +ðĿij Ħ +ðĿij º +ðĿĴ Ĺ +ðĿĵ ® +ðĿķ Ľ +ðĿķ ŀ +ðĿĸ ij +ðĿĺ ģ +ðĿĺ Ĩ +ðĿĺ ¶ +ðĿĻ ¢ +ðĿļ ľ +ðŁĮ ĥ +ðŁĮ ¦ +ðŁį Ł +ðŁİ İ +ðŁı Ļ +ðŁIJ © +ðŁIJ « +ðŁIJ ´ +ðŁij Ķ +ðŁĵ ī +ðŁĵ Ľ +ðŁĶ ī +ðŁĸ ¼ +ðŁĹ ĥ +ðŁĹ ¯ +ðŁļ ĩ +ðŁļ IJ +ðŁļ µ +ðŁ¤ ¶ +ðŁ¥ ĭ +ðŁ¥ ĵ +ðŁ¥ ® +ðŁ¦ İ +ðŁ¦ ł +ðŁ§ Ĵ +ðŁ§ ¨ +Æ IJ +Ç į +Ó Ģ +Ô Ľ +ಠ° +à´ Ļ +áĢ Ĵ +ê² Ŀ +ê¹ ¹ +ë© ¥ +ìĸ Ķ +ï¤ ģ +ï¤ ı +ï¦ ī +ï¦ ĵ +ï§ ī +ï² Ŀ +ðĿĹ ŀ +ðĿĹ ± +ðŁĮ ĭ +ðŁį ¶ +ঠļ +ìķ ľ +ðĿIJ ¯ +ðĿļ Ŀ +à° ¨ +འĺ +འł +á¡ ¥ +á¾ ° +âģ į +âĶ ° +⬠ľ +ðĿIJ ł +ðĿij ¯ +ðĿĹ Ľ +ðĿĵ » +ðĿĸ Ī +âŀ » +áŀ ł +â¡ ± +â» ij +ðŁ§ µ +ï¦ ¢ +ðŁij ĺ +ãĤ Ķ +â¼ Ł +ãĬ ¤ +ï¦ Ŀ +ãĮ ¦ +âĢ ¸ +ðŁĶ Ļ +ã ¹ +ã¹ ¦ +ï¹ ħ +ï© Į +ãī ¨ +ï¸ ½ +âį ¥ +ðŁļ ī +ðŁ¥ ľ +âĵ ľ +â» Ŀ +ï¨ ľ +ðŁĴ Ĵ +áĦ ij +â¾ ŀ +ï¨ ģ +à´ ª +áĦ İ +âŀ ´ +ঠ· +áħ ¬ +áŀ § +âĨ ¢ +âķ ¦ +âľ ij +Ë ¬ +Õ IJ +à¼ Ķ +Ê ¤ +Ë ¨ +ठŀ +à» ĥ +༠ļ +âĵ ¥ +âķ ľ +ðŁIJ ĸ +á¼ Ļ +á¼ ¤ +ìĨ ° +È Ĥ +Ê ± +à® ļ +áĥ § +á´ ĭ +á´ ® +âĿ ¡ +âŀ · +ëĿ ¡ +ï§ ¢ +ï¯ ¡ +ðĿķ ķ +ðŁħ ° +ðŁ¦ ¸ +Ç ¸ +Ó ŀ +Ô ¶ +Ö Ĩ +Ú ģ +Û ĭ +áİ ¥ +á¾ ¿ +âĶ Ń +âĶ ® +êĢ Ģ +ê± ĺ +ëIJ Ń +ë½ Ħ +ìĶ IJ +ì¸ Į +íģ ł +íĻ ± +ï¥ ī +ï¨ ĸ +ðĿij ´ +ðĿĸ Ĵ +ðĿĺ ¨ +ðĿ ļĮ +ðŁIJ ¡ +ðŁij ¢ +ðŁĵ Ķ +Å ħ +Æ İ +È © +Ò ª +Ô ĥ +áĥ « +Ḡĩ +âĽ Ł +ê» Ń +ë¨ Ħ +ìŁ Ģ +ì¤ ´ +íļ IJ +ï¤ ³ +ðŁŁ ¢ +Æ § +È ¼ +Ê Ŀ +Ë Ħ +Ë ħ +Ë į +Ë § +Ò ¥ +Õ Ķ +Ø ı +Ø ¼ +ß IJ +ß ľ +ठĵ +à¦ Ļ +à® ĵ +à¶ ´ +༠į +༠Ĵ +འ£ +áĢ Ĥ +áĢ Ĭ +áĦ Ħ +á Īĺ +áĭ Ĭ +áĮ į +áij ĭ +áŀ Ĥ +áł ¢ +á¡ Ŀ +á´ ¦ +áµ į +áµ ¨ +Ḡ¡ +Ḡ¯ +á¼ £ +âģ Ĥ +âĦ ĺ +âĦ ľ +âĦ ³ +âĦ µ +âĨ ¦ +âĩ Ĩ +âĪ · +âĬ ļ +âĮ « +âĮ ¯ +âİ Ľ +âİ ľ +âİ ¤ +âİ ¦ +âİ ® +âij ī +âĶ ī +âķ Ļ +âĸ Ĥ +âĹ Ń +âĺ Ĭ +âĺ į +âĺ Ĵ +âļ Ĩ +⼠§ +⼠² +âŀ ĺ +⥠Ħ +â´ ³ +â´ ½ +âµ Ī +ãī ¯ +ãİ ij +ã§ ¬ +êĻ ¬ +ê§ ģ +ê³ ¬ +ê´ ŀ +ê» ľ +ëħ ĵ +ëĭ ¼ +ëį ĸ +ëĸ ± +ëĿ ° +ë¡ ¹ +ë¢ ´ +ë£ Ģ +ë¤ ł +ë¨ ķ +ëŃ ¥ +ìĦ ¶ +ìħ ¤ +ìĮ ķ +ìį ª +ìı © +ìĴ Ģ +ìĶ ¯ +ìĿ Ķ +ìĿ ľ +ìł Ń +ì§ ¦ +ì¨ © +ì² ¬ +ì³ ¥ +ì¼ ¯ +íĢ « +íĢ Ń +íĥ ¸ +íĵ ģ +íķ ¬ +íĹ ¸ +íĽ ķ +íľ Ń +íĿ Ĺ +ï¤ Į +ï¤ ª +ï§ ¿ +ï¬ Ħ +ï¬ ħ +ïŃ ij +ïŃ « +ïŃ º +ï® Ĥ +ï® ¢ +ï® ¨ +ï° İ +ï° ł +ï² £ +ï³ IJ +ï³ Ĵ +ï³ ĺ +ï³ ľ +ï¹ ¼ +ï¿ ¨ +ðĿIJ © +ðĿĴ ļ +ðĿķ Ķ +ðĿķ ¤ +ðĿĸ Į +ðĿĹ £ +ðĿĹ ° +ðĿĹ ´ +ðĿĺ Ĥ +ðĿĺ ¥ +ðĿĺ ® +ðĿĺ ¸ +ðĿĻ Ģ +ðĿĽ ¾ +ðĿľ ı +ðŁĮ ģ +ðŁĮ ľ +ðŁĮ ¥ +ðŁĮ ¯ +ðŁį IJ +ðŁİ Ĵ +ðŁı Ķ +ðŁı ķ +ðŁı ® +ðŁIJ Ĥ +ðŁIJ ī +ðŁIJ ¹ +ðŁĶ ķ +ðŁĶ ļ +ðŁķ ij +ðŁķ £ +ðŁĹ ŀ +ðŁĹ ¡ +ðŁĹ ¿ +ðŁļ Ĩ +ðŁļ Ĭ +ðŁļ ĵ +ðŁļ ķ +ðŁļ ¾ +ðŁĽ ģ +ðŁĽ İ +ðŁĽ ı +ðŁ¤ ´ +ðŁ¥ ķ +ðŁ¥ ĸ +ðŁ¥ ł +ðŁ¥ ¥ +ðŁ¦ Ĩ +ðŁ¦ ī +ðŁ¦ ļ +ðŁ§ ij +ðŁ§ ¥ +ðŁ§ ¿ +Å ° +Æ º +É § +ઠĩ +à® £ +áĪ Ī +áĬ ¤ +áĭ ® +áĮ Ī +áĮ µ +ᥠ² +âĵ Ł +êĻ ³ +ê° Ĭ +ëķ ģ +ëķ ¨ +ìĬ ģ +ï¦ µ +ï¬ ² +ðĿĸ į +ðĿĺ Į +ðĿĺ ³ +ðĿĻ © +ðŁį Ļ +ðŁĸ ĸ +áī ³ +áĭ ¨ +áĸ ĩ +áŀ Į +á¹ § +âķ ª +âŀ ļ +â² ĺ +ê ķ +êķ ¥ +ï¤ · +ï® £ +ï¯ ł +ðĿĴ ĸ +ðĿķ ĺ +ðĿĸ ĩ +ðĿĹ Ł +ðĿĹ ª +ðĿĹ ¯ +ðĿĻ ł +ðŁĵ ı +à¦ Ĺ +âĴ » +â² ł +ðĿĵ µ +Ê £ +à° ľ +áĬ ¢ +áŀ IJ +Ḡ· +âĦ Ľ +âĩ Ģ +âĩ Ĭ +êĴ ¦ +ê¦ ł +ï® ¤ +ðŁį Ľ +ðŁ¤ Ľ +ᨠ¾ +âŀ º +áķ ¯ +ἠı +âĩ Ĥ +âĶ ¹ +âĻ Ĺ +ðŁĸ ¨ +ê¦ ı +ઠ° +áļ ¨ +ðŁ¤ ¥ +ðŁ§ ¢ +ãIJ Ĥ +ãĦ ¥ +ðŁĸ Į +â¼ Ĵ +ãĬ § +âį © +ðŁ¦ ij +âĶ · +ï© IJ +ï© ¡ +ðĵ Ī +ðĵĪ Ĵ +â» Ħ +ï¨ Ĵ +âĦ ª +Ò § +Ú Į +âĢ ¶ +⺠ł +â» ģ +âĨ ¸ +áĦ IJ +ãħ IJ +à» Ħ +áĹ ª +âĨ ¼ +âĩ ĭ +âĩ ĺ +âĮ ij +âĸ © +ðĿIJ Ĺ +Ä Ĭ +ঠī +ìī ł +É ¤ +ß į +ß ı +áµ Ĺ +âĤ ¥ +âĵ ī +âĶ ł +âĶ ¨ +âķ Ħ +ä ¤ +ä¤ Ģ +ê» ¸ +ï® ģ +ðĵ Ĥ +ðĵĤ ĥ +ðŁ¦ ķ +Æ Ľ +ঠĩ +ãı ĺ +ï® ¼ +Ú ĵ +Ú Ŀ +ঠĵ +à¶ ¯ +á´ ħ +á½ Ļ +âģ ¼ +âĸ İ +â¼ © +ä Ķ +äĶ Ģ +ë» ¡ +ìĽ ½ +íģ Ħ +ï¥ ¼ +ï± ī +ï¹ » +ðĿĸ ĭ +ðĿĻ Ī +ðĿĻ ª +ðĿ ϶ +ðŁIJ Ħ +ðŁIJ Ĩ +áİ ¢ +ḠĮ +âĿ ´ +ðŁı ¸ +È Ŀ +É ¸ +Î ħ +Ï ľ +Ó ¢ +Õ ¹ +à´ ħ +àº Ī +áĭ ° +áij İ +áł µ +á¡ ł +á´ ī +Ḡµ +á¿ ´ +âĵ £ +âĶ ¶ +â½ ¯ +ê² ¥ +ê¿ ĺ +ëģ İ +ëİ Ī +ëĶ ¯ +ë² ° +ìĺ ¯ +ìĽ ¸ +ìŀ Ĺ +ì§ ĺ +ì¬ ¬ +ì· ¬ +íģ ħ +íĵ Ķ +íĽ Ŀ +ï¤ ® +ï¤ ¹ +ï¥ ² +ï¯ ĸ +ðĿĵ ħ +ðĿĻ Ħ +ðŁĵ ¶ +ðŁĹ Ĵ +ðŁ¥ Ķ +ðŁ¥ Ń +Å ® +Å ´ +Æ ī +Æ « +Ç ģ +Ç £ +Ç º +Ç ¼ +È į +È ¯ +É ľ +Ê ¬ +Ë ģ +Ë ¤ +Ë µ +Ï Ľ +Ò ¤ +Ò ¬ +Ó ı +Ó Ľ +Ó ¡ +Ó ³ +Ô Į +Ô ¬ +Õ ³ +Ù » +Ú ī +Ú § +Ü ľ +ß ª +ठĿ +ঠĽ +ਠĨ +ઠķ +ઠ¡ +à® İ +à° ¬ +ൠ» +ൠ¼ +à¶ ł +à¶ Ń +à¶ ¶ +à· Ĩ +༠½ +áĢ ļ +áħ ¢ +áĨ ¸ +áĪ Ģ +áĪ ķ +áĪ ° +áī ¡ +áī ¤ +áĬ ¦ +áĬ « +áĭ ĭ +áĭ į +áİ ¯ +áij Ń +áķ Ĺ +ᣠĽ +ᥠĴ +á© ī +áŃ º +á´ ¡ +áµ ĺ +áµ Ľ +á¶ ł +Ḡģ +Ḡĭ +á¹ Ļ +á¹ Ŀ +á¹ ¦ +Ạħ +á¼ Ĥ +á½ ĥ +á½ į +á½ § +á¾ · +âĢ µ +âĤ İ +âĦ Ŀ +âħ Ģ +âĨ ŀ +âĨ § +âĩ ħ +âĪ ĥ +âī ı +âī ½ +âĬ ŀ +âĬ ¡ +âĬ § +â Ĭ¶ +âĭ Ħ +âİ Ĵ +âİ ¡ +âİ £ +âİ ª +âı İ +âĵ ĥ +âĵ ĸ +âĵ ¨ +âķ ĭ +âķ ĸ +âķ ¢ +âķ ² +âĸ Ĩ +âĸ Ĭ +âĸ į +âĸ ® +âĺ ¡ +âĺ ¦ +âĺ ± +âĺ ¿ +âĻ ĺ +âĻ Ŀ +âļ ° +⼠ij +âŀ ª +⤠Ŀ +⤠¢ +⤠· +â§ « +â¨ Ń +⨠¯ +â± £ +â² İ +âµ Ľ +ãħ Ķ +ãĪ ı +ãī ² +ãī ³ +ãĬ ij +ãĭ Ľ +ãİ IJ +ê² ¤ +ê· ¿ +ê¹ ŀ +ê» ¨ +ê¼ į +ê¿ ¸ +ëĥ ¬ +ëĩ IJ +ëĭ ł +ëį ¯ +ëĹ Į +ëĹ ij +ë¥ Ģ +ëª ĥ +ëª ¯ +ë± ¡ +ë³ ĵ +ë³ ½ +ë µľ +ìĤ ³ +ìħ ¥ +ìĩ ½ +ìı ¨ +ìı ¸ +ìķ į +ìĸ ĸ +ìŁ ¨ +ì¢ ĥ +ì¢ į +ì¥ ij +ì§ ¼ +ì© ĥ +ì® ľ +ì® ¸ +ì³ ij +ì´ ¥ +ì¾ ĥ +íħ ¦ +íĪ ¿ +íĵ ½ +íķ ³ +íĸ ı +íĹ ł +íĿ « +ï¤ ĵ +ï¤ ĺ +ï¥ İ +ï¥ ¶ +ï¦ ħ +ï¦ ½ +ï§ ĩ +ï¬ Ĩ +ï¬ ³ +ï® ĩ +ï® Ī +ï® Ŀ +ï® © +ï® ± +ï¯ ĺ +ï¯ Ļ +ï¯ ¢ +ï¯ £ +ï¯ ¤ +ï¯ ¥ +ï± Ĥ +ï² Ĩ +ï² ª +ï´ ¼ +ïº ī +ïº Ĭ +ïº ¥ +ðĿij ¨ +ðĿij © +ðĿij ² +ðĿ ĴĮ +ðĿĴ ª +ðĿĴ ® +ðĿĵ Ĥ +ðĿĵ Ī +ðĿĵ ¯ +ðĿĶ ¨ +ðĿķ Ģ +ðĿķ Ĩ +ðĿķ ¦ +ðĿķ § +ðĿķ « +ðĿķ · +ðĿĹ µ +ðĿĹ ¸ +ðĿĺ Ħ +ðĿĺ Ļ +ðĿĺ ł +ðĿĺ ¬ +ðĿĻ į +ðĿĻ ij +ðĿĻ ¡ +ðĿ ύ +ðĿĻ · +ðĿļ į +ðĿĽ ¿ +ðŁ ĥ +ðŁĥ ı +ðŁħ ĺ +ðŁ ī +ðŁī ij +ðŁİ ¡ +ðŁİ ª +ðŁİ ± +ðŁİ ³ +ðŁİ º +ðŁı İ +ðŁı Ĺ +ðŁı ļ +ðŁı ŀ +ðŁı ¦ +ðŁı § +ðŁIJ ģ +ðŁIJ ħ +ðŁIJ ĵ +ðŁĴ Ĥ +ðŁĵ ij +ðŁĵ ĵ +ðŁĵ ¨ +ðŁĵ « +ðŁĶ ĭ +ðŁĶ Ń +ðŁĶ ¯ +ðŁķ Ĺ +ðŁļ Ĥ +ðŁļ ¢ +ðŁļ ¦ +ðŁļ ¬ +ðŁĽ ĭ +ðŁĽ Į +ðŁĽ ¬ +ðŁĽ ¶ +ðŁŁ ¡ +ðŁ¥ ĺ +ðŁ¥ Ł +ðŁ¥ ¦ +ðŁ¦ ĩ +ðŁ¦ Ī +ðŁ§ Ĭ +ðŁ§ Ĺ +ðŁ§ ¤ +Ê · +Ë ¹ +á¹ ļ +á½ ¥ +âĦ Ł +ê² ¯ +ê» « +ë° · +ìĥ Ĩ +ìĽ Ŀ +ì¨ ī +ì« ı +ï¯ ķ +ðĿľ ĭ +É ² +Ò Ń +Ó Ī +འĽ +áĭ ĵ +áĻ Ń +áł © +á¹ ® +âĦ Ĵ +âĨ » +âµ ĥ +ëĢ ¨ +ëł § +ìī ¥ +ìĮ ľ +ìĹ ¶ +ì¨ Ī +ìª ¾ +íı ½ +íļ Ķ +íĽ µ +ï¤ ¸ +ï¦ IJ +ï§ Ĺ +ï§ ļ +ï¬ ¯ +ðĿIJ Ĭ +ðĿķ Ĺ +ðĿĹ ļ +ðĿļ ĸ +ðŁħ ´ +È ĥ +É Ŀ +Ï ± +Ó Ĺ +ठ¢ +áħ ł +áī ¦ +áij Į +áĴ ¼ +áŀ ¡ +áł ¨ +áł Ń +ᨠħ +á¨ Ķ +á´ ĺ +á¶ ¦ +á¸ İ +á¼ ħ +á¼ ¹ +âĨ ¯ +âĵ İ +ãı Į +ê ī +êī Ĥ +ëĨ § +ëĿ ± +ì¢ ¡ +íĪ ½ +ï¤ ĩ +ï¤ Ľ +ðĿIJ ķ +ðĿĵ ¸ +ðĿĵ ¼ +ðĿĹ ķ +ðĿĺ Ī +ðŁı £ +ðŁı ¤ +ðŁĹ Ħ +Ñ · +Ò ł +áµ ĸ +á¼ ¨ +ë¬ Ħ +ï° ´ +âĪ ½ +Õ Ń +Ú ¹ +à¥ Ł +áĢ Ĩ +áŀ Ĵ +ãĢ ¶ +ê¦ « +ï¸ ĵ +ðĿIJ Ľ +ðĿĺ Ĺ +ðŁı ľ +ì« Ń +ðŁ§ ŀ +འĤ +âĨ ¿ +âĩ ı +âĵ ģ +âĶ § +âķ ģ +âķ ¤ +ê¦ Ĺ +ê¦ ¤ +ðŁı Ī +áŀ ķ +Ô ½ +àª Ĺ +ଠĨ +âķ ķ +ï½ ł +â¼ ¦ +â¼ ¯ +â¾ · +âĶ ĸ +ଠĵ +âĺ Ĺ +âį ĭ +ï¨ Ŀ +â¼ ¥ +ï¦ ª +âĦ Ĭ +ãĢ ´ +âį ¢ +ð¡ Ī +ð¡Ī ½ +ï© ¨ +ãĢ » +ãı ĥ +ï¦ ¡ +ï¨ ĺ +ðŁIJ ĥ +ðŁĨ ĸ +ðŁĹ ¾ +ãĦ ĩ +Þ ĭ +â¼ ¼ +ï¨ Ń +Þ Ģ +Þ Ħ +Þ Ī +Þ IJ +âĮ Ħ +â» ĺ +ãŁ ¢ +á ħ§ +ðIJĮ ¿ +Ë » +à² Ĺ +áĢ ĩ +áŀ Ĭ +âķ ĩ +ãĩ ¼ +ãİ ° +Õ Ĵ +Ü Ī +ß ¥ +à¿ IJ +áĢ Ł +âĨ ¥ +âķ Į +â½ Ģ +â½ ° +â¾ Ĭ +ä Ħ +äĦ Ģ +ðĵ IJ +ðĵIJ į +ðŁİ ¦ +âĤ ¯ +âĬ ĺ +âĦ į +Ê µ +Ñ ¶ +Ú ĥ +à¦ Ķ +à´ ¦ +áİ ¶ +áĵ ķ +á¹ ¨ +âĤ ł +âĩ ° +âĹ Ĵ +â¿ Ĭ +ê· ± +ì¹ ķ +íĪ © +ïŃ Ģ +ðĿĴ ¸ +ðĿĵ Ĭ +ðĿĺ © +Ç ¦ +É « +áĬ ¨ +È ¹ +Ê ¯ +Î ª +Ú Ģ +áĮ ¸ +áİ » +áı ķ +áı ´ +á² Ĥ +á½ ¨ +âı Ŀ +âĺ Ļ +ëĥ ¨ +ëĦ ¼ +ëĪ Ļ +ë£ ħ +ìĶ ¼ +ìķ Ŀ +ìļ ¬ +ìľ ± +ï¥ Ĥ +ï¦ ¹ +ï¬ ¹ +ïŃ ģ +ï³ Ī +ðĿĶ ħ +ðĿĺ ¤ +ðĿĻ ı +ðĿĻ Ļ +ðŁķ ī +ðŁ§ Ļ +Ḡij +ê´ ¼ +ëģ į +ëĹ ´ +ëĿ ³ +ë° ŀ +ë° ¢ +ëµ ĺ +ìĤ Ķ +ìĦ Ħ +ì¼ ļ +íĢ ł +íĬ ± +íĮ ĸ +ï¤ ij +ï¦ ´ +ï¦ ¸ +ï´ į +ðĿĺ · +Ä ¬ +Å ¬ +Æ Ģ +Æ ĭ +Æ ľ +Ç ij +Ç ĺ +Ç ŀ +Ç ¥ +Ç ® +É ° +É ¶ +É · +É ½ +Ê Ī +Ê IJ +Ë İ +Ë Ł +Ë ¦ +Ë ¯ +Ï IJ +Ï ĵ +Ï ¢ +Ï ¤ +Ï ª +Ï Ń +Ï ® +Ï » +Ñ ł +Ñ Ń +Ò ¨ +Ó Ŀ +Ô ¡ +Ô · +Õ ī +Õ ĵ +Õ ĸ +Õ ļ +Õ Ŀ +Ö İ +Ø ¿ +Ú ħ +Ú į +Ú Ķ +Û Ĭ +Û ¾ +Ü Ļ +Ý Ĵ +Ý ĺ +ß Ĵ +ß ĸ +ठĬ +ठIJ +ঠı +ঠĸ +à§ Ł +ઠ® +ઠ¹ +à® ħ +à® Ĩ +à° ¡ +à° ° +ಠļ +ಠ® +ಠ¯ +à´ Ł +à´ · +ൠ¾ +à¶ ij +à¶ ŀ +༠¼ +འĵ +áĢ ĵ +áĤ ¦ +áĥ ĸ +áĥ Ń +áĥ ¯ +áħ ¨ +áħ ª +áĨ ° +áĪ ģ +áĪ İ +áĪ ĵ +áĪ ¥ +áĪ ² +áĪ ´ +áĪ » +áī ł +áī ² +áī ¶ +áĬ £ +áĬ ¥ +áĬ ª +áĭ ĺ +áĭ ² +áĭ ¶ +áĮ £ +áį ¡ +áį £ +áİ ¬ +áİ ¾ +áIJ ¡ +áķ ķ +áĸ ± +áĹ IJ +áĹ Ń +áĺ ī +áļ ± +áĽ Ł +áŀ ¥ +áŁ Ķ +áł £ +áł ª +áł ° +áł ´ +ᤠĸ +ᥠ£ +á ® +á® ł +á ¯ +á¯ Ļ +á ° +á° į +á´ Ĭ +á´ ¾ +áµ ģ +áµ İ +áµ ŀ +áµ ¤ +á¶ ħ +á¶ ĺ +á¶ Ł +á¶ ¢ +á¶ ¤ +á¶ ± +á¶ » +Ḡī +Ḡŀ +Ḡº +á¹ ĵ +á¹ Ĺ +á¹ ª +ẠĬ +Ạı +ẠĽ +á¼ ĥ +á¼ Į +á¼ ¿ +á½ Ĥ +á½ ĵ +á½ Ĺ +á½ ¦ +á¾ ± +á¾ ´ +á¿ ĺ +á¿ Ł +á¿ ¸ +âģ ĺ +âĤ ij +âĤ Ľ +âĤ ¿ +âĦ ĩ +âĦ ŀ +âĦ ± +âĩ Ł +âĩ ² +âĪ ¤ +âĪ ¶ +âī Ĥ +âī ¾ +âĬ ¨ +âĬ ³ +âĬ · +âĭ Į +âĭ ĺ +âĮ ķ +âĮ ¥ +âĮ µ +âĮ º +âį £ +âį ² +âį µ +âİ ĩ +âı ĥ +âı IJ +âı ł +âı ¤ +âı ¶ +âı ¸ +âı ¹ +âij Ĥ +âĴ · +âĴ º +âĵ ¡ +âĵ ¤ +âĶ ¾ +âĸ ĺ +âĸ µ +âĹ ª +âĹ · +âĺ ¨ +âĺ « +âĺ ² +âĺ ³ +âĻ Ĩ +âļ ¤ +âļ ¥ +⼠ĵ +⼠´ +⼠¾ +âŀ « +âŀ ¿ +⣠· +⤠ij +⤠« +⤠¶ +⤠½ +â§ ª +â¨ Ģ +â ©½ +⬠¡ +⬠¢ +⬠¤ +â² ĸ +â² ª +âµ Ģ +⸠® +⸠½ +ãĢ ł +ãĢ · +ãĦ Į +ãĦ ĺ +ãħ ij +ãĪ İ +ãĪ IJ +ãĬ ľ +ãĮ ĵ +ãĮ ł +ãİ Ł +ãİ ¤ +ãİ § +㬠® +ä Ī +äĪ Ģ +ä ° +ä° Ģ +ê ħ +êħ ī +êĩ Ĺ +ê Ī +êĪ į +ê§ Ĥ +ê§ Ĭ +êª Ģ +ê² Ī +ê² į +ê³ Ģ +êµ ł +ê½ IJ +ê¾ Ī +ê¿ ± +ëĥ ı +ëĦ ij +ëħ ¤ +ëĩ ¸ +ëĪ ¼ +ëī ħ +ëĬ £ +ëĭ º +ëį ŀ +ëIJ Į +ëķ ¸ +ëĺ ł +ëĻ ĩ +ëĻ Ī +ëľ ½ +ëŀ Ķ +ëł ľ +ë£ IJ +ë§ Ģ +ë§ Ĭ +ëª Ģ +ë¬ Ń +ë¯ ¾ +ë³ ľ +ë´ Ĭ +ëµ ī +ë· ľ +ë¸ Ģ +ë¹ ĭ +ìģ Ħ +ìĤ £ +ìĤ » +ìĦ µ +ìħ Ĵ +ìī Ī +ìī Ķ +ìĬ Į +ìĬ Ļ +ìIJ ´ +ìĵ º +ìķ ļ +ìķ º +ìĸ ľ +ìĹ ª +ìĺ ľ +ìĻ ¤ +ìļ Ľ +ìļ º +ìĿ ħ +ìĿ ı +ìĿ Ń +ìĿ ¶ +ìł Ľ +ì¡ Ī +ì¢ ī +ì¢ Ķ +ì© ł +ìŃ Į +ì¯ © +ì´ £ +ì¸ ķ +ì¹ Ł +ì¾ ¡ +ì¿ Ļ +íģ ĩ +íģ ī +íĩ Ģ +íĪ ¶ +íĸ ij +íĸ ¤ +íĹ ħ +íľ ı +íĿ Ŀ +ï¤ Ĵ +ï¤ ķ +ï¤ ¬ +ï¥ ħ +ï¥ ĩ +ï¥ ı +ï¥ ļ +ï¥ Ł +ï¦ Ħ +ï¦ Ī +ï¦ ¨ +ï¦ © +ï¦ ² +ï§ ģ +ï§ ĥ +ï§ Ķ +ï§ ł +ï§ £ +ï§ ® +ï ŃIJ +ïŃ ĸ +ïŃ ¦ +ïŃ ´ +ïŃ µ +ïŃ ¶ +ïŃ ¸ +ï® Į +ï® İ +ï® ŀ +ï® Ł +ï® ¡ +ï® ª +ï¯ Ķ +ï¯ Ĺ +ï¯ ļ +ï¯ Ľ +ï¯ Ŀ +ï¯ Ł +ï¯ § +ï¯ ¨ +ï¯ « +ï¯ ¯ +ï¯ ° +ï¯ ± +ï¯ ² +ï¯ ³ +ï¯ ´ +ï¯ µ +ï¯ ¶ +ï° Ģ +ï± ħ +ï± Ķ +ï± ´ +ï² ģ +ï³ ķ +ï· ½ +ï¸ ķ +ï¸ ± +ï¹ £ +ï¹ ½ +ï» į +ï¾ ± +ðĿIJ Ļ +ðĿIJ ½ +ðĿij ¤ +ðĿij ® +ðĿij µ +ðĿĴ ĥ +ðĿĴ Ħ +ðĿĵ Ń +ðĿĵ · +ðĿĶ ĸ +ðĿĶ ŀ +ðĿĶ ¢ +ðĿĶ ¦ +ðĿĶ ¬ +ðĿķ Ħ +ðĿķ Ĭ +ðĿķ İ +ðĿķ Ļ +ðĿķ ľ +ðĿķ Ń +ðĿķ ³ +ðĿķ ¸ +ðĿķ ¾ +ðĿ ĸī +ðĿĸ ı +ðĿĺ ĩ +ðĿĺ ī +ðĿĺ ĸ +ðĿĺ Ľ +ðĿĺ ŀ +ðĿĺ « +ðĿĺ ¾ +ðĿĻ ĩ +ðĿĻ ī +ðĿĻ ĭ +ðĿĻ İ +ðĿĻ ĺ +ðĿĻ ¥ +ðĿļ ĥ +ðĿļ IJ +ðĿļ Ķ +ðĿľ ĥ +ðŁĦ · +ðŁħ Ŀ +ðŁħ ¾ +ðŁĨ Ĥ +ðŁĨ ĵ +ðŁĮ Ĥ +ðŁĮ Ĩ +ðŁĮ ī +ðŁĮ ij +ðŁĮ ĺ +ðŁĮ © +ðŁĮ « +ðŁį ¢ +ðŁį ¥ +ðŁİ Ľ +ðŁİ ¢ +ðŁİ ´ +ðŁij ¡ +ðŁĴ ¾ +ðŁĵ Ń +ðŁĶ Ī +ðŁĶ ¦ +ðŁĶ ² +ðŁĶ ³ +ðŁķ ĵ +ðŁķ ķ +ðŁķ ĺ +ðŁķ Ł +ðŁķ · +ðŁĹ ³ +ðŁļ Ħ +ðŁļ Ķ +ðŁļ ĸ +ðŁĽ IJ +ðŁĽ ¤ +ðŁĽ ¸ +ðŁ ł +ðŁł ³ +ðŁ¤ ¹ +ðŁ¥ ĥ +ðŁ¥ ¨ +ðŁ¥ ª +ðŁ¥ ¾ +ðŁ¦ ĥ +ðŁ¦ Ĵ +ðŁ¦ Ļ +ðŁ¦ ¶ +ðŁ§ ł +ðŁ§ ª +ðŁ§ Ń +ðŁ§ ² +𣠷 +ð£· Ń +ð¦ ĺ +ð¦ĺ Ĵ +Æ ij +Ç Ļ +È ® +Ø ł +Ú Ħ +Ü Ģ +ß ¢ +áī Ģ +áĬ IJ +áİ ł +Ạŀ +ëĪ ŀ +ëķ Ł +ë£ ģ +ë¤ Ĺ +ìĦ ¥ +ìħ ij +ìĸ IJ +ìĽ Ľ +ì£ ķ +íİ ı +íĽ ĵ +ï¥ º +ï³ Ľ +ï´ « +ðĸ § +ðĸ§ · +ðĿķ ģ +ðŁIJ ª +ðŁĴ Ī +ðŁĵ ł +ðŁķ Ľ +ðŁķ ´ +Ñ Ŀ +Ó Ĭ +ॠ² +ઠª +áĥ ¤ +áį IJ +á¶ ° +á¼ Ŀ +á½ © +âĭ ĭ +âĴ ½ +âĻ ¾ +â ½Ķ +â¾ ¯ +ãĦ Ĵ +ãħ ļ +ëIJ į +ë· ģ +ìĭ Ģ +ìļ Ŀ +ì¥ ° +ìº ´ +íĭ ī +íĿ ½ +ï¦ Ģ +ï¦ ¿ +ï§ ħ +ï§ ĵ +ïŃ ¯ +ï® Ĩ +ðIJ¤ ķ +ðĿIJ Ł +ðĿĴ ħ +ðĿĵ ľ +ðĿĶ ° +ðĿĶ » +ðĿĺ į +ðĿĻ ¯ +ðŁĦ ½ +ðŁħ Ĥ +ðŁħ Ķ +ðŁħ ½ +ðŁĵ ´ +ðŁ§ ĸ +Ó Ĵ +Ḡ² +ëī ¼ +Ç ı +È ĵ +Ê ¸ +Õ Ĥ +Û ħ +ß ¡ +ß £ +à® ¯ +à° Ī +ಠ¸ +ຠ® +༠ķ +áĢ İ +áĨ ¡ +áIJ ĭ +áIJ ķ +áij ¯ +áŀ Ĩ +ᨠķ +á© Ī +âģ ħ +âĨ ļ +âĶ İ +âł © +â² Ĥ +â² Ķ +â² ¨ +ãĬ ļ +íĵ ² +ðĿij Ī +ðĿij ¬ +ðĿij ¹ +ðĿĴ ¾ +ðĿĵ ± +ðĿĵ ½ +ðĿķ ¯ +ðĿķ » +ðĿĺ ½ +ðĿļ Ĩ +ðŁĦ ° +ðŁIJ ¨ +Ò ķ +ಠħ +ï¨ Ĩ +ðĿij ° +ðŁĦ ¸ +Ô İ +Ø į +Ù µ +ಠ¶ +áĢ Ī +áĺ Ĺ +áł ¸ +á¡ ¡ +ᨠ² +á© ģ +á´ · +áµ § +âķ ¨ +âļ ģ +â¾ Ŀ +ãĢ ¼ +ãĦ ı +êĴ « +ê¦ ¥ +ê¦ © +ê¦ ² +ìĺ ¼ +íĵ IJ +ðĵ ĩ +ðĵĩ ¼ +ðĿķ ¿ +ðŁĽ ´ +ë¨ ľ +ಠµ +à´ İ +à¼ Ģ +âĩ ĸ +ãĪ « +âĵ Ģ +áħ ´ +áļ ¾ +ἠŀ +ἠ« +ᥠ´ +âĨ Ľ +âĨ ¶ +âĩ ¤ +âķ Ł +âĺ · +âļ IJ +ðŁ§ ´ +á¹ ³ +âĶ į +âĶ Ĵ +âĶ © +âĶ ¦ +â¾ µ +ઠľ +ઠ¤ +âĩ Ļ +âĶ ± +âķ Ģ +â½ Ĭ +ï½ Ł +ଠ¡ +ðł ® +ðł® · +âķ ĥ +â° Ķ +ãĬ ¦ +ðŁİ IJ +ãĩ ° +â¼ Ŀ +â¾ Ķ +â½ Ĵ +âł Ĵ +ï¨ ¦ +ï© Ĵ +ï¨ ² +ï© ĸ +ðĵı ¸ +ãĮ ĥ +ðĸ ¤ +ðĸ¤ IJ +ï¦ Ń +âĬ ħ +â¾ ³ +ä´ ¥ +ï© ķ +ðŁĮ Ķ +áŀ ĭ +âļ į +â¼ ĭ +ãİ ĺ +ðIJĮ ² +É © +áİ ij +âĨ ® +âĩ ĥ +âļ İ +ãĩ ± +ãĭ © +ãĮ ¶ +êĻ ª +ëİ ¬ +ï¨ IJ +ï¨ Ľ +ï© Ĭ +ï© į +ðĵ ħ +ðĵħ º +Ï ¡ +È ij +É Ĥ +Ô ĵ +ß İ +à´ § +áĢ ī +áĢ ĭ +áĢ ij +áĢ ł +áļ Ļ +ᨠĦ +ᨠ© +ᨠ¹ +á© ĵ +ᬠľ +á´ Ļ +áµ ij +âĤ Ń +âĨ ° +âľ ģ +â½ IJ +ãĭ ¯ +ãĮ ½ +íĨ ¢ +ï¤ ¿ +ðŁ Ĥ +ðŁĤ » +È Ĵ +Í º +Ô ¥ +Õ ij +Ú ¶ +à§ İ +à¶ ® +ຠĸ +ຠľ +ຠ½ +áĥ » +áħ ¯ +áĭ ŀ +áĸ ķ +á ´Ī +á¶ Ĩ +Ḡľ +á¹ ¼ +á¿ ¨ +âĦ ĭ +âĦ Ń +âĪ ± +âĮ ĵ +âĶ ĩ +âĶ ¢ +â± ® +â² Ħ +ãĩ ¾ +ãĪ ¬ +ë¸ ¡ +ìIJ ī +íĻ Ľ +ðĿķ ª +Æ ¹ +Í ² +Ó ģ +Û ¼ +ঠ« +áħ Ł +áī Ĩ +áį Ī +Ạĸ +á½ ī +âĶ ¸ +â½ © +ê ľ +êľ ¥ +êµ ħ +ëĤ Ķ +ëĦ ł +ëĩ Ĺ +ëĻ Ŀ +ìļ ¯ +ìļ · +ìŁ Ľ +ì· IJ +íŁ ¬ +íŁ ® +íŁ ° +ï¦ Ĩ +ï¦ ± +ï² ŀ +ï³ ¤ +ï³ ¥ +ðIJĮ ¸ +ðĿĶ ı +ðĿķ ® +ðĿĺ £ +à¦ Ī +âı ı +ãĦ ĸ +ê² ĩ +ëĸ ĺ +ëľ · +ëŀ Ĵ +ë¡ ĵ +ë¢ ī +ë£ ĥ +ë§ ĭ +ë² ĭ +ìĤ · +ìĪ ķ +ì Į¨ +ìĵ » +ìĸ Ĭ +ìĻ ¬ +ìĿ » +ì¦ ģ +ìµ ¤ +ì· ĥ +íĢ ľ +íħ ī +íį ł +íı ħ +íij ± +íķ ķ +íĸ ł +íĿ ķ +Æ Ļ +Æ ļ +Æ ŀ +Ç ĥ +Ç Ĭ +Ç ľ +Ç ¤ +Ç Ń +Ç ¹ +È Ģ +È ģ +È ħ +È ī +È Ĺ +È Ł +È ¤ +È ¥ +È ¨ +È µ +È º +È » +É Į +É ® +Ê ħ +Ê ¥ +Ê ¨ +Ë ĵ +Ë Ķ +Ë ł +Ë £ +Ë ¸ +Í ´ +Ï Ĺ +Ï ĺ +Ï Ļ +Ï ļ +Ï Ŀ +Ï ¨ +Ï ¬ +Ï ¾ +Ï ¿ +Ñ ª +Ò Ģ +Ò ľ +Ò ¼ +Ò ½ +Ó Ĥ +Ó ħ +Ó ĩ +Ó į +Ó ĸ +Ó Ł +Ó « +Ó ± +Ô Ĩ +Ô ĩ +Ô º +Õ ĭ +Ö ī +Ø Ī +Ø Ĭ +Ø ½ +Ø ¾ +Ù · +Ú Ĥ +Ú Ĭ +Ú ĸ +Ú Ĺ +Ú £ +Ú « +Ú ¸ +Û Ģ +Û į +Û ½ +Ü ī +Ü ¤ +Ý § +Ý ´ +Þ ĥ +Þ ¤ +Þ ¥ +ß ļ +ß Ľ +ß ¤ +àł į +àł ĵ +àł ³ +à¡ ¢ +ॠł +à§ ł +à§ º +ਠĬ +ਠIJ +ਠ® +ਠ¯ +ਠ° +ਠ¸ +ઠĨ +ઠ³ +ઠµ +ઠ½ +ଠĮ +ଠĺ +ଠ½ +à® ĥ +à® ¸ +à° Ĩ +à° ķ +à° ¦ +ಠĨ +ಠĬ +ಠĮ +ಠIJ +ಠĽ +ಠ¤ +ಠ¦ +ಠª +ಠ² +ಠ¹ +à´ Ĩ +à´ ı +à´ Ĺ +à´ « +à´ ¹ +ൠº +ൠ½ +à¶ ħ +à¶ Ĭ +à¶ Ķ +à¶ § +à¶ « +à¶ ° +༠Ħ +༠ħ +༠Ĭ +à½ Ļ +འ¡ +འ§ +à¿ Ģ +à¿ Ļ +áĢ Ŀ +áĢ § +áĢ © +áĢ ¿ +áģ µ +áĤ ģ +áĤ ½ +áĥ Ĥ +áĥ ª +áĦ Ĭ +áĦ ¢ +áħ ¦ +áħ Ń +áĨ ® +áĨ ± +áĨ » +á ĩ +áĩ Ĥ +áĪ ħ +áĪ ī +áĪ Į +áĪ IJ +áĪ Ĵ +áĪ Ļ +áĪ ļ +áĪ ľ +áĪ ŀ +áĪ © +áĪ ³ +áĪ º +áĪ ½ +áī ħ +áī ¢ +áī ± +áī ´ +áĬ ĥ +áĬ į +áĬ ĸ +áĬ ® +áĬ ¸ +áĭ Ľ +áĭ Ŀ +áĭ ³ +áĮ ģ +áĮ ħ +áĮ ¥ +áĮ ¦ +á Į¨ +áį Ĭ +áį į +áį ķ +áį ĸ +áį ¢ +áį ¤ +áİ Ĵ +áİ ª +áı ģ +áı IJ +áı Ł +áIJ Ĥ +áIJ ĸ +áIJ Ŀ +áIJ ŀ +áIJ Ł +áIJ ł +áij ĸ +áĴ ĭ +áĴ į +áĴ ¡ +áĵ « +áĶ ķ +áķ ĭ +áķ ij +áķ Ļ +áķ ļ +áķ Ľ +áķ ¤ +áķ ¦ +áķ ® +áķ ¼ +áĸ ĵ +áĹ Ĺ +áĹ ¢ +áĹ ¯ +áĹ · +áĺ Ħ +áĺ ij +ἠĤ +áĽ Ļ +áŀ į +áł Ĩ +áł ¡ +áł ¦ +áł ® +áł ¯ +áł ² +áł · +á¡ į +á¡ ŀ +á¡ ¤ +á ¡´ +á¡ µ +ᤠĵ +ᥠĸ +ᥠ° +ᨠ¦ +ᨠ§ +ᨠ¨ +ᨠª +ᨠ¬ +ᨠ¯ +ᨠ³ +ᨠµ +á© ĥ +ᬠķ +áŃ £ +á ± +á± ļ +á² ł +á´ ĵ +á´ ¶ +áµ Ĥ +áµ Į +áµ ¥ +áµ ´ +á¶ ĩ +á¸ Ī +Ḡł +Ḡ§ +Ḡ´ +Ḡ¾ +á¹ Ģ +á¹ ĸ +á¹ Ł +á¹ ł +á¹ « +á¹ ± +á¹ · +á¹ ¿ +ẠĦ +Ạį +Ạij +áº Ĺ +á¼ ī +á¼ ĵ +á¼ Ń +á½ ĭ +á½ Ĵ +á½ ł +á½ £ +á¾ Ħ +á¾ ı +á¾ ij +á¾ Ĺ +á¾ ¦ +á¾ § +á¾ ¾ +á¿ Ħ +á¿ ĵ +á¿ ¡ +á¿ ¬ +âģ ļ +âĤ Į +âĦ ģ +âĦ Ķ +âĦ £ +âĦ § +âĦ ¯ +âĦ ° +âĦ ´ +âħ ħ +âĨ ľ +âĨ « +âĨ Ń +âĨ ± +âĨ ¹ +âĨ ½ +âĩ ĩ +âĩ ľ +âĩ µ +âĪ ī +âĪ Ĭ +âĪ ĸ +âĪ ľ +âĪ ¾ +âī Ģ +âī ĭ +âī Į +âī ĵ +âī ľ +âī ´ +âī ¿ +âĬ Ĭ +âĬ ĭ +âĬ Ķ +âĬ ĸ +âĬ £ +âĬ ¦ +âĭ İ +âĭ ª +âĭ ² +âĮ ¦ +âĮ § +âį º +âİ Ī +âİ ¨ +âİ ¬ +âİ ³ +âİ ¼ +âİ ¾ +âı Į +âı ļ +âı « +âı ¯ +âı µ +âĴ ľ +âĴ Ŀ +âĴ « +âĵ Ħ +âĵ Ĭ +âĵ Ļ +âĵ © +âĶ ij +âĶ Ļ +âĶ ļ +âĶ ¥ +âķ ħ +âķ ī +âķ į +âķ ı +âķ ŀ +âĸ ļ +âĸ ¯ +âĹ ĥ +âĹ ļ +âĹ ¬ +âĹ ´ +âĺ Ī +âĺ ¤ +âĺ ¥ +âĺ § +âĺ ¬ +âĻ ģ +âĻ ± +âļ ĥ +âļ Ħ +âļ ħ +âļ ı +âļ ļ +âļ ŀ +âļ Ł +âļ ± +âļ ² +âľ Ģ +âľ Ł +âľ ¢ +âĿ µ +⣠¡ +⣠¦ +⣠§ +⣠³ +⣠¾ +⣠¿ +âł ĩ +⤠Ħ +⤠º +⥠Ĥ +⥠¹ +â§ ī +â§ ¼ +â§ ½ +⨠į +⬠Ĭ +â¬ Ł +âŃ ŀ +â® ŀ +â® ³ +â¯ Ī +⯠ij +â± ł +â± ± +â² Ń +â´ ¹ +âµ ķ +⸠¾ +â º« +â¼ Ĩ +â¼ ł +â½ Ł +â½ ¼ +â¾ Ľ +â¾ § +â¿ ĥ +â¿ » +ãĤ ķ +ãĤ Ł +ãĦ Ľ +ãĦ ¡ +ãĦ ¶ +ãĦ º +ãħ Ĵ +ãħ Ł +ãĨ Ģ +ãĩ » +ãĪ ij +ãĪ Ń +ãĪ ® +ãĪ ³ +ãĪ ¹ +ãī ¥ +ãī ¦ +ãī ¹ +ãī ¿ +ãĬ ŀ +ãĬ ¨ +ãĭ ij +ãĭ ¥ +ãĭ ´ +ãĭ º +ãİ Ħ +ãİ ķ +ãİ ¯ +ãı Ĥ +ãı Ī +ãı ĵ +ãı ĸ +ãı ± +ãIJ ± +ãŁ ģ +ã ¢ +㢠¨ +ã ¨ +㨠³ +ã« ª +ã« ´ +ã¶ ³ +㺠¾ +ä Ģ +äĢ Ģ +ä ĭ +äĭ Į +ä ĮĢ +äIJ Ģ +ä łĢ +ä ł +äł ¼ +ä § +ä§ ŀ +ä¨ ° +ä¨ º +ä ´Ģ +ä · +ä· ħ +ä ·¸ +ê Ĥ +êĤ « +ê Į +êĮ ¼ +ê į +êį ² +êĴ µ +ê ĵ +êĵ ½ +êĻ Ń +êĿ Ľ +êĿ ¥ +ê ŀ +êŀ Ĭ +ê¦ Ĩ +ê¦ ĩ +ê¦ Ł +ê¦ ¨ +ê§ Ī +ê © +ê© Ł +êª ĭ +êª ij +êª ķ +êª Ĺ +êª ľ +êª ® +êª ± +êª » +êª ¼ +ê« Ģ +ê« Ŀ +ê° ĥ +ê° ĺ +ê± ľ +ê² ĵ +ê² ļ +ê³ Ļ +ê³ ¾ +ê´ Ĺ +ê´ Ļ +êµ Ľ +ê¶ ĥ +ê¶ ķ +ê¶ ¨ +ê¸ © +ê¸ ¿ +ê ¹Ħ +ê¹ Ĩ +ê¹ ī +ê¹ ĵ +ê¹ ¢ +ê¹ £ +ê¹ ¸ +êº ³ +ê¿ ı +ê¿ ķ +ê¿ § +ëĢ © +ëģ ħ +ëĥ µ +ëĦ ĸ +ëĦ Ĺ +ëĦ ¢ +ëħ Ĥ +ëĨ IJ +ëĩ ľ +ëĪ ĭ +ëĪ ļ +ëī į +ëī ¨ +ëĬ ļ +ëĬ ¡ +ëĭ ľ +ëĭ ª +ëĮ ĺ +ëĮ ¤ +ëĮ ¸ +ëİ Ł +ëı ¨ +ëIJ Ħ +ëIJ ı +ëIJ ´ +ëIJ ¸ +ëij ģ +ëij ¿ +ëĴ ¨ +ëĵ · +ëĶ ® +ëĶ ² +ëķ § +ëĸ Ķ +ëĸ ª +ëĺ Ń +ëļ Ģ +ëļ ł +ëĽ Ķ +ëĽ © +ëľ ħ +ëŀ ķ +ëŀ ° +ëŁ IJ +ëł ¡ +ë¡ ŀ +ë¡ £ +ë¡ µ +ë£ Ħ +ë£ į +ë¤ ³ +ë¦ į +ë¦ ı +ë¦ ³ +ë§ Ħ +ë§ Ĩ +ë§ į +ë§ ľ +ë§ « +ë§ » +ë¨ ® +ë© Ĥ +ë© Ń +ëª ´ +ë¬ ľ +ë¬ ł +ë¬ « +ë¬ ¾ +ëŃ ¬ +ë® ĺ +ë® ¹ +ë¯ ķ +ë¯ ľ +ë° ¨ +ë° ª +ë± Ķ +ë² ĺ +ë² Ľ +ë² ± +ë² ´ +ë´ ½ +ëµ ¤ +ëµ ¨ +ë· Ĺ +ë· ĺ +ë¸ ĵ +ë¸ ľ +ë¹ ª +ëº ĥ +ëº ĺ +ëº µ +ë» ´ +ë¼ IJ +ë¾ Ķ +ìģ Ń +ìĤ ł +ìĤ ® +ìĥ ı +ìĥ Ļ +ìĦ º +ìħ ¢ +ìĨ Ģ +ìĨ ħ +ìĨ ¤ +ìĨ ¦ +ìĨ ¬ +ìĩ ± +ìĪ µ +ìĭ ¨ +ìĭ ´ +ìĮ ° +ìį ľ +ìİ Ĺ +ìİ ĺ +ìİ ¼ +ìij ī +ìij Ŀ +ìij » +ìĴ Ķ +ìĴ ¯ +ìĵ © +ìķ IJ +ìķ ĸ +ìĸ ł +ìĸ ¾ +ìĹ ĥ +ìĹ Ĺ +ìĹ ľ +ìĹ ¨ +ìĺ Ĥ +ìĺ Ħ +ìĺ ı +ìĺ ¾ +ìĺ ¿ +ìľ § +ìĿ IJ +ìĿ ĸ +ìĿ · +ìŀ į +ìŀ ı +ìŀ ¨ +ìŀ ª +ìŀ ³ +ìł ¡ +ìł ´ +ìł ¹ +ì¡ Ģ +ì¡ ª +ì¡ µ +ì¢ IJ +ì¢ ¨ +ì£ Į +ì£ Ļ +ì£ ³ +ì¦ ij +ì§ ¥ +ì§ ´ +ì§ ¾ +ì¨ ĵ +ì¨ ķ +ì© ° +ì© » +ì© ¼ +ìª Ĺ +ì¬ Ķ +ì¬ ĺ +ì® ® +ì¯ ķ +ì¯ ĺ +ì° İ +ì° ¯ +ì± ĥ +ì± µ +ì² § +ì² ® +ì² ¯ +ì³ ¬ +ì´ ĭ +ì´ ¢ +ìµ ¥ +ì¶ £ +ì¸ Ī +ì¸ Ļ +ìº ¤ +ìº Ń +ì» ½ +ì¼ Ļ +ì½ ¬ +ì¾ Ģ +ì¿ ħ +ì¿ ½ +íĢ ħ +íģ ¦ +íĤ ħ +íĥ ¶ +íĥ ¹ +íĦ Ķ +íħ £ +íĨ Ħ +íĨ § +íĨ ¹ +íĩ ¼ +íī ¤ +íĬ ½ +íĭ Ĥ +íĭ ij +íį Ī +íį Ļ +íį ¿ +íİ ¶ +íIJ Ŀ +íĴ ľ +íĵ Ŀ +íĵ ª +íĵ ± +íĵ · +íĵ ¼ +íĶ Ļ +íĶ ł +íķ ļ +íķ Ľ +íķ ŀ +íķ Ł +íķ § +íķ ¶ +íĸ Ĭ +íĸ ĭ +íĸ į +íĸ Ķ +íĸ ĺ +íĸ ¡ +íĸ ¬ +íĹ £ +íĹ ¿ +íĺ ĸ +íĺ Ń +íļ ° +íĽ į +íĽ ½ +íĿ Ł +íĿ Ń +íĿ ´ +íŀ ľ +ï¤ ī +ï¤ Ń +ï¤ ² +ï¤ µ +ï¤ ¼ +ï¥ Ģ +ï¥ ij +ï¥ Ĵ +ï¥ ķ +ï¥ ĺ +ï¥ Ļ +ï¥ « +ï¥ ¬ +ï¥ ° +ï ¥¿ +ï¦ ĭ +ï¦ ı +ï¦ Ķ +ï¦ ĸ +ï¦ ĺ +ï¦ Ľ +ï¦ ł +ï¦ ® +ï¦ ¯ +ï¦ º +ï¦ » +ï¦ ¾ +ï§ Ĩ +ï§ ĸ +ï§ Ľ +ï§ ŀ +ï§ Ł +ï§ § +ï§ ³ +ï§ º +ï§ ½ +ï¨ ĥ +ï¨ ļ +ï¨ ¢ +ï© Ł +ï¬ ¤ +ï¬ ¬ +ï¬ ¼ +ïŃ Ĵ +ïŃ ķ +ïŃ Ľ +ïŃ Ŀ +ïŃ ŀ +ïŃ Ł +ïŃ ¤ +ïŃ § +ïŃ ¨ +ïŃ ® +ïŃ ° +ïŃ ± +ïŃ · +ïŃ ¹ +ïŃ » +ï® Ģ +ï® ĥ +ï® Ħ +ï® ħ +ï® į +ï® Ĵ +ï® ĵ +ï® ķ +ï® ¦ +ï® ® +ï® ° +ï¯ ĵ +ï¯ ľ +ï¯ © +ï¯ ª +ï¯ ¬ +ï¯ Ń +ï¯ ® +ï¯ · +ï¯ ¹ +ï¯ » +ï¯ ¼ +ï° ĥ +ï° Į +ï° IJ +ï° ĺ +ï° Ļ +ï° ľ +ï° ŀ +ï° ¢ +ï° ® +ï° ° +ï° ¼ +ï° ¿ +ï± Ģ +ï± ģ +ï± Ī +ï± ĭ +ï± ı +ï± Ń +ï² Ģ +ï² ĩ +ï² Ī +ï² ĭ +ï² İ +ï² Ĵ +ï² ľ +ï² ł +ï² ¬ +ï² » +ï³ ĩ +ï³ Ķ +ï³ £ +ï³ « +ï´ ĺ +ï´ ° +ï´ ½ +ï ¶ +ï¶ ° +ï¸ ĸ +ï¸ ´ +ï¸ ¹ +ï¹ į +ï¹ Ĺ +ï¹ ¢ +ï¹ ¤ +ï¹ © +ï¹ ± +ï¾ ° +ï¿ Ĥ +ï¿ ® +ðIJĮ ° +ðIJĮ ¹ +ðIJĮ º +ðIJĮ ½ +ðIJį Ĥ +ðIJį ĥ +ðIJį Ħ +ðIJ İ +ðIJİ ¹ +ðIJ¤ Ĥ +ðIJ¤ į +ðIJ¤ ı +ðIJ¤ ĵ +ðIJŃ ī +ðIJŃ į +ðIJ° ĩ +ðIJ° ° +ðij Ĥ +ðijĤ Ħ +ðij ĺ +ðijĺ ģ +ðĴ Ģ +ðĴĢ ¸ +ðĴ ģ +ðĴģ º +ðĴ Ħ +ðĴĦ · +ðĴ Ĭ +ðĴĬ ij +ðĴ ĭ +ðĴĭ Ĺ +ð ĴĮ +ðĴĮ ¨ +ðĵĥ ¢ +ðĵĥ ° +ðĸ ł +ðĸł ļ +ðĿĦ ĥ +ðĿĦ ħ +ðĿĦ ķ +ðĿĦ Ļ +ðĿĦ ± +ðĿĦ ´ +ðĿĦ ¹ +ðĿħ İ +ðĿħ ª +ðĿĨ £ +ðĿĨ ³ +ðĿĨ ¹ +ðĿĩ Ĭ +ðĿĩ Ĺ +ðĿĩ ļ +ðĿĩ ľ +ðĿĩ ł +ðĿIJ ī +ðĿIJ ĸ +ðĿIJ ĺ +ðĿIJ £ +ðĿIJ ± +ðĿij Ĭ +ðĿij Ń +ðĿij ¼ +ðĿij ½ +ðĿĴ ° +ðĿĴ · +ðĿĴ ¿ +ðĿĵ ģ +ðĿĵ ĭ +ðĿĵ İ +ðĿĵ Ĵ +ðĿ ĵĺ +ðĿĵ ¢ +ðĿĵ ¦ +ðĿĵ « +ðĿĵ ¿ +ðĿĶ İ +ðĿĶ ± +ðĿĶ ´ +ðĿĶ · +ðĿĶ ¸ +ðĿĶ ½ +ðĿķ Ĥ +ðĿķ ĥ +ðĿķ ĭ +ðĿķ ı +ðĿķ IJ +ðĿķ ¥ +ðĿķ ´ +ðĿķ º +ðĿĸ IJ +ðĿĸ Ľ +ðĿĸ Ŀ +ðĿĸ ŀ +ðĿĹ © +ðĿĹ ³ +ðĿĹ ½ +ðĿĺ Ĭ +ðĿĺ ĭ +ðĿĺ Ķ +ðĿĺ ± +ðĿĺ ´ +ðĿĺ ¿ +ðĿĻ Ĵ +ðĿĻ Ŀ +ðĿĻ Ł +ðĿĻ ¬ +ðĿĻ Ń +ðĿĻ » +ðĿĻ ¾ +ðĿļ Ī +ðĿļ ĭ +ðĿļ ij +ðĿļ Ł +ðĿļ ł +ðĿļ £ +ðĿĽ ½ +ðĿľ Ĥ +ðĿľ Ķ +ðĿľ Ļ +ðŁ Ģ +ðŁĢ Ħ +ðŁĦ ² +ðŁĦ ¶ +ðŁħ IJ +ðŁħ ĸ +ðŁħ ļ +ðŁħ Ľ +ðŁħ ¦ +ðŁħ ¶ +ðŁħ » +ðŁħ ¼ +ðŁĨ ĥ +ðŁĨ Ĩ +ðŁĨ İ +ðŁĪ ¯ +ðŁĪ ² +ðŁĪ ¹ +ðŁĮ ĩ +ðŁĮ ĵ +ðŁį ĺ +ðŁİ ij +ðŁİ ¿ +ðŁı ı +ðŁı Ĵ +ðŁı © +ðŁı ¯ +ðŁIJ Ģ +ðŁij Ŀ +ðŁĴ ¹ +ðŁĴ º +ðŁĵ Ł +ðŁĵ ª +ðŁĵ ¼ +ðŁĶ Ģ +ðŁĶ Ĥ +ðŁĶ ĥ +ðŁĶ ĩ +ðŁĶ ĵ +ðŁĶ ¢ +ðŁĶ ¤ +ðŁĶ © +ðŁķ ĸ +ðŁķ ļ +ðŁķ ľ +ðŁķ Ŀ +ðŁķ ŀ +ðŁķ ł +ðŁķ ¢ +ðŁķ ³ +ðŁĸ ĩ +ðŁĸ ij +ðŁĸ ¶ +ðŁĹ ģ +Ñ ¨ +Ú İ +á¡ Į +Ḡ° +áº Ģ +á¼ ® +á½ Ŀ +âĦ ¬ +âļ § +⼠¤ +ã³ ¬ +êĻ ĭ +ê¸ ij +ëĶ ī +ëĹ į +ë¡ ij +ë¯ ij +ë» ħ +ë¼ Ŀ +ìĦ IJ +ìī ¡ +ìĭ ² +ìı ± +ìĹ ¤ +ìĿ © +ìĿ ¿ +ìŁ Ļ +ìł ° +ì¥ ī +íĬ Ń +íķ ® +ï® ı +ðŁħ ± +ðŁĨ Ĵ +ðŁķ ĭ +É ĺ +Ê ĵ +Õ ĥ +à´ ´ +འħ +áĨ º +áĪ Ĭ +áĪ ¨ +áĪ ¾ +áī IJ +áĮ ĥ +áĮ ½ +áĶ Ń +áł Ĥ +áł ¬ +ᨠ¸ +á© ĭ +á¶ ı +á¾ Ķ +á¿ IJ +á¿ ļ +âĻ Ļ +âļ Ĥ +âļ Ĺ +â¡ ¢ +⤠¦ +ëĸ ° +ë¤ Ĥ +ë§ ł +ë± ĭ +ë± IJ +ìĽ ¢ +ìľ ¾ +ì³ ħ +ì» ģ +íģ » +íĥ Ļ +íĵ ĸ +íĵ Ń +íķ ± +íĽ ľ +ï¤ ħ +ï¤ Ĩ +ï¦ ĥ +ï§ © +ï¨ Ĥ +ðIJ¤ Ķ +ðIJŃ ĵ +ðIJ° ¼ +ðĿĵ ŀ +ðĿĵ ° +ðĿĻ ľ +ðĿļ ģ +ðŁħ ¢ +ðŁı ĩ +È ² +Ê ¶ +Ô Ī +Ô ij +Ý ĵ +Ý ¥ +ठij +ॠ± +ଠī +à° ³ +à° µ +à² Ł +áĢ ı +áģ ¼ +áī ¨ +áĬ Ĵ +áĭ © +áĮ Ħ +áĮ Ķ +áIJ § +á ĴĮ +áĶ ħ +áĶ Ĭ +áł Ħ +ᨠģ +Ḡĥ +Ḡ» +âĶ ŀ +âĺ µ +âļ £ +â² ¢ +ãĪ ª +ä¶ µ +ê² Ļ +ê² ´ +ê³ Ĥ +ë¡ ¼ +ìĨ Ĭ +ì¼ ĩ +íĭ į +íĵ ¬ +íĵ ® +íĵ ¶ +íĵ » +ï¤ ¦ +ï¥ ł +ï¥ ± +ïŃ ² +ðIJŃ Ĭ +ðIJ ±ħ +ðĸ ¥ +ðĸ¥ ¨ +ðĿij ³ +ðĿĵ ķ +ðĿĵ ¬ +ðĿĵ ¹ +ðĿĵ ¾ +ðĿĶ ĵ +ðĿķ į +ðĿķ ¡ +ðĿķ ± +ðĿĸ ĸ +ðĿĺ ı +ðĿĺ IJ +ðĿĺ ļ +ðĿĻ ® +ðĿĻ ° +ðĿĻ ¸ +ðĿĻ º +ðĿĻ ¼ +ðĿĻ ½ +ðĿĻ ¿ +ðĿļ Ħ +ðĿļ ı +ðŁħ ħ +ðŁħ ĵ +Æ Ī +àł Į +áĻ ³ +á ļĮ +ἠħ +ἠIJ +ᤠĬ +ḠĬ +âĶ ½ +âķ Ĭ +⼠ĩ +⼠ı +âĿ ª +âĿ « +⣠° +ãĦ į +ãĦ ĵ +ãĦ § +ãħ ĸ +ãī « +ê¦ Ķ +ï± Ĭ +ຠĤ +áħ £ +á¥ Ķ +ᥠ¤ +âĨ ¤ +âĨ · +âĩ ŀ +âĸ ¤ +âŀ ¶ +ãĪ ¼ +ï¨ · +ðĵı § +âĶ ² +âĢ ´ +âĴ Ł +âĴ ¡ +â° Ĥ +â° į +â° İ +â° IJ +â° ij +â° Ł +â° ł +â° ¡ +â¼ Ń +ãĬ ¥ +âĴ ł +â½ º +ãĩ º +ãĩ ½ +ï¨ Ĭ +áķ · +âį ¨ +âº Ł +â½ Ĺ diff --git a/20250121165312/hf-593/model.safetensors.index.json b/20250121165312/hf-593/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..ac262537ce34e6cc04216f0560ee29def496b7fd --- /dev/null +++ b/20250121165312/hf-593/model.safetensors.index.json @@ -0,0 +1,970 @@ +{ + "metadata": { + "total_size": 145412407296 + }, + "weight_map": { + "lm_head.weight": "model-00031-of-00031.safetensors", + "model.embed_tokens.weight": "model-00001-of-00031.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00031.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00031.safetensors", + "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00031.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00031.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00031.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00031.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00031.safetensors", + "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00031.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00031.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00031.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00031.safetensors", + "model.layers.10.input_layernorm.weight": "model-00005-of-00031.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00005-of-00031.safetensors", + "model.layers.10.self_attn.k_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.10.self_attn.q_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.10.self_attn.v_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.11.input_layernorm.weight": "model-00005-of-00031.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00031.safetensors", + "model.layers.11.self_attn.k_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.11.self_attn.q_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.11.self_attn.v_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.12.input_layernorm.weight": "model-00006-of-00031.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00006-of-00031.safetensors", + "model.layers.12.self_attn.k_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.12.self_attn.q_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.12.self_attn.v_proj.bias": "model-00005-of-00031.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.13.input_layernorm.weight": "model-00006-of-00031.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00006-of-00031.safetensors", + "model.layers.13.self_attn.k_proj.bias": "model-00006-of-00031.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.13.self_attn.q_proj.bias": "model-00006-of-00031.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.13.self_attn.v_proj.bias": "model-00006-of-00031.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.14.input_layernorm.weight": "model-00006-of-00031.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00006-of-00031.safetensors", + "model.layers.14.self_attn.k_proj.bias": "model-00006-of-00031.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.14.self_attn.q_proj.bias": "model-00006-of-00031.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.14.self_attn.v_proj.bias": "model-00006-of-00031.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00006-of-00031.safetensors", + "model.layers.15.input_layernorm.weight": "model-00007-of-00031.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00007-of-00031.safetensors", + "model.layers.15.self_attn.k_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.15.self_attn.q_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.15.self_attn.v_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.16.input_layernorm.weight": "model-00007-of-00031.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00007-of-00031.safetensors", + "model.layers.16.self_attn.k_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.16.self_attn.q_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.16.self_attn.v_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.17.input_layernorm.weight": "model-00008-of-00031.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00008-of-00031.safetensors", + "model.layers.17.self_attn.k_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.17.self_attn.q_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.17.self_attn.v_proj.bias": "model-00007-of-00031.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00007-of-00031.safetensors", + "model.layers.18.input_layernorm.weight": "model-00008-of-00031.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00008-of-00031.safetensors", + "model.layers.18.self_attn.k_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.18.self_attn.q_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.18.self_attn.v_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.19.input_layernorm.weight": "model-00008-of-00031.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00008-of-00031.safetensors", + "model.layers.19.self_attn.k_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.19.self_attn.q_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.19.self_attn.v_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.2.input_layernorm.weight": "model-00002-of-00031.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00031.safetensors", + "model.layers.2.self_attn.k_proj.bias": "model-00002-of-00031.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.2.self_attn.q_proj.bias": "model-00002-of-00031.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.2.self_attn.v_proj.bias": "model-00002-of-00031.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.20.input_layernorm.weight": "model-00009-of-00031.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00009-of-00031.safetensors", + "model.layers.20.self_attn.k_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.20.self_attn.q_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.20.self_attn.v_proj.bias": "model-00008-of-00031.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00008-of-00031.safetensors", + "model.layers.21.input_layernorm.weight": "model-00009-of-00031.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00009-of-00031.safetensors", + "model.layers.21.self_attn.k_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.21.self_attn.q_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.21.self_attn.v_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.22.input_layernorm.weight": "model-00009-of-00031.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00009-of-00031.safetensors", + "model.layers.22.self_attn.k_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.22.self_attn.q_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.22.self_attn.v_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.23.input_layernorm.weight": "model-00010-of-00031.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00010-of-00031.safetensors", + "model.layers.23.self_attn.k_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.23.self_attn.q_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.23.self_attn.v_proj.bias": "model-00009-of-00031.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00009-of-00031.safetensors", + "model.layers.24.input_layernorm.weight": "model-00010-of-00031.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00010-of-00031.safetensors", + "model.layers.24.self_attn.k_proj.bias": "model-00010-of-00031.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.24.self_attn.q_proj.bias": "model-00010-of-00031.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.24.self_attn.v_proj.bias": "model-00010-of-00031.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.25.input_layernorm.weight": "model-00010-of-00031.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00010-of-00031.safetensors", + "model.layers.25.self_attn.k_proj.bias": "model-00010-of-00031.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.25.self_attn.q_proj.bias": "model-00010-of-00031.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.25.self_attn.v_proj.bias": "model-00010-of-00031.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00010-of-00031.safetensors", + "model.layers.26.input_layernorm.weight": "model-00011-of-00031.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00011-of-00031.safetensors", + "model.layers.26.self_attn.k_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.26.self_attn.q_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.26.self_attn.v_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.27.input_layernorm.weight": "model-00011-of-00031.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00011-of-00031.safetensors", + "model.layers.27.self_attn.k_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.27.self_attn.q_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.27.self_attn.v_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.28.input_layernorm.weight": "model-00012-of-00031.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00012-of-00031.safetensors", + "model.layers.28.self_attn.k_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.28.self_attn.q_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.28.self_attn.v_proj.bias": "model-00011-of-00031.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00011-of-00031.safetensors", + "model.layers.29.input_layernorm.weight": "model-00012-of-00031.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00012-of-00031.safetensors", + "model.layers.29.self_attn.k_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.29.self_attn.q_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.29.self_attn.v_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.3.input_layernorm.weight": "model-00002-of-00031.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00031.safetensors", + "model.layers.3.self_attn.k_proj.bias": "model-00002-of-00031.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.3.self_attn.q_proj.bias": "model-00002-of-00031.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.3.self_attn.v_proj.bias": "model-00002-of-00031.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00031.safetensors", + "model.layers.30.input_layernorm.weight": "model-00012-of-00031.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00012-of-00031.safetensors", + "model.layers.30.self_attn.k_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.30.self_attn.q_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.30.self_attn.v_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.31.input_layernorm.weight": "model-00013-of-00031.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00013-of-00031.safetensors", + "model.layers.31.self_attn.k_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.31.self_attn.q_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.31.self_attn.v_proj.bias": "model-00012-of-00031.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00012-of-00031.safetensors", + "model.layers.32.input_layernorm.weight": "model-00013-of-00031.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00013-of-00031.safetensors", + "model.layers.32.self_attn.k_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.32.self_attn.q_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.32.self_attn.v_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.33.input_layernorm.weight": "model-00013-of-00031.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00013-of-00031.safetensors", + "model.layers.33.self_attn.k_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.33.self_attn.q_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.33.self_attn.v_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.34.input_layernorm.weight": "model-00014-of-00031.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00014-of-00031.safetensors", + "model.layers.34.self_attn.k_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.34.self_attn.q_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.34.self_attn.v_proj.bias": "model-00013-of-00031.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00013-of-00031.safetensors", + "model.layers.35.input_layernorm.weight": "model-00014-of-00031.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00014-of-00031.safetensors", + "model.layers.35.self_attn.k_proj.bias": "model-00014-of-00031.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.35.self_attn.q_proj.bias": "model-00014-of-00031.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.35.self_attn.v_proj.bias": "model-00014-of-00031.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.36.input_layernorm.weight": "model-00014-of-00031.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00014-of-00031.safetensors", + "model.layers.36.self_attn.k_proj.bias": "model-00014-of-00031.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.36.self_attn.q_proj.bias": "model-00014-of-00031.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.36.self_attn.v_proj.bias": "model-00014-of-00031.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00014-of-00031.safetensors", + "model.layers.37.input_layernorm.weight": "model-00015-of-00031.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00015-of-00031.safetensors", + "model.layers.37.self_attn.k_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.37.self_attn.q_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.37.self_attn.v_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.38.input_layernorm.weight": "model-00015-of-00031.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00015-of-00031.safetensors", + "model.layers.38.self_attn.k_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.38.self_attn.q_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.38.self_attn.v_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.39.input_layernorm.weight": "model-00016-of-00031.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00016-of-00031.safetensors", + "model.layers.39.self_attn.k_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.39.self_attn.q_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.39.self_attn.v_proj.bias": "model-00015-of-00031.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00015-of-00031.safetensors", + "model.layers.4.input_layernorm.weight": "model-00003-of-00031.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00003-of-00031.safetensors", + "model.layers.4.self_attn.k_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.4.self_attn.q_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.4.self_attn.v_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.40.input_layernorm.weight": "model-00016-of-00031.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00016-of-00031.safetensors", + "model.layers.40.self_attn.k_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.40.self_attn.q_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.40.self_attn.v_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.41.input_layernorm.weight": "model-00016-of-00031.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00016-of-00031.safetensors", + "model.layers.41.self_attn.k_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.41.self_attn.q_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.41.self_attn.v_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.42.input_layernorm.weight": "model-00017-of-00031.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00017-of-00031.safetensors", + "model.layers.42.self_attn.k_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.42.self_attn.q_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.42.self_attn.v_proj.bias": "model-00016-of-00031.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00016-of-00031.safetensors", + "model.layers.43.input_layernorm.weight": "model-00017-of-00031.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00017-of-00031.safetensors", + "model.layers.43.self_attn.k_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.43.self_attn.q_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.43.self_attn.v_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.44.input_layernorm.weight": "model-00017-of-00031.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00017-of-00031.safetensors", + "model.layers.44.self_attn.k_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.44.self_attn.q_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.44.self_attn.v_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.45.input_layernorm.weight": "model-00018-of-00031.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00018-of-00031.safetensors", + "model.layers.45.self_attn.k_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.45.self_attn.q_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.45.self_attn.v_proj.bias": "model-00017-of-00031.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00017-of-00031.safetensors", + "model.layers.46.input_layernorm.weight": "model-00018-of-00031.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00018-of-00031.safetensors", + "model.layers.46.self_attn.k_proj.bias": "model-00018-of-00031.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.46.self_attn.q_proj.bias": "model-00018-of-00031.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.46.self_attn.v_proj.bias": "model-00018-of-00031.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.47.input_layernorm.weight": "model-00018-of-00031.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00018-of-00031.safetensors", + "model.layers.47.self_attn.k_proj.bias": "model-00018-of-00031.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.47.self_attn.q_proj.bias": "model-00018-of-00031.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.47.self_attn.v_proj.bias": "model-00018-of-00031.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00018-of-00031.safetensors", + "model.layers.48.input_layernorm.weight": "model-00019-of-00031.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00019-of-00031.safetensors", + "model.layers.48.self_attn.k_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.48.self_attn.q_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.48.self_attn.v_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.49.input_layernorm.weight": "model-00019-of-00031.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00019-of-00031.safetensors", + "model.layers.49.self_attn.k_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.49.self_attn.q_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.49.self_attn.v_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.5.input_layernorm.weight": "model-00003-of-00031.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00031.safetensors", + "model.layers.5.self_attn.k_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.5.self_attn.q_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.5.self_attn.v_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.50.input_layernorm.weight": "model-00020-of-00031.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00020-of-00031.safetensors", + "model.layers.50.self_attn.k_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.50.self_attn.q_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.50.self_attn.v_proj.bias": "model-00019-of-00031.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00019-of-00031.safetensors", + "model.layers.51.input_layernorm.weight": "model-00020-of-00031.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00020-of-00031.safetensors", + "model.layers.51.self_attn.k_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.51.self_attn.q_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.51.self_attn.v_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.52.input_layernorm.weight": "model-00020-of-00031.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00020-of-00031.safetensors", + "model.layers.52.self_attn.k_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.52.self_attn.q_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.52.self_attn.v_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.53.input_layernorm.weight": "model-00021-of-00031.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00021-of-00031.safetensors", + "model.layers.53.self_attn.k_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.53.self_attn.q_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.53.self_attn.v_proj.bias": "model-00020-of-00031.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00020-of-00031.safetensors", + "model.layers.54.input_layernorm.weight": "model-00021-of-00031.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00021-of-00031.safetensors", + "model.layers.54.self_attn.k_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.54.self_attn.q_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.54.self_attn.v_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.55.input_layernorm.weight": "model-00021-of-00031.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00021-of-00031.safetensors", + "model.layers.55.self_attn.k_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.55.self_attn.q_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.55.self_attn.v_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.56.input_layernorm.weight": "model-00022-of-00031.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00022-of-00031.safetensors", + "model.layers.56.self_attn.k_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.56.self_attn.q_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.56.self_attn.v_proj.bias": "model-00021-of-00031.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00021-of-00031.safetensors", + "model.layers.57.input_layernorm.weight": "model-00022-of-00031.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00022-of-00031.safetensors", + "model.layers.57.self_attn.k_proj.bias": "model-00022-of-00031.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.57.self_attn.q_proj.bias": "model-00022-of-00031.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.57.self_attn.v_proj.bias": "model-00022-of-00031.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.58.input_layernorm.weight": "model-00022-of-00031.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00022-of-00031.safetensors", + "model.layers.58.self_attn.k_proj.bias": "model-00022-of-00031.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.58.self_attn.q_proj.bias": "model-00022-of-00031.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.58.self_attn.v_proj.bias": "model-00022-of-00031.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00022-of-00031.safetensors", + "model.layers.59.input_layernorm.weight": "model-00023-of-00031.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00023-of-00031.safetensors", + "model.layers.59.self_attn.k_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.59.self_attn.q_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.59.self_attn.v_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.6.input_layernorm.weight": "model-00004-of-00031.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00004-of-00031.safetensors", + "model.layers.6.self_attn.k_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.6.self_attn.q_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.6.self_attn.v_proj.bias": "model-00003-of-00031.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00031.safetensors", + "model.layers.60.input_layernorm.weight": "model-00023-of-00031.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00023-of-00031.safetensors", + "model.layers.60.self_attn.k_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.60.self_attn.q_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.60.self_attn.v_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.61.input_layernorm.weight": "model-00024-of-00031.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00024-of-00031.safetensors", + "model.layers.61.self_attn.k_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.61.self_attn.q_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.61.self_attn.v_proj.bias": "model-00023-of-00031.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00023-of-00031.safetensors", + "model.layers.62.input_layernorm.weight": "model-00024-of-00031.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00024-of-00031.safetensors", + "model.layers.62.self_attn.k_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.62.self_attn.q_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.62.self_attn.v_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.63.input_layernorm.weight": "model-00024-of-00031.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00024-of-00031.safetensors", + "model.layers.63.self_attn.k_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.63.self_attn.q_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.63.self_attn.v_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.64.input_layernorm.weight": "model-00025-of-00031.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00025-of-00031.safetensors", + "model.layers.64.self_attn.k_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.64.self_attn.q_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.64.self_attn.v_proj.bias": "model-00024-of-00031.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00024-of-00031.safetensors", + "model.layers.65.input_layernorm.weight": "model-00025-of-00031.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00025-of-00031.safetensors", + "model.layers.65.self_attn.k_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.65.self_attn.q_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.65.self_attn.v_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.66.input_layernorm.weight": "model-00025-of-00031.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00025-of-00031.safetensors", + "model.layers.66.self_attn.k_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.66.self_attn.q_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.66.self_attn.v_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.67.input_layernorm.weight": "model-00026-of-00031.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00026-of-00031.safetensors", + "model.layers.67.self_attn.k_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.67.self_attn.q_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.67.self_attn.v_proj.bias": "model-00025-of-00031.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00025-of-00031.safetensors", + "model.layers.68.input_layernorm.weight": "model-00026-of-00031.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00026-of-00031.safetensors", + "model.layers.68.self_attn.k_proj.bias": "model-00026-of-00031.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.68.self_attn.q_proj.bias": "model-00026-of-00031.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.68.self_attn.v_proj.bias": "model-00026-of-00031.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.69.input_layernorm.weight": "model-00026-of-00031.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00026-of-00031.safetensors", + "model.layers.69.self_attn.k_proj.bias": "model-00026-of-00031.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.69.self_attn.q_proj.bias": "model-00026-of-00031.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.69.self_attn.v_proj.bias": "model-00026-of-00031.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00026-of-00031.safetensors", + "model.layers.7.input_layernorm.weight": "model-00004-of-00031.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00004-of-00031.safetensors", + "model.layers.7.self_attn.k_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.7.self_attn.q_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.7.self_attn.v_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.70.input_layernorm.weight": "model-00027-of-00031.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00027-of-00031.safetensors", + "model.layers.70.self_attn.k_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.70.self_attn.q_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.70.self_attn.v_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.71.input_layernorm.weight": "model-00027-of-00031.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00027-of-00031.safetensors", + "model.layers.71.self_attn.k_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.71.self_attn.q_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.71.self_attn.v_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.72.input_layernorm.weight": "model-00028-of-00031.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00028-of-00031.safetensors", + "model.layers.72.self_attn.k_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.72.self_attn.q_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.72.self_attn.v_proj.bias": "model-00027-of-00031.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00027-of-00031.safetensors", + "model.layers.73.input_layernorm.weight": "model-00028-of-00031.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00028-of-00031.safetensors", + "model.layers.73.self_attn.k_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.73.self_attn.q_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.73.self_attn.v_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.74.input_layernorm.weight": "model-00028-of-00031.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00028-of-00031.safetensors", + "model.layers.74.self_attn.k_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.74.self_attn.q_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.74.self_attn.v_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.75.input_layernorm.weight": "model-00029-of-00031.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00029-of-00031.safetensors", + "model.layers.75.self_attn.k_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.75.self_attn.q_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.75.self_attn.v_proj.bias": "model-00028-of-00031.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00028-of-00031.safetensors", + "model.layers.76.input_layernorm.weight": "model-00029-of-00031.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00029-of-00031.safetensors", + "model.layers.76.self_attn.k_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.76.self_attn.q_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.76.self_attn.v_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.77.input_layernorm.weight": "model-00029-of-00031.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00029-of-00031.safetensors", + "model.layers.77.self_attn.k_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.77.self_attn.q_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.77.self_attn.v_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.78.input_layernorm.weight": "model-00030-of-00031.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00030-of-00031.safetensors", + "model.layers.78.self_attn.k_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.78.self_attn.q_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.78.self_attn.v_proj.bias": "model-00029-of-00031.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00029-of-00031.safetensors", + "model.layers.79.input_layernorm.weight": "model-00030-of-00031.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00030-of-00031.safetensors", + "model.layers.79.self_attn.k_proj.bias": "model-00030-of-00031.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.79.self_attn.q_proj.bias": "model-00030-of-00031.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.79.self_attn.v_proj.bias": "model-00030-of-00031.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00030-of-00031.safetensors", + "model.layers.8.input_layernorm.weight": "model-00004-of-00031.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00004-of-00031.safetensors", + "model.layers.8.self_attn.k_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.8.self_attn.q_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.8.self_attn.v_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.9.input_layernorm.weight": "model-00005-of-00031.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00005-of-00031.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00005-of-00031.safetensors", + "model.layers.9.self_attn.k_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.9.self_attn.q_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00004-of-00031.safetensors", + "model.layers.9.self_attn.v_proj.bias": "model-00004-of-00031.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00004-of-00031.safetensors", + "model.norm.weight": "model-00030-of-00031.safetensors" + } +} diff --git a/20250121165312/hf-593/special_tokens_map.json b/20250121165312/hf-593/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..ac23c0aaa2434523c494330aeb79c58395378103 --- /dev/null +++ b/20250121165312/hf-593/special_tokens_map.json @@ -0,0 +1,31 @@ +{ + "additional_special_tokens": [ + "<|im_start|>", + "<|im_end|>", + "<|object_ref_start|>", + "<|object_ref_end|>", + "<|box_start|>", + "<|box_end|>", + "<|quad_start|>", + "<|quad_end|>", + "<|vision_start|>", + "<|vision_end|>", + "<|vision_pad|>", + "<|image_pad|>", + "<|video_pad|>" + ], + "eos_token": { + "content": "<|im_end|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "<|endoftext|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/20250121165312/hf-593/vocab.json b/20250121165312/hf-593/vocab.json new file mode 100644 index 0000000000000000000000000000000000000000..6bce3a0a3866c4791a74d83d78f6824c3af64ec3 --- /dev/null +++ b/20250121165312/hf-593/vocab.json @@ -0,0 +1,151645 @@ +{ + "!": 0, + "!!": 2928, + "!!!": 12069, + "!!!!": 17085, + "!!!!!": 69800, + "!!!!!!!!": 50667, + "!!!!ĊĊ": 81392, + "!!!Ċ": 79295, + "!!!ĊĊ": 32057, + "!!\");Ċ": 98164, + "!!)Ċ": 85892, + "!!,": 93687, + "!!.": 24426, + "!!Ċ": 50347, + "!!ĊĊ": 24796, + "!\"": 8958, + "!\")": 88783, + "!\");": 85540, + "!\");Ċ": 11142, + "!\");ĊĊ": 57073, + "!\");čĊ": 42012, + "!\")Ċ": 22988, + "!\")ĊĊ": 66506, + "!\")čĊ": 93896, + "!\",": 18789, + "!\",Ċ": 33436, + "!\".": 92993, + "!\";Ċ": 26782, + "!\";čĊ": 93397, + "!\"Ċ": 24734, + "!\"ĊĊ": 17199, + "!'": 31383, + "!');Ċ": 25490, + "!')Ċ": 36180, + "!')ĊĊ": 98973, + "!',": 38692, + "!',Ċ": 36260, + "!';Ċ": 49174, + "!'Ċ": 48727, + "!(": 10297, + "!(\"": 17223, + "!(\"{": 88928, + "!(\"{}\",": 79878, + "!(:": 49116, + "!(Ċ": 33673, + "!)": 16315, + "!),": 48643, + "!).": 42392, + "!).ĊĊ": 86779, + "!)Ċ": 27584, + "!)ĊĊ": 35786, + "!*": 68222, + "!*\\Ċ": 90237, + "!,": 17142, + "!,Ċ": 72589, + "!--": 2495, + "!.": 15365, + "!.ĊĊ": 98527, + "!/": 87980, + "!:": 35575, + "!;Ċ": 85318, + "!<": 86932, + "!": 24011, + "\").": 1827, + "\").Ċ": 38609, + "\").ĊĊ": 80984, + "\"):": 37051, + "\"):Ċ": 15146, + "\"):čĊ": 74199, + "\");": 5038, + "\");//": 40919, + "\");}Ċ": 90118, + "\");Ċ": 797, + "\");ĊĊ": 3071, + "\");ĊĊĊ": 32636, + "\");čĊ": 2815, + "\");čĊčĊ": 15746, + "\")==": 43664, + "\")[": 17135, + "\")]": 30917, + "\")]Ċ": 5422, + "\")]ĊĊ": 72389, + "\")]čĊ": 25731, + "\"){": 47189, + "\"){Ċ": 14818, + "\"){čĊ": 50938, + "\")}": 67455, + "\")},Ċ": 79583, + "\")}Ċ": 42132, + "\")Ċ": 1138, + "\")ĊĊ": 5130, + "\")ĊĊĊ": 25639, + "\")čĊ": 6060, + "\")čĊčĊ": 30218, + "\"*": 61593, + "\"+": 5572, + "\"+\"": 77728, + "\"+Ċ": 27424, + "\",": 497, + "\",\"": 2198, + "\",\"\");Ċ": 75976, + "\",\"\",": 81390, + "\",\"\",\"": 59859, + "\",\"#": 58324, + "\",\"+": 57126, + "\",\"\\": 54256, + "\",$": 17982, + "\",&": 17834, + "\",'": 38838, + "\",(": 44970, + "\",-": 65575, + "\",@\"": 37366, + "\",[": 83498, + "\",__": 78427, + "\",{": 74448, + "\",Ċ": 756, + "\",ĊĊ": 25897, + "\",čĊ": 4723, + "\"-": 34294, + "\"--": 74757, + "\".": 3263, + "\".\"": 65720, + "\".$": 8634, + "\".$_": 88706, + "\"..": 62923, + "\"...": 52570, + "\"./": 95092, + "\".Ċ": 22956, + "\".ĊĊ": 11436, + "\".ĊĊĊĊ": 80874, + "\"/": 3115, + "\"/>": 13123, + "\"/>..Ċ": 61166, + "\"/><": 75942, + "\"/>Ċ": 4639, + "\"/>ĊĊ": 73330, + "\"/>čĊ": 41578, + "\":": 788, + "\":\"": 3252, + "\":\"\"": 62366, + "\":\"\",\"": 60767, + "\":\"\",Ċ": 42398, + "\":\"'": 71835, + "\":\"+": 70318, + "\":\"/": 58528, + "\":-": 52052, + "\":@\"": 73466, + "\":[": 8899, + "\":[\"": 36799, + "\":[-": 63028, + "\":[{\"": 66582, + "\":[{Ċ": 58018, + "\":{\"": 22317, + "\":{Ċ": 12602, + "\":Ċ": 4660, + "\":ĊĊ": 51418, + "\":čĊ": 20354, + "\";": 5123, + "\";//": 52035, + "\";}": 75091, + "\";}Ċ": 65543, + "\";Ċ": 876, + "\";ĊĊ": 3302, + "\";ĊĊĊ": 29122, + "\";čĊ": 3534, + "\";čĊčĊ": 21579, + "\"<": 22476, + "\"": 11147, + "\"=>\"": 20634, + "\"=>$": 31682, + "\">": 755, + "\">#": 86143, + "\">$": 59849, + "\">${": 47705, + "\">%": 71667, + "\">&": 13366, + "\">&#": 73144, + "\">'": 17756, + "\">'+": 40622, + "\">'+Ċ": 84086, + "\">',": 78304, + "\">',Ċ": 71820, + "\">'.": 56321, + "\">'.$": 35668, + "\">';Ċ": 24352, + "\">';čĊ": 78042, + "\">'Ċ": 74897, + "\">(": 57551, + "\">*-->Ċ": 96121, + "\">--}}Ċ": 79846, + "\">//": 72006, + "\"><": 3088, + "\">": 29052, + "-->Ċ": 9992, + "-->ĊĊ": 37859, + "-->čĊ": 38942, + "--[": 91487, + "--[[": 84105, + "--}}Ċ": 27220, + "--Ċ": 7088, + "--ĊĊ": 26469, + "--čĊ": 53751, + "-.": 14523, + "-/": 66114, + "-": 405, + "->$": 36726, + "->[": 69912, + "->_": 5129, + "->__": 29899, + "->___": 85442, + "->{": 10484, + "->{$": 33647, + "->{'": 27622, + "->{_": 98393, + "->Ċ": 84472, + "-A": 6691, + "-AA": 63664, + "-API": 98773, + "-Agent": 45118, + "-Al": 64373, + "-Allow": 33348, + "-Am": 98810, + "-Americ": 50683, + "-American": 23264, + "-Americans": 59064, + "-An": 57916, + "-Apr": 66148, + "-Ass": 61155, + "-Assad": 75186, + "-Aug": 61635, + "-B": 7671, + "-BEGIN": 53945, + "-Bar": 93320, + "-Based": 60339, + "-Benz": 62221, + "-Bold": 89581, + "-C": 7658, + "-CN": 94942, + "-CS": 78313, + "-Cal": 75350, + "-Cds": 98018, + "-Ch": 29553, + "-China": 97609, + "-Christian": 88041, + "-Class": 81526, + "-Clause": 62915, + "-Co": 86171, + "-Col": 54546, + "-Cola": 65177, + "-Compatible": 37262, + "-Con": 57748, + "-Control": 25056, + "-Core": 66429, + "-Cs": 78222, + "-D": 9420, + "-DD": 40175, + "-Day": 54912, + "-Dec": 79629, + "-Disposition": 62489, + "-E": 13431, + "-END": 51107, + "-East": 93664, + "-Encoding": 67358, + "-European": 87635, + "-F": 7276, + "-FIRST": 30362, + "-Feb": 93771, + "-Fi": 26299, + "-Free": 62890, + "-G": 12010, + "-General": 57982, + "-H": 11278, + "-HT": 99085, + "-Headers": 76006, + "-I": 21477, + "-ID": 53365, + "-II": 75544, + "-IN": 80021, + "-INF": 71470, + "-INFRINGEMENT": 94673, + "-IS": 97518, + "-Identifier": 36773, + "-In": 31500, + "-Isl": 79978, + "-Israel": 70990, + "-J": 12009, + "-Jan": 66452, + "-Javadoc": 49385, + "-Jul": 91179, + "-Jun": 77734, + "-K": 15843, + "-Key": 94321, + "-L": 8125, + "-LAST": 30515, + "-La": 93916, + "-Language": 90765, + "-Le": 46168, + "-League": 84611, + "-Length": 52493, + "-Level": 92780, + "-License": 36479, + "-Life": 87808, + "-Line": 91536, + "-M": 5251, + "-MM": 18506, + "-Mail": 37513, + "-Man": 30151, + "-Mar": 47435, + "-Mart": 64156, + "-May": 61986, + "-Men": 49459, + "-Methods": 95490, + "-Mobile": 56353, + "-Muslim": 61405, + "-N": 11250, + "-NLS": 37484, + "-Nazi": 96558, + "-New": 89737, + "-No": 97976, + "-Nov": 70488, + "-O": 23620, + "-Oct": 92426, + "-On": 66354, + "-One": 90377, + "-Origin": 51636, + "-Out": 84515, + "-Owned": 90052, + "-P": 9299, + "-Pacific": 63876, + "-Pack": 76582, + "-President": 85381, + "-Pro": 57086, + "-Q": 28242, + "-Qaeda": 56885, + "-R": 10911, + "-REAL": 95262, + "-ROM": 98383, + "-Ray": 78404, + "-Re": 67960, + "-Regular": 87591, + "-Requested": 89109, + "-Russian": 84035, + "-S": 6222, + "-SA": 92214, + "-ST": 92577, + "-Sah": 92114, + "-Saharan": 96419, + "-Se": 97528, + "-Semit": 44092, + "-Semitic": 73497, + "-Semitism": 64810, + "-Sep": 95153, + "-Series": 91731, + "-Sh": 30261, + "-Shirt": 75854, + "-Smith": 96238, + "-Speed": 94477, + "-St": 39620, + "-Star": 38784, + "-State": 86515, + "-T": 9285, + "-TV": 52292, + "-Ta": 43095, + "-Th": 70526, + "-The": 65006, + "-Time": 54485, + "-To": 58308, + "-Token": 89022, + "-Tr": 88047, + "-Trump": 66920, + "-Type": 10804, + "-U": 34581, + "-UA": 36736, + "-US": 32340, + "-Un": 82245, + "-Up": 46097, + "-V": 19625, + "-Version": 83902, + "-W": 13002, + "-West": 60761, + "-With": 83156, + "-X": 30550, + "-Y": 29137, + "-Year": 70898, + "-Z": 11171, + "-Za": 39279, + "-[": 40995, + "-\\": 30529, + "-_": 70049, + "-a": 7409, + "-ab": 38030, + "-abortion": 95483, + "-about": 68105, + "-abs": 96410, + "-ac": 37598, + "-access": 42156, + "-account": 49982, + "-acre": 63334, + "-action": 25059, + "-actions": 51321, + "-active": 31244, + "-ad": 25747, + "-add": 18570, + "-added": 86342, + "-addon": 39777, + "-addons": 94759, + "-address": 41633, + "-adjust": 59584, + "-admin": 27225, + "-aff": 70160, + "-after": 44573, + "-ag": 41985, + "-age": 42941, + "-aged": 56759, + "-agent": 41935, + "-ahead": 98488, + "-air": 37535, + "-ajax": 82320, + "-al": 18779, + "-alert": 55745, + "-align": 9644, + "-aligned": 97451, + "-alist": 80978, + "-all": 22346, + "-alone": 73149, + "-alpha": 64538, + "-alt": 33350, + "-am": 32217, + "-an": 18883, + "-analysis": 55436, + "-analytics": 98910, + "-anchor": 75445, + "-and": 9777, + "-android": 97388, + "-angle": 34381, + "-angular": 50444, + "-animate": 81350, + "-animation": 63206, + "-answer": 95777, + "-any": 88427, + "-aos": 64844, + "-ap": 45156, + "-api": 23904, + "-app": 20023, + "-append": 88229, + "-application": 92479, + "-appointed": 98363, + "-approved": 66262, + "-ar": 57393, + "-archive": 95100, + "-area": 29022, + "-arm": 66818, + "-around": 45616, + "-array": 62058, + "-arrow": 29308, + "-art": 37821, + "-article": 77405, + "-as": 32434, + "-ass": 34815, + "-assets": 75158, + "-associated": 74868, + "-at": 28783, + "-au": 81997, + "-aut": 45123, + "-auth": 33503, + "-author": 42702, + "-authored": 95371, + "-auto": 19938, + "-automatic": 94473, + "-available": 93732, + "-avatar": 79087, + "-average": 77426, + "-await": 79738, + "-awaited": 89250, + "-aware": 65004, + "-away": 89624, + "-awesome": 28524, + "-axis": 35321, + "-b": 1455, + "-back": 15461, + "-backed": 45028, + "-backend": 89987, + "-background": 41934, + "-badge": 88076, + "-bal": 83125, + "-ball": 67202, + "-band": 67675, + "-banner": 70883, + "-bar": 15773, + "-bars": 82827, + "-base": 30013, + "-based": 5980, + "-basic": 51002, + "-basket": 94675, + "-be": 15150, + "-bearing": 91153, + "-bed": 48507, + "-bedroom": 66867, + "-before": 69447, + "-begin": 91673, + "-being": 32751, + "-bel": 90205, + "-best": 56601, + "-beta": 67640, + "-between": 25959, + "-bg": 35804, + "-big": 69411, + "-billion": 69073, + "-bin": 49125, + "-binary": 94839, + "-bind": 64963, + "-binding": 64400, + "-bit": 15257, + "-bl": 16278, + "-black": 36946, + "-blind": 94349, + "-block": 9425, + "-blocking": 78761, + "-blog": 59939, + "-blood": 92903, + "-blue": 31654, + "-bo": 69466, + "-board": 52855, + "-bodied": 96297, + "-body": 9350, + "-bold": 27639, + "-book": 33640, + "-books": 83091, + "-boot": 88289, + "-bootstrap": 24383, + "-border": 27019, + "-bordered": 31354, + "-born": 39160, + "-bot": 89361, + "-bottom": 11381, + "-bound": 49862, + "-box": 12625, + "-boy": 84903, + "-br": 30117, + "-brand": 33120, + "-bre": 91908, + "-break": 34349, + "-breaking": 54307, + "-browser": 31555, + "-bs": 56430, + "-btn": 15591, + "-budget": 94752, + "-buffer": 31351, + "-build": 32145, + "-builder": 71814, + "-building": 51399, + "-built": 51614, + "-business": 64013, + "-but": 69796, + "-button": 12444, + "-buttons": 61667, + "-buy": 79678, + "-by": 14319, + "-byte": 53952, + "-c": 1786, + "-ca": 50777, + "-cache": 36680, + "-cal": 48136, + "-calendar": 46997, + "-call": 58897, + "-called": 18900, + "-camera": 71944, + "-campus": 93319, + "-can": 71214, + "-cancel": 85003, + "-cap": 36689, + "-capital": 97131, + "-caption": 58317, + "-car": 23992, + "-card": 20466, + "-care": 40106, + "-caret": 82922, + "-carousel": 52237, + "-cart": 35459, + "-case": 38485, + "-cat": 83892, + "-catching": 83324, + "-categories": 95099, + "-category": 42858, + "-ce": 53212, + "-cell": 31901, + "-cent": 21217, + "-center": 6883, + "-centered": 49382, + "-central": 84081, + "-centric": 64424, + "-century": 33357, + "-cert": 61034, + "-ch": 11582, + "-chain": 65186, + "-chair": 78681, + "-chan": 77635, + "-change": 50402, + "-changing": 48861, + "-channel": 53868, + "-char": 42538, + "-character": 79225, + "-chart": 44627, + "-chat": 60535, + "-che": 36740, + "-check": 15934, + "-checkbox": 49073, + "-checked": 99131, + "-chevron": 47382, + "-chief": 73619, + "-child": 23484, + "-choice": 62626, + "-ci": 62384, + "-cigaret": 62267, + "-cigarettes": 85063, + "-circle": 18664, + "-city": 53329, + "-cl": 30617, + "-class": 14800, + "-clean": 91800, + "-clear": 67027, + "-cli": 54797, + "-click": 28118, + "-client": 30011, + "-clock": 54288, + "-close": 34462, + "-cloud": 59197, + "-cluster": 93208, + "-cmpr": 97971, + "-cn": 63264, + "-co": 22471, + "-code": 25261, + "-coded": 86311, + "-col": 19459, + "-collapse": 24487, + "-collar": 96836, + "-collection": 95609, + "-color": 7889, + "-colored": 57722, + "-cols": 97814, + "-column": 24960, + "-columns": 73569, + "-com": 11476, + "-comm": 87289, + "-command": 48419, + "-comment": 45666, + "-comments": 77154, + "-commerce": 41585, + "-commercial": 73044, + "-commit": 96398, + "-common": 45045, + "-community": 89786, + "-comp": 62150, + "-company": 91033, + "-compatible": 80215, + "-complete": 74414, + "-component": 41387, + "-components": 24594, + "-compose": 65070, + "-con": 14859, + "-cond": 83376, + "-condition": 58005, + "-conf": 50717, + "-confidence": 90880, + "-config": 25130, + "-confirm": 66525, + "-connect": 85470, + "-connected": 72188, + "-cons": 68504, + "-conscious": 65566, + "-console": 74756, + "-consuming": 69740, + "-cont": 34072, + "-contact": 52996, + "-contained": 95224, + "-container": 12578, + "-containing": 92771, + "-content": 6808, + "-context": 63633, + "-contrib": 85444, + "-control": 4465, + "-controlled": 40270, + "-controller": 67245, + "-controls": 34251, + "-cookie": 87239, + "-coordinate": 80697, + "-copy": 66140, + "-cor": 45613, + "-core": 23460, + "-corner": 87595, + "-cost": 40138, + "-count": 26580, + "-counter": 63189, + "-country": 55871, + "-course": 68429, + "-court": 97362, + "-cover": 75365, + "-covered": 83308, + "-cr": 47589, + "-create": 39679, + "-created": 70957, + "-cross": 76429, + "-css": 47859, + "-cultural": 93426, + "-cur": 94049, + "-current": 44175, + "-custom": 36898, + "-cut": 42221, + "-cycle": 77690, + "-d": 1737, + "-da": 65897, + "-danger": 12025, + "-dark": 21838, + "-dashboard": 49136, + "-dat": 92997, + "-data": 13945, + "-date": 18413, + "-datepicker": 58564, + "-day": 11228, + "-days": 95214, + "-db": 60399, + "-dd": 21541, + "-de": 6810, + "-deals": 67548, + "-death": 96486, + "-debug": 71051, + "-decoration": 30266, + "-def": 28799, + "-default": 13672, + "-defense": 57698, + "-defined": 38717, + "-definition": 89050, + "-degree": 37675, + "-del": 47126, + "-delay": 45171, + "-delete": 40904, + "-dem": 76679, + "-demand": 56264, + "-demo": 58893, + "-den": 96528, + "-density": 88031, + "-depend": 35714, + "-dependent": 42818, + "-depth": 30310, + "-derived": 71186, + "-des": 24506, + "-desc": 52547, + "-described": 50604, + "-describedby": 64762, + "-description": 42830, + "-design": 46017, + "-designed": 68856, + "-desktop": 79249, + "-dess": 88329, + "-destruct": 96222, + "-det": 90904, + "-detail": 29572, + "-details": 30433, + "-dev": 25758, + "-devel": 64103, + "-develop": 47961, + "-development": 94606, + "-device": 65607, + "-di": 50289, + "-dialog": 20641, + "-digit": 48342, + "-dimensional": 32420, + "-dir": 45283, + "-direct": 43524, + "-directed": 85755, + "-direction": 33049, + "-directory": 53634, + "-dis": 10475, + "-disable": 21400, + "-disabled": 68067, + "-disc": 51665, + "-dismiss": 18298, + "-dismissible": 77633, + "-display": 52839, + "-dist": 87259, + "-distance": 73808, + "-div": 32785, + "-divider": 69900, + "-do": 29559, + "-doc": 11527, + "-document": 82233, + "-dollar": 53035, + "-dom": 14428, + "-domain": 72794, + "-dominated": 89623, + "-done": 94288, + "-door": 57090, + "-dot": 80902, + "-double": 68620, + "-down": 14875, + "-download": 62451, + "-dr": 18643, + "-drive": 82410, + "-driven": 31405, + "-driver": 76490, + "-driving": 59711, + "-drop": 61771, + "-dropdown": 43949, + "-du": 97205, + "-duration": 64383, + "-duty": 54518, + "-e": 5655, + "-earth": 85087, + "-east": 39507, + "-economic": 76149, + "-ed": 34435, + "-edge": 47348, + "-edit": 22142, + "-editor": 46467, + "-educated": 87681, + "-eff": 52365, + "-effect": 23067, + "-effective": 52321, + "-effects": 74788, + "-efficient": 72816, + "-eight": 69715, + "-el": 83149, + "-elect": 35981, + "-elected": 95705, + "-election": 42633, + "-element": 28452, + "-elements": 64445, + "-em": 36512, + "-email": 42117, + "-employed": 87884, + "-empty": 39433, + "-en": 20767, + "-enable": 64211, + "-enabled": 54192, + "-encoded": 92262, + "-end": 13068, + "-ended": 83075, + "-ending": 66797, + "-energy": 64387, + "-eng": 96102, + "-engine": 49625, + "-enh": 83082, + "-enter": 48215, + "-entity": 93588, + "-entry": 47244, + "-env": 67462, + "-envelope": 63525, + "-equ": 62259, + "-equipped": 87102, + "-equiv": 22941, + "-era": 51343, + "-error": 18917, + "-errors": 72604, + "-es": 31854, + "-eslint": 55175, + "-esque": 78958, + "-establish": 47487, + "-established": 63768, + "-esteem": 64806, + "-et": 96010, + "-ev": 90245, + "-even": 91531, + "-event": 39687, + "-events": 53803, + "-ever": 56440, + "-ex": 10187, + "-example": 43430, + "-exc": 96933, + "-exclusive": 93058, + "-exec": 70721, + "-existent": 59828, + "-existing": 49357, + "-exp": 18376, + "-expand": 49093, + "-expanded": 30036, + "-export": 65827, + "-expression": 81493, + "-ext": 66103, + "-extension": 80539, + "-extra": 71704, + "-eye": 46697, + "-eyed": 52513, + "-f": 2220, + "-face": 29088, + "-facebook": 53587, + "-faced": 76881, + "-facing": 63306, + "-factor": 74097, + "-family": 17190, + "-fashion": 57480, + "-fashioned": 63269, + "-fast": 73668, + "-fat": 63254, + "-fe": 33203, + "-feature": 91840, + "-fed": 78000, + "-feed": 86160, + "-feedback": 34874, + "-feira": 63908, + "-fetch": 80040, + "-fi": 36631, + "-fiction": 73431, + "-field": 19130, + "-fields": 79585, + "-figure": 98996, + "-file": 14203, + "-files": 46048, + "-fill": 58987, + "-filled": 43418, + "-filter": 32448, + "-fin": 83999, + "-final": 54862, + "-finals": 97908, + "-find": 75784, + "-fire": 56269, + "-fired": 71578, + "-first": 36943, + "-fit": 49560, + "-five": 35299, + "-fix": 70913, + "-fixed": 53764, + "-fl": 12276, + "-flag": 77950, + "-flash": 92879, + "-flat": 65468, + "-flex": 18120, + "-flight": 76203, + "-floating": 89651, + "-floor": 79350, + "-flow": 61313, + "-fluid": 17422, + "-fly": 82885, + "-focus": 79464, + "-focused": 51273, + "-fold": 23432, + "-folder": 74604, + "-follow": 92585, + "-font": 30671, + "-fontawesome": 95848, + "-food": 63467, + "-foot": 20483, + "-football": 84142, + "-footer": 19019, + "-for": 15193, + "-force": 80049, + "-form": 8460, + "-format": 38380, + "-formed": 94877, + "-forward": 44804, + "-found": 26704, + "-founded": 82108, + "-founder": 35827, + "-four": 41017, + "-fr": 50378, + "-frame": 46767, + "-framework": 73397, + "-free": 12577, + "-frequency": 78312, + "-friendly": 21896, + "-from": 38051, + "-front": 63626, + "-ft": 93624, + "-full": 23642, + "-function": 33749, + "-functional": 98516, + "-functions": 77430, + "-funded": 49601, + "-fw": 39239, + "-g": 2371, + "-gallery": 51238, + "-game": 19395, + "-gap": 88557, + "-gay": 83936, + "-gen": 36464, + "-general": 67997, + "-generated": 16185, + "-generation": 42827, + "-generator": 78002, + "-generic": 82682, + "-get": 22491, + "-girl": 62870, + "-git": 81749, + "-gl": 69650, + "-global": 73319, + "-gnu": 85716, + "-go": 19926, + "-goal": 92288, + "-going": 64812, + "-good": 71700, + "-google": 62645, + "-government": 45604, + "-gr": 24321, + "-grade": 40227, + "-gradient": 42738, + "-grand": 71002, + "-graph": 73016, + "-gray": 21840, + "-green": 38268, + "-grey": 63213, + "-grid": 28083, + "-ground": 78155, + "-group": 4351, + "-groups": 90352, + "-grow": 92732, + "-growing": 55557, + "-guard": 92981, + "-guid": 83692, + "-guide": 91425, + "-gun": 75235, + "-h": 2832, + "-haired": 67379, + "-half": 33802, + "-hand": 24413, + "-handed": 36140, + "-handle": 88407, + "-handler": 72112, + "-hard": 70745, + "-has": 45898, + "-hash": 96170, + "-haspopup": 61226, + "-have": 68484, + "-he": 37335, + "-head": 26390, + "-headed": 61066, + "-header": 9351, + "-heading": 23339, + "-health": 78405, + "-heart": 42244, + "-hearted": 69295, + "-heavy": 73566, + "-height": 17068, + "-held": 57506, + "-help": 49470, + "-helper": 77664, + "-hero": 91793, + "-hidden": 13306, + "-hide": 59882, + "-high": 27561, + "-highlight": 72802, + "-history": 61374, + "-hit": 69101, + "-holder": 65409, + "-hole": 86593, + "-home": 24388, + "-hook": 33860, + "-hooks": 78387, + "-hop": 48719, + "-horizontal": 37900, + "-host": 38589, + "-hot": 59447, + "-hour": 21231, + "-hours": 76005, + "-house": 35902, + "-hover": 28893, + "-how": 57423, + "-html": 53768, + "-http": 71312, + "-human": 68995, + "-i": 17600, + "-icon": 7815, + "-icons": 18868, + "-id": 12897, + "-ie": 87327, + "-if": 44637, + "-ignore": 43171, + "-il": 56275, + "-im": 36925, + "-image": 13746, + "-images": 50991, + "-img": 21064, + "-imm": 63583, + "-import": 53315, + "-important": 97373, + "-in": 3419, + "-inc": 85619, + "-inch": 24241, + "-income": 31097, + "-ind": 17776, + "-indent": 52675, + "-independent": 97785, + "-index": 21492, + "-induced": 37873, + "-inf": 47236, + "-inflammatory": 66495, + "-info": 12505, + "-information": 80739, + "-ing": 83905, + "-init": 54773, + "-initial": 47967, + "-initialized": 68028, + "-inline": 23633, + "-inner": 30899, + "-input": 13933, + "-ins": 21853, + "-insert": 74550, + "-inspired": 52061, + "-inst": 33688, + "-instagram": 90397, + "-install": 57899, + "-instance": 73655, + "-int": 20052, + "-intensive": 86992, + "-inter": 44894, + "-interest": 64773, + "-interface": 82954, + "-inv": 87486, + "-invalid": 39740, + "-inverse": 74299, + "-io": 86269, + "-ion": 79746, + "-ios": 88119, + "-ip": 74732, + "-is": 30430, + "-ish": 66753, + "-issue": 89365, + "-it": 37511, + "-item": 6404, + "-items": 17459, + "-j": 13333, + "-java": 85339, + "-job": 69948, + "-js": 25802, + "-json": 56080, + "-k": 12646, + "-key": 16173, + "-keys": 76109, + "-kind": 59706, + "-kit": 91476, + "-know": 99120, + "-known": 21309, + "-l": 2852, + "-la": 52826, + "-label": 6942, + "-labelled": 31801, + "-labelledby": 32127, + "-lang": 75460, + "-language": 43558, + "-large": 39148, + "-largest": 66967, + "-last": 60515, + "-lasting": 64165, + "-lat": 98414, + "-launch": 85306, + "-law": 30312, + "-layer": 47335, + "-layout": 37803, + "-le": 30207, + "-leading": 68375, + "-league": 87263, + "-leaning": 77361, + "-learning": 70756, + "-led": 33954, + "-left": 7950, + "-leg": 61123, + "-legged": 94264, + "-len": 85557, + "-length": 29325, + "-less": 54909, + "-letter": 79368, + "-level": 11591, + "-lfs": 70162, + "-lg": 7510, + "-li": 75546, + "-lib": 38641, + "-library": 52421, + "-license": 58967, + "-life": 25843, + "-light": 17709, + "-like": 12681, + "-limit": 71159, + "-line": 8447, + "-linear": 61299, + "-lined": 88959, + "-lines": 83363, + "-link": 8965, + "-linked": 54414, + "-links": 46676, + "-linux": 46375, + "-list": 9029, + "-lite": 78482, + "-liter": 85594, + "-live": 73451, + "-lived": 61854, + "-ln": 27692, + "-lnd": 46751, + "-lo": 26480, + "-load": 49008, + "-loaded": 78070, + "-loader": 21350, + "-loading": 58686, + "-local": 40060, + "-location": 60870, + "-lock": 41772, + "-log": 46332, + "-login": 35048, + "-logo": 33797, + "-long": 23791, + "-looking": 30248, + "-loop": 60666, + "-loss": 76301, + "-loving": 76600, + "-low": 59269, + "-m": 1448, + "-machine": 83215, + "-made": 26877, + "-mail": 11468, + "-mails": 60220, + "-main": 29992, + "-major": 65581, + "-make": 94219, + "-maker": 83601, + "-makers": 75070, + "-making": 27746, + "-man": 20477, + "-management": 54785, + "-manager": 44896, + "-many": 80188, + "-map": 25858, + "-mar": 69998, + "-margin": 45391, + "-mark": 28904, + "-marker": 69828, + "-market": 47731, + "-mask": 76056, + "-master": 50903, + "-match": 45704, + "-material": 67318, + "-max": 44073, + "-md": 4433, + "-me": 33973, + "-mean": 82996, + "-med": 44772, + "-media": 41262, + "-mediated": 80976, + "-medium": 45817, + "-member": 43550, + "-members": 85840, + "-memory": 64096, + "-men": 87969, + "-menu": 11714, + "-message": 29332, + "-messages": 88142, + "-meta": 54017, + "-metadata": 96431, + "-metal": 88832, + "-meter": 72501, + "-method": 50297, + "-mf": 47247, + "-middle": 50067, + "-mile": 42176, + "-million": 45049, + "-min": 13398, + "-minded": 33323, + "-mini": 67354, + "-minus": 71023, + "-minute": 23498, + "-mm": 48575, + "-mobile": 41079, + "-mod": 17078, + "-modal": 25120, + "-mode": 14982, + "-model": 28244, + "-module": 46718, + "-modules": 61079, + "-mon": 77296, + "-money": 67992, + "-monitor": 91634, + "-month": 22289, + "-more": 49914, + "-mort": 89553, + "-most": 62398, + "-motion": 84249, + "-mount": 59588, + "-mounted": 77730, + "-mouth": 91039, + "-move": 95694, + "-moving": 85472, + "-ms": 64089, + "-msg": 57462, + "-mult": 96927, + "-multi": 95669, + "-muted": 30740, + "-my": 61496, + "-même": 97275, + "-n": 5279, + "-na": 93082, + "-name": 11494, + "-names": 89044, + "-national": 74529, + "-native": 15598, + "-nav": 16818, + "-navbar": 52332, + "-navigation": 38652, + "-ne": 39978, + "-neck": 95433, + "-needed": 68775, + "-negative": 60935, + "-net": 66496, + "-network": 56732, + "-neutral": 91222, + "-new": 22269, + "-news": 45251, + "-next": 22304, + "-ng": 82963, + "-night": 44327, + "-nil": 83248, + "-nine": 85603, + "-no": 28366, + "-node": 39054, + "-non": 98700, + "-none": 24887, + "-normal": 52083, + "-not": 29169, + "-notch": 77511, + "-note": 67238, + "-notes": 61867, + "-notification": 90724, + "-now": 90963, + "-null": 60341, + "-num": 31194, + "-number": 25854, + "-o": 16022, + "-ob": 92449, + "-object": 40432, + "-of": 8668, + "-off": 12462, + "-office": 81127, + "-offs": 63939, + "-offset": 23761, + "-offsetof": 54187, + "-ok": 74253, + "-old": 6284, + "-olds": 64148, + "-on": 10326, + "-one": 18589, + "-online": 68879, + "-only": 15382, + "-ons": 59126, + "-op": 29492, + "-opacity": 68059, + "-open": 25686, + "-opening": 85208, + "-oper": 58555, + "-operation": 66234, + "-operative": 88993, + "-operator": 94687, + "-opt": 56112, + "-option": 48689, + "-options": 39733, + "-or": 26511, + "-orange": 84383, + "-order": 23810, + "-orders": 85051, + "-org": 71184, + "-oriented": 35085, + "-origin": 66803, + "-original": 66958, + "-os": 82428, + "-other": 72472, + "-out": 9794, + "-outline": 29641, + "-output": 59524, + "-outs": 83975, + "-over": 28252, + "-overlay": 51495, + "-owned": 28699, + "-owner": 69809, + "-p": 2268, + "-paced": 64219, + "-pack": 56510, + "-package": 44135, + "-packages": 75761, + "-packed": 69904, + "-pad": 92819, + "-padding": 42549, + "-page": 15112, + "-pages": 72347, + "-pagination": 80750, + "-paid": 69020, + "-pane": 38452, + "-panel": 31376, + "-paper": 76719, + "-par": 56998, + "-param": 72966, + "-parameter": 89412, + "-parent": 46436, + "-parse": 85382, + "-parser": 36620, + "-part": 28037, + "-parts": 92478, + "-party": 24031, + "-pass": 46967, + "-password": 33069, + "-path": 33095, + "-pattern": 79569, + "-pay": 70088, + "-paying": 88723, + "-payment": 72443, + "-pdf": 97625, + "-pe": 95651, + "-peer": 78597, + "-pencil": 56066, + "-per": 17018, + "-percent": 70624, + "-performance": 57474, + "-period": 98081, + "-person": 28045, + "-ph": 62737, + "-phase": 81610, + "-phone": 50480, + "-photo": 66367, + "-php": 84179, + "-pic": 82773, + "-picker": 61146, + "-picture": 79658, + "-piece": 55864, + "-pill": 88387, + "-pills": 90441, + "-pin": 74443, + "-pl": 32107, + "-place": 41661, + "-placeholder": 69098, + "-placement": 51563, + "-plan": 72981, + "-plane": 89549, + "-platform": 54025, + "-play": 26544, + "-player": 42367, + "-playing": 72085, + "-plugin": 23629, + "-plugins": 83746, + "-plus": 22934, + "-po": 74388, + "-pocket": 98336, + "-point": 16574, + "-pointer": 57689, + "-points": 92458, + "-pol": 53996, + "-policy": 66420, + "-pop": 40252, + "-popup": 47536, + "-port": 41457, + "-pos": 71782, + "-position": 33524, + "-positive": 68688, + "-post": 24410, + "-posts": 88571, + "-pound": 47630, + "-power": 26526, + "-powered": 40403, + "-pr": 26650, + "-pre": 21334, + "-prefix": 67181, + "-prepend": 84159, + "-pres": 79776, + "-present": 48024, + "-president": 65988, + "-pressure": 88461, + "-prev": 67012, + "-preview": 50881, + "-price": 37308, + "-priced": 84737, + "-primary": 9799, + "-print": 42145, + "-private": 65277, + "-pro": 9838, + "-process": 50094, + "-processing": 69193, + "-produ": 45254, + "-produced": 83461, + "-producing": 89275, + "-product": 28645, + "-production": 69566, + "-products": 67873, + "-prof": 92308, + "-profile": 24194, + "-profit": 27826, + "-program": 66020, + "-progress": 53291, + "-project": 33696, + "-prom": 65171, + "-proof": 59167, + "-prop": 76571, + "-properties": 67079, + "-property": 61401, + "-provider": 81532, + "-proxy": 83701, + "-ps": 68167, + "-pt": 96051, + "-public": 56471, + "-publish": 97353, + "-purple": 94462, + "-purpose": 58238, + "-push": 98643, + "-python": 71929, + "-q": 51608, + "-qu": 73701, + "-quality": 22092, + "-quarter": 57314, + "-quarters": 82541, + "-query": 65489, + "-question": 66415, + "-quote": 92929, + "-r": 3795, + "-ra": 58301, + "-rad": 94755, + "-radio": 62693, + "-radius": 17710, + "-random": 91675, + "-range": 30508, + "-ranging": 81005, + "-ranked": 75597, + "-ranking": 71439, + "-rate": 43026, + "-rated": 54885, + "-rating": 79511, + "-ray": 29530, + "-rays": 81717, + "-re": 5504, + "-reaching": 86148, + "-react": 30596, + "-read": 28806, + "-readable": 83218, + "-reader": 90479, + "-ready": 63915, + "-real": 74795, + "-rec": 59172, + "-record": 56163, + "-red": 31598, + "-redux": 19604, + "-ref": 43007, + "-reference": 72623, + "-refresh": 75086, + "-reg": 32163, + "-regexp": 94411, + "-region": 60679, + "-register": 63018, + "-registration": 98698, + "-rel": 47612, + "-related": 13904, + "-relative": 75706, + "-release": 44724, + "-remove": 48625, + "-ren": 93009, + "-render": 58933, + "-repeat": 33743, + "-reply": 78391, + "-report": 47411, + "-reported": 84196, + "-request": 44808, + "-required": 62472, + "-res": 11588, + "-reset": 76403, + "-resistant": 45956, + "-resolution": 62936, + "-resource": 74790, + "-resources": 89811, + "-response": 58469, + "-responsive": 20964, + "-rest": 76233, + "-result": 53838, + "-results": 64770, + "-ret": 69110, + "-return": 80023, + "-review": 45721, + "-reviewed": 78704, + "-rich": 40847, + "-right": 6701, + "-rights": 86556, + "-ring": 75929, + "-rise": 88399, + "-risk": 45470, + "-ro": 51934, + "-road": 65715, + "-rock": 79337, + "-role": 54821, + "-roll": 86760, + "-room": 47740, + "-root": 39214, + "-round": 21015, + "-rounded": 64218, + "-route": 81559, + "-router": 14266, + "-routing": 52156, + "-row": 19545, + "-rule": 59301, + "-run": 22973, + "-runner": 83056, + "-running": 53489, + "-runtime": 68912, + "-s": 1331, + "-safe": 55119, + "-sale": 57537, + "-sales": 95668, + "-sama": 95176, + "-sample": 83879, + "-san": 45938, + "-save": 48553, + "-saving": 57985, + "-sc": 30319, + "-scal": 61887, + "-scalable": 74252, + "-scale": 12934, + "-scenes": 79545, + "-schema": 79684, + "-school": 34689, + "-score": 75174, + "-screen": 29260, + "-script": 65960, + "-scripts": 94148, + "-scroll": 34785, + "-scrollbar": 92920, + "-sdk": 35478, + "-se": 7806, + "-search": 19428, + "-season": 30200, + "-seat": 78081, + "-second": 43863, + "-secondary": 29383, + "-secret": 68892, + "-section": 21599, + "-sectional": 96219, + "-sector": 88111, + "-security": 62202, + "-seeking": 98438, + "-select": 20458, + "-selected": 45240, + "-selection": 75705, + "-selector": 76127, + "-self": 41476, + "-selling": 47624, + "-sem": 69689, + "-semibold": 81916, + "-send": 74118, + "-sensitive": 56667, + "-separated": 72692, + "-ser": 25397, + "-series": 56332, + "-serif": 27566, + "-server": 26300, + "-service": 23461, + "-services": 58292, + "-serving": 85986, + "-session": 57181, + "-set": 24096, + "-setting": 60456, + "-settings": 40032, + "-setup": 78097, + "-seven": 78025, + "-sex": 26028, + "-sh": 7514, + "-shadow": 26866, + "-shaped": 34731, + "-share": 59720, + "-shared": 94908, + "-sharing": 54756, + "-sheet": 88966, + "-shell": 74862, + "-shift": 89943, + "-shirt": 33668, + "-shirts": 48660, + "-shop": 74914, + "-shopping": 78442, + "-short": 68651, + "-shot": 63530, + "-show": 30537, + "-side": 24067, + "-sidebar": 34128, + "-sided": 49758, + "-sign": 27953, + "-signed": 92553, + "-simple": 65957, + "-single": 56687, + "-site": 28554, + "-six": 54441, + "-size": 7174, + "-sized": 27835, + "-sizing": 60351, + "-sk": 45307, + "-sl": 58097, + "-slide": 36457, + "-slider": 37733, + "-slot": 87219, + "-sm": 4668, + "-small": 33759, + "-sn": 75450, + "-so": 33019, + "-social": 42029, + "-solid": 57253, + "-solving": 98146, + "-song": 76640, + "-sort": 46540, + "-source": 30774, + "-sp": 22238, + "-space": 27947, + "-spacing": 52373, + "-span": 64060, + "-spe": 52975, + "-speaking": 61190, + "-spec": 57794, + "-special": 92248, + "-specific": 18906, + "-speed": 29599, + "-spin": 80906, + "-spinner": 70110, + "-split": 78412, + "-sponsored": 63550, + "-spot": 97580, + "-square": 32947, + "-src": 48942, + "-st": 5477, + "-stack": 56090, + "-stage": 50156, + "-standard": 53820, + "-standing": 55603, + "-star": 20680, + "-stars": 96906, + "-start": 18935, + "-stat": 72354, + "-state": 20733, + "-states": 89060, + "-static": 49730, + "-stats": 80013, + "-status": 34998, + "-ste": 96326, + "-step": 29208, + "-stock": 87111, + "-stop": 41211, + "-storage": 62795, + "-store": 33252, + "-story": 45099, + "-str": 41628, + "-stream": 38723, + "-string": 30881, + "-strip": 31404, + "-striped": 32775, + "-strokes": 97408, + "-strong": 95757, + "-style": 11297, + "-su": 86129, + "-sub": 17967, + "-submit": 51215, + "-success": 13102, + "-suite": 91574, + "-sum": 75811, + "-summary": 65610, + "-sup": 58515, + "-support": 55947, + "-supported": 87113, + "-sur": 67706, + "-svg": 58569, + "-sw": 61879, + "-switch": 47196, + "-symbol": 87322, + "-sync": 65034, + "-syntax": 82779, + "-system": 36648, + "-t": 2385, + "-tab": 27411, + "-table": 16785, + "-tabs": 31741, + "-tag": 38204, + "-tags": 63814, + "-taking": 79571, + "-talk": 92680, + "-tank": 95028, + "-target": 18489, + "-task": 52579, + "-tax": 57044, + "-te": 48793, + "-team": 42842, + "-tech": 41257, + "-temp": 68190, + "-template": 33409, + "-term": 9663, + "-terminal": 90623, + "-terrorism": 94896, + "-test": 16839, + "-tested": 96073, + "-testid": 89252, + "-testing": 82155, + "-tests": 81094, + "-text": 9345, + "-th": 7563, + "-than": 47654, + "-that": 93500, + "-the": 10603, + "-theme": 33185, + "-themed": 56589, + "-thinking": 98951, + "-third": 29177, + "-thirds": 44626, + "-this": 60191, + "-thread": 60804, + "-threat": 55744, + "-threatening": 61899, + "-three": 48393, + "-through": 42747, + "-thumb": 66196, + "-thumbnail": 63867, + "-thumbnails": 96401, + "-ticket": 93222, + "-tier": 57255, + "-time": 7246, + "-times": 44448, + "-tip": 83229, + "-title": 8816, + "-tm": 70535, + "-to": 4686, + "-toast": 94634, + "-toggle": 9948, + "-toggler": 45294, + "-token": 34841, + "-ton": 74635, + "-tone": 97936, + "-too": 93579, + "-tool": 40224, + "-toolbar": 72790, + "-tools": 44646, + "-tooltip": 60646, + "-top": 8175, + "-topic": 85700, + "-total": 76229, + "-touch": 38926, + "-town": 82903, + "-tr": 10188, + "-tra": 79267, + "-track": 53466, + "-tracking": 93849, + "-trade": 82029, + "-trained": 68924, + "-training": 85370, + "-trans": 38060, + "-transfer": 92691, + "-transform": 32852, + "-transition": 54501, + "-transitional": 91330, + "-translate": 98183, + "-transparent": 74218, + "-trash": 42027, + "-treated": 87086, + "-tree": 40438, + "-trigger": 57987, + "-ts": 94501, + "-turn": 63951, + "-turned": 90446, + "-tw": 97562, + "-twitter": 59090, + "-two": 37402, + "-txt": 98273, + "-type": 10604, + "-types": 20817, + "-u": 45381, + "-ui": 8723, + "-ul": 64970, + "-un": 19892, + "-under": 71107, + "-unit": 60575, + "-unstyled": 80190, + "-unused": 63412, + "-up": 5239, + "-update": 38860, + "-upload": 47169, + "-upper": 53262, + "-uppercase": 59192, + "-ups": 26761, + "-uri": 87232, + "-url": 25443, + "-urlencoded": 44557, + "-us": 25682, + "-use": 24673, + "-used": 68521, + "-user": 8694, + "-users": 43380, + "-using": 92775, + "-util": 73446, + "-utils": 46029, + "-v": 8273, + "-val": 45154, + "-valid": 84810, + "-validate": 89376, + "-validation": 58346, + "-validator": 78574, + "-valu": 80019, + "-value": 19083, + "-values": 69050, + "-var": 85415, + "-variable": 38789, + "-vars": 80633, + "-vector": 65192, + "-ver": 98056, + "-vers": 93936, + "-version": 45367, + "-vertical": 88598, + "-ves": 92809, + "-vesm": 93973, + "-video": 41311, + "-view": 22503, + "-viol": 75727, + "-vis": 81389, + "-visible": 91136, + "-volume": 66768, + "-vous": 44225, + "-vs": 97874, + "-w": 2630, + "-wage": 95876, + "-wall": 85011, + "-war": 47160, + "-warning": 24073, + "-watch": 84296, + "-water": 53951, + "-wave": 88254, + "-way": 26798, + "-we": 61341, + "-web": 29431, + "-webpack": 58799, + "-week": 29509, + "-weight": 12635, + "-west": 37602, + "-wh": 55332, + "-wheel": 61627, + "-white": 15858, + "-wide": 24472, + "-widget": 28912, + "-widgets": 81108, + "-width": 9347, + "-wife": 82835, + "-win": 47689, + "-window": 41766, + "-windows": 97194, + "-wing": 28380, + "-winning": 26777, + "-wise": 44439, + "-with": 26189, + "-word": 37328, + "-work": 28621, + "-worker": 65516, + "-workers": 62284, + "-working": 63892, + "-world": 30084, + "-worthy": 96667, + "-wow": 71532, + "-wrap": 24208, + "-wrapper": 17388, + "-write": 62062, + "-writing": 96665, + "-written": 66283, + "-wsj": 28133, + "-www": 41242, + "-x": 6558, + "-xl": 30946, + "-xs": 9941, + "-y": 12034, + "-yard": 36739, + "-year": 4666, + "-years": 56088, + "-yellow": 72912, + "-you": 45417, + "-your": 52146, + "-yyyy": 75391, + "-z": 9141, + "-zA": 21088, + "-zero": 36929, + "-zone": 77763, + "-{": 63347, + "-être": 84715, + "-Ċ": 6913, + "-ĊĊ": 10452, + "-čĊ": 69127, + ".": 13, + ".!": 61350, + ".\"": 1189, + ".\"\"": 11124, + ".\"\"\"": 74000, + ".\"\"\"Ċ": 14442, + ".\"\"\"ĊĊ": 29184, + ".\"&": 83784, + ".\"'": 34177, + ".\"'\";Ċ": 92213, + ".\"',": 61144, + ".\"','\".$": 84491, + ".\")": 38739, + ".\"));Ċ": 54136, + ".\"),": 82815, + ".\"),Ċ": 62497, + ".\").": 98401, + ".\");": 55435, + ".\");Ċ": 7320, + ".\");ĊĊ": 30458, + ".\");čĊ": 27007, + ".\")]Ċ": 51507, + ".\")Ċ": 13053, + ".\")ĊĊ": 31225, + ".\")čĊ": 63173, + ".\"+": 67047, + ".\",": 10465, + ".\",\"": 47891, + ".\",Ċ": 10346, + ".\",čĊ": 57374, + ".\".": 36911, + ".\"/": 51561, + ".\";": 77355, + ".\";Ċ": 15297, + ".\";ĊĊ": 58724, + ".\";čĊ": 54203, + ".\"<": 55408, + ".\"Ċ": 96705, + "...\"Ċ": 30438, + "...\"ĊĊ": 47486, + "...'": 73338, + "...');Ċ": 67451, + "...')Ċ": 47456, + "...',": 68386, + "...',Ċ": 74740, + "...'Ċ": 80330, + "...(": 92056, + "...)": 32574, + "...)Ċ": 31218, + "...)ĊĊ": 61827, + "...,": 39768, + "....": 1934, + ".....": 18083, + "......": 28149, + ".......": 48611, + "........": 4007, + ".........": 60973, + "..........": 143207, + "...............": 135535, + "................": 7895, + "........................": 56241, + "................................": 16562, + "................................................................": 42269, + "......ĊĊ": 70031, + ".....ĊĊ": 80634, + "....Ċ": 77828, + "....ĊĊ": 20225, + "...": 38681, + "/\"Ċ": 29555, + "/\"ĊĊ": 85312, + "/#": 26559, + "/#{": 42847, + "/$": 10749, + "/$',": 45969, + "/${": 11254, + "/%": 12627, + "/&": 83372, + "/'": 11324, + "/')": 78664, + "/');Ċ": 55979, + "/')Ċ": 48174, + "/'+": 28407, + "/',": 14351, + "/',Ċ": 30493, + "/'.": 70017, + "/'.$": 19010, + "/';Ċ": 43886, + "/';ĊĊ": 95173, + "/'Ċ": 33021, + "/'ĊĊ": 79567, + "/(": 11884, + "/(?": 53780, + "/)": 50807, + "/)Ċ": 53560, + "/*": 1057, + "/*!": 18684, + "/*!Ċ": 24021, + "/**": 3663, + "/***": 69939, + "/********": 66721, + "/****************": 72300, + "/************************": 62622, + "/********************************": 25275, + "/****************************************": 72961, + "/************************************************": 78371, + "/********************************************************": 66159, + "/****************************************************************": 20159, + "/************************************************************************": 11370, + "/****************************************************************************": 22565, + "/********************************************************************************": 75632, + "/************************************************************************************************": 86544, + "/*******************************************************************************Ċ": 36713, + "/******************************************************************************/Ċ": 84999, + "/******************************************************************************Ċ": 53196, + "/*****************************************************************************Ċ": 68132, + "/****************************************************************************Ċ": 58007, + "/***************************************************************************Ċ": 70988, + "/******/": 23996, + "/******/Ċ": 76888, + "/***/": 21724, + "/**/*.": 52565, + "/**<": 69019, + "/**Ċ": 1747, + "/**ĊĊ": 89958, + "/**čĊ": 7534, + "/*------------------------------------------------": 73062, + "/*----------------------------------------------------------------": 29865, + "/*----------------------------------------------------------------------------": 57906, + "/*.": 23540, + "/*/": 91537, + "/*================================================================": 44438, + "/*@": 82594, + "/*Ċ": 3284, + "/*ĊĊ": 49760, + "/*čĊ": 16360, + "/+": 61651, + "/,": 34545, + "/,Ċ": 58658, + "/-": 23650, + "/--": 60487, + "/.": 11930, + "/../": 78380, + "/.Ċ": 75869, + "/.ĊĊ": 83900, + "//": 322, + "//!": 13650, + "//!Ċ": 51752, + "//\"": 77215, + "//#": 14206, + "//$": 20565, + "//'": 44539, + "//(": 95438, + "//*": 21417, + "//****************************************************************": 55263, + "//************************************************************************": 65080, + "//****************************************************************************": 64226, + "//**Ċ": 94714, + "//*[": 76330, + "//*[@": 37318, + "//-": 61463, + "//--": 77517, + "//----------------": 61683, + "//--------------------------------": 48256, + "//------------------------------------------------": 30889, + "//----------------------------------------------------------------": 9888, + "//----------------------------------------------------------------------------": 18012, + "//--------------------------------------------------------------------------------": 50561, + "//------------------------------------------------------------------------------------------------": 56342, + "//----------------------------------------------------------------------------------------------------------------": 89180, + "//------------------------------------------------------------------------------Ċ": 45814, + "//------------------------------------------------------------------------------ĊĊ": 87567, + "//-----------------------------------------------------------------------------Ċ": 44142, + "//----------------------------------------------------------------------------Ċ": 78938, + "//---------------------------------------------------------------------------Ċ": 57307, + "//---------------------------------------------------------------------------ĊĊ": 94464, + "//--------------------------------------------------------------Ċ": 81042, + "///": 2575, + "////": 725, + "/////": 97656, + "////////": 1145, + "////////////": 16862, + "////////////////": 1970, + "////////////////////////////////": 3898, + "////////////////////////////////////////////////": 27405, + "////////////////////////////////////////////////////": 82845, + "////////////////////////////////////////////////////////": 60908, + "////////////////////////////////////////////////////////////": 64575, + "////////////////////////////////////////////////////////////////": 9822, + "////////////////////////////////////////////////////////////////////": 43200, + "////////////////////////////////////////////////////////////////////////": 35109, + "////////////////////////////////////////////////////////////////////////////": 27183, + "////////////////////////////////////////////////////////////////////////////////": 28226, + "////////////////////////////////////////////////////////////////////////////////////////////////": 60629, + "////////////////////////////////////////////////////////////////////////////////Ċ": 58869, + "////////////////////////////////////////////////////////////////////////////////ĊĊ": 98320, + "///////////////////////////////////////////////////////////////////////////////Ċ": 76737, + "/////////////////////////////////////////////////////////////////////////////Ċ": 83894, + "////Ċ": 59867, + "///<": 43726, + "///Ċ": 15374, + "///ĊĊ": 47144, + "///čĊ": 63458, + "//:": 91129, + "//=": 68206, + "//===": 78712, + "//================================================": 76507, + "//================================================================": 23161, + "//================================================================================": 94303, + "//@": 36585, + "//[": 65637, + "//{{": 77146, + "//{Ċ": 47603, + "//}}": 92055, + "//}Ċ": 36323, + "//}ĊĊ": 55342, + "//Ċ": 2289, + "//ĊĊ": 6320, + "//ĊĊĊ": 84741, + "//čĊ": 14015, + "//čĊčĊ": 65870, + "/:": 11315, + "/;Ċ": 79795, + "/<": 23012, + "/": 9685, + "/>\";Ċ": 67708, + "/>.": 95908, + "/>.Ċ": 21005, + "/>.ĊĊ": 51278, + "/><": 32487, + "/>\\": 98408, + "/>Ċ": 10171, + "/>ĊĊ": 68601, + "/>čĊ": 48527, + "/?": 17763, + "/@": 38662, + "/A": 10360, + "/AFP": 68315, + "/AIDS": 92546, + "/AP": 59129, + "/API": 72030, + "/Add": 98619, + "/Admin": 56556, + "/App": 42746, + "/Application": 88204, + "/Area": 85341, + "/Auth": 86280, + "/B": 16276, + "/Base": 95383, + "/Branch": 76981, + "/Button": 46196, + "/C": 11295, + "/Card": 57489, + "/Common": 78086, + "/Core": 54164, + "/Create": 62050, + "/D": 14953, + "/DD": 81961, + "/DTD": 75614, + "/Data": 50239, + "/Delete": 63675, + "/Desktop": 48499, + "/Dk": 96620, + "/Documents": 54730, + "/E": 25646, + "/Edit": 59436, + "/Error": 96259, + "/F": 12318, + "/File": 64068, + "/Footer": 84080, + "/Form": 72820, + "/Foundation": 33925, + "/Framework": 80509, + "/G": 15792, + "/GL": 82323, + "/GPL": 84006, + "/Game": 84953, + "/Gate": 88820, + "/Getty": 20890, + "/Graphics": 66250, + "/Grid": 81696, + "/H": 23354, + "/Header": 59960, + "/Home": 40815, + "/I": 38151, + "/INFO": 48957, + "/IP": 84438, + "/Image": 81265, + "/Images": 91749, + "/Index": 88151, + "/Input": 76818, + "/Instruction": 76759, + "/Internal": 70435, + "/J": 31701, + "/K": 32854, + "/L": 7434, + "/LICENSE": 10845, + "/Layout": 87155, + "/Library": 59077, + "/Linux": 94325, + "/List": 76397, + "/Login": 54803, + "/M": 10270, + "/MIT": 85163, + "/MM": 37360, + "/MPL": 63519, + "/Main": 65241, + "/Math": 98717, + "/N": 20290, + "/New": 81249, + "/O": 17532, + "/OR": 93248, + "/Object": 68526, + "/Observable": 95655, + "/Open": 96662, + "/P": 16341, + "/Page": 94639, + "/Peak": 89724, + "/Private": 74777, + "/Product": 77738, + "/Public": 64533, + "/Q": 45274, + "/R": 19382, + "/Register": 81269, + "/Resources": 76099, + "/Runtime": 82636, + "/S": 11374, + "/Search": 78893, + "/Set": 51554, + "/Sh": 76202, + "/St": 84440, + "/String": 36264, + "/Sub": 49972, + "/Subthreshold": 57844, + "/System": 91274, + "/T": 16731, + "/TR": 52327, + "/Table": 79609, + "/Test": 80527, + "/The": 47748, + "/Typography": 83539, + "/U": 62010, + "/UI": 74883, + "/UIKit": 23679, + "/User": 33511, + "/Users": 97821, + "/V": 27233, + "/W": 22175, + "/Web": 50438, + "/WebAPI": 94897, + "/X": 32979, + "/XML": 48624, + "/XMLSchema": 53266, + "/Y": 39249, + "/YYYY": 81122, + "/Z": 71960, + "/[": 44510, + "/\\": 34319, + "/]": 73977, + "/^": 70867, + "/_": 19632, + "/__": 79463, + "/a": 14186, + "/about": 42516, + "/ac": 69471, + "/access": 72356, + "/account": 47920, + "/accounts": 83743, + "/action": 54765, + "/actions": 39121, + "/activity": 97618, + "/ad": 44460, + "/add": 19627, + "/address": 90665, + "/admin": 17402, + "/ag": 64031, + "/ajax": 24221, + "/al": 52120, + "/alert": 79688, + "/all": 31406, + "/am": 55417, + "/an": 55492, + "/android": 51825, + "/angular": 40051, + "/animate": 88283, + "/animations": 77394, + "/antlr": 69630, + "/ap": 24670, + "/apache": 38737, + "/api": 10508, + "/apimachinery": 72644, + "/apis": 66152, + "/app": 10640, + "/apple": 75603, + "/application": 33032, + "/apps": 50887, + "/apt": 86665, + "/ar": 95420, + "/arch": 59973, + "/archive": 71627, + "/arm": 88544, + "/array": 95838, + "/art": 55549, + "/article": 38181, + "/articles": 54034, + "/as": 72483, + "/assert": 71625, + "/assets": 21442, + "/at": 80730, + "/audio": 71247, + "/auth": 17369, + "/authentication": 90577, + "/auto": 91250, + "/autoload": 69863, + "/avatar": 58140, + "/aws": 41273, + "/ay": 83558, + "/ayushman": 86923, + "/b": 3470, + "/back": 59948, + "/backend": 70020, + "/background": 82445, + "/banner": 84676, + "/bar": 49513, + "/base": 26090, + "/bash": 17148, + "/basic": 77909, + "/be": 58689, + "/bg": 59552, + "/big": 90936, + "/bin": 8749, + "/bind": 84480, + "/bit": 86644, + "/bl": 89193, + "/black": 98545, + "/blob": 34827, + "/block": 59981, + "/blog": 37462, + "/blue": 97916, + "/board": 97430, + "/body": 61111, + "/book": 52284, + "/books": 71263, + "/boot": 88876, + "/bootstrap": 9841, + "/bower": 85056, + "/browse": 90281, + "/browser": 43023, + "/build": 30593, + "/bus": 96916, + "/business": 84828, + "/button": 54308, + "/buttons": 72180, + "/by": 67733, + "/c": 2899, + "/ca": 80591, + "/cache": 62094, + "/cal": 91734, + "/calendar": 76717, + "/callback": 91527, + "/car": 65659, + "/card": 61309, + "/cards": 70889, + "/cart": 56677, + "/cat": 92807, + "/catalog": 76098, + "/categories": 75930, + "/category": 46333, + "/cc": 87310, + "/cgi": 70416, + "/ch": 21284, + "/change": 81586, + "/channel": 73121, + "/chart": 73941, + "/chat": 71335, + "/check": 46619, + "/cl": 55931, + "/class": 25514, + "/classes": 47058, + "/cli": 82604, + "/client": 25085, + "/close": 94308, + "/cloud": 57800, + "/cm": 69198, + "/cmd": 83033, + "/cms": 97866, + "/co": 64942, + "/code": 46928, + "/color": 43898, + "/colors": 71293, + "/column": 89274, + "/com": 25093, + "/command": 83205, + "/comment": 73980, + "/comments": 48825, + "/common": 11356, + "/commons": 91567, + "/community": 92477, + "/company": 79943, + "/compiler": 82694, + "/component": 62958, + "/components": 20261, + "/con": 31236, + "/conf": 59241, + "/config": 14730, + "/configuration": 89442, + "/connect": 78489, + "/connection": 93406, + "/console": 96269, + "/constants": 45188, + "/contact": 45936, + "/container": 83897, + "/content": 27917, + "/contentassist": 93126, + "/context": 77766, + "/control": 82667, + "/controller": 58813, + "/controllers": 51646, + "/cop": 66659, + "/copyleft": 70953, + "/core": 5320, + "/count": 81043, + "/course": 84349, + "/cpp": 88201, + "/cpu": 77317, + "/create": 25577, + "/crypto": 84229, + "/cs": 91635, + "/css": 6710, + "/csv": 92128, + "/cupertino": 86955, + "/current": 74053, + "/custom": 34665, + "/customer": 70256, + "/d": 3446, + "/dashboard": 41161, + "/dat": 37566, + "/data": 13167, + "/dataTables": 59540, + "/database": 41605, + "/datatables": 47370, + "/date": 74564, + "/day": 44739, + "/db": 29357, + "/dc": 97333, + "/dd": 44881, + "/de": 22549, + "/debug": 67479, + "/default": 28989, + "/delete": 32275, + "/demo": 55160, + "/design": 83157, + "/desktop": 97579, + "/detail": 40823, + "/details": 59265, + "/dev": 35061, + "/device": 68365, + "/devices": 80397, + "/dialog": 52446, + "/dir": 88226, + "/dis": 41510, + "/disable": 87377, + "/disc": 95454, + "/display": 73329, + "/dist": 19537, + "/div": 22013, + "/do": 82638, + "/doc": 39510, + "/docker": 61764, + "/docs": 25967, + "/document": 46388, + "/documentation": 85586, + "/documents": 97090, + "/dom": 71344, + "/domain": 71972, + "/down": 79067, + "/download": 33630, + "/downloads": 89185, + "/dr": 48054, + "/drivers": 86647, + "/e": 16546, + "/ec": 73058, + "/edit": 21345, + "/editor": 75806, + "/effects": 91708, + "/el": 88658, + "/em": 62776, + "/email": 61649, + "/embed": 58653, + "/en": 13606, + "/end": 82846, + "/engine": 82045, + "/english": 71009, + "/entities": 65885, + "/entity": 66908, + "/env": 14358, + "/environment": 54482, + "/epl": 59303, + "/error": 41737, + "/errors": 57131, + "/es": 38428, + "/etc": 47047, + "/event": 40763, + "/events": 42764, + "/ex": 25815, + "/example": 65182, + "/examples": 67020, + "/exec": 90905, + "/exp": 65459, + "/export": 66948, + "/ext": 81143, + "/extensions": 72482, + "/f": 6663, + "/fa": 87562, + "/facebook": 75988, + "/false": 91233, + "/fast": 92351, + "/favicon": 48557, + "/features": 69313, + "/feed": 93730, + "/file": 23903, + "/filepath": 96442, + "/files": 33220, + "/filter": 63524, + "/find": 81121, + "/fire": 53010, + "/firebase": 66756, + "/fixtures": 94275, + "/fl": 58303, + "/flutter": 62593, + "/font": 25123, + "/fontawesome": 97537, + "/fonts": 60667, + "/foo": 60555, + "/footer": 26844, + "/form": 20260, + "/format": 92340, + "/forms": 25424, + "/forum": 62135, + "/fr": 80870, + "/frame": 67991, + "/framework": 81724, + "/free": 55694, + "/from": 91106, + "/front": 84067, + "/frontend": 73163, + "/fs": 73036, + "/full": 62975, + "/function": 70643, + "/functions": 56810, + "/fw": 60193, + "/fwlink": 62476, + "/g": 4846, + "/gallery": 67980, + "/game": 48191, + "/games": 88311, + "/gcc": 89848, + "/ge": 68748, + "/gen": 81753, + "/general": 65961, + "/generated": 79372, + "/get": 23302, + "/gif": 90831, + "/gin": 78624, + "/git": 60590, + "/github": 71120, + "/gl": 60662, + "/global": 38984, + "/go": 25525, + "/google": 40689, + "/goto": 92020, + "/gpio": 96649, + "/gpl": 37958, + "/gr": 40336, + "/graph": 72177, + "/graphql": 93308, + "/grid": 77341, + "/group": 53389, + "/groups": 77685, + "/grpc": 96047, + "/gtest": 70611, + "/gui": 55202, + "/h": 7530, + "/hash": 76784, + "/he": 77056, + "/head": 54312, + "/header": 24655, + "/help": 78930, + "/helper": 79285, + "/helpers": 41426, + "/her": 37907, + "/high": 58135, + "/history": 81883, + "/home": 18215, + "/hooks": 92519, + "/host": 90119, + "/how": 50326, + "/hr": 74901, + "/html": 13739, + "/http": 15627, + "/i": 16438, + "/ic": 76398, + "/icon": 29298, + "/icons": 18908, + "/id": 38146, + "/if": 91183, + "/il": 97565, + "/im": 72024, + "/image": 23349, + "/images": 9737, + "/img": 13670, + "/import": 87284, + "/in": 17996, + "/inc": 72388, + "/include": 26393, + "/includes": 63659, + "/index": 9022, + "/inet": 83081, + "/info": 54976, + "/init": 42948, + "/input": 39990, + "/install": 56951, + "/int": 31114, + "/interface": 47122, + "/interfaces": 59771, + "/internal": 30968, + "/io": 41390, + "/ion": 84853, + "/ioutil": 61246, + "/ip": 54919, + "/is": 46427, + "/issues": 38745, + "/item": 43996, + "/items": 61259, + "/j": 4437, + "/jav": 86357, + "/java": 47200, + "/javascript": 9837, + "/job": 77402, + "/jpeg": 51818, + "/jquery": 9576, + "/js": 9113, + "/json": 8931, + "/k": 14109, + "/kernel": 58343, + "/key": 68864, + "/kg": 44770, + "/kubernetes": 79587, + "/l": 13328, + "/lab": 94198, + "/lang": 25253, + "/language": 88436, + "/latest": 33149, + "/layout": 40010, + "/layouts": 89117, + "/left": 97089, + "/legal": 47036, + "/lg": 79020, + "/lgpl": 83850, + "/li": 88381, + "/lib": 8194, + "/library": 45446, + "/libs": 22755, + "/lic": 6580, + "/license": 43986, + "/licenses": 6971, + "/light": 76844, + "/link": 55431, + "/linux": 63844, + "/list": 20936, + "/lists": 98281, + "/live": 94212, + "/load": 71896, + "/loader": 93128, + "/loading": 83235, + "/local": 22270, + "/locale": 84039, + "/location": 91505, + "/log": 19413, + "/logger": 70277, + "/logging": 80074, + "/login": 16543, + "/logo": 28547, + "/logout": 56097, + "/logs": 77554, + "/m": 3183, + "/mL": 92322, + "/mac": 80716, + "/mail": 60819, + "/main": 15351, + "/mainwindow": 81898, + "/man": 72192, + "/manage": 81117, + "/manual": 73095, + "/map": 41081, + "/maps": 36793, + "/mark": 73598, + "/master": 23303, + "/mat": 84600, + "/material": 14831, + "/math": 59705, + "/max": 62599, + "/md": 74823, + "/me": 50899, + "/media": 29506, + "/member": 75062, + "/memory": 79671, + "/menu": 39983, + "/message": 50624, + "/messages": 71753, + "/met": 90228, + "/meta": 69207, + "/method": 91837, + "/min": 44173, + "/misc": 77686, + "/mit": 93847, + "/ml": 59548, + "/mm": 55180, + "/mobile": 69925, + "/mock": 90386, + "/mod": 38479, + "/modal": 74201, + "/model": 24882, + "/models": 20268, + "/module": 48185, + "/modules": 22903, + "/mol": 38871, + "/moment": 83947, + "/month": 47317, + "/movie": 71777, + "/mp": 62483, + "/ms": 58634, + "/msg": 80765, + "/music": 77770, + "/my": 34198, + "/mysql": 57515, + "/n": 9612, + "/name": 75992, + "/native": 75187, + "/nav": 68236, + "/navbar": 81806, + "/navigation": 83265, + "/ne": 89310, + "/net": 38065, + "/network": 47658, + "/new": 25376, + "/news": 27355, + "/ng": 44043, + "/nginx": 69261, + "/no": 33100, + "/node": 37211, + "/non": 91130, + "/not": 62441, + "/notification": 96542, + "/npm": 63920, + "/ns": 82492, + "/null": 19293, + "/o": 20271, + "/oauth": 60703, + "/object": 59489, + "/oct": 59182, + "/octet": 79832, + "/of": 96745, + "/off": 50806, + "/on": 86537, + "/op": 52000, + "/open": 37644, + "/operator": 68341, + "/operators": 46481, + "/opt": 87196, + "/options": 72102, + "/or": 5144, + "/order": 43395, + "/orders": 82818, + "/org": 41361, + "/original": 98137, + "/os": 57892, + "/out": 48316, + "/output": 47016, + "/owl": 68618, + "/p": 4322, + "/package": 64000, + "/packages": 72919, + "/page": 32180, + "/pages": 29078, + "/parser": 61770, + "/part": 48981, + "/pass": 67070, + "/password": 59470, + "/path": 50976, + "/pay": 72423, + "/payment": 81821, + "/pdf": 47751, + "/people": 87684, + "/per": 43985, + "/perl": 77122, + "/person": 56461, + "/pg": 98234, + "/ph": 89021, + "/photo": 51392, + "/photos": 49775, + "/php": 41773, + "/pi": 62228, + "/pkg": 22523, + "/pl": 11255, + "/place": 60775, + "/plain": 36971, + "/platform": 27294, + "/play": 81092, + "/player": 59503, + "/plugin": 51372, + "/plugins": 20797, + "/pm": 97989, + "/png": 35960, + "/pol": 92750, + "/pop": 91275, + "/popper": 48152, + "/port": 71377, + "/portfolio": 87203, + "/post": 29996, + "/posts": 51311, + "/power": 86696, + "/pp": 87146, + "/pr": 44672, + "/pre": 36259, + "/preferences": 75484, + "/print": 92221, + "/private": 81001, + "/pro": 18008, + "/problem": 79997, + "/problems": 49216, + "/process": 86780, + "/product": 24788, + "/products": 33480, + "/profile": 23580, + "/project": 40118, + "/projects": 39606, + "/prom": 76951, + "/proto": 97766, + "/provider": 74997, + "/providers": 82543, + "/pub": 59492, + "/public": 22041, + "/py": 90834, + "/python": 23266, + "/q": 33894, + "/qt": 89976, + "/qu": 78845, + "/query": 66762, + "/question": 94623, + "/questions": 43319, + "/r": 7382, + "/rand": 95978, + "/random": 79442, + "/raw": 75909, + "/rc": 95502, + "/re": 10758, + "/react": 27417, + "/read": 69199, + "/rec": 84094, + "/red": 77900, + "/ref": 55763, + "/reference": 65780, + "/reg": 86858, + "/register": 37837, + "/release": 88648, + "/releases": 58187, + "/rem": 81771, + "/remove": 74930, + "/render": 83201, + "/renderer": 89429, + "/report": 62385, + "/repos": 49505, + "/repository": 97548, + "/request": 53007, + "/res": 32426, + "/reset": 70003, + "/resource": 53100, + "/resources": 38900, + "/respond": 66463, + "/response": 98804, + "/rest": 45265, + "/result": 82907, + "/results": 77174, + "/rfc": 93982, + "/right": 73101, + "/root": 72074, + "/ros": 83846, + "/router": 18546, + "/routes": 62970, + "/rs": 91885, + "/rss": 85291, + "/rules": 95489, + "/run": 48385, + "/runtime": 39298, + "/s": 2687, + "/sample": 69851, + "/save": 69027, + "/sbin": 72435, + "/sc": 33474, + "/schema": 61142, + "/screen": 97990, + "/screens": 59386, + "/script": 55678, + "/scripts": 39372, + "/sdk": 79760, + "/se": 59587, + "/search": 23167, + "/sec": 60071, + "/security": 71422, + "/select": 37346, + "/self": 68239, + "/send": 69736, + "/server": 37255, + "/service": 34186, + "/services": 22258, + "/session": 54760, + "/set": 47267, + "/settings": 29866, + "/settingsdialog": 68579, + "/setup": 77443, + "/sh": 14688, + "/share": 33914, + "/shared": 25948, + "/she": 56595, + "/shop": 61877, + "/show": 34175, + "/sidebar": 72907, + "/sign": 35763, + "/signup": 68763, + "/simple": 67195, + "/single": 96562, + "/site": 47881, + "/sites": 67502, + "/sl": 82577, + "/slick": 96224, + "/slider": 90739, + "/sm": 74037, + "/small": 96816, + "/sn": 71591, + "/social": 91508, + "/socket": 55480, + "/software": 78727, + "/source": 54373, + "/sources": 96690, + "/sp": 26734, + "/span": 90345, + "/spec": 45389, + "/sql": 59682, + "/src": 13437, + "/ss": 87602, + "/st": 14272, + "/star": 79264, + "/start": 69976, + "/stat": 42525, + "/state": 63796, + "/static": 36643, + "/stats": 94933, + "/status": 32518, + "/std": 21546, + "/stdc": 24121, + "/storage": 53345, + "/store": 31320, + "/story": 78389, + "/stream": 87434, + "/stretch": 75116, + "/stretchr": 76519, + "/string": 54544, + "/student": 82031, + "/style": 28717, + "/styles": 27788, + "/sub": 37885, + "/support": 60370, + "/svg": 32800, + "/sw": 42934, + "/swagger": 80930, + "/sweetalert": 92481, + "/sys": 41836, + "/system": 28139, + "/t": 5523, + "/tab": 78859, + "/table": 45326, + "/tag": 76196, + "/tags": 84460, + "/target": 88868, + "/task": 58186, + "/tasks": 81337, + "/tcp": 95958, + "/team": 78015, + "/temp": 70445, + "/template": 37094, + "/templates": 38941, + "/terms": 96763, + "/test": 12697, + "/testify": 76702, + "/testing": 28801, + "/tests": 62468, + "/text": 36271, + "/th": 20439, + "/the": 51257, + "/theme": 40081, + "/themes": 37065, + "/thread": 72144, + "/thumb": 79941, + "/time": 35263, + "/tiny": 72405, + "/tinyos": 84369, + "/title": 81946, + "/tmp": 57008, + "/to": 32429, + "/todo": 93214, + "/token": 54386, + "/tool": 75027, + "/tools": 45714, + "/top": 35282, + "/topic": 92969, + "/topics": 57662, + "/tos": 80945, + "/tr": 21485, + "/train": 50838, + "/trans": 80121, + "/tree": 62978, + "/trunk": 78680, + "/ts": 99046, + "/tty": 84861, + "/tutorial": 85252, + "/twitter": 82228, + "/type": 49611, + "/types": 20384, + "/u": 33566, + "/ubuntu": 79926, + "/ui": 22444, + "/umd": 83199, + "/un": 35217, + "/unit": 78311, + "/up": 71843, + "/update": 29832, + "/upload": 37173, + "/uploads": 29581, + "/url": 57254, + "/us": 62431, + "/use": 62272, + "/user": 11739, + "/users": 19178, + "/usr": 54338, + "/util": 22610, + "/utility": 95726, + "/utils": 21902, + "/v": 5457, + "/validation": 97800, + "/value": 57542, + "/values": 96985, + "/var": 92146, + "/vector": 80382, + "/vendor": 35663, + "/vendors": 72029, + "/version": 64413, + "/video": 41303, + "/videos": 72945, + "/view": 22489, + "/views": 20370, + "/vnd": 42701, + "/vue": 71597, + "/w": 6324, + "/wait": 86103, + "/watch": 26213, + "/we": 53686, + "/weather": 77758, + "/web": 21808, + "/welcome": 86697, + "/widget": 33336, + "/widgets": 42836, + "/wiki": 25502, + "/win": 97718, + "/window": 79243, + "/windows": 75349, + "/work": 35062, + "/workspace": 93381, + "/world": 78892, + "/wp": 30746, + "/write": 64264, + "/ws": 91021, + "/www": 26550, + "/x": 10776, + "/xhtml": 34132, + "/xml": 36524, + "/y": 26491, + "/year": 84900, + "/yyyy": 42651, + "/z": 31082, + "/{": 9388, + "/{$": 78658, + "/{{": 66839, + "/{{$": 68665, + "/{}/": 92509, + "/~": 53796, + "/Ċ": 5894, + "/ĊĊ": 8680, + "/ĊĊĊ": 75258, + "/ĊĊĊĊ": 75960, + "/čĊ": 29110, + "/čĊčĊ": 72534, + "0": 15, + "1": 16, + "2": 17, + "3": 18, + "4": 19, + "5": 20, + "6": 21, + "7": 22, + "8": 23, + "9": 24, + ":": 25, + ":!": 98324, + ":\"": 2974, + ":\"\",": 83131, + ":\"\",Ċ": 79389, + ":\"#": 79729, + ":\")": 90220, + ":\");Ċ": 24305, + ":\");ĊĊ": 70571, + ":\");čĊ": 67618, + ":\")Ċ": 34403, + ":\"+": 28798, + ":\",": 12147, + ":\",Ċ": 55120, + ":\"-": 86789, + ":\"-\"`Ċ": 98899, + ":\".": 95740, + ":\".$": 91920, + ":\";Ċ": 28669, + ":\";čĊ": 83359, + ":\"<<": 69034, + ":\"Ċ": 34300, + ":#": 14111, + ":$": 21701, + ":${": 37454, + ":%": 7533, + ":&": 35218, + ":'": 4884, + ":'#": 54602, + ":''": 95727, + ":'',": 96403, + ":'',Ċ": 54443, + ":')": 79895, + ":');Ċ": 93728, + ":')Ċ": 53781, + ":'+": 75316, + ":',": 17440, + ":',Ċ": 43355, + ":'.$": 75984, + ":'/": 70282, + ":';Ċ": 85568, + ":'Ċ": 91088, + ":(": 3269, + ":)": 28489, + ":)])": 95804, + ":)];Ċ": 89036, + ":)Ċ": 82630, + ":*": 53386, + ":**": 66963, + ":+": 12225, + ":,": 41239, + ":-": 10944, + ":-------------": 96178, + ":?": 76475, + ":@": 14882, + ":@\"": 5466, + ":@\"\"": 75306, + ":@\"%": 29752, + ":@\"%@": 54721, + ":@\"%@\",": 77741, + ":@{": 88877, + ":A": 55648, + ":Add": 54676, + ":Any": 97192, + ":Array": 49033, + ":B": 81268, + ":Boolean": 46084, + ":C": 55992, + ":CGPoint": 89369, + ":CGRect": 26651, + ":CGRectMake": 28509, + ":D": 63843, + ":E": 84488, + ":Event": 93164, + ":F": 69856, + ":Get": 21083, + ":H": 94192, + ":I": 57155, + ":Int": 43979, + ":Is": 46551, + ":L": 69233, + ":M": 65207, + ":N": 37694, + ":NO": 37457, + ":NS": 23787, + ":NSLayout": 89752, + ":NSLocalizedString": 93095, + ":NSMakeRange": 90282, + ":NSUTF": 75121, + ":Number": 35004, + ":Object": 65292, + ":P": 92411, + ":Register": 65339, + ":S": 75369, + ":Set": 15606, + ":SetPoint": 90305, + ":SetText": 95098, + ":String": 15816, + ":System": 69112, + ":T": 68661, + ":The": 74045, + ":UI": 16487, + ":UIAlert": 91507, + ":UIButtonType": 86261, + ":UIButtonTypeCustom": 98159, + ":UIControl": 24094, + ":UIControlEvent": 51468, + ":UIControlEventTouchUpInside": 57828, + ":UIControlState": 30959, + ":UIControlStateNormal": 34029, + ":UITableView": 66099, + ":X": 83530, + ":YES": 19566, + ":[": 7259, + ":[\"": 81262, + ":['": 69287, + ":[[": 52192, + ":[],Ċ": 90415, + ":[Ċ": 64772, + ":\\": 7190, + ":\\\"": 75035, + ":\\/\\/": 15325, + ":\\\\": 23817, + ":]": 28283, + ":])": 95849, + ":])Ċ": 43057, + ":],": 45499, + ":].": 72559, + ":]:Ċ": 75781, + ":]Ċ": 24135, + ":]ĊĊ": 64631, + ":^": 37933, + ":^(": 20281, + ":^{Ċ": 49348, + ":_": 22035, + ":`": 18736, + ":`~": 77802, + ":a": 43833, + ":absolute": 62239, + ":add": 38488, + ":aload": 60005, + ":animated": 59400, + ":any": 31466, + ":async": 92803, + ":auto": 57764, + ":b": 43973, + ":before": 79157, + ":bg": 68149, + ":black": 70686, + ":block": 57628, + ":bold": 49722, + ":boolean": 41937, + ":border": 97128, + ":both": 95716, + ":c": 48031, + ":center": 27678, + ":checked": 97381, + ":class": 91964, + ":convert": 95615, + ":create": 58584, + ":d": 40422, + ":data": 41784, + ":date": 93263, + ":def": 22582, + ":description": 93323, + ":disable": 42066, + ":e": 55976, + ":end": 25789, + ":eq": 62671, + ":error": 69195, + ":event": 18254, + ":expr": 96011, + ":f": 55477, + ":false": 22110, + ":first": 53374, + ":flex": 86885, + ":flutter": 16784, + ":focus": 88358, + ":frame": 77674, + ":function": 17934, + ":g": 70418, + ":get": 24907, + ":green": 95731, + ":h": 69651, + ":hidden": 72631, + ":host": 72361, + ":hover": 32721, + ":href": 95797, + ":http": 69459, + ":i": 14665, + ":id": 39125, + ":image": 37670, + ":index": 46088, + ":indexPath": 42293, + ":init": 48952, + ":initComponents": 56373, + ":inline": 67942, + ":int": 14337, + ":invoke": 94584, + ":item": 78852, + ":j": 57558, + ":k": 57071, + ":key": 64849, + ":l": 88397, + ":last": 67692, + ":left": 34810, + ":len": 77899, + ":length": 29296, + ":list": 78079, + ":m": 30286, + ":maj": 59873, + ":max": 82494, + ":message": 91473, + ":min": 83675, + ":mm": 20133, + ":model": 65615, + ":mysql": 85276, + ":n": 19266, + ":name": 57520, + ":new": 44755, + ":nil": 15002, + ":no": 66479, + ":none": 21869, + ":normal": 91177, + ":not": 67844, + ":nth": 37964, + ":null": 53098, + ":num": 55268, + ":number": 34040, + ":numel": 91081, + ":o": 93557, + ":on": 94876, + ":p": 44389, + ":param": 67316, + ":params": 95011, + ":path": 71796, + ":pk": 94541, + ":pointer": 71140, + ":r": 67359, + ":red": 41584, + ":relative": 84566, + ":req": 69739, + ":request": 81995, + ":result": 95161, + ":return": 66302, + ":right": 42352, + ":s": 14495, + ":selected": 54354, + ":self": 10899, + ":semicolon": 46717, + ":set": 23854, + ":size": 41621, + ":ss": 25734, + ":start": 87090, + ":str": 61296, + ":string": 22423, + ":t": 60777, + ":test": 84476, + ":text": 34952, + ":this": 37612, + ":title": 58649, + ":true": 18251, + ":type": 46445, + ":uint": 62921, + ":update": 91809, + ":url": 35730, + ":user": 64280, + ":utf": 55022, + ":v": 52649, + ":value": 62856, + ":variables": 56205, + ":view": 77380, + ":void": 28949, + ":w": 52882, + ":white": 50816, + ":x": 35254, + ":y": 69203, + ":{": 12547, + ":{}": 79411, + ":{Ċ": 17894, + ":{čĊ": 83051, + ":Ċ": 510, + ":ĊĊ": 1447, + ":ĊĊĊ": 24391, + ":ĊĊĊĊ": 14223, + ":ĊĊĊĊĊĊ": 47446, + ":čĊ": 2834, + ":čĊčĊ": 31980, + ";": 26, + ";!": 98473, + ";\"": 10892, + ";\");Ċ": 34649, + ";\")Ċ": 91812, + ";\",": 32503, + ";\",Ċ": 64497, + ";\";Ċ": 39678, + ";\">": 12350, + ";\"><": 29652, + ";\">Ċ": 12143, + ";\">čĊ": 49048, + ";\"Ċ": 41655, + ";$": 25675, + ";%": 38402, + ";&": 8009, + ";&#": 50222, + ";'": 35994, + ";');Ċ": 95674, + ";',": 47186, + ";',Ċ": 45385, + ";';Ċ": 69646, + ";'>": 46639, + ";'Ċ": 98319, + ";(": 57709, + ";)": 36174, + ";*/Ċ": 62403, + ";++": 36507, + ";,": 71554, + ";-": 54433, + ";.": 78698, + ";/": 20560, + ";/*": 86229, + ";//": 9361, + ";;": 6768, + ";;;": 36328, + ";;;;": 22789, + ";;;;;;;;": 40300, + ";;;;;;;;;;;;;;;;": 71409, + ";;Ċ": 28207, + ";;ĊĊ": 72931, + ";<": 35058, + ";": 24904, + ";?>\"": 52736, + ";?>Ċ": 39569, + ";\\": 17882, + ";\\\">": 93162, + ";\\Ċ": 55783, + ";]/": 90838, + ";a": 48222, + ";amp": 19933, + ";b": 54933, + ";background": 49954, + ";base": 81860, + ";border": 35779, + ";br": 78692, + ";break": 86879, + ";c": 78632, + ";charset": 45017, + ";color": 39691, + ";d": 99149, + ";display": 85565, + ";element": 74162, + ";font": 27056, + ";height": 43041, + ";i": 4696, + ";if": 81004, + ";j": 13061, + ";k": 34996, + ";l": 81628, + ";left": 90350, + ";line": 85173, + ";m": 67236, + ";margin": 31927, + ";n": 46703, + ";o": 93449, + ";p": 55913, + ";padding": 35129, + ";q": 60899, + ";r": 85308, + ";s": 39414, + ";set": 80389, + ";t": 53747, + ";text": 44385, + ";top": 81754, + ";width": 39336, + ";x": 54621, + ";y": 70863, + ";z": 96712, + ";}": 30031, + ";};Ċ": 86637, + ";}Ċ": 12084, + ";}ĊĊ": 33498, + ";}čĊ": 39528, + ";Ċ": 280, + ";ĊĊ": 401, + ";ĊĊĊ": 5210, + ";ĊĊĊĊ": 22116, + ";ĊĊĊĊĊ": 69494, + ";čĊ": 463, + ";čĊčĊ": 1926, + ";čĊčĊčĊ": 22367, + ";čĊčĊčĊčĊ": 86232, + ";ččĊ": 44122, + ";čččĊ": 67796, + "<": 27, + "": 71698, + "<*": 78822, + "<-": 45342, + "": 128247, + "<::": 87398, + "<<": 2442, + "<<\"": 7753, + "<<\"\\": 28777, + "<<(": 32484, + "<<<": 12603, + "<<<<": 35684, + "<<<<<<<": 66838, + "<<<<<<<<": 82128, + "<=": 8203, + "<=$": 87990, + "<=(": 80961, + "<>": 21122, + "<>(": 20591, + "<>(\"": 89039, + "<>();Ċ": 10874, + "<>();ĊĊ": 28389, + "<>();čĊ": 50721, + "<>Ċ": 62537, + "": 19542, + ">": 74094, + "": 128245, + "": 128244, + "Ċ": 76216, + "=\"\";Ċ": 26794, + "=\"\";čĊ": 67924, + "=\"\">": 18261, + "=\"\"><": 49570, + "=\"\">Ċ": 17419, + "=\"\">čĊ": 77962, + "=\"\"Ċ": 33418, + "=\"#": 4649, + "=\"#\"": 13395, + "=\"#\">": 15250, + "=\"#\"><": 27590, + "=\"#\">Ċ": 39945, + "=\"$": 20227, + "=\"$(": 46434, + "=\"${": 20451, + "=\"%": 35220, + "=\"'": 15509, + "=\"'+": 35970, + "=\"'.": 37317, + "=\"'.$": 22469, + "=\"(": 78239, + "=\").": 75244, + "=\");Ċ": 45219, + "=\"+": 14095, + "=\",": 21369, + "=\"-": 23948, + "=\"--": 98298, + "=\".": 35226, + "=\".$": 26332, + "=\".$_": 96652, + "=\"../": 6510, + "=\"../../": 15437, + "=\"../../../": 70412, + "=\"./": 26420, + "=\"/": 5914, + "=\"/\"": 44609, + "=\"/\">": 68939, + "=\"/\">Ċ": 76977, + "=\"//": 44166, + "=\";Ċ": 93175, + "=\"<": 34476, + "=\"<<": 80476, + "=\"": 84653, + "==Ċ": 20747, + "=>": 2179, + "=>\"": 44488, + "=>$": 29798, + "=>'": 39526, + "=>{Ċ": 13135, + "=?": 19884, + "=?\",": 87873, + "=?\";Ċ": 89212, + "=?,": 52650, + "=@": 33678, + "=@\"": 94806, + "=A": 46623, + "=B": 54526, + "=BitConverter": 79308, + "=C": 40917, + "=D": 48634, + "=DB": 89649, + "=E": 85234, + "=F": 82226, + "=False": 5608, + "=G": 92009, + "=Integer": 87643, + "=L": 68851, + "=M": 49682, + "=Math": 76706, + "=N": 51995, + "=NULL": 27980, + "=None": 5856, + "=P": 76778, + "=R": 72677, + "=S": 72196, + "=T": 60461, + "=True": 3618, + "=UTF": 37005, + "=User": 94255, + "=Value": 88048, + "=W": 69547, + "=X": 61345, + "=Y": 91994, + "=YES": 94834, + "=[": 5818, + "=[\"": 27965, + "=['": 13987, + "=[[": 45937, + "=[]": 30692, + "=[],": 66221, + "=[];Ċ": 39952, + "=[]Ċ": 27184, + "=[]čĊ": 73927, + "=[Ċ": 33199, + "=\\": 34433, + "=\\\"": 4070, + "=\\\"\"": 41445, + "=\\\"\";Ċ": 90983, + "=\\\"#": 67915, + "=\\\"$": 56113, + "=\\\"%": 75559, + "=\\\"/": 98185, + "=\\'": 55761, + "=_": 20904, + "=_(\"": 91347, + "=_('": 89966, + "=__": 87423, + "=`": 22250, + "=a": 24239, + "=add": 92784, + "=admin": 95890, + "=all": 88277, + "=args": 31258, + "=array": 24529, + "=ax": 71663, + "=b": 22086, + "=back": 92372, + "=batch": 71798, + "=bool": 89112, + "=c": 19535, + "=center": 74144, + "=color": 63792, + "=com": 58846, + "=config": 46290, + "=context": 70694, + "=count": 73916, + "=create": 91087, + "=current": 71923, + "=cut": 48715, + "=cv": 67470, + "=d": 25405, + "=data": 22839, + "=date": 52231, + "=datetime": 56749, + "=db": 57752, + "=default": 92628, + "=device": 19718, + "=df": 60884, + "=dict": 39597, + "=document": 33541, + "=e": 40391, + "=edge": 35940, + "=email": 76371, + "=en": 61757, + "=end": 61521, + "=event": 95260, + "=explode": 91470, + "=f": 18111, + "=false": 12219, + "=file": 59245, + "=find": 60517, + "=findViewById": 84891, + "=float": 33815, + "=fopen": 74087, + "=form": 51374, + "=format": 91662, + "=forms": 64742, + "=function": 29486, + "=g": 37162, + "=get": 28280, + "=h": 54113, + "=head": 23770, + "=headers": 54085, + "=http": 34717, + "=https": 51999, + "=i": 21397, + "=id": 46022, + "=image": 74655, + "=img": 93288, + "=in": 67384, + "=index": 62641, + "=input": 39067, + "=int": 16563, + "=is": 92310, + "=item": 32562, + "=j": 45612, + "=json": 37507, + "=k": 45157, + "=key": 86582, + "=l": 40627, + "=label": 98689, + "=lambda": 27504, + "=len": 45819, + "=length": 94362, + "=line": 78185, + "=list": 40972, + "=localhost": 84551, + "=log": 83304, + "=logging": 70228, + "=m": 27221, + "=max": 44617, + "=message": 87129, + "=min": 61506, + "=model": 62496, + "=models": 27410, + "=msg": 97378, + "=my": 80470, + "=mysql": 33991, + "=mysqli": 30772, + "=n": 21747, + "=name": 39241, + "=new": 8342, + "=nil": 63358, + "=no": 30315, + "=node": 79158, + "=np": 17418, + "=null": 18911, + "=num": 46901, + "=o": 85876, + "=obj": 77246, + "=open": 52965, + "=options": 95031, + "=os": 50864, + "=out": 84998, + "=output": 85947, + "=p": 17385, + "=params": 57643, + "=password": 73517, + "=path": 81940, + "=pd": 67295, + "=pk": 87185, + "=plt": 95933, + "=pos": 99138, + "=post": 82150, + "=q": 63048, + "=query": 79194, + "=r": 31120, + "=rand": 95595, + "=random": 79817, + "=re": 49652, + "=read": 89871, + "=req": 84016, + "=request": 25168, + "=require": 62833, + "=res": 78370, + "=response": 79090, + "=result": 74212, + "=root": 74983, + "=row": 65529, + "=s": 14149, + "=sc": 74875, + "=search": 96598, + "=self": 7667, + "=session": 83400, + "=set": 68800, + "=settings": 72652, + "=size": 74178, + "=start": 55622, + "=status": 55118, + "=str": 15932, + "=sub": 45399, + "=subprocess": 56594, + "=sum": 68815, + "=sys": 58870, + "=t": 24000, + "=target": 96778, + "=temp": 53482, + "=test": 53538, + "=text": 48306, + "=tf": 32919, + "=this": 25667, + "=time": 60282, + "=title": 90507, + "=tk": 63751, + "=tmp": 92060, + "=top": 91017, + "=torch": 66257, + "=train": 76599, + "=true": 11265, + "=u": 38633, + "=url": 64563, + "=user": 30164, + "=username": 67421, + "=utf": 22264, + "=v": 39254, + "=val": 87529, + "=value": 46538, + "=view": 91953, + "=w": 63417, + "=wx": 90712, + "=x": 25387, + "=y": 29368, + "=yes": 59565, + "=z": 82570, + "=zeros": 56544, + "={": 1165, + "={!": 64426, + "={\"": 15783, + "={\"/": 98367, + "={$": 52273, + "={'": 12854, + "={(": 16817, + "={()": 10727, + "={()=>": 55051, + "={({": 82866, + "={<": 38006, + "={[": 32752, + "={['": 95590, + "={[Ċ": 68221, + "={`": 17559, + "={`${": 37312, + "={`/": 55241, + "={{": 6780, + "={{Ċ": 18268, + "={}": 68085, + "={},": 67425, + "={}Ċ": 58627, + "={Ċ": 17554, + "=}": 77422, + "=~": 94535, + "=âĢĿ": 44121, + "=Ċ": 14750, + "=ĊĊ": 68327, + ">": 29, + ">\"": 9877, + ">\")": 82598, + ">\").": 61032, + ">\");Ċ": 15084, + ">\");ĊĊ": 67940, + ">\");čĊ": 43691, + ">\")Ċ": 42363, + ">\"+": 45957, + ">\"+Ċ": 83522, + ">\",": 21156, + ">\",Ċ": 35452, + ">\".": 52946, + ">\".$": 33643, + ">\";": 31923, + ">\";Ċ": 6734, + ">\";ĊĊ": 32423, + ">\";čĊ": 20078, + ">\"Ċ": 19134, + ">\"čĊ": 82606, + ">#": 61125, + ">$": 23586, + ">${": 37680, + ">%": 28740, + ">&": 5789, + ">'": 5592, + ">')": 79865, + ">').": 57576, + ">');": 97236, + ">');Ċ": 22085, + ">');ĊĊ": 87346, + ">');čĊ": 78916, + ">')Ċ": 39866, + ">'+": 44501, + ">'+Ċ": 41909, + ">',": 19579, + ">',Ċ": 23174, + ">'.": 44689, + ">'.$": 29541, + ">'.Ċ": 63720, + ">';": 36709, + ">';Ċ": 7165, + ">';ĊĊ": 30463, + ">';čĊ": 22857, + ">'Ċ": 23431, + ">'čĊ": 95363, + ">(": 2235, + ">(\"": 13211, + ">(&": 44784, + ">('": 67638, + ">((": 47453, + ">(()": 58789, + ">()": 13555, + ">())": 92783, + ">());Ċ": 32872, + ">());ĊĊ": 93718, + ">())Ċ": 61004, + ">(),": 39019, + ">(),Ċ": 65766, + ">()->": 88380, + ">().": 10483, + ">();": 48281, + ">();Ċ": 3913, + ">();ĊĊ": 10448, + ">();čĊ": 15921, + ">();čĊčĊ": 51622, + ">()Ċ": 18949, + ">()ĊĊ": 73953, + ">(*": 60674, + ">(_": 45609, + ">({": 95753, + ">({Ċ": 81483, + ">(Ċ": 17055, + ">)": 9231, + ">).": 71524, + ">):": 37023, + ">);Ċ": 42013, + ">)Ċ": 32279, + ">*": 19443, + ">*/Ċ": 66475, + ">*+": 95326, + ">,": 8066, + ">,,Ċ": 12520, + ">-": 38643, + ">--": 82157, + ">-->Ċ": 36360, + ">--}}Ċ": 43196, + ">.": 14276, + ">..Ċ": 29816, + ">.ĊĊ": 94367, + ">/": 18105, + ">/',": 86864, + ">//": 52593, + ">/<": 67772, + ">:": 26818, + ">::": 6831, + ">:;": 65795, + ">;Ċ": 10133, + ">;ĊĊ": 19421, + ">;čĊ": 84109, + "><": 1784, + ">Ċ": 54769, + "]->": 6294, + "].": 936, + "].\"": 70966, + "].'": 95295, + "].[": 36416, + "]._": 39981, + "].Ċ": 26126, + "].ĊĊ": 29562, + "]/": 9769, + "]:": 5669, + "]:=": 90551, + "]:Ċ": 10343, + "]:ĊĊ": 68562, + "]:ĊĊĊ": 33666, + "]:čĊ": 52386, + "];": 5265, + "];//": 55682, + "];Ċ": 935, + "];ĊĊ": 4821, + "];ĊĊĊ": 52599, + "];čĊ": 5960, + "];čĊčĊ": 25890, + "]<": 34139, + "]": 93696, + "]=[": 66128, + "]=]": 95585, + "]={": 36336, + "]={Ċ": 95525, + "]>": 25669, + "]>=": 73775, + "]>Ċ": 64077, + "]?": 49964, + "]?.": 77841, + "][": 1457, + "][\"": 9868, + "][$": 24040, + "]['": 5352, + "][(": 82776, + "][-": 62872, + "][/": 77794, + "][:": 25091, + "][]": 45725, + "][_": 94692, + "]\\": 17960, + "]\\\\": 86920, + "]]": 5053, + "]])": 29616, + "]]);Ċ": 49854, + "]])Ċ": 26731, + "]])ĊĊ": 72103, + "]],": 20492, + "]],Ċ": 18543, + "]].": 22099, + "]]:Ċ": 83320, + "]];": 97829, + "]];Ċ": 13204, + "]];ĊĊ": 86853, + "]]=": 71383, + "]]>Ċ": 87565, + "]}Ċ": 23439, + "]Ċ": 921, + "]ĊĊ": 2533, + "]ĊĊĊ": 21675, + "]ĊĊĊĊ": 46739, + "]čĊ": 6114, + "]čĊčĊ": 36818, + "^": 61, + "^(": 13268, + "^-": 49309, + "^.": 39396, + "^K": 51967, + "^[": 27736, + "^\\": 24884, + "^^": 21876, + "^^^^": 61724, + "^n": 86167, + "^{": 47822, + "^{-": 87210, + "_": 62, + "_\"": 18771, + "_\"+": 33415, + "_\",": 60102, + "_\".$": 80511, + "_##": 49131, + "_#{": 90599, + "_$": 34421, + "_$_": 62750, + "_${": 40855, + "_%": 18695, + "_'": 19333, + "_'+": 33996, + "_',": 58013, + "_'.$": 38201, + "_(": 8361, + "_(\"": 37719, + "_('": 74678, + "_();Ċ": 96060, + "_()Ċ": 87358, + "_)": 16324, + "_));Ċ": 68464, + "_),": 65016, + "_);Ċ": 15299, + "_);ĊĊ": 50640, + "_);čĊ": 62447, + "_)Ċ": 23373, + "_)čĊ": 76643, + "_*": 45797, + "_,": 6878, + "_,Ċ": 38386, + "_-": 39311, + "_->": 10462, + "_-_": 83100, + "_.": 4950, + "_/": 50295, + "_:": 23211, + "_:*": 78593, + "_;": 23182, + "_;Ċ": 8588, + "_;ĊĊ": 22663, + "_;čĊ": 41940, + "_<": 41743, + "_": 98377, + "_A": 1566, + "_AA": 81941, + "_AB": 32643, + "_ABC": 91404, + "_ABI": 86974, + "_ABORT": 71044, + "_ABS": 65246, + "_AC": 15500, + "_ACC": 25666, + "_ACCEPT": 51741, + "_ACCESS": 24420, + "_ACCOUNT": 38831, + "_ACK": 50690, + "_ACL": 97627, + "_ACT": 15923, + "_ACTION": 21866, + "_ACTIONS": 94434, + "_ACTIV": 92626, + "_ACTIVE": 25946, + "_ACTIVITY": 42445, + "_AD": 22510, + "_ADAPTER": 79602, + "_ADC": 41630, + "_ADD": 8581, + "_ADDR": 16058, + "_ADDRESS": 20135, + "_ADMIN": 30460, + "_ADV": 74438, + "_AES": 69381, + "_AF": 46735, + "_AFTER": 72339, + "_AG": 37851, + "_AGENT": 52499, + "_AHB": 95846, + "_AI": 75322, + "_AL": 8912, + "_ALARM": 87735, + "_ALERT": 64486, + "_ALIAS": 61608, + "_ALIGN": 36688, + "_ALIGNMENT": 78518, + "_ALL": 16269, + "_ALLOC": 39333, + "_ALLOW": 44324, + "_ALLOWED": 73186, + "_ALPHA": 39497, + "_ALREADY": 80554, + "_ALT": 57861, + "_ALWAYS": 62761, + "_AM": 25022, + "_AMD": 98624, + "_AMOUNT": 59993, + "_AN": 23615, + "_ANAL": 59753, + "_AND": 21767, + "_ANDROID": 73238, + "_ANGLE": 78934, + "_ANS": 67752, + "_ANT": 97119, + "_ANY": 34450, + "_AP": 21899, + "_APB": 70391, + "_API": 11415, + "_APP": 16410, + "_APPEND": 65851, + "_APPLICATION": 54206, + "_APPRO": 75001, + "_APPS": 89754, + "_AR": 23399, + "_ARB": 53400, + "_ARCH": 40049, + "_ARCHIVE": 98636, + "_AREA": 50992, + "_ARG": 13905, + "_ARGS": 24386, + "_ARGUMENT": 51922, + "_ARM": 47051, + "_ARR": 86014, + "_ARRAY": 17724, + "_ARROW": 71295, + "_ART": 59156, + "_AS": 28028, + "_ASC": 84593, + "_ASCII": 98924, + "_ASM": 86906, + "_ASS": 18676, + "_ASSERT": 11796, + "_ASSIGN": 52248, + "_ASSOC": 31593, + "_ASSUME": 50089, + "_AST": 76853, + "_ASYNC": 80909, + "_AT": 8667, + "_ATOM": 97360, + "_ATOMIC": 87843, + "_ATT": 29977, + "_ATTACH": 49797, + "_ATTACHMENT": 69448, + "_ATTACK": 63398, + "_ATTR": 12865, + "_ATTRIB": 99043, + "_ATTRIBUTE": 32245, + "_ATTRIBUTES": 64550, + "_AUD": 81697, + "_AUDIO": 38002, + "_AURA": 76922, + "_AUT": 53716, + "_AUTH": 22675, + "_AUTHOR": 78843, + "_AUTO": 29013, + "_AUX": 81634, + "_AV": 29386, + "_AVAILABLE": 51823, + "_AX": 36817, + "_AXIS": 50431, + "_Abstract": 74372, + "_Account": 86855, + "_Act": 46672, + "_Action": 59899, + "_Ad": 23834, + "_Add": 21346, + "_Address": 64899, + "_Adjust": 41704, + "_AdjustorThunk": 42901, + "_Admin": 57308, + "_Al": 40812, + "_All": 53629, + "_An": 32699, + "_And": 93846, + "_Anim": 72072, + "_Api": 91128, + "_App": 36117, + "_Application": 96579, + "_Arg": 87581, + "_Array": 47229, + "_As": 62741, + "_AspNet": 77421, + "_Ass": 62222, + "_At": 55581, + "_Att": 81797, + "_B": 1668, + "_BACK": 22005, + "_BACKEND": 77795, + "_BACKGROUND": 70128, + "_BAD": 30666, + "_BAL": 79952, + "_BAND": 78268, + "_BANK": 66407, + "_BAR": 47405, + "_BASE": 11762, + "_BASIC": 71068, + "_BATCH": 74009, + "_BB": 93929, + "_BC": 72247, + "_BE": 27168, + "_BEFORE": 82218, + "_BEGIN": 26382, + "_BG": 62290, + "_BGR": 50825, + "_BIG": 62743, + "_BIN": 55720, + "_BINARY": 49754, + "_BIND": 44555, + "_BINDING": 85172, + "_BIT": 11189, + "_BITMAP": 72376, + "_BITS": 32695, + "_BL": 21089, + "_BLACK": 56555, + "_BLE": 44300, + "_BLEND": 62763, + "_BLK": 87634, + "_BLOCK": 18756, + "_BLOCKS": 86542, + "_BLUE": 53463, + "_BO": 12551, + "_BOARD": 65256, + "_BODY": 44411, + "_BOLD": 90166, + "_BOOK": 61256, + "_BOOL": 32491, + "_BOOLEAN": 37809, + "_BOOT": 57313, + "_BORDER": 61235, + "_BOTH": 82718, + "_BOTTOM": 58333, + "_BOUND": 73644, + "_BOUNDS": 91625, + "_BOX": 40167, + "_BP": 68822, + "_BR": 26543, + "_BRANCH": 71746, + "_BREAK": 63841, + "_BROWSER": 72500, + "_BS": 92962, + "_BT": 53970, + "_BTN": 58558, + "_BU": 92218, + "_BUCKET": 74444, + "_BUF": 35207, + "_BUFF": 62128, + "_BUFFER": 14394, + "_BUILD": 37491, + "_BUS": 33473, + "_BUSY": 67283, + "_BUTTON": 20110, + "_BY": 13696, + "_BYTE": 27349, + "_BYTES": 40705, + "_Back": 64703, + "_Bar": 91069, + "_Base": 33982, + "_Begin": 93447, + "_Bl": 56947, + "_Block": 51779, + "_Blue": 97201, + "_Bool": 79948, + "_Box": 81953, + "_Buffer": 55040, + "_Build": 96686, + "_Button": 41297, + "_By": 85996, + "_C": 920, + "_CA": 54890, + "_CACHE": 29138, + "_CAL": 39415, + "_CALC": 98014, + "_CALL": 14935, + "_CALLBACK": 35773, + "_CAM": 39446, + "_CAMERA": 58379, + "_CAN": 45600, + "_CANCEL": 47459, + "_CANNOT": 90916, + "_CAP": 21879, + "_CAPACITY": 92153, + "_CAPTURE": 61801, + "_CAR": 85887, + "_CARD": 45529, + "_CART": 65172, + "_CASE": 28540, + "_CAST": 25753, + "_CAT": 64572, + "_CATEGORY": 39284, + "_CB": 48933, + "_CBC": 90764, + "_CC": 31331, + "_CD": 55200, + "_CE": 94805, + "_CELL": 59109, + "_CENTER": 34555, + "_CERT": 55298, + "_CF": 72058, + "_CFG": 21760, + "_CH": 6466, + "_CHAIN": 70232, + "_CHAN": 80324, + "_CHANGE": 32523, + "_CHANGED": 48099, + "_CHANNEL": 23688, + "_CHANNELS": 77026, + "_CHAR": 20891, + "_CHARACTER": 68877, + "_CHARS": 82319, + "_CHARSET": 95104, + "_CHAT": 70748, + "_CHECK": 11780, + "_CHILD": 59196, + "_CHIP": 79101, + "_CHK": 87855, + "_CHO": 50616, + "_CHOICES": 74995, + "_CHUNK": 70508, + "_CI": 93807, + "_CID": 87525, + "_CIPHER": 80973, + "_CITY": 85310, + "_CL": 6843, + "_CLAMP": 88309, + "_CLASS": 13735, + "_CLASSES": 69164, + "_CLEAN": 79661, + "_CLEAR": 39280, + "_CLI": 91195, + "_CLICK": 52278, + "_CLICKED": 94245, + "_CLIENT": 22521, + "_CLIP": 96353, + "_CLK": 25160, + "_CLOCK": 41296, + "_CLOSE": 24323, + "_CLOSED": 78140, + "_CLR": 71499, + "_CLUSTER": 77871, + "_CM": 42337, + "_CMD": 16344, + "_CMP": 83256, + "_CN": 56621, + "_CNT": 34202, + "_CNTL": 95575, + "_CO": 19628, + "_COD": 81505, + "_CODE": 10020, + "_CODEC": 68199, + "_CODES": 41688, + "_COL": 25332, + "_COLL": 74317, + "_COLLECTION": 68757, + "_COLOR": 14471, + "_COLORS": 85436, + "_COLS": 93714, + "_COLUMN": 36183, + "_COLUMNS": 71903, + "_COM": 6877, + "_COMBO": 95403, + "_COMM": 14025, + "_COMMAND": 22723, + "_COMMENT": 49956, + "_COMMIT": 77364, + "_COMMON": 45026, + "_COMP": 14969, + "_COMPANY": 97673, + "_COMPARE": 77898, + "_COMPAT": 76770, + "_COMPILE": 83401, + "_COMPILER": 87009, + "_COMPLETE": 50665, + "_COMPLETED": 85674, + "_COMPLEX": 97820, + "_COMPONENT": 40021, + "_CON": 4307, + "_COND": 81631, + "_CONDITION": 72815, + "_CONF": 29212, + "_CONFIG": 12568, + "_CONFIGURATION": 70752, + "_CONFIRM": 64934, + "_CONN": 60376, + "_CONNECT": 33462, + "_CONNECTED": 69434, + "_CONNECTION": 40708, + "_CONSOLE": 89292, + "_CONST": 23098, + "_CONSTANT": 55310, + "_CONT": 10710, + "_CONTACT": 54636, + "_CONTAINER": 50689, + "_CONTENT": 25560, + "_CONTEXT": 29577, + "_CONTINUE": 79066, + "_CONTROL": 23680, + "_CONTROLLER": 63995, + "_CONV": 85323, + "_CONVERT": 79121, + "_COOKIE": 76549, + "_COORD": 60508, + "_COPY": 35556, + "_COR": 47696, + "_CORE": 31310, + "_COST": 70932, + "_COUNT": 14672, + "_COUNTER": 48411, + "_COUNTRY": 94005, + "_CP": 45976, + "_CPP": 64624, + "_CPU": 29741, + "_CR": 24314, + "_CRC": 75310, + "_CRE": 68872, + "_CREAT": 70802, + "_CREATE": 25823, + "_CREATED": 57784, + "_CRITICAL": 88036, + "_CRYPTO": 87245, + "_CS": 34796, + "_CSR": 81449, + "_CSS": 83579, + "_CSV": 96594, + "_CT": 72547, + "_CTL": 37537, + "_CTRL": 19170, + "_CTX": 44233, + "_CUBE": 91969, + "_CUDA": 94861, + "_CUR": 33834, + "_CURRENT": 32938, + "_CURSOR": 58988, + "_CUSTOM": 36314, + "_CUSTOMER": 90292, + "_CY": 46431, + "_CYCLE": 86802, + "_Call": 76028, + "_Callback": 55102, + "_Camera": 89076, + "_Cancel": 97485, + "_Category": 80414, + "_Cell": 54454, + "_Ch": 27588, + "_Channel": 68858, + "_Char": 50896, + "_Character": 84529, + "_Check": 28188, + "_Checked": 48618, + "_CheckedChanged": 51911, + "_Cl": 85110, + "_Class": 47538, + "_Clear": 57744, + "_Click": 7163, + "_Client": 46102, + "_Close": 68185, + "_Cmd": 98253, + "_Code": 38084, + "_Collections": 37910, + "_Color": 43387, + "_Column": 96622, + "_Com": 16946, + "_ComCallableWrapper": 88950, + "_Comm": 59452, + "_Command": 66751, + "_Common": 77271, + "_Component": 80760, + "_Con": 15100, + "_Config": 35412, + "_Connection": 85002, + "_Construct": 41126, + "_Content": 78383, + "_Context": 71994, + "_Control": 45376, + "_Controller": 30674, + "_Copy": 77637, + "_Core": 41990, + "_Count": 50775, + "_Create": 34325, + "_Current": 40735, + "_Custom": 57402, + "_Customer": 94596, + "_D": 1557, + "_DA": 75144, + "_DAC": 85966, + "_DAMAGE": 64459, + "_DAT": 36347, + "_DATA": 7896, + "_DATABASE": 45510, + "_DATE": 18500, + "_DAY": 43566, + "_DAYS": 89256, + "_DB": 16310, + "_DBG": 37620, + "_DC": 56661, + "_DD": 95985, + "_DDR": 79838, + "_DE": 2881, + "_DEAD": 82898, + "_DEBUG": 11139, + "_DEC": 23773, + "_DECL": 22732, + "_DECLARE": 48415, + "_DECLS": 94631, + "_DECREF": 38887, + "_DEF": 21736, + "_DEFAULT": 13811, + "_DEFIN": 40503, + "_DEFINE": 46807, + "_DEFINED": 59122, + "_DEFINITION": 92326, + "_DEL": 20918, + "_DELAY": 33410, + "_DELETE": 28923, + "_DELETED": 85921, + "_DELTA": 80814, + "_DEN": 66936, + "_DENIED": 92118, + "_DEPEND": 94293, + "_DEPRECATED": 82076, + "_DEPTH": 32644, + "_DER": 96062, + "_DESC": 23456, + "_DESCRIPTION": 38168, + "_DESCRIPTOR": 55445, + "_DEST": 45874, + "_DESTROY": 79400, + "_DET": 77107, + "_DETAIL": 32905, + "_DETAILS": 56462, + "_DETECT": 96218, + "_DEV": 26419, + "_DEVICE": 18094, + "_DEVICES": 84433, + "_DF": 92864, + "_DGRAM": 92273, + "_DH": 92439, + "_DI": 44663, + "_DIAG": 76210, + "_DIALOG": 42802, + "_DICT": 54047, + "_DIFF": 53984, + "_DIG": 33362, + "_DIGEST": 94901, + "_DIM": 39302, + "_DIP": 41524, + "_DIPSETTING": 57675, + "_DIR": 8291, + "_DIRECT": 31730, + "_DIRECTION": 65065, + "_DIRECTORY": 43814, + "_DIRS": 90560, + "_DIS": 12198, + "_DISABLE": 25680, + "_DISABLED": 55701, + "_DISCONNECT": 76225, + "_DISK": 84504, + "_DISP": 85475, + "_DISPATCH": 88674, + "_DISPLAY": 37579, + "_DIST": 40953, + "_DISTANCE": 63217, + "_DIV": 35303, + "_DL": 74436, + "_DLL": 53800, + "_DM": 58796, + "_DMA": 33782, + "_DO": 21820, + "_DOC": 68911, + "_DOCUMENT": 60584, + "_DOM": 30230, + "_DOMAIN": 33729, + "_DONE": 42741, + "_DOT": 82567, + "_DOUBLE": 33839, + "_DOWN": 21157, + "_DOWNLOAD": 67748, + "_DP": 71475, + "_DR": 20488, + "_DRAW": 42731, + "_DRIVE": 92787, + "_DRIVER": 44042, + "_DROP": 64210, + "_DRV": 66173, + "_DS": 51901, + "_DSP": 94281, + "_DST": 71772, + "_DT": 60255, + "_DU": 97153, + "_DUMP": 74498, + "_DURATION": 53413, + "_DX": 92831, + "_DYNAMIC": 57356, + "_Data": 17817, + "_Date": 39564, + "_Db": 60968, + "_De": 24597, + "_Debug": 77938, + "_Dec": 78668, + "_Def": 88411, + "_Default": 60336, + "_Delay": 92726, + "_Delete": 57418, + "_Dep": 90680, + "_Des": 98054, + "_Desc": 92957, + "_Description": 79687, + "_Destroy": 79266, + "_Detail": 94762, + "_Details": 97426, + "_Device": 66780, + "_Dis": 45525, + "_Display": 80352, + "_Do": 93481, + "_Double": 84390, + "_Draw": 69154, + "_E": 2089, + "_EC": 69510, + "_ED": 32860, + "_EDEFAULT": 88948, + "_EDGE": 52043, + "_EDIT": 25106, + "_EDITOR": 52770, + "_EFFECT": 55634, + "_EL": 62283, + "_ELEM": 90704, + "_ELEMENT": 27156, + "_ELEMENTS": 66169, + "_EM": 27148, + "_EMAIL": 29548, + "_EMIT": 97763, + "_EMP": 68784, + "_EMPTY": 36640, + "_EN": 6300, + "_ENABLE": 14379, + "_ENABLED": 29276, + "_ENC": 72674, + "_ENCOD": 53269, + "_ENCODE": 89513, + "_ENCODING": 73029, + "_END": 10898, + "_ENDIAN": 57769, + "_ENDPOINT": 48756, + "_ENGINE": 62177, + "_ENSURE": 82817, + "_ENT": 16241, + "_ENTER": 41974, + "_ENTITY": 46599, + "_ENTRIES": 87695, + "_ENTRY": 22659, + "_ENUM": 31894, + "_ENV": 22027, + "_EOF": 89287, + "_EOL": 36244, + "_EP": 52194, + "_EPS": 77292, + "_EQ": 9168, + "_EQUAL": 18182, + "_EQUALS": 68581, + "_ER": 58995, + "_ERR": 9875, + "_ERROR": 5414, + "_ERRORS": 66241, + "_ES": 30275, + "_ESCAPE": 75575, + "_EST": 79816, + "_ET": 73755, + "_ETH": 79667, + "_EV": 65458, + "_EVAL": 70965, + "_EVENT": 14054, + "_EVENTS": 61357, + "_EVT": 64955, + "_EX": 4966, + "_EXCEPTION": 27285, + "_EXEC": 38235, + "_EXECUTE": 91103, + "_EXIST": 49049, + "_EXISTS": 70042, + "_EXIT": 34009, + "_EXP": 34468, + "_EXPECT": 68208, + "_EXPI": 62006, + "_EXPORT": 20789, + "_EXPR": 73051, + "_EXPRESSION": 98345, + "_EXT": 16066, + "_EXTENDED": 95952, + "_EXTENSION": 37012, + "_EXTENSIONS": 97058, + "_EXTERN": 37743, + "_EXTERNAL": 47871, + "_EXTRA": 58981, + "_Edit": 66158, + "_Element": 64099, + "_Em": 76060, + "_Email": 83168, + "_Enable": 65216, + "_Enc": 93529, + "_End": 49953, + "_Ent": 50503, + "_Enter": 91105, + "_Entity": 89618, + "_Entry": 94164, + "_Equals": 86567, + "_Err": 93623, + "_Error": 28651, + "_Event": 39354, + "_Ex": 62531, + "_Exception": 48247, + "_Execute": 83453, + "_F": 1400, + "_FA": 19058, + "_FACE": 65113, + "_FACT": 36299, + "_FACTOR": 53706, + "_FACTORY": 71152, + "_FAIL": 11878, + "_FAILED": 22775, + "_FAILURE": 16998, + "_FALL": 83962, + "_FALSE": 18125, + "_FAMILY": 80828, + "_FAR": 87961, + "_FAST": 58225, + "_FATAL": 61540, + "_FAULT": 94664, + "_FB": 64707, + "_FC": 79903, + "_FD": 86221, + "_FE": 43307, + "_FEATURE": 30827, + "_FEED": 74176, + "_FETCH": 57925, + "_FF": 86880, + "_FIELD": 17354, + "_FIELDS": 52999, + "_FIFO": 49429, + "_FILE": 8087, + "_FILENAME": 45005, + "_FILENO": 75538, + "_FILES": 48010, + "_FILL": 61000, + "_FILTER": 23728, + "_FIN": 45280, + "_FINAL": 75872, + "_FIND": 67285, + "_FINE": 90687, + "_FINISH": 87093, + "_FIRE": 76614, + "_FIRST": 36272, + "_FIX": 42635, + "_FIXED": 66300, + "_FL": 22588, + "_FLAG": 11890, + "_FLAGS": 30090, + "_FLASH": 48515, + "_FLOAT": 21496, + "_FLOW": 55818, + "_FLUSH": 80395, + "_FM": 92312, + "_FMT": 44332, + "_FN": 60122, + "_FOCUS": 96629, + "_FOLDER": 33637, + "_FOLLOW": 78869, + "_FONT": 39532, + "_FOR": 14516, + "_FORCE": 50092, + "_FORE": 47185, + "_FOREACH": 70751, + "_FORM": 12705, + "_FORMAT": 14960, + "_FORWARD": 64005, + "_FOUND": 24687, + "_FP": 70981, + "_FR": 14220, + "_FRAGMENT": 66157, + "_FRAME": 22260, + "_FRAMEBUFFER": 71551, + "_FRE": 73183, + "_FREE": 35356, + "_FREQ": 48602, + "_FREQUENCY": 93174, + "_FRIEND": 86268, + "_FROM": 27108, + "_FRONT": 51915, + "_FS": 40398, + "_FT": 84740, + "_FULL": 29822, + "_FULLSCREEN": 87688, + "_FUN": 58873, + "_FUNC": 20875, + "_FUNCTION": 18490, + "_FUNCTIONS": 95484, + "_FW": 73108, + "_FWD": 84977, + "_Field": 46272, + "_FieldOffsetTable": 88396, + "_File": 34061, + "_Filter": 68935, + "_Final": 77072, + "_Find": 95245, + "_First": 79790, + "_Flag": 98135, + "_Float": 89067, + "_Font": 74213, + "_For": 84368, + "_Form": 28121, + "_Format": 72999, + "_Frame": 68403, + "_Framework": 61995, + "_Free": 58455, + "_From": 53157, + "_Function": 59738, + "_G": 2646, + "_GAIN": 67068, + "_GAME": 40908, + "_GAP": 84721, + "_GATE": 81116, + "_GB": 98631, + "_GC": 76915, + "_GE": 59033, + "_GEN": 42362, + "_GENER": 18693, + "_GENERAL": 65336, + "_GENERIC": 25274, + "_GET": 12702, + "_GF": 79365, + "_GL": 36700, + "_GLOBAL": 36761, + "_GO": 39622, + "_GOOD": 86725, + "_GP": 85834, + "_GPIO": 17863, + "_GPS": 97123, + "_GPU": 68716, + "_GR": 17874, + "_GRA": 76774, + "_GRANTED": 63910, + "_GRAPH": 58360, + "_GRAY": 60516, + "_GRE": 72739, + "_GREEN": 53059, + "_GRID": 61435, + "_GROUP": 20130, + "_GROUPS": 89189, + "_GRP": 90023, + "_GT": 71881, + "_GU": 84513, + "_GUI": 61305, + "_GUID": 58648, + "_Game": 49223, + "_Generic": 29085, + "_GenericClass": 35974, + "_Get": 13614, + "_Global": 93132, + "_Grid": 83047, + "_Group": 52619, + "_H": 2039, + "_HAL": 43694, + "_HALF": 74611, + "_HAND": 29190, + "_HANDLE": 23683, + "_HANDLER": 42472, + "_HARD": 78149, + "_HAS": 37978, + "_HASH": 32309, + "_HAVE": 56087, + "_HC": 98991, + "_HDR": 55765, + "_HE": 7867, + "_HEAD": 27990, + "_HEADER": 20330, + "_HEADERS": 79428, + "_HEALTH": 99028, + "_HEAP": 81358, + "_HEIGHT": 17355, + "_HEL": 69235, + "_HELP": 49587, + "_HELPER": 85331, + "_HERE": 81555, + "_HERSHEY": 92534, + "_HEX": 86502, + "_HI": 72536, + "_HIDDEN": 80921, + "_HIDE": 72014, + "_HIGH": 31621, + "_HINT": 74457, + "_HISTORY": 63823, + "_HIT": 86767, + "_HOLD": 97553, + "_HOME": 28466, + "_HOOK": 82251, + "_HOR": 90462, + "_HORIZONTAL": 75867, + "_HOST": 17213, + "_HOT": 91323, + "_HOUR": 86292, + "_HP": 80588, + "_HPP": 19974, + "_HS": 81832, + "_HT": 24105, + "_HTML": 56726, + "_HTTP": 34954, + "_HW": 44013, + "_Handle": 42714, + "_HandleTypeDef": 78906, + "_Handler": 41879, + "_Header": 71353, + "_Height": 83829, + "_Helper": 67828, + "_I": 7959, + "_IA": 87490, + "_IB": 79263, + "_IC": 22960, + "_ICON": 36525, + "_ID": 3450, + "_IDENT": 40264, + "_IDENTIFIER": 49554, + "_IDLE": 46390, + "_IDS": 74942, + "_IDX": 40932, + "_IE": 60448, + "_IEnumerator": 88913, + "_IF": 19035, + "_IGNORE": 49974, + "_IL": 70862, + "_IList": 88431, + "_IM": 8351, + "_IMAGE": 19121, + "_IMAGES": 79415, + "_IMETHOD": 76779, + "_IMG": 50365, + "_IMM": 67303, + "_IMP": 97415, + "_IMPL": 51345, + "_IMPLEMENT": 62325, + "_IMPORT": 21024, + "_IMPORTED": 25267, + "_IN": 2158, + "_INC": 39901, + "_INCLUDE": 48081, + "_INCLUDED": 36078, + "_INCREF": 74347, + "_INCREMENT": 58058, + "_IND": 41003, + "_INDEX": 14515, + "_INET": 26834, + "_INF": 65487, + "_INFINITY": 91110, + "_INFO": 9068, + "_INFORMATION": 53124, + "_ING": 84485, + "_INIT": 14446, + "_INITIAL": 65872, + "_INITIALIZ": 59082, + "_INITIALIZER": 90214, + "_INLINE": 32482, + "_INPUT": 21022, + "_INS": 31650, + "_INSERT": 47924, + "_INST": 49893, + "_INSTALL": 58486, + "_INSTANCE": 42587, + "_INT": 9161, + "_INTEGER": 25757, + "_INTER": 16732, + "_INTERFACE": 31561, + "_INTERNAL": 31752, + "_INTERRUPT": 53770, + "_INTERVAL": 37353, + "_INTR": 43661, + "_INV": 40107, + "_INVALID": 15167, + "_IO": 19790, + "_IOC": 97001, + "_IOCTL": 67326, + "_IOS": 79259, + "_IP": 16607, + "_IPV": 58830, + "_IR": 17138, + "_IRQ": 21980, + "_IRQHandler": 74944, + "_IRQn": 52357, + "_IS": 12766, + "_ISO": 83858, + "_ISR": 70678, + "_ISS": 92687, + "_IT": 41681, + "_ITEM": 16461, + "_ITEMS": 56280, + "_ITER": 51347, + "_IV": 61917, + "_IW": 86421, + "_Id": 21576, + "_Il": 38563, + "_Image": 45949, + "_Impl": 60532, + "_In": 25972, + "_Index": 50361, + "_Info": 39624, + "_Init": 15644, + "_InitStruct": 42596, + "_InitStructure": 60099, + "_Input": 48653, + "_Insert": 76417, + "_Instance": 70849, + "_Int": 32054, + "_Integer": 80318, + "_Inter": 79717, + "_Interface": 74626, + "_Internal": 37238, + "_InternalArray": 72128, + "_Invoke": 79143, + "_Is": 31879, + "_Item": 27518, + "_Items": 74467, + "_J": 10598, + "_JO": 38532, + "_JOB": 59690, + "_JOIN": 69317, + "_JS": 77846, + "_JSON": 25356, + "_JUMP": 88730, + "_K": 10102, + "_KEEP": 74937, + "_KERNEL": 39637, + "_KEY": 6600, + "_KEYBOARD": 81164, + "_KEYS": 51251, + "_KEYWORD": 71908, + "_KHR": 88528, + "_KIND": 72959, + "_KP": 72849, + "_Key": 35253, + "_KeyDown": 81459, + "_KeyPress": 77834, + "_L": 2351, + "_LA": 83759, + "_LABEL": 25670, + "_LANE": 68488, + "_LANG": 39761, + "_LANGUAGE": 59990, + "_LARGE": 83861, + "_LAST": 36794, + "_LAT": 64241, + "_LAYER": 50520, + "_LAYOUT": 52685, + "_LCD": 63092, + "_LD": 72800, + "_LE": 5280, + "_LEAVE": 89741, + "_LED": 37604, + "_LEFT": 19013, + "_LEG": 72303, + "_LEN": 15536, + "_LENGTH": 15023, + "_LESS": 90664, + "_LEVEL": 17415, + "_LIB": 26207, + "_LIBRARY": 54110, + "_LICENSE": 57560, + "_LIGHT": 35614, + "_LIMIT": 29194, + "_LINE": 15657, + "_LINEAR": 51948, + "_LINES": 66744, + "_LINK": 26147, + "_LINUX": 74025, + "_LIST": 11899, + "_LITERAL": 78047, + "_LL": 70309, + "_LO": 4606, + "_LOAD": 24042, + "_LOADED": 89917, + "_LOADING": 63960, + "_LOC": 16397, + "_LOCAL": 28399, + "_LOCATION": 30202, + "_LOCK": 27661, + "_LOG": 8419, + "_LOGGER": 87673, + "_LOGIN": 38314, + "_LONG": 19903, + "_LOOK": 73686, + "_LOOKUP": 93398, + "_LOOP": 49502, + "_LOW": 25831, + "_LOWER": 90132, + "_LP": 85430, + "_LR": 94938, + "_LS": 87789, + "_LSB": 76799, + "_LT": 65917, + "_LVL": 94282, + "_Label": 53557, + "_Last": 84672, + "_Lean": 95593, + "_Left": 70785, + "_Length": 81620, + "_Level": 78861, + "_Line": 59100, + "_Link": 72420, + "_List": 27104, + "_Load": 19553, + "_Local": 71421, + "_Location": 85524, + "_Log": 44083, + "_Login": 79232, + "_Long": 88240, + "_M": 1245, + "_MA": 85311, + "_MAC": 28076, + "_MACHINE": 66702, + "_MACRO": 73361, + "_MAG": 55231, + "_MAGIC": 49194, + "_MAIL": 65082, + "_MAIN": 32276, + "_MAJOR": 52045, + "_MAKE": 77442, + "_MALLOC": 86753, + "_MAN": 25143, + "_MANAGER": 44018, + "_MANY": 97374, + "_MAP": 16306, + "_MAPPING": 73513, + "_MARGIN": 71823, + "_MARK": 36916, + "_MARKER": 82689, + "_MASK": 11720, + "_MASTER": 47068, + "_MAT": 71917, + "_MATCH": 40824, + "_MATERIAL": 92932, + "_MATH": 85944, + "_MATRIX": 58376, + "_MAX": 6806, + "_MAXIMUM": 85377, + "_MAY": 83079, + "_MB": 76344, + "_MC": 59714, + "_MD": 46497, + "_ME": 33037, + "_MED": 97550, + "_MEDIA": 49527, + "_MEDIUM": 93792, + "_MEM": 13209, + "_MEMBER": 37552, + "_MEMBERS": 96446, + "_MEMORY": 30048, + "_MENU": 27648, + "_MESH": 85388, + "_MESSAGE": 14641, + "_MESSAGES": 69485, + "_MET": 33387, + "_META": 59836, + "_METADATA": 58084, + "_METHOD": 14058, + "_MI": 59763, + "_MIC": 69478, + "_MIDDLE": 86752, + "_MIN": 11820, + "_MINOR": 61524, + "_MINUS": 68263, + "_MISC": 69035, + "_MISS": 63608, + "_MISSING": 80706, + "_MIX": 77299, + "_MM": 46957, + "_MO": 18089, + "_MOBILE": 94137, + "_MOD": 20724, + "_MODAL": 86300, + "_MODE": 8414, + "_MODEL": 27551, + "_MODIFIED": 85958, + "_MODULE": 13953, + "_MODULES": 95742, + "_MON": 27065, + "_MONITOR": 81734, + "_MONTH": 48023, + "_MORE": 75131, + "_MOUNT": 95600, + "_MOUSE": 53307, + "_MOV": 48664, + "_MOVE": 35876, + "_MP": 53052, + "_MPI": 77454, + "_MR": 82153, + "_MS": 36147, + "_MSB": 96389, + "_MSG": 14372, + "_MSK": 95317, + "_MT": 49047, + "_MUL": 91421, + "_MULT": 29207, + "_MULTI": 49794, + "_MUT": 52403, + "_MUTEX": 66994, + "_MUX": 79351, + "_MY": 84395, + "_Main": 45105, + "_Man": 71628, + "_Manager": 47573, + "_Map": 56992, + "_Master": 80681, + "_Matrix": 76933, + "_Max": 58843, + "_Menu": 54098, + "_Message": 50974, + "_Meta": 59812, + "_MetaData": 75216, + "_Metadata": 62219, + "_MetadataUsageId": 67031, + "_Method": 44844, + "_MethodInfo": 69428, + "_Min": 62122, + "_Mod": 71485, + "_Mode": 54361, + "_Model": 21626, + "_Module": 63197, + "_Mouse": 36247, + "_Move": 66352, + "_Msg": 90653, + "_Msk": 66554, + "_Msp": 67437, + "_N": 1604, + "_NAME": 4708, + "_NAMES": 24833, + "_NAMESPACE": 34552, + "_NATIVE": 55575, + "_NAV": 71148, + "_NB": 80107, + "_NC": 95942, + "_NE": 14039, + "_NEAR": 78148, + "_NEAREST": 91926, + "_NEED": 86068, + "_NEG": 62368, + "_NET": 26855, + "_NETWORK": 48119, + "_NEW": 22015, + "_NEXT": 39987, + "_NM": 96053, + "_NO": 9100, + "_NODE": 19218, + "_NODES": 92948, + "_NON": 22128, + "_NONE": 14904, + "_NONNULL": 47703, + "_NOP": 71810, + "_NORMAL": 26455, + "_NOT": 9169, + "_NOTE": 54791, + "_NOTICE": 67800, + "_NOTIFICATION": 54241, + "_NOTIFY": 57614, + "_NOW": 59401, + "_NPC": 91569, + "_NR": 68195, + "_NS": 33777, + "_NT": 76575, + "_NULL": 18119, + "_NUM": 9631, + "_NUMBER": 19364, + "_NUMERIC": 84393, + "_NV": 49631, + "_Name": 19015, + "_Native": 98233, + "_New": 39582, + "_No": 36989, + "_Node": 41340, + "_None": 52523, + "_Normal": 74900, + "_Not": 60816, + "_Null": 55534, + "_Num": 61883, + "_Number": 51799, + "_O": 2232, + "_OBJ": 27559, + "_OBJC": 63804, + "_OBJECT": 13442, + "_OBS": 86948, + "_OC": 74663, + "_OCC": 32874, + "_OCCURRED": 35746, + "_OD": 60426, + "_OF": 14234, + "_OFF": 11987, + "_OFFSET": 17626, + "_OID": 96284, + "_OK": 8375, + "_OLD": 82869, + "_OM": 83169, + "_ON": 11077, + "_ONCE": 98814, + "_ONE": 34727, + "_ONLY": 31263, + "_OP": 13908, + "_OPCODE": 73080, + "_OPEN": 24070, + "_OPENGL": 91353, + "_OPER": 27205, + "_OPERATION": 44685, + "_OPERATOR": 69678, + "_OPT": 28264, + "_OPTION": 19968, + "_OPTIONS": 36321, + "_OPTS": 84015, + "_OR": 19834, + "_ORD": 85870, + "_ORDER": 26677, + "_ORIENTATION": 94180, + "_ORIGIN": 97199, + "_OS": 30400, + "_OT": 51147, + "_OTHER": 56038, + "_OUT": 11881, + "_OUTPUT": 23981, + "_OVER": 32043, + "_OVERFLOW": 94393, + "_OVERRIDE": 49372, + "_OW": 69819, + "_OWNER": 74323, + "_Obj": 70672, + "_Object": 27839, + "_Of": 89355, + "_Off": 94377, + "_Offset": 85697, + "_On": 35482, + "_One": 68201, + "_Open": 51747, + "_Options": 86101, + "_Order": 53267, + "_Osc": 82105, + "_OscInitStruct": 96199, + "_Out": 36675, + "_Output": 65062, + "_P": 1088, + "_PA": 26191, + "_PACK": 23987, + "_PACKAGE": 55566, + "_PACKET": 40328, + "_PAD": 36451, + "_PADDING": 78359, + "_PAGE": 19971, + "_PAGES": 79873, + "_PAIR": 65765, + "_PANEL": 68988, + "_PAR": 32029, + "_PARAM": 12073, + "_PARAMETER": 33241, + "_PARAMETERS": 71463, + "_PARAMS": 37488, + "_PARENT": 46848, + "_PARSE": 69706, + "_PARSER": 78544, + "_PART": 28278, + "_PARTITION": 91664, + "_PASS": 34351, + "_PASSWORD": 23059, + "_PATCH": 76804, + "_PATH": 7944, + "_PATTERN": 44227, + "_PAUSE": 79538, + "_PAY": 36417, + "_PAYLOAD": 79493, + "_PAYMENT": 67566, + "_PB": 71556, + "_PC": 39797, + "_PCI": 71520, + "_PCIE": 93061, + "_PCM": 69071, + "_PD": 69487, + "_PE": 45784, + "_PED": 81630, + "_PEER": 92742, + "_PENDING": 56614, + "_PER": 10617, + "_PERCENT": 72223, + "_PERIOD": 43166, + "_PERMISSION": 52788, + "_PERSON": 73428, + "_PF": 79180, + "_PG": 85644, + "_PH": 17909, + "_PHASE": 70239, + "_PHONE": 59704, + "_PHOTO": 92641, + "_PHP": 81582, + "_PHY": 35519, + "_PHYS": 68356, + "_PI": 20212, + "_PICK": 74032, + "_PICTURE": 94179, + "_PID": 55794, + "_PIN": 15844, + "_PIPE": 51584, + "_PIPELINE": 96516, + "_PIX": 85372, + "_PIXEL": 50484, + "_PK": 63898, + "_PKG": 94126, + "_PKT": 69879, + "_PL": 11082, + "_PLACE": 81924, + "_PLAN": 83648, + "_PLATFORM": 43219, + "_PLAY": 38575, + "_PLAYER": 36554, + "_PLL": 46620, + "_PLUGIN": 39935, + "_PLUS": 71878, + "_PM": 39918, + "_PO": 14169, + "_POINT": 27407, + "_POINTER": 28701, + "_POINTS": 64163, + "_POL": 33165, + "_POLICY": 58999, + "_POLL": 76820, + "_POOL": 51924, + "_POP": 55508, + "_PORT": 12377, + "_PORTS": 95559, + "_POS": 18313, + "_POSITION": 33580, + "_POST": 20506, + "_POSTFIELDS": 96247, + "_POWER": 33304, + "_PP": 43258, + "_PR": 10571, + "_PRE": 17427, + "_PRED": 92180, + "_PREF": 66372, + "_PREFIX": 18519, + "_PRESENT": 68708, + "_PRESS": 61219, + "_PREVIEW": 98990, + "_PRI": 81193, + "_PRICE": 51900, + "_PRIMARY": 75019, + "_PRINT": 23402, + "_PRINTF": 70395, + "_PRIORITY": 42694, + "_PRIV": 70625, + "_PRIVATE": 30470, + "_PRO": 5756, + "_PROC": 45050, + "_PROCESS": 36199, + "_PROD": 89799, + "_PRODUCT": 31850, + "_PRODUCTS": 79881, + "_PROF": 95152, + "_PROFILE": 34533, + "_PROGRAM": 59832, + "_PROGRESS": 54446, + "_PROJECT": 43804, + "_PROM": 71346, + "_PROP": 32839, + "_PROPERTIES": 64619, + "_PROPERTY": 26021, + "_PROTO": 30265, + "_PROTOCOL": 39893, + "_PROVID": 51868, + "_PROVIDER": 62542, + "_PROXY": 59065, + "_PS": 47666, + "_PT": 65061, + "_PTR": 20657, + "_PUBLIC": 36209, + "_PULL": 60284, + "_PUR": 79006, + "_PUSH": 58811, + "_PUT": 88429, + "_PW": 96740, + "_PWM": 50459, + "_PWR": 57996, + "_PY": 78537, + "_Page": 51540, + "_Panel": 98817, + "_Param": 27426, + "_Parameter": 84458, + "_Params": 44656, + "_Parms": 78933, + "_Parse": 77337, + "_Part": 97938, + "_Password": 93302, + "_Path": 66388, + "_Per": 53918, + "_Ph": 85775, + "_Pin": 35453, + "_Pl": 41609, + "_Play": 91854, + "_Player": 51955, + "_Pods": 76861, + "_Point": 52838, + "_Port": 56328, + "_Pos": 44246, + "_Position": 55724, + "_Post": 66726, + "_Pr": 32716, + "_Pre": 79561, + "_Price": 93234, + "_Print": 45788, + "_Printf": 87257, + "_Private": 43830, + "_Pro": 16670, + "_Process": 70241, + "_Product": 44017, + "_Profile": 77294, + "_Project": 64659, + "_Property": 68353, + "_Ptr": 81822, + "_Public": 73905, + "_Q": 13337, + "_QMARK": 87343, + "_QU": 16332, + "_QUAL": 69942, + "_QUERY": 31585, + "_QUESTION": 93430, + "_QUEUE": 35292, + "_QUOTES": 53358, + "_Query": 48042, + "_R": 2568, + "_RA": 53554, + "_RAD": 90473, + "_RADIO": 74415, + "_RADIUS": 60226, + "_RAM": 62393, + "_RANDOM": 72675, + "_RANGE": 28270, + "_RANK": 70006, + "_RATE": 29108, + "_RATIO": 65679, + "_RAW": 39922, + "_RB": 89753, + "_RC": 41414, + "_RCC": 63011, + "_RD": 28694, + "_RDONLY": 66104, + "_RDWR": 66266, + "_RE": 2192, + "_READ": 13117, + "_READONLY": 93695, + "_READY": 51184, + "_REAL": 30111, + "_REALTYPE": 56261, + "_REASON": 71945, + "_REC": 62252, + "_RECE": 49816, + "_RECEIVED": 99145, + "_RECORD": 38243, + "_RECT": 69621, + "_RECV": 95042, + "_RED": 33737, + "_REDIRECT": 90279, + "_REF": 12312, + "_REFER": 70910, + "_REFERENCE": 21121, + "_REFERER": 86771, + "_REFRESH": 77272, + "_REG": 8064, + "_REGEX": 56196, + "_REGION": 39179, + "_REGISTER": 33355, + "_REGISTRY": 93895, + "_REGS": 83317, + "_REL": 41207, + "_RELEASE": 40649, + "_REMOTE": 69832, + "_REMOVE": 50046, + "_RENDER": 45867, + "_RENDERER": 96641, + "_REPEAT": 66570, + "_REPLACE": 97720, + "_REPLY": 70200, + "_REPO": 82479, + "_REPORT": 40413, + "_REQ": 30756, + "_REQUEST": 14454, + "_REQUIRE": 53927, + "_REQUIRED": 53912, + "_RES": 10983, + "_RESERVED": 67593, + "_RESET": 27028, + "_RESOLUTION": 92746, + "_RESOURCE": 38208, + "_RESOURCES": 89959, + "_RESP": 28980, + "_RESPONSE": 35901, + "_REST": 55133, + "_RESULT": 21181, + "_RESULTS": 76015, + "_RET": 38748, + "_RETRY": 77924, + "_RETURN": 21909, + "_RETURNTRANSFER": 71165, + "_REUSE": 91763, + "_REV": 65351, + "_RF": 65977, + "_RG": 47304, + "_RGB": 24815, + "_RGBA": 53336, + "_RGCTX": 60863, + "_RIGHT": 21199, + "_RING": 70695, + "_RM": 86254, + "_RO": 8302, + "_ROLE": 47298, + "_ROM": 71844, + "_ROOM": 59966, + "_ROOT": 16197, + "_ROT": 62667, + "_ROUND": 78519, + "_ROUT": 69588, + "_ROUTE": 62038, + "_ROW": 38889, + "_ROWS": 62725, + "_RPC": 76022, + "_RS": 59138, + "_RSA": 76994, + "_RSP": 85642, + "_RST": 64093, + "_RT": 53563, + "_RTC": 78945, + "_RULE": 50495, + "_RUN": 24068, + "_RUNNING": 70246, + "_RUNTIME": 52198, + "_RW": 60818, + "_RX": 20430, + "_Re": 50693, + "_Read": 38381, + "_Real": 93306, + "_Record": 75329, + "_Rect": 56469, + "_Red": 92940, + "_Ref": 81086, + "_Reference": 98790, + "_Reg": 49384, + "_Register": 73124, + "_Rel": 88922, + "_Release": 85573, + "_Rem": 54551, + "_Remove": 66843, + "_Render": 42102, + "_Renderer": 82104, + "_Report": 76428, + "_Request": 44024, + "_Res": 92815, + "_Reset": 67771, + "_Resource": 86346, + "_Response": 65873, + "_Result": 49596, + "_Return": 53316, + "_Right": 73275, + "_Row": 66932, + "_Run": 84158, + "_Runtime": 63422, + "_S": 1098, + "_SA": 81219, + "_SAFE": 57970, + "_SAMPL": 85144, + "_SAMPLE": 41439, + "_SAMPLES": 89881, + "_SAN": 63123, + "_SANITIZE": 76703, + "_SAVE": 43109, + "_SB": 72209, + "_SC": 12031, + "_SCALE": 39810, + "_SCAN": 41855, + "_SCANCODE": 83753, + "_SCENE": 89468, + "_SCHED": 86208, + "_SCHEDULE": 98118, + "_SCHEMA": 46220, + "_SCL": 92315, + "_SCOPE": 61725, + "_SCORE": 63727, + "_SCR": 89405, + "_SCREEN": 42399, + "_SCRIPT": 45968, + "_SCROLL": 60430, + "_SD": 45161, + "_SDK": 84197, + "_SE": 3620, + "_SEARCH": 33708, + "_SEC": 24672, + "_SECOND": 56224, + "_SECONDS": 75351, + "_SECRET": 31408, + "_SECTION": 44228, + "_SECURE": 77848, + "_SECURITY": 71116, + "_SEG": 52272, + "_SEGMENT": 87354, + "_SEL": 40251, + "_SELECT": 23060, + "_SELECTED": 59340, + "_SELECTION": 72489, + "_SELECTOR": 72396, + "_SELF": 38847, + "_SEND": 31872, + "_SENS": 38924, + "_SENSOR": 49836, + "_SENT": 72663, + "_SEP": 92052, + "_SEPARATOR": 32243, + "_SEQ": 50045, + "_SEQUENCE": 74018, + "_SER": 14844, + "_SERIAL": 54585, + "_SERVER": 20012, + "_SERVICE": 21590, + "_SESSION": 34361, + "_SET": 8481, + "_SETTING": 58955, + "_SETTINGS": 34189, + "_SETUP": 61178, + "_SF": 84679, + "_SH": 6834, + "_SHA": 38096, + "_SHADER": 45163, + "_SHADOW": 90943, + "_SHAPE": 93138, + "_SHARE": 68923, + "_SHARED": 54718, + "_SHIFT": 22657, + "_SHORT": 16456, + "_SHOW": 37822, + "_SI": 94421, + "_SID": 79386, + "_SIDE": 77761, + "_SIG": 55687, + "_SIGN": 36727, + "_SIGNAL": 52520, + "_SIGNATURE": 73752, + "_SIM": 30920, + "_SIMPLE": 67103, + "_SINGLE": 39028, + "_SITE": 60544, + "_SIZE": 4098, + "_SK": 34464, + "_SKIP": 67756, + "_SL": 25622, + "_SLAVE": 74418, + "_SLEEP": 64733, + "_SLOT": 53047, + "_SM": 28845, + "_SMALL": 56207, + "_SMS": 87326, + "_SN": 47260, + "_SO": 45142, + "_SOC": 67978, + "_SOCKET": 39736, + "_SOFT": 91338, + "_SOL": 90705, + "_SORT": 57187, + "_SOUND": 63327, + "_SOURCE": 25430, + "_SP": 17747, + "_SPACE": 38404, + "_SPE": 88455, + "_SPEC": 36436, + "_SPECIAL": 64240, + "_SPEED": 30391, + "_SPELL": 83668, + "_SPI": 31848, + "_SPLIT": 76415, + "_SPR": 79100, + "_SQL": 46625, + "_SR": 48163, + "_SRC": 29409, + "_SRV": 95990, + "_SS": 57207, + "_SSL": 45437, + "_ST": 2402, + "_STA": 81525, + "_STACK": 21801, + "_STAGE": 64820, + "_STANDARD": 73825, + "_STAR": 70786, + "_START": 13044, + "_STARTED": 76066, + "_STAT": 16113, + "_STATE": 10832, + "_STATES": 87437, + "_STATIC": 31352, + "_STATS": 51963, + "_STATUS": 9348, + "_STD": 47166, + "_STENCIL": 85666, + "_STEP": 39968, + "_STMT": 97963, + "_STOCK": 79581, + "_STOP": 31568, + "_STORAGE": 36024, + "_STORE": 40415, + "_STR": 7159, + "_STREAM": 23584, + "_STRING": 12283, + "_STRIP": 90054, + "_STRUCT": 30474, + "_STRUCTURE": 69906, + "_STS": 98947, + "_STYLE": 41775, + "_SU": 56615, + "_SUB": 17282, + "_SUBJECT": 89187, + "_SUCCESS": 8827, + "_SUFFIX": 63825, + "_SUITE": 84314, + "_SUM": 50369, + "_SUP": 51631, + "_SUPER": 86846, + "_SUPPLY": 89416, + "_SUPPORT": 24292, + "_SUPPORTED": 52583, + "_SUR": 60050, + "_SURFACE": 91321, + "_SUS": 78174, + "_SW": 22795, + "_SWAP": 77326, + "_SWITCH": 51183, + "_SY": 11399, + "_SYM": 82933, + "_SYMBOL": 23598, + "_SYN": 75662, + "_SYNC": 39189, + "_SYS": 26783, + "_SYSTEM": 30167, + "_SZ": 62333, + "_Save": 78746, + "_Se": 26920, + "_Search": 65913, + "_Select": 58073, + "_Selected": 38290, + "_SelectedIndexChanged": 47040, + "_Selection": 88435, + "_Send": 46267, + "_Server": 62320, + "_Service": 52548, + "_Session": 84664, + "_Set": 14812, + "_Settings": 71769, + "_Sh": 36578, + "_Show": 79665, + "_Size": 45553, + "_Source": 48122, + "_Space": 60283, + "_Speed": 87058, + "_St": 70645, + "_Start": 38056, + "_State": 24773, + "_Static": 27049, + "_StaticFields": 37735, + "_Statics": 56319, + "_Status": 36449, + "_Stop": 80308, + "_Store": 92684, + "_Str": 46171, + "_Stream": 80631, + "_String": 31777, + "_Struct": 72246, + "_Style": 88038, + "_Sub": 36359, + "_Success": 87161, + "_Surface": 69702, + "_Syntax": 48429, + "_System": 25363, + "_T": 1139, + "_TA": 61850, + "_TAB": 41708, + "_TABLE": 16039, + "_TAC": 48221, + "_TAG": 16592, + "_TAGS": 84182, + "_TARGET": 29299, + "_TASK": 26315, + "_TB": 87806, + "_TBL": 79418, + "_TC": 56738, + "_TCP": 50614, + "_TD": 95862, + "_TE": 36222, + "_TEAM": 71198, + "_TEM": 90617, + "_TEMP": 41244, + "_TEMPLATE": 34786, + "_TER": 65444, + "_TERM": 75220, + "_TERMIN": 79902, + "_TEST": 11641, + "_TESTS": 80312, + "_TEX": 84171, + "_TEXT": 10243, + "_TEXTURE": 17208, + "_TH": 10625, + "_THAN": 85868, + "_THAT": 91594, + "_THE": 76594, + "_THEME": 62281, + "_THIS": 66036, + "_THREAD": 27087, + "_THREADS": 56239, + "_THRESH": 34828, + "_THRESHOLD": 44405, + "_THROW": 43897, + "_TI": 56109, + "_TICK": 57260, + "_TILE": 75810, + "_TIM": 16746, + "_TIME": 10051, + "_TIMEOUT": 23412, + "_TIMER": 31833, + "_TIMES": 92560, + "_TIMESTAMP": 47854, + "_TITLE": 22717, + "_TLS": 69067, + "_TM": 70223, + "_TMP": 71517, + "_TO": 8650, + "_TODO": 87461, + "_TOGGLE": 79755, + "_TOKEN": 18681, + "_TOO": 74555, + "_TOOL": 57188, + "_TOOLTIP": 77852, + "_TOP": 25478, + "_TOPIC": 74077, + "_TOTAL": 50831, + "_TOUCH": 74288, + "_TP": 76000, + "_TR": 6410, + "_TRA": 39096, + "_TRACE": 24238, + "_TRACK": 44050, + "_TRAIN": 81488, + "_TRAN": 30967, + "_TRANS": 28953, + "_TRANSACTION": 78509, + "_TRANSFER": 64697, + "_TRANSFORM": 92621, + "_TRANSL": 71480, + "_TREE": 46691, + "_TRI": 41283, + "_TRIANGLE": 94175, + "_TRIANGLES": 89986, + "_TRIGGER": 56714, + "_TRNS": 92705, + "_TRUE": 14044, + "_TRUNC": 94642, + "_TRY": 85373, + "_TS": 61793, + "_TUN": 87228, + "_TURN": 94583, + "_TV": 96703, + "_TW": 95747, + "_TWO": 73892, + "_TX": 18819, + "_TXT": 84411, + "_TYP": 15109, + "_TYPE": 4189, + "_TYPED": 23982, + "_TYPEDEF": 78214, + "_TYPES": 25794, + "_Tab": 92842, + "_Table": 42544, + "_Tag": 78040, + "_Target": 75823, + "_Task": 77103, + "_Template": 57917, + "_Test": 32541, + "_TestCase": 87729, + "_Text": 20550, + "_TextChanged": 45536, + "_Texture": 72185, + "_Thread": 96076, + "_Tick": 73581, + "_Time": 39080, + "_Timer": 64026, + "_Tis": 46946, + "_Title": 72001, + "_To": 38346, + "_Tool": 88451, + "_Top": 94819, + "_Total": 85933, + "_Tr": 21038, + "_Trans": 34932, + "_Tree": 90770, + "_True": 93577, + "_Two": 82989, + "_Type": 13729, + "_TypeDef": 90720, + "_TypeInfo": 30895, + "_U": 6665, + "_UART": 32358, + "_UC": 80314, + "_UClass": 67089, + "_UD": 98265, + "_UDP": 75201, + "_UFunction": 94549, + "_UI": 30522, + "_UID": 69882, + "_UINT": 27241, + "_UL": 76228, + "_UN": 6735, + "_UNDEF": 98215, + "_UNDEFINED": 76863, + "_UNDER": 92614, + "_UNICODE": 75688, + "_UNIFORM": 98426, + "_UNIQUE": 90330, + "_UNIT": 31718, + "_UNITS": 90787, + "_UNIX": 82754, + "_UNKNOWN": 31546, + "_UNLOCK": 63068, + "_UNS": 33399, + "_UNSIGNED": 51979, + "_UNSUPPORTED": 89140, + "_UNUSED": 31916, + "_UP": 14055, + "_UPDATE": 22408, + "_UPDATED": 86048, + "_UPLOAD": 63036, + "_UPPER": 91703, + "_URI": 23116, + "_URL": 8000, + "_US": 13467, + "_USAGE": 57290, + "_USART": 68162, + "_USB": 35992, + "_USE": 22295, + "_USED": 77348, + "_USER": 9107, + "_USERNAME": 42696, + "_USERS": 52875, + "_UT": 87526, + "_UTF": 55061, + "_UTIL": 49904, + "_UTILS": 87967, + "_UUID": 57499, + "_Un": 40687, + "_Unit": 80481, + "_UnityEngine": 59938, + "_Up": 88425, + "_Update": 47393, + "_User": 31339, + "_Util": 89743, + "_Utils": 91175, + "_V": 2334, + "_VAL": 6096, + "_VALID": 27335, + "_VALIDATE": 63927, + "_VALUE": 7476, + "_VALUES": 58662, + "_VAR": 25750, + "_VARIABLE": 45873, + "_VARS": 73793, + "_VC": 84252, + "_VE": 89368, + "_VEC": 70706, + "_VECTOR": 47993, + "_VENDOR": 56448, + "_VER": 14815, + "_VERBOSE": 52102, + "_VERIFY": 56042, + "_VERSION": 10678, + "_VERTEX": 47028, + "_VERTICAL": 73559, + "_VF": 98842, + "_VIDEO": 32884, + "_VIEW": 24374, + "_VIRTUAL": 76959, + "_VIS": 54442, + "_VISIBLE": 57189, + "_VLAN": 93489, + "_VM": 65396, + "_VO": 64755, + "_VOICE": 92830, + "_VOID": 57144, + "_VOL": 70415, + "_VOLT": 81373, + "_VOLUME": 60371, + "_Val": 88271, + "_Valid": 97279, + "_Value": 27867, + "_ValueChanged": 98913, + "_Var": 61251, + "_Variable": 83932, + "_Vector": 61377, + "_Version": 85217, + "_Vert": 80810, + "_View": 50711, + "_W": 2763, + "_WAIT": 32091, + "_WAKE": 76221, + "_WALL": 93435, + "_WARN": 35148, + "_WARNING": 31962, + "_WARNINGS": 83830, + "_WATCH": 96009, + "_WATER": 97850, + "_WE": 32244, + "_WEAPON": 93279, + "_WEB": 53085, + "_WEEK": 76999, + "_WEIGHT": 60697, + "_WH": 53876, + "_WHITE": 59831, + "_WIDGET": 49094, + "_WIDTH": 13947, + "_WIFI": 75470, + "_WIN": 34377, + "_WINDOW": 24780, + "_WINDOWS": 71425, + "_WITH": 23929, + "_WM": 75452, + "_WORD": 33238, + "_WORDS": 88421, + "_WORK": 34044, + "_WORLD": 38913, + "_WP": 98149, + "_WR": 21779, + "_WRAP": 45087, + "_WRAPPER": 98851, + "_WRITE": 17475, + "_WRONG": 89607, + "_WRONLY": 88649, + "_WS": 68388, + "_Widget": 58049, + "_Width": 77609, + "_Window": 60649, + "_Word": 96544, + "_Work": 87471, + "_Write": 31825, + "_X": 6859, + "_XDECREF": 71821, + "_XML": 45617, + "_Y": 10626, + "_YEAR": 57127, + "_YELLOW": 94139, + "_YES": 76730, + "_YUV": 89046, + "_Z": 12302, + "_ZERO": 39370, + "_ZONE": 60742, + "_Zero": 97672, + "_[": 11883, + "_\\": 56014, + "_]": 74827, + "_^": 67280, + "_^(": 65018, + "__": 563, + "__\":Ċ": 21082, + "__$": 41089, + "__',": 87510, + "__':Ċ": 16179, + "__':čĊ": 82777, + "__(": 3804, + "__(\"": 51669, + "__('": 23426, + "__((": 30409, + "__()": 94786, + "__()Ċ": 32616, + "__()ĊĊ": 65264, + "__(*": 68006, + "__(/*!": 51796, + "__(Ċ": 60994, + "__)": 19141, + "__))": 63553, + "__));Ċ": 97332, + "__))Ċ": 64113, + "__),": 55087, + "__).": 59043, + "__);": 87569, + "__);Ċ": 21205, + "__);ĊĊ": 64277, + "__)Ċ": 23135, + "__)ĊĊ": 40725, + "__)ĊĊĊ": 66960, + "__*/": 50993, + "__,": 10445, + "__,__": 82930, + "__,Ċ": 48566, + "__.": 18725, + "__.'/": 70837, + "__.__": 49380, + "__/": 47372, + "__:": 80342, + "__;": 54637, + "__;Ċ": 57048, + "__[": 66190, + "__[\"": 41903, + "___": 5973, + "____": 2130, + "_____": 80517, + "________": 3979, + "____________": 41383, + "________________": 7686, + "________________________________": 17466, + "________________________________________________________________": 46932, + "_________________ĊĊ": 88252, + "___ĊĊ": 60288, + "__Ċ": 12148, + "__ĊĊ": 24197, + "__čĊ": 62564, + "_a": 4306, + "_aa": 89808, + "_ab": 22885, + "_ability": 95763, + "_abort": 70907, + "_about": 57975, + "_above": 77315, + "_abs": 31170, + "_absolute": 50874, + "_abstract": 85939, + "_ac": 14718, + "_acc": 17737, + "_accel": 52638, + "_accept": 35728, + "_accepted": 90354, + "_access": 12759, + "_accessible": 88724, + "_accessor": 33901, + "_account": 13500, + "_accounts": 55665, + "_accum": 75837, + "_accuracy": 41358, + "_ack": 48447, + "_acl": 63692, + "_acquire": 97523, + "_act": 29370, + "_action": 7931, + "_actions": 25368, + "_activ": 71684, + "_activate": 67894, + "_activation": 52404, + "_active": 12930, + "_activities": 97740, + "_activity": 30026, + "_actor": 54818, + "_actual": 40149, + "_ad": 10027, + "_adapter": 43682, + "_adc": 64840, + "_add": 2891, + "_added": 37653, + "_additional": 81742, + "_addr": 7387, + "_address": 6744, + "_addresses": 59471, + "_adj": 42559, + "_adjust": 44153, + "_admin": 12207, + "_ads": 72275, + "_adv": 35656, + "_advance": 98093, + "_advanced": 83843, + "_aes": 90958, + "_af": 67213, + "_aff": 48914, + "_after": 19844, + "_ag": 20587, + "_again": 62427, + "_age": 32053, + "_agent": 25730, + "_agents": 76347, + "_agg": 83534, + "_ai": 69415, + "_air": 50827, + "_ajax": 38119, + "_ak": 72148, + "_al": 8418, + "_alarm": 66286, + "_album": 53241, + "_alert": 35717, + "_alg": 72123, + "_algo": 76993, + "_algorithm": 59531, + "_alias": 35947, + "_aliases": 90233, + "_align": 37015, + "_aligned": 66385, + "_alignment": 51006, + "_alive": 58850, + "_all": 5705, + "_alloc": 14802, + "_allocate": 77078, + "_allocated": 82342, + "_allocation": 81267, + "_allocator": 56910, + "_allow": 55731, + "_allowed": 42155, + "_almost": 94418, + "_alpha": 26633, + "_already": 80772, + "_alt": 32105, + "_altern": 79728, + "_am": 22880, + "_amount": 13471, + "_amp": 60281, + "_amt": 54605, + "_an": 12008, + "_analysis": 42682, + "_anchor": 56532, + "_and": 8378, + "_android": 56665, + "_ang": 58060, + "_angle": 21727, + "_angles": 72643, + "_anim": 58076, + "_animation": 48487, + "_ann": 51870, + "_annotation": 61237, + "_annotations": 78896, + "_ans": 60801, + "_answer": 28534, + "_answers": 61610, + "_ant": 62460, + "_any": 37248, + "_ap": 21093, + "_api": 11697, + "_app": 8191, + "_append": 26041, + "_application": 38721, + "_apply": 36551, + "_appro": 36057, + "_approval": 95613, + "_approved": 86850, + "_approx": 90425, + "_apps": 69899, + "_ar": 25842, + "_arc": 62914, + "_arch": 34330, + "_archive": 42873, + "_are": 56855, + "_area": 15030, + "_areas": 78572, + "_arg": 6057, + "_args": 8384, + "_argument": 9025, + "_arguments": 43433, + "_argv": 73563, + "_arm": 34680, + "_armor": 84451, + "_arr": 11210, + "_array": 3858, + "_arrays": 68983, + "_arrow": 46566, + "_art": 36932, + "_article": 37321, + "_articles": 71283, + "_artist": 66840, + "_ary": 82219, + "_as": 11898, + "_asc": 88474, + "_ascii": 50238, + "_ask": 92269, + "_asm": 67529, + "_aspect": 80070, + "_ass": 12083, + "_assert": 16553, + "_asset": 42299, + "_assets": 53299, + "_assign": 20688, + "_assigned": 85515, + "_assignment": 51891, + "_assoc": 18891, + "_associ": 58665, + "_ast": 48019, + "_async": 28346, + "_at": 3752, + "_atom": 42292, + "_atomic": 51367, + "_atoms": 67755, + "_att": 13356, + "_attach": 40719, + "_attached": 93561, + "_attachment": 40868, + "_attachments": 98581, + "_attack": 45697, + "_attempt": 85951, + "_attempts": 79490, + "_attention": 75825, + "_attr": 10422, + "_attrib": 68787, + "_attribute": 16791, + "_attributes": 18240, + "_attrs": 39578, + "_atts": 95770, + "_atual": 97050, + "_auc": 75080, + "_audio": 29688, + "_audit": 84437, + "_aug": 60411, + "_aut": 32808, + "_auth": 14014, + "_authenticated": 64913, + "_authentication": 89465, + "_author": 22938, + "_auto": 27740, + "_aux": 30468, + "_av": 26173, + "_avail": 76967, + "_available": 26962, + "_avatar": 51104, + "_average": 38594, + "_avg": 26631, + "_aw": 54168, + "_ax": 42385, + "_axes": 58394, + "_axis": 23567, + "_az": 95387, + "_b": 880, + "_back": 3895, + "_backend": 40011, + "_background": 24103, + "_backup": 44710, + "_backward": 70477, + "_bad": 34199, + "_bag": 74368, + "_bal": 56441, + "_balance": 29396, + "_ball": 55808, + "_band": 45344, + "_bank": 35733, + "_banner": 46571, + "_bar": 14388, + "_barang": 54601, + "_barrier": 82355, + "_base": 7651, + "_based": 66679, + "_baseline": 82065, + "_basename": 81135, + "_bases": 96290, + "_basic": 34729, + "_basis": 62696, + "_batch": 14534, + "_batches": 57755, + "_battery": 81930, + "_bb": 50255, + "_bbox": 59390, + "_bc": 65803, + "_bd": 68499, + "_be": 21263, + "_beam": 85350, + "_bed": 89079, + "_before": 23708, + "_beg": 91167, + "_begin": 23338, + "_beh": 43241, + "_behavior": 74540, + "_bel": 96761, + "_below": 82750, + "_ber": 88165, + "_best": 33101, + "_bet": 91745, + "_beta": 49292, + "_between": 48302, + "_bg": 23122, + "_bh": 90581, + "_bi": 66814, + "_bias": 36381, + "_bid": 62509, + "_big": 36386, + "_bill": 63153, + "_billing": 87807, + "_bin": 21816, + "_binary": 31761, + "_bind": 27461, + "_binding": 60815, + "_bindings": 94516, + "_bins": 44912, + "_bio": 89641, + "_birth": 52973, + "_bit": 13996, + "_bitmap": 41527, + "_bits": 20034, + "_bl": 13141, + "_black": 40272, + "_blank": 56103, + "_ble": 70713, + "_blend": 88520, + "_blk": 51969, + "_blob": 45908, + "_bloc": 71572, + "_block": 7113, + "_blocked": 89223, + "_blocking": 91449, + "_blocks": 25201, + "_blog": 35815, + "_blue": 34991, + "_blueprint": 75327, + "_bm": 91922, + "_bn": 58112, + "_bo": 56013, + "_board": 23919, + "_body": 14114, + "_bold": 95677, + "_bonus": 59541, + "_book": 26421, + "_booking": 65165, + "_books": 73084, + "_bool": 22159, + "_boolean": 46642, + "_boost": 84913, + "_boot": 52062, + "_bootstrap": 88424, + "_border": 31940, + "_bot": 38883, + "_both": 88819, + "_bottom": 25473, + "_bound": 19447, + "_boundary": 54004, + "_bounds": 36878, + "_box": 10194, + "_boxes": 38384, + "_bp": 57106, + "_br": 17682, + "_branch": 28031, + "_brand": 54858, + "_break": 39494, + "_bridge": 54944, + "_brightness": 91028, + "_broadcast": 74923, + "_browser": 53414, + "_bs": 68550, + "_bt": 48626, + "_btn": 14839, + "_bucket": 38749, + "_buckets": 84156, + "_budget": 68867, + "_buf": 10363, + "_buff": 29328, + "_buffer": 7776, + "_buffers": 54591, + "_bug": 73232, + "_build": 20801, + "_builder": 28532, + "_building": 82397, + "_builtin": 73829, + "_bulk": 71392, + "_bullet": 94619, + "_bundle": 60986, + "_bus": 25418, + "_business": 65998, + "_busy": 75959, + "_but": 86874, + "_button": 8487, + "_buttons": 41021, + "_buy": 46348, + "_bw": 64484, + "_by": 3710, + "_byte": 19737, + "_bytes": 12524, + "_c": 666, + "_ca": 49604, + "_cache": 11529, + "_cached": 64369, + "_cal": 24005, + "_calc": 38241, + "_calendar": 57665, + "_calibration": 99086, + "_call": 13429, + "_callable": 64940, + "_callback": 12519, + "_callbacks": 60148, + "_called": 27859, + "_calls": 45636, + "_cam": 42511, + "_camera": 29843, + "_campaign": 58561, + "_can": 27421, + "_cancel": 28895, + "_candidate": 62360, + "_candidates": 73553, + "_canvas": 47136, + "_cap": 16388, + "_capabilities": 92092, + "_capability": 89686, + "_capacity": 35603, + "_caps": 52955, + "_caption": 61705, + "_capture": 55148, + "_car": 26616, + "_card": 16888, + "_cards": 41357, + "_cart": 24186, + "_case": 19096, + "_cases": 41427, + "_cash": 64363, + "_cast": 5303, + "_cat": 20825, + "_catalog": 47310, + "_cate": 92049, + "_categoria": 78736, + "_categorical": 87455, + "_categories": 27542, + "_category": 11847, + "_cats": 77370, + "_cb": 16450, + "_cc": 28955, + "_cd": 43074, + "_ce": 50727, + "_cell": 16648, + "_cells": 37796, + "_cent": 32591, + "_center": 21087, + "_centers": 80179, + "_cert": 37097, + "_certificate": 63047, + "_cf": 71143, + "_cfg": 18343, + "_cg": 98353, + "_ch": 4138, + "_chain": 30583, + "_challenge": 87406, + "_chan": 45552, + "_chance": 74127, + "_change": 15947, + "_changed": 25213, + "_changes": 47526, + "_channel": 14571, + "_channels": 19291, + "_char": 9232, + "_character": 40988, + "_characters": 79060, + "_charge": 46008, + "_chars": 37418, + "_charset": 65883, + "_chart": 40961, + "_chat": 35053, + "_che": 83527, + "_check": 7200, + "_checkbox": 55244, + "_checked": 56456, + "_checker": 62715, + "_checkout": 68186, + "_checkpoint": 55208, + "_checks": 75797, + "_checksum": 64038, + "_chg": 96276, + "_chi": 98826, + "_child": 17268, + "_children": 31206, + "_chip": 49514, + "_chk": 64673, + "_choice": 31936, + "_choices": 61909, + "_choose": 77433, + "_chr": 68993, + "_chunk": 30539, + "_chunks": 65470, + "_ci": 43514, + "_cid": 68699, + "_cipher": 76692, + "_circle": 42222, + "_city": 24971, + "_ck": 89236, + "_cl": 6794, + "_claim": 84969, + "_class": 4790, + "_classes": 16833, + "_classification": 80872, + "_classifier": 63329, + "_clause": 57530, + "_clean": 19573, + "_cleanup": 42444, + "_clear": 21811, + "_cli": 47147, + "_click": 18631, + "_clicked": 30147, + "_client": 8179, + "_cliente": 44661, + "_clients": 56899, + "_clip": 44728, + "_clk": 36500, + "_clock": 23062, + "_clone": 54742, + "_close": 12704, + "_closed": 44697, + "_closure": 72823, + "_cloud": 37356, + "_clr": 98480, + "_cls": 38656, + "_cluster": 28441, + "_clusters": 46243, + "_cm": 43619, + "_cmd": 11684, + "_cmds": 89472, + "_cmos": 81597, + "_cmp": 35193, + "_cn": 52177, + "_cnt": 15982, + "_co": 11393, + "_cod": 55290, + "_code": 4136, + "_codec": 51084, + "_codegen": 32018, + "_codes": 38482, + "_codigo": 74532, + "_coef": 63929, + "_coeff": 31990, + "_coeffs": 85950, + "_coin": 75718, + "_col": 10211, + "_coll": 43597, + "_collect": 68140, + "_collection": 25019, + "_collections": 89742, + "_collision": 70375, + "_color": 6714, + "_colors": 33670, + "_colour": 56260, + "_cols": 25347, + "_column": 8744, + "_columns": 22590, + "_com": 2965, + "_comb": 34454, + "_combine": 72374, + "_combined": 89945, + "_combo": 54781, + "_comm": 25149, + "_command": 10811, + "_commands": 44151, + "_comment": 17638, + "_comments": 30359, + "_commit": 36346, + "_common": 21107, + "_community": 97508, + "_comp": 18177, + "_company": 33403, + "_compare": 32235, + "_comparison": 90797, + "_compat": 89602, + "_compile": 74170, + "_compiler": 80910, + "_complete": 27675, + "_completed": 49068, + "_completion": 60164, + "_complex": 41522, + "_component": 23406, + "_components": 23258, + "_compress": 87845, + "_compute": 57028, + "_con": 3382, + "_concat": 57478, + "_cond": 24433, + "_condition": 27656, + "_conditions": 54099, + "_conf": 16059, + "_config": 5332, + "_configs": 59150, + "_configuration": 35726, + "_configure": 75887, + "_confirm": 23800, + "_confirmation": 37681, + "_conn": 17241, + "_connect": 15720, + "_connected": 43276, + "_connection": 15866, + "_connections": 56402, + "_connector": 76393, + "_cons": 31971, + "_console": 51724, + "_const": 13610, + "_constant": 34967, + "_constants": 55642, + "_constraint": 46973, + "_constraints": 60267, + "_construct": 64803, + "_constructor": 66210, + "_consts": 48530, + "_consum": 69870, + "_consumer": 73969, + "_cont": 10260, + "_contact": 20597, + "_contacts": 64174, + "_container": 15847, + "_contains": 63598, + "_content": 7495, + "_contents": 16682, + "_context": 8467, + "_contin": 68948, + "_continue": 75297, + "_continuous": 94400, + "_contr": 62663, + "_contract": 51499, + "_contrib": 92581, + "_control": 13436, + "_controller": 21600, + "_controls": 50936, + "_conv": 22716, + "_conversion": 64132, + "_convert": 34910, + "_converter": 74073, + "_cookie": 38663, + "_cookies": 94158, + "_coord": 30096, + "_coordinate": 68649, + "_coordinates": 50084, + "_coords": 32576, + "_copy": 16096, + "_cor": 14734, + "_core": 15467, + "_cores": 93441, + "_corner": 66884, + "_corners": 82774, + "_corpus": 94226, + "_corr": 53866, + "_correct": 31550, + "_correction": 82474, + "_cos": 61192, + "_cost": 15890, + "_costs": 85946, + "_cou": 98763, + "_count": 3180, + "_counter": 15730, + "_counters": 85632, + "_countries": 91674, + "_country": 28106, + "_counts": 25977, + "_coupon": 76049, + "_course": 31726, + "_courses": 72551, + "_cov": 50028, + "_cover": 38087, + "_coverage": 73384, + "_cp": 39811, + "_cpp": 59686, + "_cpu": 21795, + "_cpus": 90673, + "_cr": 32331, + "_crc": 59084, + "_cre": 92862, + "_create": 8657, + "_created": 27288, + "_creation": 46163, + "_creator": 68532, + "_cred": 73475, + "_credentials": 47396, + "_credit": 51569, + "_crit": 92089, + "_criteria": 80990, + "_critical": 77502, + "_crop": 51453, + "_cross": 35284, + "_crossentropy": 88246, + "_crypto": 78298, + "_cs": 35372, + "_css": 25924, + "_csv": 14020, + "_ct": 26134, + "_ctl": 42891, + "_ctor": 54185, + "_ctr": 69096, + "_ctrl": 22667, + "_ctx": 15147, + "_ctxt": 93285, + "_cu": 92045, + "_cube": 64065, + "_cuda": 62706, + "_cum": 91328, + "_cur": 18956, + "_curr": 40515, + "_currency": 34710, + "_current": 11080, + "_cursor": 28601, + "_curve": 43407, + "_cust": 87652, + "_custom": 15875, + "_customer": 28840, + "_customize": 40595, + "_cut": 41317, + "_cutoff": 74382, + "_cv": 43233, + "_cycle": 39079, + "_cycles": 68970, + "_d": 814, + "_da": 47070, + "_daily": 64530, + "_damage": 40895, + "_dark": 67937, + "_dash": 84585, + "_dashboard": 55791, + "_dat": 15353, + "_data": 1769, + "_database": 27341, + "_dataframe": 77570, + "_datas": 46530, + "_dataset": 18999, + "_datasets": 81325, + "_date": 4164, + "_dates": 46382, + "_datetime": 28943, + "_datos": 87548, + "_day": 16763, + "_days": 28353, + "_db": 8685, + "_dbg": 44124, + "_dc": 55847, + "_dd": 43174, + "_de": 2259, + "_dead": 53427, + "_death": 78294, + "_debug": 15446, + "_dec": 13783, + "_decay": 43260, + "_decimal": 74429, + "_decision": 90636, + "_deck": 83860, + "_decl": 35814, + "_declaration": 77926, + "_decode": 15227, + "_decoder": 49843, + "_decor": 70369, + "_decorator": 87252, + "_decrypt": 80764, + "_deep": 87044, + "_def": 7844, + "_default": 9993, + "_defaults": 42290, + "_define": 51049, + "_defined": 52870, + "_definition": 31698, + "_definitions": 84881, + "_defs": 56535, + "_deg": 46057, + "_degree": 49524, + "_deinit": 97435, + "_del": 18029, + "_delay": 22198, + "_delegate": 55312, + "_delete": 11353, + "_deleted": 39418, + "_delivery": 50562, + "_delta": 26710, + "_dem": 69403, + "_demand": 91295, + "_demo": 47830, + "_den": 49522, + "_dense": 72619, + "_density": 53349, + "_dep": 49258, + "_depart": 69865, + "_department": 71886, + "_depend": 34161, + "_dependencies": 71841, + "_dependency": 62387, + "_deploy": 91890, + "_deposit": 85798, + "_deps": 76489, + "_dept": 93868, + "_depth": 19061, + "_der": 35345, + "_deriv": 76546, + "_derivative": 95930, + "_des": 15768, + "_desc": 10986, + "_descr": 64622, + "_description": 11448, + "_descriptor": 33774, + "_design": 49803, + "_dest": 25264, + "_destination": 56344, + "_destroy": 18066, + "_det": 24409, + "_detach": 93067, + "_detail": 14480, + "_details": 13260, + "_detalle": 95471, + "_detect": 56457, + "_detected": 98876, + "_detection": 57505, + "_detector": 71540, + "_dev": 10433, + "_device": 9204, + "_devices": 41334, + "_df": 10894, + "_di": 29477, + "_diag": 61304, + "_dialog": 24331, + "_dic": 39154, + "_dice": 97196, + "_dict": 5243, + "_dictionary": 42605, + "_dicts": 88961, + "_die": 61588, + "_diff": 15850, + "_difference": 47525, + "_different": 82741, + "_digest": 52994, + "_digit": 48403, + "_digits": 41432, + "_dim": 10791, + "_dimension": 49619, + "_dimensions": 59699, + "_dims": 29267, + "_dir": 4334, + "_direct": 32871, + "_direction": 25438, + "_directory": 14846, + "_dirs": 44869, + "_dirty": 63642, + "_dis": 9932, + "_disable": 33842, + "_disabled": 51401, + "_disc": 37745, + "_disconnect": 67972, + "_discount": 42879, + "_disk": 41687, + "_disp": 50692, + "_dispatch": 42991, + "_dispatcher": 92189, + "_display": 14825, + "_dist": 16031, + "_distance": 19464, + "_distances": 77329, + "_distribution": 41465, + "_district": 75440, + "_div": 16237, + "_dl": 50811, + "_dll": 94626, + "_dm": 57182, + "_dma": 39476, + "_dn": 69799, + "_dns": 71125, + "_do": 26309, + "_doc": 18869, + "_docs": 49692, + "_document": 26231, + "_documento": 83061, + "_documents": 75927, + "_does": 96374, + "_dom": 35749, + "_domain": 20111, + "_domains": 70199, + "_don": 73974, + "_done": 24390, + "_door": 78962, + "_dot": 30941, + "_double": 24598, + "_down": 13998, + "_download": 35939, + "_dp": 53872, + "_dr": 26680, + "_draft": 91339, + "_drag": 72095, + "_draw": 23021, + "_drawer": 94507, + "_drive": 67151, + "_driver": 20602, + "_drop": 29584, + "_dropdown": 51024, + "_dropout": 74605, + "_drv": 39136, + "_drvdata": 79544, + "_ds": 35362, + "_dst": 33114, + "_dt": 27516, + "_dtype": 64886, + "_dual": 96772, + "_due": 75728, + "_dummy": 60321, + "_dump": 18296, + "_dup": 51932, + "_duplicate": 70434, + "_duplicates": 75051, + "_dur": 82289, + "_duration": 25454, + "_dw": 79503, + "_dx": 61984, + "_dy": 82635, + "_dyn": 69213, + "_dynamic": 45992, + "_e": 2204, + "_each": 32046, + "_easy": 59091, + "_ec": 36844, + "_echo": 68628, + "_ed": 32370, + "_edge": 17932, + "_edges": 28026, + "_edit": 13156, + "_editor": 33752, + "_ef": 79410, + "_eff": 51945, + "_effect": 27125, + "_effects": 83171, + "_eg": 90209, + "_el": 21602, + "_elapsed": 73737, + "_ele": 61365, + "_elem": 28179, + "_element": 7894, + "_elements": 22801, + "_elems": 86217, + "_elim": 84655, + "_else": 62628, + "_elt": 91034, + "_em": 22504, + "_email": 9172, + "_emails": 77773, + "_emb": 56864, + "_embed": 24007, + "_embedding": 51502, + "_embeddings": 63772, + "_emit": 69082, + "_emlrt": 92079, + "_emp": 56766, + "_employee": 51781, + "_empresa": 71787, + "_empty": 15124, + "_en": 6205, + "_enable": 18988, + "_enabled": 18220, + "_enc": 13781, + "_encode": 11224, + "_encoded": 73069, + "_encoder": 39068, + "_encoding": 37613, + "_encrypt": 66593, + "_end": 6213, + "_endian": 87193, + "_endpoint": 36699, + "_ends": 90729, + "_enemy": 73682, + "_energy": 34795, + "_eng": 62102, + "_engine": 24823, + "_enqueue": 28395, + "_ent": 27425, + "_enter": 37480, + "_entities": 47377, + "_entity": 19169, + "_entries": 26092, + "_entropy": 50374, + "_entry": 9078, + "_enum": 31054, + "_env": 15879, + "_environment": 51774, + "_eof": 90792, + "_ep": 12476, + "_epi": 41359, + "_episode": 72224, + "_episodes": 97007, + "_epoch": 20682, + "_epochs": 44127, + "_eps": 74996, + "_epsilon": 98983, + "_eq": 10714, + "_equ": 41443, + "_equal": 11478, + "_equalTo": 55522, + "_equals": 61664, + "_equiv": 71849, + "_er": 80604, + "_erase": 85516, + "_err": 9266, + "_errno": 37363, + "_error": 4096, + "_errors": 20196, + "_es": 33741, + "_esc": 92196, + "_escape": 21832, + "_est": 18583, + "_estado": 62064, + "_estimate": 78718, + "_estimator": 91968, + "_estimators": 81054, + "_et": 45668, + "_eta": 54925, + "_eth": 57757, + "_ev": 43713, + "_eval": 21296, + "_evaluation": 86505, + "_even": 68347, + "_event": 6748, + "_events": 19691, + "_every": 64048, + "_evt": 60019, + "_ex": 2702, + "_exact": 71084, + "_exam": 83334, + "_example": 39304, + "_examples": 45279, + "_exc": 45647, + "_excel": 51242, + "_except": 54633, + "_exception": 17499, + "_exceptions": 80036, + "_excerpt": 66947, + "_exchange": 59212, + "_exclude": 88668, + "_exe": 76074, + "_exec": 18430, + "_execute": 44329, + "_execution": 60948, + "_executor": 81207, + "_exempt": 95736, + "_exist": 35906, + "_existing": 62630, + "_exists": 9766, + "_exit": 16880, + "_exp": 14214, + "_expand": 67875, + "_expect": 68918, + "_expected": 32190, + "_experience": 91142, + "_experiment": 82302, + "_expire": 83409, + "_expired": 80221, + "_expiry": 96509, + "_export": 27114, + "_exports": 45309, + "_expr": 21915, + "_expression": 28068, + "_ext": 9927, + "_extend": 70265, + "_extended": 61678, + "_extension": 31035, + "_extensions": 60498, + "_extent": 70102, + "_external": 47432, + "_extra": 31858, + "_extract": 39123, + "_extraction": 94842, + "_extractor": 89821, + "_eye": 76087, + "_f": 761, + "_fa": 95864, + "_fac": 41589, + "_face": 30985, + "_facebook": 81487, + "_faces": 59848, + "_fact": 47805, + "_factor": 18588, + "_factors": 73040, + "_factory": 24269, + "_fail": 22121, + "_failed": 35060, + "_failure": 43618, + "_fake": 56881, + "_false": 36015, + "_family": 26823, + "_far": 67513, + "_fast": 35743, + "_fatal": 92188, + "_fault": 70097, + "_favorite": 96744, + "_fb": 50796, + "_fc": 46506, + "_fd": 17676, + "_fds": 90438, + "_fe": 7642, + "_feat": 46753, + "_feats": 70646, + "_feature": 17069, + "_featured": 79420, + "_features": 14965, + "_fecha": 72517, + "_fee": 34305, + "_feed": 42390, + "_feedback": 58138, + "_female": 97827, + "_fence": 81073, + "_fetch": 11803, + "_ff": 58708, + "_fft": 67059, + "_fg": 69473, + "_fh": 90808, + "_fid": 92247, + "_field": 5013, + "_fields": 12132, + "_fifo": 56590, + "_fig": 62141, + "_figure": 82731, + "_fil": 78200, + "_file": 2458, + "_filename": 13323, + "_filenames": 80632, + "_filepath": 53490, + "_files": 10931, + "_fill": 30728, + "_filled": 90592, + "_filt": 92220, + "_filter": 8727, + "_filtered": 51429, + "_filters": 22481, + "_fin": 39737, + "_final": 20676, + "_finalize": 70616, + "_find": 21814, + "_finder": 87402, + "_finish": 42980, + "_finished": 56423, + "_fire": 45772, + "_firestore": 89824, + "_first": 12978, + "_firstname": 92095, + "_fit": 32416, + "_fitness": 85346, + "_five": 95258, + "_fix": 36060, + "_fixed": 37839, + "_fixture": 74409, + "_fk": 65886, + "_fl": 5081, + "_flag": 10933, + "_flags": 14130, + "_flash": 30225, + "_flashdata": 48843, + "_flat": 49556, + "_flg": 86258, + "_flight": 90370, + "_flip": 56012, + "_float": 17586, + "_floor": 60044, + "_flow": 27441, + "_flush": 39213, + "_flutter": 68089, + "_flux": 66661, + "_fm": 78694, + "_fmt": 38128, + "_fn": 15246, + "_fname": 42334, + "_focus": 47492, + "_fold": 61187, + "_folder": 15268, + "_folders": 83928, + "_follow": 43490, + "_font": 17451, + "_fonts": 71502, + "_food": 58580, + "_foot": 75759, + "_footer": 25966, + "_for": 5478, + "_force": 40739, + "_fore": 35563, + "_foreign": 78983, + "_form": 7915, + "_format": 8955, + "_formats": 68351, + "_formatted": 86610, + "_formatter": 73965, + "_forms": 64282, + "_formula": 63869, + "_forum": 78327, + "_forward": 32121, + "_found": 21480, + "_four": 56142, + "_fp": 34160, + "_fps": 65163, + "_fr": 41537, + "_frac": 70358, + "_fraction": 59023, + "_frag": 80817, + "_fragment": 43012, + "_frame": 8929, + "_frames": 29319, + "_framework": 27565, + "_fre": 74179, + "_free": 8905, + "_freq": 21790, + "_frequency": 40132, + "_friend": 60580, + "_friends": 75430, + "_frm": 82239, + "_from": 5673, + "_front": 22926, + "_frontend": 93833, + "_fs": 34470, + "_fsm": 97069, + "_ft": 41572, + "_fu": 36467, + "_full": 16372, + "_fun": 28315, + "_func": 9596, + "_funcs": 64124, + "_function": 9174, + "_functions": 31708, + "_future": 59740, + "_fw": 57449, + "_fwd": 81083, + "_fx": 62691, + "_g": 1889, + "_gain": 39925, + "_gallery": 51867, + "_game": 18547, + "_games": 67370, + "_gamma": 61179, + "_gap": 51790, + "_gas": 82116, + "_gate": 54798, + "_gateway": 64049, + "_gb": 87935, + "_gc": 49423, + "_gchandle": 75833, + "_ge": 32933, + "_gem": 87135, + "_gen": 16322, + "_gender": 62821, + "_gene": 59602, + "_gener": 71963, + "_general": 39177, + "_generate": 48851, + "_generated": 67313, + "_generation": 64191, + "_generator": 25813, + "_generic": 41232, + "_genes": 83472, + "_genre": 91322, + "_geo": 60308, + "_geom": 72987, + "_geometry": 57022, + "_get": 3062, + "_gettime": 98221, + "_ghost": 98784, + "_gid": 72893, + "_gift": 94928, + "_git": 68801, + "_given": 57653, + "_gl": 34569, + "_glob": 95133, + "_global": 19296, + "_globals": 58775, + "_glyph": 88703, + "_go": 25515, + "_goal": 41264, + "_goals": 96244, + "_gold": 54696, + "_good": 44781, + "_goods": 45095, + "_google": 47155, + "_goto": 97732, + "_gp": 75196, + "_gpio": 32001, + "_gps": 83407, + "_gpu": 35628, + "_gr": 15682, + "_grad": 18352, + "_grade": 50515, + "_gradient": 49482, + "_gradients": 96972, + "_graph": 14738, + "_graphics": 84385, + "_gray": 51331, + "_greater": 97994, + "_green": 48254, + "_grid": 15604, + "_ground": 72492, + "_group": 6288, + "_groups": 21148, + "_growth": 49182, + "_grp": 59448, + "_grupo": 96110, + "_gs": 96930, + "_gshared": 18571, + "_gt": 37479, + "_guard": 36796, + "_guess": 54737, + "_guest": 62739, + "_gui": 46155, + "_guid": 43756, + "_guide": 80407, + "_h": 1523, + "_hal": 52580, + "_half": 40626, + "_hand": 23194, + "_handle": 10630, + "_handler": 10183, + "_handlers": 57037, + "_handles": 68017, + "_handling": 75642, + "_hard": 77712, + "_has": 21778, + "_hash": 8950, + "_hashes": 91616, + "_hat": 39773, + "_have": 67301, + "_hd": 60111, + "_hdl": 66154, + "_hdr": 28238, + "_he": 41876, + "_head": 13138, + "_header": 8757, + "_headers": 26719, + "_heading": 47242, + "_heads": 76320, + "_health": 45727, + "_heap": 33059, + "_heat": 69981, + "_height": 9561, + "_hello": 96724, + "_help": 26926, + "_helper": 10418, + "_helpers": 54473, + "_here": 73271, + "_hero": 78716, + "_hex": 32655, + "_hi": 47225, + "_hid": 87813, + "_hidden": 26739, + "_hide": 49735, + "_hierarchy": 95043, + "_high": 22680, + "_highlight": 74546, + "_hint": 45825, + "_hist": 36399, + "_histogram": 68564, + "_history": 19802, + "_hit": 37697, + "_hits": 63832, + "_hold": 65824, + "_holder": 50788, + "_hom": 57105, + "_home": 21653, + "_hook": 32005, + "_hooks": 73113, + "_hop": 98193, + "_hor": 58046, + "_horizontal": 66469, + "_host": 12848, + "_hostname": 63840, + "_hosts": 61129, + "_hot": 33433, + "_hour": 32292, + "_hours": 41371, + "_house": 64270, + "_hover": 53445, + "_hp": 48478, + "_hpp": 74277, + "_hr": 70138, + "_href": 62988, + "_hresult": 65509, + "_hs": 85956, + "_ht": 49086, + "_html": 9564, + "_http": 25888, + "_https": 26817, + "_hub": 93996, + "_human": 86247, + "_hw": 28792, + "_hyper": 93416, + "_hz": 95676, + "_i": 5318, + "_ib": 66470, + "_ic": 32747, + "_icall": 82384, + "_icon": 15648, + "_icons": 59819, + "_id": 842, + "_ident": 38399, + "_identifier": 33176, + "_identity": 46244, + "_idle": 47810, + "_ids": 8077, + "_idx": 7258, + "_idxs": 68646, + "_ie": 62203, + "_if": 11119, + "_iface": 67666, + "_iff": 36172, + "_ignore": 58493, + "_ij": 70331, + "_il": 26743, + "_im": 17895, + "_imag": 88609, + "_image": 4954, + "_images": 13283, + "_img": 8918, + "_imgs": 60825, + "_imm": 71370, + "_imp": 36788, + "_impl": 21007, + "_import": 18434, + "_in": 1243, + "_inactive": 90882, + "_inc": 31285, + "_inches": 73258, + "_include": 37878, + "_income": 67377, + "_increase": 73807, + "_increment": 51482, + "_ind": 9122, + "_indent": 69045, + "_index": 3560, + "_indexes": 50161, + "_indicator": 58070, + "_indices": 18333, + "_individual": 88873, + "_inds": 56634, + "_indx": 97587, + "_inf": 26051, + "_info": 3109, + "_information": 35212, + "_infos": 47779, + "_ing": 59112, + "_ini": 50506, + "_inicio": 65414, + "_init": 6137, + "_initial": 15809, + "_initialize": 40889, + "_initialized": 62421, + "_initializer": 36462, + "_inline": 41871, + "_inner": 34345, + "_inode": 46748, + "_inp": 78887, + "_input": 5898, + "_inputs": 28557, + "_ins": 34386, + "_insert": 17678, + "_inside": 72860, + "_insn": 79799, + "_inst": 17740, + "_install": 34245, + "_installed": 79065, + "_instance": 11904, + "_instances": 41504, + "_instr": 62574, + "_instruction": 54923, + "_instructions": 82427, + "_int": 4042, + "_integer": 31725, + "_integral": 64076, + "_integration": 90250, + "_intensity": 80187, + "_intent": 94249, + "_inter": 15318, + "_interaction": 97394, + "_interest": 62527, + "_interface": 20546, + "_interfaces": 72960, + "_internal": 23472, + "_interp": 68711, + "_interrupt": 42606, + "_intersect": 72747, + "_intersection": 82558, + "_interval": 20541, + "_intervals": 90807, + "_intf": 91290, + "_into": 45514, + "_intr": 59888, + "_intro": 68520, + "_inv": 19996, + "_invalid": 31433, + "_inventory": 50473, + "_inverse": 63333, + "_invite": 94910, + "_invoice": 39225, + "_invoke": 79779, + "_io": 16939, + "_ioctl": 59683, + "_ios": 59098, + "_ip": 10385, + "_ipc": 83806, + "_ips": 71074, + "_ipv": 49378, + "_ir": 51433, + "_irq": 22623, + "_is": 6892, + "_iso": 49660, + "_isr": 88985, + "_issue": 53340, + "_issues": 90232, + "_it": 14631, + "_item": 5634, + "_items": 12134, + "_iter": 11723, + "_iteration": 62772, + "_iterations": 55607, + "_iterator": 13491, + "_iters": 72866, + "_itr": 71938, + "_iv": 51411, + "_ix": 62686, + "_j": 5374, + "_jButton": 61750, + "_java": 77323, + "_jet": 93169, + "_job": 20298, + "_jobs": 37247, + "_join": 31017, + "_joint": 53158, + "_journal": 85075, + "_js": 26250, + "_json": 9455, + "_jump": 65436, + "_jwt": 95576, + "_k": 4698, + "_kategori": 81270, + "_kb": 95638, + "_ke": 45476, + "_keep": 50293, + "_keeper": 43404, + "_kel": 59222, + "_kelas": 95914, + "_kernel": 26876, + "_key": 3097, + "_keyboard": 58797, + "_keys": 12631, + "_keyword": 45824, + "_keywords": 51354, + "_kill": 73327, + "_kind": 33162, + "_km": 88680, + "_known": 71690, + "_kses": 96796, + "_kv": 97066, + "_kw": 64496, + "_kwargs": 36235, + "_l": 907, + "_la": 79633, + "_lab": 51828, + "_label": 6106, + "_labels": 14547, + "_lahir": 85367, + "_lambda": 51884, + "_land": 60506, + "_lane": 60302, + "_lang": 17876, + "_language": 29021, + "_languages": 77576, + "_large": 45228, + "_last": 12195, + "_lastname": 93212, + "_lat": 26174, + "_latency": 74832, + "_latest": 64880, + "_latitude": 80364, + "_launch": 51617, + "_launcher": 79277, + "_layer": 12680, + "_layers": 26091, + "_layout": 14466, + "_lazy": 49646, + "_lb": 63601, + "_lbl": 54319, + "_lc": 93057, + "_lcd": 87172, + "_ld": 50573, + "_le": 11751, + "_lead": 87052, + "_leader": 79991, + "_leaf": 38909, + "_learn": 98948, + "_learning": 67291, + "_least": 88716, + "_leave": 62666, + "_led": 38367, + "_left": 9579, + "_leg": 64122, + "_legacy": 97555, + "_legal": 82324, + "_legend": 76612, + "_len": 6043, + "_length": 5118, + "_lengths": 54416, + "_lens": 81642, + "_less": 50747, + "_letter": 46117, + "_letters": 67330, + "_level": 8274, + "_levels": 37819, + "_lex": 74547, + "_lhs": 84284, + "_li": 50450, + "_lib": 16142, + "_library": 39461, + "_license": 63839, + "_life": 72491, + "_lifetime": 98827, + "_lift": 87004, + "_light": 28663, + "_like": 25535, + "_likelihood": 95999, + "_likes": 89178, + "_lim": 58191, + "_limit": 14763, + "_limits": 31820, + "_lin": 24606, + "_line": 6528, + "_linear": 40674, + "_lineno": 31142, + "_lines": 18323, + "_link": 7233, + "_linked": 76145, + "_links": 25258, + "_linux": 77463, + "_list": 2019, + "_lista": 76386, + "_listen": 79286, + "_listener": 46493, + "_listing": 64214, + "_lists": 37288, + "_lit": 98399, + "_lite": 89191, + "_literal": 34100, + "_literals": 49643, + "_live": 55203, + "_ll": 42995, + "_lm": 90004, + "_ln": 60401, + "_lng": 89680, + "_lo": 5560, + "_load": 12411, + "_loaded": 49205, + "_loader": 22139, + "_loading": 57726, + "_loan": 94900, + "_loc": 13400, + "_local": 13564, + "_locale": 55518, + "_locals": 94552, + "_location": 13126, + "_locations": 45032, + "_locator": 75344, + "_lock": 9818, + "_locked": 60271, + "_locs": 96913, + "_log": 5224, + "_logged": 47333, + "_logger": 27413, + "_logging": 59982, + "_logic": 54335, + "_logical": 86484, + "_login": 13681, + "_logits": 59555, + "_logo": 29962, + "_logout": 69386, + "_logs": 42899, + "_lon": 57688, + "_long": 17799, + "_longitude": 79734, + "_look": 24271, + "_lookup": 27464, + "_loop": 17198, + "_loss": 11193, + "_losses": 66162, + "_lost": 88661, + "_lot": 91849, + "_low": 23767, + "_lower": 30425, + "_lowercase": 93129, + "_lp": 76529, + "_lr": 46086, + "_ls": 53174, + "_lst": 31983, + "_lstm": 96397, + "_lt": 39164, + "_lua": 76013, + "_lv": 55469, + "_lvl": 67975, + "_ly": 93337, + "_m": 717, + "_mA": 97461, + "_mB": 94665, + "_mC": 97335, + "_mD": 97955, + "_mE": 96717, + "_ma": 69586, + "_mac": 22802, + "_machine": 38695, + "_macro": 58810, + "_macros": 85904, + "_mag": 48310, + "_magic": 54612, + "_mail": 25176, + "_main": 11027, + "_major": 47916, + "_make": 28230, + "_makeConstraints": 63137, + "_maker": 96040, + "_male": 95789, + "_malloc": 38303, + "_man": 17062, + "_manage": 56473, + "_managed": 79889, + "_management": 45363, + "_manager": 12144, + "_manifest": 74269, + "_manual": 75428, + "_many": 22101, + "_map": 5376, + "_mapped": 91933, + "_mapper": 76664, + "_mapping": 26930, + "_mappings": 84153, + "_maps": 46512, + "_mar": 38063, + "_margin": 42644, + "_mark": 18924, + "_marker": 26267, + "_markers": 92082, + "_market": 68725, + "_marks": 74204, + "_markup": 54449, + "_marshaled": 81365, + "_marshall": 81007, + "_mas": 81641, + "_mask": 9999, + "_masks": 62935, + "_mass": 36714, + "_master": 24582, + "_mat": 16610, + "_match": 10708, + "_matched": 93900, + "_matches": 38344, + "_matching": 70763, + "_material": 33884, + "_math": 65561, + "_matrices": 95190, + "_matrix": 10193, + "_max": 6345, + "_maximum": 93581, + "_mb": 53812, + "_mc": 43357, + "_md": 32994, + "_me": 16612, + "_mean": 16933, + "_means": 80866, + "_meas": 95786, + "_measure": 48938, + "_measurement": 87342, + "_med": 44085, + "_media": 29173, + "_median": 83003, + "_medium": 54867, + "_mem": 12976, + "_member": 19388, + "_members": 30397, + "_membership": 84635, + "_memcpy": 78038, + "_memory": 19195, + "_ment": 26985, + "_mentions": 30063, + "_menu": 10403, + "_menus": 83333, + "_merge": 20888, + "_merged": 90702, + "_mes": 69159, + "_mesh": 29453, + "_message": 6462, + "_messages": 23428, + "_met": 44517, + "_meta": 13381, + "_metadata": 22220, + "_meter": 79338, + "_method": 9032, + "_methods": 36084, + "_metric": 41294, + "_metrics": 37686, + "_mex": 59930, + "_mgmt": 96014, + "_mgr": 42438, + "_mi": 86013, + "_micro": 73003, + "_mid": 43733, + "_middle": 62580, + "_migration": 90373, + "_mime": 83624, + "_min": 7260, + "_mini": 84821, + "_minimum": 84855, + "_minor": 70861, + "_minus": 38457, + "_minute": 76944, + "_minutes": 67655, + "_mirror": 88593, + "_misc": 69350, + "_miss": 67722, + "_missing": 40447, + "_mix": 67729, + "_mk": 64861, + "_ml": 76567, + "_mm": 35599, + "_mo": 79258, + "_mob": 85069, + "_mobile": 34491, + "_mock": 34134, + "_mod": 7480, + "_modal": 43788, + "_mode": 7302, + "_model": 5047, + "_models": 30792, + "_modes": 59724, + "_modified": 37749, + "_modifier": 61773, + "_modify": 57725, + "_module": 10750, + "_modules": 15848, + "_mon": 20737, + "_money": 34065, + "_monitor": 40112, + "_mono": 60080, + "_month": 18933, + "_months": 88462, + "_more": 36664, + "_most": 81416, + "_mot": 78584, + "_motion": 65460, + "_motor": 59789, + "_mount": 57593, + "_mouse": 39203, + "_mov": 55798, + "_move": 17134, + "_movement": 70325, + "_moves": 45390, + "_movie": 50729, + "_movies": 77631, + "_mp": 39898, + "_mpi": 83812, + "_mr": 92858, + "_ms": 21416, + "_msg": 6483, + "_msgs": 20777, + "_mt": 65944, + "_mtime": 93674, + "_mtx": 85416, + "_mu": 43521, + "_mul": 24944, + "_mult": 26290, + "_multi": 25133, + "_multip": 91802, + "_multiple": 45233, + "_multiplier": 51800, + "_multiply": 93054, + "_music": 61863, + "_mut": 29523, + "_mutex": 14204, + "_mux": 80363, + "_mv": 73187, + "_mx": 83185, + "_my": 35686, + "_mysql": 59992, + "_n": 1089, + "_na": 58631, + "_nama": 86536, + "_name": 1269, + "_named": 71834, + "_names": 9187, + "_namespace": 41571, + "_nan": 73936, + "_nat": 38169, + "_native": 44494, + "_nav": 28233, + "_navigation": 58051, + "_nb": 43675, + "_nbr": 78272, + "_nc": 52382, + "_nd": 43544, + "_ne": 13925, + "_near": 76340, + "_need": 71506, + "_needed": 57426, + "_neg": 28209, + "_negative": 53865, + "_neighbor": 71888, + "_neighbors": 55925, + "_nested": 66279, + "_net": 19722, + "_network": 20966, + "_neurons": 96156, + "_new": 5921, + "_news": 32485, + "_next": 11257, + "_nf": 78225, + "_ng": 82102, + "_nh": 93934, + "_nick": 89564, + "_nil": 36175, + "_nl": 81052, + "_nm": 57572, + "_nn": 69962, + "_no": 6536, + "_node": 5084, + "_nodes": 14896, + "_noise": 40313, + "_nom": 53475, + "_nombre": 50198, + "_nome": 76144, + "_non": 21637, + "_nonce": 48508, + "_none": 31488, + "_norm": 19181, + "_normal": 13973, + "_normalize": 80807, + "_normalized": 80006, + "_normals": 91029, + "_not": 7913, + "_note": 27207, + "_notes": 45138, + "_notice": 51807, + "_notification": 34296, + "_notifications": 67546, + "_notifier": 86179, + "_notify": 36654, + "_now": 20813, + "_np": 41760, + "_npc": 97981, + "_nr": 36513, + "_ns": 34728, + "_nsec": 64988, + "_nt": 38945, + "_nth": 78342, + "_null": 15162, + "_nullable": 69743, + "_num": 4273, + "_number": 5500, + "_numbers": 32964, + "_numer": 99001, + "_numeric": 29418, + "_numero": 96728, + "_numpy": 43034, + "_nums": 51875, + "_nv": 79430, + "_o": 14179, + "_oauth": 91193, + "_ob": 36322, + "_obj": 7328, + "_object": 5314, + "_objects": 24873, + "_objs": 62673, + "_obs": 30405, + "_observer": 81912, + "_oc": 49648, + "_occ": 57291, + "_oct": 70135, + "_od": 51797, + "_odd": 74668, + "_of": 3575, + "_off": 13651, + "_offer": 67814, + "_office": 79604, + "_offset": 6917, + "_offsets": 56924, + "_oid": 59633, + "_ok": 19817, + "_old": 21108, + "_on": 4470, + "_once": 7630, + "_one": 11667, + "_online": 51546, + "_only": 18410, + "_op": 10287, + "_opacity": 96707, + "_opcode": 71319, + "_open": 11311, + "_oper": 25634, + "_operand": 69259, + "_operation": 32565, + "_operations": 57345, + "_operator": 40594, + "_ops": 21959, + "_opt": 15032, + "_optimizer": 68168, + "_option": 9672, + "_optional": 74644, + "_options": 8743, + "_opts": 32354, + "_or": 8734, + "_ord": 67324, + "_order": 7869, + "_ordered": 75272, + "_orders": 37129, + "_org": 35583, + "_organization": 82352, + "_ori": 85525, + "_orient": 53493, + "_orientation": 66269, + "_orig": 35328, + "_origin": 34043, + "_original": 39556, + "_os": 29387, + "_ot": 65614, + "_other": 30456, + "_out": 6068, + "_outer": 67258, + "_outline": 72980, + "_output": 7645, + "_outputs": 35189, + "_ov": 83981, + "_over": 15431, + "_overflow": 79073, + "_overlap": 65794, + "_overlay": 58189, + "_override": 48576, + "_own": 82100, + "_owned": 51973, + "_owner": 29027, + "_p": 620, + "_pa": 55475, + "_pack": 32995, + "_package": 26328, + "_packages": 41874, + "_packet": 21078, + "_packets": 63569, + "_pad": 30290, + "_padding": 40726, + "_pag": 51003, + "_page": 6129, + "_pages": 21655, + "_pagination": 82231, + "_pago": 79179, + "_paid": 63207, + "_paint": 92284, + "_pair": 14445, + "_pairs": 36430, + "_pal": 73310, + "_palette": 66252, + "_pan": 71099, + "_panel": 24570, + "_paper": 88917, + "_par": 22654, + "_para": 48080, + "_paragraph": 96713, + "_parallel": 60625, + "_param": 4090, + "_parameter": 24899, + "_parameters": 18263, + "_params": 6745, + "_parent": 15960, + "_parents": 81719, + "_parm": 97705, + "_parms": 98935, + "_pars": 77113, + "_parse": 21039, + "_parsed": 75788, + "_parser": 18517, + "_part": 10495, + "_partial": 52068, + "_particle": 70203, + "_particles": 67560, + "_partition": 43840, + "_partitions": 99097, + "_partner": 72432, + "_parts": 33217, + "_party": 44021, + "_pas": 79578, + "_pass": 15464, + "_passed": 87405, + "_passwd": 88667, + "_password": 10122, + "_past": 93019, + "_pat": 55824, + "_patch": 39643, + "_patches": 98519, + "_path": 2638, + "_paths": 24152, + "_patient": 70310, + "_pattern": 21260, + "_patterns": 64923, + "_pause": 59989, + "_pay": 28925, + "_payload": 32813, + "_payment": 26696, + "_payments": 91257, + "_pb": 31409, + "_pc": 35612, + "_pci": 76384, + "_pcm": 59655, + "_pct": 71512, + "_pd": 31322, + "_pdata": 96825, + "_pdf": 39948, + "_pdu": 95249, + "_pe": 36367, + "_peak": 66709, + "_ped": 93104, + "_pedido": 96619, + "_peer": 45159, + "_pel": 82852, + "_pemb": 92883, + "_pen": 38114, + "_penalty": 82286, + "_pending": 37861, + "_peng": 65424, + "_people": 62717, + "_per": 5678, + "_perc": 82812, + "_percent": 29883, + "_percentage": 46044, + "_perf": 76776, + "_performance": 84833, + "_period": 20818, + "_periods": 95193, + "_perm": 31961, + "_permalink": 44854, + "_permission": 37971, + "_permissions": 44767, + "_perms": 98043, + "_person": 23417, + "_persona": 99118, + "_personal": 78035, + "_pes": 92195, + "_pet": 75963, + "_pf": 74961, + "_pg": 58681, + "_ph": 9782, + "_phase": 37593, + "_phi": 42009, + "_phone": 22154, + "_photo": 32385, + "_photos": 77468, + "_php": 78547, + "_phr": 57729, + "_phrase": 66696, + "_phy": 50266, + "_phys": 69319, + "_physical": 84980, + "_pi": 47771, + "_pic": 29473, + "_pick": 58168, + "_picker": 68230, + "_pickle": 73661, + "_picture": 42406, + "_pid": 30065, + "_piece": 48470, + "_pieces": 83479, + "_pin": 26296, + "_ping": 71661, + "_pins": 90089, + "_pipe": 41862, + "_pipeline": 45187, + "_pitch": 51959, + "_pix": 50646, + "_pixel": 29760, + "_pixels": 49745, + "_pk": 33321, + "_pkg": 64739, + "_pkt": 42051, + "_pl": 6317, + "_place": 34548, + "_placeholder": 60383, + "_placement": 90203, + "_places": 57446, + "_plain": 41015, + "_plan": 26564, + "_plane": 46023, + "_planes": 75340, + "_plate": 79366, + "_platform": 34260, + "_play": 22144, + "_player": 15524, + "_players": 46721, + "_playing": 83857, + "_playlist": 69267, + "_pll": 98812, + "_plot": 24351, + "_plots": 87795, + "_plugin": 23110, + "_plugins": 45658, + "_plural": 57427, + "_plus": 28043, + "_pm": 46507, + "_png": 61746, + "_po": 61814, + "_pod": 85337, + "_point": 6085, + "_pointer": 21425, + "_points": 12928, + "_pol": 47323, + "_policy": 22773, + "_poll": 40002, + "_poly": 36133, + "_polygon": 73542, + "_pool": 15709, + "_pop": 17061, + "_population": 74572, + "_popup": 44984, + "_por": 76628, + "_port": 8716, + "_portal": 97077, + "_portfolio": 80790, + "_ports": 47694, + "_pos": 6479, + "_pose": 33201, + "_position": 9661, + "_positions": 37219, + "_positive": 54160, + "_possible": 71244, + "_post": 6333, + "_posts": 20222, + "_pot": 57952, + "_pow": 56183, + "_power": 20421, + "_pp": 31600, + "_pr": 5294, + "_pre": 10442, + "_prec": 59179, + "_precision": 54618, + "_pred": 12830, + "_predicate": 97474, + "_predict": 26815, + "_predicted": 95488, + "_prediction": 51446, + "_predictions": 59887, + "_preds": 65983, + "_pref": 43331, + "_preference": 92718, + "_preferences": 65513, + "_prefix": 13974, + "_prefs": 86430, + "_prep": 83794, + "_prepare": 47460, + "_pres": 32116, + "_presence": 56403, + "_present": 36976, + "_press": 38938, + "_pressed": 46178, + "_pressure": 73595, + "_prev": 25566, + "_preview": 42383, + "_previous": 52552, + "_pri": 85639, + "_price": 9040, + "_prices": 65688, + "_prim": 87419, + "_primary": 45314, + "_prime": 38217, + "_primitive": 84087, + "_principal": 82402, + "_print": 10064, + "_printer": 91772, + "_printf": 21823, + "_prior": 58782, + "_priority": 38161, + "_priv": 24726, + "_private": 26249, + "_pro": 2540, + "_prob": 20794, + "_proba": 83022, + "_probability": 68589, + "_probe": 49108, + "_problem": 60139, + "_probs": 51405, + "_proc": 24436, + "_process": 11305, + "_processed": 63916, + "_processes": 80143, + "_processing": 58209, + "_processor": 50127, + "_processors": 73153, + "_procs": 96148, + "_prod": 34984, + "_product": 9840, + "_production": 91980, + "_producto": 58428, + "_products": 29622, + "_produk": 80752, + "_prof": 33609, + "_profile": 13789, + "_profiles": 64021, + "_profit": 72042, + "_prog": 60390, + "_program": 25096, + "_progress": 27200, + "_proj": 49453, + "_project": 16352, + "_projection": 72738, + "_projects": 58774, + "_prom": 47877, + "_prompt": 61421, + "_proof": 86757, + "_prop": 21663, + "_properties": 25158, + "_property": 16638, + "_props": 44835, + "_prot": 22357, + "_proto": 37689, + "_protocol": 34880, + "_prov": 89648, + "_provider": 29518, + "_province": 97116, + "_proxy": 29712, + "_ps": 26047, + "_pt": 20124, + "_ptr": 4348, + "_ptrs": 79533, + "_pts": 41878, + "_pub": 34014, + "_public": 27074, + "_publish": 58779, + "_published": 69524, + "_publisher": 84726, + "_pull": 65693, + "_pulse": 83645, + "_purchase": 63203, + "_push": 14218, + "_pushButton": 91468, + "_put": 15557, + "_putchar": 57969, + "_puts": 88574, + "_putstr": 74475, + "_pv": 91139, + "_pw": 60181, + "_pwd": 50957, + "_pwm": 70985, + "_px": 67162, + "_py": 40291, + "_python": 55869, + "_q": 8976, + "_qos": 78960, + "_qp": 88550, + "_qs": 80738, + "_qty": 34858, + "_qu": 11280, + "_quad": 68631, + "_qual": 64391, + "_quality": 55074, + "_quant": 53346, + "_quantity": 32097, + "_que": 47342, + "_queries": 65588, + "_query": 5738, + "_queryset": 66638, + "_quest": 98015, + "_question": 28915, + "_questions": 55048, + "_queue": 10841, + "_queues": 82413, + "_quick": 82094, + "_quit": 72424, + "_quiz": 88336, + "_quota": 97093, + "_quote": 45236, + "_quotes": 75637, + "_r": 1710, + "_ra": 59969, + "_race": 91317, + "_rad": 39764, + "_radi": 97479, + "_radio": 49740, + "_radius": 28936, + "_raise": 55383, + "_raises": 88840, + "_ram": 62124, + "_rand": 33864, + "_random": 22644, + "_range": 9698, + "_ranges": 58748, + "_rank": 20417, + "_rat": 70475, + "_rate": 9246, + "_rates": 55373, + "_rating": 30646, + "_ratings": 75188, + "_ratio": 19917, + "_raw": 16067, + "_ray": 72690, + "_rb": 68544, + "_rc": 30299, + "_rd": 54763, + "_re": 1288, + "_reaction": 96631, + "_read": 6443, + "_readable": 91232, + "_reader": 22306, + "_reading": 81859, + "_reads": 66628, + "_ready": 35456, + "_real": 15266, + "_reason": 38229, + "_rec": 7080, + "_recall": 90311, + "_receipt": 77974, + "_receive": 38557, + "_received": 40783, + "_receiver": 65691, + "_recent": 62361, + "_recipe": 62251, + "_recommend": 99084, + "_record": 14192, + "_records": 31279, + "_recovery": 91475, + "_rect": 16979, + "_rectangle": 61847, + "_recursive": 66242, + "_recv": 36118, + "_red": 26058, + "_redirect": 30043, + "_redirected": 73309, + "_redis": 87542, + "_reduce": 64596, + "_reduction": 74289, + "_ref": 7793, + "_refer": 73386, + "_reference": 25433, + "_references": 92702, + "_refl": 83623, + "_refptr": 97643, + "_refresh": 42076, + "_refs": 60638, + "_reg": 4920, + "_regeneration": 74820, + "_regex": 41832, + "_region": 20627, + "_regions": 58035, + "_register": 14000, + "_registered": 72683, + "_registers": 78360, + "_registration": 49101, + "_registro": 91834, + "_registry": 50650, + "_regression": 91144, + "_regs": 36910, + "_regular": 49443, + "_regularizer": 79059, + "_rel": 13557, + "_related": 53256, + "_relation": 46984, + "_relations": 94221, + "_relationship": 69031, + "_relative": 29286, + "_release": 24577, + "_reload": 79405, + "_relu": 76282, + "_rem": 19194, + "_remain": 86607, + "_remaining": 59244, + "_remote": 36425, + "_remove": 18193, + "_removed": 68248, + "_rename": 79399, + "_render": 22781, + "_renderer": 52345, + "_rent": 83127, + "_reordered": 66996, + "_rep": 25533, + "_repeat": 41975, + "_replace": 10633, + "_reply": 15323, + "_repo": 37784, + "_report": 14813, + "_reporting": 52376, + "_reports": 64423, + "_repository": 47301, + "_repr": 68535, + "_representation": 94263, + "_req": 17644, + "_request": 7893, + "_requested": 72709, + "_requests": 37216, + "_require": 13793, + "_required": 18743, + "_requirements": 89632, + "_requires": 54949, + "_res": 4918, + "_reservation": 96548, + "_reserve": 89591, + "_reserved": 74913, + "_reset": 18983, + "_residual": 88056, + "_resize": 53370, + "_resolution": 39849, + "_resolve": 77291, + "_resolver": 91578, + "_resource": 17962, + "_resources": 35569, + "_resp": 35160, + "_response": 9655, + "_responses": 81292, + "_rest": 32231, + "_restart": 69392, + "_restore": 62677, + "_restrict": 97684, + "_result": 5287, + "_results": 13576, + "_resume": 58132, + "_ret": 21695, + "_retry": 62745, + "_return": 12511, + "_returns": 58900, + "_rev": 38082, + "_reverse": 43277, + "_review": 38661, + "_reviews": 68706, + "_revision": 59856, + "_reward": 38260, + "_rewards": 77997, + "_rewrite": 74052, + "_rf": 65466, + "_rg": 86566, + "_rgb": 37407, + "_rgba": 95229, + "_rho": 79160, + "_rhs": 64537, + "_right": 10539, + "_rights": 98053, + "_ring": 34683, + "_rl": 98441, + "_rm": 58537, + "_rng": 66849, + "_rnn": 98600, + "_ro": 26608, + "_robot": 59116, + "_roi": 72391, + "_role": 19792, + "_roles": 43061, + "_roll": 57578, + "_rom": 81383, + "_room": 25003, + "_rooms": 76037, + "_root": 12993, + "_ros": 79215, + "_rot": 38256, + "_rotate": 60834, + "_rotation": 44813, + "_round": 29896, + "_route": 28109, + "_router": 55587, + "_routes": 64314, + "_routing": 97710, + "_row": 8530, + "_rows": 10949, + "_rp": 83440, + "_rpc": 60799, + "_rq": 79638, + "_rr": 56285, + "_rs": 47115, + "_rsa": 82522, + "_rsp": 64381, + "_rt": 40169, + "_ru": 73584, + "_rule": 21124, + "_rules": 21407, + "_run": 14007, + "_runner": 54828, + "_running": 37333, + "_runs": 67352, + "_runtime": 33232, + "_rw": 49566, + "_rwlock": 98850, + "_rx": 24330, + "_s": 643, + "_sa": 66443, + "_safe": 34067, + "_saida": 76420, + "_sal": 62491, + "_salary": 71222, + "_sale": 47114, + "_sales": 47067, + "_salt": 92280, + "_same": 33574, + "_sample": 17491, + "_sampler": 72812, + "_samples": 18297, + "_sampling": 77716, + "_san": 87866, + "_sat": 51303, + "_save": 15376, + "_saved": 53709, + "_sb": 55540, + "_sc": 13171, + "_scal": 28656, + "_scalar": 41652, + "_scale": 16727, + "_scaled": 61715, + "_scaling": 79216, + "_scan": 28857, + "_scenario": 89712, + "_scene": 37296, + "_sched": 67394, + "_schedule": 34530, + "_scheduler": 53646, + "_schema": 25371, + "_scheme": 53293, + "_school": 62165, + "_scope": 23199, + "_score": 10405, + "_scores": 27198, + "_scr": 59728, + "_screen": 17649, + "_script": 14660, + "_scripts": 35789, + "_scroll": 41407, + "_sd": 40168, + "_sdk": 61783, + "_se": 3453, + "_search": 10716, + "_season": 78725, + "_seat": 91381, + "_sec": 17242, + "_second": 29644, + "_secondary": 77759, + "_seconds": 34825, + "_secret": 21962, + "_secs": 68718, + "_section": 16221, + "_sections": 59485, + "_sector": 62651, + "_secure": 73088, + "_security": 48726, + "_seed": 33809, + "_seek": 74473, + "_seen": 60668, + "_seg": 35325, + "_segment": 28061, + "_segments": 55735, + "_sel": 34153, + "_select": 13051, + "_selected": 23755, + "_selection": 23672, + "_selector": 28890, + "_self": 25637, + "_sell": 71072, + "_sem": 30442, + "_semaphore": 90255, + "_send": 13565, + "_sender": 54356, + "_sensitive": 65324, + "_sensor": 28899, + "_sent": 24115, + "_sentence": 48332, + "_sentences": 75903, + "_sep": 54775, + "_separator": 58204, + "_seq": 14486, + "_seqs": 93261, + "_sequence": 23735, + "_sequences": 58732, + "_ser": 75861, + "_serial": 25602, + "_serialize": 88686, + "_serializer": 67441, + "_series": 35015, + "_serv": 44053, + "_server": 12015, + "_servers": 67507, + "_service": 12267, + "_services": 39846, + "_sess": 83856, + "_session": 12316, + "_sessions": 71385, + "_set": 2602, + "_setopt": 23533, + "_sets": 21289, + "_setting": 20313, + "_settings": 10853, + "_setup": 21363, + "_sex": 82113, + "_sf": 56182, + "_sg": 76318, + "_sh": 3712, + "_sha": 48836, + "_shader": 35485, + "_shadow": 53120, + "_shape": 13597, + "_shapes": 71315, + "_share": 38272, + "_shared": 20405, + "_sheet": 43112, + "_shell": 48945, + "_shift": 22230, + "_ship": 31042, + "_shipping": 53553, + "_shop": 38690, + "_short": 16673, + "_shortcode": 48259, + "_shot": 81226, + "_should": 43378, + "_show": 15267, + "_shuffle": 73484, + "_shutdown": 54804, + "_si": 40983, + "_sibling": 96328, + "_sid": 53044, + "_side": 30862, + "_sidebar": 31628, + "_sig": 29252, + "_sigma": 49005, + "_sign": 11172, + "_signal": 21137, + "_signals": 76698, + "_signature": 39859, + "_signed": 55617, + "_signup": 82005, + "_sim": 18314, + "_similarity": 76236, + "_simple": 30015, + "_simps": 79572, + "_simulation": 82487, + "_sin": 78710, + "_since": 56262, + "_single": 19487, + "_singleton": 82119, + "_singular": 93240, + "_sink": 51567, + "_site": 24507, + "_sites": 68600, + "_size": 2368, + "_sizes": 32159, + "_sk": 33811, + "_skb": 94486, + "_skill": 45624, + "_skills": 92980, + "_skin": 66052, + "_skip": 44830, + "_sku": 79647, + "_sl": 11886, + "_slave": 65320, + "_sleep": 49393, + "_slice": 26488, + "_slices": 87288, + "_slide": 61833, + "_slider": 36007, + "_slope": 79743, + "_slot": 27563, + "_slots": 54161, + "_slow": 82447, + "_slug": 31024, + "_sm": 15874, + "_small": 31966, + "_smart": 88825, + "_smooth": 71655, + "_sms": 58567, + "_sn": 28022, + "_snap": 74175, + "_snapshot": 53265, + "_snd": 88518, + "_so": 39833, + "_soc": 54114, + "_social": 51196, + "_sock": 33103, + "_socket": 19555, + "_soft": 38326, + "_softc": 75302, + "_softmax": 99108, + "_sol": 53668, + "_sold": 88193, + "_solution": 50274, + "_solve": 86246, + "_solver": 53156, + "_some": 61855, + "_song": 57102, + "_sort": 18435, + "_sorted": 41277, + "_sound": 37676, + "_soup": 97352, + "_source": 10347, + "_sources": 52896, + "_sp": 10123, + "_space": 14663, + "_spaces": 67883, + "_spacing": 63328, + "_span": 37382, + "_sparse": 71123, + "_spawn": 76026, + "_spec": 13594, + "_special": 41629, + "_species": 74130, + "_specific": 56592, + "_specs": 71200, + "_spectrum": 95463, + "_speed": 16944, + "_spell": 64717, + "_sphere": 86973, + "_spi": 44920, + "_spin": 47965, + "_spinner": 65114, + "_split": 17052, + "_splits": 73000, + "_spot": 84766, + "_sprite": 45489, + "_sprites": 89629, + "_sq": 47252, + "_sql": 18063, + "_sqrt": 92199, + "_square": 39794, + "_squared": 54641, + "_sr": 61248, + "_src": 16274, + "_srv": 63921, + "_ss": 41848, + "_ssh": 82805, + "_ssl": 48210, + "_st": 1261, + "_sta": 79951, + "_stack": 15528, + "_staff": 47060, + "_stage": 35238, + "_stamp": 52978, + "_stand": 87957, + "_standard": 48688, + "_star": 31681, + "_start": 4906, + "_started": 55679, + "_starts": 86552, + "_startup": 80858, + "_stat": 16271, + "_state": 4387, + "_statement": 37404, + "_states": 22972, + "_static": 25360, + "_station": 44798, + "_statistics": 49569, + "_stats": 15381, + "_status": 4773, + "_statuses": 83702, + "_std": 15656, + "_stderr": 93033, + "_stdio": 37227, + "_stdout": 67416, + "_ste": 65937, + "_step": 11946, + "_steps": 22731, + "_stmt": 21824, + "_stock": 30541, + "_stop": 19039, + "_storage": 23310, + "_store": 14809, + "_story": 66342, + "_str": 2895, + "_strategy": 49308, + "_strcmp": 74523, + "_strdup": 45901, + "_stream": 12673, + "_streams": 74872, + "_street": 78562, + "_strength": 71893, + "_strerror": 81444, + "_stride": 41513, + "_strike": 71717, + "_string": 3904, + "_strings": 33500, + "_strip": 66130, + "_strlen": 44288, + "_struct": 15126, + "_structure": 38283, + "_stub": 62781, + "_student": 36753, + "_students": 71644, + "_study": 85581, + "_stuff": 95396, + "_style": 15117, + "_styles": 37857, + "_stylesheet": 84724, + "_su": 89927, + "_sub": 5228, + "_subject": 28834, + "_submenu": 96733, + "_submission": 75866, + "_submit": 31674, + "_subnet": 95681, + "_subplot": 56978, + "_subs": 82767, + "_subscribe": 88935, + "_subscription": 59838, + "_subset": 47532, + "_substr": 83578, + "_subtitle": 94927, + "_subtype": 96228, + "_succ": 70758, + "_success": 18632, + "_successful": 92951, + "_suffix": 37151, + "_suite": 57239, + "_sum": 10160, + "_summary": 27251, + "_sun": 86623, + "_sup": 23723, + "_super": 38886, + "_superuser": 85754, + "_supp": 91480, + "_supplier": 75438, + "_supply": 65630, + "_support": 25827, + "_supported": 57885, + "_sur": 49946, + "_surf": 88323, + "_surface": 31030, + "_survey": 88345, + "_suspend": 64196, + "_sv": 76808, + "_svc": 78198, + "_svg": 76849, + "_sw": 32581, + "_swap": 40991, + "_switch": 27652, + "_sy": 58530, + "_sym": 26825, + "_symbol": 21179, + "_symbols": 55752, + "_syn": 51393, + "_sync": 23008, + "_syntax": 78894, + "_sys": 20344, + "_system": 17687, + "_sz": 33945, + "_t": 528, + "_tA": 81612, + "_tC": 84772, + "_tD": 88741, + "_tE": 98304, + "_tF": 79798, + "_ta": 91710, + "_tab": 17344, + "_table": 5237, + "_tables": 35632, + "_tabs": 57953, + "_tac": 30626, + "_tag": 9372, + "_tags": 16333, + "_tail": 35471, + "_take": 73261, + "_taken": 72933, + "_tar": 66608, + "_target": 11123, + "_targets": 41997, + "_task": 12184, + "_tasks": 32823, + "_tau": 75760, + "_tax": 24579, + "_taxonomy": 75794, + "_tb": 23481, + "_tbl": 34572, + "_tc": 63126, + "_tcb": 98725, + "_tcp": 45562, + "_td": 57051, + "_te": 38678, + "_teacher": 70553, + "_team": 26532, + "_teams": 86299, + "_tel": 67180, + "_tele": 58253, + "_tem": 28613, + "_temp": 11771, + "_temperature": 53525, + "_template": 8693, + "_templates": 49526, + "_ten": 75827, + "_tensor": 23188, + "_tensors": 91067, + "_ter": 72493, + "_term": 17464, + "_terminal": 64309, + "_terms": 37498, + "_test": 4452, + "_testing": 70962, + "_tests": 32509, + "_tex": 48187, + "_text": 4326, + "_texts": 79646, + "_texture": 28165, + "_tf": 47719, + "_tgt": 93085, + "_th": 5854, + "_than": 51613, + "_that": 70197, + "_the": 16068, + "_theme": 19864, + "_then": 68367, + "_theta": 39363, + "_thickness": 91231, + "_third": 79519, + "_this": 24868, + "_thr": 72071, + "_thread": 10814, + "_threads": 29725, + "_three": 50016, + "_thresh": 49562, + "_threshold": 21858, + "_through": 87399, + "_throw": 72161, + "_thumb": 46352, + "_thumbnail": 37824, + "_ti": 98799, + "_tick": 43612, + "_ticket": 45718, + "_tickets": 91836, + "_ticks": 49961, + "_tid": 73719, + "_tile": 29844, + "_tiles": 60857, + "_tim": 29087, + "_time": 3009, + "_timeline": 77560, + "_timeout": 20537, + "_timer": 16255, + "_times": 22353, + "_timestamp": 23073, + "_timezone": 53193, + "_timing": 70973, + "_tip": 48014, + "_tipo": 43659, + "_title": 6112, + "_titles": 72199, + "_tl": 82970, + "_tls": 71262, + "_tm": 44045, + "_tmp": 16125, + "_to": 2346, + "_today": 68717, + "_todo": 87686, + "_toggle": 48860, + "_tok": 76162, + "_token": 6458, + "_tokenize": 86508, + "_tokens": 28838, + "_tol": 70121, + "_tolerance": 92069, + "_tool": 22785, + "_toolbar": 65695, + "_tools": 39723, + "_tooltip": 76265, + "_top": 10426, + "_topic": 31414, + "_topics": 64358, + "_topology": 90234, + "_tot": 36138, + "_total": 10784, + "_totals": 96892, + "_touch": 60840, + "_tp": 55191, + "_tpl": 14107, + "_tr": 3547, + "_tra": 27242, + "_trace": 23575, + "_track": 28545, + "_tracker": 50264, + "_tracking": 66105, + "_tracks": 66953, + "_trade": 52162, + "_traffic": 90012, + "_train": 7591, + "_training": 32891, + "_trait": 78491, + "_traits": 39693, + "_traj": 83823, + "_trajectory": 86031, + "_trampoline": 92942, + "_tran": 98941, + "_trans": 7965, + "_transaction": 28884, + "_transactions": 68182, + "_transaksi": 95661, + "_transfer": 35403, + "_transform": 18449, + "_transient": 97758, + "_transition": 53593, + "_translate": 66381, + "_translation": 49273, + "_transport": 46398, + "_trap": 88536, + "_travel": 91319, + "_tree": 11663, + "_trees": 75791, + "_tri": 42120, + "_trial": 63486, + "_trials": 75845, + "_triangle": 70575, + "_trigger": 32925, + "_triggered": 78678, + "_trim": 70025, + "_trip": 63883, + "_true": 16082, + "_truth": 49186, + "_try": 53283, + "_ts": 25023, + "_tt": 66740, + "_ttl": 87157, + "_tunnel": 95692, + "_tuple": 21773, + "_tuples": 89269, + "_turn": 37274, + "_tv": 46132, + "_tw": 54178, + "_tweet": 75419, + "_tweets": 75920, + "_twitter": 83338, + "_two": 23241, + "_tx": 17805, + "_txn": 92299, + "_txt": 22436, + "_ty": 53171, + "_typ": 42111, + "_type": 1819, + "_typeDefinition": 66605, + "_typeDefinitionSize": 66650, + "_typeof": 93852, + "_types": 9763, + "_u": 7300, + "_uart": 56789, + "_ub": 66794, + "_uc": 68887, + "_ud": 71939, + "_udp": 69432, + "_ue": 91064, + "_ui": 22412, + "_uid": 25396, + "_uint": 15807, + "_ul": 61039, + "_ulong": 72288, + "_um": 72063, + "_un": 4907, + "_unc": 76576, + "_under": 58228, + "_undo": 98940, + "_unicode": 54662, + "_uniform": 45066, + "_union": 51621, + "_unique": 21218, + "_unit": 14832, + "_units": 28345, + "_unix": 80572, + "_unknown": 57507, + "_unlock": 19465, + "_unpack": 54889, + "_unref": 49484, + "_unregister": 68992, + "_unset": 98109, + "_unsigned": 67830, + "_until": 44352, + "_unused": 66944, + "_up": 8237, + "_upd": 83550, + "_update": 8882, + "_updated": 33492, + "_updates": 57829, + "_upgrade": 67794, + "_upload": 21691, + "_uploaded": 54411, + "_upper": 34445, + "_ur": 64879, + "_uri": 15572, + "_url": 2903, + "_urls": 32822, + "_us": 11306, + "_usage": 31507, + "_usb": 50863, + "_use": 15951, + "_usec": 43047, + "_used": 27803, + "_user": 3317, + "_userdata": 59315, + "_userid": 84568, + "_username": 21588, + "_users": 16348, + "_using": 75180, + "_usr": 79078, + "_usuario": 29246, + "_ut": 60363, + "_utc": 84259, + "_utf": 39453, + "_util": 18974, + "_utilities": 94044, + "_utils": 17309, + "_uuid": 25540, + "_uv": 67474, + "_v": 2273, + "_va": 73959, + "_val": 6189, + "_valid": 8337, + "_validate": 42681, + "_validation": 19416, + "_validator": 64959, + "_valor": 93131, + "_vals": 29550, + "_value": 3142, + "_values": 9146, + "_var": 4612, + "_variable": 14635, + "_variables": 28182, + "_variance": 77450, + "_variant": 46112, + "_variation": 95409, + "_vars": 11168, + "_vc": 73396, + "_ve": 73285, + "_vec": 13251, + "_vect": 59402, + "_vector": 12247, + "_vectors": 49158, + "_vehicle": 64008, + "_vel": 34504, + "_velocity": 39666, + "_vendor": 58320, + "_venta": 87621, + "_ver": 26042, + "_verbose": 67921, + "_verification": 84245, + "_verified": 71871, + "_verify": 35638, + "_version": 9438, + "_versions": 65148, + "_vert": 54000, + "_vertex": 26611, + "_vertical": 61203, + "_vertices": 37720, + "_verts": 98035, + "_vi": 97319, + "_via": 80710, + "_vid": 68385, + "_video": 19815, + "_videos": 83791, + "_view": 7122, + "_viewer": 80344, + "_views": 39498, + "_virtual": 58730, + "_vis": 15887, + "_visibility": 71256, + "_visible": 34371, + "_visit": 56484, + "_visited": 96729, + "_visitor": 81364, + "_visual": 64630, + "_vk": 89852, + "_vlan": 71296, + "_vlog": 92145, + "_vm": 39008, + "_vocab": 52823, + "_voice": 76765, + "_void": 37775, + "_vol": 42784, + "_voltage": 62756, + "_volume": 26941, + "_vote": 54360, + "_votes": 65116, + "_vp": 87960, + "_vs": 42434, + "_vue": 69147, + "_w": 1670, + "_wait": 18760, + "_waiting": 84683, + "_walk": 56131, + "_wall": 50731, + "_wallet": 62308, + "_war": 90539, + "_warn": 55045, + "_warning": 38395, + "_warnings": 87658, + "_was": 91612, + "_watch": 58562, + "_water": 54550, + "_wave": 60063, + "_way": 48936, + "_wc": 75871, + "_we": 45922, + "_weak": 86356, + "_weapon": 69475, + "_weather": 69364, + "_web": 25960, + "_website": 79032, + "_week": 31277, + "_weight": 15876, + "_weights": 21114, + "_wf": 86382, + "_wh": 36225, + "_wheel": 82012, + "_when": 47636, + "_where": 36814, + "_while": 86069, + "_white": 44431, + "_whitespace": 86175, + "_widget": 13585, + "_widgets": 51815, + "_width": 7927, + "_wifi": 71495, + "_win": 25672, + "_wind": 87438, + "_window": 12571, + "_windows": 58220, + "_winner": 92588, + "_wire": 75206, + "_with": 6615, + "_within": 72385, + "_without": 39904, + "_wo": 96058, + "_word": 13533, + "_words": 18981, + "_work": 11498, + "_worker": 40385, + "_workers": 43557, + "_workflow": 86418, + "_working": 81101, + "_workspace": 75560, + "_world": 31792, + "_wp": 39287, + "_wr": 44074, + "_wrap": 38550, + "_wrapper": 23561, + "_write": 9165, + "_writer": 28908, + "_written": 72403, + "_wrong": 75198, + "_ws": 50211, + "_x": 3212, + "_xlabel": 52698, + "_xlim": 74561, + "_xml": 23855, + "_xor": 76462, + "_xpath": 36595, + "_xs": 80646, + "_xt": 60069, + "_xx": 95968, + "_xy": 40920, + "_xyz": 64974, + "_y": 4178, + "_yaml": 64380, + "_yaw": 80183, + "_year": 14645, + "_years": 74490, + "_yellow": 93053, + "_yes": 59744, + "_yield": 83709, + "_ylabel": 48189, + "_ylim": 67646, + "_you": 70456, + "_z": 6415, + "_zero": 19359, + "_zeros": 83761, + "_zip": 42131, + "_zone": 28692, + "_zones": 93002, + "_zoom": 56736, + "_{": 15159, + "_|": 35395, + "_Ċ": 12924, + "_ĊĊ": 18797, + "_čĊ": 57891, + "_čĊčĊ": 89950, + "`": 63, + "`\"]Ċ": 91595, + "`${": 37753, + "`(": 81556, + "`()": 67680, + "`)": 32881, + "`),Ċ": 89746, + "`).": 59210, + "`);Ċ": 20625, + "`);ĊĊ": 74526, + "`)Ċ": 24183, + "`,": 7808, + "`,`": 40181, + "`,Ċ": 12892, + "`.": 28587, + "`.`": 28905, + "`.Ċ": 18639, + "`.ĊĊ": 62338, + "`:": 44622, + "`;": 77582, + "`;Ċ": 16909, + "`;ĊĊ": 18259, + "`=": 90260, + "`='$": 76116, + "`\\": 61069, + "`]": 80709, + "`](": 84702, + "``": 13874, + "``,": 90716, + "```": 73594, + "``Ċ": 84274, + "`s": 39829, + "`t": 79115, + "`}": 74768, + "`}Ċ": 53136, + "`Ċ": 3989, + "`ĊĊ": 19324, + "`čĊ": 75679, + "a": 64, + "aData": 94950, + "aN": 55019, + "aa": 5305, + "aaS": 64610, + "aaa": 32646, + "aaaa": 28458, + "aaaaaaaa": 69440, + "aab": 88607, + "aac": 64705, + "aad": 86629, + "aal": 54342, + "aan": 41702, + "aar": 73609, + "aat": 38267, + "ab": 370, + "aba": 12004, + "abad": 27347, + "abaj": 78984, + "abajo": 73268, + "abal": 62794, + "abama": 18884, + "aban": 38165, + "abant": 79283, + "abar": 94560, + "abay": 97186, + "abb": 12523, + "abbage": 56489, + "abbit": 19949, + "abbix": 95442, + "abble": 72191, + "abbo": 53550, + "abbr": 71276, + "abbrev": 44272, + "abbreviation": 76857, + "abby": 78636, + "abc": 13683, + "abcd": 68644, + "abcdef": 41202, + "abcdefgh": 56747, + "abcdefghijkl": 63168, + "abcdefghijklmnop": 65102, + "abcdefghijklmnopqrstuvwxyz": 67512, + "abd": 85920, + "abe": 8229, + "abee": 79112, + "abei": 38972, + "abel": 780, + "abela": 37385, + "abeled": 22320, + "abella": 71590, + "abelle": 54167, + "aben": 40308, + "aber": 43139, + "abet": 10237, + "abetes": 16231, + "abeth": 17406, + "abetic": 44838, + "abez": 83478, + "abh": 66366, + "abhäng": 138644, + "abhängig": 138645, + "abi": 25084, + "abic": 68291, + "abies": 97041, + "abil": 12762, + "abile": 57109, + "abileceÄŁi": 132216, + "abilia": 93350, + "abilidad": 54262, + "abilidade": 79916, + "abilir": 81758, + "abilit": 45400, + "abilities": 8456, + "ability": 2897, + "abilité": 86682, + "abin": 8892, + "abinet": 75512, + "abis": 15367, + "abit": 38927, + "abl": 61473, + "able": 480, + "ableObject": 79176, + "ableOpacity": 22529, + "ableView": 7938, + "ableViewController": 94906, + "abled": 2312, + "abler": 91967, + "ables": 4788, + "abling": 18768, + "ablish": 5102, + "ablo": 31460, + "ably": 2845, + "ablytyped": 80898, + "abo": 47421, + "abol": 52804, + "abolic": 56465, + "abor": 4324, + "abort": 46699, + "about": 9096, + "abouts": 63299, + "above": 48432, + "abox": 76734, + "abr": 77571, + "abra": 43932, + "abras": 53888, + "abric": 28897, + "abs": 3435, + "absolute": 17182, + "abstract": 16249, + "abstractmethod": 64951, + "abus": 57541, + "abwe": 42431, + "aby": 6115, + "abyrin": 57337, + "abyrinth": 61530, + "abyte": 66711, + "abytes": 71129, + "abı": 126148, + "ac": 580, + "aca": 17106, + "acad": 49857, + "academic": 90256, + "acades": 18686, + "acak": 78665, + "acakt": 127384, + "acaktır": 127385, + "acam": 95608, + "acao": 15108, + "acas": 77183, + "acaģı": 125207, + "acaģını": 131225, + "acaģız": 136232, + "acb": 97471, + "acc": 4475, + "accel": 43888, + "accent": 77548, + "accept": 10330, + "acceptable": 37016, + "accepted": 54574, + "access": 5211, + "accessToken": 41167, + "accessible": 62347, + "acci": 34603, + "accine": 88021, + "accion": 45422, + "acciones": 97707, + "acción": 66047, + "acco": 19500, + "accom": 71987, + "according": 73846, + "accordion": 35337, + "account": 4608, + "accountId": 75906, + "accounts": 26206, + "acct": 95081, + "accum": 38931, + "accumulate": 98387, + "accumulator": 94810, + "accur": 35921, + "accuracy": 32729, + "accès": 135341, + "acd": 94112, + "ace": 578, + "acea": 82582, + "acebook": 4497, + "aced": 4435, + "acellular": 64341, + "acemark": 97594, + "acement": 6346, + "acements": 33078, + "acen": 90487, + "acency": 40624, + "acent": 18181, + "acente": 94860, + "aceous": 76040, + "acer": 9584, + "acerb": 50819, + "acers": 72697, + "aces": 2434, + "acet": 67223, + "aceut": 24309, + "aceutical": 25134, + "acey": 55920, + "acf": 62594, + "ach": 610, + "acha": 51576, + "achable": 77283, + "achat": 96164, + "ache": 1777, + "ached": 3854, + "achel": 14770, + "achelor": 24736, + "achen": 45699, + "acher": 11007, + "achers": 26520, + "aches": 14242, + "achi": 30364, + "achie": 44520, + "achievement": 74608, + "achine": 3814, + "achinery": 62954, + "achines": 70142, + "aching": 11829, + "achment": 8888, + "acho": 63429, + "achs": 86873, + "achsen": 70487, + "acht": 15938, + "achte": 84642, + "achten": 53776, + "achts": 81006, + "achu": 64918, + "achuset": 20041, + "achusetts": 20051, + "aci": 24464, + "acia": 27041, + "acial": 32111, + "acias": 42442, + "acic": 93157, + "acid": 57056, + "acidad": 69545, + "acie": 46283, + "aciente": 58203, + "acier": 53453, + "acies": 26030, + "acific": 40483, + "acija": 83903, + "acimiento": 54366, + "acin": 39388, + "acing": 4527, + "acio": 35870, + "acion": 5806, + "acional": 21598, + "acionales": 84139, + "aciones": 12430, + "acios": 79845, + "acious": 19430, + "acist": 49322, + "acists": 92612, + "acity": 4018, + "ación": 5721, + "acja": 51689, + "acje": 69209, + "acji": 33378, + "acjÄĻ": 135541, + "ack": 473, + "ackBar": 32629, + "ackage": 1434, + "ackages": 22211, + "ackbar": 27691, + "acked": 11191, + "acker": 9683, + "ackers": 87412, + "acket": 5709, + "ackets": 18382, + "acking": 8985, + "ackle": 51522, + "acks": 7821, + "ackson": 19387, + "acky": 79957, + "acl": 47736, + "aclass": 67528, + "acle": 7902, + "acles": 18209, + "acman": 94114, + "aco": 16428, + "acob": 38951, + "acobian": 62518, + "acock": 80278, + "acoes": 60126, + "acom": 44570, + "acomment": 75215, + "acon": 22379, + "acons": 75444, + "acos": 41647, + "acr": 64748, + "acre": 28474, + "acro": 49507, + "acs": 19252, + "act": 531, + "actable": 66660, + "actal": 80148, + "acted": 22167, + "acter": 2481, + "acteria": 77752, + "acterial": 69934, + "acters": 4983, + "actic": 23170, + "actical": 36019, + "actice": 33001, + "actices": 89190, + "actics": 67969, + "acting": 34922, + "action": 1311, + "actionDate": 89107, + "actions": 4020, + "activ": 8859, + "activate": 16856, + "activated": 30162, + "activation": 23002, + "active": 3028, + "actively": 63019, + "activex": 36231, + "activities": 31978, + "activity": 7175, + "activité": 133850, + "activo": 83070, + "actly": 32739, + "actor": 5621, + "actories": 52893, + "actoring": 75407, + "actors": 21161, + "actory": 2919, + "acts": 11359, + "actual": 11944, + "actually": 73028, + "actus": 60959, + "acula": 85193, + "acular": 23006, + "acus": 95305, + "acute": 31490, + "acy": 2757, + "acyj": 95553, + "acz": 41830, + "acı": 125403, + "ad": 329, + "ada": 2584, + "adal": 50488, + "adalafil": 93086, + "adam": 50954, + "adamente": 76055, + "adan": 53640, + "adapt": 88071, + "adapter": 19731, + "adaptive": 41562, + "adaptiveStyles": 49225, + "adar": 36080, + "adas": 11107, + "adastrar": 66170, + "adastro": 56566, + "adata": 4602, + "aday": 64626, + "adays": 33976, + "adaÅŁ": 126531, + "adb": 39591, + "adc": 33385, + "add": 718, + "addAction": 57772, + "addAll": 67809, + "addChild": 44852, + "addClass": 12287, + "addColumn": 47792, + "addComponent": 11270, + "addContainerGap": 33710, + "addElement": 76293, + "addError": 88122, + "addEventListener": 32374, + "addField": 51855, + "addGap": 23714, + "addGroup": 15488, + "addItem": 49738, + "addListener": 84587, + "addOn": 58067, + "addPreferredGap": 34275, + "addTo": 50301, + "addWidget": 36805, + "addafi": 93312, + "added": 16828, + "adden": 70095, + "adder": 20808, + "adders": 80208, + "addin": 97006, + "adding": 2998, + "addir": 44525, + "additional": 35499, + "addle": 20137, + "addock": 79434, + "addon": 51099, + "addons": 77270, + "addr": 6214, + "address": 4995, + "addresses": 53789, + "adds": 72994, + "addtogroup": 84286, + "addy": 22478, + "ade": 1021, + "adecimal": 49326, + "aded": 14255, + "adel": 15151, + "adelphia": 17802, + "adem": 59372, + "ademic": 69758, + "aden": 21140, + "adena": 44347, + "adeon": 32709, + "ader": 998, + "adera": 66601, + "adero": 73230, + "aders": 6393, + "ades": 3452, + "adesh": 22902, + "adf": 96289, + "adge": 60705, + "adget": 50931, + "adh": 51587, + "adi": 2767, + "adia": 52831, + "adian": 10066, + "adians": 20736, + "adiator": 67808, + "adic": 36214, + "adiens": 82695, + "adients": 36318, + "adier": 95265, + "adies": 17054, + "adin": 30984, + "ading": 2228, + "adio": 4022, + "adioButton": 17253, + "adir": 41173, + "adium": 13478, + "adius": 4037, + "adj": 24739, + "adjust": 38719, + "adjusted": 87797, + "adle": 71355, + "adm": 46929, + "admin": 2882, + "administr": 68849, + "administration": 90590, + "administrator": 66966, + "admins": 74600, + "ado": 2123, + "adol": 86769, + "adoo": 77112, + "adoop": 25268, + "adopt": 98238, + "ador": 5364, + "adora": 29813, + "adoras": 73438, + "adores": 18244, + "adors": 66887, + "ados": 5553, + "adow": 5843, + "adows": 21060, + "adox": 39097, + "adr": 14666, + "adratic": 88678, + "adress": 82236, + "adresse": 60275, + "adro": 89676, + "ads": 7664, + "adt": 25720, + "adult": 67856, + "adv": 14290, + "advance": 59320, + "advanced": 41779, + "advert": 40240, + "advertisement": 93148, + "advisor": 81794, + "adx": 38285, + "ady": 7604, + "ae": 5918, + "aea": 70335, + "aec": 71221, + "aed": 61454, + "aeda": 34660, + "ael": 5891, + "aeper": 73665, + "aepernick": 73732, + "aes": 76584, + "af": 2577, + "afa": 35834, + "afari": 59639, + "afb": 96834, + "afc": 94298, + "afd": 92139, + "afe": 5645, + "afen": 67558, + "afety": 38618, + "aff": 2649, + "affe": 37780, + "affected": 31057, + "affen": 71378, + "affer": 72153, + "affiliate": 78340, + "affle": 44090, + "affles": 80535, + "afi": 45910, + "afia": 48451, + "afil": 42962, + "afka": 21883, + "afone": 79294, + "afort": 52713, + "afs": 71498, + "after": 10694, + "afx": 34178, + "afür": 60119, + "ag": 351, + "aga": 12466, + "again": 32771, + "against": 68749, + "agal": 68112, + "agan": 18345, + "agar": 37712, + "agara": 59722, + "agas": 85052, + "agascar": 78675, + "agate": 46836, + "agation": 27137, + "age": 424, + "aged": 3279, + "agedList": 87765, + "agem": 14661, + "agement": 3448, + "agements": 61096, + "agen": 8535, + "agency": 55565, + "agenda": 79723, + "agens": 38796, + "agent": 8092, + "agenta": 47062, + "agento": 28053, + "agents": 54875, + "ager": 1409, + "agers": 11218, + "ages": 1134, + "aget": 84063, + "agg": 15718, + "aggable": 84896, + "agged": 96476, + "agger": 10114, + "aggi": 31616, + "aggio": 34159, + "aggregate": 59606, + "agh": 33756, + "aghan": 96169, + "agher": 63574, + "aghetti": 63965, + "agi": 36035, + "agic": 12780, + "agina": 80601, + "agination": 10353, + "aginator": 33128, + "agine": 23392, + "aging": 4118, + "agini": 90069, + "agit": 73872, + "agle": 32674, + "agli": 79056, + "agma": 5716, + "agment": 15601, + "agn": 3246, + "agna": 54957, + "agnar": 80753, + "agne": 23870, + "agner": 39438, + "agnet": 63233, + "agnetic": 38000, + "agnitude": 34715, + "agnosis": 49815, + "agnost": 8208, + "agnostic": 11953, + "agnostics": 15550, + "ago": 6304, + "agog": 54080, + "agogue": 70936, + "agon": 6113, + "agonal": 23450, + "agoon": 67413, + "agos": 80607, + "agr": 68305, + "agra": 12655, + "agram": 5745, + "agrams": 68772, + "agrant": 63487, + "agraph": 9216, + "agree": 68100, + "agrid": 45327, + "ags": 2032, + "agt": 62713, + "agu": 46720, + "agua": 54648, + "aguay": 83142, + "ague": 4663, + "agues": 13488, + "agus": 58738, + "agy": 81670, + "ah": 1466, + "aha": 13546, + "ahaha": 73535, + "ahan": 22051, + "ahas": 91966, + "ahead": 42787, + "ahi": 51106, + "ahir": 50998, + "ahkan": 74286, + "ahl": 16722, + "ahlen": 58193, + "ahn": 29560, + "aho": 28582, + "ahoma": 27840, + "ahoo": 16967, + "ahr": 15113, + "ahrain": 57958, + "ahren": 21176, + "ahrenheit": 47910, + "ahrung": 65224, + "ahrungen": 74374, + "ahu": 26695, + "ahun": 29727, + "ai": 2143, + "aic": 61388, + "aid": 3779, + "aida": 35444, + "aidu": 35801, + "aight": 7386, + "aign": 4109, + "aigned": 69969, + "ail": 604, + "ailability": 12535, + "ailable": 2094, + "ailand": 25178, + "ailed": 5687, + "ailer": 38782, + "ailing": 14277, + "aille": 50909, + "ailles": 86355, + "ails": 6209, + "ailure": 9373, + "aily": 5069, + "aim": 2640, + "aimassage": 55950, + "aims": 25676, + "ain": 466, + "aina": 72622, + "aincontri": 74472, + "aine": 8346, + "ained": 2627, + "ainen": 55664, + "ainer": 1743, + "ainers": 20568, + "aines": 32803, + "aining": 2056, + "ainless": 16459, + "ainment": 72612, + "ains": 1735, + "aint": 1641, + "ainted": 30229, + "aintenance": 24559, + "ainter": 73541, + "ainties": 61024, + "ainting": 70859, + "ainty": 18271, + "air": 1310, + "aira": 89780, + "aire": 11998, + "aired": 24274, + "aires": 17276, + "airie": 74590, + "airo": 25549, + "airobi": 83452, + "airport": 86615, + "airro": 37658, + "airs": 4720, + "airy": 21939, + "ais": 2782, + "aisal": 91020, + "aise": 4016, + "aised": 23964, + "aiser": 39266, + "aises": 21868, + "aising": 50126, + "aison": 56188, + "aisy": 49056, + "ait": 1315, + "aits": 56479, + "aj": 1630, + "aja": 19042, + "ajan": 96286, + "ajar": 50904, + "ajaran": 84033, + "ajas": 66822, + "ajax": 7242, + "aje": 11058, + "ajes": 35450, + "aji": 41788, + "ajo": 18734, + "ajor": 3035, + "ajs": 23720, + "aju": 51413, + "ajÄħ": 32873, + "ajÄħc": 95251, + "ajÄħce": 134287, + "ajÄħcy": 135390, + "ak": 585, + "aka": 13334, + "akah": 76427, + "akan": 19262, + "akash": 93995, + "ake": 726, + "akeFromNib": 69729, + "aked": 7741, + "akedirs": 35567, + "akedown": 77245, + "aken": 3366, + "akening": 73362, + "akens": 70737, + "aker": 4407, + "akers": 8312, + "akes": 2050, + "akespeare": 36871, + "akest": 74141, + "aket": 60942, + "akeup": 62744, + "akh": 21758, + "akhir": 64858, + "akhstan": 63747, + "aki": 14624, + "akin": 42230, + "aking": 1765, + "akis": 75777, + "akistan": 69293, + "akit": 97401, + "akk": 30041, + "aklı": 131920, + "ako": 28786, + "akov": 88630, + "akra": 51403, + "aks": 9810, + "aksi": 37679, + "akt": 9913, + "akte": 28815, + "akter": 51767, + "aktion": 67726, + "aktiv": 73607, + "aktu": 50654, + "aku": 23557, + "akukan": 49995, + "aky": 28100, + "akÄ±ÅŁ": 138676, + "al": 278, + "ala": 6053, + "alach": 71384, + "alah": 29393, + "alam": 16823, + "alamat": 53218, + "alan": 32283, + "alance": 4978, + "aland": 15337, + "alar": 7934, + "alaria": 56916, + "alarm": 56780, + "alars": 39977, + "alary": 39459, + "alborg": 68202, + "album": 21011, + "albums": 61471, + "alc": 16927, + "alchemy": 34905, + "alcon": 46214, + "alculate": 45441, + "ald": 4747, + "aldi": 76916, + "aldo": 39132, + "ale": 1574, + "aleb": 63930, + "aled": 5838, + "aleigh": 54830, + "alem": 21589, + "alement": 53860, + "alen": 35649, + "alendar": 12082, + "alent": 68161, + "aler": 13111, + "alers": 55785, + "alert": 5083, + "alertView": 86609, + "alerts": 91133, + "ales": 3831, + "alesce": 73250, + "alette": 28510, + "aleur": 59947, + "alex": 55875, + "aley": 41927, + "alez": 42945, + "aleza": 79152, + "alf": 3104, + "alfa": 84426, + "alg": 23881, + "algo": 72464, + "algorithm": 19688, + "ali": 7956, + "alia": 18848, + "alian": 10480, + "alias": 14956, + "aliases": 33924, + "alic": 31513, + "alice": 63195, + "alie": 74518, + "alien": 55318, + "align": 6612, + "aligned": 47142, + "alignment": 44598, + "alin": 35107, + "aling": 6132, + "alink": 66823, + "alion": 49741, + "alis": 34865, + "alist": 53072, + "ality": 2719, + "alité": 131275, + "alive": 50961, + "alk": 1692, + "alker": 36808, + "alking": 29755, + "all": 541, + "alla": 20905, + "allah": 76478, + "allas": 15863, + "allax": 44168, + "allback": 3420, + "alle": 5054, + "alled": 4736, + "allee": 53382, + "allel": 7957, + "allele": 94520, + "allen": 57301, + "alleng": 33769, + "allenge": 15832, + "allenges": 42370, + "aller": 13956, + "alleries": 97226, + "allery": 7447, + "alles": 95929, + "allest": 19236, + "allet": 7464, + "alley": 92166, + "alli": 95793, + "alling": 16740, + "allis": 95144, + "allo": 35937, + "alloc": 4742, + "alloca": 90594, + "allocate": 31191, + "allocated": 57372, + "allocation": 93239, + "allocator": 57631, + "allon": 92414, + "allow": 7183, + "allowed": 20967, + "alloween": 22685, + "allows": 55867, + "alls": 5583, + "ally": 745, + "alm": 7673, + "almart": 26952, + "almost": 59201, + "almö": 90756, + "alnum": 93362, + "alo": 12529, + "aload": 54396, + "alog": 30951, + "alogy": 79278, + "alom": 82028, + "alon": 43611, + "alone": 31576, + "along": 38293, + "alore": 43921, + "alous": 29443, + "alph": 30223, + "alpha": 7141, + "alphabet": 89453, + "already": 45484, + "als": 1127, + "alsa": 41826, + "alse": 710, + "alsex": 69488, + "also": 18654, + "alsy": 59284, + "alt": 3145, + "alta": 68051, + "altar": 70843, + "alte": 92320, + "alted": 61234, + "alten": 59493, + "alter": 37277, + "altern": 40412, + "alternate": 75362, + "alternative": 76995, + "altet": 93908, + "alth": 1864, + "although": 36141, + "alties": 23487, + "altimore": 87132, + "altitude": 98011, + "alto": 75677, + "altung": 61806, + "altura": 68491, + "alty": 10026, + "altı": 132917, + "alu": 36986, + "alue": 540, + "alus": 86127, + "always": 32122, + "aly": 5774, + "alysis": 9092, + "alytics": 17555, + "alyze": 55856, + "alyzed": 67842, + "alyzer": 27165, + "alık": 127781, + "alım": 131736, + "alÄ±ÅŁ": 72969, + "am": 309, + "ama": 3029, + "amac": 32607, + "amacare": 33824, + "amage": 9617, + "amaged": 91820, + "amaha": 85213, + "aman": 12715, + "amanho": 46305, + "amar": 15554, + "amarin": 30038, + "amas": 28089, + "amat": 41911, + "amate": 91066, + "amation": 26187, + "amax": 89627, + "amazon": 72753, + "amaño": 62313, + "amb": 2969, + "amba": 41908, + "ambah": 38187, + "ambda": 5698, + "amber": 15221, + "ambi": 88402, + "ambia": 52837, + "ambient": 59614, + "ambiguous": 90213, + "ambil": 77071, + "ambio": 64658, + "ambique": 89829, + "amble": 62934, + "ambre": 57466, + "amburg": 31448, + "amburger": 46675, + "amd": 67913, + "ame": 373, + "amed": 3606, + "ameda": 98071, + "amedi": 91906, + "amel": 35562, + "ameleon": 96811, + "amen": 24774, + "ament": 2838, + "amental": 43086, + "amentals": 77014, + "amente": 8980, + "amenti": 63480, + "amento": 10411, + "amentos": 36779, + "aments": 23558, + "amer": 15232, + "amera": 3436, + "amerate": 76966, + "american": 77718, + "ameron": 90612, + "ames": 971, + "amespace": 1680, + "amework": 3345, + "ami": 10606, + "amic": 3991, + "amics": 78178, + "amide": 65696, + "amient": 66328, + "amiento": 22820, + "amientos": 78309, + "amil": 20933, + "amiliar": 40405, + "amilies": 58904, + "amily": 5174, + "amin": 8608, + "amina": 34050, + "amination": 26837, + "amine": 19991, + "aminer": 64153, + "amines": 95981, + "aming": 6469, + "amins": 37825, + "aml": 9467, + "amm": 8666, + "amma": 13099, + "ammable": 99110, + "ammad": 53396, + "ammed": 24794, + "ammen": 36772, + "amment": 63556, + "ammer": 66599, + "ammers": 42057, + "ammo": 98327, + "ammu": 87039, + "amo": 21100, + "amodel": 40259, + "amon": 30836, + "among": 76305, + "amos": 10585, + "amoto": 56860, + "amount": 6045, + "amous": 22517, + "amp": 1121, + "ampa": 22639, + "ampaign": 18168, + "amped": 32198, + "amphetamine": 84790, + "ampie": 48298, + "ampil": 51628, + "ampilkan": 96313, + "amping": 30457, + "ampion": 13277, + "ampions": 9132, + "ampionship": 12241, + "ampire": 32033, + "ample": 1516, + "ampled": 65007, + "ampler": 34501, + "amples": 4023, + "ampling": 29709, + "ampo": 53730, + "ampoline": 59352, + "ampoo": 47595, + "ampp": 45726, + "amps": 14647, + "ampton": 42001, + "ampus": 43551, + "ams": 4122, + "amsung": 14441, + "amt": 35225, + "amu": 51972, + "amura": 83717, + "amus": 55355, + "amy": 26228, + "améli": 142829, + "améliorer": 142830, + "an": 276, + "ana": 3362, + "anagan": 89163, + "anager": 8184, + "anal": 48821, + "analy": 43703, + "analysis": 34484, + "analytics": 77652, + "analyze": 93221, + "aname": 93694, + "anan": 28618, + "anas": 25908, + "anax": 84612, + "anc": 1129, + "anca": 63245, + "ance": 681, + "anced": 4874, + "ancel": 4949, + "anceled": 38392, + "ancell": 14241, + "ancellable": 83603, + "ancellation": 49222, + "ancellationToken": 23927, + "ancellor": 34146, + "ancement": 16480, + "ancements": 38581, + "ancer": 11705, + "ancers": 31742, + "ances": 3020, + "ancestor": 66878, + "anch": 3497, + "anche": 27997, + "anches": 82181, + "anchise": 65690, + "anchor": 17109, + "anchors": 70499, + "ancia": 24158, + "ancial": 4898, + "ancias": 79641, + "ancies": 31637, + "ancing": 8974, + "anco": 18557, + "ancock": 74081, + "ancode": 87115, + "ancouver": 20471, + "ancy": 6572, + "ancybox": 61046, + "and": 437, + "andExpect": 40793, + "andFilterWhere": 84697, + "andReturn": 65805, + "andWhere": 65147, + "anda": 9817, + "andal": 16672, + "andalone": 84112, + "andan": 82808, + "andard": 7648, + "andas": 55433, + "andatory": 62088, + "andbox": 31536, + "ande": 22925, + "anded": 6465, + "andel": 37121, + "andelier": 78121, + "anden": 61700, + "ander": 8199, + "andering": 60280, + "anders": 83951, + "andes": 61352, + "andest": 90856, + "andex": 44526, + "andez": 37222, + "andi": 64574, + "andid": 4551, + "andidate": 17050, + "andidates": 26222, + "andin": 50601, + "anding": 25591, + "andise": 30670, + "andle": 3316, + "andler": 5252, + "andles": 20125, + "ando": 4883, + "andom": 2206, + "andon": 11037, + "andoned": 90144, + "andr": 73876, + "andra": 23274, + "andre": 79181, + "andro": 41724, + "android": 5954, + "ands": 2844, + "andscape": 71027, + "andum": 57285, + "andy": 13331, + "ane": 2145, + "anean": 32461, + "aned": 51430, + "anel": 2387, + "anela": 77440, + "aneous": 17666, + "aneously": 20315, + "anes": 14655, + "aney": 81186, + "anford": 94103, + "ang": 524, + "anga": 19372, + "angan": 18633, + "anganese": 65864, + "ange": 844, + "anged": 3726, + "angel": 17281, + "angelo": 87910, + "angelog": 61873, + "angement": 56633, + "angen": 59141, + "angent": 66451, + "angep": 55851, + "angepicker": 75321, + "anger": 4003, + "angered": 96309, + "angers": 14052, + "anges": 5520, + "anggal": 30509, + "anggan": 77257, + "anghai": 30070, + "angi": 78328, + "angible": 60086, + "anging": 8595, + "angkan": 86012, + "angle": 4044, + "angled": 38940, + "angler": 65548, + "angles": 17248, + "angling": 90104, + "ango": 5170, + "angs": 64487, + "angstrom": 79023, + "angu": 2325, + "anguage": 2616, + "anguages": 10513, + "anguard": 52190, + "angular": 4234, + "anh": 27924, + "ani": 5559, + "ania": 9166, + "anian": 74512, + "anic": 31270, + "anical": 44883, + "anie": 19151, + "aniel": 12046, + "aniem": 96308, + "anim": 15968, + "animal": 47899, + "animals": 76796, + "animate": 22030, + "animated": 19700, + "animation": 23050, + "animations": 52927, + "anime": 94670, + "anine": 91256, + "aning": 43059, + "anio": 92708, + "anism": 67813, + "anitize": 58652, + "anity": 38270, + "aniu": 72431, + "anium": 34722, + "anj": 52091, + "anja": 93951, + "anje": 84957, + "ank": 1180, + "anka": 26671, + "ankan": 75082, + "anke": 88729, + "anked": 40772, + "anken": 89289, + "anker": 98248, + "ankind": 69270, + "anking": 32334, + "anko": 92852, + "anks": 4039, + "anky": 91816, + "anmar": 45716, + "ann": 1020, + "anna": 12642, + "annabin": 96753, + "annah": 43004, + "anne": 20368, + "anned": 7295, + "anneer": 90485, + "annel": 2594, + "annels": 6680, + "anner": 4887, + "anners": 24003, + "annes": 41156, + "anness": 82033, + "anni": 42661, + "annie": 71076, + "annies": 93242, + "anning": 5947, + "annis": 76283, + "anno": 21552, + "annon": 17436, + "annonce": 97623, + "annot": 3401, + "annotate": 91350, + "annotation": 24674, + "annotations": 39626, + "announce": 65512, + "announcement": 80309, + "annt": 93339, + "annual": 63609, + "anny": 12888, + "annya": 71098, + "année": 84235, + "ano": 5652, + "anoi": 72703, + "anoia": 69895, + "anol": 38610, + "anon": 58910, + "anonymous": 25280, + "anooga": 87002, + "anos": 43605, + "another": 41963, + "anova": 85463, + "ans": 596, + "ansa": 93602, + "ansas": 13805, + "anse": 69654, + "ansen": 60865, + "anship": 34302, + "ansi": 52067, + "ansible": 91742, + "ansion": 10501, + "ansk": 40774, + "anske": 87211, + "ansom": 47231, + "anson": 36545, + "ansson": 80165, + "anst": 61194, + "answer": 9217, + "answered": 56724, + "answers": 24710, + "ant": 517, + "anta": 8260, + "antage": 24806, + "antaged": 74140, + "antages": 37786, + "antal": 59693, + "antan": 66189, + "antanamo": 95623, + "antar": 88965, + "antas": 76814, + "antasy": 96937, + "antd": 36881, + "ante": 4942, + "anted": 7566, + "antee": 13843, + "anteed": 58250, + "anten": 54834, + "anter": 80030, + "antes": 15477, + "anth": 31229, + "antha": 57869, + "anti": 15359, + "antiago": 42974, + "antic": 8159, + "antically": 80949, + "anticipated": 92778, + "antics": 36233, + "antidad": 26793, + "antine": 38357, + "antino": 95364, + "antis": 37866, + "antity": 5759, + "antium": 97925, + "antlr": 44964, + "antly": 17548, + "anto": 13561, + "antom": 30002, + "anton": 30954, + "antor": 88167, + "antro": 86190, + "antry": 31210, + "ants": 1783, + "antt": 88688, + "antu": 91306, + "antwort": 69046, + "antz": 89497, + "anus": 57963, + "anut": 37818, + "anuts": 72752, + "anvas": 11947, + "any": 3767, + "anya": 24074, + "anyahu": 41559, + "anyak": 50303, + "anych": 44453, + "anye": 48283, + "anyl": 89914, + "anything": 72154, + "anz": 12070, + "anza": 26466, + "anzeigen": 88860, + "anzi": 71818, + "ança": 90998, + "ao": 3441, + "aoke": 68505, + "aos": 64866, + "ap": 391, + "apGestureRecognizer": 54614, + "apa": 18321, + "apache": 58072, + "apan": 5359, + "apanese": 28689, + "apas": 88027, + "apat": 25760, + "apatkan": 90543, + "ape": 2027, + "apeake": 96556, + "aped": 10185, + "apel": 81100, + "apellido": 55303, + "aper": 3191, + "apers": 9532, + "apes": 9337, + "apesh": 77931, + "apest": 27624, + "apeut": 26498, + "apeutic": 31386, + "apeutics": 87786, + "apg": 63360, + "apgolly": 64407, + "aph": 1342, + "aphael": 74240, + "aphore": 25545, + "aphrag": 94796, + "api": 2068, + "apiKey": 57371, + "apia": 79370, + "apid": 43121, + "apikey": 72174, + "aping": 14216, + "apiro": 69479, + "apis": 13725, + "apist": 59229, + "apixel": 94195, + "apk": 93812, + "apl": 72289, + "apo": 65552, + "apol": 16657, + "apolis": 23814, + "apollo": 53610, + "apon": 10736, + "apons": 25778, + "apor": 20715, + "aporan": 71739, + "aporation": 95549, + "apore": 17665, + "apos": 30415, + "apot": 89901, + "app": 676, + "appId": 74829, + "appName": 85102, + "appa": 27180, + "appable": 86384, + "appe": 33431, + "appear": 40766, + "appearance": 96655, + "appeared": 98467, + "apped": 5677, + "appen": 63760, + "append": 5090, + "appendChild": 42609, + "appendTo": 46603, + "apper": 3106, + "appers": 27821, + "apphire": 55105, + "appid": 58342, + "appiness": 66291, + "apping": 3629, + "appings": 23036, + "appl": 76096, + "apple": 22377, + "applicant": 90264, + "application": 5132, + "applications": 82453, + "apply": 10280, + "appoint": 11671, + "appointed": 70116, + "appointment": 51101, + "appro": 15707, + "appropri": 75471, + "appropriate": 28128, + "approval": 54439, + "approve": 49221, + "approved": 33912, + "approx": 48053, + "approximately": 96736, + "apps": 27635, + "appy": 11144, + "apr": 59817, + "apro": 58848, + "après": 139320, + "aps": 2625, + "apse": 7477, + "apsed": 9200, + "apses": 78290, + "apsible": 68561, + "apsulation": 86115, + "apt": 2689, + "aptcha": 25431, + "apter": 2873, + "apters": 17425, + "aptic": 52174, + "aption": 10162, + "aptive": 27781, + "aptop": 16386, + "aptops": 40973, + "aptor": 32657, + "aptors": 57412, + "apture": 11850, + "aptured": 80228, + "apult": 72806, + "apur": 99040, + "apus": 67074, + "apy": 9667, + "apyrus": 89194, + "aq": 36306, + "aqu": 71571, + "aque": 19659, + "ar": 277, + "arDown": 59342, + "arLayout": 16814, + "arParams": 63752, + "arResult": 59081, + "ara": 5059, + "arah": 42418, + "arak": 27885, + "aram": 637, + "aramel": 46604, + "aran": 21320, + "araoh": 81197, + "aras": 72842, + "arat": 46857, + "aravel": 61653, + "arb": 75270, + "arbeit": 59908, + "arbon": 51645, + "arbonate": 96242, + "arc": 8198, + "arcer": 42986, + "arch": 1113, + "archical": 45234, + "arching": 82285, + "architecture": 80419, + "archive": 16019, + "archives": 95812, + "archivo": 61947, + "archs": 86050, + "archy": 15272, + "arcy": 96379, + "ard": 567, + "arda": 71526, + "ardash": 55688, + "arde": 41341, + "arded": 20958, + "arden": 8341, + "ardi": 36389, + "ardin": 53790, + "arding": 28410, + "ardless": 29820, + "ardo": 20473, + "ardon": 45296, + "ardown": 37496, + "ards": 2347, + "ardu": 82061, + "arduino": 90201, + "ardware": 37750, + "ardy": 90574, + "are": 546, + "area": 4798, + "areas": 32537, + "ared": 1605, + "arefa": 75251, + "arehouse": 20256, + "arel": 30431, + "arella": 74931, + "arem": 68368, + "aremos": 80184, + "aren": 9151, + "arena": 30527, + "arence": 68677, + "arend": 52845, + "arendra": 63494, + "arent": 1390, + "arently": 30127, + "arer": 60470, + "ares": 5403, + "arest": 15432, + "aret": 11302, + "areth": 56265, + "arez": 38514, + "arf": 59226, + "arg": 858, + "arga": 17922, + "argar": 33892, + "argas": 74079, + "argc": 23307, + "arge": 2744, + "arged": 92249, + "argent": 76117, + "arger": 41776, + "arges": 58240, + "argest": 32381, + "arget": 1284, + "argin": 2904, + "arging": 74716, + "argins": 81839, + "argo": 12088, + "argon": 70821, + "argout": 60490, + "args": 2116, + "argument": 14479, + "arguments": 16370, + "argv": 6508, + "arhus": 91089, + "ari": 2780, + "aria": 10432, + "arial": 42721, + "arian": 8821, + "ariance": 36905, + "arians": 29527, + "ariant": 15341, + "arias": 47113, + "ariat": 40879, + "ariate": 49659, + "arie": 66445, + "aries": 5431, + "arih": 93783, + "arily": 6613, + "arin": 42740, + "arine": 74214, + "aring": 3249, + "ario": 3290, + "arios": 10100, + "arious": 53271, + "aris": 41860, + "arith": 56636, + "arity": 10748, + "arium": 33665, + "arius": 40221, + "ark": 838, + "arkan": 61886, + "arked": 42061, + "arker": 12856, + "arkers": 90041, + "arket": 54560, + "arkin": 92321, + "arking": 33452, + "arks": 7193, + "arl": 48258, + "arlo": 60928, + "arm": 2178, + "arma": 19726, + "armac": 71382, + "armacy": 75242, + "arme": 73860, + "armed": 30272, + "arming": 32902, + "armor": 42548, + "arms": 15914, + "arn": 1885, + "arna": 39215, + "arnation": 54049, + "arness": 23518, + "arning": 3177, + "arnings": 14857, + "aro": 17165, + "aron": 12765, + "aroo": 75765, + "around": 19454, + "arov": 93918, + "arp": 7876, + "arpa": 81897, + "arr": 1118, + "arra": 79030, + "arrant": 6598, + "arrants": 51080, + "arranty": 43190, + "arrass": 20920, + "array": 1653, + "arrays": 66893, + "arrera": 82131, + "arring": 19859, + "arrings": 51498, + "arris": 58322, + "arrison": 50003, + "arriv": 141345, + "arrival": 77079, + "arrivée": 141346, + "arrière": 134183, + "arro": 81201, + "arrow": 6044, + "arry": 11433, + "arrêt": 134455, + "ars": 1561, + "arse": 2583, + "arsed": 18112, + "arseille": 61160, + "arser": 10425, + "arsers": 40488, + "arshal": 28423, + "arsi": 60014, + "arsimp": 52209, + "arsing": 28598, + "arsity": 70880, + "art": 471, + "arta": 24969, + "arte": 19840, + "arten": 23118, + "arter": 5295, + "arters": 14361, + "arth": 46501, + "arti": 67541, + "artial": 20894, + "artic": 80887, + "article": 7058, + "articles": 16243, + "artifact": 63722, + "artin": 34845, + "artisan": 36251, + "artisanlib": 80159, + "artist": 18622, + "artists": 76635, + "artment": 3195, + "artner": 92307, + "arto": 35943, + "arton": 79496, + "arts": 7038, + "arty": 6720, + "artz": 21593, + "aru": 30144, + "arus": 50319, + "ary": 658, + "arya": 72168, + "aryana": 92974, + "aryawan": 75627, + "aryl": 77808, + "ará": 30741, + "arÃŃa": 82701, + "arı": 67128, + "as": 300, + "asInstanceOf": 49044, + "asString": 75422, + "asa": 15428, + "asad": 76611, + "asaki": 56409, + "asan": 45388, + "asant": 15596, + "asar": 66938, + "asbourg": 83150, + "asc": 5061, + "asca": 98440, + "ascade": 19977, + "ascal": 35840, + "ascar": 65888, + "ascending": 71900, + "ascii": 23324, + "ascimento": 56961, + "asco": 79357, + "ascript": 5825, + "asctime": 77077, + "ascular": 49704, + "ascus": 52643, + "asd": 79102, + "asdf": 76615, + "ase": 519, + "ased": 1475, + "asel": 83497, + "aseline": 49314, + "asename": 27100, + "aser": 12080, + "asers": 59730, + "ases": 2264, + "aset": 5028, + "aseña": 42523, + "ash": 988, + "asha": 30443, + "ashboard": 7349, + "ashed": 13571, + "asher": 33767, + "ashes": 14051, + "ashi": 30378, + "ashing": 19049, + "ashington": 31699, + "ashion": 8782, + "ashire": 96314, + "ashtra": 57381, + "asi": 10215, + "asia": 35159, + "asic": 5971, + "asics": 67981, + "aside": 28512, + "asier": 76683, + "asil": 29049, + "asily": 37267, + "asin": 50310, + "asing": 4422, + "asio": 39513, + "asion": 17766, + "asionally": 68391, + "asions": 87671, + "asis": 10718, + "asiswa": 55624, + "asive": 26547, + "asje": 26846, + "asjon": 93546, + "ask": 1073, + "aska": 16883, + "askan": 78351, + "askell": 42758, + "asket": 9553, + "askets": 48140, + "asking": 70705, + "asks": 4604, + "asl": 75366, + "asley": 73011, + "asm": 10530, + "asma": 22572, + "asmine": 38864, + "asmus": 81788, + "asn": 65536, + "aso": 38955, + "ason": 1497, + "asonic": 42275, + "asonry": 50793, + "asons": 98849, + "asp": 13367, + "aspberry": 35962, + "aspect": 57058, + "asper": 32261, + "aspers": 87629, + "ass": 395, + "assa": 55592, + "assador": 24500, + "assadors": 73253, + "assage": 38236, + "assandra": 70093, + "asse": 13256, + "assed": 59004, + "assel": 80561, + "assemble": 69108, + "assembly": 14993, + "assen": 27277, + "asser": 33758, + "assert": 2207, + "assertCount": 94031, + "assertEquals": 16422, + "assertFalse": 59041, + "assertInstanceOf": 63180, + "assertSame": 37774, + "assertTrue": 35637, + "asses": 50990, + "assessment": 64672, + "asset": 9852, + "assets": 5160, + "assi": 81543, + "assic": 64344, + "assign": 6983, + "assigned": 39021, + "assignment": 29951, + "assignments": 96310, + "assin": 43723, + "assing": 72832, + "assis": 32467, + "assist": 50966, + "assistant": 77091, + "assium": 48063, + "asso": 28863, + "assoc": 46877, + "associ": 24313, + "associate": 71531, + "associated": 49087, + "association": 54465, + "assume": 45051, + "assuming": 65363, + "assword": 1938, + "assy": 26703, + "ast": 559, + "asta": 14300, + "aste": 5525, + "asted": 15036, + "aster": 2300, + "astered": 87373, + "asters": 14199, + "asterxml": 27453, + "astery": 53571, + "astes": 26908, + "asthan": 71549, + "asti": 68460, + "astic": 5064, + "astically": 35319, + "astics": 38071, + "asticsearch": 27791, + "asting": 14725, + "astle": 74430, + "asto": 53385, + "aston": 94991, + "astos": 52993, + "astr": 20467, + "astreet": 61398, + "astro": 22975, + "asts": 11757, + "asty": 14980, + "astype": 21754, + "asu": 95277, + "asurable": 55812, + "asure": 3970, + "asured": 39312, + "asurement": 24359, + "asurer": 56294, + "asures": 23471, + "asuring": 68674, + "asury": 26102, + "asy": 6405, + "asyarak": 90322, + "asyarakat": 91280, + "async": 7692, + "asz": 52340, + "ası": 92285, + "at": 266, + "atLng": 32738, + "ata": 459, + "atab": 2096, + "atabase": 2211, + "atabases": 23822, + "atable": 15086, + "atables": 29375, + "ataire": 65649, + "ataires": 92451, + "atak": 68728, + "ataka": 60124, + "atal": 4212, + "ataloader": 70286, + "atalog": 7750, + "atan": 16225, + "atar": 6392, + "atars": 43858, + "atas": 19346, + "ataset": 8369, + "atasets": 76649, + "atat": 92484, + "atatype": 91162, + "atch": 754, + "atched": 34244, + "atcher": 28058, + "atches": 9118, + "atchet": 83992, + "atchewan": 58747, + "atching": 31924, + "ate": 349, + "ateau": 76625, + "ated": 657, + "atedRoute": 27195, + "atee": 95680, + "ateful": 20840, + "ateg": 1328, + "ategic": 89367, + "ategies": 69388, + "atego": 94072, + "ategor": 7593, + "ategori": 31234, + "ategoria": 16758, + "ategorias": 50382, + "ategorical": 46047, + "ategorie": 37473, + "ategories": 5268, + "ategorized": 57611, + "ategory": 2031, + "ategy": 10228, + "atel": 83326, + "ately": 2652, + "atem": 37246, + "atemala": 53450, + "atement": 5605, + "aten": 13518, + "ater": 962, + "ateral": 19165, + "aterangepicker": 85107, + "ateria": 28183, + "aterial": 2180, + "atern": 13375, + "aternion": 16351, + "aternity": 64043, + "aterno": 79556, + "aters": 11535, + "ates": 973, + "atest": 13893, + "atetime": 27662, + "ateur": 10831, + "ateurs": 35206, + "atever": 7415, + "ateway": 12043, + "atform": 3329, + "ath": 587, + "atha": 65726, + "atham": 68376, + "athan": 13981, + "athe": 68936, + "athed": 69637, + "ather": 1856, + "atherine": 27009, + "athering": 82795, + "athers": 19030, + "athi": 66531, + "athing": 43561, + "athlete": 89944, + "athlon": 76917, + "atholic": 95330, + "athom": 72455, + "athon": 23941, + "athroom": 77832, + "aths": 26286, + "athy": 19135, + "ati": 9307, + "atial": 22433, + "atic": 774, + "atica": 53297, + "atical": 61072, + "atically": 7022, + "atican": 43062, + "aticon": 85049, + "atics": 28370, + "atie": 25852, + "atient": 16389, + "atif": 49748, + "atile": 9010, + "atility": 31156, + "atin": 14768, + "ating": 1095, + "atings": 27689, + "atinum": 29090, + "atio": 6266, + "ation": 367, + "ationToken": 17044, + "ational": 1663, + "ationale": 37035, + "ationally": 29054, + "ations": 804, + "ationship": 15471, + "atis": 3605, + "atisation": 97671, + "atisch": 81591, + "atische": 69022, + "atisf": 7478, + "atisfaction": 51270, + "atisfied": 46265, + "atism": 71110, + "atitis": 81701, + "atitude": 17145, + "ativ": 19488, + "ativa": 27852, + "ativas": 51541, + "ative": 1388, + "atively": 7887, + "atives": 5859, + "ativity": 43415, + "ativo": 28250, + "ativos": 63445, + "atk": 53809, + "atl": 59335, + "ato": 4330, + "atoes": 20187, + "atoi": 51071, + "atoire": 66663, + "atology": 73914, + "atom": 21855, + "atomic": 6618, + "atoms": 65550, + "atomy": 40795, + "aton": 23535, + "atonin": 98412, + "atoon": 51675, + "ator": 850, + "atore": 44009, + "atori": 50637, + "atoria": 95310, + "atorial": 37936, + "atories": 37170, + "atorio": 36590, + "atorium": 65744, + "ators": 2973, + "atory": 5269, + "atos": 14030, + "atown": 87333, + "atr": 60978, + "atra": 39558, + "atre": 15715, + "atri": 2361, + "atrib": 75567, + "atrice": 67536, + "atrigesimal": 42487, + "atrix": 2555, + "atro": 35336, + "atron": 80089, + "ats": 1862, + "atsapp": 71444, + "atsby": 35514, + "atsu": 50743, + "att": 1587, + "atta": 31919, + "attach": 16330, + "attached": 67497, + "attachment": 21981, + "attachments": 45878, + "attack": 20566, + "attacks": 80110, + "atte": 62634, + "atted": 12127, + "attempt": 55505, + "atten": 14456, + "attend": 56458, + "attendance": 62625, + "attended": 84906, + "attention": 53103, + "atter": 1650, + "attered": 21924, + "attering": 30336, + "attern": 3227, + "atters": 10175, + "attery": 18670, + "atti": 30955, + "attice": 31791, + "attivitÃł": 138961, + "attle": 4626, + "attles": 90127, + "atto": 16755, + "attr": 2991, + "attrib": 44113, + "attribute": 9116, + "attributes": 12340, + "attro": 76947, + "attrs": 20468, + "atts": 20983, + "atty": 22908, + "atu": 35309, + "atum": 26253, + "atur": 2628, + "atura": 24296, + "atural": 4585, + "aturally": 71147, + "aturas": 86529, + "aturated": 34367, + "aturday": 6970, + "aturdays": 69704, + "ature": 1568, + "atures": 2789, + "aturing": 39228, + "aturity": 37854, + "atus": 1000, + "aty": 49971, + "atype": 20157, + "atypes": 60913, + "atz": 20175, + "atég": 93387, + "ató": 44028, + "atório": 64694, + "atÄĥ": 75480, + "au": 2863, + "aub": 43506, + "auc": 20436, + "auce": 75790, + "aucoup": 44885, + "auction": 79762, + "aud": 7880, + "audi": 15523, + "audio": 16736, + "audit": 48545, + "auen": 45258, + "auer": 27097, + "auf": 50528, + "aug": 19179, + "auga": 89524, + "auge": 19392, + "aukee": 31987, + "aul": 4943, + "auled": 74667, + "ault": 945, + "aunch": 18423, + "aupt": 51898, + "aur": 4110, + "aura": 21563, + "aurant": 14334, + "aurants": 33984, + "aurus": 42513, + "aus": 11855, + "ausal": 79074, + "ause": 3454, + "auses": 35143, + "ausible": 42636, + "auss": 63051, + "aussian": 46972, + "aut": 2717, + "auth": 3242, + "authenticate": 50060, + "authenticated": 57707, + "authentication": 45890, + "author": 3094, + "authority": 84454, + "authorization": 39554, + "authorize": 52022, + "authorized": 19084, + "authors": 47005, + "autical": 67167, + "autiful": 34677, + "auto": 3902, + "autocomplete": 56514, + "autoload": 38513, + "autom": 27073, + "automatic": 83456, + "automation": 67080, + "autop": 87321, + "autoplay": 90323, + "autor": 62045, + "autorelease": 77386, + "autoreleasepool": 87422, + "autos": 79301, + "auty": 57951, + "auté": 93843, + "auważ": 139309, + "aux": 11981, + "av": 402, + "ava": 2907, + "avad": 37055, + "avadoc": 43425, + "avage": 67156, + "avail": 70263, + "availability": 53062, + "available": 10334, + "aval": 21761, + "avan": 30905, + "avana": 28245, + "avanaugh": 34714, + "avar": 36232, + "avascript": 6173, + "avatar": 11962, + "avatars": 74814, + "avax": 10914, + "avaÅŁ": 127498, + "ave": 523, + "avec": 87323, + "aved": 4141, + "avel": 3878, + "aveled": 28185, + "avelength": 34861, + "aven": 5276, + "avenous": 80343, + "avenport": 96876, + "aver": 7255, + "average": 17202, + "avern": 33718, + "avers": 21565, + "aversable": 97513, + "aversal": 45545, + "averse": 22439, + "avery": 24698, + "aves": 4693, + "avez": 51559, + "avg": 13961, + "avi": 6190, + "avia": 34002, + "avian": 59178, + "avic": 88445, + "avicon": 27921, + "avid": 15212, + "avier": 30648, + "aviest": 98362, + "avig": 2957, + "avigate": 10379, + "avigation": 4372, + "avigator": 10325, + "avin": 34616, + "aving": 2317, + "avings": 45751, + "aviolet": 84211, + "avior": 5640, + "aviors": 22176, + "aviour": 8897, + "aviours": 59256, + "avirus": 23779, + "avis": 22354, + "avit": 43110, + "avity": 11718, + "avl": 67311, + "avn": 30597, + "avo": 27217, + "avoid": 47856, + "avor": 3292, + "avored": 75386, + "avorite": 15622, + "avorites": 29379, + "avors": 50485, + "avour": 25998, + "avourite": 51388, + "avourites": 69316, + "avr": 69584, + "avra": 62867, + "avras": 88123, + "avs": 38751, + "avy": 5663, + "aw": 672, + "awa": 14077, + "awah": 78528, + "awai": 35969, + "awaii": 70954, + "await": 11421, + "awaiter": 72976, + "awan": 40929, + "award": 68606, + "aware": 19541, + "away": 13757, + "aways": 39590, + "awe": 84177, + "awei": 37585, + "awesome": 16875, + "awi": 40878, + "awk": 23397, + "awks": 29278, + "awl": 87448, + "awn": 6379, + "awner": 67193, + "awning": 50846, + "awns": 39996, + "aws": 8635, + "awy": 95719, + "ax": 706, + "axe": 49739, + "axed": 51451, + "axes": 19992, + "axies": 56988, + "axios": 21325, + "axis": 7184, + "axon": 51836, + "axter": 72539, + "axy": 13773, + "ay": 352, + "aya": 12589, + "ayah": 84074, + "ayan": 43857, + "ayar": 75146, + "ayaran": 72902, + "ayas": 97340, + "aybe": 49791, + "aycast": 51287, + "aydın": 139959, + "aye": 59955, + "ayed": 42895, + "ayer": 1135, + "ayers": 5074, + "ayette": 48289, + "aygı": 143877, + "aying": 17270, + "aylight": 93864, + "ayload": 6989, + "aylor": 14023, + "ayment": 6074, + "ayne": 96521, + "ayo": 75980, + "ayout": 1407, + "ays": 942, + "ayscale": 50348, + "aysia": 58085, + "ayı": 126379, + "az": 1370, + "aza": 12707, + "azaar": 68338, + "azar": 33044, + "azard": 57657, + "aze": 9832, + "azed": 27011, + "azeera": 72671, + "azel": 68326, + "azen": 82821, + "azer": 22850, + "azers": 60307, + "azes": 73785, + "azi": 17812, + "azine": 9918, + "azines": 28741, + "azing": 6657, + "azio": 56846, + "azione": 15013, + "azioni": 31651, + "azo": 40184, + "azole": 75901, + "azon": 5522, + "azor": 42866, + "azu": 39129, + "azure": 39495, + "azy": 13619, + "azz": 9802, + "azzi": 71318, + "azzo": 39673, + "azı": 131035, + "azÄĥ": 86903, + "ação": 12967, + "ações": 33465, + "añ": 32369, + "aña": 41261, + "aÄĩ": 78942, + "aÄŁ": 143786, + "aÄŁa": 131106, + "aÄŁlı": 143787, + "aģı": 125494, + "aģını": 141336, + "aż": 59919, + "aÅĤ": 126256, + "aÅĤa": 129948, + "aÅĤem": 133325, + "b": 65, + "ba": 4645, + "baar": 70217, + "bab": 47722, + "babel": 31211, + "bable": 59051, + "baby": 77964, + "bac": 55877, + "bach": 34538, + "back": 1419, + "backend": 20942, + "background": 6742, + "backgroundColor": 39989, + "backs": 24113, + "backup": 31371, + "backward": 81196, + "bad": 13855, + "badge": 31524, + "bag": 21250, + "bagai": 45497, + "bage": 20652, + "bags": 52394, + "bah": 68143, + "bai": 50407, + "bairro": 58383, + "bak": 69822, + "bal": 13313, + "balance": 21571, + "balanced": 58402, + "balances": 75427, + "ball": 3959, + "balls": 45518, + "bam": 80567, + "ban": 6850, + "banana": 87747, + "band": 7053, + "bands": 41906, + "bane": 35914, + "bang": 27174, + "bank": 17033, + "banks": 78946, + "banner": 27128, + "bao": 64841, + "bar": 2257, + "bara": 24095, + "barang": 86468, + "barcode": 64286, + "bard": 65919, + "bare": 54102, + "bars": 24950, + "bart": 40579, + "baru": 89360, + "bas": 17797, + "base": 3152, + "basePath": 79163, + "baseUrl": 37691, + "based": 29939, + "baseline": 75958, + "basename": 42953, + "bases": 78267, + "bash": 46216, + "basic": 22342, + "basis": 88826, + "basket": 59764, + "bast": 28011, + "bastian": 36497, + "bat": 13718, + "batch": 14049, + "bate": 70529, + "batim": 54748, + "batis": 36588, + "battery": 74359, + "battle": 64954, + "bau": 71832, + "baugh": 61639, + "baum": 58511, + "bay": 46264, + "baz": 42573, + "baÅŁÄ±": 137805, + "bb": 6066, + "bbb": 53151, + "bbbb": 87609, + "bbc": 84185, + "bbe": 64073, + "bben": 36039, + "bbie": 54871, + "bbing": 67341, + "bble": 59880, + "bbox": 58456, + "bc": 8904, + "bcc": 69018, + "bcd": 91490, + "bcm": 91185, + "bcrypt": 65498, + "bd": 8940, + "bdb": 85760, + "bdd": 50442, + "be": 1371, + "beam": 53871, + "bean": 17479, + "beans": 27340, + "bear": 67660, + "bearer": 95131, + "bearing": 87350, + "beat": 22227, + "beautiful": 86944, + "beb": 90805, + "bec": 16692, + "because": 27653, + "becca": 37196, + "beck": 54077, + "becue": 51985, + "bed": 2721, + "bedPane": 80543, + "bedo": 88987, + "bedtls": 76032, + "bee": 32031, + "beeld": 78931, + "been": 81750, + "beer": 47795, + "bef": 57647, + "before": 14801, + "beg": 51153, + "begin": 7265, + "beginTransaction": 58110, + "beh": 29898, + "behavior": 55866, + "bei": 62401, + "being": 34114, + "beit": 19985, + "beiten": 50656, + "beiter": 72964, + "beits": 64002, + "bek": 76514, + "bel": 9779, + "belie": 31798, + "belief": 59776, + "bell": 17250, + "bellion": 85763, + "belongs": 80897, + "belongsTo": 42867, + "below": 53177, + "belt": 75848, + "belum": 78244, + "ben": 7964, + "bench": 27024, + "benchmark": 79186, + "benef": 67144, + "benh": 38795, + "benhavn": 40977, + "bens": 55874, + "ber": 652, + "berapa": 74402, + "bere": 61286, + "berg": 7725, + "berger": 48020, + "berman": 80195, + "bern": 76712, + "bero": 80784, + "beros": 92795, + "berra": 62341, + "berries": 35889, + "berry": 15357, + "bers": 1902, + "bersome": 93424, + "bert": 9160, + "bery": 30204, + "bes": 9433, + "besch": 143729, + "beschäft": 143730, + "beschäftig": 143731, + "best": 15862, + "bestos": 64164, + "bet": 28551, + "beta": 19127, + "beth": 82044, + "bett": 90957, + "better": 57134, + "between": 41867, + "bew": 81936, + "bez": 143213, + "bezpieczeÅĦst": 143214, + "bezpieczeÅĦstw": 143215, + "bf": 13233, + "bfd": 81664, + "bff": 92106, + "bg": 12220, + "bgcolor": 95900, + "bh": 51871, + "bi": 8221, + "bia": 14500, + "bial": 46959, + "bian": 13162, + "bias": 60442, + "biased": 64301, + "bib": 57280, + "bic": 93498, + "bid": 20648, + "bidden": 22108, + "bie": 21745, + "bies": 56673, + "big": 16154, + "bigint": 84674, + "bih": 40116, + "bij": 61269, + "bike": 55155, + "bil": 48838, + "bild": 60365, + "bilder": 86389, + "bildung": 93590, + "bilit": 126947, + "bilité": 133296, + "bilitÃł": 130670, + "bill": 29642, + "billing": 38637, + "bilt": 69724, + "bin": 6863, + "binary": 25891, + "binations": 73629, + "bind": 7666, + "bindParam": 33816, + "bindValue": 40424, + "binding": 31036, + "bindings": 65495, + "bindung": 90226, + "bine": 46561, + "bing": 7132, + "bins": 39330, + "bio": 37323, + "biology": 80062, + "bios": 69628, + "bir": 43855, + "bird": 22592, + "birds": 66361, + "birth": 27713, + "birthdate": 98531, + "birthday": 51265, + "bis": 53581, + "bish": 57699, + "bishop": 60278, + "bit": 4489, + "bitcoin": 83910, + "bite": 89935, + "bitmap": 40954, + "bitrary": 87851, + "bits": 11516, + "bitset": 77615, + "biz": 54673, + "bió": 139153, + "bj": 35751, + "bject": 800, + "bjerg": 91878, + "bk": 40029, + "bl": 2024, + "bla": 64726, + "black": 11453, + "blade": 62727, + "blah": 70614, + "blank": 10189, + "blas": 49297, + "blast": 63317, + "ble": 891, + "bled": 37659, + "blem": 10121, + "blems": 21699, + "blend": 82632, + "bler": 77462, + "blers": 70378, + "bles": 38763, + "blick": 83820, + "blind": 92909, + "bling": 9695, + "blings": 97463, + "blink": 54316, + "blk": 34989, + "blo": 38145, + "blob": 35112, + "block": 4574, + "blocked": 61348, + "blocking": 70356, + "blockquote": 48775, + "blocks": 21928, + "blog": 11659, + "blogs": 53903, + "blood": 50005, + "blr": 22134, + "blue": 12203, + "blur": 33542, + "bm": 29307, + "bmp": 52696, + "bn": 11081, + "bnb": 44002, + "bo": 749, + "boBox": 9405, + "boa": 69014, + "board": 2482, + "boarding": 37569, + "boards": 19270, + "boat": 37765, + "boats": 94358, + "bob": 47086, + "body": 2599, + "bohydr": 33558, + "bol": 30046, + "bold": 14824, + "bole": 60075, + "bolt": 52433, + "bomb": 78344, + "bon": 5970, + "bond": 64239, + "bone": 19756, + "bones": 80991, + "bons": 46520, + "bonus": 57469, + "boo": 32993, + "book": 2190, + "booking": 21278, + "bookmark": 60375, + "books": 12110, + "bool": 2641, + "boolean": 6117, + "boom": 95316, + "boost": 13000, + "boot": 4619, + "bootstrap": 6281, + "bor": 9368, + "borah": 60558, + "border": 9461, + "borg": 24210, + "born": 15998, + "borne": 30996, + "boro": 34867, + "borough": 34079, + "borrow": 71141, + "bos": 78448, + "bose": 8297, + "boss": 33314, + "bot": 6331, + "bote": 92918, + "both": 21028, + "bots": 61905, + "bottom": 14860, + "bounce": 64302, + "bound": 10891, + "boundary": 65180, + "bounded": 65686, + "bounding": 61579, + "bounds": 34019, + "bour": 15385, + "bourg": 67545, + "bourne": 21436, + "bout": 72901, + "bove": 30924, + "bow": 15439, + "bower": 56949, + "bows": 54531, + "box": 2011, + "boxed": 79075, + "boxes": 22204, + "boxing": 89174, + "boy": 17184, + "boys": 29023, + "bp": 17808, + "bpp": 71145, + "bps": 31180, + "br": 1323, + "bra": 14383, + "brace": 31237, + "bracht": 85426, + "braco": 96967, + "brahim": 57932, + "brain": 53060, + "brains": 32625, + "brakk": 47540, + "bral": 41643, + "branch": 17940, + "branches": 67839, + "brand": 13473, + "brands": 69194, + "brane": 87454, + "bras": 70761, + "braska": 33788, + "brate": 64116, + "brates": 98768, + "bre": 20512, + "bread": 47516, + "breadcrumb": 24234, + "breadcrumbs": 35029, + "break": 8960, + "breaker": 64121, + "breaking": 36657, + "bred": 90742, + "brero": 77085, + "brew": 21632, + "brick": 69673, + "brid": 16223, + "bridge": 13709, + "brids": 83588, + "brief": 6658, + "brig": 82032, + "bright": 72116, + "brightness": 99056, + "bring": 81377, + "bringing": 79988, + "bris": 28172, + "brit": 31314, + "brities": 34695, + "bro": 14877, + "broadcast": 54162, + "broken": 48909, + "broker": 69309, + "bron": 67457, + "brook": 42266, + "brown": 64461, + "browse": 61213, + "browser": 22468, + "brtc": 78536, + "bru": 32613, + "bruar": 93289, + "bruary": 6812, + "brush": 36061, + "bs": 1279, + "bsd": 51835, + "bserv": 4840, + "bservable": 8293, + "bservice": 64693, + "bsite": 3222, + "bsites": 11712, + "bsolute": 16855, + "bson": 91723, + "bsp": 5417, + "bst": 24585, + "bstract": 48494, + "bsub": 69683, + "bt": 12755, + "btc": 92634, + "btn": 3905, + "bu": 19644, + "buah": 86461, + "buat": 69842, + "bubble": 77489, + "buch": 72043, + "bucket": 30410, + "buckets": 92417, + "bucks": 39788, + "bud": 69927, + "budget": 48406, + "buf": 5909, + "buff": 25976, + "buffer": 7573, + "buffers": 60433, + "bufio": 97551, + "bug": 2313, + "bugs": 56176, + "buie": 86460, + "build": 5834, + "builder": 17850, + "builders": 98507, + "building": 23296, + "built": 46239, + "builtin": 42457, + "bul": 16110, + "bulan": 86326, + "bulk": 67142, + "bull": 42004, + "bullet": 39460, + "bum": 5377, + "bundle": 34518, + "bundles": 86660, + "bung": 83592, + "buquerque": 68176, + "bur": 11240, + "burg": 10269, + "burger": 62452, + "burgh": 17816, + "burn": 21719, + "burse": 37982, + "bursement": 51073, + "burst": 57738, + "bury": 19603, + "bus": 10338, + "buscar": 79757, + "business": 26151, + "buster": 55552, + "busters": 75518, + "busy": 78467, + "but": 8088, + "butt": 92796, + "button": 2159, + "buttonShape": 84054, + "buttons": 24086, + "buy": 19783, + "buyer": 59307, + "buzz": 86126, + "bv": 54929, + "bw": 39824, + "bx": 21861, + "by": 1694, + "bye": 28374, + "byn": 39085, + "byname": 94878, + "byss": 55072, + "byte": 3782, + "byter": 71607, + "byterian": 79336, + "bytes": 9651, + "bz": 71701, + "bé": 91074, + "bö": 135514, + "bü": 129578, + "c": 66, + "ca": 924, + "caa": 87734, + "cab": 54793, + "cac": 83110, + "cache": 9360, + "cached": 32918, + "cad": 34455, + "cade": 32796, + "cadena": 86863, + "cae": 97987, + "caf": 68796, + "caff": 23778, + "caffe": 87760, + "caffold": 27864, + "cai": 95150, + "cairo": 62383, + "cake": 47384, + "cakes": 61992, + "cal": 5416, + "cala": 93380, + "calar": 59153, + "calc": 26586, + "calcul": 80630, + "calculate": 35597, + "calculator": 88821, + "cale": 2246, + "caled": 25228, + "calendar": 26258, + "caler": 63084, + "cales": 30196, + "caling": 81552, + "call": 6659, + "callable": 95192, + "callback": 13494, + "callbacks": 68311, + "called": 43882, + "callee": 86476, + "caller": 56126, + "calling": 73726, + "calloc": 65488, + "calls": 44620, + "cam": 11599, + "camatan": 71050, + "came": 6014, + "camel": 93321, + "camera": 24910, + "camp": 25238, + "campaign": 37339, + "campo": 56762, + "cams": 74250, + "can": 4814, + "cancel": 18515, + "cancellationToken": 95470, + "cancelled": 95282, + "candidate": 46274, + "candidates": 91345, + "canf": 20393, + "cannot": 33260, + "cano": 68277, + "canonical": 77942, + "cant": 65263, + "cantidad": 49331, + "canvas": 19943, + "cao": 28339, + "cap": 11346, + "capabilities": 94392, + "capability": 81251, + "capacity": 37387, + "cape": 5330, + "capital": 65063, + "capitalize": 81341, + "caps": 48233, + "capt": 43203, + "captcha": 54347, + "caption": 23347, + "capture": 45070, + "captures": 66326, + "car": 6918, + "caracter": 135523, + "caracterÃŃsticas": 135524, + "carbon": 73341, + "card": 4951, + "cards": 25024, + "care": 10506, + "career": 94537, + "caret": 54227, + "cargo": 66715, + "carousel": 23454, + "carrier": 65985, + "carry": 73765, + "cars": 50708, + "cart": 11452, + "cas": 24307, + "cascade": 49900, + "case": 5638, + "casecmp": 48072, + "cased": 91126, + "cases": 23910, + "cash": 41271, + "cassert": 66922, + "cast": 3829, + "castHit": 82720, + "caster": 32020, + "casting": 76332, + "castle": 35108, + "casts": 50034, + "cat": 4616, + "catalog": 26539, + "catch": 7173, + "cate": 42772, + "categoria": 43244, + "categorias": 93005, + "categorie": 86695, + "categories": 15497, + "category": 5471, + "categoryId": 61378, + "catid": 78677, + "cation": 64882, + "cats": 37452, + "caught": 61208, + "cause": 1563, + "cação": 90523, + "cb": 7221, + "cba": 93829, + "cbc": 80766, + "cbd": 90749, + "cc": 638, + "cca": 24441, + "ccak": 84262, + "ccb": 81317, + "ccc": 37054, + "cccc": 55597, + "ccd": 95840, + "cce": 86921, + "ccess": 1322, + "cci": 56708, + "ccion": 14059, + "ccione": 66319, + "cciones": 54703, + "cción": 22004, + "cco": 57661, + "ccoli": 59703, + "ccording": 6641, + "ccount": 4513, + "cctor": 88605, + "cd": 4385, + "cdb": 85301, + "cdc": 72026, + "cdecl": 66916, + "cdf": 58348, + "cdn": 12254, + "cdnjs": 38762, + "cdot": 50853, + "cdr": 64170, + "ce": 346, + "cea": 72938, + "cean": 11206, + "ceans": 42220, + "ceased": 94204, + "ceb": 84797, + "cec": 68955, + "ced": 1998, + "cede": 94874, + "cedes": 25701, + "cedure": 13196, + "cedures": 53691, + "cee": 70596, + "ceed": 4635, + "ceeded": 94206, + "cef": 65041, + "ceil": 22058, + "ceipt": 24130, + "ceive": 8238, + "ceived": 8771, + "ceiver": 12862, + "ceiving": 46344, + "cej": 57245, + "cek": 45700, + "cel": 3672, + "celain": 59142, + "celand": 35216, + "cele": 69765, + "celed": 33183, + "cell": 5873, + "cellent": 24746, + "cells": 36691, + "celona": 22365, + "cem": 53821, + "cen": 47322, + "cence": 35534, + "cene": 59351, + "cent": 1168, + "centage": 16133, + "centaje": 70399, + "cente": 77584, + "center": 3057, + "centers": 85441, + "central": 51811, + "centration": 93276, + "centre": 74117, + "centroid": 96026, + "cents": 38328, + "cep": 31652, + "ceph": 58722, + "cept": 1484, + "ceptar": 74958, + "ception": 995, + "ceptions": 34928, + "ceptive": 58274, + "ceptor": 15349, + "ceptors": 57550, + "cepts": 57771, + "cer": 3828, + "cerer": 68881, + "ceries": 55182, + "cern": 28544, + "cerpt": 34028, + "cerr": 23960, + "cers": 16832, + "cert": 12246, + "certificate": 63586, + "cery": 21505, + "ces": 1603, + "ceso": 54198, + "cess": 1120, + "cession": 25461, + "cessive": 86628, + "cesso": 29312, + "cest": 15184, + "cestor": 81340, + "ceu": 81130, + "ceÄŁi": 125329, + "cf": 9792, + "cff": 57823, + "cfg": 14072, + "cg": 27446, + "cgi": 72415, + "ch": 331, + "cha": 6447, + "chaft": 20542, + "chai": 45081, + "chain": 8819, + "chains": 58358, + "chair": 34196, + "chal": 85567, + "chalk": 86388, + "challenge": 57365, + "chan": 5658, + "chandle": 72621, + "chang": 77982, + "change": 3373, + "changed": 17353, + "changer": 72831, + "changes": 19365, + "changing": 51713, + "channel": 10119, + "channels": 32425, + "chant": 11681, + "chantment": 68889, + "chants": 98806, + "chap": 92319, + "chapter": 40226, + "char": 1762, + "charAt": 70857, + "character": 19190, + "characters": 47485, + "charg": 17685, + "charge": 13891, + "charged": 36961, + "charger": 41456, + "charges": 62299, + "charging": 62359, + "charm": 98409, + "chars": 19255, + "charset": 25327, + "chart": 15941, + "chartInstance": 96246, + "charted": 90818, + "charts": 36584, + "chas": 58679, + "chat": 9686, + "che": 1528, + "cheap": 61690, + "check": 2028, + "checkBox": 72230, + "checkbox": 10263, + "checked": 7549, + "checker": 69955, + "checking": 57854, + "checkout": 33035, + "checkpoint": 69134, + "checks": 49383, + "checksum": 70740, + "ched": 2397, + "chedule": 8796, + "cheduled": 26644, + "cheduler": 15222, + "chedulers": 36212, + "chedules": 49613, + "cheduling": 44356, + "chef": 75079, + "cheid": 57926, + "cheiden": 98975, + "chein": 63859, + "chem": 2407, + "chema": 3416, + "chemas": 31126, + "chematic": 81049, + "cheme": 8058, + "chemes": 66346, + "chemical": 30956, + "chemist": 74088, + "chemistry": 51655, + "chemy": 24088, + "chen": 7522, + "cheng": 67184, + "cheon": 75169, + "cher": 9034, + "cherche": 29510, + "chers": 98623, + "ches": 8528, + "chest": 26320, + "chester": 14819, + "chestra": 37160, + "chet": 39113, + "chez": 34015, + "chg": 65940, + "chi": 14604, + "chia": 89546, + "chie": 5467, + "chied": 31722, + "chief": 61526, + "child": 3048, + "childNodes": 75773, + "children": 5864, + "childs": 86500, + "chimp": 90340, + "chin": 59071, + "china": 80136, + "chine": 87895, + "ching": 26977, + "chio": 52401, + "chip": 29662, + "chk": 35896, + "chl": 67056, + "chlor": 71867, + "chluss": 93383, + "chmod": 56274, + "chn": 2125, + "chner": 85577, + "chnitt": 88718, + "cho": 958, + "chod": 128030, + "chodzÄħ": 128031, + "chodzÄħc": 138993, + "choice": 11746, + "choices": 24789, + "chool": 2135, + "choose": 27052, + "chooser": 67360, + "chor": 6150, + "chos": 45248, + "chosen": 63075, + "chr": 16789, + "chrift": 82850, + "christ": 82132, + "chrom": 40384, + "chrome": 31902, + "chron": 7410, + "chrono": 25768, + "chs": 55457, + "chsel": 90830, + "cht": 13920, + "chte": 38486, + "chten": 42888, + "chter": 66685, + "chts": 54327, + "chu": 65752, + "chunk": 25979, + "chunks": 84263, + "church": 64194, + "chure": 58334, + "chw": 65845, + "chwitz": 90931, + "chy": 57964, + "chè": 134284, + "ché": 74017, + "ci": 5855, + "cia": 24119, + "cial": 15781, + "cid": 20558, + "cidade": 52443, + "cido": 54351, + "cie": 24161, + "cient": 12295, + "cies": 69007, + "cil": 81023, + "cimal": 87099, + "cimiento": 66037, + "cin": 18103, + "cing": 6125, + "cio": 60830, + "cion": 22613, + "cip": 79657, + "cipher": 67586, + "cir": 57032, + "circ": 43298, + "circle": 25857, + "cis": 78055, + "cision": 54704, + "cit": 53861, + "citation": 83147, + "cite": 67075, + "cities": 53009, + "city": 8926, + "cité": 135719, + "civil": 93130, + "ció": 87100, + "ción": 76931, + "ciÄħ": 129554, + "ciÄħg": 129555, + "ciÅĽni": 140045, + "cj": 79446, + "cjÄħ": 131176, + "cjÄĻ": 129165, + "ck": 377, + "cka": 96257, + "cke": 60273, + "ckeditor": 62324, + "cken": 40188, + "cker": 15160, + "cket": 3996, + "ckett": 84829, + "cki": 97738, + "ckill": 94068, + "cko": 31463, + "ckpt": 68175, + "cks": 14553, + "cksÃ¥": 86586, + "ckt": 68412, + "cl": 564, + "cla": 53661, + "clado": 96881, + "claim": 7859, + "claimed": 21275, + "claimer": 13640, + "claims": 48561, + "clair": 48172, + "clamation": 32984, + "clamp": 85453, + "clang": 25380, + "clar": 12821, + "claration": 16490, + "clarations": 49121, + "clare": 67554, + "clared": 86251, + "clarsimp": 78160, + "clas": 86263, + "class": 1040, + "classCallCheck": 80198, + "classList": 33305, + "className": 13871, + "classNames": 83874, + "classed": 96737, + "classes": 8855, + "classic": 79709, + "classification": 65381, + "classified": 64874, + "classifier": 78100, + "classify": 94290, + "classmethod": 26755, + "classname": 62806, + "classnames": 55113, + "classpath": 79814, + "clause": 82082, + "clave": 40533, + "clazz": 82914, + "clc": 78839, + "cle": 9148, + "clean": 18377, + "cleanup": 55235, + "clear": 7422, + "clearfix": 32453, + "clerosis": 90611, + "cles": 18638, + "clf": 69126, + "cli": 19521, + "click": 3678, + "clicked": 47013, + "clid": 75044, + "clide": 45323, + "clidean": 50078, + "client": 2972, + "clientId": 67683, + "cliente": 37771, + "clientes": 81644, + "clients": 31869, + "cliffe": 79456, + "climate": 93774, + "clin": 46581, + "cline": 87934, + "cling": 62098, + "clinic": 89265, + "clinical": 90799, + "clip": 7974, + "clipboard": 70848, + "clipse": 9473, + "clist": 84009, + "cljs": 32934, + "clk": 32583, + "clo": 32578, + "cloak": 88751, + "clock": 20666, + "clone": 19982, + "close": 5552, + "closed": 34087, + "closest": 36461, + "closing": 85777, + "closure": 46916, + "cloth": 87954, + "cloud": 12361, + "clr": 60947, + "cls": 18074, + "club": 19862, + "clubs": 76285, + "clud": 1485, + "clude": 857, + "cluded": 10181, + "cludes": 7396, + "cluding": 10910, + "cluir": 44721, + "clus": 4163, + "clusion": 8957, + "clusions": 23529, + "clusive": 8336, + "cluster": 18855, + "clusters": 78521, + "clusão": 142528, + "cly": 84145, + "clé": 133861, + "cm": 6226, + "cmath": 33725, + "cmb": 95055, + "cmc": 94866, + "cmd": 8710, + "cmds": 92407, + "cmp": 7293, + "cmpeq": 99021, + "cms": 28711, + "cn": 14271, + "cname": 78488, + "cnt": 15853, + "cntl": 75017, + "co": 1015, + "coach": 88093, + "coal": 40465, + "coat": 67412, + "cob": 85556, + "cobra": 59410, + "coc": 77312, + "cock": 37153, + "cod": 20116, + "code": 1851, + "codec": 34607, + "coded": 65530, + "codegen": 95859, + "coder": 40170, + "codes": 25814, + "codigo": 33466, + "codile": 82674, + "coding": 48367, + "coe": 72772, + "coef": 64650, + "coeff": 48638, + "coes": 79850, + "coffee": 78117, + "coh": 39275, + "cohol": 11106, + "coholic": 44394, + "coil": 85737, + "coin": 7160, + "coins": 29609, + "col": 2074, + "cola": 55452, + "cold": 87072, + "cole": 55645, + "coles": 84838, + "coli": 91380, + "coll": 17222, + "collapse": 17908, + "collapsed": 42028, + "collect": 17384, + "collection": 13421, + "collectionView": 42440, + "collections": 51137, + "collector": 91469, + "college": 67434, + "collision": 54567, + "colm": 41837, + "coln": 22033, + "colo": 58248, + "colon": 76824, + "color": 3423, + "colors": 21036, + "colour": 47339, + "cols": 21380, + "column": 6229, + "columnName": 86030, + "columns": 16369, + "com": 874, + "coma": 81845, + "comb": 21873, + "combat": 85643, + "combe": 91057, + "combination": 75028, + "combine": 63136, + "combined": 66307, + "combo": 61492, + "comboBox": 68586, + "come": 2020, + "comed": 29990, + "comes": 6579, + "comfort": 19411, + "comic": 89889, + "coming": 4959, + "comings": 69143, + "comm": 3621, + "comma": 45386, + "command": 5631, + "commands": 24270, + "commend": 30528, + "commended": 35871, + "comment": 6182, + "comments": 14727, + "commerce": 43194, + "commercial": 83607, + "commission": 80064, + "commit": 17413, + "committed": 97446, + "committee": 56785, + "common": 5464, + "commons": 52361, + "communic": 25579, + "communication": 50171, + "communications": 43310, + "community": 28402, + "como": 96296, + "comp": 5689, + "compact": 39760, + "compan": 44249, + "companies": 64751, + "company": 10139, + "companyId": 92706, + "compare": 18948, + "compareTo": 55583, + "comparison": 54705, + "compass": 30829, + "compat": 18331, + "compatible": 34842, + "compet": 27245, + "competition": 78023, + "competitive": 92399, + "compile": 20433, + "compileComponents": 43998, + "compiled": 50845, + "compiler": 33620, + "complete": 14737, + "completed": 34735, + "completion": 43312, + "complex": 23247, + "complexContent": 97715, + "complexType": 79419, + "component": 8571, + "components": 5149, + "compose": 52706, + "composed": 92734, + "composer": 74894, + "composite": 89504, + "composition": 76807, + "compound": 86204, + "compress": 38360, + "compressed": 45703, + "compression": 83192, + "comput": 64947, + "compute": 27706, + "computed": 71011, + "computer": 43111, + "con": 443, + "cona": 90424, + "conc": 40446, + "concat": 20164, + "concept": 68487, + "concert": 88305, + "conciliation": 98240, + "cond": 1297, + "conda": 63753, + "conde": 94859, + "condition": 9056, + "conditional": 45368, + "conditionally": 95967, + "conditions": 16495, + "conds": 7848, + "conduct": 76652, + "cone": 58082, + "conexao": 85680, + "conexion": 43653, + "conf": 6135, + "conference": 78490, + "confidence": 81929, + "config": 1676, + "configs": 53978, + "configuration": 21138, + "configure": 21002, + "configured": 81443, + "confirm": 13800, + "confirmation": 53412, + "confirmed": 41028, + "coni": 85210, + "conj": 92440, + "conn": 5148, + "connect": 6459, + "connected": 15288, + "connecting": 90811, + "connection": 7742, + "connections": 49986, + "connector": 53700, + "connexion": 74878, + "cono": 85858, + "conom": 44217, + "conomic": 31007, + "conomics": 80092, + "conomy": 70071, + "cons": 6254, + "conscious": 63489, + "consider": 24712, + "consin": 18864, + "consistent": 78399, + "console": 5354, + "const": 1024, + "constant": 18077, + "constants": 15763, + "constexpr": 42281, + "constitution": 43003, + "constitutional": 48574, + "constraint": 48057, + "constraints": 65312, + "construct": 7596, + "constructed": 81578, + "construction": 47197, + "constructor": 21846, + "consts": 95773, + "consult": 48412, + "consulta": 42168, + "consum": 66676, + "consume": 84494, + "consumer": 46764, + "cont": 772, + "contact": 6287, + "contacts": 29063, + "contador": 67248, + "contain": 52095, + "contained": 52515, + "container": 3586, + "containers": 39399, + "contains": 13372, + "conte": 45352, + "contenido": 76268, + "content": 1796, + "contentType": 44963, + "contents": 17610, + "contest": 65717, + "context": 2147, + "contexts": 71927, + "conti": 82910, + "contin": 21319, + "continent": 74595, + "continental": 82851, + "continue": 9534, + "continued": 85857, + "continuous": 78589, + "conto": 65244, + "contr": 8222, + "contra": 96743, + "contract": 20257, + "contracts": 63322, + "contrast": 84224, + "contre": 30300, + "contres": 74850, + "contri": 12855, + "contrib": 47110, + "contributors": 84355, + "contro": 43748, + "control": 2865, + "controlled": 58607, + "controller": 7152, + "controllers": 21611, + "controls": 28517, + "conut": 26261, + "conv": 12027, + "conversation": 61637, + "conversion": 73725, + "convert": 14166, + "convertView": 94267, + "converted": 76204, + "converter": 73746, + "cook": 86795, + "cooked": 98697, + "cookie": 16236, + "cookies": 44317, + "cool": 42196, + "cooldown": 90396, + "coon": 72235, + "coop": 86857, + "coord": 26402, + "coordinate": 62526, + "coordinates": 34739, + "coords": 35030, + "cop": 37728, + "cope": 2417, + "copies": 77601, + "copy": 8560, + "copyright": 15492, + "cor": 6005, + "coration": 26452, + "cord": 30989, + "cordova": 76297, + "core": 2153, + "cores": 7701, + "corev": 98645, + "corlib": 67430, + "corn": 38630, + "corner": 73425, + "corp": 80727, + "corr": 59277, + "correct": 19928, + "correo": 66171, + "cors": 53972, + "cos": 9407, + "cosa": 97456, + "cosity": 85118, + "cost": 16440, + "cosystem": 23287, + "cot": 64498, + "cott": 50705, + "cou": 22249, + "could": 28077, + "couldn": 90962, + "count": 1830, + "counter": 8292, + "countries": 46708, + "country": 11141, + "counts": 44859, + "county": 64926, + "coupon": 38300, + "cour": 66871, + "course": 11856, + "courses": 41324, + "court": 62758, + "cout": 6104, + "cov": 56412, + "cover": 3688, + "coverage": 54250, + "covered": 20804, + "covering": 91913, + "covers": 55858, + "covery": 7449, + "cow": 18921, + "cox": 82000, + "cp": 4672, + "cpf": 64553, + "cplusplus": 24926, + "cpp": 10821, + "cps": 72576, + "cpt": 97995, + "cpu": 16475, + "cpy": 7641, + "cq": 95418, + "cr": 5082, + "craft": 7712, + "crafted": 79794, + "cran": 71985, + "crap": 98721, + "craper": 86282, + "crast": 85474, + "cratch": 86284, + "crate": 61711, + "crawl": 94899, + "crawler": 74929, + "crc": 66083, + "cre": 837, + "cream": 46000, + "crear": 98753, + "crease": 19947, + "creasing": 75462, + "creat": 94739, + "create": 3182, + "createClass": 87979, + "createCommand": 97380, + "createElement": 58008, + "createForm": 84658, + "createFrom": 90095, + "createQuery": 95867, + "createQueryBuilder": 88085, + "createTime": 83186, + "createUrl": 97218, + "createView": 75613, + "created": 7120, + "createdAt": 42765, + "creates": 58519, + "creating": 44902, + "creation": 37375, + "creative": 51717, + "creativecommons": 85231, + "creator": 32398, + "creature": 64127, + "cred": 10844, + "credential": 66799, + "credentials": 32353, + "credible": 57941, + "credit": 23311, + "credited": 66209, + "credits": 55769, + "creds": 85734, + "creen": 2191, + "creens": 15539, + "creenshot": 31109, + "crement": 13477, + "crest": 44451, + "cret": 4423, + "crete": 9353, + "cretion": 88690, + "crets": 52710, + "crew": 38617, + "cri": 740, + "cribe": 3114, + "cribed": 17433, + "cribes": 55136, + "cribing": 43531, + "crime": 84354, + "criminal": 98395, + "cripcion": 20667, + "cript": 1228, + "cription": 1453, + "criptions": 24685, + "criptive": 40319, + "criptor": 6820, + "criptors": 25705, + "cripts": 24227, + "crire": 66919, + "crit": 36996, + "criteria": 20519, + "critical": 41541, + "crm": 61738, + "cro": 37884, + "croft": 81693, + "croll": 15241, + "crollView": 41088, + "cron": 69387, + "crop": 34147, + "cross": 28842, + "crow": 51805, + "crud": 53569, + "cry": 61138, + "crypt": 48125, + "crypto": 35772, + "cré": 129532, + "cs": 4837, + "csi": 63229, + "csr": 77851, + "csrf": 24102, + "css": 5143, + "cstdint": 96975, + "cstdio": 37212, + "cstdlib": 42695, + "cstring": 38906, + "csv": 18104, + "ct": 302, + "cta": 77519, + "ctal": 93981, + "cter": 74069, + "ctest": 67880, + "cth": 91821, + "ctic": 25555, + "ctica": 49029, + "ctime": 33600, + "ction": 407, + "ctions": 5136, + "ctl": 12373, + "ctor": 4904, + "ctors": 12532, + "ctp": 77941, + "ctr": 10597, + "ctrine": 13743, + "ctrl": 11933, + "cts": 60903, + "ctx": 3773, + "ctxt": 77492, + "ctype": 17211, + "cu": 20199, + "cuador": 49262, + "cube": 45411, + "cuda": 33926, + "cue": 31724, + "cul": 3314, + "cular": 21366, + "culate": 11207, + "culated": 49804, + "culator": 20022, + "culo": 63387, + "culos": 78034, + "cult": 92013, + "culture": 69805, + "culus": 41349, + "cum": 59253, + "cup": 36665, + "cur": 2352, + "cura": 53535, + "curacy": 21500, + "curities": 30602, + "curity": 3666, + "curl": 19925, + "curr": 15427, + "currency": 15973, + "current": 3231, + "currentColor": 60745, + "currentIndex": 59314, + "currentPage": 46383, + "currentState": 84870, + "currentTime": 66900, + "currentUser": 38880, + "currently": 58202, + "curring": 45739, + "curso": 42201, + "cursor": 17437, + "curve": 50051, + "cury": 25455, + "cus": 56340, + "cust": 44414, + "custom": 9163, + "customer": 11049, + "customerId": 82019, + "customers": 40154, + "customize": 94385, + "cut": 10242, + "cuts": 22497, + "cv": 13122, + "cw": 62543, + "cwd": 32165, + "cx": 25844, + "cxx": 89073, + "cy": 11130, + "cyan": 69746, + "cych": 73046, + "cycl": 97484, + "cycle": 21621, + "cycles": 48421, + "cz": 13769, + "czas": 66637, + "czeÅĦ": 125936, + "cznie": 88494, + "cznÄħ": 133477, + "czy": 47772, + "czyÄĩ": 130912, + "czyÅĤ": 127618, + "czÄħ": 126580, + "czÄħc": 132574, + "czÄĻ": 125846, + "cá": 132438, + "cé": 49984, + "có": 129133, + "ców": 126613, + "cÃŃ": 130497, + "cÃŃa": 129467, + "cı": 126283, + "cılar": 134830, + "cılı": 135295, + "cılık": 135296, + "d": 67, + "dB": 40955, + "dT": 90363, + "da": 3235, + "dac": 82549, + "dad": 55844, + "dados": 42589, + "dae": 89184, + "daemon": 75841, + "daf": 85733, + "dag": 50641, + "daily": 48074, + "dain": 62499, + "dal": 51547, + "dale": 34183, + "dam": 15408, + "damage": 42865, + "damn": 88619, + "dan": 35155, + "dana": 45683, + "dance": 98875, + "danger": 31450, + "dao": 41521, + "dap": 91294, + "daq": 73207, + "dar": 34123, + "dark": 22624, + "dart": 35594, + "darwin": 98765, + "das": 34889, + "dash": 43519, + "dashboard": 18641, + "dat": 5911, + "data": 691, + "dataArray": 81568, + "dataGridView": 76488, + "dataProvider": 35498, + "dataSource": 64860, + "dataTable": 66769, + "dataType": 57694, + "database": 12216, + "datable": 88831, + "datagrid": 84990, + "datal": 82553, + "datap": 93253, + "datas": 27240, + "dataset": 21378, + "datasets": 65546, + "datatable": 80884, + "datatype": 62409, + "date": 1028, + "dateFormat": 90848, + "dateTime": 48006, + "dated": 3577, + "datepicker": 36263, + "dater": 27463, + "dates": 6394, + "datetime": 15450, + "dating": 14110, + "dato": 71105, + "datos": 27615, + "datum": 55933, + "daughter": 59778, + "day": 1292, + "days": 13778, + "daÄĩ": 133107, + "daÄŁ": 136768, + "db": 1999, + "dbContext": 69762, + "dbName": 90881, + "dba": 89455, + "dbc": 14899, + "dbcTemplate": 48972, + "dbe": 83406, + "dbg": 54318, + "dbh": 40800, + "dbl": 72602, + "dbname": 35265, + "dbo": 20852, + "dbuf": 87314, + "dbus": 78986, + "dc": 7628, + "dcc": 86964, + "dd": 631, + "dda": 71487, + "ddb": 89723, + "ddd": 40360, + "dddd": 64100, + "dde": 88089, + "dden": 10792, + "ddf": 96482, + "ddie": 28454, + "ddit": 20090, + "ddl": 78372, + "dds": 33650, + "ddy": 53510, + "de": 450, + "dea": 55088, + "dead": 33754, + "deadline": 78788, + "deal": 29639, + "dealer": 78235, + "dealloc": 75660, + "death": 61907, + "deaux": 61915, + "deb": 37935, + "debian": 90784, + "debit": 89380, + "debug": 8349, + "dec": 8169, + "decay": 84499, + "decess": 26911, + "decimal": 23289, + "decision": 63938, + "deck": 33425, + "decl": 10005, + "declar": 56305, + "declaration": 80383, + "declare": 18471, + "declaring": 64565, + "declspec": 56318, + "decltype": 74364, + "decode": 18196, + "decoded": 62913, + "decoder": 48110, + "decor": 32093, + "decorate": 47976, + "decorators": 94700, + "decrypt": 73295, + "ded": 9789, + "dee": 57894, + "deen": 64481, + "deep": 32880, + "deer": 96262, + "def": 750, + "default": 2258, + "defaultValue": 53102, + "defaults": 26756, + "defense": 96992, + "defer": 62095, + "defgroup": 86029, + "define": 1289, + "defined": 9711, + "defines": 57043, + "definition": 18375, + "definitions": 48563, + "defs": 48485, + "deg": 16508, + "degree": 41027, + "degrees": 69749, + "dehy": 79560, + "dehyde": 85736, + "dek": 76273, + "del": 9588, + "delay": 22263, + "delegate": 28227, + "delete": 4542, + "deleted": 26521, + "delimiter": 58116, + "deliver": 74728, + "delivery": 32390, + "delivr": 62701, + "delta": 20255, + "dem": 27431, + "demand": 82931, + "demo": 25762, + "den": 5183, + "dens": 52221, + "dense": 80286, + "density": 69818, + "dent": 54823, + "deny": 89963, + "deo": 94514, + "dep": 14891, + "depart": 32048, + "department": 27314, + "departments": 91795, + "departure": 58171, + "departureday": 95637, + "depend": 19510, + "dependence": 90646, + "dependencies": 54274, + "dependency": 53690, + "dependent": 37555, + "depending": 80480, + "depends": 57947, + "deploy": 35794, + "deployment": 82213, + "deposit": 60514, + "deprecated": 38189, + "deps": 72587, + "dept": 40585, + "depth": 17561, + "deque": 53498, + "der": 1107, + "dera": 97903, + "derabad": 56700, + "dere": 96393, + "deriv": 82495, + "derive": 27098, + "derived": 50082, + "ders": 12321, + "des": 5799, + "desc": 8614, + "descending": 88032, + "descr": 66515, + "describe": 12332, + "descricao": 63662, + "descripcion": 39238, + "description": 4684, + "descriptor": 53132, + "deserialize": 66777, + "design": 24852, + "designation": 88116, + "desired": 82782, + "desk": 50069, + "desktop": 36869, + "dess": 33007, + "dest": 4979, + "destination": 17997, + "destroy": 15518, + "destruct": 60268, + "det": 19285, + "detach": 56764, + "detail": 14585, + "details": 14871, + "detalle": 84118, + "detect": 61243, + "detector": 91544, + "dev": 3583, + "develop": 15840, + "developer": 34401, + "developers": 64305, + "development": 29571, + "device": 6111, + "deviceId": 94202, + "devices": 46966, + "devil": 80152, + "dex": 91769, + "dez": 86524, + "df": 2940, + "dfa": 97549, + "dfd": 91230, + "dff": 98714, + "dfs": 34378, + "dfunding": 63891, + "dg": 35138, + "dge": 91691, + "dh": 30621, + "dhcp": 96220, + "di": 8579, + "dia": 31006, + "diag": 60214, + "dialog": 11817, + "dialogs": 41454, + "diamond": 88576, + "dic": 29680, + "dice": 76053, + "dict": 8477, + "dictionary": 35671, + "dictions": 52390, + "did": 22920, + "didn": 96749, + "die": 26493, + "diet": 93842, + "dif": 87464, + "diff": 13490, + "difference": 59251, + "different": 63650, + "difficulty": 71864, + "dig": 44861, + "digest": 36339, + "digit": 34934, + "digital": 57269, + "digits": 60770, + "dim": 12927, + "dimension": 42539, + "dimensions": 59239, + "dims": 54490, + "din": 72811, + "ding": 6968, + "dings": 24602, + "dio": 71763, + "dir": 3741, + "dire": 63584, + "direccion": 62689, + "direct": 19798, + "direction": 19071, + "directive": 72863, + "director": 69795, + "directories": 75701, + "directory": 23037, + "dirname": 14434, + "dirs": 57623, + "dirty": 53488, + "dis": 4243, + "disable": 18015, + "disabled": 11978, + "disc": 16822, + "discard": 76492, + "disciplinary": 70100, + "disconnect": 51992, + "discord": 42579, + "discount": 27359, + "discover": 97629, + "discussion": 69869, + "dish": 72833, + "disk": 30496, + "dismiss": 81334, + "disp": 29795, + "dispatch": 18274, + "dispatcher": 60492, + "display": 5493, + "displayName": 46630, + "displayText": 81733, + "dispose": 83865, + "disposed": 75529, + "disposing": 19286, + "dist": 12211, + "distance": 19348, + "distinct": 53021, + "distributed": 62375, + "distribution": 81716, + "district": 35391, + "dit": 17752, + "dition": 2345, + "ditor": 68731, + "div": 611, + "dives": 90514, + "divide": 59394, + "divider": 50794, + "division": 28764, + "diÄŁ": 125284, + "diÄŁi": 126276, + "diÄŁim": 135065, + "diÄŁimiz": 135066, + "diÄŁinde": 139902, + "diÄŁini": 130988, + "diÄŁiniz": 134758, + "dj": 77504, + "django": 13575, + "dk": 7584, + "dl": 8736, + "dle": 90385, + "dlg": 59040, + "dling": 60316, + "dll": 67159, + "dm": 13849, + "dma": 44981, + "dn": 17395, + "dna": 92877, + "dney": 18787, + "dni": 96280, + "dns": 45226, + "do": 2982, + "dob": 51016, + "doc": 5236, + "doch": 72293, + "dock": 77055, + "docker": 28648, + "docs": 14120, + "doctor": 36983, + "doctrine": 91165, + "doctype": 50139, + "document": 6062, + "documentation": 76303, + "documento": 97665, + "documents": 50778, + "does": 27057, + "doesn": 71914, + "dog": 18457, + "dogs": 80034, + "doi": 47786, + "doing": 44259, + "dojo": 72611, + "doll": 77189, + "dom": 5600, + "domain": 12204, + "domains": 59621, + "don": 15007, + "done": 10438, + "dong": 75211, + "donnees": 98553, + "dont": 77025, + "door": 10787, + "doors": 27304, + "dorf": 89999, + "dos": 40525, + "dot": 16119, + "dotenv": 50893, + "dots": 67816, + "double": 4331, + "doubleValue": 82078, + "down": 2923, + "download": 12885, + "downloads": 81329, + "doÄŁ": 127522, + "doÄŁan": 127523, + "dp": 9796, + "dpi": 78029, + "dq": 85646, + "dr": 3612, + "drFc": 70266, + "dra": 22248, + "draft": 44997, + "drag": 33931, + "dragon": 55419, + "dration": 49691, + "draul": 30338, + "draulic": 32742, + "draw": 7633, + "drawable": 91015, + "drawer": 80127, + "drawing": 80849, + "dre": 62840, + "dream": 56191, + "dress": 68234, + "drink": 97327, + "drive": 31967, + "driver": 12521, + "drivers": 62125, + "drm": 98558, + "dro": 11263, + "drop": 6719, + "dropIfExists": 43692, + "dropdown": 11138, + "dropout": 84620, + "drops": 90512, + "drs": 93246, + "drug": 72721, + "drv": 54813, + "dry": 53210, + "ds": 5356, + "dsa": 96780, + "dsl": 81874, + "dsn": 75136, + "dsp": 94724, + "dst": 15658, + "dt": 8047, + "dto": 58978, + "dtype": 47727, + "du": 1054, + "duc": 7545, + "ducation": 33885, + "duce": 13373, + "duced": 53378, + "ducer": 28938, + "ducers": 33375, + "ducible": 78285, + "duck": 72970, + "duct": 1058, + "ducted": 53915, + "duction": 23113, + "ductive": 67143, + "ductor": 35769, + "ductory": 50338, + "due": 23646, + "dued": 78133, + "duino": 31186, + "duit": 46248, + "duk": 56571, + "dummy": 31390, + "dump": 27660, + "dup": 21912, + "duplicate": 63826, + "dur": 70692, + "duration": 17021, + "during": 81877, + "duto": 27943, + "dux": 12138, + "duÄŁ": 140114, + "duÄŁu": 130565, + "duÄŁunu": 140115, + "dv": 37261, + "dw": 29406, + "dx": 12719, + "dy": 10258, + "dyn": 43085, + "dynamic": 21544, + "dz": 37877, + "dzi": 74816, + "dziaÅĤ": 128706, + "dziÄĩ": 131681, + "dziÅĤ": 136300, + "dzÄħ": 132828, + "dão": 133150, + "dé": 128505, + "dì": 134839, + "dü": 126427, + "dür": 126700, + "dÃ¼ÄŁÃ¼": 131803, + "dÃŃ": 133489, + "dÃŃa": 135505, + "dı": 126220, + "dık": 132046, + "dıkl": 127712, + "dıklar": 142409, + "dıkları": 135015, + "dıklarını": 142410, + "dıkt": 140082, + "dıktan": 140083, + "dım": 130412, + "dın": 128099, + "dır": 125020, + "dıģ": 125257, + "dıģı": 128815, + "dıģımız": 137530, + "dıģında": 138571, + "dıģını": 130478, + "dıģınız": 138306, + "dÄĻ": 127209, + "dź": 130682, + "dÅĤug": 136853, + "e": 68, + "ea": 12508, + "each": 9547, + "eacher": 49200, + "ead": 3149, + "eah": 10427, + "eam": 14580, + "ean": 5307, + "ear": 682, + "earable": 76858, + "earch": 2902, + "earchBar": 92336, + "earer": 20786, + "earing": 13678, + "early": 22140, + "earn": 10118, + "earned": 57600, + "earning": 16526, + "ears": 7444, + "earth": 27541, + "ease": 46614, + "east": 60501, + "easy": 45022, + "eat": 32066, + "eated": 67822, + "eating": 90855, + "eature": 15208, + "eatures": 22462, + "eax": 40144, + "eb": 3065, + "eba": 70753, + "ebb": 85302, + "ebile": 137343, + "ebileceÄŁi": 137344, + "ebin": 76157, + "ebp": 65347, + "ebra": 50213, + "ebx": 81363, + "ec": 757, + "eca": 22794, + "ecake": 78256, + "ecal": 97198, + "ecast": 91438, + "ecause": 6631, + "ecc": 71325, + "ecd": 37206, + "ece": 74407, + "eced": 58357, + "ecedor": 65035, + "ecer": 93118, + "ecera": 90227, + "ecess": 4005, + "ecessarily": 73653, + "ecessary": 20122, + "eceÄŁi": 126841, + "eceÄŁini": 136757, + "ech": 4737, + "echa": 17772, + "echo": 3047, + "eci": 75732, + "ecided": 78641, + "ecimal": 42090, + "ecn": 30567, + "eco": 54297, + "ecom": 65174, + "ecome": 51421, + "economic": 48844, + "ecret": 50856, + "ecs": 53717, + "ect": 439, + "ectar": 38816, + "ected": 1569, + "ection": 11915, + "ections": 37605, + "ective": 95878, + "ectl": 52669, + "ectomy": 71683, + "ector": 1256, + "ectors": 10605, + "ecture": 62193, + "ecure": 76108, + "ecurity": 18429, + "ecute": 3068, + "ecx": 76084, + "ecycle": 16696, + "ecz": 68668, + "eczy": 96495, + "ed": 291, + "edBy": 97322, + "edException": 18826, + "edImage": 57472, + "edList": 38301, + "edReader": 17120, + "edTextBox": 97917, + "eda": 13830, + "edad": 75498, + "edar": 33215, + "edata": 72885, + "eday": 45112, + "edb": 93727, + "edback": 19733, + "edd": 53537, + "eddar": 77616, + "ede": 15326, + "eded": 29935, + "edef": 4219, + "edelta": 46854, + "eden": 14134, + "eder": 7294, + "ederal": 5077, + "ederation": 96013, + "edere": 85019, + "ederland": 39355, + "edes": 58526, + "edge": 7186, + "edges": 16900, + "edi": 19048, + "edia": 4495, + "edian": 28567, + "ediator": 87694, + "edic": 33283, + "edicine": 90901, + "edics": 80711, + "edido": 53605, + "edin": 30401, + "eding": 16100, + "edio": 61959, + "edis": 41825, + "edish": 25950, + "edit": 3587, + "editable": 43646, + "editar": 83175, + "editary": 94354, + "edited": 61403, + "edith": 81063, + "editing": 61353, + "edition": 77937, + "editor": 8868, + "edium": 23090, + "edly": 52323, + "edm": 89525, + "edo": 65881, + "edom": 8769, + "edor": 24358, + "edores": 68219, + "edral": 35520, + "edriver": 71039, + "eds": 6767, + "edu": 55101, + "educ": 24770, + "educated": 65349, + "education": 36738, + "educt": 80500, + "edula": 68855, + "edx": 71093, + "edy": 9313, + "ee": 2127, + "eea": 84408, + "eec": 95834, + "eed": 12051, + "eee": 34063, + "eeee": 68616, + "eel": 95451, + "een": 14911, + "eenth": 61555, + "eeper": 43031, + "eer": 34756, + "ees": 5516, + "ef": 823, + "efa": 75631, + "efault": 11301, + "efd": 93678, + "efe": 44953, + "efeller": 80612, + "eff": 6445, + "effect": 26061, + "effective": 77895, + "effects": 80424, + "efficient": 42770, + "efined": 4006, + "efore": 4642, + "efs": 43962, + "eft": 33573, + "eful": 51121, + "efully": 72992, + "eg": 791, + "ega": 11188, + "egade": 95237, + "egal": 61321, + "egan": 15756, + "egas": 14842, + "egasus": 86033, + "egative": 15060, + "egen": 19981, + "egend": 87045, + "eger": 1878, + "egers": 67592, + "egg": 28368, + "eggies": 55560, + "egie": 60057, + "egin": 3013, + "egis": 89967, + "egl": 63502, + "egment": 71080, + "ego": 13243, + "egot": 65978, + "egov": 74187, + "egr": 14412, + "egra": 87575, + "egral": 26821, + "egrate": 57017, + "egrated": 47172, + "egration": 17376, + "egrator": 96435, + "egree": 26776, + "egrity": 67212, + "egt": 50659, + "eguard": 81871, + "eh": 2636, + "ehen": 40697, + "ehicle": 15970, + "ehicles": 41865, + "ehler": 57596, + "ehr": 54988, + "ei": 19972, + "eid": 61878, + "eight": 67532, + "eil": 58791, + "ein": 71620, + "eing": 31775, + "either": 49898, + "eiÃŁ": 131095, + "ej": 35468, + "ejs": 91783, + "ek": 1225, + "eka": 52313, + "eken": 53894, + "eker": 51216, + "eki": 78753, + "eking": 58852, + "eko": 97664, + "eks": 21167, + "ekt": 17149, + "ekte": 66362, + "ekyll": 96809, + "el": 301, + "ela": 8294, + "elage": 80792, + "elah": 74977, + "elaide": 44093, + "elan": 76100, + "eland": 18123, + "elapsed": 53118, + "elas": 39743, + "elastic": 61964, + "elay": 6895, + "elcome": 5717, + "eld": 783, + "elda": 36908, + "elden": 57588, + "elder": 66976, + "eldig": 98411, + "eldo": 83455, + "eldom": 52634, + "eldon": 55043, + "eldorf": 69875, + "ele": 10068, + "elect": 18957, + "election": 61472, + "electric": 63365, + "electron": 74503, + "electronics": 78566, + "eled": 40089, + "elem": 18871, + "element": 6101, + "elementGuidId": 88375, + "elementType": 93287, + "elements": 21423, + "elems": 94822, + "elen": 8398, + "elenium": 14784, + "eler": 7865, + "elerik": 35100, + "elf": 490, + "elfare": 19251, + "elfast": 46025, + "elfth": 87698, + "elho": 85999, + "eli": 12294, + "elia": 35929, + "eliac": 94605, + "elial": 58444, + "elian": 69564, + "elib": 61218, + "elic": 42623, + "elier": 67149, + "elif": 12458, + "elig": 38467, + "elige": 95946, + "elight": 88243, + "eligible": 68050, + "elihood": 19906, + "elijk": 32238, + "elijke": 65046, + "elim": 39795, + "eliminar": 88256, + "elin": 32730, + "eline": 4834, + "elines": 10999, + "eliness": 54698, + "eling": 20058, + "elist": 19230, + "elite": 70291, + "elivery": 72871, + "elix": 67718, + "elize": 19412, + "ell": 613, + "ella": 6842, + "ellan": 84786, + "ellaneous": 40142, + "ellant": 77141, + "ellar": 26880, + "ellas": 74495, + "ellation": 42213, + "elle": 6712, + "elled": 15201, + "elleicht": 66360, + "ellen": 15671, + "eller": 7073, + "ellers": 24777, + "ellery": 49876, + "elles": 36807, + "elli": 20508, + "ellido": 24118, + "ellidos": 50128, + "ellig": 6703, + "elligence": 8509, + "elligent": 20509, + "ellij": 35259, + "elling": 6293, + "ellipse": 68004, + "ellipsis": 71922, + "ellite": 18155, + "ellites": 41660, + "ello": 4791, + "ellow": 5299, + "elloworld": 95292, + "ells": 6436, + "ellschaft": 69701, + "ellt": 32863, + "ellular": 53130, + "ellung": 82017, + "ellungen": 76213, + "elly": 11896, + "elm": 23162, + "elman": 63541, + "elmet": 74772, + "eln": 17454, + "elo": 20172, + "eload": 76897, + "elocity": 19054, + "elog": 50571, + "elon": 64367, + "elong": 38022, + "elope": 81565, + "elor": 69247, + "elow": 49381, + "elp": 1267, + "elper": 2947, + "elpers": 17527, + "elps": 83619, + "elry": 24044, + "els": 2010, + "else": 1503, + "elsea": 23278, + "elseif": 26016, + "elsen": 41145, + "elsey": 92467, + "elsif": 65967, + "elsing": 96588, + "elsinki": 53917, + "elsius": 40247, + "elson": 21395, + "elt": 3818, + "elta": 5964, + "eltas": 71906, + "elter": 17872, + "elters": 46783, + "elts": 84028, + "elu": 78278, + "elve": 21210, + "elves": 4275, + "ely": 974, + "elyn": 91509, + "em": 336, + "ema": 9176, + "emaakt": 93127, + "emachine": 98469, + "emacs": 47935, + "emade": 28432, + "email": 2332, + "emailer": 88822, + "emails": 51376, + "emain": 92558, + "emaker": 58675, + "emale": 11749, + "emales": 86403, + "eman": 15977, + "emand": 20096, + "emann": 79898, + "emap": 42040, + "emark": 21115, + "emarks": 36166, + "emas": 34162, + "emat": 42598, + "ematic": 12240, + "ematics": 33705, + "emax": 92297, + "emb": 8858, + "embali": 68705, + "embed": 12250, + "embedded": 69864, + "embedding": 94611, + "ember": 1377, + "embers": 7062, + "emble": 91629, + "embourg": 55018, + "embr": 59063, + "embrance": 71569, + "embre": 22530, + "embro": 48910, + "embros": 90468, + "emd": 94110, + "eme": 3894, + "emean": 46890, + "emed": 16404, + "emen": 16335, + "emens": 72737, + "ement": 1114, + "emente": 38861, + "ementia": 45644, + "emento": 75284, + "ements": 3723, + "emer": 40896, + "emergency": 98808, + "emes": 15660, + "emet": 34826, + "emetery": 36538, + "emi": 21780, + "emia": 21925, + "emiah": 64845, + "emic": 8111, + "emics": 37205, + "emie": 96146, + "emies": 11041, + "emin": 25684, + "eming": 73814, + "emit": 18712, + "emm": 72419, + "emma": 33793, + "emme": 96592, + "emo": 6726, + "emode": 91546, + "emodel": 95648, + "emoji": 37523, + "emon": 7291, + "emonic": 38801, + "emons": 58296, + "emony": 18867, + "emory": 4731, + "emos": 14946, + "emoth": 95574, + "emotion": 73353, + "emouth": 86041, + "emp": 3262, + "emperature": 93725, + "emphasis": 86192, + "empl": 25892, + "emplace": 93592, + "emplate": 3708, + "emplates": 76793, + "emple": 44434, + "emplo": 30210, + "emploi": 96575, + "employ": 7668, + "employed": 35142, + "employee": 12946, + "employees": 34539, + "employer": 65201, + "employment": 20461, + "empo": 22114, + "empor": 52371, + "emporary": 13766, + "empre": 36839, + "empresa": 35838, + "empt": 20968, + "emption": 33106, + "empty": 3194, + "ems": 11852, + "emsp": 91137, + "emu": 33063, + "emy": 4524, + "emás": 41841, + "en": 268, + "ena": 7157, + "enable": 12552, + "enabled": 15868, + "enade": 86039, + "enaire": 96720, + "enal": 74397, + "ename": 1840, + "enames": 32954, + "enan": 61841, + "enance": 10220, + "enant": 25121, + "enaries": 84275, + "enario": 20413, + "enarios": 60494, + "enary": 46128, + "enas": 26945, + "enate": 31123, + "enberg": 55836, + "enburg": 54385, + "enc": 954, + "ence": 763, + "enced": 5767, + "encent": 26365, + "encer": 19529, + "encers": 62294, + "ences": 2380, + "ench": 19762, + "enchmark": 39381, + "enci": 45257, + "encia": 8792, + "encial": 44312, + "encias": 26799, + "encies": 5946, + "encil": 17005, + "encing": 11373, + "encion": 92773, + "ención": 43631, + "enco": 73385, + "encode": 6180, + "encodeURIComponent": 97683, + "encoded": 19329, + "encoder": 27008, + "encoding": 17159, + "encrypt": 32826, + "encrypted": 36444, + "encryption": 79937, + "encv": 34384, + "ency": 2251, + "end": 408, + "endDate": 45618, + "endTime": 49688, + "enda": 9696, + "endale": 69060, + "endance": 29719, + "endant": 20372, + "endants": 28310, + "endar": 5153, + "endars": 87614, + "endas": 41507, + "endcode": 89691, + "enddate": 98589, + "ende": 10938, + "ended": 2883, + "endedor": 61167, + "enden": 19875, + "endencies": 37331, + "endency": 21213, + "endent": 28193, + "ender": 1659, + "endereco": 91146, + "enderit": 97468, + "enderror": 41742, + "enders": 14506, + "endet": 54081, + "endez": 72502, + "endforeach": 22095, + "endi": 40024, + "endid": 50403, + "endif": 2330, + "endimento": 79859, + "ending": 2459, + "endir": 66473, + "endl": 5336, + "endment": 14646, + "endo": 8691, + "endon": 90986, + "endor": 8029, + "endors": 32885, + "endoza": 95428, + "endphp": 83662, + "endpoint": 32540, + "endra": 60699, + "endregion": 7715, + "ends": 1412, + "endsWith": 84474, + "endtime": 90861, + "endum": 24530, + "ene": 1952, + "ened": 6758, + "enedor": 76121, + "enef": 50465, + "enefit": 88629, + "eneg": 77529, + "enegro": 87821, + "enemy": 27828, + "enen": 33370, + "ener": 798, + "eneral": 18451, + "enerate": 13220, + "enerated": 10543, + "enerating": 93107, + "eneration": 17252, + "enerative": 74889, + "enerator": 15312, + "energy": 17489, + "eneric": 3469, + "enerima": 92682, + "eners": 18223, + "enery": 89233, + "enes": 21649, + "eness": 23709, + "enet": 69336, + "enever": 30745, + "enez": 20556, + "eng": 826, + "enga": 74575, + "engage": 84739, + "engan": 18262, + "enge": 6309, + "engeance": 55379, + "enger": 10009, + "engers": 14886, + "enght": 70078, + "engine": 8512, + "engineering": 97915, + "engkap": 87272, + "engl": 43624, + "english": 29120, + "engo": 64653, + "ength": 996, + "engu": 27004, + "enguin": 45872, + "enguins": 55358, + "enh": 39867, + "enha": 62026, + "enheim": 91610, + "eni": 33129, + "enia": 58286, + "eniable": 78811, + "enic": 55889, + "enido": 32458, + "enie": 61558, + "ening": 6019, + "enis": 38585, + "enity": 55585, + "enk": 78571, + "enkins": 57006, + "enko": 55055, + "enment": 51961, + "enn": 2667, + "enna": 14950, + "ennai": 50924, + "enne": 25136, + "ennen": 40193, + "ennent": 88710, + "ennes": 15209, + "ennessee": 21464, + "ennial": 31231, + "ennie": 98028, + "ennifer": 58970, + "ennis": 16156, + "ennon": 62925, + "enny": 17835, + "eno": 11790, + "enor": 98731, + "enos": 70870, + "enqueue": 61373, + "ens": 724, + "ensa": 57825, + "ensagem": 67859, + "ensaje": 26007, + "ensation": 41625, + "ensburg": 76929, + "ensch": 46745, + "enschaft": 52028, + "ense": 1117, + "ensed": 28940, + "ensely": 60032, + "ensem": 44399, + "ensemble": 48475, + "ensen": 34017, + "enser": 59444, + "enses": 4280, + "ensex": 95536, + "ensi": 55487, + "ensible": 36764, + "ensibly": 69772, + "ensing": 48105, + "ension": 2645, + "ensions": 4664, + "ensis": 86678, + "ensitive": 18723, + "ensitivity": 47837, + "ensity": 7968, + "ensive": 4025, + "ensively": 26916, + "enso": 71689, + "enson": 33137, + "ensor": 3805, + "ensored": 55778, + "ensors": 26529, + "enstein": 61656, + "ensual": 66729, + "ensure": 27289, + "ensus": 13626, + "ent": 306, + "enta": 16576, + "entai": 35130, + "ental": 6296, + "entanyl": 94291, + "entar": 79680, + "entario": 62836, + "entarios": 81546, + "ente": 6817, + "ented": 15864, + "entence": 18380, + "enter": 1950, + "entered": 86043, + "enterprise": 77934, + "entes": 21243, + "enth": 20684, + "enthal": 77175, + "enti": 22396, + "ential": 2283, + "entialAction": 57971, + "entially": 30568, + "entials": 8476, + "entic": 4256, + "enticate": 22661, + "enticated": 17942, + "entication": 7164, + "enticator": 61393, + "enties": 57042, + "entieth": 51131, + "entifier": 7189, + "entiful": 78720, + "entimes": 91355, + "entin": 43409, + "entina": 23909, + "entine": 26872, + "enting": 91199, + "ention": 2939, + "entionPolicy": 78165, + "entions": 63701, + "entious": 97881, + "entities": 10499, + "entity": 2996, + "entityManager": 47112, + "entlich": 44275, + "ently": 4402, + "ento": 17537, + "enton": 73972, + "entr": 23745, + "entrada": 62240, + "entral": 45086, + "entrant": 69558, + "entre": 39897, + "entreprise": 87888, + "entric": 39409, + "entries": 12940, + "entropy": 49150, + "entry": 4085, + "entrée": 134944, + "ents": 805, + "enty": 7166, + "enu": 1481, + "enuine": 63212, + "enuity": 89489, + "enum": 9018, + "enumer": 39501, + "enumerate": 76569, + "enumerator": 96616, + "enums": 92572, + "enuous": 82022, + "enus": 37865, + "env": 3160, + "enville": 67632, + "environment": 23294, + "environments": 78291, + "eny": 32395, + "enz": 16597, + "enza": 23353, + "enze": 84091, + "enzhen": 78411, + "enzie": 56863, + "enzyme": 64628, + "ença": 86680, + "enção": 97298, + "enÃŃ": 96402, + "eo": 24612, + "eof": 77594, + "eon": 65030, + "eor": 23385, + "eos": 84399, + "eous": 33072, + "ep": 747, + "epad": 69541, + "epam": 88929, + "epar": 11584, + "eparator": 91640, + "epend": 12769, + "ependency": 40529, + "eper": 10436, + "eph": 23544, + "ephir": 80673, + "ephy": 96253, + "eping": 54223, + "episode": 39291, + "epoch": 20197, + "epochs": 46529, + "eprom": 52945, + "eps": 7124, + "epsilon": 31767, + "epy": 70548, + "eq": 11006, + "equ": 25310, + "equal": 25795, + "equalTo": 90585, + "equality": 81638, + "equals": 7176, + "equalsIgnoreCase": 34633, + "equip": 68095, + "equipment": 67518, + "equiv": 57317, + "er": 261, + "era": 2416, + "erable": 26959, + "erah": 96208, + "erais": 80256, + "eral": 3253, + "erala": 49981, + "erald": 24380, + "erals": 17368, + "eras": 9247, + "erase": 18688, + "erate": 58668, + "eration": 20927, + "erator": 39522, + "eração": 74744, + "erb": 40499, + "erc": 2962, + "erca": 77893, + "erce": 25641, + "erchant": 59943, + "ercial": 7991, + "ercicio": 47000, + "ercise": 18359, + "ercises": 47715, + "ercul": 68193, + "erculosis": 73454, + "erd": 14348, + "erdale": 72320, + "erde": 86805, + "erdem": 87783, + "erdings": 68627, + "ere": 485, + "erea": 67109, + "ereal": 43253, + "ereco": 33994, + "ered": 12336, + "eree": 74960, + "eref": 43970, + "ereg": 52633, + "erek": 27626, + "eren": 17392, + "ereo": 64853, + "ereotype": 15742, + "erequisite": 90375, + "erequisites": 82301, + "erer": 90017, + "eres": 12917, + "eresa": 37586, + "ereum": 23923, + "erez": 33786, + "ereço": 77956, + "erg": 2375, + "ergarten": 47572, + "erge": 10080, + "erged": 51525, + "ergency": 65301, + "ergic": 74339, + "erging": 95296, + "erglass": 75469, + "ergus": 32931, + "erguson": 34908, + "ergy": 41943, + "eri": 30703, + "eria": 49829, + "erial": 2848, + "erialization": 50563, + "erialize": 9050, + "erialized": 67577, + "erializer": 41939, + "eric": 11632, + "erie": 26105, + "eries": 4699, + "erin": 84409, + "ering": 4671, + "erior": 14575, + "erk": 16754, + "erland": 86954, + "erm": 4195, + "ermal": 18067, + "ermalink": 42699, + "erman": 4753, + "ermann": 91450, + "ermen": 59158, + "ermint": 93832, + "ermo": 40716, + "ern": 932, + "erna": 46528, + "ernal": 2978, + "ernals": 89320, + "ernational": 27135, + "ernaut": 91189, + "erne": 67395, + "ernel": 5454, + "ernels": 42329, + "ernen": 49072, + "erner": 41372, + "ernes": 73625, + "erness": 28568, + "ernet": 13845, + "ernetes": 29823, + "ernity": 33619, + "erno": 24935, + "ernote": 87450, + "ero": 2328, + "erokee": 59439, + "eron": 19540, + "eroon": 79732, + "eros": 6264, + "erosis": 57781, + "erot": 88992, + "erox": 98158, + "erp": 22632, + "err": 615, + "errMsg": 98871, + "erra": 13886, + "erral": 28178, + "errals": 55143, + "errar": 50049, + "erras": 84384, + "errat": 59609, + "errated": 83656, + "erre": 66000, + "errer": 30929, + "erreur": 70556, + "errick": 60400, + "erring": 14443, + "errmsg": 69932, + "errno": 26379, + "erro": 40296, + "error": 841, + "errorCode": 60329, + "errorMessage": 43169, + "errorMsg": 76044, + "errors": 7650, + "errs": 84735, + "errupt": 10579, + "errupted": 80853, + "erry": 5400, + "ers": 388, + "ersed": 39919, + "ersen": 87989, + "erset": 44146, + "ersh": 54990, + "ersion": 1325, + "ersions": 35279, + "ersist": 4975, + "ersistence": 8026, + "ersistent": 13931, + "ersive": 52353, + "erson": 1270, + "ersonic": 94557, + "erspective": 85091, + "ert": 529, + "erta": 22637, + "ertain": 7615, + "ertainment": 11205, + "ertainty": 80146, + "ertas": 52430, + "ertation": 75304, + "erte": 13820, + "erten": 50958, + "ertest": 83386, + "ertext": 37220, + "erti": 34113, + "ertia": 40122, + "ertiary": 66530, + "erties": 2707, + "ertificate": 20962, + "ertil": 85695, + "ertility": 61487, + "ertime": 69741, + "ertino": 90600, + "erto": 13878, + "ertoire": 61604, + "erton": 36633, + "ertools": 36316, + "ertos": 67119, + "erts": 15546, + "ertura": 55870, + "erture": 41847, + "erty": 1344, + "ertype": 45069, + "ertz": 57937, + "erule": 79393, + "erus": 45009, + "erusform": 69216, + "erv": 648, + "erva": 45865, + "erval": 6152, + "ervals": 42198, + "ervas": 95064, + "ervation": 8768, + "ervations": 45821, + "ervative": 53168, + "ervatives": 78529, + "erve": 5852, + "erved": 2771, + "ervention": 46598, + "erver": 2836, + "ervers": 18729, + "erves": 13280, + "ervice": 1017, + "ervices": 2161, + "erview": 12873, + "erville": 50772, + "erving": 19505, + "ervised": 77990, + "ervisor": 31396, + "ervlet": 4560, + "ervo": 77044, + "ervoir": 75850, + "ery": 722, + "eryl": 45025, + "erÃł": 132617, + "erÃŃa": 131397, + "es": 288, + "esModule": 30012, + "esa": 22371, + "esan": 62424, + "esar": 32240, + "esc": 3300, + "escal": 81082, + "escape": 12998, + "escaped": 65826, + "escaping": 42480, + "esch": 59378, + "esco": 93834, + "escort": 78652, + "escription": 2037, + "esda": 68626, + "ese": 2367, + "esehen": 81184, + "eselect": 51630, + "esen": 65935, + "eses": 83736, + "esh": 4288, + "eshire": 89245, + "esi": 33083, + "esian": 29221, + "esign": 13985, + "esimal": 29472, + "esion": 55981, + "esis": 13774, + "esity": 27123, + "esium": 40830, + "esk": 66819, + "esktop": 62204, + "eslint": 85502, + "eso": 28830, + "esome": 7057, + "eson": 60917, + "esor": 47835, + "esp": 24610, + "especially": 34946, + "esper": 69038, + "espère": 143899, + "ess": 433, + "essa": 25496, + "essage": 801, + "essages": 3737, + "essaging": 32140, + "essay": 79977, + "esse": 23318, + "essed": 21712, + "essel": 35548, + "essen": 38809, + "essenger": 40137, + "essential": 48482, + "esser": 36370, + "esseract": 82538, + "esses": 64870, + "essian": 74329, + "essim": 66733, + "ession": 1338, + "essional": 15774, + "essions": 8551, + "essler": 76949, + "essment": 23391, + "esso": 9823, + "essoa": 27587, + "esson": 71405, + "essor": 8309, + "est": 477, + "esta": 29179, + "establish": 33400, + "estado": 22016, + "estar": 89327, + "estate": 64115, + "estation": 63978, + "estatus": 98525, + "este": 17750, + "estead": 72337, + "ested": 9980, + "esteem": 63617, + "ester": 5191, + "esterday": 11282, + "esters": 75702, + "esthes": 53020, + "esthesia": 62523, + "esthetic": 70009, + "esti": 53483, + "estic": 9826, + "estimate": 40130, + "estimated": 44229, + "estinal": 63950, + "estination": 20646, + "esting": 59855, + "estion": 42498, + "estival": 10863, + "esto": 51425, + "estone": 97933, + "estr": 15111, + "estre": 72775, + "estring": 65040, + "estro": 54556, + "estroy": 9478, + "estruct": 29294, + "estruction": 68702, + "estructor": 42922, + "ests": 17966, + "esture": 15992, + "estureRecognizer": 81920, + "esty": 40239, + "estyle": 37383, + "esub": 71636, + "esus": 42393, + "esy": 64641, + "esz": 74036, + "et": 295, + "etAddress": 53862, + "etCode": 89260, + "etSocketAddress": 80507, + "eta": 1915, + "etable": 35064, + "etadata": 7603, + "etag": 80465, + "etail": 1628, + "etailed": 10111, + "etails": 2237, + "etak": 88721, + "etal": 21323, + "etary": 15859, + "etas": 35917, + "etat": 96615, + "etc": 12107, + "etch": 2995, + "etched": 36780, + "etchup": 90897, + "etcode": 83145, + "ete": 7714, + "etect": 13176, + "etection": 22280, + "etections": 89183, + "eted": 24481, + "eten": 33748, + "eteor": 59466, + "eter": 1404, + "eterangan": 53311, + "eteria": 77822, + "etermin": 16483, + "eterminate": 64224, + "etermination": 81037, + "etermine": 24308, + "etermined": 46637, + "eters": 2424, + "etes": 95181, + "etest": 57824, + "eth": 769, + "ethe": 18522, + "etheless": 26071, + "ether": 2723, + "ethereum": 67122, + "etherlands": 23612, + "etheus": 39705, + "ethical": 74842, + "ething": 2298, + "ethnic": 80418, + "ethod": 1068, + "ethoven": 92522, + "ethyl": 41872, + "ethylene": 87540, + "ethyst": 87303, + "eti": 50898, + "etic": 5298, + "etical": 44508, + "etically": 36674, + "etics": 24279, + "eties": 25559, + "etime": 4107, + "etimes": 6889, + "etine": 85228, + "eting": 11619, + "eto": 11023, + "eton": 34694, + "etr": 17366, + "etrain": 95416, + "etration": 73854, + "etre": 46887, + "etri": 93123, + "etric": 16340, + "etrics": 13468, + "etrize": 56953, + "etro": 8804, + "etrofit": 70308, + "etroit": 18328, + "etros": 34466, + "etry": 15149, + "ets": 1415, + "etsk": 85774, + "etsy": 62627, + "ett": 7066, + "etta": 26527, + "ette": 6535, + "ettel": 84793, + "etten": 94101, + "etter": 21405, + "ettes": 45903, + "etti": 27937, + "etting": 51089, + "ettings": 81183, + "ettle": 47626, + "etto": 52879, + "etty": 88798, + "etu": 94417, + "etur": 31841, + "eturn": 2489, + "etus": 63376, + "etween": 10053, + "etwork": 2349, + "etxt": 88900, + "ety": 2611, + "etyl": 86248, + "etype": 16214, + "etypes": 67553, + "etz": 42189, + "etzt": 61007, + "etÃł": 94871, + "eu": 20128, + "euillez": 87925, + "eur": 12559, + "europ": 143830, + "europä": 143831, + "europäische": 143832, + "eurs": 37547, + "eus": 89702, + "eut": 60053, + "ev": 5120, + "eva": 73418, + "eval": 14170, + "evaluate": 47291, + "evaluation": 74014, + "eve": 82048, + "even": 16788, + "evenodd": 83627, + "event": 3087, + "eventId": 90356, + "eventName": 85827, + "eventType": 94066, + "evento": 83954, + "events": 12389, + "ever": 2054, + "every": 29015, + "everyone": 91957, + "everything": 72726, + "evice": 63924, + "evil": 58867, + "evin": 14281, + "evity": 51679, + "evt": 28734, + "ew": 365, + "ewan": 52298, + "eward": 24035, + "eware": 54069, + "ewart": 26497, + "ewater": 47444, + "eway": 24723, + "eways": 69271, + "ewe": 87723, + "ewear": 98128, + "ewed": 92076, + "ewhat": 80475, + "ewhere": 60652, + "ewidth": 28357, + "ewing": 78410, + "ewire": 70545, + "ewis": 82152, + "ewise": 24350, + "ewish": 65829, + "ewith": 97809, + "ewitness": 70896, + "ewn": 94450, + "ewnÄĻtr": 128160, + "ewnÄĻtrzn": 128161, + "ewolf": 76699, + "ewood": 75265, + "ework": 2445, + "eworld": 95772, + "eworthy": 56138, + "ewriter": 72480, + "ews": 27744, + "ex": 327, + "exact": 46385, + "exam": 41616, + "example": 8687, + "exampleInput": 57692, + "exampleInputEmail": 82001, + "exampleModal": 58390, + "exampleModalLabel": 82403, + "examples": 51668, + "exao": 39539, + "exas": 7467, + "exc": 39441, + "excel": 40524, + "except": 11683, + "exception": 7803, + "exceptions": 61752, + "excerpt": 71570, + "exchange": 39568, + "exclude": 32413, + "excluding": 78995, + "exclusive": 89122, + "exe": 46184, + "exec": 11748, + "executable": 95924, + "execute": 10257, + "execution": 52721, + "executor": 80787, + "exels": 94727, + "exemple": 70734, + "exercise": 63462, + "exion": 16132, + "exist": 28575, + "existence": 92672, + "existent": 64085, + "existing": 36895, + "exists": 16304, + "exit": 13652, + "exo": 77209, + "exp": 4580, + "expand": 32317, + "expanded": 35483, + "expect": 17119, + "expectException": 87714, + "expected": 7325, + "expects": 37631, + "expense": 58757, + "expenses": 92210, + "experience": 49559, + "experiment": 59429, + "experimental": 86703, + "expert": 73887, + "expiration": 63643, + "expire": 52161, + "expired": 75532, + "expires": 48203, + "expiry": 75717, + "expl": 29892, + "explain": 94344, + "explained": 80850, + "explicit": 93632, + "explo": 68231, + "explode": 54842, + "expo": 71131, + "export": 1533, + "exports": 4986, + "expr": 9413, + "express": 13788, + "expression": 28099, + "expérience": 132808, + "ext": 427, + "extAlignment": 58749, + "extField": 25082, + "extView": 64981, + "extend": 25048, + "extended": 41098, + "extends": 15231, + "extension": 12427, + "extensions": 27609, + "extent": 70172, + "exter": 37565, + "exterity": 58816, + "extern": 4301, + "external": 20921, + "externalActionCode": 90098, + "extr": 38649, + "extra": 15460, + "extracomment": 75513, + "extract": 23493, + "extras": 72575, + "extrême": 144003, + "extérieur": 138186, + "exual": 45179, + "exus": 23666, + "ey": 1195, + "eyJ": 84609, + "eya": 77889, + "eye": 18958, + "eyed": 47959, + "eyer": 34372, + "eyes": 57231, + "eyond": 22650, + "ez": 10125, + "ezier": 98616, + "eña": 131080, + "eño": 131011, + "eÄŁ": 127129, + "eÄŁe": 138375, + "eÄŁi": 129230, + "eÄŁin": 131490, + "eÅĦ": 129299, + "f": 69, + "fa": 3632, + "fab": 36855, + "fabric": 85154, + "fabs": 50147, + "fac": 22185, + "face": 1564, + "facebook": 20944, + "faces": 7605, + "facet": 83523, + "facility": 78683, + "fact": 33110, + "faction": 90309, + "factor": 37591, + "factory": 21618, + "facts": 68053, + "faculty": 85168, + "fad": 83059, + "fade": 17517, + "fadeIn": 34993, + "fadeOut": 39619, + "faf": 76476, + "fahren": 96922, + "fail": 18403, + "failed": 16091, + "fails": 59631, + "failure": 28939, + "fair": 40900, + "fait": 63865, + "faith": 74610, + "fake": 30570, + "faker": 36419, + "fal": 95588, + "fall": 13464, + "fallback": 73311, + "falls": 32595, + "false": 3849, + "family": 18984, + "fan": 30178, + "fang": 42477, + "fans": 85877, + "fant": 60727, + "faq": 45523, + "far": 23559, + "fare": 23051, + "farm": 65811, + "fas": 14847, + "fast": 9349, + "fasta": 94466, + "fastcall": 42403, + "fat": 44183, + "fatal": 74394, + "father": 23017, + "fault": 59572, + "fav": 80864, + "favicon": 54557, + "favor": 73960, + "favorite": 38490, + "favorites": 78503, + "fax": 39053, + "fb": 10798, + "fbe": 93264, + "fc": 8316, + "fcc": 72975, + "fce": 97614, + "fclose": 78054, + "fcn": 92041, + "fcntl": 49676, + "fd": 6902, + "fdb": 94228, + "fdc": 92168, + "fdf": 80583, + "fds": 64834, + "fe": 1859, + "fea": 94214, + "feas": 89277, + "feat": 44644, + "feature": 12753, + "featured": 36215, + "features": 20304, + "feb": 77371, + "fec": 76526, + "fecha": 19808, + "fect": 3751, + "fection": 13120, + "fections": 24384, + "fed": 51123, + "fee": 30017, + "feed": 11184, + "feedback": 20999, + "feeding": 45419, + "feeds": 64542, + "feel": 56576, + "fef": 97804, + "feit": 61121, + "feito": 89118, + "fel": 58267, + "feld": 30102, + "fell": 66543, + "felt": 65822, + "female": 42634, + "fen": 30353, + "feof": 85670, + "fer": 802, + "ference": 2202, + "ferences": 4901, + "fern": 78733, + "ferred": 5554, + "fers": 49793, + "fest": 52764, + "fetch": 9641, + "fetchAll": 39015, + "few": 70730, + "ff": 542, + "ffa": 71942, + "ffb": 88549, + "ffc": 77887, + "ffd": 72554, + "ffe": 27969, + "ffect": 1705, + "ffective": 39684, + "ffects": 39927, + "ffee": 53125, + "ffen": 25651, + "ffer": 2945, + "fff": 14092, + "ffff": 7238, + "ffffff": 26886, + "fffffff": 82692, + "ffffffff": 53697, + "ffi": 53799, + "ffic": 20615, + "ffield": 53239, + "ffiti": 55326, + "ffmpeg": 72422, + "ffset": 55340, + "fft": 29175, + "fg": 4817, + "fgang": 75313, + "fgets": 98993, + "fh": 31104, + "fi": 9983, + "fib": 75326, + "fiber": 82945, + "fic": 58628, + "fiction": 57062, + "fid": 32439, + "fidf": 80954, + "field": 2566, + "fieldName": 44594, + "fieldname": 86098, + "fields": 9007, + "fieldset": 23566, + "fifo": 74031, + "fig": 904, + "figcaption": 79212, + "fight": 21143, + "fighter": 72331, + "fighters": 78061, + "figure": 17781, + "figures": 98027, + "fil": 12441, + "fila": 46136, + "file": 1192, + "fileName": 20464, + "filePath": 36697, + "filename": 8404, + "filepath": 44157, + "filer": 72555, + "files": 7198, + "filesize": 91378, + "filesystem": 41897, + "fill": 7559, + "fillType": 54329, + "fillable": 41386, + "filled": 27511, + "fillna": 95096, + "fills": 66167, + "film": 30155, + "filme": 78335, + "films": 65402, + "filt": 87631, + "filter": 5315, + "filtered": 41131, + "filters": 24608, + "filtr": 84746, + "filtro": 93749, + "fin": 5472, + "final": 11822, + "finalize": 94405, + "finally": 39176, + "finance": 62675, + "financial": 74987, + "find": 3903, + "findAll": 51389, + "findBy": 61236, + "findById": 94009, + "findFirst": 98042, + "findOne": 44381, + "findOrFail": 54290, + "findViewById": 25782, + "finder": 33339, + "finding": 67187, + "fine": 62057, + "finger": 79116, + "finish": 30150, + "finished": 12129, + "finite": 25722, + "finity": 19888, + "fir": 91636, + "fire": 10796, + "firebase": 29310, + "firefox": 97912, + "fires": 55071, + "firm": 8802, + "first": 3896, + "firstName": 27987, + "firstname": 28425, + "fish": 18170, + "fit": 6276, + "fitness": 73521, + "fits": 29222, + "five": 52670, + "fix": 5743, + "fixed": 22021, + "fixture": 59612, + "fixtures": 45247, + "fk": 41718, + "fl": 1489, + "flag": 9903, + "flags": 11161, + "flake": 63456, + "flamm": 44764, + "flammatory": 53997, + "flare": 31605, + "flash": 19268, + "flashdata": 71934, + "flat": 26229, + "flatMap": 57591, + "flate": 16716, + "flater": 11729, + "flation": 64149, + "flatten": 47066, + "fld": 48311, + "fleet": 72698, + "flen": 58467, + "flex": 12018, + "flies": 53919, + "flight": 38390, + "flip": 39017, + "flix": 17528, + "flo": 89806, + "float": 3649, + "floating": 82510, + "floor": 30449, + "flow": 4965, + "flower": 38753, + "flowers": 88670, + "flows": 38140, + "flt": 79209, + "flu": 26164, + "fluence": 40016, + "fluid": 66025, + "flush": 21664, + "flutter": 14972, + "flux": 36706, + "fly": 21642, + "fläche": 135266, + "fm": 21112, + "fmt": 12501, + "fn": 8822, + "fname": 23993, + "fo": 824, + "focus": 17414, + "focused": 62157, + "foil": 98957, + "fois": 85533, + "fol": 8400, + "fold": 19961, + "folder": 17668, + "folders": 67355, + "folio": 10988, + "folios": 57998, + "folk": 40931, + "follow": 18559, + "follower": 93742, + "followers": 65866, + "following": 42918, + "fon": 21470, + "fone": 31341, + "fono": 21812, + "font": 4026, + "fontName": 77101, + "fontSize": 35095, + "fontWeight": 61881, + "fonts": 20203, + "fontsize": 45801, + "fony": 13172, + "foo": 7975, + "foobar": 50267, + "food": 13915, + "foods": 95320, + "foon": 94776, + "foot": 5334, + "football": 84336, + "footer": 6956, + "for": 1958, + "forEach": 17969, + "forall": 33629, + "forc": 10458, + "force": 8833, + "forced": 24246, + "forcement": 10927, + "forcements": 87393, + "forcer": 82010, + "forces": 48796, + "forcing": 41496, + "ford": 8187, + "forder": 87227, + "fore": 1324, + "foreach": 8808, + "forecast": 58984, + "foreground": 34919, + "foregroundColor": 81144, + "foreign": 28443, + "forest": 50655, + "forg": 96844, + "forge": 24112, + "forget": 40019, + "forgettable": 56446, + "forgot": 67803, + "fork": 44738, + "form": 627, + "formData": 52873, + "forma": 47571, + "formance": 11795, + "format": 2243, + "formation": 1627, + "formats": 63482, + "formatted": 49878, + "formatter": 32941, + "forme": 75601, + "formed": 10155, + "formedURLException": 64771, + "former": 34527, + "formerly": 69514, + "formik": 76607, + "forming": 54757, + "forms": 9807, + "formula": 59499, + "formulario": 93493, + "fort": 3969, + "fortawesome": 56663, + "forth": 70527, + "fortunate": 32074, + "fortunately": 11029, + "fortune": 83512, + "forum": 22585, + "forums": 54059, + "forward": 13435, + "fos": 50688, + "foto": 42138, + "found": 15105, + "foundation": 72307, + "foundland": 76633, + "four": 34024, + "fout": 96805, + "fov": 85582, + "fox": 15011, + "foy": 93717, + "fp": 10852, + "fprintf": 31617, + "fps": 43428, + "fq": 63919, + "fr": 1626, + "fra": 42289, + "frac": 37018, + "fraction": 81090, + "frag": 33198, + "frage": 83301, + "fragistics": 53625, + "fragment": 42202, + "fram": 91504, + "frame": 6763, + "frames": 23719, + "framework": 3794, + "frared": 81382, + "frau": 96115, + "frauen": 98058, + "fre": 24716, + "fred": 27993, + "free": 10593, + "freeze": 51638, + "frei": 97369, + "freq": 33540, + "frequency": 46521, + "fresh": 71308, + "frey": 75231, + "frica": 29880, + "frican": 47793, + "fried": 97525, + "friend": 10701, + "friendly": 81530, + "friends": 29462, + "frm": 31636, + "frog": 85982, + "from": 1499, + "fromJson": 97756, + "front": 6951, + "frontend": 28181, + "fruit": 35598, + "fs": 3848, + "fsp": 92113, + "fst": 49494, + "fstream": 30516, + "ft": 723, + "ftar": 46201, + "fte": 63319, + "ften": 25061, + "fter": 1046, + "ftime": 17789, + "ftp": 25068, + "fts": 44444, + "ftware": 2110, + "fty": 36958, + "ftype": 80390, + "fu": 32621, + "fuck": 70474, + "fuel": 63375, + "ful": 1262, + "fulWidget": 35385, + "full": 8878, + "fullName": 67158, + "fullname": 44846, + "fullscreen": 52449, + "fully": 3641, + "fulness": 30414, + "fun": 11894, + "func": 2830, + "funcs": 87482, + "function": 1688, + "functional": 49228, + "functions": 21409, + "fund": 57603, + "fung": 78606, + "fur": 82901, + "furt": 38186, + "fusc": 51589, + "fuscated": 80247, + "fuse": 95592, + "fusion": 54565, + "future": 21055, + "fv": 27890, + "fw": 20564, + "fwrite": 96727, + "fx": 8298, + "fy": 30595, + "fä": 131660, + "fäh": 136507, + "fähig": 136508, + "fäll": 143155, + "fälle": 142845, + "fällig": 143156, + "fällt": 135475, + "fé": 58858, + "férence": 135387, + "för": 94961, + "förder": 138154, + "fü": 30166, + "füg": 60409, + "füh": 68545, + "fühl": 133666, + "führ": 72313, + "führen": 131527, + "führer": 142884, + "führt": 96373, + "führung": 132975, + "für": 133238, + "fÃŃcio": 137548, + "g": 70, + "gL": 80130, + "gMaps": 83804, + "ga": 6743, + "gaard": 94489, + "gabe": 61229, + "gaben": 83543, + "gable": 35034, + "gado": 77640, + "gae": 66278, + "gage": 14943, + "gages": 55044, + "gain": 59146, + "gal": 16470, + "galement": 42210, + "gallery": 26205, + "gam": 39329, + "gambar": 84069, + "game": 5804, + "gameObject": 42763, + "games": 34568, + "gamma": 32214, + "gan": 29428, + "gang": 46014, + "gap": 41410, + "gar": 12164, + "gard": 53492, + "gars": 63402, + "gart": 45246, + "gary": 33488, + "gas": 39525, + "gate": 24601, + "gateway": 46473, + "gather": 95725, + "gatsby": 49148, + "gay": 74493, + "gb": 9511, + "gba": 56380, + "gboolean": 86962, + "gc": 20669, + "gcc": 70983, + "gcd": 91289, + "gd": 28584, + "ge": 709, + "gear": 54727, + "geb": 42153, + "geben": 42316, + "geber": 95905, + "gebn": 66198, + "gebra": 26083, + "gebung": 92135, + "ged": 3556, + "gee": 98394, + "gef": 140600, + "gefähr": 140601, + "geführt": 133429, + "geh": 94873, + "geist": 78400, + "gel": 28852, + "gem": 34103, + "geme": 64649, + "gement": 29564, + "gements": 83027, + "gems": 99053, + "gemÃ¤ÃŁ": 142334, + "gen": 4370, + "gence": 28255, + "gency": 21640, + "gend": 57064, + "gende": 58773, + "genden": 85953, + "gender": 12968, + "gene": 42371, + "gener": 7490, + "general": 24595, + "generate": 19366, + "generated": 16187, + "generation": 80057, + "generator": 35851, + "generic": 35787, + "genes": 58175, + "genesis": 77894, + "genic": 88205, + "genome": 87945, + "genre": 33613, + "genres": 63911, + "gens": 58305, + "gent": 15772, + "geo": 13052, + "geois": 46309, + "geom": 41916, + "geometry": 15281, + "ger": 1389, + "gerald": 56797, + "gere": 87027, + "geries": 60388, + "gers": 10637, + "gerät": 137143, + "ges": 4188, + "gesch": 45061, + "gest": 6969, + "gesture": 85357, + "get": 455, + "getActiveSheet": 49732, + "getAll": 45071, + "getApplication": 91869, + "getAs": 91346, + "getAttribute": 25635, + "getBlock": 97497, + "getBody": 58720, + "getBytes": 56809, + "getC": 59515, + "getCell": 89514, + "getChild": 68350, + "getClass": 42660, + "getClient": 53893, + "getClientOriginal": 74885, + "getCode": 57934, + "getColor": 47660, + "getColumn": 53909, + "getConfig": 52536, + "getConnection": 52414, + "getContent": 46679, + "getContext": 49233, + "getCurrent": 44454, + "getData": 25940, + "getDate": 74280, + "getDb": 95158, + "getDefault": 75987, + "getDescription": 94813, + "getDisplay": 99033, + "getDoctrine": 50018, + "getDrawable": 88498, + "getElement": 75657, + "getElementsByTagName": 70620, + "getEmail": 89512, + "getError": 69681, + "getExtension": 57860, + "getField": 40430, + "getFile": 76595, + "getFullYear": 96785, + "getHeight": 66773, + "getID": 66031, + "getId": 15612, + "getImage": 63308, + "getIndex": 95934, + "getInstance": 19720, + "getInt": 48880, + "getItem": 51753, + "getJSON": 79031, + "getKey": 63147, + "getLast": 71557, + "getList": 74689, + "getLocale": 88582, + "getLocation": 87972, + "getManager": 56884, + "getMessage": 20383, + "getMethod": 90161, + "getMock": 46228, + "getMockBuilder": 91424, + "getModel": 49430, + "getName": 13875, + "getNext": 79202, + "getNode": 69984, + "getNum": 83044, + "getObject": 61949, + "getOption": 69023, + "getOrElse": 91702, + "getPage": 96472, + "getParam": 44775, + "getParameter": 62649, + "getParent": 59169, + "getPath": 50174, + "getPlayer": 94972, + "getPosition": 50408, + "getPost": 64997, + "getProperty": 58636, + "getQuery": 57098, + "getReference": 72265, + "getRepository": 41332, + "getRequest": 39397, + "getResource": 34560, + "getResponse": 59341, + "getResult": 56656, + "getRoot": 93190, + "getRow": 87476, + "getService": 71510, + "getSession": 60849, + "getSimpleName": 73317, + "getSingleton": 79166, + "getSize": 59353, + "getSource": 87804, + "getState": 62353, + "getStatus": 68873, + "getStatusCode": 76537, + "getStore": 80579, + "getString": 20769, + "getStringExtra": 86681, + "getStyle": 74743, + "getTable": 59193, + "getText": 31298, + "getTime": 25278, + "getTitle": 66305, + "getToken": 84731, + "getType": 28549, + "getUrl": 66270, + "getUser": 28440, + "getValue": 18499, + "getVar": 71049, + "getView": 86449, + "getWidth": 65731, + "getWindow": 92594, + "getX": 51439, + "getY": 52568, + "getattr": 81732, + "getc": 71303, + "gether": 3439, + "getic": 35224, + "getitem": 59912, + "getline": 56692, + "gets": 18691, + "gett": 76445, + "getter": 52891, + "gettext": 57259, + "getti": 96372, + "getting": 50110, + "getto": 63622, + "gew": 37285, + "gewater": 85030, + "gf": 45124, + "gfx": 55111, + "gg": 14398, + "gger": 91776, + "gh": 866, + "ghan": 15708, + "ghest": 57242, + "ghi": 75076, + "ghost": 38654, + "ght": 2076, + "gi": 8212, + "gia": 54801, + "giatan": 78209, + "gid": 34849, + "gie": 21512, + "gien": 87264, + "giene": 42212, + "gies": 67369, + "gif": 33186, + "gift": 52330, + "gil": 60787, + "gin": 8163, + "ginas": 78907, + "ging": 3173, + "gings": 52833, + "gio": 45145, + "girl": 28552, + "girls": 60196, + "gis": 51098, + "gist": 95294, + "git": 12882, + "github": 5204, + "gium": 89239, + "give": 46430, + "given": 41968, + "gl": 6072, + "glas": 26383, + "glass": 58436, + "gle": 3408, + "gles": 85609, + "glfw": 90445, + "gli": 45288, + "glich": 28494, + "glm": 42219, + "glob": 58926, + "global": 9752, + "globals": 53167, + "glomer": 75046, + "gly": 22945, + "glyph": 65481, + "glyphicon": 22135, + "glÄħd": 131772, + "gm": 26186, + "gmail": 76209, + "gment": 9070, + "gments": 26139, + "gmt": 46063, + "gn": 4905, + "gne": 67155, + "gni": 46242, + "gnore": 10289, + "gnu": 68828, + "go": 3346, + "goal": 33939, + "goals": 84157, + "god": 39711, + "going": 9546, + "gold": 34537, + "golden": 97235, + "gom": 36083, + "gomery": 38230, + "gon": 11670, + "gone": 45596, + "gons": 43874, + "goo": 62048, + "good": 18536, + "goods": 19693, + "goog": 39858, + "google": 17485, + "gor": 5628, + "gorit": 86415, + "gorith": 6063, + "gorithm": 7095, + "gorithms": 18980, + "gorm": 73281, + "gos": 34073, + "got": 22390, + "gota": 70153, + "goto": 28535, + "gov": 53203, + "government": 86763, + "gow": 36186, + "gp": 21888, + "gpio": 44877, + "gps": 74511, + "gpu": 42594, + "gr": 901, + "gra": 31950, + "grab": 58212, + "grad": 6759, + "gradable": 77030, + "gradation": 48141, + "grade": 6937, + "graded": 23343, + "grades": 22340, + "gradient": 26394, + "grading": 32259, + "graduate": 27670, + "graf": 76039, + "gram": 1520, + "grammar": 41094, + "grams": 50770, + "gran": 67063, + "grand": 52666, + "grant": 51627, + "graph": 4439, + "graphic": 85872, + "graphics": 31685, + "graphql": 38294, + "graphs": 86186, + "grass": 53354, + "gratis": 94746, + "grav": 35390, + "grave": 76631, + "gravity": 70791, + "gray": 11650, + "gre": 15525, + "greSQL": 59796, + "great": 46891, + "greater": 65235, + "gree": 22490, + "green": 13250, + "greens": 81726, + "greg": 7998, + "gregar": 58932, + "gregate": 14240, + "gregated": 93040, + "gregation": 34442, + "gregator": 58131, + "gren": 74265, + "grep": 38205, + "gres": 17818, + "grese": 48450, + "greso": 72156, + "gresql": 81624, + "gress": 2483, + "gression": 29740, + "gressive": 46719, + "gressor": 73231, + "grey": 34571, + "gricult": 12835, + "grid": 4203, + "grily": 79955, + "grim": 86097, + "gro": 41411, + "gross": 93802, + "ground": 1951, + "groundColor": 56716, + "grounds": 62111, + "group": 4074, + "groupBox": 63618, + "groupBy": 71288, + "groupId": 51854, + "groupName": 90631, + "groupid": 93799, + "groupon": 69914, + "groups": 16753, + "grow": 66218, + "grown": 61276, + "growth": 73089, + "grp": 42854, + "grpc": 56585, + "grund": 59785, + "grunt": 61224, + "grupo": 71977, + "gráf": 137434, + "gráfica": 137435, + "gráfico": 135912, + "gré": 89866, + "grÃ¶ÃŁe": 143721, + "grün": 136613, + "gründe": 137609, + "gs": 5857, + "gst": 63205, + "gsub": 94548, + "gt": 5178, + "gte": 55067, + "gtest": 82038, + "gtk": 40749, + "gu": 8717, + "gua": 84269, + "guard": 26098, + "guarded": 96303, + "guards": 78738, + "gue": 37668, + "guess": 51751, + "guest": 30177, + "gui": 19109, + "guid": 23448, + "guide": 48370, + "guided": 65419, + "guild": 57877, + "gulp": 23328, + "gun": 13259, + "guna": 61275, + "gunakan": 74705, + "guns": 51821, + "gunta": 59910, + "gué": 136418, + "gv": 21404, + "gw": 54133, + "gx": 59957, + "gy": 4577, + "gypt": 13539, + "gz": 46589, + "gzip": 70996, + "gänge": 140299, + "gé": 126295, + "géni": 137583, + "gência": 132439, + "gó": 129253, + "gü": 126547, + "gün": 133372, + "gı": 128137, + "gÅĤ": 125797, + "gÅĤo": 135142, + "gÅĤos": 130315, + "gÅĤow": 133542, + "gÅĤoÅĽ": 135143, + "gÅĤÄĻb": 136092, + "h": 71, + "ha": 4223, + "haar": 68129, + "hab": 10573, + "habi": 61595, + "habit": 96702, + "hack": 65972, + "had": 31245, + "haft": 81760, + "hai": 25214, + "hair": 50629, + "hait": 97708, + "hal": 11866, + "halb": 49662, + "haled": 59885, + "half": 37006, + "hall": 42241, + "halt": 39416, + "halten": 35505, + "ham": 5604, + "hammad": 33057, + "hammer": 45334, + "hamster": 94265, + "han": 9917, + "hana": 83103, + "hand": 10661, + "handle": 8192, + "handleChange": 48694, + "handleRequest": 91237, + "handleSubmit": 65560, + "handled": 38821, + "handler": 17905, + "handlers": 55627, + "handles": 43931, + "handling": 90097, + "hands": 88012, + "hang": 20658, + "hani": 86811, + "hap": 37810, + "hape": 29788, + "happy": 56521, + "haps": 6224, + "hapus": 67955, + "har": 12982, + "hard": 18703, + "hardt": 59282, + "hardware": 68215, + "hare": 76711, + "harga": 59862, + "hari": 75907, + "harma": 77238, + "hart": 46389, + "has": 4648, + "hasClass": 40172, + "hasMany": 48296, + "hasOne": 80347, + "hash": 8296, + "hashCode": 60695, + "hashed": 95647, + "hashtags": 30209, + "hasil": 32956, + "hat": 9198, + "hatt": 23104, + "hattan": 25337, + "haul": 43156, + "haus": 31551, + "hausen": 76147, + "haust": 15074, + "have": 19016, + "haven": 70912, + "having": 68566, + "hawk": 81154, + "hawks": 70287, + "hay": 67312, + "hazi": 63585, + "hb": 49039, + "hc": 38052, + "hci": 73810, + "hcp": 62169, + "hd": 15990, + "hdl": 90374, + "hdr": 28785, + "he": 383, + "hea": 39933, + "head": 1983, + "headed": 62500, + "header": 2708, + "headers": 7713, + "heading": 11412, + "headline": 63193, + "heads": 35810, + "health": 12120, + "healthy": 37028, + "heap": 15961, + "hear": 86060, + "heard": 54503, + "heart": 17735, + "heartbeat": 72471, + "heat": 19963, + "heatmap": 94061, + "heavy": 88845, + "hec": 41650, + "heck": 1227, + "hecy": 49731, + "hed": 41009, + "hee": 49053, + "heed": 56554, + "heel": 8431, + "heels": 72510, + "heet": 3674, + "heets": 20306, + "hei": 60852, + "heid": 44797, + "height": 2563, + "heim": 20572, + "heimer": 34080, + "heiro": 59477, + "heit": 22385, + "heiten": 89249, + "heits": 87818, + "hek": 91080, + "hel": 49122, + "held": 27608, + "hell": 56095, + "hello": 14990, + "helm": 51899, + "help": 8653, + "helper": 18764, + "helpers": 21723, + "hem": 29032, + "hen": 12032, + "her": 1923, + "here": 6739, + "herence": 51361, + "herent": 37538, + "herit": 9263, + "heritance": 62443, + "herited": 48394, + "herits": 38693, + "hero": 17732, + "heroes": 53476, + "hes": 16668, + "hesion": 58638, + "hesive": 40818, + "hest": 6402, + "het": 13803, + "heten": 93933, + "hetic": 26863, + "hetics": 48366, + "hetto": 76242, + "heure": 74776, + "hev": 80813, + "hevik": 80908, + "hex": 17308, + "hexdigest": 97224, + "hey": 35561, + "hf": 44754, + "hg": 66602, + "hh": 20367, + "hhh": 90064, + "hi": 6023, + "hib": 5803, + "hiba": 78476, + "hibit": 65817, + "hibited": 92517, + "hibition": 58973, + "hic": 82513, + "hid": 46453, + "hidden": 6263, + "hide": 8559, + "hift": 47833, + "high": 11892, + "higher": 89347, + "highest": 74154, + "highlight": 35198, + "hil": 75323, + "hill": 27507, + "him": 39517, + "hin": 41557, + "hind": 73938, + "hint": 46125, + "hints": 95348, + "hip": 2151, + "hips": 33222, + "hipster": 95988, + "hir": 29628, + "hire": 34448, + "hiro": 85138, + "his": 25235, + "hist": 21158, + "histoire": 89346, + "histor": 43441, + "historic": 73316, + "history": 18844, + "hit": 22492, + "hits": 50205, + "hive": 88568, + "hiên": 136081, + "hiếu": 137503, + "hk": 85511, + "hl": 17958, + "hlen": 70667, + "hm": 34301, + "hma": 40353, + "hn": 24134, + "ho": 6161, + "hoa": 98661, + "hoc": 88473, + "hoe": 75935, + "hof": 75858, + "hoff": 93304, + "hog": 60622, + "hol": 8452, + "hold": 6282, + "holder": 4251, + "holders": 16662, + "holding": 42491, + "holds": 53119, + "hole": 30420, + "holes": 74690, + "holiday": 70540, + "holm": 33505, + "hom": 34840, + "home": 5117, + "homepage": 52019, + "homes": 85272, + "homme": 86613, + "hon": 81549, + "hone": 25256, + "hood": 5694, + "hook": 20873, + "hooks": 38560, + "hoot": 90321, + "hop": 25115, + "hope": 60793, + "hopefully": 96425, + "hor": 19530, + "hora": 46364, + "hores": 82071, + "horia": 95629, + "horizontal": 30629, + "horn": 50884, + "horse": 59675, + "hort": 22007, + "hos": 41835, + "hospital": 70027, + "host": 3790, + "hostname": 27806, + "hosts": 44692, + "hot": 10622, + "hotel": 38817, + "hotmail": 55454, + "hots": 34413, + "hou": 18166, + "houette": 45625, + "hound": 75096, + "hour": 24677, + "hours": 30382, + "house": 7675, + "houses": 36741, + "housing": 80615, + "hover": 17583, + "how": 5158, + "hower": 83018, + "however": 97836, + "hp": 21197, + "hpp": 90843, + "hq": 81899, + "hr": 4079, + "hra": 71910, + "hread": 73839, + "href": 13335, + "hrs": 65262, + "hs": 4997, + "hsi": 67494, + "ht": 426, + "hta": 75809, + "htable": 26686, + "htag": 34311, + "htags": 25285, + "htaking": 52243, + "htar": 84644, + "htdocs": 39748, + "hte": 55499, + "hti": 94848, + "htm": 87372, + "html": 1551, + "htmlspecialchars": 79596, + "hton": 64712, + "htt": 96336, + "http": 1254, + "httpClient": 83417, + "https": 2428, + "htub": 65413, + "hu": 16739, + "hua": 91236, + "hub": 26682, + "hud": 80388, + "hue": 92923, + "huge": 95370, + "hum": 27300, + "human": 25312, + "humidity": 93046, + "hung": 56832, + "hunt": 80075, + "hunter": 98272, + "hurst": 69100, + "hus": 12800, + "hv": 81979, + "hw": 27827, + "hwnd": 97318, + "hx": 58424, + "hy": 8503, + "hya": 86223, + "hydr": 25172, + "hydrate": 87378, + "hydration": 80724, + "hydro": 66129, + "hyp": 78243, + "hyper": 68192, + "hyth": 22576, + "hythm": 60154, + "hz": 37259, + "hä": 127533, + "hält": 130655, + "hänge": 139101, + "häuser": 143492, + "hé": 128959, + "hôtel": 133068, + "hões": 93230, + "hör": 129602, + "hế": 140652, + "há»ı": 137899, + "há»ķ": 137390, + "hợ": 137129, + "i": 72, + "iOS": 62760, + "iPad": 66293, + "iParam": 69908, + "iPhone": 44740, + "iT": 83589, + "iVar": 81885, + "ia": 685, + "iability": 78343, + "iable": 2156, + "iac": 17569, + "iad": 62944, + "iado": 96709, + "iae": 71978, + "iage": 64440, + "iagnostics": 18938, + "iah": 19310, + "iais": 56467, + "ial": 530, + "iale": 19893, + "ialect": 55056, + "iales": 32788, + "iali": 81064, + "ialias": 88214, + "ialis": 17691, + "ialized": 76974, + "ially": 33475, + "ialog": 2344, + "ials": 10309, + "iam": 4932, + "iameter": 36044, + "iami": 15264, + "iamo": 34214, + "iamond": 18479, + "iams": 12382, + "ian": 1103, + "iana": 12417, + "iance": 5284, + "iances": 21813, + "iane": 35035, + "iang": 27224, + "iangle": 63850, + "iani": 46447, + "ianne": 63352, + "iano": 13088, + "ians": 5380, + "iant": 5372, + "iants": 62278, + "iao": 22516, + "iap": 67721, + "iar": 13572, + "iard": 67865, + "iards": 80460, + "iare": 95186, + "ias": 3473, + "iasco": 97283, + "iasi": 97554, + "iasm": 32890, + "iat": 10358, + "iate": 6493, + "iated": 10029, + "iates": 42298, + "iating": 22897, + "iation": 7101, + "iationException": 74027, + "iations": 17144, + "iative": 53579, + "iator": 36122, + "iators": 80290, + "iatric": 22520, + "iatrics": 70428, + "iaux": 82291, + "iaz": 26975, + "iazza": 89414, + "iação": 67643, + "iaÅĤ": 34284, + "iaÅĤa": 94590, + "ib": 579, + "ibBundleOrNil": 89390, + "ibName": 59344, + "ibNameOrNil": 89326, + "iba": 34938, + "ibaba": 28260, + "ibal": 50529, + "iban": 33779, + "iband": 87363, + "ibase": 98746, + "ibbean": 32059, + "ibble": 82418, + "ibbon": 35080, + "ibe": 23549, + "ibel": 68595, + "iben": 63236, + "iber": 8629, + "ibern": 17660, + "ibernate": 18818, + "ibi": 39393, + "ibia": 66237, + "ibil": 30517, + "ibile": 41417, + "ibili": 85948, + "ibilidad": 56235, + "ibilidade": 86585, + "ibilit": 39848, + "ibilities": 13450, + "ibility": 3147, + "ibir": 82460, + "ible": 1238, + "ibles": 20410, + "ibli": 68613, + "ibling": 19876, + "iblings": 24229, + "ibly": 6487, + "ibo": 28846, + "ibold": 58395, + "ibox": 70283, + "ibr": 10665, + "ibraltar": 84796, + "ibraries": 72616, + "ibrary": 27666, + "ibrate": 43092, + "ibrated": 60566, + "ibration": 18350, + "ibrator": 84894, + "ibre": 63591, + "ibri": 57834, + "ibs": 54341, + "ibt": 20319, + "ibu": 56982, + "ibur": 85198, + "ibus": 33395, + "ic": 292, + "ica": 3001, + "icable": 74338, + "icago": 9440, + "icaid": 31484, + "ical": 938, + "ically": 2673, + "icals": 52603, + "icamente": 51677, + "ican": 7065, + "icana": 98505, + "icans": 37373, + "icao": 36982, + "icap": 97350, + "icare": 80038, + "icas": 15185, + "icast": 35446, + "icate": 8463, + "icated": 10089, + "icates": 24821, + "ication": 20285, + "icator": 13254, + "icators": 42052, + "icação": 96192, + "icc": 47638, + "ice": 558, + "iced": 7572, + "icemail": 86283, + "icens": 6314, + "icense": 1545, + "icensed": 7296, + "icensing": 23411, + "iceps": 74569, + "icer": 12999, + "icerca": 72662, + "icers": 34980, + "ices": 1216, + "icester": 50677, + "ich": 713, + "ichTextBox": 52459, + "icha": 61387, + "ichael": 59708, + "ichage": 72165, + "iche": 12063, + "ichel": 90240, + "ichen": 28324, + "icher": 28064, + "ichern": 88581, + "ichert": 92473, + "ichever": 45569, + "ichi": 40552, + "ichick": 98599, + "ichier": 62796, + "icho": 77115, + "icht": 4865, + "ichte": 37347, + "ichten": 28698, + "ichtet": 89730, + "ichtig": 34966, + "ici": 3375, + "icia": 24024, + "icial": 7090, + "ician": 12452, + "icians": 9621, + "iciar": 72451, + "iciary": 26813, + "icias": 49685, + "icide": 14502, + "icides": 75095, + "icie": 76277, + "iciel": 77393, + "iciencies": 42590, + "iciency": 10387, + "icient": 5385, + "iciente": 65680, + "icients": 28142, + "icies": 44339, + "icina": 72865, + "icine": 54003, + "icing": 10124, + "icio": 11267, + "icion": 13898, + "icional": 62793, + "icionar": 53793, + "iciones": 41224, + "icions": 77110, + "icios": 25345, + "icious": 9627, + "icip": 8442, + "icipant": 21757, + "icipants": 27925, + "icipation": 76353, + "icism": 41814, + "icit": 18677, + "icits": 50550, + "icity": 61691, + "ición": 25742, + "ick": 865, + "icked": 18504, + "icken": 9692, + "icker": 5215, + "ickerView": 76256, + "ickers": 39089, + "ickest": 69752, + "icket": 5897, + "ickets": 17480, + "ickey": 74939, + "icking": 16272, + "ickle": 25606, + "ickname": 41052, + "ickness": 25900, + "icks": 5788, + "ickt": 70980, + "icky": 18964, + "icl": 87367, + "iclass": 69503, + "icle": 2002, + "icles": 4342, + "icmp": 50774, + "ico": 3955, + "icode": 13634, + "icol": 26463, + "icolon": 38517, + "icolor": 73809, + "icom": 53994, + "icon": 1924, + "iconductor": 51492, + "icone": 36173, + "icons": 10670, + "icont": 51160, + "icontains": 99029, + "icontrol": 77006, + "icopt": 23951, + "icopter": 87223, + "icorn": 50356, + "icos": 16627, + "icot": 86437, + "icro": 2754, + "icrobial": 81959, + "icrosoft": 3929, + "icrous": 83189, + "ics": 1211, + "ict": 849, + "icted": 12770, + "icter": 79639, + "ictim": 51144, + "iction": 2479, + "ictionaries": 44192, + "ictionary": 3916, + "ictions": 21439, + "ictory": 90033, + "icts": 29995, + "icture": 3826, + "ictureBox": 10201, + "ictured": 29275, + "ictures": 7489, + "icular": 23632, + "icularly": 48157, + "iculo": 32653, + "iculos": 67277, + "icult": 3866, + "icultural": 52815, + "iculture": 92728, + "iculty": 25990, + "icum": 85131, + "icus": 58244, + "icut": 27306, + "icy": 2842, + "icycle": 26165, + "icz": 52222, + "id": 307, + "idUser": 99100, + "ida": 4744, + "idable": 36754, + "idad": 5845, + "idade": 9075, + "idades": 13596, + "idak": 60389, + "idal": 25880, + "idan": 55204, + "idar": 92902, + "idas": 17015, + "idata": 81080, + "iday": 48878, + "idd": 1772, + "idden": 2854, + "idders": 99129, + "idding": 26500, + "iddle": 3310, + "iddled": 87762, + "iddles": 77455, + "iddleware": 11603, + "iddy": 75570, + "ide": 577, + "idea": 41777, + "ideal": 62953, + "ideas": 93684, + "idebar": 12683, + "ided": 4490, + "idel": 26802, + "idelberg": 91919, + "idelity": 45101, + "iden": 12487, + "idenav": 49263, + "idence": 7087, + "idences": 83996, + "idency": 25775, + "idend": 68850, + "ident": 1713, + "idental": 61623, + "identally": 56122, + "idente": 96278, + "idential": 11234, + "identification": 91577, + "identified": 42260, + "identifier": 15909, + "identify": 80396, + "identity": 16912, + "idents": 6880, + "ideo": 1888, + "ideographic": 94007, + "ideon": 99142, + "ideos": 5188, + "idepress": 58833, + "ider": 1776, + "iders": 13269, + "ides": 3341, + "ideshow": 44700, + "idf": 77004, + "idge": 6478, + "idges": 29163, + "idget": 2006, + "idi": 12278, + "idia": 86700, + "idian": 35642, + "idine": 89973, + "iding": 6577, + "idious": 79127, + "idis": 84242, + "idity": 18518, + "idl": 55464, + "idle": 47273, + "ido": 5249, + "idon": 90456, + "idor": 28756, + "idores": 80421, + "idos": 13349, + "idot": 93269, + "ids": 3365, + "idth": 1104, + "idual": 3340, + "idue": 60607, + "idunt": 56043, + "idx": 6361, + "idy": 42586, + "idée": 131742, + "ie": 645, + "ieber": 60241, + "iec": 66492, + "iece": 9108, + "ieces": 29765, + "iect": 91542, + "ied": 1122, + "iedad": 39029, + "iedade": 95166, + "iedades": 73913, + "ieder": 22181, + "iedo": 87051, + "iedy": 81619, + "ieee": 82154, + "ief": 4738, + "iefs": 29821, + "ieg": 35010, + "iegel": 74947, + "iego": 62287, + "iei": 75999, + "iej": 56547, + "iek": 35007, + "iel": 13029, + "ield": 816, + "ielding": 98992, + "ields": 6907, + "iele": 24619, + "iem": 26862, + "iembre": 53471, + "ieme": 88030, + "ien": 3591, + "iena": 86651, + "ience": 1835, + "iences": 11975, + "ienda": 77921, + "iendo": 36016, + "iene": 42568, + "ienen": 31261, + "ienes": 56862, + "ienia": 51458, + "ienie": 74322, + "ienne": 25503, + "iens": 79363, + "ient": 1167, + "iente": 12844, + "ientes": 26240, + "iento": 10873, + "ientos": 31960, + "ientras": 49892, + "ients": 4075, + "ienza": 65502, + "ier": 1268, + "iera": 25834, + "ieran": 81381, + "ierarchical": 98950, + "ierarchy": 33880, + "ierce": 88968, + "iere": 18849, + "ieren": 16030, + "ieres": 64488, + "ierge": 86003, + "ieri": 53939, + "iero": 43588, + "ieron": 39508, + "ierre": 44423, + "ierrez": 95080, + "iers": 4813, + "iership": 76078, + "iert": 17029, + "ierte": 65161, + "ierten": 77399, + "ierung": 36616, + "ierz": 65343, + "ies": 550, + "iese": 81053, + "iesel": 82130, + "iesen": 93840, + "iest": 13438, + "iesta": 58449, + "iesz": 49806, + "iet": 3891, + "iete": 66860, + "ieten": 33254, + "ietet": 56908, + "ietf": 74068, + "ieties": 31506, + "iets": 80781, + "ieu": 25173, + "ieur": 60905, + "ieurs": 45038, + "ieux": 52719, + "iev": 43091, + "ieval": 31886, + "ieve": 19155, + "ieved": 30977, + "iever": 81021, + "ieves": 38135, + "ieving": 83346, + "iew": 570, + "iez": 73366, + "ież": 55723, + "ieÅĦ": 130694, + "if": 333, + "ifa": 33818, + "iface": 52674, + "ifact": 20754, + "ifacts": 26401, + "ifar": 92927, + "ifax": 53617, + "ifdef": 34732, + "ife": 1612, + "ifecycle": 19517, + "ifen": 55339, + "ifer": 11437, + "iferay": 33204, + "ifers": 98813, + "ifes": 82295, + "ifest": 6962, + "ifestyle": 16479, + "ifestyles": 72525, + "ifetime": 28515, + "iff": 3092, + "iffany": 55082, + "iffe": 43414, + "iffer": 14320, + "ifference": 18192, + "ifferences": 58326, + "ifferent": 18277, + "ifferential": 68657, + "iffies": 80571, + "iffin": 83864, + "iffs": 18852, + "ifi": 12919, + "ifiable": 22619, + "ifiant": 83454, + "ific": 1086, + "ifica": 29488, + "ificacion": 37703, + "ificaciones": 89086, + "ificación": 52407, + "ificado": 47138, + "ificador": 54900, + "ificados": 90762, + "ificance": 99004, + "ificant": 33728, + "ificantly": 97764, + "ificar": 23244, + "ificate": 9261, + "ificates": 24405, + "ification": 2404, + "ifications": 7029, + "ificação": 81451, + "ifice": 21898, + "ificent": 36143, + "ificial": 16488, + "ificio": 92638, + "ifie": 56509, + "ified": 1870, + "ifier": 3049, + "ifiers": 11836, + "ifies": 9606, + "ifik": 48672, + "ifikasi": 77350, + "ifique": 51660, + "ifix": 96322, + "ifié": 89759, + "ifle": 19092, + "iflower": 76773, + "ifndef": 56919, + "ifo": 31497, + "ifold": 48285, + "iform": 7250, + "iforn": 6417, + "ifornia": 6574, + "ifr": 96716, + "iframe": 23207, + "ifs": 21835, + "ifstream": 49497, + "ift": 2085, + "ifter": 37700, + "ifth": 57610, + "ifting": 17680, + "ifton": 90261, + "ifty": 38624, + "ifu": 20850, + "iful": 4985, + "ify": 1437, + "ifying": 7766, + "ig": 343, + "igDecimal": 30183, + "igInteger": 47950, + "iga": 16551, + "igail": 84206, + "igan": 10881, + "igans": 75757, + "igar": 51284, + "igaret": 43275, + "igate": 63956, + "igated": 82241, + "igation": 17930, + "igator": 57082, + "igators": 64983, + "ige": 7256, + "igel": 49539, + "igen": 6433, + "igenous": 25354, + "iger": 7272, + "igeria": 63717, + "igers": 32114, + "iges": 51663, + "igest": 15153, + "igg": 20218, + "igger": 4500, + "iggers": 21385, + "iggins": 51926, + "iggs": 61147, + "igh": 1090, + "igham": 85222, + "ighb": 4681, + "ighbor": 19247, + "ighborhood": 46185, + "ighbors": 24101, + "ighbour": 46818, + "ighbours": 52117, + "ighest": 7504, + "ighet": 81508, + "ight": 491, + "ighted": 64507, + "ighter": 85200, + "ighth": 21877, + "ighthouse": 57909, + "ighting": 44750, + "ightly": 71948, + "ighton": 87773, + "ights": 2796, + "igi": 29537, + "igid": 20926, + "igidBody": 69181, + "igidbody": 29690, + "igin": 9143, + "iginal": 11040, + "igion": 31618, + "igious": 22153, + "igit": 12597, + "igital": 10233, + "igits": 18667, + "igkeit": 50499, + "igli": 76161, + "iglia": 79996, + "igm": 24129, + "igma": 11491, + "igmat": 97571, + "igmatic": 98730, + "igmoid": 52111, + "ign": 622, + "ignKey": 13466, + "ignal": 25719, + "ignant": 58010, + "igne": 18857, + "igned": 1542, + "igner": 77656, + "ignet": 84858, + "ignite": 95408, + "ignment": 3531, + "ignon": 61699, + "ignore": 13130, + "ignored": 58471, + "ignty": 42743, + "ignum": 92160, + "igo": 7836, + "igon": 73881, + "igor": 69970, + "igos": 32239, + "igr": 5233, + "igram": 49127, + "igrams": 83109, + "igrant": 56019, + "igrants": 70236, + "igraph": 53867, + "igraphy": 92070, + "igrate": 34479, + "igrated": 94830, + "igration": 5033, + "igrationBuilder": 52255, + "igrations": 17824, + "igroup": 74658, + "igs": 14462, + "igsaw": 86924, + "igslist": 48922, + "igt": 17917, + "igte": 92172, + "igth": 89690, + "igu": 26745, + "igua": 90746, + "igue": 21945, + "iguiente": 89776, + "igung": 42603, + "iguous": 27029, + "igure": 9811, + "igy": 92493, + "ih": 6996, + "ihad": 30613, + "ihan": 56888, + "ihanna": 78357, + "ihar": 71641, + "ihat": 80877, + "ihil": 51426, + "ihilation": 91241, + "ihn": 90550, + "ihu": 81884, + "ii": 3808, + "iid": 54483, + "iii": 34594, + "ij": 3172, + "ija": 28130, + "ijd": 32223, + "ije": 31276, + "iji": 34873, + "ijing": 23649, + "ijk": 18750, + "ijke": 50405, + "ijken": 56902, + "ijkl": 59779, + "ijkstra": 87177, + "ijn": 14153, + "ijo": 46900, + "iju": 63174, + "ik": 1579, + "ika": 11496, + "ikal": 87460, + "ikan": 22222, + "ike": 2970, + "iked": 42608, + "ikel": 28855, + "iken": 36433, + "iker": 24803, + "ikers": 53113, + "ikes": 12450, + "ikh": 30503, + "ikhail": 82256, + "iki": 7570, + "iking": 19233, + "ikip": 14628, + "ikipedia": 14939, + "ikit": 60403, + "ikk": 29887, + "ikki": 83541, + "iko": 23630, + "ikon": 61614, + "iks": 67891, + "ikt": 31580, + "iktig": 96433, + "iku": 38242, + "ikut": 88994, + "il": 321, + "ila": 10524, + "ilage": 86976, + "ilan": 89359, + "iland": 78691, + "ilar": 3569, + "ilarity": 48225, + "ilate": 75357, + "ilated": 92483, + "ilater": 90795, + "ilateral": 42939, + "ilation": 13455, + "ild": 695, + "ilda": 55161, + "ilde": 34276, + "ilded": 90254, + "ilden": 51865, + "ildenafil": 65086, + "ilder": 20776, + "ildo": 29945, + "ile": 457, + "ileaks": 83964, + "iled": 2181, + "ilee": 63066, + "ileen": 90694, + "ilege": 42769, + "ileged": 68431, + "ileges": 70838, + "ilen": 22837, + "ilename": 4033, + "ilenames": 52768, + "ilent": 68414, + "ileo": 84134, + "iler": 5769, + "ilers": 27158, + "iles": 3658, + "iless": 72141, + "ilestone": 93028, + "ilet": 20619, + "iley": 17801, + "ilha": 84110, + "ili": 3921, + "ilia": 24078, + "ilian": 68083, + "ilians": 25625, + "iliar": 9538, + "iliary": 34446, + "iliate": 34850, + "iliated": 53773, + "iliation": 28009, + "ilib": 29959, + "ilibrium": 45680, + "ilies": 7597, + "ilig": 73276, + "ilight": 34900, + "ilih": 46968, + "ilihan": 87162, + "ilik": 96159, + "iliki": 65023, + "iline": 26560, + "ilinear": 70933, + "iling": 7979, + "ilingual": 49823, + "ilinx": 98033, + "ilio": 33329, + "ilion": 48912, + "ilit": 3664, + "ilitary": 20247, + "ilitating": 75154, + "ilitation": 32849, + "ilities": 4298, + "ility": 1403, + "ilk": 33163, + "ill": 483, + "illa": 6241, + "illac": 63287, + "illage": 47064, + "illance": 18832, + "illard": 66259, + "illary": 34505, + "illas": 33244, + "illation": 66084, + "illator": 70084, + "illaume": 98012, + "ille": 4517, + "illed": 4374, + "illegal": 75998, + "iller": 15252, + "illery": 35369, + "illes": 30255, + "illet": 40884, + "illez": 67390, + "illi": 57168, + "illian": 64721, + "illin": 60497, + "illing": 9400, + "illion": 14726, + "illions": 90287, + "illis": 56212, + "illisecond": 65358, + "illiseconds": 56194, + "illo": 21782, + "illon": 42488, + "illos": 64937, + "illow": 95923, + "ills": 3305, + "illum": 76565, + "illus": 86337, + "illusion": 81162, + "illustr": 82618, + "illy": 14385, + "illé": 139041, + "ilm": 9484, + "ilmington": 74534, + "ilmiÅŁ": 133021, + "ilo": 18045, + "iloc": 25953, + "ilog": 50773, + "ilogue": 90492, + "ilogy": 40058, + "ilon": 16917, + "ilor": 49538, + "ilos": 89319, + "ilot": 23958, + "ils": 8669, + "ilst": 19282, + "ilt": 2963, + "ilter": 2044, + "ilters": 8612, + "ilton": 16296, + "iltr": 36604, + "iltro": 46120, + "ilty": 15054, + "ilver": 47030, + "ily": 1541, + "ilyn": 68482, + "im": 318, + "ima": 7523, + "imachinery": 70355, + "imag": 28016, + "image": 1805, + "imageName": 79778, + "imageUrl": 67673, + "imageView": 70245, + "imagem": 84696, + "imagen": 39347, + "imagenes": 85285, + "images": 3642, + "imagin": 79124, + "imal": 2861, + "imals": 24163, + "iman": 39210, + "imap": 39942, + "imar": 87356, + "imary": 3111, + "imas": 49363, + "imat": 77332, + "imate": 3426, + "imated": 7145, + "imately": 7108, + "imates": 47471, + "imating": 78348, + "imation": 5465, + "imations": 96376, + "imator": 13689, + "imators": 71036, + "imax": 75109, + "imb": 20828, + "imbabwe": 43181, + "imbledon": 84663, + "imbus": 95216, + "imd": 87493, + "imdi": 127540, + "ime": 545, + "imeInterval": 80932, + "imeType": 34872, + "imed": 75485, + "imedia": 26936, + "imei": 97861, + "imeline": 54147, + "imen": 26144, + "imens": 45597, + "imension": 18161, + "imensional": 28221, + "iment": 3819, + "imentary": 74543, + "imenti": 83297, + "imento": 15027, + "imentos": 55273, + "iments": 12558, + "imeo": 50888, + "imer": 3134, + "imers": 67897, + "imes": 1733, + "imest": 4255, + "imestamp": 4702, + "imestep": 48747, + "imesteps": 76632, + "imestone": 35441, + "imet": 85266, + "imeter": 25334, + "imeters": 54236, + "imetype": 72777, + "img": 1892, + "imgs": 58575, + "imi": 24836, + "imid": 98494, + "imiento": 16320, + "imientos": 44962, + "imilar": 78862, + "imin": 60234, + "iminal": 8150, + "iminary": 77873, + "iming": 59076, + "imir": 30104, + "imit": 2353, + "imited": 31511, + "imiter": 17700, + "imiters": 67645, + "imitive": 13654, + "imitives": 35209, + "imits": 22866, + "imity": 30935, + "imité": 98381, + "imization": 65964, + "imize": 11853, + "imizeBox": 64534, + "imized": 45706, + "imizer": 41699, + "imm": 12543, + "immel": 64864, + "immer": 18982, + "immers": 72583, + "imming": 40033, + "immune": 69055, + "immutable": 91240, + "imo": 11366, + "imon": 25417, + "imonial": 49325, + "imonials": 57335, + "imony": 64456, + "imore": 21382, + "imos": 24107, + "imoto": 96158, + "imp": 6664, + "impact": 57300, + "impan": 89888, + "impl": 6383, + "imple": 6456, + "implement": 94474, + "implementation": 14369, + "implemented": 54071, + "implements": 93958, + "implicit": 30940, + "implicitly": 87295, + "implified": 73837, + "implify": 70206, + "implode": 55577, + "import": 474, + "importDefault": 86479, + "important": 15333, + "importe": 93796, + "imports": 62888, + "imps": 66658, + "impse": 39427, + "ims": 5742, + "imshow": 65447, + "imson": 45445, + "imu": 78326, + "imulation": 60361, + "imulator": 91503, + "imum": 3295, + "imus": 76928, + "imuth": 55207, + "in": 258, + "ina": 2210, + "inactive": 61137, + "inaire": 67876, + "inal": 977, + "inalg": 33231, + "inality": 80777, + "inally": 37645, + "inals": 23695, + "iname": 79856, + "inan": 73176, + "inance": 24387, + "inand": 75731, + "inant": 85296, + "inar": 13762, + "inars": 46337, + "inary": 3287, + "inas": 20114, + "inate": 3277, + "inated": 15479, + "inati": 97040, + "inating": 15873, + "ination": 2554, + "inations": 12634, + "inator": 17272, + "inbox": 84699, + "inburgh": 36786, + "inc": 2840, + "incare": 65903, + "ince": 1701, + "incer": 69798, + "incerely": 86091, + "inces": 30076, + "incess": 19570, + "inceton": 41590, + "inch": 36750, + "inci": 48597, + "incible": 76872, + "incident": 75011, + "incinn": 31997, + "incinnati": 32176, + "incip": 5706, + "incipal": 15702, + "inciple": 15996, + "incl": 91157, + "include": 997, + "included": 82622, + "includes": 14526, + "including": 16169, + "inclusive": 98361, + "income": 56966, + "incoming": 79677, + "incorrect": 61954, + "incr": 98428, + "incre": 42742, + "increase": 78501, + "increment": 35744, + "increments": 56437, + "inct": 7800, + "inction": 21362, + "inctions": 73590, + "incy": 75586, + "ind": 484, + "inda": 17416, + "inde": 26976, + "inded": 16593, + "inden": 58217, + "indent": 32840, + "inder": 5744, + "indered": 82203, + "inders": 31451, + "index": 1252, + "indexOf": 29849, + "indexPath": 18887, + "indexed": 97473, + "indexes": 71227, + "indh": 65989, + "indhoven": 82058, + "indi": 28685, + "indic": 85084, + "indicator": 56151, + "indice": 73417, + "indices": 14987, + "inding": 3961, + "individual": 54877, + "indle": 57763, + "indo": 34999, + "indow": 985, + "indows": 1491, + "indr": 80587, + "indre": 70833, + "indrical": 62406, + "indrome": 36058, + "inds": 8673, + "indsay": 53977, + "indsight": 78860, + "industry": 70764, + "indx": 86707, + "indy": 48990, + "ine": 482, + "ineTransform": 49781, + "inea": 34418, + "inear": 88921, + "inecraft": 12086, + "ined": 1589, + "inee": 84799, + "inel": 29708, + "inely": 22094, + "inem": 68413, + "inema": 93423, + "inement": 37184, + "inen": 16676, + "inent": 13847, + "iner": 10453, + "inerary": 51553, + "iners": 39699, + "inery": 16323, + "ines": 1543, + "inese": 7346, + "inesis": 82789, + "iness": 1880, + "inet": 29876, + "inez": 38295, + "inf": 13573, + "infeld": 98836, + "infer": 89559, + "infile": 74833, + "infinity": 80016, + "inflate": 37455, + "info": 2733, + "inform": 40440, + "informatics": 97488, + "information": 25069, + "infos": 55965, + "infra": 92317, + "ing": 287, + "inge": 17707, + "inged": 56558, + "ingen": 27918, + "inger": 5137, + "ingerprint": 47918, + "ingers": 14105, + "inges": 40399, + "ingga": 65401, + "ingham": 23246, + "inging": 24030, + "ingle": 2173, + "ingles": 49603, + "ingleton": 75710, + "ingly": 11307, + "ingo": 27908, + "ingredient": 38525, + "ingredients": 38120, + "ingroup": 41281, + "ings": 819, + "ingt": 74919, + "ington": 4805, + "ingu": 12810, + "ingular": 41880, + "inh": 20868, + "inha": 43238, + "inherit": 12825, + "inheritDoc": 32469, + "inheritdoc": 21047, + "inherits": 77836, + "inho": 27576, + "ini": 6591, + "inia": 90450, + "inic": 92335, + "inicio": 64053, + "inidad": 68155, + "inin": 64444, + "inine": 82234, + "ining": 5740, + "inion": 36300, + "inions": 83896, + "inis": 84603, + "inish": 18176, + "inished": 24657, + "init": 2327, + "initWith": 24214, + "inite": 15856, + "initely": 7475, + "initial": 9426, + "initialize": 21641, + "initialized": 36161, + "initializer": 46774, + "inition": 4054, + "inity": 13489, + "inium": 63890, + "iniz": 69639, + "inja": 41420, + "inject": 32133, + "ink": 766, + "inka": 74122, + "inke": 95873, + "inkel": 81176, + "inker": 41112, + "inki": 49409, + "inking": 15736, + "inkle": 35144, + "inks": 15504, + "inky": 38034, + "inline": 5057, + "inn": 6130, + "innacle": 68953, + "innamon": 39335, + "inne": 93232, + "inned": 20561, + "innen": 34711, + "inner": 4382, + "innerHTML": 15609, + "innerText": 55017, + "inness": 70673, + "innie": 98189, + "inning": 19770, + "innitus": 97339, + "ino": 3313, + "inoa": 78994, + "inode": 52239, + "inois": 16669, + "inos": 15215, + "inosaur": 88147, + "inous": 60698, + "inox": 73079, + "inp": 42092, + "input": 1355, + "inputEmail": 97232, + "inputs": 24941, + "inq": 5885, + "inqu": 52888, + "ins": 1330, + "inscription": 98546, + "insert": 4208, + "insi": 31284, + "insic": 27592, + "insics": 67102, + "inside": 41350, + "insk": 88392, + "inski": 52877, + "insky": 51441, + "insn": 85817, + "inson": 28873, + "inspace": 30313, + "inspect": 82054, + "inspection": 53395, + "inst": 6308, + "instagram": 46266, + "install": 12248, + "installation": 80816, + "installed": 49573, + "installer": 63216, + "instanc": 33005, + "instance": 4851, + "instances": 47825, + "instancetype": 33684, + "instant": 85158, + "instead": 64837, + "instein": 36394, + "institution": 81309, + "inston": 38505, + "instr": 60511, + "instruction": 54974, + "instructions": 62295, + "instrument": 66368, + "insula": 35373, + "insurance": 69954, + "int": 396, + "intColor": 38492, + "intValue": 46040, + "inta": 33469, + "intage": 16153, + "inte": 45329, + "integer": 11662, + "integr": 80729, + "integral": 81035, + "integration": 60168, + "intel": 79543, + "intelligence": 92275, + "intendent": 43402, + "intendo": 20323, + "intent": 56431, + "inter": 2245, + "interaction": 68075, + "interactive": 37540, + "interest": 12724, + "interested": 83698, + "interesting": 87557, + "interface": 4970, + "interfaces": 28965, + "intern": 55444, + "internal": 10481, + "international": 97597, + "internet": 85743, + "interop": 31068, + "interopRequire": 32899, + "interopRequireDefault": 35364, + "interp": 63721, + "interpre": 74714, + "interpret": 16897, + "interpreted": 93461, + "interpreter": 90554, + "interrupt": 54805, + "inters": 49940, + "intersect": 96018, + "intersection": 73648, + "interval": 27904, + "intestinal": 70458, + "inth": 47021, + "intl": 97642, + "into": 18122, + "inton": 7306, + "intosh": 61594, + "intptr": 91091, + "intr": 58788, + "intree": 88583, + "intro": 36442, + "intros": 94728, + "ints": 21042, + "intval": 57602, + "inté": 133889, + "intégr": 133890, + "intér": 136207, + "intérieur": 134920, + "intérêt": 136208, + "inue": 6232, + "inus": 70856, + "inux": 19559, + "inv": 14057, + "invalid": 11808, + "invalidate": 90835, + "inventory": 31093, + "inverse": 61482, + "invert": 83729, + "invest": 42014, + "investment": 96012, + "invisible": 90751, + "invitation": 97530, + "invite": 56279, + "invoice": 22217, + "invoices": 88596, + "invoke": 22430, + "inx": 20014, + "iny": 6441, + "inya": 61087, + "inyin": 61505, + "inz": 40522, + "io": 815, + "ioc": 73352, + "ioctl": 76704, + "iod": 3127, + "iode": 94726, + "iol": 24903, + "iola": 80439, + "iolet": 30912, + "iological": 40214, + "iology": 30126, + "iom": 33595, + "iomanip": 59623, + "ioms": 90169, + "ion": 290, + "iona": 41690, + "ionage": 63690, + "ional": 3914, + "ionale": 55697, + "ionales": 82423, + "ionario": 48466, + "ionate": 84899, + "ione": 6339, + "ioned": 96557, + "ioneer": 55398, + "iones": 11431, + "ioni": 18820, + "ionic": 20764, + "ions": 908, + "ior": 2462, + "iore": 86007, + "iores": 45796, + "iors": 21867, + "ios": 3530, + "iosa": 66617, + "iosis": 90160, + "iosity": 32179, + "iosk": 74917, + "ioso": 58448, + "iostream": 9665, + "iot": 11098, + "iotic": 61014, + "iotics": 82200, + "iou": 98255, + "ious": 1223, + "iously": 13307, + "iov": 30835, + "iox": 23691, + "ioxid": 41551, + "ioxide": 36801, + "ip": 573, + "ipa": 85150, + "ipated": 62232, + "ipation": 48586, + "ipay": 43912, + "ipc": 68281, + "ipe": 3444, + "iped": 31721, + "ipeg": 46884, + "ipel": 81079, + "ipeline": 8790, + "ipelines": 93997, + "iper": 12858, + "ipers": 58835, + "ipes": 8923, + "iph": 15261, + "ipher": 10558, + "ipheral": 23298, + "ipherals": 68972, + "iphers": 82077, + "iphertext": 45043, + "iphery": 93548, + "iphone": 53359, + "iphy": 60230, + "ipi": 90921, + "ipient": 14117, + "ipients": 47647, + "iping": 46095, + "ipl": 10352, + "iple": 3778, + "iples": 37458, + "iplina": 82721, + "iplinary": 43450, + "ipline": 47429, + "ipmap": 50428, + "ipment": 5402, + "ipo": 6943, + "ipop": 79382, + "ipp": 2807, + "ipped": 6450, + "ippers": 31897, + "ippet": 21581, + "ippets": 42870, + "ippi": 24255, + "ippines": 63906, + "ipping": 5654, + "ipple": 51164, + "ipples": 50560, + "ippo": 87353, + "ippy": 45749, + "ipro": 48789, + "ips": 3077, + "ipse": 27332, + "ipsis": 47402, + "ipsoid": 96321, + "ipt": 11442, + "iptables": 95289, + "ipur": 71877, + "ipv": 42676, + "ipy": 22947, + "ipzig": 56200, + "iq": 23740, + "iqu": 5011, + "ique": 2372, + "iquement": 80135, + "iquer": 71584, + "iques": 8303, + "iqueta": 97110, + "iquid": 31989, + "iqué": 75266, + "ir": 404, + "ira": 8832, + "irable": 24966, + "iral": 36378, + "iram": 66708, + "iran": 54690, + "iras": 50085, + "irate": 88340, + "irates": 24557, + "iration": 19629, + "irc": 2437, + "ircle": 7074, + "ircles": 74263, + "ircon": 88276, + "ircraft": 12615, + "ircuit": 37268, + "ircular": 21472, + "ird": 2603, + "ire": 554, + "irebase": 13341, + "ireccion": 31505, + "irect": 1226, + "irected": 74612, + "irection": 2866, + "irectional": 44670, + "irector": 65476, + "irectory": 4758, + "ired": 2690, + "irement": 46568, + "iren": 46335, + "ires": 3861, + "irez": 59736, + "iri": 20679, + "irical": 66866, + "irie": 61216, + "irim": 61372, + "iring": 6191, + "iris": 41946, + "irit": 6906, + "iriÅŁ": 127045, + "irk": 14136, + "irl": 2811, + "irlines": 24536, + "irling": 50768, + "irm": 2853, + "irma": 44011, + "irmed": 8434, + "irmingham": 32600, + "irms": 14404, + "irmware": 53260, + "iro": 8698, + "iron": 2475, + "iropr": 54987, + "iros": 47228, + "irq": 16329, + "irror": 28812, + "irs": 16838, + "irsch": 81865, + "irse": 82614, + "irst": 864, + "irt": 2106, + "irteen": 44904, + "irth": 4809, + "irthday": 19951, + "irting": 66939, + "irts": 19666, + "irtschaft": 61811, + "irtual": 2901, + "irty": 16531, + "irus": 16972, + "irut": 86305, + "iry": 33770, + "irá": 83562, + "is": 285, + "isActive": 31814, + "isAdmin": 72658, + "isArray": 67113, + "isChecked": 53882, + "isContained": 80469, + "isEmpty": 18972, + "isEnabled": 74452, + "isEqual": 97886, + "isEqualTo": 32651, + "isFunction": 94144, + "isLoading": 54584, + "isLoggedIn": 93968, + "isNaN": 43958, + "isNew": 59333, + "isNull": 66811, + "isObject": 88633, + "isOk": 79274, + "isOpen": 63837, + "isRequired": 68542, + "isSelected": 21753, + "isValid": 29192, + "isVisible": 72632, + "isa": 10761, + "isable": 81485, + "isan": 17132, + "isans": 66399, + "isas": 85934, + "isateur": 32126, + "isation": 7923, + "isations": 55176, + "isbn": 66438, + "isbury": 84606, + "isc": 3427, + "iscal": 15787, + "iscard": 47560, + "isce": 89268, + "isch": 16053, + "ische": 10559, + "ischen": 17739, + "ischer": 32708, + "isches": 86957, + "isci": 92286, + "iscing": 52754, + "isclosed": 73690, + "isco": 9464, + "iscopal": 85205, + "iscrim": 41610, + "iscrimination": 85913, + "isd": 20640, + "isdiction": 25623, + "isdigit": 69829, + "ise": 1064, + "isease": 55307, + "isecond": 45866, + "iseconds": 16683, + "ised": 4056, + "isel": 65294, + "iselect": 74310, + "isempty": 46064, + "isen": 23129, + "iser": 12059, + "isers": 61280, + "ises": 4909, + "iset": 49189, + "iseum": 93465, + "isex": 75253, + "ish": 812, + "isha": 34940, + "ished": 3304, + "isher": 38572, + "ishes": 20527, + "ishi": 38376, + "ishing": 10976, + "ishlist": 48211, + "ishly": 64823, + "ishment": 16025, + "ishments": 98953, + "ishop": 20158, + "ishops": 50109, + "isi": 13229, + "isia": 47939, + "isible": 23066, + "isicing": 55946, + "isiert": 87481, + "isify": 90293, + "isin": 57404, + "ising": 3791, + "ision": 1816, + "isions": 6805, + "isis": 9479, + "isión": 70255, + "isk": 3187, + "iska": 48560, + "iske": 32209, + "iskey": 46090, + "isko": 68543, + "isks": 92834, + "isky": 70704, + "isl": 22457, + "islation": 58221, + "isle": 40105, + "ism": 2142, + "isma": 38152, + "isman": 62570, + "ismatch": 24976, + "ismatic": 56145, + "isme": 43518, + "ismet": 90158, + "ismic": 54094, + "ismo": 17001, + "isms": 13664, + "isnan": 78129, + "iso": 15420, + "isode": 6549, + "isodes": 14739, + "isoft": 61018, + "isol": 59923, + "ison": 3335, + "isoner": 97020, + "isons": 19379, + "isor": 15841, + "isors": 41214, + "isory": 85903, + "isos": 69304, + "isp": 13090, + "ispens": 53620, + "isper": 27370, + "ispers": 68835, + "isphere": 46794, + "ispiel": 68057, + "isplay": 2316, + "isposable": 46907, + "isque": 95543, + "iss": 1038, + "issa": 21429, + "issan": 33097, + "issance": 23965, + "issant": 65162, + "isse": 15398, + "issement": 49653, + "issen": 36964, + "issenschaft": 79748, + "issent": 78638, + "isser": 73233, + "isses": 84360, + "isset": 4888, + "isseur": 73508, + "issing": 13577, + "ission": 7558, + "issional": 89580, + "issions": 16528, + "issippi": 26035, + "isson": 48263, + "issor": 70350, + "issors": 41622, + "isspace": 81770, + "issue": 11159, + "issued": 58226, + "issuer": 66817, + "issues": 18427, + "issy": 60215, + "ist": 380, + "ista": 9087, + "istan": 8944, + "istance": 3924, + "istani": 39322, + "istant": 11202, + "istar": 83474, + "istas": 26673, + "iste": 16776, + "isted": 13236, + "istem": 66421, + "istema": 41448, + "isten": 37122, + "istence": 80524, + "istencia": 88200, + "istency": 47094, + "istent": 18128, + "ister": 1571, + "isters": 9303, + "istes": 74360, + "isti": 61517, + "istic": 4532, + "istica": 86259, + "istical": 39496, + "istically": 37110, + "istics": 5589, + "istik": 95393, + "istine": 53065, + "isting": 11083, + "istingu": 71169, + "istinguish": 88647, + "istinguished": 79282, + "istique": 96396, + "istle": 28300, + "istles": 65943, + "isto": 25339, + "istogram": 28499, + "istol": 91768, + "iston": 58819, + "istor": 5323, + "istorical": 95698, + "istory": 2579, + "istr": 3758, + "istra": 74963, + "istrar": 34765, + "istrate": 60301, + "istrates": 93100, + "istration": 4048, + "istrator": 18690, + "istream": 40633, + "istrib": 4789, + "istribute": 80133, + "istributed": 25146, + "istribution": 13107, + "istributions": 17994, + "istributor": 78388, + "istrict": 8391, + "istring": 73756, + "istringstream": 82416, + "istro": 15561, + "istros": 81950, + "istry": 4944, + "ists": 1671, + "istung": 83214, + "isty": 87523, + "isu": 62748, + "isure": 32137, + "iswa": 67220, + "isy": 28013, + "isz": 69728, + "isé": 67968, + "isée": 137123, + "it": 275, + "ita": 6255, + "itable": 15368, + "itably": 34668, + "itag": 94540, + "itage": 18443, + "itaire": 59124, + "ital": 2174, + "italic": 34406, + "italize": 34313, + "itals": 16265, + "itamin": 75552, + "itan": 12849, + "itant": 50944, + "itar": 12355, + "itarian": 20030, + "itary": 44806, + "itas": 21226, + "itat": 47777, + "itate": 19652, + "itated": 32237, + "itates": 81729, + "itating": 49544, + "itation": 7556, + "itational": 49824, + "itations": 30667, + "itative": 21312, + "itbart": 49199, + "itch": 1610, + "itched": 92403, + "itchen": 7454, + "itchens": 48621, + "itches": 25220, + "ite": 632, + "iteDatabase": 54251, + "itech": 60782, + "itect": 8360, + "itecture": 41134, + "ited": 1608, + "iteit": 68519, + "itel": 56643, + "itelist": 57645, + "item": 1203, + "itemId": 54382, + "itemName": 95720, + "itemap": 25330, + "itemid": 84127, + "itempty": 19918, + "items": 3615, + "iten": 14668, + "iter": 2015, + "iteral": 9953, + "iterals": 37802, + "iterate": 41512, + "iterated": 50486, + "iteration": 36722, + "iterations": 67584, + "iterator": 6854, + "itere": 41846, + "iterr": 33825, + "ites": 3611, + "itespace": 36074, + "itesse": 86860, + "itest": 97105, + "itet": 41538, + "iteur": 82948, + "itez": 54645, + "ith": 410, + "ithe": 82756, + "ither": 2485, + "ithmetic": 25922, + "ithub": 3827, + "iti": 12303, + "itia": 87475, + "itial": 1559, + "itian": 78444, + "itic": 47986, + "itical": 31260, + "itics": 46979, + "ities": 1361, + "itim": 126690, + "itimate": 70750, + "itime": 38504, + "itin": 84886, + "iting": 5853, + "ition": 680, + "itional": 3005, + "itionally": 17412, + "itioner": 71246, + "itions": 5930, + "itious": 64695, + "itis": 19435, + "itive": 3404, + "itives": 50354, + "itivity": 18055, + "itize": 26310, + "itized": 83443, + "itizen": 93134, + "itizer": 63818, + "itle": 1002, + "itled": 15263, + "itledBorder": 85121, + "itlement": 45337, + "itles": 41243, + "itm": 72284, + "itmap": 19244, + "itness": 8091, + "ito": 6357, + "itol": 26003, + "iton": 62437, + "itone": 84066, + "itor": 1919, + "itore": 92916, + "itored": 36201, + "itori": 93357, + "itories": 13427, + "itorio": 95986, + "itoris": 79289, + "itors": 11905, + "itory": 10618, + "itos": 25593, + "itous": 49755, + "itr": 21561, + "itra": 36429, + "itre": 48734, + "itrust": 84871, + "its": 1199, + "itsu": 49557, + "itt": 1442, + "itta": 45024, + "ittal": 82123, + "ittance": 87191, + "itte": 96673, + "itted": 3762, + "ittel": 95283, + "itten": 22449, + "itter": 3248, + "itters": 28063, + "ittest": 14267, + "itti": 71766, + "itting": 14810, + "ittings": 52259, + "ittle": 2377, + "itto": 86187, + "itty": 22814, + "itu": 32362, + "itud": 12842, + "itude": 3906, + "itudes": 20590, + "itulo": 46378, + "itunes": 84482, + "itung": 50158, + "itur": 51930, + "itura": 76766, + "iture": 9280, + "itures": 41104, + "itus": 35791, + "itution": 33690, + "ity": 487, + "ityEngine": 6082, + "itz": 11042, + "itzer": 20480, + "itzerland": 26423, + "itä": 64321, + "ität": 36614, + "ité": 12815, + "ités": 34833, + "itÃł": 23940, + "iu": 18738, + "ium": 2356, + "ius": 9156, + "iv": 344, + "iva": 9924, + "ivable": 42572, + "ival": 3936, + "ivalence": 88790, + "ivalent": 11769, + "ivals": 18233, + "ivamente": 56158, + "ivan": 27322, + "ivant": 96005, + "ivar": 56398, + "ivariate": 55422, + "ivas": 38824, + "ivate": 731, + "ivated": 54686, + "ivating": 64143, + "ivation": 39127, + "ivative": 47158, + "ive": 533, + "iveau": 39219, + "ivec": 68580, + "ived": 2221, + "ivel": 20528, + "ively": 3132, + "ivement": 90540, + "iven": 2071, + "iveness": 12781, + "ivent": 80441, + "iver": 1524, + "ivered": 43056, + "ivering": 83940, + "ivers": 1945, + "iversal": 33952, + "iversary": 17577, + "iverse": 8034, + "iversity": 2977, + "ivery": 6497, + "ives": 1886, + "ivet": 98654, + "ivi": 32820, + "ivia": 20809, + "ivial": 26658, + "ivic": 39406, + "ivicrm": 96752, + "ivid": 63923, + "ividad": 94166, + "ividual": 16826, + "ivil": 6230, + "iving": 2249, + "ivirus": 57913, + "ivism": 72950, + "ivist": 79144, + "ivities": 42379, + "ivity": 1927, + "ivityManager": 69557, + "ivité": 142566, + "ivitÃł": 78443, + "ivo": 6496, + "ivol": 80506, + "ivor": 49675, + "ivors": 85393, + "ivos": 24224, + "ivot": 16084, + "ivr": 57766, + "ivre": 83495, + "ivy": 64505, + "iw": 37081, + "ix": 941, + "ixa": 46384, + "ixe": 85300, + "ixed": 3286, + "ixedReality": 56804, + "ixel": 5941, + "ixels": 10274, + "ixer": 39014, + "ixin": 20456, + "ixmap": 24593, + "ixo": 51362, + "ixon": 28472, + "ixture": 12735, + "iy": 16220, + "iya": 79474, + "iyor": 83342, + "iz": 449, + "iza": 17090, + "izabeth": 18540, + "izable": 8335, + "izacao": 97190, + "izacion": 51470, + "ización": 41500, + "izada": 51211, + "izado": 25846, + "izador": 90589, + "izados": 72158, + "izando": 91858, + "izar": 11155, + "izard": 13722, + "izards": 34091, + "izarre": 29482, + "ization": 2022, + "izational": 76391, + "izations": 8040, + "ização": 47376, + "ize": 551, + "ized": 1506, + "izedName": 20167, + "izen": 23678, + "izens": 28960, + "izer": 3135, + "izers": 12230, + "izes": 4756, + "izi": 33235, + "izia": 62041, + "izin": 76971, + "izing": 4849, + "izio": 44982, + "izione": 38361, + "izioni": 78159, + "izo": 42894, + "izon": 16973, + "izona": 15371, + "izons": 63847, + "izont": 6186, + "izontal": 6506, + "izontally": 50401, + "izoph": 45178, + "izophren": 45312, + "izr": 65362, + "izu": 91615, + "izz": 8759, + "izza": 13741, + "izzard": 38148, + "izzare": 76598, + "izzas": 67304, + "izzato": 91838, + "izzazione": 86108, + "izzer": 93329, + "izzes": 58327, + "izzie": 87987, + "izziness": 96645, + "izzle": 43137, + "izzlies": 94695, + "izzling": 89119, + "izzly": 72725, + "izzo": 60206, + "izzy": 90491, + "ião": 77328, + "ição": 38080, + "ième": 89172, + "ière": 20108, + "ièrement": 66782, + "ières": 58207, + "ié": 138522, + "ién": 23538, + "iéndo": 138523, + "iêu": 132728, + "ië": 127180, + "ió": 34894, + "ión": 3655, + "iÃŁ": 36947, + "iÄħ": 46892, + "iÄĩ": 67829, + "iÄį": 83095, + "iÄĻ": 62602, + "iÄĻÄĩ": 138476, + "iÄŁi": 127625, + "iÅĤ": 130274, + "iÅŁ": 97802, + "iếc": 135075, + "iết": 133545, + "iếu": 141010, + "iá»ĥu": 137444, + "iá»ĩm": 141071, + "iá»ĩn": 133682, + "iá»ĩu": 130241, + "j": 73, + "jQuery": 28004, + "ja": 5580, + "jab": 38916, + "jabi": 91884, + "jac": 86131, + "jack": 27134, + "jad": 62086, + "jadi": 92046, + "jah": 97396, + "jak": 62402, + "jal": 47692, + "jam": 43711, + "jamin": 25246, + "jan": 22838, + "jandro": 72059, + "jang": 37527, + "jango": 5846, + "jar": 19033, + "jas": 29726, + "jav": 61459, + "java": 10042, + "javascript": 14073, + "javax": 45148, + "jaw": 73992, + "jax": 14518, + "jay": 70037, + "jb": 38978, + "jc": 62987, + "jd": 43390, + "jdbc": 52132, + "jdk": 88851, + "je": 3756, + "jec": 73201, + "ject": 583, + "jected": 28303, + "jection": 7606, + "jections": 24575, + "jective": 50491, + "jectives": 86578, + "jectories": 65572, + "jectory": 23363, + "jee": 67896, + "jej": 76810, + "jejer": 56932, + "jem": 62903, + "jen": 23166, + "jenis": 56520, + "jer": 26360, + "jerne": 75347, + "jes": 20645, + "jest": 49015, + "jet": 15407, + "jeta": 64234, + "jets": 73904, + "jf": 96423, + "jh": 74306, + "ji": 7754, + "jian": 81891, + "jiang": 63547, + "jid": 72876, + "jie": 96237, + "jin": 49800, + "jing": 98168, + "jis": 82726, + "jit": 58809, + "jj": 38811, + "jk": 41808, + "jl": 43870, + "jm": 65939, + "jmp": 61055, + "jn": 93808, + "jni": 79114, + "jo": 7305, + "job": 8799, + "jobs": 31105, + "john": 47817, + "join": 5987, + "joined": 71154, + "joining": 65205, + "joint": 32850, + "jom": 60218, + "jon": 34165, + "jong": 98460, + "jos": 39161, + "jour": 29262, + "jourd": 68472, + "journal": 42518, + "jours": 39806, + "joy": 4123, + "jp": 46054, + "jpeg": 30921, + "jpg": 17974, + "jq": 43915, + "jquery": 26836, + "jr": 58913, + "js": 2519, + "jsc": 75527, + "json": 2236, + "jsonData": 91803, + "jsonp": 57045, + "jsonwebtoken": 76474, + "jsp": 93340, + "jspb": 92947, + "jspx": 30477, + "jsx": 45290, + "jt": 57205, + "ju": 8613, + "jual": 65029, + "jualan": 81925, + "juan": 94706, + "juana": 13264, + "jud": 18138, + "judge": 92361, + "judul": 67612, + "jug": 47702, + "jumbotron": 89334, + "jumlah": 48616, + "jump": 43296, + "jun": 57681, + "junction": 72428, + "jur": 86877, + "jure": 52302, + "jured": 79391, + "juries": 89827, + "jury": 41761, + "just": 4250, + "justice": 38768, + "justify": 34706, + "juven": 62131, + "jwt": 41592, + "jylland": 75161, + "jähr": 135453, + "jährige": 135454, + "jÃł": 43324, + "jÄħ": 36837, + "jÄħc": 125749, + "jÄħce": 127269, + "jÄħcego": 133123, + "jÄħcej": 135225, + "jÄħcy": 133188, + "jÄħcych": 129954, + "jÄħcym": 134200, + "jÄħcÄħ": 136413, + "jÄĻ": 35310, + "jÅĽÄĩ": 136981, + "k": 74, + "kB": 81524, + "kHz": 61756, + "kJ": 42330, + "ka": 4554, + "kaar": 87151, + "kad": 93275, + "kah": 83502, + "kal": 35541, + "kan": 8656, + "kap": 38834, + "kar": 28134, + "kart": 74847, + "kas": 62899, + "kat": 33755, + "kategori": 67447, + "kaza": 133821, + "kazaÅĤ": 133822, + "kb": 21310, + "kbd": 48456, + "kc": 31378, + "kd": 37744, + "kdir": 12438, + "ke": 440, + "kea": 42428, + "ked": 47159, + "kee": 21822, + "keep": 13096, + "keeper": 18861, + "keepers": 55538, + "keeping": 32394, + "kees": 40748, + "keh": 70617, + "kehr": 81112, + "keit": 26612, + "keiten": 51707, + "kek": 85207, + "kel": 17659, + "kelas": 89952, + "keletal": 85155, + "keleton": 27094, + "keley": 27299, + "kelig": 85010, + "ken": 2712, + "kening": 89273, + "kept": 97920, + "ker": 7052, + "kerja": 74918, + "kernel": 23248, + "kers": 26177, + "kes": 12556, + "kest": 39323, + "ket": 25475, + "keterangan": 85325, + "ketøy": 80785, + "key": 792, + "keyCode": 91367, + "keyboard": 41713, + "keydown": 44191, + "keypress": 60406, + "keys": 10563, + "keyup": 37224, + "keyword": 19863, + "keywords": 28995, + "kf": 81869, + "kg": 7351, + "kh": 30664, + "ki": 6642, + "kich": 70776, + "kick": 55793, + "kid": 63281, + "kids": 73896, + "kie": 30322, + "kiego": 93737, + "kiem": 87150, + "kil": 85526, + "kill": 10851, + "killer": 72078, + "kills": 60925, + "kim": 57893, + "kin": 7989, + "kind": 15314, + "king": 10566, + "kins": 11704, + "kinson": 50447, + "kip": 13389, + "kir": 67931, + "kish": 73620, + "kit": 8226, + "kits": 89417, + "kiye": 95203, + "kj": 92697, + "kk": 19196, + "kke": 90761, + "kker": 52598, + "kl": 10561, + "klass": 54422, + "kle": 23089, + "klä": 63730, + "klär": 77131, + "km": 16017, + "kn": 19095, + "knife": 42720, + "know": 32034, + "knowledge": 89053, + "known": 5278, + "ko": 9325, + "koa": 88288, + "kode": 37136, + "kok": 45264, + "kol": 51452, + "kom": 22097, + "kommen": 42789, + "kon": 39537, + "koneksi": 88782, + "kont": 39720, + "kontakte": 58868, + "kop": 83598, + "kor": 40717, + "kos": 56678, + "kov": 51667, + "kowski": 70124, + "koÅĦ": 126852, + "koÅĦcz": 130302, + "koÅĦczyÅĤ": 141213, + "kp": 48495, + "kr": 9855, + "kraine": 77736, + "kre": 54942, + "kreÅĽl": 135099, + "krit": 87550, + "krä": 144004, + "kräfte": 144005, + "krÄĻ": 141698, + "krÄĻc": 141699, + "ks": 2787, + "ksam": 91647, + "ksen": 85261, + "kses": 54009, + "ksi": 31060, + "ksiÄħż": 143972, + "ksiÄħżka": 143973, + "ksz": 139438, + "ksztaÅĤ": 139439, + "ksztaÅĤc": 139440, + "kt": 5840, + "kte": 40951, + "ktion": 39470, + "ktop": 8617, + "ktor": 75278, + "ktör": 127334, + "ku": 12133, + "kube": 97717, + "kubectl": 79971, + "kul": 90195, + "kur": 68317, + "kus": 78466, + "kv": 43408, + "kw": 28600, + "kwargs": 9674, + "ky": 7891, + "ków": 93457, + "kö": 73052, + "köp": 137316, + "körper": 142223, + "köy": 131751, + "kü": 129056, + "kür": 127356, + "kı": 126348, + "kıl": 132544, + "kır": 130071, + "kÄħ": 128578, + "kÄĻ": 84444, + "kÅĤad": 126416, + "l": 75, + "la": 4260, + "lab": 14380, + "label": 1502, + "labels": 16873, + "labs": 70271, + "lac": 92429, + "lace": 26536, + "laces": 76588, + "lad": 18107, + "laden": 88551, + "ladesh": 34121, + "ladı": 126763, + "ladıģı": 133786, + "lag": 13363, + "lage": 55754, + "lagen": 54708, + "lags": 22851, + "lah": 14992, + "lahoma": 22088, + "lain": 51971, + "lake": 62110, + "lam": 23771, + "lama": 80001, + "lambda": 12935, + "lamp": 94757, + "lan": 10715, + "land": 1933, + "lander": 34003, + "landers": 84967, + "landing": 48341, + "lando": 23693, + "lands": 8166, + "landscape": 88391, + "landı": 131181, + "landır": 128112, + "landırma": 142498, + "landırıl": 140119, + "lane": 37847, + "lang": 5205, + "langle": 70800, + "langs": 63917, + "language": 11528, + "languages": 42824, + "lanması": 135501, + "lanmÄ±ÅŁ": 137736, + "lap": 15958, + "lapping": 90619, + "laps": 89722, + "lar": 13796, + "lararas": 128171, + "lararası": 128172, + "lard": 133946, + "lardır": 133947, + "larg": 26897, + "large": 16767, + "ları": 45837, + "larımız": 129545, + "ların": 95448, + "larına": 128767, + "larınd": 129956, + "larında": 128759, + "larından": 129957, + "larını": 128445, + "larının": 128650, + "larınız": 127882, + "larınızı": 136810, + "larıyla": 133006, + "las": 14493, + "lash": 24686, + "lashes": 46545, + "lass": 447, + "lassen": 70489, + "lasses": 33868, + "lassian": 89597, + "last": 4259, + "lastName": 29156, + "lastic": 51179, + "lasting": 69429, + "lastname": 29206, + "lat": 5524, + "latable": 91147, + "late": 5075, + "lated": 22076, + "latent": 50023, + "later": 67576, + "lates": 48246, + "latesAutoresizingMaskIntoConstraints": 60107, + "latest": 19350, + "latex": 64680, + "latin": 60085, + "lation": 2302, + "lations": 29656, + "latitude": 23718, + "latlong": 61499, + "lator": 21829, + "lauf": 97761, + "laughs": 72456, + "laughter": 51783, + "launch": 33499, + "launcher": 77191, + "laus": 57229, + "lav": 78836, + "law": 19915, + "laws": 67537, + "lay": 6651, + "layan": 128045, + "layer": 10333, + "layers": 44198, + "layout": 8399, + "layouts": 20054, + "lays": 63359, + "layui": 41335, + "layın": 138649, + "lazy": 49013, + "laÅŁ": 124447, + "laÅŁma": 127697, + "laÅŁtır": 137850, + "laÅŁtırma": 142099, + "lb": 21123, + "lbl": 31998, + "lbrace": 75233, + "lbrakk": 74161, + "lbs": 52944, + "lc": 17257, + "lcd": 68416, + "ld": 507, + "lda": 73426, + "ldap": 38665, + "ldata": 52681, + "ldb": 56925, + "lder": 76869, + "ldr": 72377, + "ldre": 63196, + "lds": 71033, + "le": 273, + "leDb": 89506, + "lea": 88291, + "lead": 26060, + "leader": 37391, + "leaders": 78386, + "leading": 20654, + "leads": 98370, + "leaf": 24153, + "leaflet": 81707, + "league": 46731, + "lean": 2675, + "leaning": 58951, + "leanor": 76680, + "leans": 21708, + "leanup": 60639, + "lear": 1547, + "leared": 92333, + "learn": 12675, + "learner": 93595, + "learning": 20981, + "lease": 1623, + "leased": 4673, + "leases": 28299, + "leasing": 83552, + "least": 55271, + "leave": 21510, + "lec": 34753, + "leccion": 42630, + "leck": 93876, + "lect": 767, + "lected": 2209, + "lectic": 67449, + "lection": 1170, + "lections": 67679, + "lector": 27669, + "lectric": 46162, + "lectron": 69229, + "lectual": 58417, + "lecture": 76761, + "led": 832, + "ledge": 4890, + "ledged": 92624, + "ledger": 50704, + "ledo": 60655, + "ledon": 82457, + "lee": 8501, + "leen": 39662, + "leep": 3499, + "leet": 18973, + "leetcode": 35326, + "lef": 59238, + "left": 2359, + "leftJoin": 75661, + "leftright": 80375, + "leftrightarrow": 81819, + "leg": 1937, + "legacy": 39884, + "legal": 6428, + "legalArgumentException": 97009, + "legant": 58802, + "legate": 4989, + "legates": 53379, + "legation": 87566, + "lege": 4757, + "legen": 48414, + "legend": 14505, + "legg": 83028, + "leground": 64766, + "legs": 72236, + "legt": 71168, + "leh": 73790, + "lehem": 87572, + "lei": 62785, + "leich": 54371, + "leigh": 62969, + "lein": 78369, + "leine": 91833, + "leitung": 74462, + "lek": 75776, + "lem": 3433, + "lemen": 41307, + "lement": 986, + "lements": 90455, + "lemetry": 35958, + "lemma": 9633, + "lemn": 61030, + "lems": 39616, + "len": 2892, + "lena": 48021, + "lename": 29118, + "lene": 34191, + "leneck": 59013, + "leness": 76219, + "length": 4129, + "lenmiÅŁ": 140313, + "leo": 81763, + "leon": 79398, + "lep": 62967, + "ler": 1536, + "leri": 46779, + "lernen": 78877, + "lero": 65311, + "lers": 9254, + "les": 642, + "lesai": 97531, + "lesc": 63531, + "lescope": 38113, + "lesen": 56027, + "lesh": 17469, + "leshoot": 96408, + "leshooting": 50240, + "lesi": 75895, + "lesia": 92237, + "less": 1717, + "lessly": 15740, + "lessness": 31928, + "lesson": 27495, + "lessons": 72312, + "lest": 59144, + "leston": 45568, + "let": 1149, + "leta": 48151, + "letal": 45490, + "letcher": 60520, + "lete": 1617, + "leted": 6873, + "letes": 13881, + "letic": 21973, + "letics": 43028, + "leting": 35935, + "letion": 52625, + "leton": 15616, + "lets": 9942, + "lett": 27064, + "lette": 9809, + "letter": 9451, + "letters": 21053, + "lettes": 65592, + "leur": 34977, + "leurs": 76560, + "lev": 3449, + "levance": 32094, + "levant": 8367, + "levard": 42243, + "levation": 43757, + "levator": 79200, + "level": 3294, + "leveland": 21478, + "levelname": 75641, + "levels": 42564, + "lever": 71805, + "levision": 10864, + "lew": 74964, + "lex": 2571, + "lexer": 38815, + "lexible": 94544, + "lexical": 93364, + "lexport": 83353, + "ley": 3179, + "leys": 47679, + "lez": 40781, + "leÅŁ": 124968, + "leÅŁme": 127495, + "leÅŁtir": 125269, + "leÅŁtirme": 139670, + "lf": 11008, + "lfw": 96106, + "lg": 11819, + "lh": 63045, + "lhs": 31895, + "li": 742, + "lia": 23350, + "liable": 50330, + "liament": 12403, + "lian": 91798, + "lias": 71829, + "lib": 2740, + "libc": 55576, + "libft": 56523, + "libraries": 63049, + "library": 18065, + "libs": 35719, + "lic": 415, + "lica": 15317, + "licable": 46114, + "lical": 31248, + "licant": 35237, + "licants": 61832, + "licas": 52210, + "licate": 48795, + "licated": 13724, + "lication": 1693, + "lications": 10709, + "licative": 65266, + "lice": 4754, + "liced": 74630, + "licence": 93206, + "license": 13266, + "licensed": 71711, + "licenses": 77914, + "licer": 97243, + "lices": 37414, + "lich": 6772, + "liche": 20005, + "lichen": 21673, + "licher": 50095, + "liches": 95285, + "lichkeit": 88185, + "licht": 37633, + "licing": 89114, + "licit": 6026, + "licity": 24779, + "lick": 1206, + "licken": 93790, + "lickr": 31803, + "lico": 36011, + "licos": 88844, + "lict": 21242, + "licted": 41343, + "licting": 78208, + "licts": 56445, + "lid": 61206, + "lide": 16655, + "lider": 19118, + "lie": 11567, + "lien": 85036, + "lient": 1451, + "lients": 7041, + "lier": 5623, + "liers": 40376, + "lies": 48101, + "liest": 23138, + "lieÃŁ": 56400, + "lieÃŁlich": 83090, + "life": 14450, + "lifetime": 97057, + "lift": 34969, + "lifting": 51903, + "lify": 55500, + "lig": 7708, + "liga": 44870, + "lige": 52790, + "light": 4145, + "lightbox": 98504, + "lightly": 79498, + "lights": 13826, + "ligne": 90030, + "ligt": 58156, + "lij": 22953, + "lijah": 75368, + "lijk": 47292, + "lijke": 81934, + "lik": 7637, + "like": 4803, + "liked": 43236, + "likelihood": 61130, + "likely": 14282, + "likes": 25039, + "lim": 4659, + "lime": 38943, + "limit": 9506, + "limitations": 73168, + "limited": 18235, + "limits": 41078, + "lin": 3732, + "line": 1056, + "lineEdit": 92547, + "linear": 22763, + "lined": 15121, + "lineno": 79983, + "liner": 61798, + "lines": 7969, + "liness": 48858, + "linewidth": 57212, + "ling": 2718, + "lingen": 76307, + "linger": 71560, + "lings": 18812, + "lington": 42647, + "linha": 66820, + "linik": 79870, + "lining": 34881, + "link": 2080, + "linked": 43133, + "linkedin": 76170, + "linkplain": 68640, + "links": 15880, + "linky": 51858, + "lint": 14957, + "linux": 14210, + "lio": 25147, + "lion": 78151, + "lip": 33115, + "liqu": 39993, + "lique": 36426, + "liquid": 53637, + "lis": 47203, + "lish": 1672, + "list": 1607, + "lista": 32660, + "liste": 64646, + "listed": 31240, + "listen": 36337, + "listener": 35039, + "listeners": 76950, + "listening": 74819, + "listing": 34596, + "lists": 17575, + "lit": 31635, + "lite": 68078, + "liter": 68091, + "literal": 35785, + "little": 55392, + "lius": 86419, + "liv": 33567, + "live": 24822, + "living": 63175, + "liwo": 141935, + "liwoÅĽci": 141936, + "lix": 15051, + "lixir": 97772, + "liÄŁ": 125364, + "liÄŁe": 139261, + "liÄŁi": 126326, + "liÄŁin": 133752, + "liÄŁinde": 137161, + "liÄŁini": 133350, + "liÅĽmy": 129083, + "lj": 52735, + "lk": 41748, + "lke": 126732, + "ll": 654, + "lla": 35782, + "lland": 73412, + "lld": 32459, + "lle": 65501, + "llen": 79788, + "ller": 54956, + "lli": 39976, + "llib": 20745, + "llu": 60749, + "llum": 5448, + "lluminate": 5585, + "llvm": 29042, + "llx": 94990, + "lm": 17771, + "ln": 2261, + "lname": 50228, + "lng": 20810, + "lo": 385, + "load": 1078, + "loadModel": 93931, + "loaded": 15589, + "loader": 8355, + "loading": 10628, + "loads": 32685, + "loan": 38329, + "loat": 1239, + "loating": 28493, + "lob": 1684, + "lobal": 2042, + "lobals": 16616, + "lobber": 91397, + "lobby": 71775, + "lobe": 62997, + "lobs": 68164, + "loc": 1074, + "local": 2438, + "localObject": 53207, + "localStorage": 43453, + "localctx": 54474, + "locale": 14484, + "locales": 80232, + "localhost": 8301, + "localized": 71594, + "locals": 64130, + "locate": 22401, + "located": 39463, + "location": 2527, + "locations": 31309, + "locator": 68033, + "lock": 1023, + "locked": 24633, + "locker": 93414, + "locking": 89482, + "locks": 34936, + "locs": 83629, + "lod": 82482, + "lodash": 30279, + "loe": 34498, + "loff": 93038, + "log": 839, + "logan": 93925, + "logen": 85845, + "logfile": 75566, + "logg": 93369, + "logged": 24867, + "loggedIn": 80149, + "loggedin": 96227, + "logger": 9786, + "logging": 25263, + "logic": 24225, + "logical": 30256, + "login": 3673, + "logo": 10129, + "logout": 15937, + "logradouro": 67840, + "logs": 22081, + "logue": 67078, + "logy": 35681, + "loh": 82348, + "loi": 66392, + "loid": 51096, + "loit": 78549, + "lok": 60931, + "lol": 59319, + "lon": 12212, + "long": 4825, + "longitude": 25446, + "longleftrightarrow": 87459, + "longrightarrow": 55042, + "loo": 47133, + "lood": 4556, + "look": 7201, + "looking": 20363, + "looks": 94273, + "lookup": 21020, + "loom": 18474, + "loomberg": 90072, + "loon": 31335, + "loop": 10498, + "loops": 82451, + "loor": 3993, + "lop": 22288, + "lope": 35327, + "lopedia": 42373, + "lopen": 41899, + "lops": 53689, + "loquent": 20522, + "lor": 9566, + "lord": 25598, + "lords": 47765, + "lore": 20156, + "los": 2301, + "lose": 1469, + "losed": 9259, + "loser": 69215, + "loses": 49341, + "losing": 17831, + "losion": 40455, + "loss": 9379, + "lossen": 76166, + "lost": 54337, + "losure": 11653, + "losures": 41206, + "lot": 9184, + "loth": 69552, + "lots": 65776, + "lotte": 24573, + "loud": 52178, + "lov": 35147, + "lovak": 86110, + "love": 30053, + "lover": 56497, + "low": 10303, + "lower": 14772, + "lowest": 89898, + "lox": 55728, + "loy": 1989, + "loyd": 25796, + "loyment": 39130, + "loys": 33298, + "lp": 13545, + "lpVtbl": 55138, + "lparr": 88887, + "lr": 19018, + "lrt": 55868, + "ls": 4730, + "lsa": 77683, + "lse": 87782, + "lsen": 59988, + "lsi": 65533, + "lsru": 90655, + "lsruhe": 93409, + "lst": 36687, + "lsx": 29017, + "lt": 4832, + "lte": 48791, + "ltk": 43671, + "ltr": 67375, + "ltra": 39369, + "ltre": 78758, + "ltür": 127667, + "lu": 9835, + "lua": 27623, + "lucent": 59054, + "luck": 85376, + "lude": 52538, + "luent": 56067, + "luet": 56316, + "luetooth": 57076, + "lug": 43213, + "lund": 84733, + "lung": 37949, + "lus": 82645, + "lush": 34604, + "lust": 70979, + "lut": 86854, + "lux": 62859, + "luÄŁu": 131977, + "luÄŁun": 136253, + "lv": 21827, + "lvl": 58160, + "lw": 57053, + "lx": 31447, + "ly": 398, + "lya": 96099, + "lycer": 67490, + "lying": 6711, + "lymp": 13842, + "lyn": 18013, + "lyph": 9525, + "lyphicon": 13684, + "lys": 60392, + "lz": 95505, + "lá": 129788, + "lâ": 131618, + "lä": 43183, + "läss": 139376, + "lässig": 139377, + "läuft": 141136, + "lève": 137379, + "lé": 68228, + "lée": 133607, + "légi": 137728, + "lóg": 142312, + "lógica": 142313, + "lö": 131807, + "lös": 141850, + "lösung": 141851, + "lü": 125462, + "lük": 130343, + "lÃ¼ÄŁÃ¼": 132489, + "lÃł": 132368, + "lÃŃ": 128910, + "lı": 25611, + "lık": 91733, + "lıkl": 126108, + "lıklar": 135388, + "lıkları": 135347, + "lıģ": 131236, + "lıģa": 138386, + "lıģı": 125508, + "lıģın": 131237, + "lıģını": 133182, + "m": 76, + "mA": 69265, + "mAh": 78535, + "mPid": 83779, + "ma": 1728, + "maal": 78791, + "mac": 11948, + "macen": 64601, + "mach": 54277, + "machine": 32056, + "macro": 32606, + "macros": 86398, + "mad": 20302, + "made": 26912, + "maf": 98061, + "mag": 30506, + "mage": 52310, + "magic": 37451, + "magnitude": 96264, + "mah": 74119, + "mai": 76485, + "maid": 45242, + "maids": 97150, + "mail": 3711, + "mailbox": 92447, + "mailer": 58403, + "mailto": 38526, + "main": 3817, + "maint": 66813, + "maintenance": 98078, + "mainwindow": 41580, + "maj": 95195, + "major": 36505, + "mak": 48562, + "make": 6927, + "maker": 25766, + "makers": 19887, + "makes": 79685, + "making": 27836, + "makt": 126389, + "maktad": 126390, + "maktadır": 126391, + "mal": 14649, + "male": 36476, + "malink": 46618, + "mall": 29532, + "malloc": 16166, + "mallow": 74380, + "malı": 127245, + "malıdır": 133088, + "man": 1515, + "mana": 47476, + "manage": 25117, + "managed": 25270, + "managedType": 65248, + "management": 42473, + "manager": 13297, + "mand": 1928, + "manda": 35545, + "mandatory": 80116, + "mando": 79744, + "manent": 32754, + "mani": 77198, + "mania": 51568, + "manifest": 42315, + "mann": 17562, + "mans": 69891, + "manship": 57732, + "mant": 41451, + "manual": 19730, + "manuel": 36164, + "manufact": 97450, + "manufacturer": 60914, + "many": 34576, + "manın": 132549, + "manız": 133370, + "map": 2186, + "mapped": 65371, + "mapper": 38076, + "mapping": 40792, + "maps": 17640, + "mar": 5612, + "marca": 84718, + "mare": 27655, + "mares": 62536, + "margin": 8936, + "marginLeft": 81637, + "marginTop": 66541, + "maries": 89333, + "marine": 84443, + "mark": 3987, + "markdown": 60073, + "marked": 46362, + "marker": 27742, + "markers": 60773, + "market": 19301, + "marketing": 91818, + "markets": 59460, + "marks": 15544, + "markt": 84590, + "markup": 80827, + "married": 69482, + "mars": 94926, + "marsh": 36239, + "marshal": 27121, + "marshall": 84161, + "marshaller": 52541, + "mart": 33472, + "marvin": 50276, + "mary": 1534, + "mas": 6899, + "mask": 11258, + "masked": 81535, + "mass": 26331, + "massage": 78585, + "mast": 96513, + "master": 13629, + "masters": 63611, + "ması": 75924, + "masına": 132585, + "masını": 126890, + "mat": 8470, + "match": 6347, + "matchCondition": 23972, + "matched": 36734, + "matcher": 70826, + "matches": 19914, + "matching": 89043, + "mate": 18052, + "material": 8243, + "materials": 77572, + "mates": 16457, + "math": 10374, + "mathrm": 91550, + "matic": 37244, + "matplotlib": 80427, + "matrix": 18110, + "matter": 58965, + "max": 2810, + "maxLength": 60992, + "maxcdn": 43273, + "maximum": 39187, + "maxlength": 50456, + "may": 18358, + "mayacaģı": 140645, + "maybe": 36760, + "mayı": 130298, + "mayın": 136469, + "maze": 97950, + "mb": 3096, + "mbH": 31098, + "mc": 12887, + "md": 2277, + "mdat": 80621, + "mdb": 78127, + "mdi": 40423, + "mdir": 69695, + "mdl": 47214, + "mdp": 90669, + "me": 2660, + "meal": 41780, + "mean": 14287, + "meaning": 56765, + "means": 46221, + "measure": 47799, + "measurement": 81425, + "med": 2061, + "medi": 4404, + "media": 7399, + "median": 55651, + "mediate": 14636, + "mediately": 43187, + "medical": 68116, + "medicine": 82443, + "medium": 26086, + "meer": 89925, + "meet": 63410, + "meeting": 61249, + "meg": 70276, + "mega": 42510, + "mek": 73753, + "mel": 27127, + "meld": 36601, + "meldung": 87697, + "melon": 71118, + "mem": 10536, + "member": 9597, + "memberOf": 68981, + "memberof": 49592, + "members": 22974, + "membership": 69328, + "memcmp": 61534, + "memcpy": 81084, + "memo": 55409, + "memory": 17269, + "memset": 98066, + "men": 5676, + "meno": 69091, + "mens": 48874, + "mensagem": 88841, + "mensaje": 46345, + "ment": 478, + "mental": 92121, + "mentation": 31099, + "mente": 12541, + "mented": 26980, + "mention": 29248, + "mentioned": 36591, + "mentions": 97355, + "mentor": 97645, + "ments": 1368, + "menu": 6064, + "menuItem": 88932, + "menus": 57627, + "mer": 1174, + "merc": 64422, + "merce": 10190, + "merchant": 39011, + "mercial": 18839, + "mere": 59931, + "merge": 19052, + "merged": 40354, + "meric": 2117, + "merican": 15717, + "mers": 22780, + "mes": 8828, + "mesh": 23506, + "mess": 49746, + "message": 1994, + "messages": 16325, + "met": 4059, + "meta": 5490, + "metadata": 17637, + "metal": 54008, + "metatable": 93589, + "meteor": 97774, + "meter": 59422, + "meth": 68891, + "method": 4393, + "methodName": 58550, + "methodPointerType": 95556, + "methodVisitor": 24962, + "methods": 35742, + "metic": 17990, + "metics": 45266, + "metis": 56944, + "metric": 15903, + "metrical": 58302, + "metrics": 43262, + "metro": 57047, + "metros": 67116, + "metry": 32242, + "meye": 142635, + "meyeceÄŁi": 142636, + "mez": 82156, + "mf": 28124, + "mg": 12311, + "mgr": 48292, + "mh": 51916, + "mi": 8155, + "mia": 90199, + "mic": 21220, + "micro": 40443, + "mid": 16106, + "midd": 84649, + "middle": 19656, + "middleware": 26624, + "middlewares": 88632, + "midi": 96519, + "midt": 42301, + "mie": 73596, + "mieÅĽci": 134938, + "might": 44968, + "mighty": 67305, + "migration": 80227, + "mil": 72270, + "mile": 66597, + "mill": 25013, + "million": 58313, + "milliseconds": 60772, + "mime": 49588, + "min": 1065, + "mina": 75435, + "minate": 34016, + "minated": 51199, + "minating": 64283, + "mination": 32096, + "mind": 37724, + "mine": 5967, + "minecraft": 72538, + "minent": 46407, + "miner": 92501, + "ming": 5311, + "mini": 36411, + "minimal": 92607, + "minimum": 37988, + "mino": 31824, + "minor": 45670, + "mins": 46696, + "minster": 40378, + "mint": 67791, + "minus": 38365, + "minute": 42557, + "minutes": 37037, + "mir": 19936, + "mirror": 73225, + "mis": 33865, + "misc": 47240, + "mise": 73206, + "misión": 132884, + "miss": 1831, + "missible": 44064, + "missing": 30616, + "mission": 2728, + "missions": 5176, + "missive": 56928, + "missão": 132192, + "mist": 78924, + "mit": 1763, + "mite": 52895, + "mites": 91499, + "mith": 43674, + "mits": 44703, + "mitt": 23692, + "mitted": 5483, + "mittel": 91184, + "mitter": 16126, + "mitters": 82089, + "mium": 50218, + "mix": 35071, + "mixed": 56685, + "miÄĻ": 131281, + "miÅŁ": 124801, + "miÅŁti": 132747, + "miÅŁtir": 129479, + "mj": 77830, + "mk": 24452, + "mkdir": 26117, + "ml": 1014, + "mland": 85694, + "mlin": 40926, + "mlink": 44243, + "mlx": 66903, + "mm": 3821, + "mma": 15895, + "mmas": 89536, + "mmc": 92008, + "mmm": 48886, + "mmo": 77675, + "mn": 21775, + "mnop": 63500, + "mnt": 40882, + "mo": 6355, + "mob": 33935, + "mobile": 14933, + "mobx": 88415, + "mock": 16712, + "mod": 2593, + "modal": 5667, + "modation": 68511, + "mode": 8516, + "model": 2528, + "modelName": 93374, + "modele": 76925, + "modelo": 78451, + "models": 6507, + "modern": 49789, + "modes": 89822, + "modifiable": 84079, + "modification": 97596, + "modified": 27162, + "modifier": 49742, + "modifiers": 91090, + "modify": 34053, + "modity": 51202, + "modo": 75943, + "mods": 60790, + "module": 4352, + "moduleId": 85843, + "moduleName": 73340, + "modules": 11525, + "modulo": 87966, + "modx": 95630, + "moid": 45059, + "moil": 60047, + "moire": 96395, + "mojom": 82336, + "mol": 44344, + "mom": 60774, + "moment": 28599, + "mon": 1645, + "monary": 54792, + "mond": 12388, + "monds": 54620, + "money": 28259, + "mong": 71155, + "mongo": 53070, + "mongodb": 37197, + "mongoose": 22774, + "monic": 73589, + "monitor": 32225, + "monkey": 96016, + "mono": 56401, + "mons": 23570, + "monster": 50519, + "mont": 20416, + "month": 10249, + "monthly": 69138, + "months": 49714, + "monto": 94073, + "monton": 40765, + "mony": 64427, + "moon": 67269, + "mor": 56802, + "more": 6384, + "morgan": 96391, + "morph": 89833, + "mort": 92611, + "mos": 8631, + "most": 3562, + "mostat": 54725, + "mostly": 62895, + "mot": 46828, + "moth": 67404, + "mother": 26412, + "motion": 48875, + "motor": 57988, + "mount": 16557, + "mounted": 81516, + "mour": 83084, + "mouse": 13237, + "mousedown": 56638, + "mouseenter": 86328, + "mouseleave": 81446, + "mousemove": 60077, + "mouseout": 55711, + "mouseover": 46158, + "mouseup": 67790, + "mouth": 32680, + "mov": 29961, + "move": 3397, + "moved": 94818, + "movement": 61174, + "moves": 66082, + "movie": 20032, + "movies": 39420, + "moving": 65115, + "moz": 46517, + "mozilla": 50345, + "mp": 1307, + "mpeg": 47551, + "mph": 55002, + "mpi": 39479, + "mpjes": 74554, + "mpl": 54010, + "mploy": 7111, + "mployee": 9458, + "mpp": 93368, + "mpr": 78373, + "mps": 93470, + "mpz": 82705, + "mq": 27674, + "mqtt": 86576, + "mr": 20946, + "ms": 1011, + "mscorlib": 71605, + "mse": 96582, + "msg": 3236, + "msgid": 86509, + "msgs": 77845, + "msp": 92545, + "mt": 2501, + "mtime": 60976, + "mtree": 70422, + "mts": 91326, + "mtx": 90310, + "mu": 15128, + "much": 58078, + "mue": 91138, + "mui": 67466, + "mul": 24160, + "mult": 25875, + "multi": 26268, + "multiline": 92760, + "multip": 77976, + "multipart": 29542, + "multiple": 35673, + "multiply": 64648, + "mun": 24997, + "mund": 35314, + "munition": 33586, + "mur": 65106, + "mus": 37727, + "music": 31161, + "must": 24812, + "mut": 6984, + "mutable": 30473, + "mutation": 30971, + "mutations": 51755, + "mute": 52214, + "mutex": 22366, + "mux": 75066, + "muÅŁ": 126661, + "muÅŁtur": 128085, + "mv": 26002, + "mvc": 94847, + "mw": 44128, + "mx": 18085, + "my": 2408, + "myModal": 46033, + "myModalLabel": 81102, + "myfile": 94921, + "myp": 86134, + "mys": 8209, + "mysql": 12272, + "mysqli": 15647, + "mysÅĤ": 134160, + "myÅĽl": 129772, + "mát": 143643, + "mática": 130739, + "máticos": 143644, + "mÃ¤ÃŁig": 140448, + "mètre": 132754, + "mé": 128613, + "méd": 130054, + "média": 130055, + "même": 132861, + "mó": 137905, + "móvel": 137906, + "mówi": 132953, + "mö": 127846, + "möglich": 127847, + "möglichkeiten": 142432, + "mü": 130187, + "mÃ¼ÅŁ": 131441, + "mı": 130329, + "mıyor": 135207, + "mız": 125673, + "mızda": 137332, + "mızı": 131174, + "mÄ±ÅŁ": 125042, + "mÄ±ÅŁtı": 132862, + "mÄ±ÅŁtır": 129577, + "n": 77, + "nEnter": 78941, + "nP": 86749, + "nThe": 88230, + "na": 3376, + "nable": 16251, + "nad": 80944, + "nage": 67055, + "nah": 38130, + "nahme": 72164, + "naire": 28483, + "naires": 68563, + "naissance": 45202, + "nak": 42874, + "nal": 41978, + "nam": 12400, + "nama": 18425, + "name": 606, + "named": 30245, + "naments": 73815, + "nameof": 68442, + "names": 11400, + "namese": 43419, + "namespace": 2231, + "nan": 18759, + "nano": 93625, + "nant": 51885, + "nants": 62263, + "nap": 6861, + "napshot": 9601, + "nar": 51657, + "nard": 96932, + "nas": 45423, + "nat": 33297, + "natal": 73339, + "nation": 98795, + "national": 41140, + "native": 29738, + "natural": 52880, + "nature": 66712, + "nav": 3722, + "navbar": 9839, + "navbarDropdown": 98311, + "navbarSupportedContent": 75722, + "navigate": 70839, + "navigation": 22608, + "navigationBar": 82653, + "navigator": 63098, + "naz": 87296, + "nb": 18080, + "nbr": 64401, + "nbsp": 5674, + "nc": 1016, + "nce": 16443, + "nces": 98162, + "ncia": 18953, + "ncias": 54679, + "ncmp": 32801, + "ncoder": 66943, + "ncpy": 35250, + "ncy": 20735, + "nd": 303, + "nda": 27896, + "ndata": 82603, + "nde": 42341, + "ndef": 5037, + "nder": 900, + "ndern": 93016, + "ndl": 88370, + "ndo": 89214, + "ndon": 85352, + "nds": 81003, + "ndx": 93256, + "ne": 811, + "nea": 32152, + "neapolis": 38476, + "near": 51659, + "nearest": 69511, + "neas": 97257, + "neath": 27817, + "neau": 82253, + "necessary": 94217, + "neck": 62340, + "nect": 2321, + "nection": 1821, + "ned": 18694, + "nee": 33091, + "need": 17046, + "needed": 40817, + "needle": 59519, + "needs": 55245, + "nees": 68719, + "neg": 28775, + "negative": 42224, + "nego": 52099, + "neh": 25402, + "nehmen": 37060, + "nehmer": 88487, + "neider": 47611, + "neighbor": 36469, + "neighbors": 79488, + "nej": 35800, + "nek": 46945, + "nel": 8138, + "nell": 48465, + "nelle": 63729, + "nelly": 86631, + "nels": 35427, + "nement": 39180, + "nemonic": 70775, + "nen": 12495, + "neo": 51260, + "ner": 1194, + "nerRadius": 33151, + "nerg": 38524, + "nergie": 63274, + "nergy": 9997, + "nero": 68952, + "ners": 4972, + "nes": 4873, + "nesc": 86596, + "nesday": 7316, + "nesia": 97475, + "nesota": 17434, + "ness": 2090, + "nest": 21414, + "nested": 59271, + "nestjs": 44587, + "nesty": 47285, + "net": 4711, + "netinet": 69699, + "nets": 52861, + "nett": 82724, + "nette": 78728, + "network": 17511, + "neutral": 59568, + "nev": 97393, + "never": 36493, + "new": 931, + "newInstance": 84036, + "newValue": 52830, + "newline": 89202, + "news": 9984, + "newsletter": 71368, + "nex": 73226, + "next": 3600, + "nextInt": 94096, + "ney": 3437, + "neys": 35271, + "nez": 54406, + "nf": 31737, + "ng": 968, + "nga": 62530, + "nge": 84490, + "ngen": 99011, + "nger": 64067, + "ngine": 70891, + "nginx": 73561, + "ngle": 66230, + "ngo": 70590, + "ngoing": 85071, + "ngr": 58192, + "ngrx": 75184, + "ngth": 6717, + "ngthen": 18465, + "ngx": 33046, + "nh": 16719, + "ni": 7751, + "nia": 20887, + "nic": 11590, + "nicas": 73346, + "nice": 52796, + "nick": 20406, + "nickname": 39413, + "nicos": 84430, + "nid": 57344, + "nie": 10810, + "niej": 59904, + "niejs": 91262, + "nier": 88983, + "niest": 84780, + "night": 9287, + "nih": 86065, + "nik": 21490, + "nika": 89065, + "ników": 129408, + "nil": 8385, + "nilai": 75591, + "nim": 56751, + "nin": 78460, + "nine": 93223, + "ning": 1229, + "ningar": 92575, + "ningen": 72108, + "ninger": 87421, + "nings": 38469, + "nio": 92633, + "nip": 69773, + "nis": 25151, + "nish": 96839, + "nist": 64759, + "nit": 44068, + "nite": 61564, + "nivel": 89000, + "nivers": 127515, + "niÄħ": 130961, + "niÄĻ": 125847, + "niÄĻcia": 141588, + "niÄĻcie": 139896, + "nj": 56108, + "nk": 49436, + "nl": 15643, + "nm": 19638, + "nn": 7370, + "nnen": 19294, + "no": 2152, + "noDB": 82240, + "nob": 55023, + "noc": 47991, + "noch": 76018, + "nock": 87121, + "nod": 34566, + "node": 3509, + "nodeName": 93451, + "nodeValue": 89985, + "nodes": 20008, + "nodiscard": 82334, + "nodoc": 93569, + "nof": 26654, + "nofollow": 28530, + "noinspection": 89267, + "noise": 52218, + "nom": 16687, + "nombre": 14236, + "nome": 17081, + "non": 6280, + "nonatomic": 10439, + "nonce": 39593, + "none": 6697, + "nonnull": 61436, + "noon": 12402, + "noop": 40162, + "noopener": 66106, + "nop": 62813, + "nor": 44707, + "nore": 93066, + "norm": 20011, + "normal": 8252, + "normalize": 30590, + "normalized": 38460, + "normally": 85236, + "north": 61895, + "nos": 36391, + "nosis": 75105, + "nost": 31457, + "nosti": 73852, + "nostic": 86424, + "not": 1921, + "notEmpty": 89517, + "nota": 67217, + "notated": 87029, + "notation": 10930, + "notations": 14297, + "note": 9974, + "noteq": 45299, + "notes": 18286, + "nothing": 41212, + "nothrow": 79381, + "notice": 23556, + "noticed": 74275, + "notif": 85046, + "notification": 18553, + "notifications": 38188, + "notify": 21948, + "notin": 83780, + "noun": 90109, + "nov": 38042, + "nova": 93184, + "novation": 98748, + "now": 3328, + "nowled": 50022, + "nowledge": 51186, + "nown": 4169, + "nowrap": 55423, + "nox": 41403, + "noxious": 94732, + "noÅĽci": 44323, + "noÅĽÄĩ": 71619, + "np": 6199, + "npc": 45383, + "npj": 62245, + "npm": 40039, + "npos": 38992, + "nqu": 139033, + "nquête": 139034, + "nr": 19618, + "ns": 4412, + "nsic": 42211, + "nst": 57718, + "nt": 406, + "ntag": 94905, + "ntax": 8842, + "nte": 23446, + "nten": 42469, + "nth": 51738, + "ntity": 89258, + "ntl": 44456, + "nton": 55353, + "ntp": 89340, + "nts": 65677, + "nty": 40833, + "nu": 8933, + "null": 2921, + "nullable": 12902, + "nullptr": 41132, + "num": 2413, + "number": 4082, + "numberOf": 37823, + "numbers": 37378, + "numer": 75596, + "numeric": 19600, + "numero": 32615, + "numerusform": 70719, + "nummer": 67919, + "numpy": 35083, + "nums": 26350, + "nung": 46621, + "nut": 23333, + "nutrition": 82564, + "nuts": 63766, + "nv": 36941, + "nvarchar": 63565, + "nw": 61227, + "nx": 23357, + "ny": 3834, + "nya": 22891, + "nych": 27284, + "nyder": 49255, + "nym": 48121, + "nz": 50824, + "nze": 84449, + "nç": 131164, + "nça": 131059, + "né": 51135, + "née": 130577, + "nü": 129894, + "nÃŃ": 54022, + "nı": 128658, + "nın": 128337, + "nız": 125288, + "nızda": 137722, + "nızı": 134502, + "nÄħ": 56350, + "nÄħÄĩ": 131264, + "nÄħÅĤ": 137253, + "nÄĻ": 126919, + "nÄĻÅĤa": 141716, + "o": 78, + "oS": 72743, + "oa": 19533, + "oad": 2731, + "oader": 39966, + "oard": 33386, + "oauth": 34363, + "ob": 674, + "oba": 26833, + "obao": 69131, + "obar": 31393, + "obb": 20415, + "obbies": 40312, + "obble": 95850, + "obbled": 90946, + "obby": 10317, + "obe": 15422, + "obel": 77106, + "ober": 5923, + "obi": 18337, + "obia": 40063, + "obic": 30806, + "obierno": 51582, + "obil": 27929, + "obile": 3372, + "obili": 94823, + "obj": 2295, + "objPHPExcel": 77500, + "objc": 28826, + "object": 1700, + "objectId": 88170, + "objectManager": 98047, + "objective": 84511, + "objects": 19210, + "objet": 66716, + "objs": 88327, + "obl": 37734, + "oble": 49993, + "oblin": 47061, + "oblins": 78814, + "obo": 38519, + "obody": 42340, + "obook": 73328, + "obot": 44639, + "obox": 32460, + "obra": 28856, + "obre": 37608, + "obs": 5481, + "observ": 22764, + "observable": 33495, + "observation": 77960, + "observations": 70384, + "observe": 48033, + "observeOn": 95059, + "observer": 30730, + "obsolete": 55015, + "obsÅĤug": 140863, + "obuf": 18464, + "obutton": 82377, + "oby": 27015, + "oc": 509, + "oca": 16846, + "ocab": 20497, + "ocabulary": 43527, + "ocache": 85672, + "ocado": 46182, + "ocal": 3683, + "ocale": 36311, + "ocaly": 66680, + "ocalypse": 48308, + "ocalyptic": 68648, + "ocard": 73278, + "ocate": 81515, + "ocation": 2276, + "ocations": 55199, + "ocator": 74328, + "ocaust": 37926, + "occ": 13983, + "occan": 84011, + "occasion": 93311, + "occer": 96188, + "occo": 40842, + "occup": 23785, + "occupation": 58262, + "occupied": 40356, + "occus": 91311, + "oce": 80556, + "ocene": 77682, + "ocese": 57644, + "och": 4953, + "ocha": 58729, + "ochastic": 65954, + "ochen": 54536, + "ochond": 46417, + "ochrome": 91721, + "ocht": 63767, + "oci": 2119, + "ocial": 2705, + "ocide": 42177, + "ociety": 6052, + "ocio": 46410, + "ocious": 77187, + "ocities": 80449, + "ocity": 7453, + "ock": 1176, + "ocked": 67385, + "ocker": 13659, + "ocket": 4537, + "ockets": 19601, + "ockey": 17887, + "ocking": 70091, + "ocks": 25183, + "ocl": 61768, + "oco": 13140, + "ocoa": 33302, + "ocode": 43688, + "ocoder": 50722, + "ocol": 4547, + "ocolate": 13816, + "ocols": 22018, + "ocom": 11986, + "ocommerce": 23181, + "ocomplete": 20104, + "ocop": 83521, + "ocos": 41388, + "ocr": 4215, + "ocracy": 45260, + "ocrat": 72532, + "ocrates": 76946, + "ocratic": 37241, + "ocrats": 54106, + "ocre": 62418, + "ocrin": 90053, + "ocrine": 77638, + "ocrisy": 67145, + "ocs": 14128, + "oct": 41692, + "ocular": 67697, + "ocument": 1452, + "ocumented": 36684, + "ocup": 72125, + "ocurrency": 77480, + "ocus": 3384, + "ocused": 20387, + "ocusing": 86495, + "ocy": 33846, + "ocyte": 78659, + "ocytes": 56778, + "ocz": 83895, + "ocê": 45759, + "od": 347, + "oda": 13993, + "odable": 69129, + "odafone": 79477, + "odal": 57597, + "odash": 25601, + "odata": 56169, + "odate": 87011, + "oday": 6289, + "odb": 16853, + "odcast": 98511, + "odd": 14556, + "ode": 534, + "oded": 6737, + "odef": 97577, + "odega": 89162, + "odel": 720, + "odeled": 96479, + "odelist": 84013, + "odem": 91461, + "oden": 54736, + "oder": 4316, + "oders": 52498, + "odes": 2539, + "odesk": 71224, + "odge": 15542, + "odi": 30459, + "odia": 46328, + "odiac": 56886, + "odial": 89069, + "odian": 63849, + "odic": 52760, + "odie": 85349, + "odied": 71106, + "odies": 10028, + "odigo": 13687, + "oding": 3700, + "odings": 58335, + "odium": 46776, + "odka": 56540, + "odo": 6004, + "odom": 84767, + "odon": 76052, + "odont": 85715, + "odor": 48917, + "odore": 67465, + "odos": 16385, + "odox": 30062, + "odoxy": 92298, + "ods": 29697, + "odu": 75955, + "odule": 1756, + "odus": 49477, + "ody": 1076, + "odyn": 59670, + "odynam": 71831, + "odynamic": 59902, + "odynamics": 78911, + "odzi": 75311, + "oe": 4644, + "oeff": 20719, + "oen": 15975, + "oenix": 18399, + "oes": 7072, + "of": 1055, + "ofday": 57826, + "off": 1847, + "offee": 21180, + "offer": 25077, + "offers": 71879, + "office": 26516, + "official": 32812, + "offline": 63529, + "offs": 32483, + "offset": 3176, + "offsetof": 19214, + "ofi": 92023, + "ofil": 46799, + "ofile": 65273, + "ofilm": 78576, + "ofire": 85482, + "ofs": 36081, + "ofstream": 52930, + "oft": 14118, + "often": 60817, + "oftware": 73105, + "og": 538, + "oga": 14034, + "ogan": 29536, + "ogany": 91732, + "oge": 40532, + "ogen": 11706, + "ogene": 33124, + "ogeneity": 75630, + "ogeneous": 48022, + "ogenerated": 58605, + "ogenesis": 51279, + "ogenic": 28469, + "ogenous": 52495, + "ogens": 56018, + "oger": 55440, + "ogg": 16108, + "oggle": 5216, + "oggled": 63900, + "oggler": 39178, + "oggles": 50852, + "ogh": 63936, + "ogi": 63877, + "ogie": 77672, + "ogl": 97247, + "ogle": 2443, + "oglob": 75425, + "oglobin": 93755, + "ogn": 3934, + "ogne": 95341, + "ognition": 30656, + "ognitive": 50449, + "ognito": 63441, + "ogo": 23500, + "ogonal": 65971, + "ogr": 73546, + "ogra": 91141, + "ograd": 66911, + "ograf": 25058, + "ografia": 49801, + "ografÃŃa": 141742, + "ogram": 12958, + "ograms": 55385, + "ograph": 3178, + "ographed": 90573, + "ographer": 41506, + "ographers": 78354, + "ographic": 12679, + "ographical": 31177, + "ographically": 63931, + "ographics": 44145, + "ographies": 66145, + "ographs": 66032, + "ography": 5696, + "ográf": 82481, + "ogs": 26307, + "ogue": 27423, + "ogui": 90470, + "ogy": 15661, + "oh": 2267, + "oha": 65083, + "ohan": 56472, + "ohana": 84342, + "ohen": 26440, + "ohl": 39337, + "ohn": 3103, + "oho": 95286, + "ohon": 98592, + "oi": 6728, + "oice": 8834, + "oid": 588, + "oidal": 70816, + "oids": 16960, + "oil": 73813, + "oin": 1961, + "oine": 62816, + "oined": 73506, + "oins": 68798, + "oint": 781, + "ointed": 83979, + "ointment": 27919, + "ointments": 59517, + "oints": 14399, + "oir": 13300, + "oire": 41727, + "ois": 29048, + "oise": 68189, + "oit": 47796, + "oj": 21265, + "oji": 26902, + "ojis": 83428, + "ok": 562, + "oka": 30766, + "okable": 63734, + "okableCall": 76566, + "okane": 79586, + "okay": 93217, + "oke": 4740, + "oked": 10823, + "okedex": 86331, + "okemon": 13514, + "oken": 1679, + "okens": 9713, + "oker": 10451, + "okers": 40681, + "okes": 7940, + "oki": 38191, + "okia": 27552, + "okie": 12502, + "okies": 30818, + "okin": 78017, + "oking": 10746, + "okino": 82434, + "okit": 96363, + "oklyn": 82187, + "oko": 27243, + "oks": 34222, + "oksen": 62764, + "oku": 16493, + "okus": 90076, + "oky": 69881, + "ol": 337, + "ola": 7924, + "olah": 90786, + "olan": 37772, + "oland": 96440, + "olang": 37287, + "olar": 7417, + "olare": 76361, + "olarity": 72527, + "olars": 26242, + "olas": 19186, + "olate": 33066, + "olated": 80519, + "olatile": 25517, + "olation": 34962, + "old": 813, + "oldem": 73718, + "oldemort": 75063, + "older": 2018, + "olders": 32120, + "olding": 14995, + "oldown": 24766, + "olds": 18431, + "oldt": 93044, + "ole": 1263, + "olean": 1738, + "oleans": 96999, + "olec": 12465, + "olecular": 42843, + "olecule": 54169, + "olecules": 81106, + "oled": 41731, + "olem": 70858, + "olen": 17205, + "oleon": 59103, + "oler": 22072, + "olerance": 31661, + "oles": 7151, + "olesale": 42507, + "olest": 60995, + "olesterol": 34144, + "oley": 48999, + "olf": 8095, + "olg": 74063, + "oli": 14225, + "olia": 38160, + "olian": 88008, + "olib": 56603, + "oliberal": 71216, + "olic": 7762, + "olicies": 42038, + "olicit": 97048, + "olicited": 91104, + "olicitud": 58399, + "olics": 45907, + "olicy": 8018, + "olid": 5192, + "oliday": 10092, + "olidays": 19891, + "olie": 61017, + "olin": 36637, + "olina": 57904, + "oline": 21908, + "oling": 84569, + "olini": 94794, + "olio": 72167, + "olis": 68758, + "olist": 34675, + "olith": 47118, + "olithic": 65370, + "olk": 22604, + "olkata": 62692, + "olkien": 75979, + "oll": 965, + "olla": 42683, + "ollah": 47855, + "olland": 42332, + "ollapse": 10786, + "ollapsed": 66974, + "ollar": 21295, + "olle": 54048, + "ollect": 24605, + "ollection": 1908, + "ollections": 3107, + "ollectors": 29751, + "ollen": 32289, + "oller": 69396, + "olley": 34519, + "olleyError": 93384, + "ollider": 31255, + "ollipop": 89544, + "ollision": 29037, + "ollo": 27961, + "ollow": 1544, + "ollower": 29034, + "olls": 67878, + "olly": 8618, + "ollywood": 14115, + "olo": 10011, + "oload": 28112, + "olocation": 43206, + "olog": 1609, + "ologi": 85635, + "ologia": 37973, + "ologic": 38127, + "ological": 5729, + "ologically": 29350, + "ologie": 37369, + "ologies": 9090, + "ologist": 16155, + "ologists": 21306, + "ologna": 65790, + "ologne": 67444, + "ologue": 76728, + "ology": 2449, + "ologÃŃa": 53333, + "olon": 72269, + "olor": 789, + "olph": 42970, + "ols": 3069, + "olson": 71538, + "olt": 6181, + "olta": 59854, + "oltage": 26489, + "oltip": 13379, + "olu": 43806, + "olucion": 91560, + "olución": 77076, + "olulu": 72046, + "olum": 1132, + "olumbia": 73070, + "olume": 4661, + "olumes": 19705, + "olumn": 1286, + "olut": 9041, + "olute": 6268, + "olutely": 9689, + "olution": 3214, + "olutions": 20201, + "olv": 35315, + "olvable": 88097, + "olve": 3948, + "olved": 8731, + "olvency": 93930, + "olver": 7921, + "olvers": 39435, + "olves": 18186, + "olvimento": 78669, + "olving": 19648, + "oly": 5730, + "olygon": 19582, + "olynomial": 76309, + "ológ": 60094, + "ológica": 134411, + "ológico": 133397, + "om": 316, + "oma": 7786, + "omain": 3121, + "omal": 52811, + "omaly": 83092, + "oman": 6908, + "omanip": 55915, + "omap": 74811, + "omas": 22267, + "omat": 80768, + "omatic": 13487, + "omb": 2855, + "omba": 76930, + "ombat": 77628, + "ombie": 23342, + "ombies": 35694, + "ombine": 26197, + "ombo": 80387, + "ombok": 16424, + "ombre": 5509, + "ombres": 30032, + "ombs": 85704, + "ome": 635, + "omed": 24139, + "omedical": 60760, + "omega": 32696, + "omen": 6297, + "omencl": 88847, + "omens": 55541, + "oment": 12913, + "omentum": 58311, + "omer": 25359, + "omers": 68538, + "omes": 20347, + "omet": 4413, + "ometer": 20408, + "ometers": 32404, + "omething": 11532, + "ometimes": 12693, + "ometown": 36718, + "ometric": 23375, + "ometrics": 91791, + "ometry": 6988, + "omez": 53594, + "omi": 21441, + "omial": 20855, + "omic": 3075, + "omics": 23320, + "omid": 43019, + "omidou": 53341, + "omin": 7970, + "ominated": 49515, + "omination": 80380, + "ominator": 30206, + "oming": 17765, + "omit": 77968, + "omite": 85803, + "omitempty": 19967, + "omm": 19786, + "ommen": 46646, + "omnia": 49305, + "omo": 15900, + "omon": 34585, + "omore": 43984, + "omorphic": 69917, + "omp": 14435, + "ompiler": 87485, + "ompson": 95001, + "oms": 6940, + "omy": 5533, + "on": 263, + "onChange": 51485, + "onClick": 48897, + "onDelete": 61171, + "ona": 6721, + "onacci": 39345, + "onal": 24202, + "onald": 79694, + "onas": 55301, + "onation": 90883, + "onaut": 36064, + "once": 13184, + "onces": 42798, + "onclick": 39799, + "oncé": 94610, + "ond": 2111, + "onda": 18194, + "onde": 17231, + "onden": 72216, + "onder": 14184, + "ondere": 70439, + "onders": 27313, + "ondheim": 82303, + "ondo": 34971, + "ondon": 6474, + "ondrous": 93550, + "onds": 49958, + "one": 603, + "oned": 19684, + "onedDateTime": 83459, + "oneksi": 64242, + "onen": 41928, + "onent": 84547, + "onents": 98341, + "onenumber": 73266, + "oner": 26576, + "ones": 3154, + "onesia": 56672, + "onest": 35124, + "onestly": 80472, + "onet": 68686, + "oney": 2534, + "ong": 644, + "onga": 93924, + "ongan": 82288, + "ongo": 6363, + "ongodb": 21225, + "ongoose": 10872, + "ongs": 13181, + "ongsTo": 34225, + "ongyang": 56898, + "oni": 20783, + "onia": 21251, + "onian": 75491, + "onic": 14011, + "onica": 63945, + "onical": 22391, + "oning": 19744, + "oningen": 97177, + "onio": 81413, + "onis": 96694, + "onitor": 30314, + "onium": 89244, + "online": 25649, + "only": 3243, + "onn": 26378, + "onna": 13459, + "onne": 28952, + "onnement": 89980, + "onnen": 98103, + "onné": 134824, + "ono": 10148, + "onom": 14274, + "onomic": 47128, + "onomies": 79346, + "onomous": 29846, + "onomy": 16974, + "onomÃŃa": 141019, + "onor": 54037, + "ons": 2382, + "onsense": 97733, + "onso": 61767, + "ont": 544, + "onta": 49414, + "onte": 52277, + "onth": 6039, + "onto": 9935, + "ontology": 62233, + "ontvangst": 53209, + "onus": 27698, + "ony": 3549, + "onym": 7831, + "onymous": 9757, + "onyms": 45603, + "onz": 27045, + "oo": 2624, + "ood": 1386, + "oodle": 30607, + "oodles": 97121, + "oodoo": 73401, + "oods": 9065, + "oogle": 2671, + "ook": 1941, + "ooke": 85458, + "ookeeper": 94171, + "ookie": 9619, + "ookies": 25874, + "ooks": 14685, + "ooky": 66724, + "ool": 1749, + "oolStrip": 10094, + "oola": 45980, + "ooled": 45073, + "ools": 6178, + "oom": 4191, + "oomla": 45836, + "oon": 9009, + "oons": 28539, + "ooo": 38621, + "oooo": 25761, + "oooooooo": 60293, + "oop": 47306, + "oops": 64562, + "oor": 10692, + "oord": 81437, + "oose": 13752, + "oot": 1905, + "ooter": 9746, + "ooth": 6491, + "oothing": 38413, + "ooting": 27462, + "op": 453, + "opa": 39852, + "opacity": 20570, + "opal": 56979, + "opaque": 95480, + "opard": 31929, + "opath": 35111, + "opathic": 61109, + "opathy": 52942, + "opause": 83010, + "opc": 67858, + "opcion": 92781, + "opcode": 48128, + "ope": 2792, + "oped": 16367, + "open": 2508, + "opencv": 43195, + "opened": 56128, + "openh": 42070, + "openhagen": 44829, + "openid": 48551, + "opening": 67685, + "opens": 24175, + "opensource": 44408, + "openssl": 53612, + "oper": 3296, + "operand": 40609, + "operands": 92593, + "operate": 85199, + "operation": 9262, + "operations": 38163, + "operative": 42619, + "operator": 7884, + "operators": 32838, + "opers": 51087, + "opes": 18523, + "opez": 41259, + "oph": 5127, + "ophage": 97954, + "ophe": 80409, + "opher": 16940, + "ophil": 97535, + "ophile": 67446, + "ophilia": 92555, + "ophobia": 54508, + "ophobic": 60700, + "ophon": 93806, + "ophone": 77203, + "ophys": 84304, + "ophysical": 89904, + "opi": 71087, + "opia": 29551, + "opian": 47648, + "opic": 24810, + "opies": 75541, + "oping": 33707, + "opl": 55078, + "oplan": 93460, + "oplast": 91235, + "oplay": 36265, + "oplayer": 93108, + "ople": 1164, + "opleft": 76252, + "oples": 26424, + "oplevel": 67780, + "opo": 55709, + "opol": 27665, + "opolitan": 84909, + "opoly": 39763, + "opor": 88666, + "opot": 88390, + "opoulos": 56895, + "opp": 4775, + "oppable": 77539, + "opped": 17573, + "oppel": 66522, + "opper": 18487, + "oppers": 31442, + "opping": 7008, + "oppins": 79396, + "opport": 143008, + "opportunitÃł": 143009, + "opr": 46288, + "opro": 45926, + "oproject": 61351, + "ops": 3721, + "opsis": 32806, + "opsy": 48192, + "opt": 2912, + "optgroup": 96330, + "optic": 93683, + "optim": 19133, + "optimize": 94332, + "optimized": 98868, + "optimizer": 33296, + "option": 2047, + "optional": 12807, + "options": 2875, + "opts": 10518, + "opup": 27735, + "opus": 45870, + "opy": 1266, + "opyright": 2171, + "opération": 135016, + "oque": 60552, + "or": 269, + "orElse": 54088, + "orThunk": 42844, + "orWhere": 83328, + "ora": 6215, + "orable": 24715, + "oracle": 69631, + "orado": 96755, + "orage": 2680, + "oral": 9819, + "oralType": 95904, + "oram": 93496, + "orama": 48016, + "oran": 54404, + "orang": 84168, + "orange": 34164, + "orarily": 96154, + "orary": 7592, + "oras": 64647, + "orate": 37895, + "oration": 7614, + "orative": 62365, + "orb": 29886, + "orbit": 74341, + "orc": 49822, + "orca": 92159, + "orce": 16316, + "orch": 21584, + "ord": 539, + "orda": 57080, + "ordable": 35170, + "ordan": 13396, + "orde": 52118, + "orden": 27784, + "order": 1358, + "orderBy": 22197, + "orderId": 53491, + "orderby": 45757, + "ordered": 10544, + "orderid": 97561, + "ordering": 51869, + "orders": 7917, + "ordes": 80392, + "ordial": 92673, + "ordin": 5986, + "ordinal": 45151, + "ordinary": 21031, + "ordinate": 8730, + "ordinated": 82469, + "ordinates": 8742, + "ordination": 98244, + "ordinator": 17442, + "ording": 2678, + "ordion": 25711, + "ordo": 51272, + "ordon": 21508, + "ordova": 76163, + "ords": 2260, + "ore": 460, + "orea": 60048, + "oreach": 8539, + "oreal": 89768, + "orean": 45195, + "ored": 3018, + "oredProcedure": 45517, + "oref": 46752, + "oreferrer": 57940, + "orelease": 45214, + "orem": 13173, + "oren": 23647, + "orent": 68525, + "orer": 14827, + "ores": 4589, + "orest": 41419, + "orestation": 96793, + "oret": 18174, + "oretical": 90767, + "orf": 32842, + "org": 1775, + "organ": 8462, + "organic": 60594, + "organisation": 57804, + "organisms": 75891, + "organization": 23899, + "organizations": 69253, + "organized": 62216, + "orge": 6670, + "orgen": 39563, + "orgeous": 20708, + "orges": 80775, + "orget": 55755, + "orgetown": 59468, + "orgh": 88526, + "orghini": 95319, + "orgia": 46949, + "orgot": 43685, + "orgt": 82751, + "ori": 13601, + "oria": 10782, + "orial": 21913, + "orian": 22090, + "orianCalendar": 63793, + "orias": 67263, + "oriasis": 91903, + "oric": 26359, + "orical": 69748, + "orida": 39757, + "orie": 29203, + "orient": 14881, + "orientation": 24294, + "oriented": 32436, + "ories": 2433, + "orig": 4670, + "origin": 8611, + "original": 9889, + "oring": 5503, + "orio": 49831, + "orious": 21995, + "oriously": 69024, + "oris": 50280, + "orizontal": 8101, + "ork": 669, + "orks": 73302, + "orld": 1385, + "orm": 493, + "ormal": 2492, + "orman": 62875, + "ormap": 39031, + "ormsg": 82027, + "orn": 1512, + "orna": 33264, + "ornado": 62943, + "orne": 16907, + "ornecedor": 76935, + "orners": 44280, + "orney": 9430, + "orneys": 26812, + "orning": 79575, + "ornings": 51685, + "ornment": 94475, + "orno": 11121, + "orns": 43029, + "orny": 79577, + "oro": 18307, + "oron": 84850, + "orough": 14438, + "orous": 20244, + "orp": 30115, + "orph": 16347, + "orphic": 40869, + "orphism": 52807, + "orpion": 72475, + "orpor": 38282, + "orption": 65081, + "orque": 85964, + "orr": 17391, + "orra": 83038, + "orraine": 90022, + "orrar": 57648, + "orre": 92433, + "orrect": 27034, + "orrent": 48709, + "orrh": 53208, + "orris": 97316, + "orro": 67569, + "orrow": 7768, + "orry": 8468, + "ors": 1087, + "orsch": 79985, + "orsche": 43476, + "orse": 10836, + "orses": 22434, + "orsi": 73087, + "orsk": 81587, + "orst": 65296, + "ort": 371, + "orta": 88539, + "ortal": 33372, + "ortality": 74952, + "orte": 16568, + "orted": 13595, + "ortex": 44592, + "orth": 2364, + "orthand": 61679, + "ortho": 90824, + "orthy": 33494, + "ortic": 89338, + "orting": 51472, + "ortion": 25886, + "orton": 37100, + "orts": 18955, + "oru": 65005, + "orum": 33006, + "orus": 29869, + "ory": 679, + "orz": 67323, + "os": 436, + "osa": 11983, + "osaic": 44783, + "osal": 32556, + "osals": 85975, + "osas": 86809, + "osate": 87870, + "osaur": 57976, + "osaurs": 52256, + "osc": 23469, + "osci": 84628, + "oscope": 62653, + "oscopic": 83567, + "ose": 960, + "oseconds": 75889, + "osed": 19586, + "osemite": 83013, + "osen": 8313, + "oser": 23438, + "oses": 19696, + "osex": 19791, + "osexual": 48200, + "osg": 91412, + "osh": 9267, + "oshi": 31845, + "osi": 30724, + "osing": 14421, + "osis": 10704, + "osit": 34054, + "osite": 87742, + "osition": 2113, + "osity": 22053, + "osl": 47458, + "oso": 27147, + "osomal": 95008, + "osome": 31454, + "osomes": 57275, + "osoph": 11343, + "osopher": 45359, + "osos": 94328, + "osp": 4785, + "ospace": 80014, + "ospel": 26156, + "ospels": 97998, + "osph": 23609, + "osphate": 92373, + "osphere": 65122, + "ospital": 6298, + "oss": 3662, + "ossa": 43836, + "ossal": 78816, + "ossed": 87850, + "ossible": 23653, + "ossier": 39229, + "ossil": 83961, + "ossip": 41473, + "ost": 535, + "osta": 27364, + "oste": 84123, + "osten": 14012, + "oster": 11713, + "osterone": 32474, + "osti": 70679, + "osto": 35790, + "oston": 9424, + "ostream": 7617, + "ostringstream": 59006, + "ot": 354, + "ota": 6089, + "otal": 2370, + "otas": 53524, + "otate": 20223, + "otation": 11606, + "otch": 93607, + "ote": 1272, + "otec": 88162, + "oteca": 64390, + "otech": 58096, + "otechn": 44760, + "otechnology": 51436, + "oted": 9253, + "otel": 40785, + "oten": 65628, + "otence": 97560, + "otent": 63532, + "oter": 25392, + "oteric": 82675, + "oters": 39796, + "otes": 6295, + "oth": 8767, + "other": 1575, + "otherapy": 41711, + "othermal": 89996, + "others": 58861, + "otherwise": 59936, + "othy": 28571, + "oti": 88292, + "otic": 14212, + "otics": 49903, + "otide": 67944, + "otify": 37253, + "otime": 19175, + "otine": 39153, + "oting": 11519, + "otion": 5956, + "otionEvent": 63420, + "otional": 40864, + "otions": 40256, + "otive": 23827, + "otland": 77111, + "otle": 50526, + "otlin": 15345, + "oto": 2072, + "otomy": 84895, + "oton": 25849, + "otonin": 66955, + "otope": 50682, + "otor": 9887, + "otos": 8001, + "otoxic": 90576, + "otp": 48708, + "otr": 86379, + "otre": 51162, + "otron": 51005, + "otropic": 78332, + "otros": 47045, + "ots": 2412, + "ott": 1716, + "otta": 22193, + "ottage": 33695, + "otte": 50011, + "otted": 15521, + "otten": 10291, + "ottenham": 49703, + "ottes": 48253, + "ottesville": 60570, + "otti": 41227, + "ottie": 82229, + "ottle": 62118, + "otto": 35626, + "ottom": 2826, + "otton": 16881, + "otts": 87701, + "otty": 57639, + "otyp": 36841, + "otype": 4156, + "otypes": 21830, + "otypical": 87067, + "otyping": 66147, + "ou": 283, + "oub": 5482, + "ouble": 1201, + "oubles": 85613, + "oubted": 33300, + "oubtedly": 34110, + "ouce": 69779, + "ouch": 3026, + "ouched": 33070, + "oucher": 23937, + "ouchers": 52755, + "oud": 2950, + "ouden": 67527, + "oueur": 71166, + "oufl": 61740, + "ouflage": 72231, + "oug": 48304, + "ough": 1384, + "ought": 2450, + "oui": 73844, + "ouis": 47990, + "oul": 10965, + "ould": 616, + "oulder": 37634, + "oulos": 53947, + "oulouse": 66630, + "oultry": 54019, + "oun": 1624, + "ounc": 2981, + "ounce": 9734, + "ounced": 19453, + "ouncement": 40171, + "ouncements": 54264, + "ouncer": 79745, + "ounces": 29944, + "ouncil": 6690, + "ouncill": 45403, + "ouncing": 36757, + "ouncy": 81301, + "ound": 795, + "oundary": 72466, + "oundation": 4068, + "ounded": 13082, + "ounder": 92906, + "ounding": 13586, + "oundingBox": 57504, + "ounds": 3095, + "ounge": 25564, + "ouns": 58986, + "ounsel": 93641, + "ount": 629, + "ountain": 34524, + "ountains": 92306, + "ounter": 39914, + "ounters": 92002, + "ountries": 20126, + "ountry": 4976, + "ounty": 36725, + "oup": 13346, + "ouple": 84389, + "ouples": 68426, + "oupon": 32987, + "oupper": 36888, + "our": 413, + "ourage": 60040, + "ource": 919, + "ourced": 53392, + "ourcem": 96831, + "ources": 2360, + "ourcing": 42562, + "ourd": 76392, + "oure": 79184, + "oured": 20394, + "ourg": 48441, + "ouri": 21467, + "ourke": 84434, + "ourmet": 50702, + "ourn": 3431, + "ournal": 4931, + "ournals": 32614, + "ournament": 9783, + "ournaments": 33568, + "ournemouth": 96979, + "ourney": 32990, + "ournée": 71394, + "ouro": 49826, + "ours": 2471, + "ourse": 5253, + "ourses": 16349, + "ourt": 28586, + "ous": 782, + "ousand": 51849, + "ousands": 39037, + "ouse": 1530, + "oused": 36488, + "ousedown": 48066, + "ousel": 12374, + "ouser": 88005, + "ouses": 19757, + "ousing": 21738, + "ously": 7017, + "ousse": 45551, + "oust": 25112, + "oustic": 34315, + "ouston": 14415, + "ousy": 93523, + "out": 411, + "outcome": 86194, + "outdir": 94853, + "oute": 2133, + "outed": 18533, + "outedEventArgs": 29072, + "outer": 2676, + "outers": 66095, + "outes": 5495, + "outfile": 59625, + "outh": 2898, + "outil": 30158, + "outine": 14159, + "outines": 28628, + "outing": 10909, + "outline": 44130, + "outlined": 49133, + "output": 3006, + "outputs": 41006, + "outs": 11672, + "outside": 66990, + "outu": 50032, + "outube": 10221, + "ouv": 65767, + "ouve": 82284, + "ouver": 14826, + "oux": 40570, + "ov": 859, + "ova": 12660, + "ovable": 31955, + "ovah": 71094, + "oval": 13516, + "ovan": 42657, + "ovation": 22874, + "ove": 994, + "oved": 4941, + "ovel": 43335, + "ovement": 14458, + "oven": 16253, + "ovenant": 38602, + "over": 1975, + "overall": 74455, + "overe": 68930, + "overflow": 20823, + "overlap": 80061, + "overlay": 21118, + "overn": 6706, + "overnment": 25057, + "override": 9199, + "overrides": 83358, + "overs": 8969, + "overview": 49278, + "overwrite": 64915, + "oves": 10088, + "ovi": 45088, + "ovic": 40045, + "ovich": 50114, + "ovid": 25582, + "ovie": 6327, + "ovies": 12546, + "oving": 9130, + "ovo": 21715, + "ovolta": 92491, + "ovsky": 78578, + "ovy": 61515, + "ow": 363, + "owa": 13472, + "owan": 80231, + "owane": 73349, + "owania": 56029, + "owanie": 54389, + "owany": 94791, + "oward": 81971, + "owards": 70739, + "owaÄĩ": 49147, + "owaÅĤ": 128958, + "owaÅĤa": 132233, + "owe": 23733, + "owed": 12817, + "owego": 77280, + "owej": 60382, + "owel": 67057, + "owell": 79499, + "ower": 1202, + "owered": 79213, + "owering": 89614, + "owers": 15965, + "owi": 71338, + "owie": 28986, + "owied": 59428, + "owing": 23184, + "owitz": 52331, + "owl": 9605, + "owler": 64328, + "owment": 75455, + "own": 779, + "owned": 18332, + "owner": 8118, + "owners": 22980, + "ownership": 78147, + "ownik": 72732, + "ownload": 37702, + "owns": 42356, + "ownt": 12901, + "owntown": 15508, + "owo": 57715, + "owoÅĽci": 128124, + "owoÅĽÄĩ": 140615, + "ows": 4241, + "owski": 28284, + "owy": 36875, + "owych": 45911, + "owÄħ": 128891, + "ox": 5131, + "oxel": 52391, + "oxetine": 90603, + "oxic": 81229, + "oxid": 83978, + "oxide": 54089, + "oxy": 60163, + "oy": 2253, + "oya": 46062, + "oyal": 47102, + "oyer": 41119, + "oyo": 52378, + "oystick": 74132, + "oz": 9510, + "oze": 90153, + "ozilla": 27342, + "ozo": 95514, + "ozy": 75423, + "oÃłi": 133046, + "oÄŁ": 129294, + "oÄŁlu": 129295, + "oÅĽci": 89056, + "oÅĽciÄħ": 130795, + "oÅĽÄĩ": 128428, + "p": 79, + "pData": 57881, + "pNet": 49929, + "pNext": 89281, + "pa": 6595, + "paRepository": 41972, + "pac": 44951, + "pace": 1306, + "paced": 68552, + "paces": 27338, + "pack": 4748, + "package": 1722, + "packageName": 91063, + "packages": 43141, + "packed": 50321, + "packet": 24829, + "packing": 95124, + "pad": 13242, + "padding": 9571, + "padÅĤ": 140793, + "pag": 18419, + "page": 2893, + "pageIndex": 91877, + "pageNum": 81297, + "pageNumber": 84727, + "pageSize": 39069, + "pageTitle": 85786, + "paged": 63115, + "pagen": 82613, + "pager": 46200, + "pages": 10781, + "pagesize": 84917, + "pagina": 63897, + "paginate": 38465, + "pagination": 25584, + "paginator": 78356, + "paging": 72746, + "pai": 84959, + "paid": 34678, + "pain": 91624, + "paint": 33617, + "pair": 12670, + "paired": 76167, + "pairs": 77256, + "pais": 87098, + "pak": 70989, + "pal": 19308, + "palette": 59674, + "pan": 848, + "pane": 25325, + "panel": 7429, + "panels": 82698, + "panic": 19079, + "panied": 22065, + "panies": 17010, + "panion": 40790, + "panse": 94419, + "pants": 91937, + "pany": 1981, + "paper": 19166, + "papers": 47293, + "paque": 53422, + "par": 1732, + "para": 14794, + "parable": 49156, + "paragraph": 27727, + "paragus": 93689, + "paralle": 64761, + "parallel": 46103, + "paralleled": 67801, + "param": 903, + "parameter": 16181, + "parameters": 13786, + "paramref": 43015, + "params": 3519, + "parate": 67598, + "parated": 49600, + "paration": 35135, + "parator": 12356, + "parcel": 95920, + "pard": 61643, + "pare": 3380, + "pared": 7212, + "paredStatement": 22422, + "paren": 41064, + "parency": 21279, + "parent": 3765, + "parentId": 53548, + "parentNode": 80704, + "parents": 25456, + "parer": 61119, + "paring": 44152, + "parison": 35542, + "parity": 94368, + "park": 28736, + "parm": 57453, + "parms": 77508, + "pars": 72380, + "parse": 6400, + "parseFloat": 64872, + "parseInt": 33587, + "parsed": 41030, + "parser": 9657, + "parsers": 88564, + "part": 4480, + "partial": 37420, + "partials": 49458, + "participant": 58513, + "participants": 91135, + "particle": 41979, + "particles": 68498, + "particularly": 72939, + "partition": 40998, + "partment": 35991, + "partner": 36258, + "partners": 97031, + "parts": 18252, + "party": 32957, + "pas": 41101, + "pass": 6385, + "passed": 35422, + "passes": 94323, + "passport": 42669, + "passwd": 52500, + "password": 3833, + "past": 52420, + "paste": 56868, + "pat": 4676, + "patch": 3400, + "patches": 96878, + "path": 2343, + "pathname": 66281, + "paths": 21623, + "patial": 32414, + "patibility": 53053, + "patible": 37079, + "patient": 22722, + "patients": 80337, + "patrick": 64220, + "pattern": 14339, + "patterns": 25971, + "pause": 27448, + "paused": 78293, + "pawn": 68596, + "pay": 13890, + "payer": 93586, + "payload": 19427, + "payment": 13755, + "payments": 63043, + "paypal": 71205, + "pañ": 142298, + "pb": 16650, + "pbs": 26015, + "pc": 3992, + "pcb": 90870, + "pcf": 90143, + "pch": 70348, + "pci": 55387, + "pcion": 16270, + "pciones": 77368, + "pción": 43328, + "pcl": 77853, + "pcm": 76484, + "pcodes": 71719, + "pcs": 47313, + "pd": 15360, + "pdata": 57796, + "pdb": 84202, + "pdev": 49419, + "pdf": 11828, + "pdo": 29714, + "pe": 375, + "peace": 54125, + "peak": 22792, + "peaker": 65474, + "pear": 8015, + "pearance": 13024, + "peare": 36552, + "peat": 10979, + "peated": 41954, + "peater": 65242, + "peating": 64877, + "peats": 82439, + "pec": 992, + "pecia": 79960, + "pecial": 2964, + "pecially": 4794, + "pecies": 88621, + "pecific": 15564, + "pecified": 44040, + "pect": 987, + "pected": 4046, + "pecting": 65348, + "pection": 16076, + "pections": 48713, + "pective": 9085, + "pectives": 32201, + "pector": 18997, + "pectral": 86575, + "pectrum": 38706, + "pects": 7973, + "ped": 6874, + "pedia": 82050, + "pedido": 80662, + "pedo": 52494, + "pee": 61498, + "peech": 39586, + "peed": 4508, + "peek": 29107, + "peer": 16537, + "pees": 81300, + "peg": 10311, + "pegawai": 72958, + "pei": 63874, + "pek": 87176, + "pekt": 78881, + "pel": 47524, + "pell": 37522, + "pellier": 91654, + "pen": 2752, + "pend": 3740, + "pendicular": 68551, + "pending": 28210, + "peng": 63083, + "pent": 45125, + "peon": 90036, + "people": 16069, + "peq": 82338, + "per": 712, + "perPage": 94622, + "perate": 79113, + "peration": 91774, + "perator": 76866, + "perature": 14365, + "perc": 48135, + "percent": 24422, + "percentage": 40550, + "pered": 42768, + "perf": 59545, + "perfect": 63085, + "perfil": 81071, + "perform": 28488, + "performance": 61244, + "perhaps": 65272, + "peri": 31736, + "peria": 74099, + "perial": 67227, + "perience": 12970, + "perienced": 44833, + "periences": 68497, + "periment": 14329, + "perimental": 25513, + "periments": 49632, + "pering": 60718, + "period": 19304, + "perl": 64936, + "perm": 19913, + "permalink": 81104, + "permanent": 97424, + "permission": 22618, + "permissions": 29900, + "permit": 39681, + "perms": 87772, + "peror": 25819, + "pers": 31290, + "perse": 81244, + "persist": 39826, + "persistent": 69389, + "person": 8987, + "persona": 70594, + "personal": 29612, + "persons": 76651, + "pert": 76368, + "pery": 61379, + "pes": 20352, + "pesan": 96285, + "pest": 29123, + "pet": 6862, + "petition": 55767, + "pets": 64201, + "petto": 82243, + "pez": 39959, + "peÅĤ": 126739, + "peÅĤni": 129887, + "pf": 15897, + "pg": 3517, + "pga": 74669, + "pgsql": 53298, + "ph": 759, + "pha": 3967, + "phabet": 18485, + "phalt": 35174, + "phan": 9943, + "phans": 84311, + "phant": 26924, + "phants": 52567, + "phanumeric": 65788, + "phas": 50380, + "phase": 25141, + "phasis": 54149, + "phe": 80806, + "phem": 59941, + "phen": 14769, + "pher": 28082, + "pherd": 42531, + "phere": 8023, + "pheres": 64633, + "pheric": 32249, + "pherical": 44745, + "phet": 21951, + "phetamine": 65604, + "phi": 16827, + "phia": 17653, + "phies": 47072, + "phil": 45085, + "phin": 54753, + "phins": 39672, + "phinx": 49561, + "phis": 36789, + "phon": 51701, + "phone": 4844, + "phoneNumber": 58173, + "phones": 16729, + "phony": 29648, + "phoon": 77049, + "phot": 51299, + "photo": 11556, + "photos": 24603, + "php": 1208, + "phrase": 27710, + "phthalm": 80837, + "phy": 12700, + "phys": 41205, + "physical": 46592, + "physics": 66765, + "pi": 2493, + "piar": 56526, + "pic": 15587, + "picable": 94018, + "pick": 29245, + "picked": 93499, + "picker": 14868, + "pickle": 85905, + "pickup": 84684, + "pics": 73457, + "picture": 26483, + "pictureBox": 59933, + "pictured": 68289, + "pictures": 75403, + "pid": 10412, + "pie": 20283, + "piece": 22362, + "pieces": 46468, + "pieczeÅĦ": 128184, + "pieczeÅĦst": 128185, + "piel": 44236, + "pies": 44652, + "pig": 89500, + "pii": 93921, + "pike": 64446, + "pile": 12192, + "piler": 28055, + "pill": 49846, + "pillar": 76502, + "pin": 13273, + "pine": 38038, + "ping": 9989, + "pink": 63249, + "pins": 74558, + "pio": 11917, + "pios": 92397, + "pip": 51501, + "pipe": 13768, + "pipeline": 51258, + "pipes": 86720, + "pir": 5565, + "piration": 28479, + "pire": 22922, + "pired": 11724, + "pires": 18968, + "piring": 78763, + "pirit": 52729, + "piry": 48209, + "pis": 56896, + "pisa": 139525, + "pisaÄĩ": 139526, + "pit": 32586, + "pitch": 53238, + "pite": 5972, + "pivot": 68441, + "pix": 35484, + "pixel": 29684, + "pixels": 53319, + "pizza": 92438, + "piÄĻ": 142029, + "piÄĻt": 142030, + "pj": 91116, + "pk": 20819, + "pkg": 30069, + "pkt": 46775, + "pl": 500, + "pla": 56538, + "place": 2007, + "placed": 36369, + "placeholder": 12384, + "placeholders": 82670, + "placement": 16101, + "placements": 63535, + "placer": 90212, + "places": 26078, + "placing": 47933, + "plain": 20772, + "plaint": 31297, + "plaintext": 71223, + "plan": 10393, + "planation": 35890, + "plane": 19025, + "planes": 38057, + "planet": 50074, + "plans": 78524, + "plant": 20827, + "plants": 63705, + "plash": 19493, + "plat": 42339, + "plate": 1750, + "plates": 7278, + "platform": 15734, + "platz": 57548, + "play": 1363, + "played": 50808, + "player": 3434, + "players": 18312, + "playing": 41912, + "playlist": 34233, + "plays": 27797, + "ple": 694, + "pleado": 32266, + "pleasant": 43907, + "please": 30021, + "pled": 49085, + "plement": 2764, + "plementary": 65982, + "plementation": 9874, + "plemented": 14231, + "plements": 4674, + "plen": 86535, + "pler": 12910, + "plers": 88409, + "plet": 10819, + "plete": 5171, + "pleted": 11028, + "pletely": 49168, + "pletion": 14386, + "plets": 89492, + "plevel": 88805, + "plex": 9111, + "pliance": 31612, + "pliant": 59852, + "plib": 81682, + "plied": 3440, + "plier": 7875, + "pliers": 15178, + "plies": 7202, + "pline": 31439, + "pling": 11863, + "plings": 80302, + "plist": 39934, + "plit": 2292, + "plits": 39033, + "plitude": 30050, + "pll": 80438, + "plode": 70744, + "plorer": 79825, + "plot": 4469, + "plotlib": 15406, + "plots": 32797, + "ploy": 2693, + "ployment": 52799, + "plr": 94845, + "pls": 56986, + "plt": 9476, + "pluck": 58496, + "plug": 47474, + "plugin": 9138, + "plugins": 18716, + "plural": 66546, + "plus": 7138, + "plusplus": 24428, + "plx": 78854, + "ply": 2541, + "plätze": 142454, + "pm": 5187, + "pmat": 94663, + "pn": 19958, + "pname": 82662, + "png": 14066, + "po": 5368, + "pod": 39073, + "poi": 82263, + "point": 2768, + "pointer": 16044, + "points": 7706, + "poke": 48098, + "pokemon": 57380, + "pol": 9896, + "polate": 45429, + "polation": 44686, + "polator": 57736, + "pole": 68072, + "policy": 34790, + "polit": 90527, + "political": 74685, + "poll": 30141, + "pollo": 85890, + "poly": 33435, + "polygon": 65012, + "pon": 618, + "ponce": 88055, + "pond": 3511, + "ponde": 75049, + "ponder": 20328, + "pone": 80917, + "ponent": 1146, + "ponential": 59825, + "ponents": 2700, + "pong": 59102, + "ponge": 36254, + "ponible": 96851, + "pons": 2737, + "ponse": 1025, + "ponses": 22648, + "ponsible": 42319, + "ponsive": 14221, + "ponsor": 30241, + "ponsored": 34741, + "ponsors": 50876, + "pool": 10285, + "poon": 31976, + "poons": 82088, + "pop": 8374, + "popover": 68842, + "popular": 43248, + "populate": 47721, + "population": 44441, + "popup": 23706, + "por": 4308, + "pora": 70664, + "porate": 60621, + "porn": 56650, + "porno": 78385, + "port": 403, + "portal": 37953, + "porte": 44708, + "ported": 35526, + "porter": 89232, + "portfolio": 27159, + "portion": 15388, + "portlet": 84543, + "porto": 92574, + "portrait": 95641, + "ports": 3394, + "portun": 55886, + "portunity": 67064, + "pos": 966, + "posable": 17421, + "posables": 95966, + "posal": 24155, + "pose": 2900, + "posed": 3865, + "poser": 20071, + "poses": 8285, + "posición": 130858, + "posing": 8314, + "posit": 2724, + "posite": 13607, + "posites": 80139, + "position": 3487, + "positions": 27436, + "positive": 30487, + "positor": 60950, + "positories": 17286, + "pository": 3099, + "posium": 53137, + "posix": 80804, + "posição": 131280, + "poss": 8831, + "possible": 10031, + "possibly": 53128, + "post": 2203, + "postData": 65139, + "postId": 71716, + "posta": 38531, + "postal": 33170, + "postalcode": 76481, + "postcode": 72438, + "posted": 43082, + "poster": 45401, + "postgres": 43070, + "postgresql": 88470, + "posting": 64564, + "posts": 12664, + "posure": 12044, + "posé": 132189, + "pot": 19099, + "potential": 91936, + "pour": 54519, + "pow": 21743, + "power": 13175, + "powered": 51516, + "powers": 77304, + "powied": 128070, + "powiedzie": 139570, + "powiedzieÄĩ": 139571, + "powiedź": 140849, + "pożycz": 143442, + "pp": 602, + "ppard": 98004, + "ppe": 71905, + "pped": 6924, + "ppelin": 68008, + "pper": 6922, + "ppers": 35523, + "pping": 10732, + "ppo": 39480, + "pps": 59543, + "ppt": 97428, + "ppv": 98597, + "ppy": 33963, + "pq": 64016, + "pr": 649, + "pra": 31659, + "praak": 79086, + "practice": 59521, + "pragma": 6015, + "prar": 93950, + "pras": 81222, + "pray": 90604, + "prd": 86222, + "pre": 1726, + "pread": 20717, + "prec": 10645, + "precated": 19889, + "precation": 69593, + "preced": 26152, + "precedented": 28691, + "prech": 61231, + "preci": 62453, + "precio": 47774, + "precision": 27182, + "pred": 23464, + "predicate": 61785, + "predict": 34698, + "predicted": 94300, + "prediction": 68931, + "predictions": 91327, + "pref": 30552, + "prefer": 80060, + "preferences": 60989, + "preferred": 73486, + "prefix": 11849, + "prefs": 60963, + "preg": 29485, + "preh": 30782, + "prehensive": 52899, + "preload": 96007, + "premium": 84594, + "prend": 58876, + "prene": 16534, + "preneur": 55305, + "prenom": 66348, + "prep": 71774, + "prepare": 13609, + "prepared": 60112, + "prepend": 55906, + "pres": 23623, + "presence": 73419, + "present": 28744, + "presentation": 43925, + "preserve": 92882, + "preset": 85508, + "presión": 134252, + "press": 1873, + "pressed": 14318, + "pression": 4011, + "pressions": 39873, + "pressive": 68463, + "presso": 34460, + "pressor": 56220, + "pressure": 45974, + "pressão": 135353, + "prestashop": 91292, + "pret": 8043, + "preter": 28637, + "pretty": 32955, + "prev": 9883, + "prevState": 66743, + "prevent": 92655, + "preview": 27171, + "previous": 19702, + "pri": 68608, + "price": 6555, + "prices": 58769, + "pricing": 74370, + "prim": 17893, + "primary": 6545, + "primaryKey": 54513, + "prime": 32338, + "primer": 47570, + "primir": 49067, + "primitive": 73448, + "principal": 66450, + "pring": 2637, + "print": 1350, + "printStats": 82261, + "printed": 52213, + "printer": 62956, + "printf": 2517, + "printing": 98831, + "println": 33655, + "prints": 25738, + "printw": 94095, + "prior": 64233, + "priority": 23582, + "prise": 9671, + "prises": 25425, + "prising": 33654, + "prisingly": 48164, + "prit": 39707, + "prite": 5734, + "prites": 28879, + "priv": 11887, + "privacy": 46821, + "private": 1996, + "privation": 64493, + "privileged": 97288, + "prix": 91713, + "prm": 94043, + "pro": 776, + "prob": 19899, + "probability": 87440, + "probably": 54440, + "probante": 87937, + "probe": 52329, + "problem": 34586, + "problems": 95340, + "proc": 15782, + "procedure": 21141, + "process": 4630, + "processable": 57240, + "processed": 34022, + "processing": 20660, + "processor": 29474, + "processors": 81748, + "procs": 90087, + "prod": 19748, + "produ": 8555, + "produce": 97274, + "producer": 58912, + "product": 3031, + "productId": 46166, + "productName": 88426, + "production": 22627, + "productive": 33270, + "producto": 46707, + "productos": 74905, + "products": 10144, + "produk": 88812, + "produto": 66195, + "prof": 21826, + "profession": 96135, + "professional": 65786, + "profil": 90992, + "profile": 5365, + "profiles": 56195, + "profit": 25394, + "profits": 76501, + "prog": 32992, + "program": 14906, + "programming": 91626, + "programs": 72953, + "progress": 14366, + "progressbar": 81583, + "proj": 30386, + "project": 4987, + "projectId": 62562, + "projection": 60471, + "projects": 17161, + "prom": 24468, + "promise": 49550, + "promo": 75465, + "promotion": 74142, + "prompt": 40581, + "pron": 71139, + "proof": 15780, + "prop": 2674, + "propTypes": 86993, + "proper": 80668, + "properties": 13193, + "property": 3699, + "propertyName": 62548, + "proposal": 73874, + "propri": 70521, + "props": 4761, + "prot": 4391, + "protect": 45979, + "protected": 5764, + "protein": 78465, + "proto": 15110, + "protobuf": 32921, + "protocol": 17014, + "protocols": 85412, + "prototype": 45654, + "prov": 42947, + "prove": 34463, + "proved": 95508, + "provide": 61456, + "provided": 63425, + "provider": 19979, + "providers": 47287, + "province": 26335, + "prowad": 128132, + "prowadzi": 128133, + "prowadziÄĩ": 140699, + "prowadziÅĤ": 142311, + "prox": 41498, + "proxy": 22803, + "prs": 84305, + "prt": 88697, + "prung": 92352, + "prus": 49887, + "przed": 128198, + "przedsi": 128199, + "przedsiÄĻb": 128200, + "przedsiÄĻbior": 128201, + "przedsiÄĻbiorc": 143957, + "przedsiÄĻbiorst": 142625, + "przedsiÄĻbiorstw": 142626, + "pré": 133938, + "prés": 130311, + "présent": 130312, + "prü": 66693, + "prüf": 133618, + "ps": 1690, + "psc": 81984, + "pseudo": 68893, + "psi": 30765, + "psilon": 59892, + "pson": 18014, + "pst": 88795, + "psy": 45146, + "psych": 91009, + "psz": 44688, + "pt": 417, + "pta": 51591, + "pte": 24151, + "pter": 73137, + "ptest": 70334, + "pth": 8490, + "pthread": 36283, + "ptic": 26235, + "ptide": 46209, + "ptides": 73389, + "ptime": 28941, + "ption": 560, + "ptions": 1300, + "ptive": 53735, + "pto": 57991, + "ptom": 79697, + "ptoms": 51173, + "pton": 18748, + "ptr": 3505, + "ptrdiff": 88269, + "ptron": 94710, + "pts": 12754, + "ptune": 51914, + "pty": 1595, + "ptype": 72814, + "pu": 5584, + "pub": 9585, + "public": 888, + "publication": 71160, + "publish": 27502, + "published": 32866, + "publisher": 51191, + "pull": 23441, + "pulse": 78207, + "pun": 64907, + "punk": 75509, + "punkt": 71865, + "pur": 24998, + "purchase": 33877, + "pure": 51733, + "purple": 56507, + "purpose": 76663, + "pus": 17820, + "push": 9077, + "pushViewController": 97271, + "put": 628, + "putation": 13320, + "putc": 89040, + "pute": 8492, + "puted": 19292, + "puter": 11281, + "puties": 43933, + "puts": 16920, + "puty": 16983, + "pués": 42209, + "pv": 30168, + "pw": 31862, + "pwd": 25565, + "px": 1767, + "py": 3288, + "pygame": 51060, + "pytest": 26721, + "python": 12669, + "pyx": 85359, + "pz": 66565, + "pé": 128937, + "péri": 137823, + "pósito": 143269, + "póÅĤ": 129666, + "põ": 139565, + "põe": 139566, + "pÅĤat": 130088, + "pÅĤyw": 135506, + "q": 80, + "qa": 15445, + "qb": 49698, + "qc": 59833, + "qd": 76727, + "qe": 78124, + "qed": 51233, + "qh": 91900, + "qi": 40187, + "qid": 88587, + "qing": 89584, + "ql": 1470, + "qli": 6905, + "qm": 77398, + "qml": 83612, + "qn": 48350, + "qp": 32763, + "qq": 27579, + "qr": 23004, + "qrst": 61739, + "qrstuvwxyz": 64467, + "qrt": 8140, + "qry": 61255, + "qs": 26358, + "qt": 23044, + "qtt": 32267, + "qty": 29303, + "qu": 446, + "qua": 43832, + "quad": 31610, + "quake": 66718, + "qual": 1751, + "qualification": 71909, + "qualified": 36335, + "qualities": 82365, + "quality": 10473, + "quals": 2366, + "quam": 86701, + "quan": 88437, + "quant": 30448, + "quantity": 13523, + "quare": 5151, + "quared": 34249, + "quares": 40856, + "quarter": 32015, + "quartered": 67620, + "quarters": 19670, + "quat": 72983, + "quate": 19137, + "que": 591, + "queda": 41360, + "quee": 17110, + "queen": 93114, + "queeze": 19340, + "quel": 41914, + "quelle": 87165, + "quence": 4375, + "quences": 43945, + "quent": 16769, + "quential": 15838, + "quer": 15959, + "queries": 42835, + "querque": 67608, + "query": 1631, + "queryParams": 80720, + "querySelector": 86267, + "queryString": 95888, + "ques": 13968, + "quest": 719, + "question": 7841, + "questions": 17348, + "quests": 86322, + "quet": 23300, + "quete": 60354, + "quets": 90591, + "quette": 51706, + "queue": 4584, + "queueReusable": 37185, + "queueReusableCell": 38559, + "queued": 66547, + "queues": 76089, + "quez": 41121, + "qui": 46291, + "quia": 83622, + "quick": 27763, + "quier": 32471, + "quierda": 86381, + "quiet": 43650, + "quil": 87584, + "quila": 74129, + "quin": 35344, + "quina": 53406, + "quine": 47258, + "quip": 35747, + "quipe": 60573, + "quipment": 48879, + "quir": 11388, + "quire": 984, + "quired": 2931, + "quirer": 50792, + "quires": 17694, + "quiries": 31900, + "quiring": 70094, + "quirrel": 86640, + "quiry": 17343, + "quis": 9202, + "quisa": 58495, + "quisar": 95235, + "quisite": 32424, + "quisites": 48250, + "quisition": 17204, + "quisitions": 52829, + "quist": 55309, + "quit": 31537, + "quite": 83170, + "quito": 41096, + "quiv": 15418, + "quivo": 28944, + "quivos": 86231, + "quiz": 31949, + "quo": 26512, + "quoi": 61898, + "quoise": 58670, + "quot": 13492, + "quota": 92604, + "quote": 2949, + "quoted": 63725, + "quotelev": 30860, + "quotes": 53282, + "qus": 61504, + "qué": 63491, + "quÃŃ": 130532, + "qv": 69578, + "qw": 69291, + "qx": 88735, + "r": 81, + "rPid": 83893, + "ra": 956, + "rab": 50105, + "rabbit": 86345, + "rac": 19959, + "race": 41183, + "racial": 48925, + "rack": 71685, + "raction": 16199, + "racuse": 59769, + "rad": 13281, + "rade": 49076, + "rades": 59535, + "radi": 69212, + "radient": 66737, + "radio": 12636, + "radius": 26715, + "rado": 76827, + "rador": 65251, + "radouro": 54640, + "rael": 6196, + "raf": 15453, + "raft": 2944, + "rafted": 60769, + "rag": 4101, + "rage": 47368, + "ragen": 56364, + "ragment": 6017, + "ragments": 41956, + "ragon": 55150, + "rah": 32507, + "raham": 16602, + "rai": 67862, + "raid": 13831, + "raig": 72316, + "rail": 75635, + "rails": 37345, + "rain": 29093, + "rais": 39998, + "raisal": 64259, + "raise": 18704, + "raised": 92513, + "raises": 32271, + "raising": 93680, + "raison": 76503, + "rait": 14350, + "raith": 85782, + "raits": 26287, + "raj": 55386, + "ral": 3461, + "rale": 45040, + "ram": 2396, + "rama": 30373, + "rame": 2373, + "ramento": 35756, + "ramer": 46369, + "rames": 15615, + "ramework": 5794, + "ramid": 31703, + "ramids": 86373, + "rams": 94701, + "ran": 6576, + "rance": 34106, + "rances": 94497, + "rand": 11335, + "random": 11463, + "rang": 92583, + "range": 9669, + "ranges": 66562, + "rangle": 71182, + "rank": 13129, + "ranking": 96546, + "rans": 33389, + "ransition": 84954, + "rant": 34434, + "rap": 4611, + "rape": 19842, + "raph": 1935, + "raphic": 78073, + "raphics": 6348, + "rapid": 85712, + "rapped": 56289, + "rapper": 5518, + "rapy": 60826, + "raq": 10686, + "raquo": 85260, + "rar": 43983, + "rarian": 94920, + "raries": 14876, + "rary": 3451, + "ras": 12784, + "rase": 10632, + "rases": 26565, + "rasing": 96478, + "rastructure": 10314, + "rasında": 127920, + "rat": 17606, + "rate": 7698, + "rated": 54273, + "rates": 66485, + "rather": 73203, + "rating": 21931, + "ratings": 73136, + "ratio": 45358, + "ration": 2165, + "rats": 62937, + "ratulations": 37416, + "raud": 97957, + "raw": 1041, + "rawer": 95112, + "rawid": 139722, + "rawidÅĤow": 139723, + "rawing": 1696, + "rawl": 33683, + "rawler": 33369, + "rawn": 67568, + "rawtypes": 97261, + "rax": 68765, + "ray": 11611, + "rays": 65143, + "raz": 8703, + "razier": 91047, + "razil": 14042, + "razione": 63031, + "razy": 12079, + "rb": 10681, + "rbrace": 75283, + "rbrakk": 74084, + "rc": 1287, + "rch": 49066, + "rchive": 96191, + "rcode": 54796, + "rd": 6498, + "rdf": 77998, + "re": 265, + "rea": 5213, + "reach": 22606, + "reachable": 46550, + "react": 2934, + "reaction": 59207, + "reactstrap": 78352, + "read": 878, + "readOnly": 87703, + "readcr": 12370, + "readcrumb": 17032, + "readcrumbs": 24466, + "reader": 11149, + "reading": 6154, + "readonly": 22569, + "reads": 30358, + "ready": 2307, + "readystatechange": 59811, + "reak": 1108, + "real": 7951, + "realDonaldTrump": 73405, + "realloc": 81590, + "really": 53660, + "realm": 69897, + "ream": 1237, + "reamble": 86706, + "rear": 58408, + "reas": 50418, + "reason": 19895, + "reasonable": 46692, + "reat": 1222, + "reate": 964, + "reated": 2850, + "reater": 28362, + "reatest": 11175, + "reating": 72810, + "reation": 26453, + "reative": 86366, + "reatment": 34220, + "reator": 36886, + "reature": 52251, + "reau": 89949, + "reb": 31475, + "rebbe": 66491, + "rec": 2758, + "recall": 93828, + "recated": 38777, + "rece": 53387, + "receipt": 62623, + "receive": 41893, + "received": 41823, + "receiver": 37553, + "recent": 46643, + "recht": 47631, + "recio": 35658, + "recipe": 25208, + "recipes": 49859, + "recipient": 42610, + "recision": 21079, + "reck": 24638, + "recogn": 33451, + "recognized": 46067, + "recommend": 66589, + "recommended": 85347, + "record": 8548, + "records": 26203, + "recover": 74283, + "rect": 2851, + "rectangle": 55316, + "rection": 30366, + "rections": 67343, + "recur": 75762, + "recursive": 49512, + "recv": 33977, + "red": 1151, + "redd": 32774, + "reddit": 36933, + "rede": 91661, + "redential": 30320, + "redentials": 15735, + "redi": 64599, + "redicate": 16874, + "redict": 8861, + "redient": 14342, + "redients": 10377, + "redirect": 8117, + "redirectTo": 65799, + "redirectToRoute": 89279, + "redis": 21748, + "redit": 10827, + "redits": 29502, + "redni": 127752, + "redo": 63561, + "reds": 53369, + "reduce": 26273, + "reducers": 66398, + "redux": 21178, + "ree": 765, + "reeNode": 31583, + "reece": 89031, + "reed": 28516, + "reek": 10726, + "reement": 17150, + "reements": 57775, + "reen": 4442, + "rees": 7858, + "reesome": 57305, + "reet": 3744, + "reeting": 43632, + "reetings": 56517, + "reeze": 27633, + "ref": 1097, + "refer": 41160, + "reference": 16291, + "references": 35909, + "referer": 88963, + "referrer": 97949, + "reff": 61644, + "reffen": 64260, + "refix": 5060, + "refixer": 89348, + "reflect": 34913, + "reflection": 76998, + "refresh": 17168, + "refs": 16149, + "refund": 68483, + "reg": 1580, + "rega": 38876, + "regar": 33730, + "regation": 95795, + "regex": 26387, + "regexp": 55796, + "region": 3943, + "regions": 57708, + "regist": 53287, + "register": 6343, + "registered": 34909, + "registers": 76937, + "registr": 48464, + "registration": 25862, + "registro": 55346, + "registry": 29172, + "reglo": 52739, + "regn": 39281, + "rego": 67717, + "regon": 17103, + "regor": 65921, + "regs": 30886, + "regular": 22308, + "regulated": 80622, + "regunta": 86273, + "reh": 11063, + "rei": 55314, + "reib": 38013, + "reiben": 60322, + "reibung": 75399, + "reich": 40334, + "reinterpret": 52521, + "reira": 77242, + "reject": 46996, + "rejected": 95353, + "rek": 41861, + "rel": 3748, + "reland": 86466, + "related": 9721, + "relation": 22221, + "relations": 65334, + "relationship": 36095, + "relationships": 85824, + "relative": 20432, + "relay": 90891, + "release": 22998, + "released": 76938, + "relevant": 97573, + "religious": 96150, + "rell": 16285, + "rella": 33042, + "rellas": 72923, + "rello": 74133, + "reload": 38982, + "rels": 53783, + "relu": 31208, + "relude": 85068, + "rem": 1826, + "rema": 89718, + "remain": 47584, + "remainder": 77381, + "remaining": 43436, + "remark": 37448, + "remarks": 14447, + "reme": 9634, + "remely": 53258, + "remember": 29280, + "remen": 61869, + "rement": 54655, + "reminder": 77905, + "remium": 84032, + "remote": 18147, + "remove": 5399, + "removeAttr": 56817, + "removeClass": 14142, + "removed": 45756, + "ren": 1440, + "rena": 22691, + "renal": 46638, + "rename": 51058, + "renc": 62834, + "rence": 15892, + "rench": 6655, + "renched": 53971, + "rend": 9300, + "rende": 85527, + "render": 7322, + "renderer": 26948, + "renders": 54159, + "rending": 61994, + "rends": 78617, + "rene": 46909, + "rength": 17842, + "reno": 54883, + "renom": 44843, + "rens": 77340, + "rent": 7976, + "rente": 71723, + "reo": 29583, + "reon": 49790, + "reopen": 81507, + "rep": 9995, + "repair": 76439, + "repeat": 30624, + "repid": 93594, + "replace": 8156, + "replaceAll": 55860, + "replacement": 83631, + "reply": 21034, + "repo": 23476, + "report": 11736, + "reported": 75733, + "reports": 36906, + "repos": 68354, + "repositories": 81907, + "repository": 23319, + "repr": 30837, + "represent": 35269, + "representation": 83116, + "represented": 52759, + "req": 2958, + "requ": 11019, + "requencies": 71731, + "requency": 15845, + "requent": 69691, + "requently": 69841, + "request": 2035, + "requestCode": 74535, + "requestData": 93478, + "requested": 67105, + "requests": 36242, + "requete": 94342, + "require": 4310, + "required": 6279, + "requirements": 70126, + "requires": 41375, + "rer": 37049, + "res": 416, + "resa": 14315, + "resar": 91198, + "resas": 37954, + "rescia": 97512, + "resco": 71418, + "resden": 57009, + "rese": 70438, + "research": 60464, + "resent": 2695, + "resentation": 13584, + "resenter": 74646, + "resents": 9578, + "reservation": 47684, + "reserve": 56985, + "reserved": 51102, + "reset": 9716, + "resh": 2488, + "reshape": 16137, + "resher": 85424, + "reshold": 7161, + "resident": 39531, + "resize": 17347, + "resizing": 41467, + "reso": 49764, + "resolution": 55756, + "resolve": 17325, + "resolved": 39747, + "resolver": 48943, + "resource": 9233, + "resources": 12745, + "resp": 18243, + "respect": 44634, + "respond": 6818, + "respons": 81589, + "response": 2322, + "responseData": 89278, + "responseObject": 86006, + "responses": 71000, + "responsive": 51297, + "respuesta": 57660, + "ress": 673, + "resse": 21797, + "ressed": 14378, + "resses": 43257, + "ressing": 56781, + "ression": 11185, + "ressive": 47953, + "rest": 3927, + "restart": 50215, + "restaurant": 33226, + "restaurants": 95137, + "reste": 98934, + "restore": 30804, + "restrial": 52586, + "restrict": 27898, + "restricted": 50797, + "restriction": 71331, + "result": 1382, + "resultCode": 87124, + "resultSet": 71451, + "resultado": 32361, + "results": 8074, + "resume": 41395, + "ret": 2122, + "reta": 65698, + "retain": 55372, + "retch": 10209, + "retched": 61737, + "rete": 83505, + "reten": 85885, + "retorno": 96814, + "retrieve": 55527, + "retry": 44848, + "rets": 75330, + "rett": 17261, + "rette": 66928, + "retty": 21322, + "return": 689, + "returnValue": 53015, + "returned": 77591, + "returns": 4216, + "retval": 19843, + "retweeted": 86658, + "reu": 64786, + "reur": 34240, + "reuse": 53049, + "rev": 7282, + "reve": 90976, + "reveal": 94026, + "reverse": 25903, + "review": 19417, + "reviews": 39475, + "revision": 28342, + "revolution": 95722, + "rew": 4266, + "reward": 49007, + "rewrite": 52473, + "rex": 37539, + "rey": 8073, + "rez": 22741, + "rf": 8052, + "rg": 1984, + "rgan": 8443, + "rganization": 68364, + "rgb": 16509, + "rgba": 20400, + "rgctx": 61488, + "rgyz": 98862, + "rh": 40096, + "rho": 40322, + "rhs": 26408, + "ri": 461, + "ria": 4204, + "riad": 46539, + "riage": 8775, + "riages": 37032, + "rial": 6619, + "rian": 7266, + "riangle": 84191, + "rians": 35168, + "rias": 47831, + "rib": 1897, + "riba": 84244, + "ribbon": 54580, + "rible": 12280, + "ribly": 36752, + "ric": 2216, + "rica": 61696, + "rical": 12108, + "ricane": 25053, + "ricanes": 55195, + "ricao": 27673, + "rice": 23120, + "rices": 24419, + "rich": 13851, + "richText": 77498, + "richTextPanel": 82215, + "richt": 36702, + "ricia": 45191, + "ricing": 38806, + "rick": 12649, + "ricks": 77789, + "rico": 55247, + "rics": 6198, + "ricula": 45482, + "ricular": 56233, + "riculum": 23136, + "rid": 1869, + "riday": 6121, + "ridden": 42185, + "ride": 1399, + "rides": 18245, + "ridge": 11183, + "ridged": 79128, + "riding": 49680, + "rido": 69549, + "ridor": 82436, + "rie": 7231, + "rieb": 44914, + "rieben": 73461, + "ried": 4487, + "rief": 4364, + "rieg": 82491, + "riel": 22036, + "rien": 34738, + "riend": 5039, + "riendly": 17784, + "riends": 12067, + "rient": 12986, + "rientation": 17663, + "rients": 31830, + "rier": 7253, + "riere": 84467, + "riers": 17291, + "ries": 4019, + "riet": 64570, + "rieve": 45004, + "rieved": 82612, + "rieving": 67704, + "rif": 92266, + "rift": 41380, + "rig": 14408, + "riger": 63944, + "rigesimal": 40359, + "right": 1291, + "rightarrow": 19491, + "rightness": 36630, + "rights": 76187, + "rij": 20279, + "rijk": 86674, + "rik": 20411, + "rika": 40454, + "rike": 40652, + "rikes": 67631, + "ril": 30560, + "rim": 6283, + "riminal": 57040, + "riminator": 94124, + "rimon": 48759, + "rimp": 43894, + "rin": 25288, + "rina": 39971, + "rine": 39001, + "ring": 12640, + "ringe": 84024, + "rink": 34365, + "rint": 1586, + "rio": 10383, + "rior": 3254, + "riority": 8773, + "riors": 21175, + "rios": 27851, + "riot": 84050, + "riott": 66771, + "rious": 83618, + "rip": 4561, + "ripe": 51091, + "ripp": 54689, + "ripper": 93265, + "ripple": 82373, + "rippling": 90771, + "rips": 75716, + "ripsi": 66683, + "rique": 33295, + "rir": 37215, + "rire": 95413, + "ris": 5963, + "rise": 31509, + "rish": 93907, + "rising": 58923, + "risk": 80943, + "rist": 2819, + "ristol": 40615, + "rists": 93959, + "rit": 1003, + "rita": 97727, + "ritable": 22533, + "ritch": 91589, + "rite": 1247, + "ritel": 58962, + "riteln": 51097, + "riter": 2542, + "riteria": 5094, + "riterion": 12334, + "riters": 31829, + "rites": 23262, + "ritic": 49208, + "ritical": 14509, + "riting": 12201, + "rition": 17026, + "ritional": 36742, + "ritis": 36990, + "rito": 28041, + "ritos": 94012, + "ritt": 54913, + "ritte": 98280, + "ritten": 12153, + "ritz": 56721, + "rium": 40452, + "riv": 76367, + "rive": 56935, + "river": 5469, + "rix": 18356, + "riz": 24741, + "rière": 64120, + "rió": 97301, + "rk": 47272, + "rl": 2381, + "rlen": 69069, + "rm": 8719, + "rms": 90869, + "rn": 35622, + "rna": 89231, + "rnd": 64869, + "rne": 65920, + "rng": 69890, + "ro": 299, + "roach": 64500, + "road": 8546, + "roadcast": 19105, + "roads": 42691, + "roat": 17804, + "rob": 22740, + "robat": 75101, + "robe": 32836, + "robot": 18247, + "robots": 60115, + "roc": 27687, + "rocess": 40848, + "rocessing": 48687, + "rock": 20521, + "rocket": 46790, + "rod": 23505, + "rodu": 4550, + "roduce": 47845, + "roduced": 29217, + "roducing": 59537, + "roduction": 16741, + "roe": 34704, + "rog": 11918, + "rogate": 48373, + "rogen": 25188, + "rogram": 86120, + "roi": 50460, + "roid": 1577, + "roids": 40094, + "roit": 68496, + "rok": 53561, + "roke": 7441, + "roken": 81709, + "roker": 45985, + "rokes": 55037, + "rol": 1080, + "role": 5778, + "roleId": 97491, + "roles": 14643, + "roleum": 39999, + "roll": 1100, + "rollable": 68755, + "rollback": 33559, + "rolled": 20947, + "roller": 1468, + "rollers": 7273, + "rolley": 74043, + "rolling": 16210, + "rollment": 25864, + "rollo": 43684, + "rolls": 38274, + "rom": 441, + "roma": 56984, + "roman": 62014, + "romatic": 98766, + "rome": 6648, + "rometer": 87271, + "romise": 5273, + "romium": 77859, + "romo": 98539, + "romosome": 75118, + "rompt": 14749, + "ron": 2248, + "rone": 77221, + "rones": 31821, + "rong": 14347, + "ronic": 8370, + "ronics": 27329, + "rons": 25610, + "ront": 9411, + "ronym": 46880, + "roof": 75636, + "room": 2966, + "roomId": 86394, + "rooms": 9750, + "root": 2888, + "rootScope": 23553, + "roots": 37165, + "rop": 887, + "ropa": 27557, + "ropdown": 13271, + "roperties": 9249, + "roperty": 8218, + "roph": 21990, + "rophe": 47883, + "rophic": 41710, + "rophy": 57075, + "ropic": 43936, + "ropical": 50072, + "ropol": 22341, + "ropolis": 54322, + "ropolitan": 30511, + "ropp": 75056, + "ropped": 41716, + "ropping": 80204, + "ropri": 6160, + "ropriate": 24278, + "rops": 3702, + "ropsych": 97338, + "ropy": 17764, + "ror": 7632, + "rored": 47643, + "roring": 89579, + "rors": 31686, + "ros": 3630, + "roscope": 89779, + "rose": 24847, + "rosis": 62312, + "ross": 2128, + "rosse": 52356, + "rosso": 32212, + "rossover": 37172, + "rost": 41502, + "rot": 4640, + "rotate": 16213, + "rotation": 13538, + "rotch": 91478, + "rote": 5529, + "rotein": 90648, + "rots": 89229, + "rott": 46689, + "rottle": 27535, + "rou": 581, + "rouch": 74927, + "roud": 42276, + "rough": 1432, + "rought": 6344, + "roulette": 98148, + "round": 1049, + "rounded": 43991, + "roup": 886, + "roupe": 37715, + "roupon": 47003, + "roups": 88410, + "rous": 26522, + "route": 8966, + "routeParams": 70029, + "routeProvider": 83660, + "router": 9937, + "routes": 19794, + "routine": 52980, + "routing": 73320, + "rouw": 71623, + "rov": 39394, + "rove": 50415, + "rovers": 12563, + "row": 651, + "rowCount": 48149, + "rowData": 97959, + "rowable": 14941, + "rowad": 82287, + "rowave": 33703, + "rowing": 24181, + "rown": 4830, + "rowned": 44084, + "rowning": 50420, + "rows": 1811, + "rowsable": 51111, + "rowse": 13397, + "rowser": 4012, + "rowsers": 90558, + "rowsing": 86266, + "rowth": 19089, + "rox": 54789, + "roy": 3352, + "roys": 49193, + "roz": 63989, + "rozen": 42240, + "rp": 22252, + "rparr": 88920, + "rpc": 29414, + "rpm": 73782, + "rq": 52189, + "rr": 634, + "rray": 846, + "rrha": 90495, + "rror": 1275, + "rs": 5428, + "rsa": 60869, + "rschein": 94213, + "rsp": 59976, + "rss": 50623, + "rst": 58025, + "rstrip": 70398, + "rt": 3342, + "rtc": 45892, + "rtl": 48468, + "rtle": 92436, + "rts": 77724, + "rtype": 32513, + "ru": 2672, + "rub": 59430, + "ruby": 46275, + "ruc": 79355, + "rud": 66355, + "rug": 71947, + "ruise": 92653, + "ruit": 21026, + "ruitment": 46133, + "ruits": 67898, + "ruk": 76667, + "rule": 12937, + "rules": 21977, + "rum": 10721, + "run": 6108, + "runner": 41736, + "running": 27173, + "runs": 53310, + "runtime": 22255, + "rup": 30622, + "rupt": 6585, + "ruptcy": 95219, + "rupted": 85954, + "ruption": 14123, + "ruptions": 54681, + "rus": 20341, + "rush": 77177, + "rust": 35788, + "rut": 69446, + "ruta": 80907, + "ruz": 62323, + "rv": 10553, + "rvine": 66574, + "rw": 31768, + "rx": 12651, + "rxjs": 19209, + "ry": 884, + "ryan": 56758, + "ryfall": 73398, + "rying": 27509, + "rylic": 43180, + "ryn": 60243, + "ryo": 96134, + "rypt": 3571, + "rypted": 14026, + "ryption": 15597, + "rypto": 9444, + "ryptography": 46663, + "rypton": 53536, + "rys": 41326, + "rysler": 55356, + "rz": 51960, + "rze": 54074, + "rzy": 87988, + "rzÄħ": 133676, + "rzÄħd": 128705, + "rzÄĻ": 132141, + "rá": 37368, + "ráf": 79043, + "rán": 129216, + "rão": 131200, + "rä": 125245, + "räg": 133053, + "rän": 134291, + "ränk": 134292, + "räu": 137994, + "räume": 141292, + "rÃ¥": 96921, + "rès": 127404, + "ré": 42611, + "rés": 131526, + "rêt": 140104, + "ró": 82037, + "rób": 132822, + "ród": 90945, + "rów": 129491, + "róż": 127420, + "różni": 133109, + "rô": 136033, + "rö": 131315, + "rü": 126354, + "rück": 129579, + "rün": 125334, + "rÃł": 126289, + "rÃŃ": 129358, + "rÃŃa": 128446, + "rı": 128136, + "rÄĻ": 132244, + "rÄĻcz": 132245, + "s": 82, + "sPid": 83815, + "sWith": 16056, + "sa": 9081, + "sad": 82114, + "safe": 18675, + "said": 83259, + "saida": 85750, + "sal": 19222, + "salary": 51610, + "saldo": 92512, + "sale": 32654, + "sales": 29041, + "salt": 42226, + "sam": 45996, + "same": 24063, + "samp": 92303, + "sample": 13611, + "samples": 41118, + "sampling": 96959, + "san": 33082, + "sand": 91246, + "sandbox": 76756, + "sanitize": 65704, + "sans": 76758, + "sanız": 131300, + "sap": 64613, + "sar": 98082, + "sass": 60326, + "sat": 36468, + "sav": 94251, + "save": 6628, + "saved": 32702, + "saving": 83057, + "say": 36790, + "says": 96936, + "sb": 16892, + "sbin": 92263, + "sburg": 90101, + "sburgh": 25254, + "sc": 2388, + "scal": 93207, + "scala": 26822, + "scalar": 35342, + "scale": 12445, + "scaled": 73947, + "scaling": 93052, + "scan": 16405, + "scanf": 26395, + "scanner": 67941, + "scape": 57518, + "scar": 24513, + "scatter": 70250, + "scenario": 61422, + "scene": 22483, + "scenes": 86965, + "sch": 21062, + "sched": 72243, + "schedule": 28630, + "scheduled": 93585, + "scheduler": 63122, + "schema": 17349, + "schemas": 56543, + "scheme": 46141, + "schlie": 138663, + "schlieÃŁen": 138664, + "schlä": 143753, + "schläge": 143754, + "schlü": 143340, + "schlüsse": 143341, + "school": 21850, + "schools": 90656, + "schüt": 137711, + "schütz": 137712, + "sci": 53827, + "science": 39557, + "scient": 55347, + "scientific": 84830, + "scious": 23121, + "sciously": 70788, + "scope": 4186, + "scoped": 78740, + "scopes": 80925, + "scopic": 57410, + "scopy": 50756, + "score": 12338, + "scores": 46695, + "scp": 63063, + "scr": 25065, + "scratch": 54055, + "screen": 8174, + "screens": 48816, + "scri": 11592, + "scribe": 6273, + "scribed": 49785, + "scriber": 20351, + "scribers": 46361, + "scripción": 141275, + "script": 2282, + "scriptId": 48855, + "scription": 12124, + "scriptions": 29966, + "scripts": 19831, + "scrição": 136407, + "scroll": 12605, + "scrollTop": 46724, + "scrollView": 49214, + "scss": 63298, + "sd": 13446, + "sdale": 83682, + "sdk": 51295, + "se": 325, + "sea": 36441, + "sealed": 75940, + "search": 1836, + "searchModel": 64584, + "season": 16798, + "seat": 43058, + "seau": 69399, + "sec": 5024, + "second": 5569, + "secondary": 18699, + "seconds": 17403, + "secret": 20474, + "secs": 53281, + "sect": 9687, + "section": 2809, + "sections": 24661, + "sector": 44757, + "sects": 54429, + "secure": 25132, + "secured": 84308, + "security": 17039, + "secutive": 85780, + "sed": 31324, + "see": 4060, + "seealso": 50762, + "seed": 22602, + "seeing": 65054, + "seek": 25713, + "seen": 28312, + "seg": 14607, + "segment": 23169, + "segments": 56829, + "segue": 46106, + "sehen": 65423, + "sei": 72768, + "sein": 48435, + "seite": 61065, + "sek": 56267, + "sel": 9507, + "select": 1742, + "selectAll": 81473, + "selected": 4525, + "selectedIndex": 74826, + "selection": 23617, + "selector": 8925, + "selectorMethod": 89312, + "selectors": 95264, + "self": 721, + "sell": 34835, + "seller": 34099, + "selling": 37737, + "sem": 6998, + "semantic": 47858, + "semb": 4838, + "sembl": 27013, + "semble": 15790, + "sembled": 25905, + "sembler": 35401, + "sembles": 40694, + "semblies": 45338, + "sembling": 41964, + "sembly": 5642, + "semester": 61536, + "semi": 86598, + "semicolon": 44761, + "sen": 11759, + "senal": 25499, + "send": 6681, + "sendKeys": 63054, + "sendMessage": 61957, + "sender": 11644, + "sending": 79696, + "senha": 61409, + "sense": 97835, + "sensor": 29511, + "sent": 24512, + "sentence": 51889, + "seo": 44666, + "sep": 28036, + "separator": 40120, + "seq": 13262, + "sequ": 6603, + "sequelize": 57856, + "sequence": 15512, + "sequences": 61708, + "sequent": 71357, + "sequential": 92350, + "sequently": 38642, + "ser": 799, + "serde": 47024, + "serial": 10182, + "serialization": 74835, + "serialize": 24166, + "serialized": 75277, + "serializer": 52718, + "serie": 61143, + "series": 19880, + "serious": 93327, + "serir": 98487, + "serrat": 63064, + "sert": 6047, + "serter": 90727, + "serv": 30775, + "serve": 13267, + "server": 4030, + "servername": 69760, + "servers": 67696, + "service": 7936, + "serviceName": 93485, + "services": 12779, + "ses": 9275, + "sess": 39057, + "session": 5920, + "sessionId": 70515, + "sessions": 50964, + "set": 746, + "setAttribute": 31317, + "setBackground": 90525, + "setCellValue": 41269, + "setChecked": 64945, + "setColor": 54091, + "setContent": 68121, + "setCurrent": 62767, + "setData": 43891, + "setDefault": 92174, + "setDescription": 63542, + "setDisplay": 76355, + "setEnabled": 37109, + "setError": 98130, + "setFlash": 62389, + "setFont": 89402, + "setIcon": 69001, + "setId": 63820, + "setImage": 88756, + "setItem": 85023, + "setLabel": 88553, + "setLayout": 58270, + "setMax": 99017, + "setMessage": 73387, + "setName": 40730, + "setOnClickListener": 63628, + "setParameter": 52004, + "setPosition": 49310, + "setProperty": 80307, + "setQuery": 73228, + "setScale": 79674, + "setSize": 96907, + "setState": 78858, + "setStatus": 83451, + "setText": 20989, + "setTimeout": 78700, + "setTitle": 40244, + "setType": 70341, + "setUp": 62852, + "setValue": 22541, + "setVisibility": 95792, + "setVisible": 38262, + "setWidth": 89766, + "setattr": 87170, + "setq": 32526, + "sets": 4917, + "sett": 66514, + "setter": 88479, + "setting": 15320, + "settings": 6511, + "setup": 15188, + "setw": 51822, + "seud": 39412, + "seudo": 21952, + "seven": 80185, + "severity": 77019, + "sex": 11553, + "sexo": 83180, + "sexual": 43587, + "sexy": 98254, + "sey": 8506, + "seys": 36423, + "sez": 51970, + "sf": 17246, + "sg": 1991, + "sgi": 51612, + "sgiving": 31627, + "sh": 927, + "sha": 15247, + "shade": 86054, + "shader": 45534, + "shadow": 32952, + "shaft": 96413, + "shake": 29661, + "shal": 12214, + "shaled": 75303, + "shall": 19012, + "shan": 82127, + "shape": 12231, + "shapes": 92059, + "share": 19368, + "shared": 6100, + "shares": 75047, + "sharing": 83646, + "sharp": 52917, + "shaw": 50004, + "she": 31058, + "shed": 69461, + "sheet": 15119, + "sheets": 91437, + "shelf": 53950, + "shell": 21384, + "sher": 89505, + "shi": 59421, + "shield": 47952, + "shift": 13418, + "shima": 91477, + "shine": 30238, + "shint": 64623, + "ship": 5270, + "shipment": 69087, + "shipping": 26508, + "ships": 17675, + "shire": 14909, + "shirt": 48372, + "shit": 40053, + "shiv": 64730, + "shm": 92137, + "shoot": 67292, + "shop": 8675, + "shopping": 50566, + "shops": 26770, + "shore": 28426, + "short": 8676, + "shortcode": 83387, + "shortcut": 46357, + "shot": 6340, + "shots": 27634, + "should": 5445, + "shouldBe": 77261, + "shouldReceive": 85708, + "show": 3445, + "shown": 69363, + "shows": 59456, + "shr": 66039, + "shuffle": 65355, + "shutdown": 59547, + "si": 6321, + "siblings": 51920, + "sic": 51777, + "sid": 21027, + "side": 2929, + "sidebar": 18989, + "sie": 78612, + "sig": 21339, + "sigma": 24912, + "sigmoid": 88927, + "sign": 7752, + "signIn": 60747, + "signal": 26622, + "signals": 60615, + "signature": 34140, + "signed": 2215, + "significant": 90545, + "signin": 55612, + "signup": 28725, + "sil": 34804, + "silent": 68658, + "silver": 81914, + "sim": 14781, + "simd": 96258, + "similar": 64287, + "simp": 20845, + "simple": 22944, + "simulate": 86711, + "simulation": 83772, + "sin": 15940, + "since": 11284, + "sing": 79307, + "single": 15338, + "singleton": 59915, + "singular": 92309, + "sink": 66738, + "sip": 88250, + "sis": 14406, + "sist": 139224, + "sistência": 139225, + "sit": 46865, + "site": 9513, + "sitemap": 37963, + "sites": 54764, + "six": 50364, + "size": 2141, + "sizei": 60643, + "sizeof": 10318, + "sizes": 53155, + "sión": 128610, + "sj": 22910, + "sk": 4886, + "skb": 66109, + "ski": 32539, + "skill": 29454, + "skills": 53240, + "skin": 36182, + "skins": 49318, + "skip": 20599, + "skirts": 66975, + "sku": 39929, + "sky": 26684, + "skór": 133745, + "skÄħ": 131944, + "sl": 3226, + "sla": 51642, + "slack": 88548, + "slash": 50256, + "slashes": 47629, + "slaught": 77068, + "slave": 52505, + "sleep": 25809, + "slice": 24963, + "slick": 69158, + "slide": 19341, + "slideDown": 89847, + "slideUp": 78274, + "slider": 20349, + "slides": 65821, + "slope": 96512, + "slot": 21537, + "slots": 49767, + "slow": 35211, + "slt": 48893, + "slug": 15125, + "sm": 3563, + "small": 9004, + "smart": 39803, + "smarty": 11818, + "smith": 33017, + "smooth": 56866, + "smouth": 76577, + "sms": 41392, + "smtp": 39854, + "sn": 9613, + "snake": 72139, + "snap": 48854, + "snapshot": 35501, + "snd": 46395, + "snippet": 57615, + "snow": 74478, + "sns": 72494, + "so": 704, + "soap": 35012, + "soc": 57021, + "social": 22386, + "sock": 13199, + "socket": 9556, + "sockets": 76729, + "sockopt": 58576, + "soever": 34554, + "sofar": 93571, + "soft": 3510, + "softmax": 77499, + "software": 36651, + "sol": 38198, + "sold": 62478, + "sole": 1546, + "soles": 40400, + "solete": 27246, + "solid": 31027, + "solution": 49460, + "solve": 59419, + "solver": 75501, + "som": 56739, + "some": 14689, + "someone": 89479, + "something": 33331, + "sometimes": 56653, + "son": 930, + "song": 31751, + "songs": 75052, + "sonian": 77613, + "sono": 73122, + "sons": 37348, + "soon": 66056, + "sorry": 67597, + "sort": 6860, + "sortBy": 65032, + "sortable": 65598, + "sorted": 28298, + "sorting": 67039, + "sound": 27532, + "sounds": 70531, + "soup": 89542, + "source": 2427, + "sources": 39651, + "south": 66484, + "sov": 84293, + "sp": 2154, + "spNet": 14349, + "spa": 89198, + "space": 8746, + "spacer": 97065, + "spaces": 44285, + "spacing": 23436, + "spam": 75545, + "span": 1480, + "spar": 77884, + "spark": 53782, + "sparse": 98223, + "spath": 30089, + "spawn": 46087, + "spb": 43467, + "spd": 90763, + "spe": 33374, + "speaker": 64476, + "spec": 9535, + "special": 15144, + "specialchars": 25520, + "species": 42490, + "specific": 51240, + "specified": 53434, + "specifier": 67251, + "specs": 94531, + "spect": 67082, + "speech": 88225, + "speed": 20374, + "spell": 37655, + "spender": 92825, + "spent": 94364, + "sphere": 87410, + "spi": 39157, + "spiel": 80885, + "spin": 41582, + "spinner": 62091, + "spir": 49077, + "spiracy": 69723, + "spirit": 85290, + "spl": 69095, + "splash": 48632, + "splice": 94960, + "split": 6960, + "spm": 68622, + "spo": 82859, + "spoken": 51758, + "sponsor": 95811, + "sport": 60901, + "sports": 83560, + "spot": 18648, + "spotify": 87388, + "spots": 67010, + "spoÅĤ": 132531, + "spoÅĤec": 132532, + "spoÅĤeczn": 132533, + "spr": 51673, + "spread": 58195, + "spring": 29223, + "sprintf": 49727, + "sprite": 32641, + "sprites": 87470, + "spy": 28191, + "sq": 28343, + "sql": 3544, + "sqlite": 37042, + "sqrt": 26888, + "squ": 97407, + "square": 37476, + "squeeze": 51913, + "sr": 15094, + "src": 3548, + "srv": 40020, + "ss": 778, + "ssa": 52107, + "ssc": 89091, + "sse": 64513, + "ssel": 89182, + "ssf": 83174, + "ssh": 25537, + "ssi": 45656, + "ssid": 61743, + "ssize": 77460, + "ssl": 24635, + "sson": 29931, + "ssp": 67233, + "ssql": 79713, + "sss": 84160, + "sst": 50900, + "sstream": 38366, + "ssue": 83890, + "ssé": 131875, + "st": 267, + "sta": 20491, + "stab": 67488, + "stable": 27992, + "stack": 7693, + "stackoverflow": 42035, + "stackpath": 68194, + "stacle": 43598, + "stacles": 73979, + "stad": 46840, + "stadt": 44278, + "staff": 27122, + "stag": 46680, + "stage": 20743, + "stairs": 30134, + "stakes": 95060, + "staking": 88885, + "stalk": 89749, + "stamp": 49113, + "stan": 24025, + "stanbul": 45116, + "stance": 1309, + "stances": 9436, + "stand": 2685, + "standard": 20270, + "standen": 84884, + "standing": 10070, + "stands": 96715, + "stant": 4706, + "stantial": 75957, + "stantiate": 42251, + "stantiateViewController": 82783, + "stants": 6886, + "star": 11870, + "stars": 36638, + "start": 2468, + "startDate": 40807, + "startIndex": 95311, + "startTime": 40906, + "startdate": 95398, + "started": 46723, + "starter": 31374, + "starting": 39289, + "starts": 65876, + "startsWith": 57936, + "startswith": 69325, + "starttime": 88786, + "startup": 80665, + "stash": 49771, + "stashop": 83074, + "stasy": 72755, + "stat": 9878, + "state": 2454, + "stateParams": 55069, + "stateProvider": 79349, + "statement": 24184, + "statements": 94587, + "states": 32069, + "static": 1978, + "staticmethod": 22215, + "station": 20155, + "stations": 74628, + "statistics": 54120, + "stats": 16260, + "statt": 96346, + "status": 2829, + "statusCode": 44480, + "statuses": 65353, + "staw": 49011, + "stay": 58921, + "std": 1834, + "stdClass": 78848, + "stdafx": 56306, + "stdarg": 88268, + "stdbool": 62556, + "stdcall": 60789, + "stddef": 66397, + "stderr": 36422, + "stdexcept": 94513, + "stdin": 51602, + "stdint": 31435, + "stdio": 10345, + "stdlib": 13149, + "stdout": 36358, + "ste": 5342, + "stead": 42180, + "steady": 71666, + "steam": 46590, + "sted": 23892, + "steder": 41889, + "steel": 61554, + "steen": 93957, + "stein": 12429, + "stell": 41137, + "stellar": 77293, + "stellen": 58158, + "steller": 72511, + "stellung": 54377, + "stem": 64088, + "sten": 15795, + "step": 9520, + "steps": 24080, + "ster": 3667, + "sterdam": 26302, + "stered": 82880, + "sterol": 32182, + "sterreich": 68398, + "sters": 12570, + "stery": 75970, + "sth": 30077, + "sthrough": 86901, + "sti": 83419, + "stial": 50606, + "stice": 93703, + "stick": 29033, + "sticks": 68945, + "sticky": 55442, + "stile": 46504, + "still": 43366, + "stim": 53654, + "stime": 86094, + "stin": 47461, + "stinence": 94721, + "stinian": 87357, + "stit": 3696, + "stitial": 48170, + "stitución": 139535, + "stituição": 136647, + "stitute": 7660, + "stitution": 10446, + "stitutions": 92678, + "stk": 75309, + "stm": 24855, + "stmt": 9215, + "sto": 33052, + "stock": 13479, + "stocks": 68027, + "stoff": 79956, + "stoi": 87241, + "ston": 7720, + "stone": 10812, + "stones": 32510, + "stood": 51684, + "stop": 9495, + "stopped": 89863, + "stops": 98020, + "stor": 47365, + "storage": 16172, + "store": 4314, + "storeId": 84397, + "stored": 71249, + "stores": 43469, + "stories": 68235, + "storm": 26413, + "storms": 82009, + "story": 26485, + "storybook": 90661, + "stown": 42958, + "stp": 91462, + "str": 495, + "stra": 13342, + "stract": 4339, + "straction": 42602, + "stractions": 64690, + "straight": 88192, + "strain": 77626, + "strained": 57727, + "straint": 4655, + "straints": 7458, + "stral": 54970, + "strand": 85881, + "strap": 5092, + "strar": 22189, + "stras": 58787, + "strate": 69915, + "strategy": 61914, + "stration": 54581, + "straÃŁe": 135786, + "strcasecmp": 68852, + "strcmp": 14222, + "strconv": 51848, + "strcpy": 89729, + "stre": 66423, + "stream": 4027, + "streams": 60975, + "street": 27351, + "strength": 74237, + "stress": 94190, + "stretch": 72084, + "strftime": 43737, + "stri": 29872, + "strict": 6627, + "stricted": 73072, + "striction": 51130, + "stride": 47057, + "strike": 73004, + "string": 917, + "stringLiteral": 59305, + "stringValue": 78883, + "strings": 18594, + "stringstream": 28803, + "strip": 13105, + "stripe": 61233, + "strlen": 20877, + "strncmp": 46787, + "stro": 95941, + "stroke": 28241, + "strom": 34055, + "strong": 4519, + "stroy": 10966, + "strpos": 48808, + "strstr": 71266, + "strt": 39126, + "strtolower": 84469, + "strtotime": 50962, + "struct": 1235, + "struction": 3024, + "structions": 19657, + "structor": 3086, + "structors": 91366, + "structural": 95697, + "structure": 7837, + "structured": 51143, + "structures": 46094, + "structuring": 62172, + "struk": 95194, + "strument": 19474, + "struments": 37718, + "strup": 89399, + "stry": 4617, + "sts": 36279, + "stu": 60682, + "stub": 59398, + "stud": 75343, + "student": 12038, + "students": 37218, + "studio": 59019, + "study": 54965, + "stuff": 57304, + "sty": 78755, + "styl": 85588, + "style": 3528, + "styleType": 49509, + "styled": 22849, + "styles": 3954, + "stylesheet": 6793, + "stype": 73520, + "stä": 132684, + "ständ": 47635, + "stände": 138087, + "stär": 139522, + "stärke": 139523, + "stÃ¥": 96490, + "stück": 135522, + "stüt": 57849, + "stÄħ": 131728, + "stÄħpi": 131729, + "stÄĻp": 128860, + "su": 27051, + "sub": 1966, + "subcategory": 79875, + "subclass": 99117, + "subj": 78014, + "subject": 11501, + "subjects": 64782, + "submenu": 60984, + "submission": 57108, + "submit": 5955, + "submitButton": 82047, + "submitted": 72463, + "subnet": 88636, + "subpackage": 42837, + "subplot": 49856, + "subs": 64798, + "subscribe": 9384, + "subscriber": 59205, + "subscription": 34404, + "subscriptions": 72560, + "subsection": 68874, + "subset": 38253, + "subseteq": 52396, + "subst": 52802, + "substr": 28614, + "substring": 33874, + "subtitle": 40417, + "subtotal": 73483, + "subtract": 59442, + "subtype": 77241, + "succ": 64441, + "success": 5630, + "successful": 30950, + "successfully": 60505, + "such": 20805, + "sudo": 18881, + "suffix": 26786, + "suggest": 95761, + "suit": 72040, + "suite": 49992, + "sum": 1242, + "suma": 94809, + "sume": 31323, + "sumer": 11761, + "suming": 61117, + "summ": 69544, + "summary": 1708, + "summer": 62566, + "sun": 39519, + "sup": 12776, + "super": 9522, + "supplier": 39830, + "supply": 94683, + "support": 23362, + "supported": 18216, + "supports": 77709, + "suppress": 96960, + "sur": 19784, + "sure": 19098, + "surf": 77860, + "surface": 39746, + "surname": 51415, + "survey": 47454, + "sus": 85421, + "suspend": 98611, + "sut": 99020, + "sv": 3492, + "svc": 58094, + "svg": 15196, + "sville": 49928, + "svm": 96251, + "svn": 63589, + "svp": 80606, + "sw": 2280, + "swagen": 43577, + "swagger": 74755, + "swana": 93710, + "swap": 25741, + "sweet": 94828, + "swer": 3559, + "swers": 9409, + "swick": 56072, + "swift": 70642, + "swing": 85284, + "swiper": 53227, + "swire": 79725, + "switch": 17338, + "swith": 22177, + "sword": 79038, + "sworth": 60731, + "sx": 48612, + "sy": 22860, + "sylvania": 18480, + "sym": 23802, + "symbol": 18785, + "symbols": 67426, + "syn": 20339, + "sync": 12996, + "syntax": 56193, + "synthesize": 30883, + "sys": 7791, + "syscall": 71069, + "system": 8948, + "systems": 45554, + "sz": 14357, + "szczÄĻ": 132230, + "szed": 139632, + "szedÅĤ": 139633, + "szÄħ": 129343, + "szÄĻ": 130620, + "são": 128791, + "sätze": 140053, + "sı": 126203, + "sına": 129561, + "sında": 128892, + "sını": 129489, + "sının": 130182, + "sınız": 131038, + "sıyla": 131720, + "sız": 128894, + "sÄħd": 141820, + "sÅĤ": 142202, + "sÅĤaw": 132224, + "sÅĤuch": 142203, + "t": 83, + "ta": 2565, + "tab": 6192, + "tabl": 73099, + "tabla": 57354, + "table": 2005, + "tableFuture": 69629, + "tableName": 35977, + "tableView": 14430, + "tablename": 42679, + "tables": 32110, + "tablet": 98693, + "taboola": 48489, + "tabpanel": 73246, + "tabs": 30993, + "tag": 4578, + "tagName": 73273, + "tags": 14082, + "tahun": 65996, + "taient": 84706, + "tail": 14586, + "tails": 39960, + "tain": 22921, + "taire": 70137, + "take": 22769, + "taken": 82161, + "takes": 77979, + "taking": 88794, + "tal": 50877, + "talk": 35241, + "tam": 72707, + "tan": 52591, + "tanggal": 48699, + "tank": 85171, + "tant": 60412, + "tap": 30047, + "tape": 63520, + "tar": 26737, + "target": 5657, + "targets": 45547, + "tas": 47542, + "task": 8202, + "taskId": 80688, + "tasks": 24760, + "tat": 67138, + "tatus": 78321, + "tau": 29143, + "tax": 17616, + "taxonomy": 53004, + "taÅĤ": 128055, + "taÅŁ": 132318, + "tb": 18387, + "tbl": 18064, + "tbody": 10095, + "tc": 10413, + "tcp": 27161, + "td": 1296, + "tdown": 71546, + "te": 665, + "tea": 81594, + "teacher": 23360, + "teachers": 80980, + "team": 9196, + "teams": 38496, + "tearDown": 82892, + "teborg": 77429, + "tec": 54200, + "tech": 17785, + "techn": 25444, + "technical": 72137, + "technology": 58242, + "tection": 39141, + "ted": 6565, + "tee": 6547, + "teen": 14901, + "teenth": 45387, + "tees": 29800, + "tega": 91905, + "teil": 55632, + "tein": 38240, + "tek": 42713, + "tel": 22924, + "tele": 15410, + "telefone": 78130, + "telefono": 51208, + "telegram": 81669, + "telephone": 51927, + "tell": 72357, + "tem": 870, + "tement": 78407, + "temp": 3888, + "temperature": 34558, + "templ": 74860, + "template": 4214, + "templates": 15463, + "temporary": 78090, + "temps": 66280, + "tempt": 4213, + "tempts": 33272, + "tems": 90837, + "ten": 1960, + "tenant": 43919, + "tener": 73500, + "teness": 63097, + "tensor": 46111, + "tensorflow": 84282, + "tent": 72368, + "tep": 90898, + "ter": 465, + "tera": 50037, + "terdam": 50129, + "tere": 54604, + "tered": 33159, + "tering": 59385, + "terior": 40544, + "term": 4991, + "termin": 22969, + "terminal": 36327, + "terminate": 48061, + "terminated": 68659, + "termination": 88676, + "terms": 18347, + "tern": 4160, + "ternal": 15342, + "tero": 90389, + "terr": 68669, + "terra": 47821, + "terraform": 61385, + "terrain": 46782, + "territ": 95215, + "terror": 37388, + "terrorism": 55563, + "ters": 5045, + "tery": 17615, + "tes": 2338, + "test": 1944, + "testCase": 66194, + "testdata": 92425, + "teste": 88999, + "tested": 74936, + "tester": 73358, + "testimonial": 80745, + "testing": 8840, + "tests": 23841, + "tesy": 18882, + "tet": 72442, + "tex": 33344, + "text": 1318, + "textAlign": 81175, + "textBox": 39598, + "textContent": 43435, + "textField": 37782, + "textInput": 66997, + "textTheme": 93723, + "textView": 84228, + "textarea": 11741, + "textbox": 77375, + "textfield": 61763, + "texto": 57003, + "texts": 85942, + "texture": 27496, + "textures": 70030, + "tf": 8935, + "tfoot": 58484, + "tg": 41428, + "tgl": 45078, + "tgt": 85226, + "th": 339, + "tha": 22410, + "thag": 95362, + "thal": 30292, + "thalm": 77643, + "than": 53795, + "thane": 85085, + "thank": 57417, + "thanks": 45493, + "that": 9033, + "the": 1782, + "theValue": 73277, + "thead": 11417, + "theast": 14759, + "thed": 78271, + "theid": 68355, + "their": 49728, + "theless": 19986, + "them": 48718, + "theme": 9047, + "themes": 40715, + "then": 3391, + "thenReturn": 26842, + "theon": 63010, + "theorem": 31515, + "theory": 93204, + "ther": 696, + "therapy": 45655, + "there": 18532, + "thermal": 68118, + "thern": 6322, + "theros": 90782, + "thers": 2985, + "thes": 6375, + "these": 43910, + "theses": 38322, + "thesis": 13189, + "thesize": 26887, + "thesized": 82879, + "thest": 60843, + "theta": 15976, + "thetic": 17556, + "thew": 15765, + "they": 20069, + "thic": 81564, + "thickness": 90967, + "thin": 63671, + "thing": 1596, + "things": 57074, + "think": 26865, + "thinkable": 90103, + "thinking": 82260, + "third": 31727, + "this": 574, + "thon": 4587, + "thood": 31868, + "thora": 64683, + "those": 53619, + "though": 4535, + "thought": 60565, + "thouse": 55034, + "thr": 60024, + "thread": 4528, + "threads": 27286, + "threat": 96129, + "three": 27856, + "thren": 62264, + "thresh": 94382, + "threshold": 28902, + "thritis": 39385, + "thro": 92110, + "throat": 86064, + "throp": 97543, + "through": 19742, + "throw": 7119, + "thrown": 86445, + "throws": 8100, + "ths": 16896, + "thumb": 25036, + "thumbnail": 27786, + "thumbnails": 86109, + "thumbs": 95390, + "thur": 21456, + "thus": 33132, + "thy": 26422, + "thè": 128213, + "thèque": 143536, + "thèse": 143387, + "thé": 131157, + "ti": 10251, + "tic": 28050, + "tica": 81955, + "tical": 34658, + "tick": 34155, + "ticker": 68345, + "ticket": 26534, + "tickets": 68727, + "ticks": 35078, + "tics": 12734, + "tid": 24449, + "tie": 48731, + "tier": 49335, + "ties": 28546, + "tif": 84340, + "tight": 74182, + "til": 1646, + "tile": 21094, + "tiles": 60882, + "tility": 12189, + "tim": 20091, + "time": 1678, + "timeline": 48960, + "timeofday": 59523, + "timeout": 14150, + "timer": 19278, + "times": 15136, + "timestamp": 13035, + "timestamps": 33591, + "timezone": 43271, + "timing": 89760, + "tin": 55971, + "ting": 1280, + "tingham": 59135, + "tings": 2065, + "tiny": 46116, + "tion": 27391, + "tip": 13145, + "tipo": 21194, + "tips": 45454, + "tit": 16903, + "title": 2102, + "titleLabel": 88446, + "titles": 34523, + "titre": 69581, + "titulo": 34963, + "tivitÃł": 138580, + "tiÄŁi": 126570, + "tk": 6242, + "tl": 11544, + "tle": 11239, + "tlement": 44489, + "tls": 34488, + "tm": 13730, + "tml": 3864, + "tmp": 5173, + "tmpl": 71359, + "tn": 1517, + "to": 983, + "toArray": 27774, + "toBe": 8898, + "toBeDefined": 67538, + "toBeFalsy": 87453, + "toBeInTheDocument": 99039, + "toBeTruthy": 28033, + "toContain": 75804, + "toDate": 79459, + "toDouble": 64483, + "toEqual": 13226, + "toFixed": 36529, + "toFloat": 91297, + "toHave": 24720, + "toHaveBeenCalled": 33606, + "toHaveBeenCalledTimes": 82544, + "toHaveBeenCalledWith": 56147, + "toHaveLength": 87947, + "toISOString": 89988, + "toInt": 30960, + "toJson": 73985, + "toList": 41907, + "toLocale": 77385, + "toLowerCase": 27216, + "toMatch": 43548, + "toMatchSnapshot": 61426, + "toPromise": 81659, + "toString": 6575, + "toThrow": 88110, + "toUpperCase": 37086, + "toa": 49095, + "toast": 52101, + "tober": 52144, + "toc": 54780, + "tod": 67887, + "today": 30113, + "todo": 17370, + "todos": 49188, + "tog": 67173, + "toggle": 17703, + "toggleClass": 48279, + "togroup": 82861, + "toi": 52609, + "tok": 29594, + "token": 5839, + "tokenId": 98214, + "tokenizer": 85593, + "tokens": 30566, + "tol": 20576, + "tolist": 52141, + "tolower": 38993, + "tolua": 94681, + "tom": 37401, + "ton": 777, + "tone": 57929, + "tones": 91645, + "tons": 34411, + "too": 36127, + "took": 90353, + "tool": 14172, + "toolStrip": 65539, + "toolbar": 37530, + "toolbox": 87183, + "tools": 15918, + "tooltip": 21539, + "top": 3481, + "topic": 16411, + "topics": 55152, + "topl": 98605, + "tor": 10980, + "torch": 27414, + "torrent": 70230, + "tors": 97166, + "tos": 42581, + "tot": 27779, + "total": 5035, + "totalCount": 97036, + "totals": 84846, + "touch": 22020, + "touches": 76953, + "toupper": 87732, + "tour": 51339, + "tournament": 88079, + "tout": 66333, + "tower": 77578, + "town": 27831, + "toy": 81225, + "tp": 790, + "tparam": 68071, + "tpl": 12620, + "tplib": 94030, + "tps": 96031, + "tr": 376, + "tra": 2172, + "trace": 15067, + "tracer": 94941, + "track": 13131, + "tracked": 58381, + "tracker": 50395, + "tracking": 44553, + "tracks": 49324, + "tract": 2144, + "tracted": 54435, + "traction": 26425, + "tractive": 69078, + "tractor": 54766, + "tracts": 36188, + "trad": 47026, + "trade": 36071, + "traditional": 86908, + "traffic": 53987, + "trag": 52015, + "trail": 77875, + "train": 10397, + "trained": 35722, + "trainer": 83776, + "training": 30654, + "trait": 29432, + "traits": 62633, + "trajectory": 83787, + "trak": 70019, + "tram": 97940, + "tran": 43369, + "trand": 94574, + "trans": 1458, + "transaction": 13528, + "transactions": 41844, + "transfer": 24188, + "transform": 4701, + "transforms": 94833, + "transition": 20454, + "translate": 14045, + "translated": 53242, + "translation": 3053, + "translations": 58450, + "translator": 44357, + "transparent": 28269, + "transport": 26445, + "transpose": 54487, + "trap": 55914, + "tras": 13503, + "trash": 68826, + "trasound": 53765, + "travel": 48510, + "traî": 138377, + "traîn": 138378, + "tre": 10157, + "trecht": 93870, + "tree": 9344, + "trees": 90412, + "treeview": 79205, + "trer": 94820, + "tres": 44532, + "trfs": 80012, + "tri": 22446, + "trial": 47347, + "triangle": 55114, + "trib": 1332, + "tribute": 2199, + "tributes": 3604, + "tribution": 31140, + "tridge": 36877, + "tridges": 62831, + "trie": 8927, + "tries": 4543, + "trieve": 34216, + "trigger": 18468, + "tright": 68976, + "trim": 10666, + "trimmed": 98101, + "tring": 561, + "trinsic": 45002, + "trip": 32981, + "trl": 9599, + "tro": 32442, + "tron": 34685, + "trs": 85385, + "tru": 65253, + "truck": 98318, + "true": 1866, + "truncate": 66054, + "trust": 56655, + "trusted": 83837, + "truth": 58577, + "trx": 72461, + "try": 1539, + "trying": 84034, + "trys": 39607, + "tryside": 44255, + "trzym": 128122, + "trzymaÄĩ": 142411, + "trzymaÅĤ": 140350, + "trä": 140463, + "trägt": 140464, + "tré": 133349, + "tréal": 98177, + "ts": 2576, + "tsky": 81236, + "tsx": 87233, + "tsy": 58960, + "tt": 5566, + "tte": 39836, + "tti": 97465, + "ttl": 62858, + "ttp": 1209, + "tty": 81711, + "tu": 24494, + "tube": 73738, + "tuple": 24590, + "tur": 65367, + "ture": 4974, + "tures": 18513, + "turn": 412, + "turned": 41186, + "turnstile": 56190, + "turtle": 81109, + "tutorial": 91782, + "tv": 22209, + "tw": 15560, + "tweet": 35387, + "tweets": 79721, + "twenty": 93965, + "twig": 71653, + "twitter": 14679, + "two": 19789, + "twor": 142676, + "tworzyÄĩ": 142677, + "twór": 140086, + "tx": 3998, + "txn": 73370, + "txt": 8586, + "ty": 1881, + "tyard": 55536, + "typ": 3653, + "type": 1313, + "typeName": 85491, + "typed": 44181, + "typedef": 5286, + "typeid": 88342, + "typename": 9031, + "typeof": 10222, + "typeorm": 80059, + "typeparam": 26946, + "types": 9242, + "typescript": 49552, + "typically": 86084, + "typing": 89802, + "typings": 85735, + "tz": 37592, + "tá": 128786, + "té": 128294, + "téri": 131783, + "tó": 128644, + "tón": 134823, + "tória": 132327, + "tório": 129741, + "tów": 128771, + "tör": 130457, + "tü": 126439, + "tür": 125601, + "türk": 134631, + "tÃł": 126318, + "tÃŃ": 129459, + "tı": 125453, + "tık": 129763, + "tım": 132948, + "tır": 125501, + "tıģ": 127929, + "tıģı": 129962, + "tıģımız": 142221, + "tıģını": 137550, + "tÄħ": 130124, + "u": 84, + "uC": 91011, + "uD": 71872, + "uParam": 91587, + "ua": 4284, + "uable": 13128, + "uably": 31664, + "uada": 85100, + "uae": 75517, + "uai": 84695, + "uais": 72991, + "ual": 928, + "uala": 70120, + "uale": 49311, + "uales": 40752, + "uali": 51166, + "uality": 10733, + "ually": 1832, + "uan": 10386, + "uang": 68610, + "uania": 62511, + "uant": 27574, + "uar": 18731, + "uard": 11034, + "uards": 54608, + "uario": 6080, + "uarios": 20240, + "uars": 66387, + "uart": 27170, + "uary": 3536, + "uat": 54182, + "uate": 6292, + "uated": 12852, + "uates": 50985, + "uating": 53543, + "uation": 4002, + "uations": 37070, + "uator": 45162, + "uação": 72326, + "ub": 392, + "uba": 30429, + "ubah": 81324, + "ubar": 44497, + "ubb": 20683, + "ubber": 91886, + "ubble": 14600, + "ubbles": 34295, + "ubbo": 97737, + "ubby": 54854, + "ube": 3760, + "ubectl": 53380, + "uben": 75967, + "uber": 29870, + "ubern": 76259, + "ubernetes": 29827, + "ubes": 50625, + "ubi": 54233, + "ubic": 41181, + "ubishi": 62220, + "ubit": 59220, + "ubits": 81402, + "uble": 40472, + "ublic": 475, + "ublish": 2538, + "ublished": 11669, + "ublisher": 15182, + "ubo": 96119, + "ubre": 63259, + "ubs": 15738, + "ubuntu": 37979, + "uby": 20028, + "ubyte": 70388, + "uc": 1754, + "ucas": 48846, + "ucc": 14570, + "ucceed": 29264, + "ucceeded": 43805, + "uccess": 2557, + "ucch": 70566, + "ucchini": 85570, + "ucci": 52369, + "ucción": 67736, + "uce": 10521, + "uced": 23967, + "ucene": 58054, + "ucer": 59396, + "uces": 62271, + "uch": 1387, + "ucha": 54381, + "uchar": 29272, + "uche": 33827, + "uchen": 44977, + "ucher": 73612, + "uchi": 54917, + "uchos": 67250, + "uchs": 53152, + "uchsia": 73391, + "ucht": 32030, + "uci": 42409, + "ucid": 72687, + "ucing": 58044, + "ucion": 42277, + "uciones": 66247, + "ución": 35223, + "uck": 1942, + "ucked": 39358, + "ucken": 96990, + "ucker": 24370, + "ucket": 11152, + "uckets": 38551, + "ucking": 46170, + "uckland": 47878, + "uckle": 55975, + "uckles": 83696, + "ucks": 15582, + "ucky": 10073, + "ucle": 22147, + "uclear": 9637, + "uco": 94414, + "ucose": 93597, + "ucs": 42089, + "ucson": 59285, + "uct": 667, + "uction": 2654, + "uctions": 91463, + "uctive": 58823, + "uctor": 79222, + "uctose": 86773, + "ucumber": 40912, + "ucursal": 82230, + "ucus": 37501, + "ucz": 97567, + "ud": 661, + "uda": 8053, + "udad": 24776, + "udades": 92823, + "udas": 53741, + "udd": 8347, + "udded": 85792, + "udden": 37995, + "uddenly": 43706, + "udder": 56129, + "uddle": 74984, + "uddled": 84101, + "uddy": 33917, + "ude": 793, + "udeau": 40496, + "uded": 37377, + "udem": 87205, + "uden": 60127, + "udence": 73093, + "udent": 4881, + "udents": 16286, + "uder": 32619, + "udes": 28146, + "udev": 77297, + "udge": 19561, + "udget": 6312, + "udging": 87312, + "udi": 20536, + "udiant": 40839, + "udiante": 49663, + "udiantes": 68754, + "udies": 57687, + "uding": 50767, + "udio": 3230, + "udios": 88785, + "udit": 33887, + "udo": 7680, + "udoku": 68302, + "udos": 45079, + "udp": 31101, + "uds": 27988, + "udson": 31678, + "udu": 87207, + "udy": 17858, + "ue": 361, + "ueba": 42412, + "uebas": 90683, + "ueblo": 60750, + "ued": 3260, + "uede": 63667, + "uego": 26681, + "uegos": 86206, + "ueil": 51808, + "uel": 4000, + "uela": 23737, + "uele": 74343, + "ueling": 78997, + "uell": 41658, + "uelle": 29937, + "uellement": 76676, + "uellen": 70628, + "uelles": 64732, + "uels": 60144, + "uelve": 46856, + "uely": 66410, + "uem": 69847, + "uen": 10316, + "uencia": 63972, + "uent": 11680, + "uenta": 38352, + "uentes": 68737, + "uento": 55288, + "uer": 8801, + "uerdo": 52470, + "uers": 83425, + "ues": 1137, + "uesday": 7047, + "uese": 35454, + "uess": 33301, + "uest": 3045, + "uesta": 19441, + "uestas": 57147, + "uestion": 8678, + "uesto": 31491, + "uestos": 64182, + "uestra": 75364, + "uet": 13807, + "uetooth": 16412, + "uetype": 98340, + "ueue": 6318, + "ueur": 78523, + "ueva": 50658, + "uevo": 32321, + "uez": 27600, + "uf": 1704, + "ufact": 4725, + "ufacturer": 31545, + "ufe": 66075, + "ufen": 28421, + "uff": 1362, + "uffed": 40573, + "uffer": 2040, + "uffers": 20342, + "ufficient": 26683, + "uffix": 13554, + "uffle": 13208, + "uffled": 42335, + "uffles": 77453, + "uffling": 79690, + "uffman": 77708, + "uffs": 32568, + "uffy": 29877, + "ufig": 81548, + "uforia": 69754, + "ufreq": 91001, + "ufs": 87083, + "uft": 89131, + "ug": 768, + "uga": 36960, + "ugador": 81420, + "ugal": 44184, + "ugar": 8566, + "ugas": 57831, + "uge": 4733, + "ugen": 68989, + "ugeot": 93422, + "uger": 58420, + "uges": 50159, + "ugg": 2596, + "uggage": 43665, + "ugged": 20146, + "uggest": 3799, + "uggested": 53276, + "uggestion": 42116, + "uggestions": 37882, + "ugging": 35268, + "uggle": 49328, + "uggling": 61931, + "uggy": 60232, + "ugh": 7443, + "ughs": 31343, + "ught": 3863, + "ughter": 7340, + "ughters": 26505, + "ughty": 35516, + "ugi": 50540, + "ugin": 3631, + "ugins": 6840, + "uglify": 86532, + "ugo": 33501, + "ugs": 13299, + "ugu": 29785, + "uguay": 59203, + "uh": 12540, + "uhan": 69886, + "uhe": 91547, + "uhl": 95174, + "uhn": 92164, + "ui": 1963, + "uib": 79623, + "uibModal": 91109, + "uid": 2423, + "uida": 84320, + "uide": 34705, + "uido": 53076, + "uids": 91216, + "uien": 87074, + "uil": 57697, + "uild": 1498, + "uilder": 2511, + "uilt": 11227, + "uiltin": 25628, + "uin": 74621, + "uing": 7471, + "uint": 2496, + "uintptr": 51380, + "uir": 20911, + "uire": 51650, + "uis": 9241, + "uish": 50120, + "uisine": 55140, + "uisse": 91548, + "uit": 3083, + "uitable": 85481, + "uitar": 34746, + "uite": 9302, + "uiten": 86980, + "uition": 37061, + "uitive": 34151, + "uitka": 58840, + "uito": 25572, + "uits": 11797, + "uity": 34494, + "uição": 77023, + "uj": 9635, + "uja": 88708, + "uje": 33867, + "ujemy": 82328, + "ujet": 67907, + "uju": 72301, + "ujÄħ": 48078, + "ujÄħc": 129668, + "ujÄħce": 132869, + "ujÄħcy": 132154, + "ujÄħcych": 132467, + "ujÄĻ": 130813, + "uk": 3101, + "uka": 25196, + "ukan": 27724, + "uke": 10430, + "ukes": 64750, + "uki": 21504, + "ukkan": 67204, + "ukkit": 22793, + "uko": 71058, + "ukt": 37667, + "uktur": 84956, + "uku": 35734, + "ul": 360, + "ula": 5607, + "ulado": 84429, + "ulaire": 73675, + "ulan": 50735, + "ulance": 40832, + "ulant": 68633, + "ular": 1276, + "ulares": 88029, + "ulario": 32330, + "ularity": 28979, + "ulary": 23216, + "ulas": 25541, + "ulate": 6334, + "ulated": 7757, + "ulates": 23156, + "ulating": 15486, + "ulation": 2914, + "ulations": 7455, + "ulative": 22160, + "ulator": 10511, + "ulators": 41291, + "ulatory": 37120, + "ulação": 90503, + "uld": 36568, + "ule": 1111, + "uled": 8448, + "ulen": 94211, + "ulence": 58466, + "ulent": 28480, + "uler": 8479, + "ulerAngles": 58752, + "ulers": 26800, + "ules": 2425, + "ulet": 48418, + "ulf": 14308, + "ulfill": 86516, + "ulfilled": 84993, + "ulg": 54999, + "uli": 23602, + "ulia": 69745, + "uliar": 41618, + "ulin": 23401, + "uling": 12269, + "ulings": 67249, + "ulis": 64030, + "ulist": 96074, + "ulk": 21785, + "ulkan": 60538, + "ull": 617, + "ulla": 34078, + "ullah": 40797, + "ullan": 41120, + "ulle": 40054, + "ulled": 91022, + "ullen": 41474, + "ullet": 22376, + "ullets": 65430, + "ulling": 60636, + "ullivan": 36041, + "ullo": 64588, + "ulls": 70423, + "ully": 21965, + "ulner": 58401, + "ulnerable": 92907, + "ulo": 8763, + "ulong": 32832, + "ulos": 28652, + "ulous": 13540, + "ulously": 37562, + "ulp": 12840, + "ulpt": 23396, + "uls": 14295, + "ulse": 17217, + "ulses": 66016, + "ulsion": 32903, + "ulsive": 57821, + "ult": 494, + "ulta": 27157, + "ultan": 60093, + "ulti": 7068, + "ultimate": 66566, + "ultimo": 84316, + "ultip": 9845, + "ultipart": 18204, + "ultipartFile": 63300, + "ultiple": 12229, + "ultiply": 21352, + "ulton": 75476, + "ultur": 76431, + "ultural": 43447, + "ulture": 10523, + "ultureInfo": 52928, + "ulty": 14728, + "ultz": 54426, + "ulu": 24411, + "ului": 31648, + "ulum": 16496, + "ulumi": 65482, + "ulus": 19425, + "uly": 3901, + "um": 372, + "uma": 13416, + "umably": 74951, + "uman": 7136, + "umann": 63507, + "umar": 30867, + "umas": 53924, + "umat": 45213, + "umatic": 31930, + "umb": 3551, + "umba": 70569, + "umbai": 29855, + "umbed": 98596, + "umber": 890, + "umberland": 68759, + "umbing": 32446, + "umble": 22733, + "umbled": 25817, + "umbledore": 75318, + "umbles": 57099, + "umbling": 41632, + "umblr": 29601, + "umbn": 9670, + "umbnail": 11026, + "umbnails": 31312, + "umbo": 48609, + "umbotron": 57848, + "umbs": 15775, + "umd": 60245, + "ume": 3885, + "umed": 38155, + "umen": 27899, + "ument": 1123, + "uments": 2831, + "umer": 3389, + "umerator": 60969, + "umeric": 12572, + "umericUpDown": 41246, + "umes": 18618, + "umi": 24336, + "umidity": 73926, + "umin": 10110, + "uming": 29489, + "uminium": 39851, + "uminum": 18814, + "uml": 42231, + "umlah": 30327, + "umm": 25964, + "ummer": 62273, + "ummies": 49179, + "ummings": 82433, + "ummy": 8574, + "umn": 21371, + "umni": 56865, + "umno": 48347, + "umnos": 61636, + "umo": 64625, + "umont": 89817, + "umor": 68261, + "ump": 1510, + "umped": 60192, + "umper": 24027, + "umph": 30661, + "umping": 54332, + "umps": 11793, + "umpt": 49874, + "umptech": 62517, + "umption": 60574, + "umpy": 6752, + "ums": 6237, + "umu": 83209, + "un": 359, + "una": 8565, + "unable": 45928, + "unakan": 94104, + "unal": 25581, + "uname": 40379, + "unami": 61003, + "unan": 64634, + "unar": 84992, + "unas": 58264, + "unate": 21769, + "unately": 9505, + "unbind": 73111, + "unc": 1347, + "uncan": 37466, + "uncate": 26900, + "uncated": 38007, + "unce": 96693, + "unch": 3185, + "unchecked": 31684, + "unched": 10653, + "unci": 18424, + "uncia": 94485, + "unciation": 46310, + "uncio": 66044, + "uncios": 62156, + "unct": 19931, + "unction": 598, + "unctuation": 72299, + "uncture": 62874, + "und": 1241, + "unda": 26892, + "undai": 49681, + "undance": 97581, + "unday": 6676, + "unde": 28865, + "unded": 36053, + "undef": 16135, + "undefined": 9614, + "unden": 22157, + "under": 7995, + "underline": 45673, + "unders": 31009, + "underscore": 53933, + "undi": 55898, + "unding": 37189, + "undle": 4206, + "undler": 85365, + "undles": 49204, + "undo": 13499, + "undos": 64854, + "undra": 78730, + "undred": 7114, + "undreds": 45713, + "undry": 27247, + "unds": 42950, + "undy": 59185, + "une": 2886, + "uned": 48883, + "unei": 95223, + "uner": 47037, + "unes": 8531, + "unexpected": 53859, + "unfinished": 15092, + "unfold": 83704, + "unft": 60730, + "ung": 2185, + "unga": 91845, + "ungal": 57167, + "ungalow": 83919, + "ungan": 51188, + "unge": 13884, + "ungen": 11595, + "ungeon": 23452, + "ungeons": 50017, + "unger": 71622, + "ungi": 81490, + "ungkin": 97934, + "ungle": 28647, + "ungs": 28122, + "uni": 15705, + "uniacid": 88558, + "unic": 3525, + "unication": 39195, + "unicip": 13603, + "unicipio": 65167, + "unicode": 24519, + "unicorn": 92357, + "unidad": 82730, + "uniform": 38351, + "unifu": 22241, + "unik": 70449, + "uning": 37202, + "union": 16192, + "uniq": 54398, + "unique": 9587, + "unist": 15779, + "unistd": 27483, + "unit": 3843, + "unitOfWork": 73116, + "unities": 6037, + "units": 25643, + "unittest": 91005, + "unity": 2439, + "unitÃł": 128195, + "universal": 95300, + "unix": 56646, + "unj": 68914, + "unk": 3122, + "unken": 53929, + "unker": 79929, + "unknown": 16088, + "unks": 15296, + "unkt": 43226, + "unky": 69205, + "unless": 38770, + "unlikely": 55013, + "unlink": 67042, + "unload": 92665, + "unlock": 55021, + "unn": 14931, + "unnable": 82211, + "unned": 37990, + "unnel": 40292, + "unner": 54415, + "unning": 11216, + "unny": 27297, + "uno": 16311, + "unordered": 30419, + "unos": 53443, + "unpack": 80774, + "unprocessable": 67760, + "unque": 36527, + "unread": 96818, + "uns": 10862, + "unsafe": 38157, + "unsch": 80791, + "unset": 36804, + "unsigned": 3626, + "unsqueeze": 93636, + "unsubscribe": 90423, + "unsupported": 81145, + "unt": 3850, + "unta": 42050, + "untary": 58956, + "untas": 54572, + "unte": 10830, + "unted": 27636, + "unteer": 40905, + "unteers": 62684, + "unter": 29159, + "until": 38730, + "untime": 4466, + "unting": 26324, + "untlet": 85061, + "unto": 25420, + "untos": 49374, + "unts": 36940, + "untu": 18230, + "unu": 62253, + "unused": 24713, + "unwrap": 15454, + "unya": 81271, + "uo": 23137, + "uong": 68795, + "uos": 83708, + "uous": 9193, + "uously": 65635, + "up": 454, + "upa": 45831, + "upakan": 73736, + "upal": 20534, + "upaten": 74895, + "upd": 76618, + "update": 2386, + "updated": 11768, + "updatedAt": 69289, + "updates": 49661, + "updating": 92054, + "upe": 85477, + "uper": 3466, + "upert": 49480, + "upertino": 78454, + "upgrade": 44230, + "upid": 15884, + "upil": 93645, + "upiter": 19519, + "uple": 6061, + "uples": 28323, + "uplic": 9846, + "uplicate": 14070, + "uplicated": 98984, + "uplicates": 40614, + "upload": 6120, + "uploaded": 56883, + "uploader": 86790, + "uploads": 19374, + "upo": 20106, + "upon": 58623, + "upos": 86447, + "upp": 31832, + "uppe": 58498, + "uppen": 89097, + "upper": 13574, + "uppercase": 79882, + "uppet": 38836, + "uppies": 94747, + "upplier": 71691, + "upply": 83923, + "upport": 2800, + "upportInitialize": 12682, + "upported": 12513, + "uppy": 64029, + "upro": 96101, + "ups": 8602, + "upt": 7564, + "uptime": 74659, + "uptools": 51281, + "upuncture": 68462, + "upy": 57498, + "upyter": 72852, + "ur": 324, + "ura": 5690, + "uraa": 89495, + "urable": 18329, + "uracion": 58768, + "uracy": 29384, + "urahan": 95422, + "urai": 47913, + "ural": 4176, + "urally": 41924, + "uran": 26014, + "urance": 5771, + "urances": 63896, + "urança": 87156, + "urar": 75859, + "uras": 19319, + "urat": 90198, + "urate": 61159, + "uration": 2017, + "urations": 21449, + "urator": 57383, + "urb": 9048, + "urban": 59059, + "urbation": 64816, + "urbed": 74225, + "urch": 2574, + "urchase": 12877, + "urchased": 72250, + "urchases": 74102, + "urd": 16081, + "urdy": 39118, + "ure": 552, + "ureau": 17169, + "ured": 3073, + "ureen": 95572, + "ureka": 60522, + "urement": 38246, + "uren": 21211, + "urence": 86937, + "urent": 75965, + "urer": 7733, + "urers": 11807, + "ures": 1413, + "urette": 72143, + "urf": 29363, + "urface": 10509, + "urg": 5556, + "urga": 90776, + "urge": 39823, + "urgence": 71996, + "urgent": 85053, + "urgeon": 80462, + "urger": 34298, + "urgery": 84292, + "urgical": 56573, + "urgy": 61529, + "uri": 6070, + "uria": 74784, + "uridad": 45098, + "uries": 10396, + "urile": 92348, + "uring": 1677, + "urious": 27526, + "uristic": 34456, + "urities": 37233, + "urity": 5514, + "url": 1085, + "urlencode": 79472, + "urlpatterns": 39584, + "urls": 20502, + "urm": 52370, + "urma": 86579, + "urn": 399, + "urnal": 62450, + "urname": 19312, + "urning": 53344, + "urnished": 83439, + "urniture": 24578, + "uro": 2798, + "uron": 36090, + "urons": 53669, + "urope": 4084, + "uropean": 73287, + "uros": 43386, + "urous": 54609, + "urovision": 63009, + "urple": 35705, + "urpose": 29249, + "urr": 8027, + "urray": 23717, + "urre": 843, + "urrect": 58000, + "urrection": 39204, + "urred": 8130, + "urrenc": 4451, + "urrence": 20559, + "urrences": 33246, + "urrencies": 19607, + "urrency": 5088, + "urrent": 951, + "urret": 41354, + "urrets": 87334, + "urring": 20449, + "urry": 39856, + "urs": 1723, + "ursal": 54683, + "ursday": 7181, + "urse": 29285, + "ursed": 51732, + "urses": 56857, + "ursion": 34280, + "ursions": 75239, + "ursive": 16514, + "ursively": 79737, + "ursor": 3823, + "ursors": 33191, + "ursos": 47922, + "urst": 32612, + "urt": 5639, + "urther": 3917, + "urtle": 21744, + "urtles": 51114, + "uru": 20915, + "urus": 55166, + "urv": 96760, + "urve": 73047, + "urved": 79527, + "urvey": 38278, + "ury": 3350, + "us": 355, + "usa": 30753, + "usable": 22779, + "usage": 17698, + "usaha": 67030, + "usahaan": 75576, + "usal": 35414, + "usalem": 24260, + "usan": 40212, + "usat": 87432, + "usb": 24343, + "usband": 90981, + "usc": 25034, + "usch": 51078, + "use": 810, + "useRal": 88371, + "useRalative": 88372, + "useRalativeImagePath": 88373, + "useState": 55670, + "used": 2591, + "usement": 42665, + "useppe": 92003, + "user": 872, + "userData": 42526, + "userID": 37506, + "userId": 13508, + "userInfo": 36939, + "userManager": 56838, + "userName": 28841, + "userRepository": 85741, + "userService": 72870, + "usercontent": 42417, + "userdata": 23496, + "userid": 20085, + "userinfo": 61510, + "username": 5113, + "users": 4218, + "uses": 4776, + "uset": 18187, + "useum": 11509, + "ush": 1116, + "ushed": 51978, + "usher": 51680, + "ushi": 37074, + "ushima": 75059, + "ushing": 40813, + "ushman": 75481, + "ushort": 40375, + "usi": 52813, + "usic": 11524, + "usiness": 2063, + "using": 970, + "usion": 7560, + "usions": 35707, + "usive": 22010, + "usize": 51878, + "usk": 32779, + "uslim": 43712, + "uso": 46435, + "usp": 29763, + "uspend": 12758, + "uspendLayout": 15475, + "uspended": 66349, + "usr": 7063, + "usra": 89444, + "uss": 1854, + "ussed": 58842, + "ussels": 31864, + "ussen": 42961, + "usses": 82872, + "ussia": 28478, + "ussian": 15579, + "ussion": 17405, + "ussions": 38733, + "ussy": 13447, + "ust": 590, + "usta": 81879, + "ustain": 30773, + "ustainability": 73485, + "ustainable": 41241, + "uste": 85660, + "usted": 27145, + "uster": 4993, + "ustering": 36694, + "usterity": 54155, + "usters": 14605, + "ustin": 44235, + "usting": 37487, + "usto": 60462, + "ustom": 1450, + "ustomed": 49686, + "ustomer": 4394, + "ustr": 3414, + "ustral": 5408, + "ustralia": 28039, + "ustralian": 57149, + "ustria": 84812, + "ustrial": 46379, + "ustry": 17674, + "ustum": 75308, + "usty": 46325, + "usu": 63868, + "usual": 80224, + "usually": 42966, + "usuario": 15219, + "usuarios": 49644, + "usz": 69391, + "ut": 332, + "uta": 15999, + "utable": 5922, + "utan": 48524, + "utar": 40197, + "utas": 16726, + "utation": 22705, + "utations": 82141, + "utc": 28355, + "utch": 14061, + "utches": 87349, + "utdown": 18452, + "ute": 1070, + "uted": 2774, + "utedString": 39530, + "utely": 97366, + "uten": 13160, + "utenant": 37257, + "utenberg": 44433, + "uter": 27951, + "uters": 15558, + "utes": 2095, + "uteur": 38995, + "utex": 9371, + "utf": 4762, + "uth": 940, + "uther": 25487, + "utherford": 92994, + "utherland": 91832, + "uthor": 1553, + "uti": 31873, + "util": 1314, + "utilities": 61134, + "utility": 30900, + "utils": 6031, + "utilus": 90628, + "uting": 10607, + "ution": 1488, + "utions": 3977, + "utive": 6704, + "utivo": 87576, + "utm": 40220, + "uto": 1535, + "utoff": 27440, + "utom": 58624, + "utomation": 85263, + "utor": 4831, + "utorial": 13888, + "utorials": 54927, + "utors": 9475, + "utory": 36145, + "utos": 20435, + "utow": 9616, + "utowired": 9647, + "utr": 79281, + "utra": 77881, + "utral": 14253, + "utron": 74148, + "uts": 6128, + "utsch": 23747, + "utsche": 28932, + "utschein": 75799, + "utschen": 60135, + "utt": 20826, + "utta": 75516, + "utter": 6207, + "utters": 73770, + "utterstock": 62596, + "uttgart": 54143, + "utting": 73981, + "uttle": 98243, + "utto": 55472, + "utton": 959, + "utura": 78707, + "uture": 2976, + "utures": 74606, + "uty": 20237, + "utz": 32997, + "utzer": 59522, + "utzt": 89002, + "uu": 30072, + "uuid": 17128, + "uum": 20434, + "uur": 50381, + "uv": 12058, + "uve": 56023, + "uvian": 98612, + "uvo": 93463, + "uvre": 60380, + "uvw": 61810, + "uvwxyz": 64213, + "uw": 42110, + "ux": 2200, + "uxe": 35322, + "uxt": 76400, + "uxtap": 92362, + "uy": 4076, + "uya": 44544, + "uye": 55611, + "uyen": 37755, + "uyo": 86688, + "uyên": 124527, + "uyến": 125283, + "uyết": 126033, + "uyá»ģn": 125593, + "uyá»ĥn": 125550, + "uyá»ĩn": 124853, + "uz": 5197, + "uze": 90211, + "uzione": 54346, + "uzu": 86325, + "uzz": 8889, + "uzzer": 91447, + "uzzi": 78471, + "uzzle": 14945, + "uzzy": 34758, + "uç": 127177, + "ução": 60967, + "ué": 95459, + "uÃŁ": 134662, + "uÃŁer": 132894, + "uÃŃ": 64125, + "uÄŁu": 127718, + "uÅŁ": 95532, + "v": 85, + "va": 6586, + "vable": 23760, + "vably": 88134, + "vac": 53703, + "vae": 70242, + "vais": 98963, + "vak": 53621, + "val": 831, + "vale": 77729, + "valid": 1891, + "validate": 7067, + "validated": 59590, + "validation": 12284, + "validator": 16112, + "validators": 59424, + "valor": 24782, + "vals": 25596, + "valu": 25510, + "valuate": 19642, + "valuation": 23819, + "valuator": 52389, + "value": 957, + "valueOf": 93013, + "values": 3661, + "van": 15649, + "vana": 67459, + "vanced": 88087, + "vang": 27761, + "vangst": 32943, + "vanished": 87757, + "vanized": 67124, + "vant": 76182, + "var": 947, + "vara": 95950, + "varchar": 32922, + "vard": 21601, + "vari": 29165, + "variable": 9813, + "variables": 18616, + "variably": 69991, + "variant": 15969, + "variants": 54611, + "variation": 75785, + "vars": 15380, + "vary": 82976, + "vas": 4281, + "vascular": 32845, + "vasion": 57993, + "vasive": 77034, + "vat": 49713, + "vation": 60328, + "vature": 56705, + "vault": 82983, + "vb": 42405, + "vc": 7362, + "vd": 16598, + "ve": 586, + "veal": 40611, + "veau": 33467, + "vec": 4083, + "vecs": 94654, + "vect": 75008, + "vection": 79198, + "vector": 3215, + "ved": 2066, + "veedor": 51699, + "veget": 82471, + "veh": 15200, + "vehicle": 19764, + "vehicles": 57115, + "veillance": 60300, + "veis": 43699, + "vel": 889, + "veled": 93964, + "vell": 72212, + "velle": 88481, + "velocity": 14640, + "velop": 1813, + "velope": 18853, + "veloper": 78806, + "velopment": 75831, + "velt": 42421, + "velte": 97679, + "vely": 16239, + "vement": 7830, + "vements": 12477, + "ven": 1037, + "vendor": 19213, + "vendors": 63258, + "venes": 90801, + "venge": 52205, + "venida": 87195, + "venience": 39547, + "venient": 47139, + "venile": 39104, + "venir": 42306, + "vens": 21097, + "vent": 684, + "venta": 52919, + "ventario": 76346, + "ventas": 97101, + "vented": 73128, + "venth": 44607, + "venting": 81820, + "vention": 7459, + "ventional": 43222, + "ventions": 30659, + "ventory": 7223, + "vents": 47664, + "venture": 44943, + "ventus": 66460, + "venue": 7140, + "venues": 45369, + "ver": 423, + "verage": 4355, + "verages": 42868, + "veral": 17539, + "verb": 22328, + "verbatim": 84005, + "verbose": 14883, + "verbosity": 65210, + "verbs": 54939, + "vere": 19289, + "vergence": 80748, + "verification": 50632, + "verified": 21037, + "verify": 12446, + "verige": 74195, + "verity": 26472, + "vern": 75752, + "vero": 88810, + "vers": 3004, + "versation": 22323, + "versations": 72995, + "verse": 4450, + "versed": 43776, + "versible": 64298, + "version": 4366, + "versions": 28290, + "verständ": 142674, + "verständlich": 142675, + "vert": 1621, + "verte": 64832, + "verted": 21630, + "verter": 8721, + "vertex": 12085, + "vertical": 15292, + "vertices": 13030, + "vertime": 80238, + "verting": 49417, + "vertis": 10218, + "vertise": 67787, + "vertisement": 44424, + "vertiser": 84815, + "vertising": 29924, + "verts": 28163, + "verture": 64830, + "verty": 16684, + "very": 1204, + "ves": 2342, + "vest": 7019, + "vester": 82738, + "vestment": 93439, + "vet": 17115, + "vetica": 31212, + "vette": 57230, + "vey": 5617, + "veyor": 68869, + "veys": 49269, + "vez": 19069, + "vf": 45241, + "vfs": 82586, + "vg": 7239, + "vh": 29919, + "vi": 9971, + "via": 20105, + "viar": 68330, + "vic": 44675, + "vice": 1768, + "victim": 90939, + "vid": 1301, + "vida": 75500, + "vide": 19448, + "vided": 42957, + "vidence": 27480, + "video": 9986, + "videos": 34023, + "vider": 3734, + "viders": 17761, + "vides": 67154, + "vidia": 40098, + "vido": 76543, + "vie": 63419, + "vier": 46716, + "view": 1050, + "viewController": 95412, + "viewModel": 73610, + "viewer": 47622, + "viewport": 18272, + "views": 5072, + "vig": 91247, + "vik": 43081, + "vil": 45252, + "villa": 63338, + "ville": 7919, + "vim": 41194, + "vin": 9603, + "vinc": 34133, + "vince": 14748, + "vinces": 78565, + "vincia": 57164, + "vincial": 64327, + "vine": 61201, + "vinfos": 80661, + "ving": 4405, + "vio": 72722, + "viol": 55151, + "violent": 89062, + "vious": 2469, + "viously": 18281, + "vip": 64197, + "vir": 45857, + "viron": 2772, + "vironment": 3105, + "vironments": 17866, + "virt": 47381, + "virtual": 25668, + "vis": 2682, + "visa": 93533, + "vised": 25379, + "visibility": 28176, + "visible": 12601, + "vision": 13013, + "visions": 40015, + "visit": 27460, + "visited": 28868, + "visitor": 39985, + "visión": 133337, + "viso": 72449, + "visor": 38012, + "visual": 29218, + "visualization": 88981, + "visão": 132321, + "vit": 54811, + "viz": 69840, + "vj": 72773, + "vk": 48363, + "vl": 14536, + "vla": 79563, + "vlan": 89622, + "vlc": 98404, + "vm": 7338, + "vn": 28138, + "vo": 3334, + "vocab": 86847, + "vod": 91010, + "voice": 9862, + "voices": 40683, + "void": 1004, + "voie": 87035, + "voir": 34473, + "voke": 7621, + "vol": 12536, + "volatile": 27307, + "volent": 71037, + "volt": 96271, + "volución": 140012, + "volume": 25060, + "volution": 64172, + "volução": 143170, + "von": 58170, + "voor": 86368, + "vor": 36115, + "vore": 90502, + "vos": 47619, + "vote": 29358, + "votes": 37835, + "voucher": 79013, + "vous": 57445, + "vox": 61397, + "voy": 79797, + "vp": 15260, + "vpn": 59589, + "vr": 18920, + "vre": 48506, + "vrier": 90037, + "vrir": 62246, + "vro": 36662, + "vrolet": 37320, + "vron": 29479, + "vs": 11562, + "vsp": 33850, + "vt": 9708, + "vtColor": 48955, + "vtk": 60087, + "vu": 45690, + "vue": 21499, + "vv": 33075, + "vvm": 66682, + "vw": 34431, + "vx": 47280, + "vy": 13714, + "vä": 72050, + "vé": 51174, + "vê": 131514, + "vÃŃ": 129587, + "w": 86, + "wa": 9991, + "wagon": 93119, + "wahl": 71965, + "wait": 11489, + "waitFor": 86205, + "waiting": 49534, + "wake": 89919, + "wal": 26397, + "wald": 52906, + "walk": 19059, + "walker": 44252, + "walking": 80649, + "wall": 16431, + "wallet": 35735, + "walls": 86296, + "wan": 16556, + "wand": 79810, + "wang": 70552, + "want": 52657, + "wanted": 86592, + "wap": 21726, + "war": 11455, + "ward": 1606, + "wards": 4014, + "ware": 1664, + "warehouse": 45546, + "wares": 37903, + "warf": 32634, + "wargs": 6674, + "warm": 82597, + "warn": 33481, + "warning": 18928, + "warnings": 60447, + "wart": 35608, + "warts": 63056, + "warz": 44704, + "was": 16123, + "wash": 69445, + "washer": 47297, + "washing": 97091, + "wat": 58047, + "watch": 14321, + "water": 12987, + "waters": 98659, + "waukee": 32479, + "wav": 52046, + "wave": 30398, + "waves": 57265, + "way": 3117, + "wayne": 89562, + "ways": 2284, + "waÄĩ": 127216, + "waż": 126149, + "ważnie": 142049, + "waÅĤ": 126575, + "wb": 20211, + "wc": 24028, + "wchar": 81604, + "wcs": 34943, + "wcsstore": 39170, + "wd": 6377, + "wdx": 87944, + "we": 896, + "weak": 13150, + "weakSelf": 57832, + "wealth": 30533, + "weapon": 22873, + "weapons": 84765, + "wear": 22744, + "weather": 15206, + "weathermap": 90862, + "web": 2911, + "webElement": 23874, + "webElementProperties": 32686, + "webElementX": 45972, + "webElementXpaths": 45973, + "webView": 70687, + "webdriver": 95477, + "webkit": 25400, + "webpack": 11380, + "webs": 88895, + "website": 19485, + "websocket": 83208, + "webtoken": 64836, + "wechat": 78495, + "wed": 67367, + "weed": 85327, + "week": 10264, + "weekday": 77989, + "weekly": 71767, + "weeks": 80989, + "ween": 1780, + "weeney": 90269, + "weep": 48542, + "weet": 4488, + "weetalert": 66178, + "weeted": 59608, + "weets": 25415, + "weg": 28129, + "wegian": 36565, + "wei": 27105, + "weigh": 56640, + "weight": 4765, + "weighted": 79796, + "weights": 13327, + "weis": 73869, + "weise": 41210, + "weit": 73411, + "wel": 89463, + "welcome": 34084, + "well": 9157, + "wen": 16948, + "wend": 40231, + "wendung": 66671, + "went": 41600, + "wer": 6566, + "were": 51798, + "werk": 58864, + "werp": 94319, + "wers": 23660, + "wert": 54594, + "west": 11039, + "western": 30282, + "wf": 43083, + "wg": 45540, + "wget": 57357, + "wh": 1312, + "what": 12555, + "whatever": 68286, + "wheel": 38590, + "whel": 15631, + "whelming": 93079, + "when": 9309, + "where": 2870, + "whereIn": 85493, + "whether": 48764, + "which": 8206, + "while": 3472, + "white": 5782, + "who": 14623, + "whole": 66633, + "whose": 85143, + "why": 34634, + "wi": 16632, + "wich": 16889, + "wick": 21188, + "wicklung": 69936, + "wid": 43615, + "wide": 9150, + "widget": 9797, + "widgets": 29819, + "width": 3098, + "wie": 45702, + "wife": 47741, + "wifi": 67281, + "wig": 36922, + "wij": 86083, + "wik": 48437, + "wiki": 29707, + "wil": 64636, + "wild": 67874, + "will": 14387, + "willReturn": 44853, + "win": 7526, + "wind": 19154, + "window": 5507, + "windows": 27077, + "wine": 71437, + "wing": 23593, + "wingConstants": 65657, + "winner": 52145, + "wins": 86052, + "winter": 82928, + "wipe": 70109, + "wire": 35531, + "wis": 48043, + "wise": 4482, + "wish": 81099, + "wishlist": 85789, + "wit": 88519, + "witch": 5539, + "with": 4197, + "withErrors": 98211, + "withdraw": 62446, + "within": 55850, + "without": 28996, + "withstanding": 48944, + "witter": 3801, + "wives": 58388, + "wizard": 58593, + "wiÄħ": 59870, + "wiÄħz": 76169, + "wiÄħzan": 132205, + "wiÄĻ": 63998, + "wiÄĻks": 126565, + "wiÄĻksz": 129373, + "wiÄĻt": 127968, + "wiÅĽcie": 128029, + "wj": 66954, + "wjgl": 83680, + "wk": 50522, + "wl": 26417, + "wm": 19557, + "wn": 1522, + "wnd": 23152, + "wner": 4721, + "wo": 1126, + "woff": 68585, + "woke": 95328, + "wolf": 76230, + "woman": 22028, + "women": 64662, + "won": 54667, + "woo": 48774, + "woocommerce": 26470, + "wood": 6660, + "woods": 81029, + "woord": 72506, + "wor": 49710, + "word": 1158, + "wordpress": 58215, + "words": 5761, + "work": 1778, + "worked": 70114, + "worker": 21462, + "workers": 54958, + "workflow": 56249, + "working": 21152, + "works": 18525, + "worksheet": 78937, + "workspace": 42909, + "world": 14615, + "worm": 55641, + "wort": 31151, + "worth": 19501, + "worthy": 42529, + "would": 40350, + "wow": 57454, + "wp": 8421, + "wpdb": 26780, + "wr": 18718, + "wrap": 10097, + "wrapped": 75704, + "wrapper": 21364, + "wright": 52752, + "write": 4934, + "writeln": 93456, + "writer": 18189, + "writers": 61612, + "writes": 59079, + "writing": 25304, + "written": 25569, + "wrong": 34870, + "wró": 127772, + "wróci": 142215, + "wróciÄĩ": 142216, + "ws": 8915, + "wstring": 43890, + "wt": 9306, + "wu": 65465, + "ww": 1250, + "www": 2136, + "wx": 20984, + "wy": 22545, + "wyn": 53656, + "wyż": 126991, + "wä": 129059, + "wód": 127687, + "wództ": 143484, + "wództw": 143485, + "wój": 132444, + "wüns": 141805, + "wünsche": 141806, + "wÄĻ": 130511, + "wÅĤ": 132258, + "wÅĤa": 135251, + "wÅĤasn": 132259, + "wÅĤaÅĽci": 135252, + "wÅĤaÅĽciw": 135253, + "x": 87, + "xA": 14673, + "xAA": 60051, + "xAB": 78146, + "xAC": 79429, + "xAD": 79447, + "xAE": 89728, + "xAF": 83721, + "xB": 14377, + "xBA": 88483, + "xBB": 80197, + "xBC": 80979, + "xBD": 81803, + "xBE": 85449, + "xBF": 71181, + "xC": 12125, + "xCA": 86443, + "xCB": 87480, + "xCC": 72255, + "xCD": 82661, + "xCE": 81695, + "xCF": 77996, + "xD": 15764, + "xDA": 92570, + "xDB": 82344, + "xDC": 83989, + "xDD": 82331, + "xDE": 78979, + "xDF": 73334, + "xE": 12606, + "xEA": 89751, + "xEB": 85211, + "xEC": 86599, + "xED": 83223, + "xEE": 77124, + "xEF": 65769, + "xF": 9770, + "xFA": 72068, + "xFB": 69624, + "xFC": 53473, + "xFD": 63012, + "xFE": 41381, + "xFF": 9264, + "xFFF": 69420, + "xFFFF": 14384, + "xFFFFFF": 72402, + "xFFFFFFFF": 22620, + "xa": 9591, + "xaa": 43300, + "xab": 52616, + "xac": 52642, + "xad": 55254, + "xae": 57052, + "xaf": 54795, + "xampp": 51792, + "xb": 7929, + "xba": 55368, + "xbb": 48939, + "xbc": 43416, + "xbd": 33564, + "xbe": 42459, + "xbf": 47659, + "xc": 8148, + "xca": 56071, + "xcb": 49391, + "xcc": 47380, + "xcd": 53864, + "xce": 55963, + "xcf": 55649, + "xd": 9703, + "xda": 56524, + "xdb": 54555, + "xdc": 48361, + "xdd": 53432, + "xde": 56185, + "xdf": 53833, + "xe": 8371, + "xea": 56570, + "xeb": 54468, + "xec": 52350, + "xed": 51319, + "xee": 50138, + "xef": 47510, + "xes": 89674, + "xf": 5848, + "xfa": 48890, + "xfb": 46071, + "xfc": 40428, + "xfd": 41483, + "xfe": 31469, + "xff": 9020, + "xfff": 87812, + "xffff": 20518, + "xffffff": 68247, + "xffffffff": 41798, + "xford": 22566, + "xhr": 58983, + "xi": 11665, + "xia": 67194, + "xic": 14331, + "xico": 53693, + "xies": 80934, + "xiety": 16308, + "xima": 56492, + "ximity": 91728, + "ximo": 38884, + "xin": 58766, + "xing": 89551, + "xious": 29251, + "xis": 7191, + "xit": 20241, + "xito": 74389, + "xiv": 88733, + "xl": 24310, + "xlabel": 49715, + "xlim": 97414, + "xls": 59520, + "xlsx": 65245, + "xm": 55771, + "xmax": 94032, + "xmin": 86949, + "xml": 6455, + "xmlns": 74155, + "xmm": 75088, + "xn": 42967, + "xo": 40822, + "xon": 61485, + "xoops": 83828, + "xor": 71311, + "xp": 35725, + "xpath": 50151, + "xr": 50997, + "xs": 18561, + "xsd": 82370, + "xt": 2252, + "xx": 4146, + "xxx": 24048, + "xxxx": 18516, + "xxxxxxxx": 44102, + "xy": 4130, + "xygen": 18936, + "xyz": 28854, + "xz": 92955, + "y": 88, + "ya": 7755, + "yah": 94680, + "yahoo": 45792, + "yal": 16453, + "yalty": 24771, + "yaml": 41466, + "yan": 8339, + "yang": 40245, + "yar": 52757, + "yard": 17428, + "yards": 56428, + "yas": 81453, + "yat": 98615, + "yaw": 98586, + "yb": 84307, + "ybrid": 93390, + "yc": 3337, + "ycastle": 93476, + "ych": 5641, + "ycin": 64056, + "ycl": 6179, + "ycle": 5449, + "ycled": 36576, + "ycler": 10760, + "yclerView": 11897, + "yclerview": 33174, + "ycles": 15805, + "yclic": 65304, + "ycling": 16021, + "yclopedia": 45126, + "ycop": 66766, + "ycopg": 71725, + "ycz": 38134, + "yd": 40121, + "ydro": 67007, + "ydı": 132794, + "ye": 9011, + "yeah": 75415, + "year": 3157, + "years": 41720, + "yect": 23516, + "yecto": 29746, + "yectos": 62148, + "yellow": 27869, + "yen": 76151, + "yer": 26161, + "yers": 48472, + "yes": 9693, + "yet": 46392, + "yg": 48765, + "yh": 73801, + "yi": 38072, + "yield": 29696, + "yii": 37164, + "ying": 7024, + "yk": 72540, + "yl": 3923, + "yla": 60722, + "ylabel": 49761, + "ylan": 36478, + "ylation": 78833, + "yle": 967, + "yled": 7601, + "yleft": 62648, + "ylene": 63551, + "yles": 2491, + "ylie": 66477, + "ylim": 47798, + "ylinder": 22868, + "yling": 98607, + "ylko": 49576, + "yll": 24705, + "ylland": 51606, + "ylon": 26004, + "ylum": 27333, + "ylv": 14753, + "ylvania": 17952, + "ym": 1600, + "ymax": 56288, + "ymb": 3356, + "ymbol": 4001, + "ymbols": 11786, + "ymce": 68979, + "yme": 30118, + "ymes": 78709, + "ymi": 76975, + "ymm": 64320, + "ymmetric": 29459, + "ymology": 98074, + "ymoon": 68860, + "ymous": 89392, + "ymph": 31700, + "yms": 75025, + "yn": 1872, + "yna": 52344, + "ynam": 11440, + "ynamic": 5040, + "ynamics": 50148, + "ynamo": 85608, + "ynamodb": 83348, + "ync": 1721, + "ynch": 68481, + "ynchron": 14113, + "ynchronization": 67929, + "ynchronize": 52787, + "ynchronized": 42164, + "ynchronous": 27834, + "ynchronously": 54194, + "yne": 81708, + "ynec": 73930, + "ynes": 83167, + "ynet": 97988, + "ynn": 58260, + "ynom": 61120, + "ynomial": 25358, + "ynomials": 76961, + "ynos": 80393, + "yntax": 13662, + "yntaxException": 76493, + "ynth": 45809, + "ynthesis": 73667, + "ynthia": 63539, + "yny": 56404, + "ynı": 126542, + "yo": 16032, + "yon": 24990, + "yor": 47220, + "you": 9330, + "young": 81179, + "your": 21732, + "youtu": 54860, + "youtube": 43978, + "yp": 1082, + "ypad": 54295, + "ypass": 49911, + "ype": 499, + "yped": 32501, + "yper": 39252, + "ypes": 1804, + "ypi": 80977, + "ypical": 88115, + "ypo": 59300, + "ypress": 47601, + "yps": 75199, + "ypse": 47478, + "ypsum": 67626, + "ypsy": 88267, + "ypy": 54234, + "yr": 10920, + "yre": 71440, + "yro": 66810, + "yrı": 126005, + "yrıca": 127368, + "ys": 1047, + "yscale": 82909, + "yses": 72179, + "ysi": 74284, + "ysical": 18969, + "ysics": 17243, + "ysis": 4448, + "ysize": 69359, + "ysql": 14869, + "ysqli": 34150, + "yssey": 53247, + "yst": 597, + "ystack": 93176, + "ystal": 13946, + "ystals": 91375, + "ystate": 42904, + "ystatechange": 59573, + "ystem": 612, + "ystems": 94790, + "yster": 21848, + "ysterious": 96399, + "ysters": 99063, + "ystery": 95257, + "ystick": 30413, + "ystone": 41266, + "ystore": 63373, + "ysts": 98235, + "ysz": 72345, + "yt": 16415, + "yte": 78361, + "yth": 76896, + "ythe": 95860, + "ython": 26895, + "ytic": 69404, + "ytt": 90138, + "ytut": 88486, + "yu": 40001, + "yum": 89652, + "yun": 42420, + "yw": 84414, + "ywaÄĩ": 135776, + "yx": 51826, + "yy": 4807, + "yyval": 52646, + "yyvsp": 38631, + "yyy": 39319, + "yyyy": 14767, + "yyyyMMdd": 78231, + "yz": 18348, + "yü": 127279, + "yük": 88555, + "yı": 128624, + "yız": 135471, + "yÄĩ": 129460, + "yá»ĩn": 132949, + "z": 89, + "zM": 65215, + "za": 4360, + "zac": 95626, + "zag": 92674, + "zahl": 46728, + "zan": 49126, + "zano": 98998, + "zap": 92371, + "zar": 55900, + "zas": 50355, + "zb": 25329, + "zbek": 75013, + "zbollah": 58504, + "zburg": 78202, + "zc": 34968, + "zcze": 74589, + "zd": 48655, + "ze": 2986, + "zech": 25751, + "zed": 15905, + "zee": 97966, + "zego": 89188, + "zeich": 50591, + "zeichnet": 85740, + "zeigen": 44919, + "zeigt": 96695, + "zeit": 29413, + "zeitig": 97973, + "zej": 45794, + "zek": 63023, + "zel": 27393, + "zelf": 86612, + "zell": 97229, + "zellik": 125961, + "zem": 94133, + "zen": 5679, + "zend": 85348, + "zenia": 35088, + "zenie": 38418, + "zens": 18549, + "zent": 37137, + "zept": 54749, + "zer": 7070, + "zerbai": 56760, + "zero": 14154, + "zeros": 35716, + "zers": 59619, + "zes": 31793, + "zet": 60728, + "zeug": 37400, + "zew": 95204, + "zeÅĦ": 64887, + "zeÅĽ": 127578, + "zf": 69394, + "zh": 23815, + "zheimer": 41260, + "zhou": 51765, + "zi": 8345, + "zia": 67051, + "zial": 59380, + "ziaÅĤ": 69117, + "zie": 13536, + "ziehung": 74991, + "ziej": 77412, + "ziel": 64177, + "zier": 37968, + "zig": 35363, + "zik": 75474, + "zilla": 19330, + "zimmer": 64474, + "zin": 57097, + "zing": 20371, + "zion": 89975, + "zione": 29667, + "zioni": 69182, + "zip": 9964, + "zipcode": 59629, + "zk": 40696, + "zl": 67957, + "zm": 31489, + "zman": 80916, + "zn": 20308, + "zo": 12738, + "zoek": 81613, + "zon": 51619, + "zone": 8684, + "zones": 67534, + "zoom": 28716, + "zos": 53701, + "zp": 93769, + "zs": 40950, + "zsche": 84728, + "zt": 11687, + "zte": 89282, + "zu": 25655, + "zug": 72101, + "zung": 38132, + "zure": 50203, + "zw": 30781, + "zx": 66700, + "zy": 4246, + "zych": 57531, + "zym": 20773, + "zyst": 53981, + "zyÄĩ": 69600, + "zz": 10400, + "zza": 46814, + "zzarella": 95882, + "zzle": 48961, + "zzo": 45648, + "zá": 88543, + "zähl": 134128, + "zÄħ": 27282, + "zÄħd": 76681, + "zÄĻ": 59605, + "zÅĤ": 131511, + "zÅij": 85727, + "{": 90, + "{!!": 54466, + "{\"": 4913, + "{$": 7003, + "{%": 66365, + "{'": 13608, + "{(": 96065, + "{*": 97806, + "{-": 19999, + "{-#": 34376, + "{/": 67678, + "{/*": 94087, + "{//": 55897, + "{:": 44986, + "{@": 59221, + "{EIF": 30727, + "{Name": 63121, + "{T": 76025, + "{\\": 35702, + "{\\\"": 64238, + "{_": 87657, + "{i": 24204, + "{id": 61761, + "{j": 71043, + "{k": 69094, + "{l": 94617, + "{lng": 88754, + "{n": 91362, + "{name": 47006, + "{o": 32531, + "{return": 54678, + "{s": 84386, + "{text": 83021, + "{x": 45340, + "{{": 2979, + "{{$": 29283, + "{{--": 38998, + "{{{": 90691, + "{|": 46115, + "{}": 6257, + "{}\",": 42351, + "{}\".": 42025, + "{}'.": 38238, + "{})": 28875, + "{})Ċ": 37790, + "{},": 22655, + "{},Ċ": 38837, + "{}.": 46391, + "{};Ċ": 45982, + "{}\\": 87963, + "{}]": 78134, + "{}_": 65797, + "{}{Ċ": 67066, + "{}Ċ": 16094, + "{}ĊĊ": 31483, + "{Ċ": 515, + "{ĊĊ": 4257, + "{ĊĊĊ": 53632, + "{čĊ": 1666, + "{čĊčĊ": 25289, + "|": 91, + "|\"": 47576, + "|#": 79450, + "|$": 80050, + "|%": 80166, + "|(": 60016, + "|()Ċ": 66627, + "|(Ċ": 37663, + "|)Ċ": 37275, + "|,Ċ": 50188, + "|-": 50854, + "|--": 80495, + "|--------------------------------------------------------------------------Ċ": 42080, + "|.": 84626, + "|.Ċ": 26280, + "|;Ċ": 34026, + "|=": 87803, + "|=Ċ": 70634, + "|M": 86878, + "|R": 86698, + "|RF": 98578, + "|[": 74723, + "|\\": 22957, + "|^": 96750, + "|`Ċ": 61081, + "|array": 80862, + "|h": 91584, + "|i": 96136, + "|int": 92004, + "|m": 90563, + "|max": 43755, + "|min": 59583, + "|null": 22643, + "|r": 62640, + "|required": 68588, + "|string": 36050, + "|unique": 92983, + "|wx": 83378, + "|x": 62827, + "|{Ċ": 70285, + "||": 8484, + "||(": 82432, + "||||": 76934, + "||Ċ": 90675, + "|}Ċ": 70269, + "|Ċ": 7360, + "|ĊĊ": 43738, + "|čĊ": 84539, + "}": 92, + "}\"": 9863, + "}\")": 55961, + "}\");Ċ": 20305, + "}\");ĊĊ": 64139, + "}\");čĊ": 95140, + "}\")]Ċ": 44665, + "}\")Ċ": 14451, + "}\")ĊĊ": 55266, + "}\")čĊ": 81608, + "}\",": 9545, + "}\",Ċ": 24375, + "}\".": 21901, + "}\";Ċ": 26259, + "}\";ĊĊ": 71612, + "}\">Ċ": 94976, + "}\"Ċ": 11195, + "}\"ĊĊ": 42273, + "}$": 31716, + "}$/": 69153, + "}${": 31192, + "}%": 43715, + "}&": 24778, + "}'": 11661, + "}'\",": 93708, + "}')": 67183, + "}');Ċ": 86321, + "}')Ċ": 32868, + "}')ĊĊ": 74384, + "}',": 16843, + "}','": 49483, + "}',Ċ": 36502, + "}'.": 23495, + "}';Ċ": 83485, + "}'Ċ": 43341, + "}(": 25547, + "}());Ċ": 84693, + "}()Ċ": 69826, + "}()ĊĊ": 66816, + "})": 5410, + "})\",": 78608, + "})\"Ċ": 92349, + "})(": 31888, + "})();": 81174, + "})();Ċ": 52411, + "})();ĊĊ": 94345, + "}))": 65305, + "}));Ċ": 33626, + "}));ĊĊ": 44316, + "}))Ċ": 44194, + "}))ĊĊ": 93596, + "}),": 38842, + "}),Ċ": 30793, + "}).": 16630, + "}):": 90688, + "});": 14088, + "});Ċ": 2960, + "});ĊĊ": 4546, + "});ĊĊĊ": 28348, + "});ĊĊĊĊ": 78137, + "});čĊ": 16240, + "});čĊčĊ": 34083, + "})}Ċ": 92753, + "})Ċ": 3518, + "})ĊĊ": 8824, + "})ĊĊĊ": 43060, + "})čĊ": 35099, + "})čĊčĊ": 70641, + "}*": 92814, + "}*/Ċ": 45434, + "}*/ĊĊ": 67977, + "},": 2137, + "},\"": 51193, + "},${": 91448, + "},'": 76042, + "},{": 16561, + "},{\"": 36828, + "},{Ċ": 65763, + "},Ċ": 1583, + "},ĊĊ": 15766, + "},čĊ": 11557, + "},čĊčĊ": 82979, + "}-": 19732, + "}-${": 54982, + "}->": 57415, + "}->{": 63118, + "}-{": 72013, + "}.": 7810, + "}.${": 77197, + "}.{": 65884, + "}.Ċ": 27275, + "}/": 4472, + "}/#{": 95016, + "}/${": 22805, + "}//": 29450, + "}/>": 87556, + "}/>Ċ": 18769, + "}/{": 22472, + "}:": 16097, + "}:${": 56074, + "}:{": 58146, + "};": 11061, + "};Ċ": 2440, + "};ĊĊ": 2315, + "};ĊĊĊ": 14711, + "};ĊĊĊĊ": 56831, + "};čĊ": 15414, + "};čĊčĊ": 16863, + "};čĊčĊčĊ": 73534, + "}<": 70793, + "}": 13507, + "}><": 71020, + "}>{": 23503, + "}>Ċ": 4877, + "}>čĊ": 59049, + "}?": 52402, + "}@": 80728, + "}[": 43389, + "}\\": 11035, + "}\\\"": 46869, + "}\\.[": 87995, + "}\\\\": 85539, + "}]": 25439, + "}],": 72441, + "}],Ċ": 64054, + "}];Ċ": 75053, + "}]Ċ": 57320, + "}_": 19444, + "}_${": 67661, + "}_{": 51535, + "}`": 5541, + "}`).": 59979, + "}`);Ċ": 18811, + "}`);ĊĊ": 71496, + "}`)Ċ": 27085, + "}`,": 28350, + "}`,Ċ": 27249, + "}`;Ċ": 24840, + "}`;ĊĊ": 72764, + "}`}": 53944, + "}`}>Ċ": 56291, + "}`}Ċ": 44099, + "}`Ċ": 31257, + "}catch": 35262, + "}else": 7825, + "}elseif": 64123, + "}px": 47250, + "}s": 60618, + "}while": 84298, + "}{": 15170, + "}{$": 58373, + "}{Ċ": 59403, + "}|": 52398, + "}}": 3417, + "}}\"": 23386, + "}}\">{{$": 90411, + "}}\">Ċ": 42179, + "}}\"Ċ": 95063, + "}});Ċ": 71230, + "}})Ċ": 71362, + "}},": 38154, + "}},Ċ": 22050, + "}};Ċ": 52481, + "}}": 83697, + "}}>Ċ": 42109, + "}}],Ċ": 83280, + "}}{{": 77286, + "}}}": 75542, + "}}},Ċ": 74869, + "}}Ċ": 11248, + "}}ĊĊ": 47449, + "}Ċ": 532, + "}ĊĊ": 630, + "}ĊĊĊ": 3733, + "}ĊĊĊĊ": 13246, + "}ĊĊĊĊĊ": 33862, + "}ĊĊĊĊĊĊ": 57351, + "}čĊ": 1771, + "}čĊčĊ": 2570, + "}čĊčĊčĊ": 17255, + "}čĊčĊčĊčĊ": 49217, + "}ččĊ": 75531, + "~": 93, + "~\":\"": 73470, + "~,": 54497, + "~-": 49994, + "~-~-": 56746, + "~-~-~-~-": 90007, + "~/": 22714, + "~=": 50633, + "~~": 5817, + "~~~~": 10313, + "~~~~~~~~": 17886, + "~~~~~~~~~~~~~~~~": 32543, + "~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~": 65648, + "~Ċ": 88141, + "~ĊĊ": 58629, + "¡": 94, + "¡°": 48800, + "¡´": 90557, + "¢": 95, + "¢åįķ": 43720, + "£": 96, + "£¼": 35146, + "¤": 97, + "¤æĸŃ": 79326, + "¤í": 44680, + "¥": 98, + "¥¿": 58098, + "¦": 99, + "¦¬": 56823, + "¦Ĥ": 24447, + "§": 100, + "§è¡Į": 58522, + "¨": 101, + "¨¡": 44401, + "©": 102, + "©½": 101867, + "©ëĭĪëĭ¤": 48458, + "ª": 103, + "ªĮ": 36677, + "«": 104, + "«ìŀIJ": 92187, + "«ĺ": 40419, + "¬": 105, + "¬´": 90734, + "¬¸": 33543, + "¬ģ": 70731, + "®": 106, + "¯": 107, + "°": 108, + "±": 109, + "±Ð¾ÑĤ": 35720, + "±ħ": 127964, + "²": 110, + "³": 111, + "³»": 36097, + "´": 112, + "´Ģ": 70297, + "´Ī": 63046, + "µ": 113, + "µ¬": 44706, + "µľ": 71481, + "¶": 114, + "¶Ī": 27757, + "¶ļ": 89083, + "·": 115, + "·¨": 71933, + "·¸": 32926, + "·»": 34811, + "·»åĬł": 35125, + "·æĸ°": 91050, + "¸": 116, + "¹": 117, + "¹Ħ": 46282, + "º": 118, + "º«": 83643, + "»": 119, + "»¿": 3219, + "¼": 120, + "½": 121, + "½æķ°": 29941, + "½Ķ": 65291, + "½Ķëĵľ": 77047, + "¾": 122, + "¾ç¤º": 46871, + "¿": 123, + "À": 124, + "Á": 125, + "Â": 126, + "¡": 39832, + "¢": 56394, + "£": 21651, + "£o": 84291, + "¤": 81538, + "Â¥": 81093, + "¦": 64621, + "§": 17851, + "§Ãĥ": 96784, + "¨": 75005, + "©": 20043, + "ª": 63905, + "«": 23703, + "¬": 82093, + "®": 11909, + "®,": 66357, + "¯": 33985, + "¯¯": 38195, + "¯¯¯¯": 61707, + "°": 11616, + "°C": 30937, + "°F": 58472, + "±": 37021, + "²": 29456, + "³": 43201, + "´": 28111, + "´s": 55376, + "´t": 96865, + "µ": 73048, + "¶": 54509, + "·": 13935, + "··": 60076, + "¸": 144404, + "¹": 59497, + "º": 28200, + "»": 12992, + "»,": 60212, + "».": 74508, + "»ĊĊ": 60543, + "¼": 41873, + "½": 26062, + "¾": 67396, + "¿": 30182, + "ÂĢ": 41416, + "ÂĢÂ": 42476, + "ÂĢÂĻ": 74709, + "ÂĴ": 75151, + "Âł": 4102, + "³³": 9238, + "³³³": 45393, + "³³³³": 19070, + "³³³³³³³": 86210, + "³³³³³³³³": 45449, + "ÂłĊ": 50950, + "ÂŃ": 5760, + "ÂŃi": 96182, + "ÂŃing": 45331, + "ÂŃs": 71780, + "ÂŃt": 49043, + "ÂŃtion": 68605, + "Ã": 127, + "á": 1953, + "áb": 58560, + "ábado": 80408, + "ább": 80699, + "ác": 29341, + "ách": 53061, + "ácil": 55604, + "áf": 63401, + "áfico": 94183, + "ág": 27693, + "ágenes": 86008, + "ágina": 39190, + "ái": 124900, + "ák": 71442, + "ál": 19003, + "ález": 96365, + "álido": 58066, + "áln": 90165, + "ált": 97620, + "ám": 35234, + "ámara": 84001, + "án": 11390, + "ánchez": 98534, + "ánd": 47580, + "ández": 77046, + "ándose": 139230, + "áng": 124397, + "ánh": 124415, + "ánt": 85433, + "ány": 75353, + "ánÃŃ": 86762, + "áo": 77689, + "áp": 125478, + "ár": 41254, + "ária": 57942, + "ário": 19774, + "ários": 37085, + "ás": 7061, + "ása": 93099, + "ást": 55403, + "át": 16685, + "ática": 48502, + "áticas": 91255, + "ático": 66549, + "átis": 71606, + "áveis": 83012, + "ável": 43315, + "áz": 59821, + "â": 8835, + "âce": 64996, + "âm": 86728, + "ân": 39392, + "ância": 65596, + "ând": 78066, + "ât": 63172, + "âte": 68528, + "âu": 124692, + "ây": 124317, + "âÌ": 138673, + "ậ": 138674, + "âÌģ": 134932, + "ã": 3202, + "ãeste": 84651, + "ãi": 126175, + "ãn": 128211, + "ãng": 135076, + "ão": 3413, + "ä": 2305, + "äch": 31573, + "äche": 94650, + "ächst": 92975, + "ächt": 135573, + "äd": 30875, + "ädchen": 41014, + "äft": 56539, + "äge": 70668, + "äh": 21519, + "ähl": 90785, + "ählen": 85242, + "ählt": 83049, + "ähr": 36956, + "äl": 29802, + "äll": 32251, + "älle": 128209, + "ällt": 78120, + "ält": 41203, + "äm": 66897, + "ämp": 141943, + "ämpfe": 141944, + "än": 15141, + "änd": 21545, + "ände": 90360, + "änder": 57396, + "ändig": 130337, + "äng": 25423, + "änge": 76141, + "änger": 61673, + "änn": 62968, + "änner": 42524, + "är": 13977, + "äre": 46686, + "ären": 96049, + "äs": 54840, + "äsent": 90930, + "ät": 14033, + "äter": 70551, + "ätig": 137186, + "ätt": 49191, + "ätz": 59638, + "ätze": 89493, + "ätzlich": 86759, + "äu": 129076, + "ää": 42443, + "ään": 59346, + "Ã¤ÃŁ": 68947, + "Ã¤ÃŁig": 128123, + "Ã¥": 3785, + "Ã¥de": 66336, + "Ã¥l": 86504, + "Ã¥n": 38731, + "Ã¥r": 17900, + "æ": 9191, + "æk": 47096, + "ække": 76123, + "ælland": 86133, + "ænd": 59577, + "ær": 33471, + "ære": 91016, + "ç": 3131, + "ça": 17472, + "ças": 45278, + "çe": 124806, + "çek": 124726, + "çi": 126325, + "çiler": 139750, + "ço": 20210, + "çois": 67353, + "çon": 52410, + "ços": 47919, + "çu": 78784, + "çãeste": 139684, + "ção": 5903, + "ções": 15249, + "çı": 126435, + "è": 4458, + "ège": 72799, + "èle": 50353, + "èles": 84698, + "ème": 24267, + "èmes": 65374, + "ène": 46999, + "ère": 12068, + "èrent": 140142, + "ères": 30439, + "ès": 12142, + "ète": 65422, + "ètre": 98646, + "ètres": 74004, + "ève": 78251, + "é": 963, + "éal": 62397, + "éc": 19667, + "échange": 139300, + "écial": 69777, + "école": 134670, + "économ": 139301, + "économie": 139302, + "écoute": 140673, + "écran": 140348, + "éd": 15083, + "éducation": 143829, + "ée": 7888, + "ées": 13700, + "éfono": 72397, + "ég": 18863, + "égor": 93729, + "égorie": 96093, + "éis": 87018, + "ék": 74373, + "él": 18503, + "élection": 139980, + "éli": 72411, + "élé": 28882, + "ém": 16647, + "ément": 41525, + "éments": 97842, + "émon": 24149, + "én": 10394, + "énergie": 136348, + "énom": 96978, + "ény": 72291, + "éné": 31353, + "éo": 88477, + "éparation": 137834, + "épo": 140253, + "époque": 140254, + "équipe": 129862, + "ér": 13763, + "éra": 75643, + "ération": 52201, + "érc": 95014, + "érer": 51324, + "éri": 27055, + "éric": 38865, + "érica": 70181, + "érie": 34185, + "érience": 64579, + "érieur": 56563, + "ério": 127785, + "érique": 77029, + "éro": 55671, + "ért": 86353, + "éré": 67762, + "és": 5397, + "ése": 91152, + "ést": 86092, + "ész": 77879, + "ét": 10302, + "éta": 142993, + "établ": 137913, + "établissement": 137914, + "étais": 142994, + "était": 131327, + "état": 133775, + "ética": 87929, + "ético": 86138, + "étique": 98088, + "étr": 142754, + "étranger": 142755, + "étude": 132573, + "été": 38783, + "év": 127990, + "éviter": 139685, + "évolution": 138216, + "évén": 138535, + "événement": 138536, + "ê": 5498, + "êm": 38830, + "ême": 22239, + "ên": 27021, + "ência": 23696, + "ências": 63809, + "ês": 36830, + "êt": 48430, + "ête": 35993, + "êtes": 89176, + "être": 37716, + "êu": 124634, + "êÌ": 127369, + "ệ": 132840, + "êÌģ": 132515, + "êÌī": 138068, + "ë": 12179, + "ël": 73453, + "ële": 135586, + "ën": 87555, + "ër": 72474, + "ì": 23531, + "ìm": 126303, + "ìn": 126368, + "ình": 58506, + "î": 24137, + "ît": 76322, + "ître": 64914, + "ï": 37572, + "ð": 67700, + "ñ": 5653, + "ña": 18324, + "ñana": 87084, + "ñas": 59925, + "ñe": 136936, + "ño": 14529, + "ños": 18219, + "ò": 22190, + "òa": 125832, + "òn": 125437, + "òng": 63775, + "ó": 1794, + "óa": 58875, + "ób": 51109, + "óc": 125877, + "ód": 20511, + "ódigo": 32743, + "óg": 34397, + "ói": 125444, + "ój": 61324, + "ól": 51937, + "ólica": 140091, + "óln": 87209, + "ólogo": 139428, + "óm": 49980, + "ómo": 71461, + "ón": 3165, + "óng": 124890, + "ónica": 91963, + "ónico": 59795, + "ópez": 85726, + "ór": 59377, + "ória": 51077, + "ório": 89466, + "óry": 127947, + "ós": 28732, + "ósito": 98854, + "óst": 94852, + "ót": 50358, + "ów": 12263, + "ówi": 126750, + "ówki": 134036, + "ówn": 41313, + "óż": 57162, + "óÅĤ": 47392, + "ô": 9574, + "ôi": 124312, + "ôle": 56232, + "ôm": 94365, + "ôme": 98657, + "ôn": 61452, + "ông": 24015, + "ônica": 140717, + "ôt": 34557, + "ôte": 89386, + "ôtel": 127412, + "õ": 11151, + "ões": 12652, + "ö": 2956, + "öff": 51176, + "öffent": 91712, + "öffentlich": 132149, + "ög": 61897, + "öh": 84251, + "ök": 76791, + "öl": 49802, + "ölü": 134981, + "öm": 85584, + "ömür": 143714, + "ön": 23348, + "öne": 133406, + "önü": 139399, + "ör": 9416, + "ören": 88287, + "örper": 75024, + "ört": 51817, + "örü": 134126, + "ös": 32454, + "öst": 97682, + "öt": 44226, + "ött": 86764, + "öyle": 127688, + "öz": 125367, + "Ã¶ÃŁe": 78940, + "Ã¶ÄŁ": 139381, + "Ã¶ÄŁret": 139382, + "Ã¶ÄŁretim": 139383, + "÷": 144422, + "ø": 6151, + "ød": 74720, + "øj": 64794, + "ør": 17107, + "øre": 66027, + "øy": 50248, + "ù": 14926, + "ùng": 88771, + "ú": 6654, + "úa": 89639, + "úb": 26790, + "ública": 68074, + "úc": 124411, + "úde": 67388, + "údo": 86702, + "úmer": 21606, + "úmero": 25052, + "úmeros": 63918, + "ún": 24180, + "úng": 125446, + "ús": 33762, + "úsqueda": 86304, + "út": 124931, + "û": 29772, + "ût": 59668, + "ü": 2391, + "üb": 130977, + "über": 48985, + "üc": 127771, + "ück": 19358, + "ücken": 91639, + "ücü": 135636, + "ügen": 82040, + "üh": 21558, + "ühl": 93110, + "ühr": 50753, + "ührung": 92182, + "ük": 39857, + "ül": 35524, + "üler": 92071, + "ült": 89541, + "ülü": 135740, + "üm": 48521, + "ümüz": 127783, + "ün": 16074, + "ünc": 143388, + "ünchen": 51831, + "üncü": 143389, + "ünd": 34271, + "ünde": 97310, + "ündig": 136869, + "ünk": 96307, + "ünkü": 127529, + "üns": 66107, + "ünst": 59391, + "ünü": 129782, + "ünün": 132508, + "üp": 131481, + "ür": 5186, + "ürger": 94858, + "ürlich": 56530, + "ürn": 89220, + "ürnberg": 93356, + "ürü": 125665, + "üs": 126827, + "üss": 30203, + "üsseldorf": 85035, + "üssen": 43961, + "üst": 138417, + "üstü": 138418, + "üsü": 133094, + "üt": 28658, + "üyor": 126645, + "üyü": 142860, + "üz": 78591, + "üzü": 131887, + "Ã¼ÃŁ": 134215, + "Ã¼ÄŁ": 124958, + "Ã¼ÄŁÃ¼": 124959, + "Ã¼ÅŁ": 124911, + "ý": 19622, + "ých": 97855, + "ýt": 96600, + "þ": 134629, + "ÿ": 145059, + "ÃĢ": 71953, + "Ãģ": 43701, + "ÃĤ": 32795, + "ÃĤN": 141570, + "Ãĥ": 18791, + "ÃĥO": 52128, + "ÃĦ": 86975, + "Ãħ": 144044, + "ÃĨ": 144839, + "Ãĩ": 45913, + "ÃĩÃĥO": 74634, + "ÃĪ": 142505, + "Ãī": 26789, + "Ãītat": 135212, + "ÃĬ": 127229, + "Ãĭ": 144679, + "ÃĮ": 144201, + "Ãį": 52220, + "Ãİ": 71807, + "Ãı": 144865, + "ÃIJ": 95199, + "Ãij": 71354, + "ÃĴ": 144086, + "Ãĵ": 39062, + "ÃĵN": 51769, + "ÃĶ": 131371, + "Ãķ": 144335, + "Ãĸ": 63361, + "ÃĸZ": 141452, + "ÃĹ": 17568, + "ÃĹ": 88720, + "Ġ\".": 5933, + "Ġ\".\"": 22760, + "Ġ\".\");Ċ": 87260, + "Ġ\".\",": 68514, + "Ġ\".\";Ċ": 91223, + "Ġ\".$": 14786, + "Ġ\".$_": 88947, + "Ġ\"..": 32213, + "Ġ\"...": 38713, + "Ġ\"../": 7005, + "Ġ\"../../": 10208, + "Ġ\"../../../": 30630, + "Ġ\"../../../../": 63392, + "Ġ\"./": 5924, + "Ġ\".Ċ": 80410, + "Ġ\"/": 3521, + "Ġ\"/\"": 16891, + "Ġ\"/\");Ċ": 97539, + "Ġ\"/\",": 64657, + "Ġ\"/\";Ċ": 79480, + "Ġ\"/\"Ċ": 80455, + "Ġ\"//": 35574, + "Ġ\":": 13022, + "Ġ\":\"": 23427, + "Ġ\"::": 70154, + "Ġ\";": 18963, + "Ġ\";\"": 66150, + "Ġ\";Ċ": 7620, + "Ġ\";ĊĊ": 41620, + "Ġ\";čĊ": 24881, + "Ġ\"<": 4055, + "Ġ\"": 28957, + "Ġ\">\"": 90013, + "Ġ\"><": 96334, + "Ġ\">Ċ": 23478, + "Ġ\"?": 27244, + "Ġ\"?\"": 82729, + "Ġ\"@": 8428, + "Ġ\"@\"": 96848, + "Ġ\"@/": 73735, + "Ġ\"[": 10545, + "Ġ\"[\"": 50681, + "Ġ\"[%": 94090, + "Ġ\"\\": 2917, + "Ġ\"\\\"": 15898, + "Ġ\"\\\"\"": 67622, + "Ġ\"\\\",": 77885, + "Ġ\"\\\">": 83213, + "Ġ\"\\(": 43146, + "Ġ\"\\<": 23435, + "Ġ\"\\\\": 26468, + "Ġ\"\\\\\"": 68937, + "Ġ\"\\Ċ": 93317, + "Ġ\"]": 21299, + "Ġ\"]\"": 87596, + "Ġ\"]\");Ċ": 92723, + "Ġ\"]\";Ċ": 57130, + "Ġ\"^": 39915, + "Ġ\"_": 9000, + "Ġ\"_\"": 32725, + "Ġ\"__": 13265, + "Ġ\"`": 35973, + "Ġ\"{": 13868, + "Ġ\"{\"": 97049, + "Ġ\"{$": 54977, + "Ġ\"{\\\"": 54734, + "Ġ\"{{": 47219, + "Ġ\"{}": 35503, + "Ġ\"|": 25203, + "Ġ\"|\"": 60420, + "Ġ\"}": 80548, + "Ġ\"}\";Ċ": 98274, + "Ġ\"}\\": 90105, + "Ġ\"}Ċ": 43270, + "Ġ\"~": 51269, + "Ġ\"~/": 48188, + "Ġ\"Ċ": 6228, + "Ġ\"ĊĊ": 22746, + "Ġ\"čĊ": 31487, + "Ġ#": 671, + "Ġ#\"": 55050, + "Ġ##": 7704, + "Ġ###": 16600, + "Ġ####": 26274, + "Ġ#####": 68684, + "Ġ########": 26352, + "Ġ############": 76527, + "Ġ################": 37615, + "Ġ########################": 96515, + "Ġ################################": 54301, + "Ġ################################################": 67161, + "Ġ################################################################": 48497, + "Ġ########################################################################": 93858, + "Ġ############################################################################": 86200, + "Ġ########.": 34173, + "Ġ###Ċ": 65287, + "Ġ##Ċ": 47726, + "Ġ#$": 81685, + "Ġ#%": 50594, + "Ġ#'": 42037, + "Ġ#(": 89166, + "Ġ#+#": 20505, + "Ġ#-": 29907, + "Ġ#-}Ċ": 41196, + "Ġ#-}ĊĊ": 75499, + "Ġ#:": 46198, + "Ġ#=>": 98915, + "Ġ#@": 91027, + "Ġ#[": 11506, + "Ġ#__": 96807, + "Ġ#{": 11273, + "Ġ#{@": 54628, + "Ġ#Ċ": 9541, + "Ġ#ĊĊ": 58606, + "Ġ#čĊ": 55223, + "Ġ$": 400, + "Ġ$\"": 23440, + "Ġ$\"{": 37758, + "Ġ$#": 65915, + "Ġ$$": 26107, + "Ġ$$$": 97700, + "Ġ$(": 4930, + "Ġ$(\"": 24148, + "Ġ$(\"#": 5945, + "Ġ$(\"#\"": 82994, + "Ġ$(\".": 13005, + "Ġ$(\"<": 50134, + "Ġ$($": 93977, + "Ġ$('": 15716, + "Ġ$('#": 5484, + "Ġ$('#'": 81036, + "Ġ$('.": 9607, + "Ġ$('<": 37084, + "Ġ$('[": 65678, + "Ġ$.": 8672, + "Ġ$?": 50061, + "Ġ$\\": 57960, + "Ġ$_": 3320, + "Ġ$_[": 73929, + "Ġ$__": 24972, + "Ġ${": 3570, + "Ġ${(": 62067, + "Ġ${({": 54131, + "Ġ${Ċ": 98524, + "Ġ$Ċ": 24437, + "Ġ$ĊĊ": 84500, + "Ġ$čĊ": 95884, + "Ġ%": 1018, + "Ġ%\"": 68868, + "Ġ%#": 52406, + "Ġ%%": 23277, + "Ġ%%Ċ": 64058, + "Ġ%(": 21227, + "Ġ%+": 68524, + "Ġ%-": 50953, + "Ġ%.": 18524, + "Ġ%=": 49132, + "Ġ%>": 79036, + "Ġ%@": 50969, + "Ġ%@\",": 31452, + "Ġ%[": 64353, + "Ġ%{": 63737, + "Ġ%}Ċ": 67914, + "Ġ%Ċ": 32548, + "Ġ%ĊĊ": 91697, + "Ġ&": 609, + "Ġ&#": 44533, + "Ġ&$": 26950, + "Ġ&&": 1009, + "Ġ&&Ċ": 6561, + "Ġ&&čĊ": 41075, + "Ġ&'": 30136, + "Ġ&(": 22796, + "Ġ&)": 65387, + "Ġ&);Ċ": 79394, + "Ġ&,": 51953, + "Ġ&:": 51245, + "Ġ&=": 14581, + "Ġ&[": 44590, + "Ġ&_": 30795, + "Ġ&___": 25483, + "Ġ&Ċ": 27271, + "Ġ'": 364, + "Ġ'!": 61926, + "Ġ'\"": 7127, + "Ġ'\"'": 52090, + "Ġ'\"';Ċ": 94335, + "Ġ'\"+": 52125, + "Ġ'\",": 90949, + "Ġ'\".": 64050, + "Ġ'\".$": 26954, + "Ġ'\".$_": 88181, + "Ġ'\">": 96929, + "Ġ'\">'": 50446, + "Ġ'#": 5875, + "Ġ'#'": 57250, + "Ġ'#{": 77491, + "Ġ'$": 8248, + "Ġ'${": 35165, + "Ġ'%": 7677, + "Ġ'%\"": 81403, + "Ġ'%$": 81388, + "Ġ'%'": 65976, + "Ġ'&": 14723, + "Ġ'&#": 58539, + "Ġ'&'": 85099, + "Ġ''": 3355, + "Ġ'''": 11456, + "Ġ'''Ċ": 11166, + "Ġ'''ĊĊ": 44732, + "Ġ'''čĊ": 53895, + "Ġ'')": 13453, + "Ġ''),": 87775, + "Ġ''),Ċ": 62631, + "Ġ'').": 40691, + "Ġ'');": 91515, + "Ġ'');Ċ": 25249, + "Ġ'');ĊĊ": 75910, + "Ġ''){Ċ": 53995, + "Ġ'')Ċ": 22312, + "Ġ'')ĊĊ": 98422, + "Ġ'',": 8981, + "Ġ'',Ċ": 6870, + "Ġ'',čĊ": 48644, + "Ġ''.": 42863, + "Ġ'':Ċ": 48115, + "Ġ'';": 36803, + "Ġ'';Ċ": 7547, + "Ġ'';ĊĊ": 25501, + "Ġ'';čĊ": 32386, + "Ġ''}Ċ": 75352, + "Ġ''Ċ": 11776, + "Ġ''ĊĊ": 45320, + "Ġ''čĊ": 65138, + "Ġ'(": 22022, + "Ġ'('": 37880, + "Ġ'((": 87114, + "Ġ')": 16667, + "Ġ')'": 79375, + "Ġ')';Ċ": 85490, + "Ġ'))Ċ": 59509, + "Ġ'),": 89384, + "Ġ').": 49884, + "Ġ');": 83853, + "Ġ');Ċ": 25624, + "Ġ');ĊĊ": 93873, + "Ġ');čĊ": 97309, + "Ġ')[": 63481, + "Ġ')Ċ": 21686, + "Ġ')ĊĊ": 87141, + "Ġ')čĊ": 77009, + "Ġ'*": 14879, + "Ġ'*'": 30701, + "Ġ'*',": 80174, + "Ġ'**": 77165, + "Ġ'*.": 72718, + "Ġ'+": 13902, + "Ġ'+'": 33388, + "Ġ',": 6614, + "Ġ','": 22262, + "Ġ',',": 63026, + "Ġ','.": 97680, + "Ġ',Ċ": 31418, + "Ġ'-": 7788, + "Ġ'-'": 20672, + "Ġ'-')": 69163, + "Ġ'-')Ċ": 90096, + "Ġ'-',": 50349, + "Ġ'-';Ċ": 83852, + "Ġ'--": 17304, + "Ġ'.": 6256, + "Ġ'.$": 15808, + "Ġ'.'": 24361, + "Ġ'.')": 70235, + "Ġ'.',": 45669, + "Ġ'.';Ċ": 94502, + "Ġ'..": 45121, + "Ġ'..',": 88542, + "Ġ'../": 4927, + "Ġ'../../": 6356, + "Ġ'../../../": 18554, + "Ġ'../../../../": 37449, + "Ġ'../../../../../": 94645, + "Ġ'./": 2756, + "Ġ'./../": 55100, + "Ġ'./../../": 68106, + "Ġ'/": 3353, + "Ġ'/'": 14853, + "Ġ'/')": 64433, + "Ġ'/');Ċ": 85342, + "Ġ'/')Ċ": 93075, + "Ġ'/',": 56427, + "Ġ'/',Ċ": 59611, + "Ġ'/';Ċ": 77311, + "Ġ'/'Ċ": 68133, + "Ġ'/../": 56512, + "Ġ'//": 59660, + "Ġ':": 13592, + "Ġ':'": 36304, + "Ġ'::": 97752, + "Ġ';": 27825, + "Ġ';'": 81954, + "Ġ';Ċ": 19901, + "Ġ';ĊĊ": 93244, + "Ġ';čĊ": 72473, + "Ġ'<": 3857, + "Ġ'<%=": 95095, + "Ġ'": 55907, + "Ġ'>'": 77774, + "Ġ'?": 31067, + "Ġ'?'": 73516, + "Ġ'@": 3458, + "Ġ'@/": 30110, + "Ġ'[": 18309, + "Ġ'['": 48290, + "Ġ'\\": 5196, + "Ġ'\\\"": 84849, + "Ġ'\\'": 58355, + "Ġ'\\''": 34109, + "Ġ'\\\\": 28078, + "Ġ'\\\\'": 93103, + "Ġ']": 42877, + "Ġ'^": 49438, + "Ġ'_": 9235, + "Ġ'_'": 35302, + "Ġ'_',": 90924, + "Ġ'__": 12112, + "Ġ'`": 55995, + "Ġ'{": 11573, + "Ġ'{\"": 52098, + "Ġ'{$": 44600, + "Ġ'{@": 54073, + "Ġ'{{": 42542, + "Ġ'{}": 32912, + "Ġ'{}'": 69422, + "Ġ'|": 36357, + "Ġ'|'": 84639, + "Ġ'}": 40074, + "Ġ'}';Ċ": 57956, + "Ġ'}Ċ": 52462, + "Ġ'~": 63470, + "Ġ'~/": 71404, + "Ġ'Ċ": 7885, + "Ġ'ĊĊ": 55144, + "Ġ'čĊ": 26861, + "Ġ(": 320, + "Ġ(!": 1505, + "Ġ(!!": 77980, + "Ġ(!$": 11378, + "Ġ(!(": 13836, + "Ġ(!((": 95278, + "Ġ(![": 65831, + "Ġ(!_": 24076, + "Ġ(\"": 3489, + "Ġ(\"%": 50534, + "Ġ(\"-": 98670, + "Ġ(\"/": 88889, + "Ġ(\"<": 90276, + "Ġ(\"\\": 51658, + "Ġ(#": 29083, + "Ġ($": 1711, + "Ġ($(": 39137, + "Ġ($(\"#": 72946, + "Ġ($('#": 70044, + "Ġ($)": 75346, + "Ġ($.": 84330, + "Ġ($_": 17435, + "Ġ(${": 62789, + "Ġ(%": 13456, + "Ġ(%)": 33955, + "Ġ(&": 15899, + "Ġ('": 4319, + "Ġ('$": 59877, + "Ġ('\\": 88744, + "Ġ((": 1781, + "Ġ((!": 44824, + "Ġ((\"": 89101, + "Ġ(($": 26856, + "Ġ(('": 92833, + "Ġ(((": 11600, + "Ġ((((": 51100, + "Ġ(()": 66647, + "Ġ((*": 36456, + "Ġ((_": 63887, + "Ġ((__": 54323, + "Ġ()": 1719, + "Ġ())": 44481, + "Ġ());Ċ": 49323, + "Ġ());ĊĊ": 89684, + "Ġ())Ċ": 58389, + "Ġ(),": 38104, + "Ġ(),Ċ": 84201, + "Ġ()->": 78838, + "Ġ().": 33711, + "Ġ():": 40548, + "Ġ();": 39147, + "Ġ();Ċ": 9563, + "Ġ();ĊĊ": 25162, + "Ġ();čĊ": 54453, + "Ġ()=>": 71630, + "Ġ()=>{Ċ": 73162, + "Ġ(){": 93083, + "Ġ(){Ċ": 40653, + "Ġ()Ċ": 12668, + "Ġ()ĊĊ": 32805, + "Ġ()čĊ": 69540, + "Ġ(*": 4609, + "Ġ(*(": 38707, + "Ġ(*((": 82379, + "Ġ(*)": 44048, + "Ġ(*)(": 74794, + "Ġ(**": 76496, + "Ġ(*.": 57018, + "Ġ(+": 17973, + "Ġ(++": 46338, + "Ġ(-": 10293, + "Ġ(--": 57619, + "Ġ(.": 29465, + "Ġ(...": 48967, + "Ġ(...)": 65777, + "Ġ(/": 46225, + "Ġ(:": 29254, + "Ġ(::": 68590, + "Ġ(;": 30375, + "Ġ(;;": 52489, + "Ġ(;;)": 75600, + "Ġ(<": 22438, + "Ġ(=": 38738, + "Ġ(>": 76852, + "Ġ(?": 44959, + "Ġ(?)": 76540, + "Ġ(?,": 63985, + "Ġ(@": 8006, + "Ġ([": 11743, + "Ġ([[": 60525, + "Ġ([]": 34923, + "Ġ(\\": 19788, + "Ġ(\\<": 24333, + "Ġ(^": 41491, + "Ġ(^)(": 58265, + "Ġ(_": 5453, + "Ġ(_)": 56144, + "Ġ(_,": 39464, + "Ġ(_.": 64360, + "Ġ(__": 22558, + "Ġ(`": 28654, + "Ġ({": 9469, + "Ġ({Ċ": 13861, + "Ġ(~": 30757, + "Ġ(~(": 71265, + "Ġ(£": 53815, + "Ġ(âĢľ": 26087, + "Ġ(âĪ": 94825, + "Ġ(Ċ": 2399, + "Ġ(ĊĊ": 52054, + "Ġ(čĊ": 25315, + "Ġ)": 873, + "Ġ)\"": 90798, + "Ġ)(": 40612, + "Ġ))": 11985, + "Ġ)))": 98952, + "Ġ)),Ċ": 39290, + "Ġ));": 59792, + "Ġ));Ċ": 16518, + "Ġ));ĊĊ": 33351, + "Ġ));čĊ": 77314, + "Ġ))}Ċ": 31715, + "Ġ))Ċ": 19245, + "Ġ))ĊĊ": 48622, + "Ġ)*": 89401, + "Ġ),": 6882, + "Ġ),Ċ": 2837, + "Ġ),ĊĊ": 36984, + "Ġ),čĊ": 21406, + "Ġ)->": 35538, + "Ġ).": 7457, + "Ġ).Ċ": 95605, + "Ġ).ĊĊ": 49270, + "Ġ):": 16512, + "Ġ):Ċ": 21035, + "Ġ):ĊĊ": 58418, + "Ġ);": 6903, + "Ġ);Ċ": 1439, + "Ġ);ĊĊ": 3475, + "Ġ);ĊĊĊ": 35833, + "Ġ);čĊ": 7144, + "Ġ);čĊčĊ": 18772, + "Ġ)[": 74029, + "Ġ)]Ċ": 85925, + "Ġ){": 40023, + "Ġ){Ċ": 10959, + "Ġ){ĊĊ": 58177, + "Ġ){čĊ": 41191, + "Ġ)}Ċ": 17566, + "Ġ)}ĊĊ": 90222, + "Ġ)Ċ": 1727, + "Ġ)ĊĊ": 5125, + "Ġ)ĊĊĊ": 25571, + "Ġ)ĊĊĊĊĊĊĊĊ": 28038, + "Ġ)čĊ": 8616, + "Ġ)čĊčĊ": 52531, + "Ġ*": 353, + "Ġ*\"": 91204, + "Ġ*&": 77745, + "Ġ*(": 12861, + "Ġ*((": 15852, + "Ġ*(*": 86724, + "Ġ*)": 2586, + "Ġ*)\"": 93652, + "Ġ*)&": 22158, + "Ġ*)(": 23585, + "Ġ*)((": 78774, + "Ġ*))": 97591, + "Ġ*));Ċ": 72615, + "Ġ*);Ċ": 27042, + "Ġ*);ĊĊ": 62610, + "Ġ*)[": 78082, + "Ġ*)__": 70247, + "Ġ*)Ċ": 17400, + "Ġ*)ĊĊ": 46644, + "Ġ**": 3070, + "Ġ**)": 27876, + "Ġ**)&": 73593, + "Ġ***": 17160, + "Ġ***!Ċ": 83555, + "Ġ****": 30704, + "Ġ*****": 41423, + "Ġ********": 34145, + "Ġ****************": 46846, + "Ġ************************": 54176, + "Ġ********************************": 22958, + "Ġ****************************************": 61360, + "Ġ************************************************": 71836, + "Ġ********************************************************": 65195, + "Ġ****************************************************************": 22292, + "Ġ************************************************************************": 10534, + "Ġ**************************************************************************": 17535, + "Ġ****************************************************************************": 28644, + "Ġ********************************************************************************": 76550, + "Ġ******************************************************************************/Ċ": 79371, + "Ġ******************************************************************************/ĊĊ": 93682, + "Ġ******************************************************************************Ċ": 85785, + "Ġ***/Ċ": 81554, + "Ġ***Ċ": 75241, + "Ġ**/Ċ": 19546, + "Ġ**/ĊĊ": 68612, + "Ġ**/čĊ": 88490, + "Ġ**Ċ": 36914, + "Ġ*,": 11777, + "Ġ*,Ċ": 74792, + "Ġ*----------------------------------------------------------------": 77149, + "Ġ*.": 31530, + "Ġ*/": 639, + "Ġ*/)": 89855, + "Ġ*/,": 36572, + "Ġ*/,Ċ": 47101, + "Ġ*/;Ċ": 37994, + "Ġ*/}Ċ": 21654, + "Ġ*/Ċ": 735, + "Ġ*/ĊĊ": 2791, + "Ġ*/ĊĊĊ": 21668, + "Ġ*/ĊĊĊĊ": 78435, + "Ġ*/čĊ": 3456, + "Ġ*/čĊčĊ": 15452, + "Ġ*/čĊčĊčĊ": 87442, + "Ġ*": 22948, + "Ġ*>(": 28565, + "Ġ*@": 45637, + "Ġ*[": 97932, + "Ġ*\\": 87787, + "Ġ*_": 19332, + "Ġ*__": 37062, + "Ġ*}": 85779, + "Ġ*}ĊĊ": 98372, + "Ġ*Ċ": 1213, + "Ġ*ĊĊ": 19799, + "Ġ*ĊĊĊ": 71515, + "Ġ*čĊ": 7613, + "Ġ+": 488, + "Ġ+\"": 28184, + "Ġ+#": 94743, + "Ġ+#+": 13223, + "Ġ+#+#+#+": 33856, + "Ġ+#+#+#+#+#+": 33857, + "Ġ+%": 74524, + "Ġ+'": 49210, + "Ġ+(": 61083, + "Ġ++": 3443, + "Ġ++$": 72306, + "Ġ++)": 55589, + "Ġ++)Ċ": 83532, + "Ġ+++": 59092, + "Ġ++;Ċ": 40840, + "Ġ++Ċ": 97347, + "Ġ+-": 77545, + "Ġ+/-": 51615, + "Ġ+:+": 13109, + "Ġ+=": 1421, + "Ġ+=Ċ": 88176, + "Ġ+Ċ": 3610, + "Ġ+ĊĊ": 59454, + "Ġ+čĊ": 19587, + "Ġ,": 1154, + "Ġ,\"": 33821, + "Ġ,$": 86794, + "Ġ,'": 38339, + "Ġ,(": 80592, + "Ġ,,": 80258, + "Ġ,-": 90504, + "Ġ,[": 80280, + "Ġ,Ċ": 13493, + "Ġ,ĊĊ": 21174, + "Ġ,čĊ": 55819, + "Ġ-": 481, + "Ġ-$": 82358, + "Ġ-(": 28406, + "Ġ-*": 8195, + "Ġ-*-": 12017, + "Ġ-*-Ċ": 18754, + "Ġ-*-ĊĊ": 34499, + "Ġ-*-čĊ": 82294, + "Ġ-,": 85922, + "Ġ--": 1177, + "Ġ---": 12448, + "Ġ----": 11304, + "Ġ-----": 34703, + "Ġ------": 55460, + "Ġ-------": 60158, + "Ġ--------": 56136, + "Ġ---------": 80822, + "Ġ----------": 43789, + "Ġ-----------": 92586, + "Ġ------------": 32558, + "Ġ-------------": 96803, + "Ġ----------------": 14393, + "Ġ--------------------": 90781, + "Ġ--------------------------------": 19728, + "Ġ------------------------------------------------": 18036, + "Ġ------------------------------------------------------------": 86632, + "Ġ----------------------------------------------------------------": 8466, + "Ġ-------------------------------------------------------------------------": 90915, + "Ġ----------------------------------------------------------------------------": 23852, + "Ġ--------------------------------------------------------------------------------": 39373, + "Ġ------------------------------------------------------------------------------------------------": 59809, + "Ġ----------------------------------------------------------------------------------------------------------------": 66207, + "Ġ-----------------------------------------------------------------------------Ċ": 51814, + "Ġ----------------------------------------------------------------------------Ċ": 63538, + "Ġ---------------------------------------------------------------------------Ċ": 78757, + "Ġ--------------------------------------------------------------------------Ċ": 97942, + "Ġ-------------------------------------------------------------------------Ċ": 80550, + "Ġ------------------------------------------------------------------------Ċ": 96581, + "Ġ----------------------------------------------------------------------Ċ": 89694, + "Ġ----------Ċ": 30323, + "Ġ--------Ċ": 89138, + "Ġ-------Ċ": 38607, + "Ġ------>": 97200, + "Ġ-----Ċ": 79949, + "Ġ--->": 69726, + "Ġ---Ċ": 53923, + "Ġ-->": 3844, + "Ġ-->Ċ": 3659, + "Ġ-->ĊĊ": 16117, + "Ġ-->ĊĊĊ": 70652, + "Ġ-->čĊ": 15956, + "Ġ-->čĊčĊ": 71610, + "Ġ--}}Ċ": 40304, + "Ġ--Ċ": 39514, + "Ġ--ĊĊ": 59581, + "Ġ-.": 42745, + "Ġ-/Ċ": 66870, + "Ġ-:": 52173, + "Ġ-=": 5982, + "Ġ->": 1464, + "Ġ->Ċ": 12381, + "Ġ-Ċ": 18221, + "Ġ-ĊĊ": 21974, + "Ġ.": 659, + "Ġ.\"": 41348, + "Ġ.$": 45377, + "Ġ.'": 43584, + "Ġ.'": 6206, + "Ġ/>\";Ċ": 60338, + "Ġ/>'": 71042, + "Ġ/>';Ċ": 56045, + "Ġ/>);Ċ": 63191, + "Ġ/>)Ċ": 98436, + "Ġ/>,": 77551, + "Ġ/>,Ċ": 55983, + "Ġ/>;Ċ": 41289, + "Ġ/><": 23976, + "Ġ/>\\": 76246, + "Ġ/>}": 85043, + "Ġ/>}Ċ": 36781, + "Ġ/>Ċ": 2597, + "Ġ/>ĊĊ": 18544, + "Ġ/>čĊ": 17989, + "Ġ/[": 95329, + "Ġ/\\": 23536, + "Ġ/\\.": 45472, + "Ġ/\\.(": 84863, + "Ġ/^": 34510, + "Ġ/^(": 93296, + "Ġ/^[": 75336, + "Ġ/^\\": 80957, + "Ġ/Ċ": 38981, + "Ġ/ĊĊ": 91545, + "Ġ:": 549, + "Ġ:\"": 34638, + "Ġ:\",": 71863, + "Ġ:\";Ċ": 91898, + "Ġ:'": 53546, + "Ġ:(": 39524, + "Ġ:)": 25956, + "Ġ:)Ċ": 89063, + "Ġ:)ĊĊ": 32623, + "Ġ:+:": 15950, + "Ġ:,": 27782, + "Ġ:-": 20955, + "Ġ:-)": 91841, + "Ġ:-Ċ": 62344, + "Ġ::": 3504, + "Ġ:::": 21222, + "Ġ::::::::": 33258, + "Ġ::=": 64154, + "Ġ::Ċ": 80687, + "Ġ:": 47879, + "Ġ<%": 59230, + "Ġ<%=": 90194, + "Ġ<*": 86202, + "Ġ<*>": 69483, + "Ġ<+": 85703, + "Ġ<-": 9119, + "Ġ<--": 70386, + "Ġ<->": 66832, + "ĠĊ": 23600, + "Ġ<:": 47926, + "Ġ<<": 1115, + "Ġ<<\"": 41554, + "Ġ<<-": 75674, + "Ġ<<<": 43662, + "Ġ<<=": 51178, + "Ġ<<Ċ": 48503, + "Ġ<=": 2651, + "Ġ<=>": 71505, + "Ġ<>": 14392, + "Ġ<>Ċ": 20758, + "Ġ": 94331, + "Ġ==>": 32296, + "Ġ==Ċ": 47761, + "Ġ=>": 589, + "Ġ=>$": 60554, + "Ġ=>'": 49807, + "Ġ=>{Ċ": 30634, + "Ġ=>Ċ": 6274, + "Ġ=>čĊ": 48668, + "Ġ=[": 67610, + "Ġ={": 72142, + "Ġ={Ċ": 83724, + "Ġ=~": 20495, + "Ġ=Ċ": 4035, + "Ġ=ĊĊ": 79483, + "Ġ=čĊ": 32203, + "Ġ>": 861, + "Ġ>\",": 82876, + "Ġ>&": 36498, + "Ġ>(": 36720, + "Ġ>/": 61359, + "Ġ>::": 33813, + "Ġ><": 29348, + "Ġ>=": 2604, + "Ġ>>": 3578, + "Ġ>>=": 32971, + "Ġ>>>": 12109, + "Ġ>>ĊĊ": 56948, + "Ġ>Ċ": 6107, + "Ġ>ĊĊ": 25321, + "Ġ>čĊ": 39206, + "Ġ?": 937, + "Ġ?\");Ċ": 90192, + "Ġ?\",": 42313, + "Ġ?\";Ċ": 64253, + "Ġ?',": 57950, + "Ġ?,": 22473, + "Ġ?.": 79928, + "Ġ?:": 27257, + "Ġ?>": 3370, + "Ġ?>\"": 9543, + "Ġ?>\"/>Ċ": 93744, + "Ġ?>\">": 37751, + "Ġ?>\"><": 48782, + "Ġ?>\">\">\">Ċ": 19306, + "Ġ?>\">čĊ": 70931, + "Ġ?>\"Ċ": 64607, + "Ġ?>&": 66706, + "Ġ?>'": 91566, + "Ġ?>/": 27967, + "Ġ?>:;Ċ": 98187, + "Ġ?><": 43269, + "Ġ?>>": 39246, + "Ġ?>>>Ċ": 54130, + "Ġ?>Ċ": 3870, + "Ġ?>ĊĊ": 12279, + "Ġ?>ĊĊĊ": 70193, + "Ġ?>čĊ": 15861, + "Ġ?>čĊčĊ": 54616, + "Ġ??": 9415, + "Ġ???": 51317, + "Ġ?????": 73478, + "Ġ?Ċ": 17607, + "Ġ?ĊĊ": 23754, + "Ġ@": 569, + "Ġ@\"": 7098, + "Ġ@\"\";Ċ": 70878, + "Ġ@\"\\": 67976, + "Ġ@\"Ċ": 98688, + "Ġ@$": 36017, + "Ġ@$_": 98570, + "Ġ@(": 42181, + "Ġ@@": 22307, + "Ġ@@Ċ": 87120, + "Ġ@[": 95800, + "Ġ@_": 64677, + "Ġ@_;Ċ": 42574, + "Ġ@_;ĊĊ": 61163, + "Ġ@{": 33267, + "Ġ@{$": 78600, + "Ġ@{@\"": 92625, + "Ġ@{Ċ": 56247, + "Ġ@}": 93998, + "ĠA": 362, + "ĠAA": 28234, + "ĠAAA": 47097, + "ĠAABB": 88580, + "ĠAAC": 75348, + "ĠAAP": 74752, + "ĠAB": 14137, + "ĠABC": 19360, + "ĠABI": 63214, + "ĠABOUT": 51812, + "ĠABOVE": 92514, + "ĠABS": 36481, + "ĠAC": 10584, + "ĠACA": 64938, + "ĠACC": 25840, + "ĠACCEPT": 54906, + "ĠACCESS": 50552, + "ĠACCOUNT": 62358, + "ĠACE": 40806, + "ĠACK": 53763, + "ĠACL": 43461, + "ĠACLU": 64868, + "ĠACM": 89215, + "ĠACPI": 70118, + "ĠACS": 78084, + "ĠACT": 21116, + "ĠACTION": 16054, + "ĠACTIONS": 63315, + "ĠACTIVE": 74775, + "ĠAD": 9630, + "ĠADA": 78325, + "ĠADC": 22083, + "ĠADD": 15814, + "ĠADDR": 80514, + "ĠADDRESS": 64428, + "ĠADHD": 51464, + "ĠADM": 85001, + "ĠADMIN": 50235, + "ĠADS": 64235, + "ĠADV": 24540, + "ĠADVISED": 29321, + "ĠAE": 42108, + "ĠAES": 38841, + "ĠAF": 19885, + "ĠAFC": 63536, + "ĠAFF": 68938, + "ĠAFL": 56918, + "ĠAFP": 26648, + "ĠAFTER": 48264, + "ĠAG": 15082, + "ĠAGAIN": 66894, + "ĠAGE": 90588, + "ĠAGRE": 75490, + "ĠAH": 47288, + "ĠAHL": 89918, + "ĠAI": 15235, + "ĠAIDS": 44960, + "ĠAIM": 86768, + "ĠAIR": 45894, + "ĠAIS": 94100, + "ĠAJ": 32950, + "ĠAJAX": 57316, + "ĠAK": 30572, + "ĠAL": 8753, + "ĠALERT": 90548, + "ĠALIGN": 63769, + "ĠALL": 13097, + "ĠALLOW": 72260, + "ĠALOG": 79803, + "ĠALPHA": 82797, + "ĠALS": 38423, + "ĠALSO": 49821, + "ĠALT": 49412, + "ĠALTER": 75397, + "ĠALWAYS": 67414, + "ĠAM": 6769, + "ĠAMA": 51789, + "ĠAMAZ": 76520, + "ĠAMC": 74309, + "ĠAMD": 24311, + "ĠAMP": 62908, + "ĠAMS": 97270, + "ĠAN": 2100, + "ĠANAL": 65757, + "ĠANC": 85014, + "ĠAND": 3567, + "ĠANN": 76406, + "ĠANSI": 58014, + "ĠANSW": 96704, + "ĠANT": 86914, + "ĠANY": 4137, + "ĠAO": 65269, + "ĠAOL": 93151, + "ĠAP": 10106, + "ĠAPA": 83540, + "ĠAPC": 86241, + "ĠAPI": 5333, + "ĠAPIs": 33356, + "ĠAPK": 62486, + "ĠAPP": 17912, + "ĠAPPLE": 90336, + "ĠAPPLICATION": 59237, + "ĠAPPLY": 88986, + "ĠAPPRO": 93654, + "ĠAPR": 52426, + "ĠAPS": 91781, + "ĠAQ": 82455, + "ĠAR": 6261, + "ĠARC": 70344, + "ĠARCH": 55213, + "ĠARE": 15824, + "ĠAREA": 69632, + "ĠARG": 32746, + "ĠARGS": 84149, + "ĠARISING": 17549, + "ĠARM": 30486, + "ĠARP": 91337, + "ĠARR": 82079, + "ĠARRAY": 31112, + "ĠART": 33879, + "ĠARTICLE": 91915, + "ĠAS": 5752, + "ĠASA": 88600, + "ĠASAP": 66490, + "ĠASC": 19796, + "ĠASCII": 39316, + "ĠASD": 77194, + "ĠASE": 98867, + "ĠASF": 46268, + "ĠASIC": 93352, + "ĠASM": 65576, + "ĠASN": 54855, + "ĠASP": 48409, + "ĠASS": 35560, + "ĠASSERT": 12208, + "ĠASSIGN": 95571, + "ĠAST": 22465, + "ĠASTM": 61035, + "ĠASUS": 74670, + "ĠAT": 7369, + "ĠATA": 91622, + "ĠATF": 92851, + "ĠATH": 87089, + "ĠATI": 78776, + "ĠATK": 65335, + "ĠATL": 77523, + "ĠATM": 46035, + "ĠATP": 66556, + "ĠATS": 95375, + "ĠATT": 41285, + "ĠATTACK": 98450, + "ĠATTR": 37918, + "ĠATTRIBUTE": 84811, + "ĠATV": 95336, + "ĠAU": 39056, + "ĠAUD": 57580, + "ĠAUDIO": 52124, + "ĠAUG": 94881, + "ĠAUT": 53641, + "ĠAUTH": 16777, + "ĠAUTHOR": 40809, + "ĠAUTHORS": 26688, + "ĠAUTO": 35340, + "ĠAUX": 95755, + "ĠAV": 12155, + "ĠAVAILABLE": 67719, + "ĠAVC": 69318, + "ĠAVCapture": 91409, + "ĠAVG": 70414, + "ĠAVL": 89667, + "ĠAVR": 81313, + "ĠAW": 33338, + "ĠAWS": 23245, + "ĠAX": 40711, + "ĠAXIS": 87226, + "ĠAZ": 30876, + "ĠAad": 79272, + "ĠAaron": 25673, + "ĠAb": 3680, + "ĠAbb": 25973, + "ĠAbbas": 70145, + "ĠAbbey": 64455, + "ĠAbbott": 42127, + "ĠAbby": 84448, + "ĠAbd": 29638, + "ĠAbdel": 92291, + "ĠAbdul": 51598, + "ĠAbdullah": 63432, + "ĠAbe": 55139, + "ĠAbed": 98264, + "ĠAbel": 86849, + "ĠAber": 36574, + "ĠAberdeen": 79251, + "ĠAbilities": 79377, + "ĠAbility": 35983, + "ĠAble": 73850, + "ĠAboriginal": 55410, + "ĠAbort": 76955, + "ĠAbout": 9975, + "ĠAbove": 43610, + "ĠAbr": 36401, + "ĠAbraham": 36388, + "ĠAbram": 90003, + "ĠAbrams": 74231, + "ĠAbs": 22080, + "ĠAbsolute": 48881, + "ĠAbsolutely": 55547, + "ĠAbstract": 13513, + "ĠAbu": 30129, + "ĠAbuse": 45684, + "ĠAbyss": 64958, + "ĠAc": 6381, + "ĠAcad": 12364, + "ĠAcademic": 41070, + "ĠAcademy": 15815, + "ĠAcc": 11429, + "ĠAcceler": 46905, + "ĠAccent": 81809, + "ĠAccept": 20829, + "ĠAccepted": 63289, + "ĠAccess": 9549, + "ĠAccessToken": 94341, + "ĠAccessed": 97300, + "ĠAccessibility": 80272, + "ĠAccessories": 39719, + "ĠAccident": 75504, + "ĠAccom": 52964, + "ĠAccord": 79557, + "ĠAccording": 10548, + "ĠAccordingly": 62809, + "ĠAccount": 8615, + "ĠAccountability": 83690, + "ĠAccounting": 44244, + "ĠAccounts": 40655, + "ĠAccred": 96692, + "ĠAccum": 79525, + "ĠAccuracy": 50175, + "ĠAce": 37707, + "ĠAcer": 75977, + "ĠAch": 45565, + "ĠAchie": 41124, + "ĠAchievement": 65595, + "ĠAchilles": 87636, + "ĠAcid": 49134, + "ĠAck": 50982, + "ĠAcquisition": 72371, + "ĠAcres": 82459, + "ĠAcrobat": 97547, + "ĠAcross": 57137, + "ĠAct": 3218, + "ĠActing": 72689, + "ĠAction": 5586, + "ĠActionBar": 79093, + "ĠActionController": 73452, + "ĠActionListener": 40219, + "ĠActionResult": 16831, + "ĠActionType": 81762, + "ĠActionTypes": 89206, + "ĠActions": 26722, + "ĠActiv": 14708, + "ĠActivate": 51714, + "ĠActivatedRoute": 31417, + "ĠActivation": 49647, + "ĠActive": 9905, + "ĠActiveForm": 80964, + "ĠActiveRecord": 17780, + "ĠActiveSupport": 73001, + "ĠActivities": 43282, + "ĠActivity": 14981, + "ĠActivityCompat": 92444, + "ĠActor": 24718, + "ĠActors": 81361, + "ĠActress": 78439, + "ĠActs": 48818, + "ĠActual": 33359, + "ĠActually": 33763, + "ĠAd": 2410, + "ĠAda": 51149, + "ĠAdam": 15037, + "ĠAdams": 26234, + "ĠAdapt": 58431, + "ĠAdapter": 28807, + "ĠAdapterView": 70497, + "ĠAdaptive": 86408, + "ĠAdd": 2691, + "ĠAdded": 17684, + "ĠAddiction": 78690, + "ĠAdding": 30370, + "ĠAddison": 98994, + "ĠAddition": 78646, + "ĠAdditional": 23208, + "ĠAdditionally": 22406, + "ĠAddr": 72585, + "ĠAddress": 9177, + "ĠAddresses": 91430, + "ĠAdds": 24475, + "ĠAde": 62040, + "ĠAdelaide": 49201, + "ĠAdemás": 87483, + "ĠAdidas": 69889, + "ĠAdj": 62916, + "ĠAdjust": 27196, + "ĠAdjustable": 66728, + "ĠAdjustment": 91949, + "ĠAdler": 98618, + "ĠAdmin": 7582, + "ĠAdministr": 32851, + "ĠAdministration": 16713, + "ĠAdministrative": 51841, + "ĠAdministrator": 28593, + "ĠAdmir": 49469, + "ĠAdmiral": 57994, + "ĠAdmission": 62346, + "ĠAdobe": 28866, + "ĠAdoles": 62637, + "ĠAdolescent": 97482, + "ĠAdolf": 89485, + "ĠAdopt": 85280, + "ĠAdoption": 92017, + "ĠAdresse": 89470, + "ĠAdri": 61343, + "ĠAdrian": 43592, + "ĠAds": 42183, + "ĠAdult": 22133, + "ĠAdults": 57238, + "ĠAdv": 12325, + "ĠAdvance": 46296, + "ĠAdvanced": 21159, + "ĠAdvances": 90858, + "ĠAdvantage": 70359, + "ĠAdvent": 20540, + "ĠAdventure": 32012, + "ĠAdventures": 50579, + "ĠAdvertisement": 66002, + "ĠAdvertising": 46323, + "ĠAdvice": 54720, + "ĠAdvis": 23924, + "ĠAdvisor": 53332, + "ĠAdvisors": 90685, + "ĠAdvisory": 43913, + "ĠAdvoc": 47259, + "ĠAdvocate": 79007, + "ĠAe": 89795, + "ĠAer": 63050, + "ĠAero": 88423, + "ĠAeros": 63150, + "ĠAerospace": 80865, + "ĠAf": 12907, + "ĠAff": 9748, + "ĠAffairs": 22485, + "ĠAffero": 24304, + "ĠAffiliate": 63116, + "ĠAffordable": 42506, + "ĠAfghan": 16795, + "ĠAfghanistan": 20500, + "ĠAfr": 6432, + "ĠAfrica": 10174, + "ĠAfrican": 11643, + "ĠAfricans": 65638, + "ĠAfro": 93600, + "ĠAfter": 4636, + "ĠAfterwards": 86540, + "ĠAg": 4598, + "ĠAgain": 13759, + "ĠAgainst": 29141, + "ĠAge": 13081, + "ĠAgencies": 90186, + "ĠAgency": 16379, + "ĠAgenda": 57562, + "ĠAgent": 20713, + "ĠAgents": 50254, + "ĠAges": 48993, + "ĠAggregate": 56922, + "ĠAgile": 82184, + "ĠAgility": 94719, + "ĠAging": 76088, + "ĠAgo": 96603, + "ĠAgr": 71708, + "ĠAgree": 83203, + "ĠAgreement": 22499, + "ĠAgricult": 27900, + "ĠAgricultural": 59034, + "ĠAgriculture": 36863, + "ĠAgu": 76308, + "ĠAh": 16366, + "ĠAhead": 76444, + "ĠAhmad": 54317, + "ĠAhmed": 40669, + "ĠAi": 55986, + "ĠAid": 37405, + "ĠAim": 70615, + "ĠAin": 81151, + "ĠAir": 6553, + "ĠAirbnb": 60101, + "ĠAirbus": 80374, + "ĠAircraft": 58845, + "ĠAires": 64617, + "ĠAirlines": 34130, + "ĠAirport": 20691, + "ĠAirways": 69402, + "ĠAj": 40657, + "ĠAjax": 41162, + "ĠAk": 16358, + "ĠAkron": 94793, + "ĠAkt": 49096, + "ĠAl": 1674, + "ĠAla": 88662, + "ĠAlabama": 20623, + "ĠAlam": 64950, + "ĠAlamofire": 94074, + "ĠAlan": 25282, + "ĠAlarm": 46284, + "ĠAlaska": 27267, + "ĠAlb": 31572, + "ĠAlban": 56891, + "ĠAlbania": 94633, + "ĠAlbany": 54954, + "ĠAlbert": 17513, + "ĠAlberta": 32554, + "ĠAlberto": 64681, + "ĠAlbum": 25665, + "ĠAlbums": 86723, + "ĠAlbuquerque": 75812, + "ĠAlcohol": 44296, + "ĠAld": 30347, + "ĠAle": 19081, + "ĠAlec": 85016, + "ĠAlejandro": 84919, + "ĠAleks": 89578, + "ĠAleppo": 57121, + "ĠAlert": 13975, + "ĠAlertDialog": 31176, + "ĠAlerts": 68308, + "ĠAless": 69785, + "ĠAlex": 8515, + "ĠAlexa": 49612, + "ĠAlexand": 37477, + "ĠAlexander": 20042, + "ĠAlexandra": 74016, + "ĠAlexandre": 89054, + "ĠAlexandria": 56133, + "ĠAlexis": 58961, + "ĠAlf": 65768, + "ĠAlfred": 41492, + "ĠAlg": 81199, + "ĠAlgebra": 76443, + "ĠAlger": 58509, + "ĠAlgeria": 80241, + "ĠAlgorithm": 40325, + "ĠAlgorithms": 85759, + "ĠAli": 14583, + "ĠAlias": 58040, + "ĠAlibaba": 54364, + "ĠAlic": 63158, + "ĠAlice": 29405, + "ĠAlicia": 78902, + "ĠAlien": 46747, + "ĠAlign": 49472, + "ĠAlignment": 32265, + "ĠAlison": 66363, + "ĠAlive": 65426, + "ĠAll": 2009, + "ĠAlla": 90633, + "ĠAllah": 27371, + "ĠAllan": 60225, + "ĠAlle": 42161, + "ĠAlleg": 57724, + "ĠAllen": 20060, + "ĠAlley": 85989, + "ĠAlliance": 22751, + "ĠAllied": 60755, + "ĠAllies": 80674, + "ĠAllison": 62408, + "ĠAlloc": 46685, + "ĠAllocate": 44625, + "ĠAllocation": 72667, + "ĠAllocator": 80332, + "ĠAllow": 26530, + "ĠAllowed": 57622, + "ĠAllows": 45771, + "ĠAlloy": 60330, + "ĠAlly": 81470, + "ĠAlma": 98472, + "ĠAlmighty": 91314, + "ĠAlmost": 34303, + "ĠAlo": 94454, + "ĠAlone": 69308, + "ĠAlong": 31844, + "ĠAlonso": 74067, + "ĠAlpha": 24708, + "ĠAlphabet": 62797, + "ĠAlpine": 81713, + "ĠAlps": 92900, + "ĠAlready": 39512, + "ĠAlright": 97593, + "ĠAls": 44462, + "ĠAlso": 7281, + "ĠAlt": 23682, + "ĠAlta": 82060, + "ĠAlter": 42851, + "ĠAltern": 19489, + "ĠAlternate": 74724, + "ĠAlternative": 41109, + "ĠAlternatively": 38478, + "ĠAlthough": 10328, + "ĠAlto": 53270, + "ĠAltın": 143109, + "ĠAluminium": 88482, + "ĠAluminum": 45853, + "ĠAlumni": 75426, + "ĠAlv": 68946, + "ĠAlvarez": 76715, + "ĠAlways": 23240, + "ĠAly": 68310, + "ĠAlzheimer": 43431, + "ĠAlém": 133036, + "ĠAm": 3303, + "ĠAmanda": 41759, + "ĠAmar": 73657, + "ĠAmateur": 37983, + "ĠAmazing": 36533, + "ĠAmazon": 8176, + "ĠAmb": 19833, + "ĠAmbassador": 44572, + "ĠAmber": 46664, + "ĠAmbient": 92179, + "ĠAmelia": 90879, + "ĠAmen": 61396, + "ĠAmend": 76713, + "ĠAmendment": 21711, + "ĠAmendments": 98217, + "ĠAmer": 49773, + "ĠAmeric": 2288, + "ĠAmerica": 5159, + "ĠAmerican": 3693, + "ĠAmericans": 8877, + "ĠAmericas": 51148, + "ĠAmes": 95094, + "ĠAmid": 88689, + "ĠAmir": 89151, + "ĠAmit": 83763, + "ĠAmmo": 79481, + "ĠAmnesty": 77696, + "ĠAmong": 21658, + "ĠAmount": 25783, + "ĠAmp": 53687, + "ĠAmph": 92161, + "ĠAmpl": 71615, + "ĠAmsterdam": 37741, + "ĠAmy": 28693, + "ĠAmérica": 97092, + "ĠAn": 1527, + "ĠAna": 32138, + "ĠAnaheim": 71164, + "ĠAnal": 19452, + "ĠAnalog": 63446, + "ĠAnaly": 37427, + "ĠAnalysis": 18320, + "ĠAnalyst": 40470, + "ĠAnalytics": 32427, + "ĠAnalyzer": 78705, + "ĠAnast": 93325, + "ĠAnat": 95564, + "ĠAnatomy": 79182, + "ĠAnc": 30277, + "ĠAnch": 71484, + "ĠAnchor": 48772, + "ĠAncient": 36950, + "ĠAnd": 1597, + "ĠAndAlso": 69390, + "ĠAnda": 39138, + "ĠAnders": 47593, + "ĠAndersen": 85551, + "ĠAnderson": 20642, + "ĠAndr": 77878, + "ĠAndre": 26427, + "ĠAndrea": 40084, + "ĠAndreas": 50062, + "ĠAndres": 95346, + "ĠAndrew": 13615, + "ĠAndrews": 51851, + "ĠAndroid": 8514, + "ĠAndré": 84800, + "ĠAndy": 24832, + "ĠAng": 7416, + "ĠAnge": 10991, + "ĠAngebot": 81020, + "ĠAngel": 23080, + "ĠAngela": 37143, + "ĠAngeles": 11902, + "ĠAngelo": 89377, + "ĠAngels": 42045, + "ĠAngie": 98929, + "ĠAngle": 36237, + "ĠAnglic": 90534, + "ĠAnglo": 59159, + "ĠAngola": 97692, + "ĠAngry": 77818, + "ĠAngular": 28578, + "ĠAngularFire": 67087, + "ĠAngus": 83955, + "ĠAnim": 45465, + "ĠAnimal": 21292, + "ĠAnimals": 46866, + "ĠAnimated": 46888, + "ĠAnimation": 20752, + "ĠAnimator": 45754, + "ĠAnime": 51115, + "ĠAnita": 75626, + "ĠAnk": 56815, + "ĠAnkara": 70244, + "ĠAnn": 9305, + "ĠAnna": 23223, + "ĠAnne": 27926, + "ĠAnnex": 88620, + "ĠAnnie": 51989, + "ĠAnniversary": 46798, + "ĠAnnotation": 45823, + "ĠAnnotations": 93332, + "ĠAnnouncement": 86830, + "ĠAnnual": 24947, + "ĠAnnunci": 63166, + "ĠAnonymous": 30437, + "ĠAnother": 13293, + "ĠAns": 38453, + "ĠAnsi": 72022, + "ĠAnswer": 21806, + "ĠAnswers": 37243, + "ĠAnt": 6755, + "ĠAntar": 47523, + "ĠAntarctic": 79741, + "ĠAntarctica": 71687, + "ĠAnth": 16580, + "ĠAnthem": 89132, + "ĠAnthony": 20696, + "ĠAnthrop": 69284, + "ĠAnti": 22992, + "ĠAntib": 94851, + "ĠAntique": 70070, + "ĠAntoine": 90154, + "ĠAnton": 17500, + "ĠAntonio": 22437, + "ĠAntony": 96628, + "ĠAntwort": 65754, + "ĠAntworten": 95817, + "ĠAnxiety": 80369, + "ĠAny": 5765, + "ĠAnyObject": 41543, + "ĠAnyone": 32534, + "ĠAnything": 40933, + "ĠAnyway": 41569, + "ĠAo": 49364, + "ĠAp": 5232, + "ĠApache": 8914, + "ĠApart": 34702, + "ĠApartment": 52789, + "ĠApartments": 66551, + "ĠApex": 83355, + "ĠAph": 97308, + "ĠApi": 13443, + "ĠApiController": 87826, + "ĠApiException": 82793, + "ĠApiResponse": 71300, + "ĠApiService": 78751, + "ĠApocalypse": 79197, + "ĠApollo": 35150, + "ĠApost": 46759, + "ĠApostle": 80959, + "ĠApp": 1845, + "ĠAppBar": 41301, + "ĠAppBundle": 69135, + "ĠAppCompatActivity": 24125, + "ĠAppComponent": 30492, + "ĠAppConfig": 63385, + "ĠAppDelegate": 35846, + "ĠAppMethodBeat": 83476, + "ĠAppModule": 48537, + "ĠAppRoutingModule": 43207, + "ĠAppState": 74891, + "ĠAppalach": 77727, + "ĠAppalachian": 97927, + "ĠApparel": 81600, + "ĠApparently": 40874, + "ĠAppe": 28054, + "ĠAppeal": 60906, + "ĠAppeals": 47107, + "ĠAppearance": 60616, + "ĠAppears": 93017, + "ĠAppend": 29807, + "ĠAppendix": 64714, + "ĠAppet": 92279, + "ĠAppl": 57845, + "ĠApple": 8162, + "ĠAppleWebKit": 59688, + "ĠAppliances": 96333, + "ĠApplicant": 91653, + "ĠApplicants": 97689, + "ĠApplication": 7323, + "ĠApplicationContext": 91518, + "ĠApplicationController": 30938, + "ĠApplicationDbContext": 73224, + "ĠApplicationException": 95818, + "ĠApplicationRecord": 54826, + "ĠApplicationUser": 93143, + "ĠApplications": 31525, + "ĠApplied": 42508, + "ĠApplies": 82165, + "ĠApply": 20552, + "ĠApplying": 89601, + "ĠAppointment": 56313, + "ĠApprec": 92821, + "ĠApprent": 68064, + "ĠApprentice": 83891, + "ĠAppro": 18642, + "ĠApproach": 53084, + "ĠAppropri": 85722, + "ĠApproval": 63777, + "ĠApproved": 51075, + "ĠApprox": 43819, + "ĠApproximately": 78804, + "ĠApps": 33020, + "ĠApr": 5076, + "ĠApril": 5813, + "ĠAprès": 133846, + "ĠApt": 96987, + "ĠApós": 140227, + "ĠAqu": 22959, + "ĠAqua": 71407, + "ĠAquarium": 80322, + "ĠAr": 1644, + "ĠAra": 87369, + "ĠArab": 10756, + "ĠArabia": 23061, + "ĠArabian": 72598, + "ĠArabic": 34117, + "ĠArabs": 57609, + "ĠAralık": 135680, + "ĠAraÅŁtırma": 142965, + "ĠArb": 86887, + "ĠArbeit": 57803, + "ĠArbeits": 70165, + "ĠArbit": 58795, + "ĠArbitrary": 97898, + "ĠArbor": 64539, + "ĠArc": 19689, + "ĠArcade": 64207, + "ĠArch": 9494, + "ĠArchae": 65002, + "ĠArchbishop": 81211, + "ĠArcher": 58879, + "ĠArchie": 90606, + "ĠArchitect": 24018, + "ĠArchitects": 77013, + "ĠArchitecture": 37843, + "ĠArchive": 29702, + "ĠArchived": 97750, + "ĠArchives": 37229, + "ĠArctic": 36418, + "ĠArd": 73907, + "ĠArduino": 46669, + "ĠAre": 8713, + "ĠArea": 12030, + "ĠAreas": 55716, + "ĠAren": 73504, + "ĠArena": 27047, + "ĠArg": 7638, + "ĠArgentina": 31064, + "ĠArgentine": 81722, + "ĠArgs": 17693, + "ĠArgument": 13818, + "ĠArgumentError": 76884, + "ĠArgumentException": 39808, + "ĠArgumentNullException": 29335, + "ĠArgumentOutOfRangeException": 76521, + "ĠArguments": 27702, + "ĠAri": 41526, + "ĠArial": 52771, + "ĠAriel": 75826, + "ĠArist": 51929, + "ĠAristotle": 77901, + "ĠArithmetic": 92984, + "ĠArizona": 16938, + "ĠArk": 22017, + "ĠArkadaÅŁ": 142026, + "ĠArkansas": 31690, + "ĠArlington": 58696, + "ĠArm": 12990, + "ĠArmed": 41394, + "ĠArmen": 38378, + "ĠArmenia": 71530, + "ĠArmenian": 66742, + "ĠArmor": 31183, + "ĠArmour": 60585, + "ĠArms": 43616, + "ĠArmstrong": 44866, + "ĠArmy": 13011, + "ĠArn": 35094, + "ĠArnold": 43212, + "ĠAround": 32816, + "ĠArr": 18418, + "ĠArrange": 40580, + "ĠArray": 2910, + "ĠArrayAdapter": 49640, + "ĠArrayBuffer": 83005, + "ĠArrayCollection": 98502, + "ĠArrayList": 5056, + "ĠArrays": 22966, + "ĠArrest": 71317, + "ĠArrival": 92098, + "ĠArrow": 33712, + "ĠArs": 72838, + "ĠArsenal": 32002, + "ĠArt": 5166, + "ĠArte": 97500, + "ĠArtem": 78266, + "ĠArthur": 27586, + "ĠArticle": 13355, + "ĠArticles": 28361, + "ĠArtifact": 86082, + "ĠArtificial": 58194, + "ĠArtikel": 50201, + "ĠArtist": 28359, + "ĠArtists": 52104, + "ĠArts": 17520, + "ĠArtÃŃculo": 140308, + "ĠAry": 86310, + "ĠAs": 1634, + "ĠAsc": 39560, + "ĠAsh": 14595, + "ĠAshe": 75179, + "ĠAsheville": 96538, + "ĠAshley": 37118, + "ĠAshton": 86618, + "ĠAsi": 87710, + "ĠAsia": 13622, + "ĠAsian": 14533, + "ĠAsians": 74308, + "ĠAside": 56094, + "ĠAsk": 20437, + "ĠAsked": 52663, + "ĠAsking": 95149, + "ĠAspect": 80053, + "ĠAspen": 93228, + "ĠAsphalt": 89314, + "ĠAss": 2688, + "ĠAssad": 37516, + "ĠAssange": 59407, + "ĠAssass": 98571, + "ĠAssassin": 60001, + "ĠAssault": 48971, + "ĠAssembly": 11738, + "ĠAssemblyCompany": 93891, + "ĠAssemblyCopyright": 94017, + "ĠAssemblyDescription": 91391, + "ĠAssemblyFileVersion": 98143, + "ĠAssemblyProduct": 92435, + "ĠAssemblyTitle": 89590, + "ĠAssemblyTrademark": 95827, + "ĠAssemblyVersion": 65675, + "ĠAssert": 5319, + "ĠAssertion": 46730, + "ĠAssertionError": 66226, + "ĠAssertions": 59806, + "ĠAssess": 81835, + "ĠAssessment": 36257, + "ĠAsset": 22605, + "ĠAssetImage": 92206, + "ĠAssets": 40638, + "ĠAssign": 31639, + "ĠAssigned": 62624, + "ĠAssignment": 34427, + "ĠAssist": 53249, + "ĠAssistance": 45765, + "ĠAssistant": 21388, + "ĠAssoc": 92376, + "ĠAssoci": 7156, + "ĠAssociate": 32368, + "ĠAssociated": 25403, + "ĠAssociates": 39101, + "ĠAssociation": 10024, + "ĠAssociations": 96089, + "ĠAssume": 62197, + "ĠAssumes": 98107, + "ĠAssuming": 63964, + "ĠAssurance": 83782, + "ĠAssy": 97013, + "ĠAst": 20113, + "ĠAster": 67498, + "ĠAston": 65674, + "ĠAstr": 38516, + "ĠAstro": 64129, + "ĠAstronomy": 94703, + "ĠAstroph": 93102, + "ĠAstros": 71458, + "ĠAsus": 75173, + "ĠAsync": 21433, + "ĠAsyncCallback": 91436, + "ĠAsyncStorage": 71621, + "ĠAsyncTask": 65697, + "ĠAsÃŃ": 133920, + "ĠAt": 2411, + "ĠAtari": 83876, + "ĠAtat": 140195, + "ĠAtatür": 140196, + "ĠAtatürk": 140197, + "ĠAth": 19698, + "ĠAthe": 84837, + "ĠAthen": 58552, + "ĠAthena": 78689, + "ĠAthens": 45826, + "ĠAthletic": 50406, + "ĠAthletics": 69731, + "ĠAtkins": 95984, + "ĠAtl": 10843, + "ĠAtlanta": 19440, + "ĠAtlantic": 22375, + "ĠAtlantis": 86935, + "ĠAtlas": 42343, + "ĠAtmos": 53828, + "ĠAtmospheric": 86497, + "ĠAtom": 39516, + "ĠAtomic": 30316, + "ĠAtomicInteger": 85038, + "ĠAtt": 7711, + "ĠAttach": 48384, + "ĠAttached": 83436, + "ĠAttachment": 55748, + "ĠAttack": 20790, + "ĠAttacks": 66142, + "ĠAttempt": 43517, + "ĠAttempts": 83300, + "ĠAttend": 78773, + "ĠAttendance": 78957, + "ĠAttention": 62020, + "ĠAttorney": 18412, + "ĠAttr": 50943, + "ĠAttribute": 16752, + "ĠAttributeError": 53631, + "ĠAttributeSet": 57440, + "ĠAttributes": 34127, + "ĠAttribution": 58318, + "ĠAu": 32050, + "ĠAub": 38587, + "ĠAuburn": 52349, + "ĠAuch": 55448, + "ĠAuckland": 53246, + "ĠAuction": 44100, + "ĠAud": 15066, + "ĠAudi": 39828, + "ĠAudience": 64900, + "ĠAudio": 12352, + "ĠAudioClip": 69137, + "ĠAudioManager": 95881, + "ĠAudioSource": 69221, + "ĠAudit": 47646, + "ĠAuditor": 61392, + "ĠAudrey": 90049, + "ĠAuf": 25714, + "ĠAug": 4928, + "ĠAugust": 6156, + "ĠAugusta": 83475, + "ĠAugustine": 73465, + "ĠAunt": 70222, + "ĠAur": 34631, + "ĠAura": 62286, + "ĠAure": 88215, + "ĠAurora": 46792, + "ĠAus": 20004, + "ĠAuschwitz": 92290, + "ĠAuss": 50244, + "ĠAussie": 86463, + "ĠAust": 12926, + "ĠAustin": 19260, + "ĠAustral": 5998, + "ĠAustralia": 8330, + "ĠAustralian": 13369, + "ĠAustralians": 55837, + "ĠAustria": 34898, + "ĠAustrian": 56969, + "ĠAuswahl": 95830, + "ĠAut": 9460, + "ĠAuth": 7366, + "ĠAuthService": 38598, + "ĠAuthentic": 46367, + "ĠAuthenticate": 93487, + "ĠAuthentication": 23537, + "ĠAuthenticationService": 94733, + "ĠAuthor": 6885, + "ĠAuthorities": 67315, + "ĠAuthority": 21914, + "ĠAuthorization": 30562, + "ĠAuthorized": 90569, + "ĠAuthors": 24124, + "ĠAutism": 77663, + "ĠAuto": 8979, + "ĠAutoMapper": 68082, + "ĠAutodesk": 90235, + "ĠAutof": 93822, + "ĠAutom": 19739, + "ĠAutomated": 94723, + "ĠAutomatic": 33981, + "ĠAutomatically": 70150, + "ĠAutomation": 53778, + "ĠAutomobile": 85141, + "ĠAutomotive": 54418, + "ĠAutonomous": 96448, + "ĠAutor": 56872, + "ĠAutos": 91552, + "ĠAutumn": 59802, + "ĠAux": 50059, + "ĠAuxiliary": 95500, + "ĠAuÃŁen": 142791, + "ĠAuÃŁerdem": 141196, + "ĠAv": 7519, + "ĠAva": 91657, + "ĠAvailability": 51810, + "ĠAvailable": 16136, + "ĠAval": 48978, + "ĠAvalanche": 98707, + "ĠAvatar": 37609, + "ĠAve": 21294, + "ĠAvec": 61193, + "ĠAvengers": 43097, + "ĠAvenue": 17131, + "ĠAverage": 23565, + "ĠAvery": 67480, + "ĠAvg": 62080, + "ĠAviation": 48947, + "ĠAviv": 65847, + "ĠAvoid": 34006, + "ĠAw": 17889, + "ĠAwait": 98195, + "ĠAwake": 41222, + "ĠAwakening": 93657, + "ĠAward": 17318, + "ĠAwards": 22658, + "ĠAware": 54540, + "ĠAwareness": 62035, + "ĠAway": 41481, + "ĠAwesome": 38305, + "ĠAws": 40083, + "ĠAx": 25171, + "ĠAxe": 71851, + "ĠAxel": 86679, + "ĠAxes": 89704, + "ĠAxios": 50214, + "ĠAxis": 34474, + "ĠAy": 23905, + "ĠAydın": 142228, + "ĠAynı": 139673, + "ĠAyrıca": 132839, + "ĠAz": 15395, + "ĠAzerbai": 64323, + "ĠAzerbaijan": 73477, + "ĠAzure": 34119, + "ĠAç": 139657, + "ĠAçık": 139658, + "ĠAÄŁ": 136777, + "ĠAÄŁust": 136778, + "ĠAÄŁustos": 136779, + "ĠB": 425, + "ĠBA": 33489, + "ĠBAB": 59311, + "ĠBACK": 32082, + "ĠBACKGROUND": 96614, + "ĠBAD": 47718, + "ĠBAL": 69145, + "ĠBALL": 83848, + "ĠBAM": 92127, + "ĠBANK": 66010, + "ĠBAR": 44187, + "ĠBAS": 10765, + "ĠBASE": 22194, + "ĠBASIC": 80048, + "ĠBASIS": 11389, + "ĠBAT": 77981, + "ĠBATCH": 98791, + "ĠBB": 18270, + "ĠBBB": 94458, + "ĠBBC": 18096, + "ĠBBQ": 45277, + "ĠBC": 18040, + "ĠBCE": 78577, + "ĠBCH": 61172, + "ĠBCHP": 78726, + "ĠBCM": 48220, + "ĠBD": 39310, + "ĠBDS": 59827, + "ĠBDSM": 84750, + "ĠBE": 7206, + "ĠBED": 82264, + "ĠBEEN": 74653, + "ĠBEFORE": 56482, + "ĠBEGIN": 22203, + "ĠBEL": 68259, + "ĠBELOW": 64393, + "ĠBEN": 82834, + "ĠBER": 93574, + "ĠBEST": 45782, + "ĠBET": 40349, + "ĠBETWEEN": 56639, + "ĠBF": 50504, + "ĠBFS": 90351, + "ĠBG": 43011, + "ĠBH": 67747, + "ĠBI": 47053, + "ĠBIG": 36854, + "ĠBILL": 92646, + "ĠBIN": 82609, + "ĠBIND": 91370, + "ĠBIO": 72066, + "ĠBIOS": 52863, + "ĠBIT": 24247, + "ĠBITS": 91149, + "ĠBJ": 86499, + "ĠBJP": 40067, + "ĠBK": 76782, + "ĠBL": 14850, + "ĠBLACK": 36647, + "ĠBLE": 51718, + "ĠBLL": 87933, + "ĠBLOCK": 28677, + "ĠBLUE": 55892, + "ĠBM": 19800, + "ĠBMC": 81706, + "ĠBMI": 46124, + "ĠBMP": 75194, + "ĠBMW": 29764, + "ĠBN": 45316, + "ĠBO": 7811, + "ĠBOARD": 51353, + "ĠBODY": 68583, + "ĠBOOK": 47098, + "ĠBOOL": 16573, + "ĠBOOLEAN": 59393, + "ĠBOOST": 21203, + "ĠBOOT": 83676, + "ĠBORDER": 84272, + "ĠBOT": 82330, + "ĠBOTH": 85489, + "ĠBOTTOM": 90571, + "ĠBOX": 52683, + "ĠBP": 29067, + "ĠBPM": 88219, + "ĠBR": 18803, + "ĠBRA": 73763, + "ĠBRAND": 89954, + "ĠBRE": 80562, + "ĠBREAK": 69647, + "ĠBRO": 77587, + "ĠBS": 27618, + "ĠBSD": 23102, + "ĠBSON": 76235, + "ĠBSP": 60417, + "ĠBST": 43541, + "ĠBT": 21111, + "ĠBTC": 36045, + "ĠBTN": 80415, + "ĠBTS": 82068, + "ĠBU": 54029, + "ĠBUF": 73178, + "ĠBUFF": 94924, + "ĠBUFFER": 52135, + "ĠBUG": 68534, + "ĠBUILD": 47697, + "ĠBUS": 22673, + "ĠBUSINESS": 26595, + "ĠBUT": 10915, + "ĠBUTTON": 41389, + "ĠBUY": 51762, + "ĠBV": 40100, + "ĠBW": 51120, + "ĠBX": 63275, + "ĠBY": 7710, + "ĠBYTE": 35377, + "ĠBYU": 92354, + "ĠBa": 14322, + "ĠBab": 25369, + "ĠBaba": 96352, + "ĠBabe": 73348, + "ĠBabies": 92692, + "ĠBaby": 20617, + "ĠBabylon": 63578, + "ĠBac": 68296, + "ĠBach": 41158, + "ĠBachelor": 40990, + "ĠBack": 6841, + "ĠBackbone": 53907, + "ĠBackend": 55260, + "ĠBackground": 24800, + "ĠBackpack": 71660, + "ĠBackup": 43438, + "ĠBacon": 61384, + "ĠBad": 11461, + "ĠBadRequest": 36403, + "ĠBadge": 59083, + "ĠBag": 19708, + "ĠBaghd": 52241, + "ĠBaghdad": 55218, + "ĠBags": 63723, + "ĠBah": 31329, + "ĠBahamas": 83129, + "ĠBahrain": 65639, + "ĠBai": 84908, + "ĠBail": 82144, + "ĠBailey": 42214, + "ĠBain": 81253, + "ĠBair": 97678, + "ĠBak": 35669, + "ĠBakan": 127762, + "ĠBakanlıģı": 135490, + "ĠBakanı": 138935, + "ĠBake": 52074, + "ĠBaker": 28392, + "ĠBakery": 90615, + "ĠBal": 19420, + "ĠBalance": 30846, + "ĠBalanced": 97109, + "ĠBald": 39102, + "ĠBaldwin": 54918, + "ĠBale": 83130, + "ĠBali": 63228, + "ĠBalk": 59869, + "ĠBall": 12836, + "ĠBallard": 97716, + "ĠBallet": 86636, + "ĠBalls": 70603, + "ĠBalt": 20581, + "ĠBaltic": 71989, + "ĠBaltimore": 23947, + "ĠBam": 55007, + "ĠBamboo": 96928, + "ĠBan": 22730, + "ĠBanana": 75824, + "ĠBanc": 83077, + "ĠBanco": 76687, + "ĠBand": 16936, + "ĠBands": 87302, + "ĠBang": 16914, + "ĠBangalore": 58125, + "ĠBangkok": 59016, + "ĠBangladesh": 38501, + "ĠBank": 8547, + "ĠBanking": 49499, + "ĠBanks": 37766, + "ĠBanner": 39614, + "ĠBannon": 54165, + "ĠBans": 93961, + "ĠBapt": 34876, + "ĠBaptist": 42648, + "ĠBar": 4716, + "ĠBarack": 23539, + "ĠBarang": 87607, + "ĠBarb": 46042, + "ĠBarbar": 90070, + "ĠBarbara": 31107, + "ĠBarber": 68135, + "ĠBarbie": 83239, + "ĠBarcelona": 26937, + "ĠBarcl": 67912, + "ĠBarclays": 88745, + "ĠBarcode": 86732, + "ĠBard": 63204, + "ĠBare": 60792, + "ĠBarg": 86714, + "ĠBark": 59529, + "ĠBarker": 80404, + "ĠBarn": 22421, + "ĠBarnes": 43854, + "ĠBarnett": 85320, + "ĠBarney": 93429, + "ĠBaron": 52907, + "ĠBarr": 31717, + "ĠBarrel": 63971, + "ĠBarrett": 55759, + "ĠBarrier": 70987, + "ĠBarry": 30758, + "ĠBars": 56106, + "ĠBart": 28527, + "ĠBarth": 88664, + "ĠBarton": 69217, + "ĠBas": 14662, + "ĠBase": 5351, + "ĠBaseActivity": 74645, + "ĠBaseController": 55461, + "ĠBaseEntity": 74082, + "ĠBaseModel": 64605, + "ĠBaseService": 98563, + "ĠBaseType": 88338, + "ĠBaseball": 37158, + "ĠBased": 20205, + "ĠBasel": 89458, + "ĠBasement": 92043, + "ĠBash": 38530, + "ĠBashar": 70260, + "ĠBasic": 14625, + "ĠBasically": 40712, + "ĠBasics": 67176, + "ĠBasil": 65970, + "ĠBasin": 60258, + "ĠBasis": 69028, + "ĠBasket": 33117, + "ĠBasketball": 46489, + "ĠBass": 35953, + "ĠBast": 45068, + "ĠBat": 16098, + "ĠBatch": 33904, + "ĠBates": 73862, + "ĠBath": 19161, + "ĠBathroom": 36832, + "ĠBaths": 74442, + "ĠBatman": 26707, + "ĠBaton": 85169, + "ĠBatt": 32167, + "ĠBattalion": 73639, + "ĠBatter": 72918, + "ĠBattery": 33612, + "ĠBattle": 16115, + "ĠBattlefield": 70635, + "ĠBattles": 70156, + "ĠBau": 73360, + "ĠBauer": 78601, + "ĠBaum": 82220, + "ĠBav": 76980, + "ĠBaxter": 91747, + "ĠBay": 9154, + "ĠBayer": 85420, + "ĠBayern": 55591, + "ĠBayesian": 98134, + "ĠBaylor": 70047, + "ĠBaz": 91711, + "ĠBaÅŁ": 97505, + "ĠBaÅŁb": 140160, + "ĠBaÅŁbakan": 140161, + "ĠBaÅŁkan": 125685, + "ĠBaÅŁkanlıģı": 142877, + "ĠBaÅŁkanı": 129580, + "ĠBbw": 85597, + "ĠBd": 73560, + "ĠBe": 2823, + "ĠBea": 60714, + "ĠBeach": 12721, + "ĠBeacon": 58620, + "ĠBeam": 50130, + "ĠBean": 33723, + "ĠBeans": 80736, + "ĠBear": 23985, + "ĠBeard": 87430, + "ĠBearing": 38668, + "ĠBearings": 64316, + "ĠBears": 35968, + "ĠBeast": 33182, + "ĠBeat": 26796, + "ĠBeatles": 54857, + "ĠBeats": 84520, + "ĠBeau": 79227, + "ĠBeaut": 14469, + "ĠBeautiful": 19490, + "ĠBeautifulSoup": 35910, + "ĠBeauty": 30585, + "ĠBeaver": 91434, + "ĠBec": 31046, + "ĠBecame": 90044, + "ĠBecause": 9211, + "ĠBeck": 28718, + "ĠBecker": 77384, + "ĠBeckham": 80003, + "ĠBecky": 87601, + "ĠBecome": 50278, + "ĠBed": 13093, + "ĠBedford": 72668, + "ĠBedroom": 32855, + "ĠBedrooms": 67730, + "ĠBeds": 66254, + "ĠBee": 37430, + "ĠBeef": 63125, + "ĠBeen": 50578, + "ĠBeer": 33384, + "ĠBeet": 73949, + "ĠBeetle": 91696, + "ĠBefore": 13235, + "ĠBeg": 43693, + "ĠBegin": 18597, + "ĠBeginner": 92175, + "ĠBeginners": 90168, + "ĠBeginning": 51850, + "ĠBegins": 91308, + "ĠBeh": 14612, + "ĠBehavior": 26675, + "ĠBehaviorSubject": 61638, + "ĠBehavioral": 71984, + "ĠBehaviour": 79254, + "ĠBehind": 42374, + "ĠBei": 47436, + "ĠBeijing": 26549, + "ĠBeing": 20690, + "ĠBeirut": 94311, + "ĠBeispiel": 79192, + "ĠBeit": 75030, + "ĠBeitrag": 98552, + "ĠBek": 70219, + "ĠBel": 7828, + "ĠBelarus": 69506, + "ĠBelfast": 54239, + "ĠBelg": 22475, + "ĠBelgi": 143391, + "ĠBelgian": 48062, + "ĠBelgium": 32961, + "ĠBelgië": 143392, + "ĠBelieve": 58021, + "ĠBelize": 97004, + "ĠBell": 17884, + "ĠBella": 54500, + "ĠBelle": 50508, + "ĠBellev": 85605, + "ĠBelly": 96585, + "ĠBelmont": 94966, + "ĠBelow": 21193, + "ĠBelt": 32893, + "ĠBen": 7355, + "ĠBench": 35258, + "ĠBenchmark": 61622, + "ĠBend": 47738, + "ĠBender": 96545, + "ĠBene": 86591, + "ĠBened": 51620, + "ĠBenedict": 61145, + "ĠBenef": 26276, + "ĠBenefit": 66350, + "ĠBenefits": 38095, + "ĠBeng": 25250, + "ĠBengal": 49020, + "ĠBengals": 63829, + "ĠBenghazi": 67768, + "ĠBenjamin": 29311, + "ĠBenn": 29780, + "ĠBennett": 38286, + "ĠBenny": 94185, + "ĠBenson": 71188, + "ĠBent": 40514, + "ĠBentley": 73420, + "ĠBenton": 92683, + "ĠBenz": 77542, + "ĠBer": 8907, + "ĠBere": 43157, + "ĠBereich": 74219, + "ĠBerg": 30682, + "ĠBergen": 78999, + "ĠBerger": 68625, + "ĠBerk": 50191, + "ĠBerkeley": 32008, + "ĠBerkshire": 90275, + "ĠBerlin": 19846, + "ĠBerm": 75072, + "ĠBermuda": 90266, + "ĠBern": 14168, + "ĠBernard": 34252, + "ĠBernardino": 85757, + "ĠBernie": 29224, + "ĠBernstein": 81315, + "ĠBerry": 43585, + "ĠBers": 86275, + "ĠBert": 50675, + "ĠBes": 18078, + "ĠBesch": 77814, + "ĠBesides": 30809, + "ĠBest": 7107, + "ĠBeste": 69351, + "ĠBet": 13017, + "ĠBeta": 34243, + "ĠBeth": 28003, + "ĠBethesda": 78331, + "ĠBethlehem": 94958, + "ĠBett": 67588, + "ĠBetter": 23434, + "ĠBetting": 83985, + "ĠBetty": 55011, + "ĠBetween": 27133, + "ĠBever": 37999, + "ĠBeverage": 92035, + "ĠBeverly": 57674, + "ĠBew": 37210, + "ĠBewert": 85651, + "ĠBey": 53892, + "ĠBeyond": 30786, + "ĠBez": 96889, + "ĠBh": 30830, + "ĠBhar": 66592, + "ĠBi": 12100, + "ĠBian": 66945, + "ĠBias": 82990, + "ĠBiaÅĤ": 139425, + "ĠBib": 29445, + "ĠBible": 16947, + "ĠBibli": 61826, + "ĠBiblical": 67101, + "ĠBicycle": 78347, + "ĠBid": 40402, + "ĠBiden": 37080, + "ĠBieber": 76220, + "ĠBien": 73556, + "ĠBig": 6164, + "ĠBigDecimal": 20618, + "ĠBigInt": 62608, + "ĠBigInteger": 34042, + "ĠBigNumber": 53673, + "ĠBiggest": 85521, + "ĠBihar": 92274, + "ĠBij": 79566, + "ĠBik": 75378, + "ĠBike": 37830, + "ĠBil": 38058, + "ĠBild": 44492, + "ĠBilder": 56580, + "ĠBill": 8596, + "ĠBillboard": 66193, + "ĠBilling": 72450, + "ĠBillion": 65709, + "ĠBills": 45640, + "ĠBilly": 32819, + "ĠBin": 29344, + "ĠBinary": 17718, + "ĠBinaryTree": 98070, + "ĠBind": 29189, + "ĠBinder": 90163, + "ĠBinding": 25901, + "ĠBindingFlags": 58678, + "ĠBing": 53487, + "ĠBingo": 92048, + "ĠBio": 23525, + "ĠBiography": 78169, + "ĠBiol": 80794, + "ĠBiological": 62611, + "ĠBiology": 38923, + "ĠBiom": 90508, + "ĠBios": 76848, + "ĠBip": 97848, + "ĠBir": 36819, + "ĠBirch": 86988, + "ĠBird": 23331, + "ĠBirds": 56528, + "ĠBirliÄŁi": 141302, + "ĠBirmingham": 35837, + "ĠBirth": 35027, + "ĠBirthday": 36240, + "ĠBis": 42584, + "ĠBishop": 33242, + "ĠBison": 88043, + "ĠBit": 6495, + "ĠBitConverter": 72884, + "ĠBitSet": 57227, + "ĠBitcoin": 13127, + "ĠBitcoins": 81825, + "ĠBite": 88208, + "ĠBitmap": 17533, + "ĠBitmapFactory": 69896, + "ĠBits": 49457, + "ĠBitte": 84276, + "ĠBiz": 61450, + "ĠBj": 68741, + "ĠBl": 2502, + "ĠBlack": 5235, + "ĠBlackBerry": 63428, + "ĠBlackburn": 88353, + "ĠBlackhawks": 92392, + "ĠBlackjack": 87092, + "ĠBlacks": 66488, + "ĠBlade": 35980, + "ĠBlades": 88960, + "ĠBlair": 41869, + "ĠBlake": 30894, + "ĠBlanc": 68925, + "ĠBlanch": 91429, + "ĠBlanco": 98823, + "ĠBlank": 42441, + "ĠBlasio": 92393, + "ĠBlast": 57336, + "ĠBlaze": 83427, + "ĠBlazers": 89689, + "ĠBle": 38185, + "ĠBlend": 54148, + "ĠBlender": 87568, + "ĠBless": 43261, + "ĠBlessed": 75997, + "ĠBlick": 81287, + "ĠBlind": 54270, + "ĠBlink": 73679, + "ĠBliss": 91570, + "ĠBlitz": 88128, + "ĠBlizzard": 52033, + "ĠBlo": 39391, + "ĠBlob": 49439, + "ĠBloc": 71396, + "ĠBlock": 8362, + "ĠBlockPos": 79064, + "ĠBlockchain": 42885, + "ĠBlocked": 63996, + "ĠBlocking": 81945, + "ĠBlockly": 81918, + "ĠBlocks": 45678, + "ĠBlog": 14163, + "ĠBlogger": 87675, + "ĠBlogs": 75236, + "ĠBlond": 91998, + "ĠBlonde": 52456, + "ĠBlood": 20070, + "ĠBloody": 93572, + "ĠBloom": 24503, + "ĠBloomberg": 36553, + "ĠBloss": 88759, + "ĠBlow": 45852, + "ĠBlowjob": 84010, + "ĠBlu": 37769, + "ĠBlue": 8697, + "ĠBlueprint": 30003, + "ĠBlueprintReadOnly": 93125, + "ĠBlues": 32325, + "ĠBluetooth": 23843, + "ĠBlur": 98160, + "ĠBlvd": 49441, + "ĠBo": 2516, + "ĠBoard": 8751, + "ĠBoards": 71086, + "ĠBoat": 44232, + "ĠBob": 14261, + "ĠBobby": 37381, + "ĠBoca": 98557, + "ĠBod": 44602, + "ĠBodies": 80346, + "ĠBody": 13958, + "ĠBoeh": 69197, + "ĠBoehner": 77167, + "ĠBoeing": 41859, + "ĠBog": 41548, + "ĠBoh": 51350, + "ĠBoiler": 45665, + "ĠBoise": 78573, + "ĠBol": 24043, + "ĠBold": 46002, + "ĠBolivia": 76610, + "ĠBollywood": 64186, + "ĠBols": 67257, + "ĠBolshevik": 91401, + "ĠBolt": 45947, + "ĠBolton": 61577, + "ĠBom": 47352, + "ĠBomb": 32809, + "ĠBombay": 96613, + "ĠBon": 13481, + "ĠBond": 23618, + "ĠBonds": 84026, + "ĠBone": 45601, + "ĠBones": 71299, + "ĠBonnie": 73195, + "ĠBonus": 28198, + "ĠBoo": 73684, + "ĠBook": 5893, + "ĠBooker": 64995, + "ĠBooking": 44176, + "ĠBookmark": 55418, + "ĠBooks": 17833, + "ĠBool": 12608, + "ĠBoolean": 6992, + "ĠBoom": 53411, + "ĠBoone": 90595, + "ĠBoost": 33407, + "ĠBooster": 74192, + "ĠBoot": 15004, + "ĠBooth": 63270, + "ĠBoots": 45468, + "ĠBootstrap": 26059, + "ĠBor": 22918, + "ĠBord": 97885, + "ĠBordeaux": 83957, + "ĠBorder": 13992, + "ĠBorderLayout": 48357, + "ĠBorderRadius": 31594, + "ĠBorderSide": 72401, + "ĠBorders": 80712, + "ĠBorg": 64482, + "ĠBoris": 51671, + "ĠBorn": 37816, + "ĠBorough": 84203, + "ĠBorrow": 77947, + "ĠBos": 27971, + "ĠBosch": 71225, + "ĠBose": 93458, + "ĠBosnia": 76328, + "ĠBoss": 31569, + "ĠBoston": 10196, + "ĠBot": 23007, + "ĠBoth": 11733, + "ĠBott": 36230, + "ĠBottle": 53985, + "ĠBottom": 25737, + "ĠBou": 29954, + "ĠBought": 77162, + "ĠBoulder": 55643, + "ĠBoulevard": 47440, + "ĠBound": 37176, + "ĠBoundary": 76416, + "ĠBounds": 69130, + "ĠBounty": 94189, + "ĠBour": 38375, + "ĠBourbon": 92504, + "ĠBout": 63103, + "ĠBoutique": 90117, + "ĠBow": 18436, + "ĠBowen": 95520, + "ĠBowie": 67247, + "ĠBowl": 20288, + "ĠBowling": 82969, + "ĠBowman": 84596, + "ĠBox": 8261, + "ĠBoxDecoration": 36093, + "ĠBoxFit": 69354, + "ĠBoxes": 71267, + "ĠBoxing": 71583, + "ĠBoy": 16180, + "ĠBoyd": 62621, + "ĠBoyle": 82059, + "ĠBoys": 29757, + "ĠBr": 3240, + "ĠBra": 25453, + "ĠBrace": 69206, + "ĠBracket": 85497, + "ĠBrad": 17041, + "ĠBradford": 59829, + "ĠBradley": 36448, + "ĠBrady": 35370, + "ĠBrah": 76984, + "ĠBrain": 30317, + "ĠBrake": 75261, + "ĠBram": 70784, + "ĠBran": 75918, + "ĠBranch": 25119, + "ĠBrand": 16430, + "ĠBrandon": 31766, + "ĠBrands": 54232, + "ĠBras": 61124, + "ĠBrasil": 41925, + "ĠBrass": 58029, + "ĠBraun": 69603, + "ĠBrave": 69208, + "ĠBraves": 70677, + "ĠBravo": 70723, + "ĠBray": 87820, + "ĠBraz": 72311, + "ĠBrazil": 15948, + "ĠBrazilian": 34983, + "ĠBre": 11427, + "ĠBread": 49041, + "ĠBreak": 15623, + "ĠBreakfast": 43191, + "ĠBreaking": 51524, + "ĠBreast": 56498, + "ĠBreath": 57192, + "ĠBreed": 74152, + "ĠBreitbart": 55876, + "ĠBren": 43708, + "ĠBrend": 38964, + "ĠBrenda": 89034, + "ĠBrendan": 55843, + "ĠBrennan": 58750, + "ĠBrent": 46331, + "ĠBret": 71677, + "ĠBrett": 39103, + "ĠBrew": 20923, + "ĠBrewer": 82967, + "ĠBrewers": 80411, + "ĠBrewery": 65627, + "ĠBrewing": 43113, + "ĠBrexit": 23907, + "ĠBri": 54758, + "ĠBrian": 17083, + "ĠBrick": 44614, + "ĠBrid": 31337, + "ĠBride": 77060, + "ĠBridge": 19874, + "ĠBridges": 76239, + "ĠBrief": 36518, + "ĠBrig": 37789, + "ĠBrigade": 66214, + "ĠBriggs": 93117, + "ĠBrigham": 95185, + "ĠBright": 26410, + "ĠBrighton": 57419, + "ĠBrill": 66644, + "ĠBrilliant": 93274, + "ĠBring": 39124, + "ĠBringing": 97073, + "ĠBris": 43731, + "ĠBrisbane": 46235, + "ĠBristol": 43939, + "ĠBrit": 5450, + "ĠBritain": 13224, + "ĠBritann": 97420, + "ĠBritish": 7855, + "ĠBritt": 63017, + "ĠBrittany": 88154, + "ĠBro": 5906, + "ĠBroad": 15894, + "ĠBroadcast": 31965, + "ĠBroadcastReceiver": 96064, + "ĠBroadcasting": 63460, + "ĠBroadcom": 95917, + "ĠBroadway": 36676, + "ĠBrock": 47577, + "ĠBroken": 53583, + "ĠBroker": 52701, + "ĠBrom": 64598, + "ĠBron": 19820, + "ĠBroncos": 41594, + "ĠBronx": 65136, + "ĠBronze": 44867, + "ĠBrook": 15605, + "ĠBrooke": 78461, + "ĠBrooklyn": 25748, + "ĠBrooks": 38019, + "ĠBros": 33221, + "ĠBrother": 26348, + "ĠBrotherhood": 49236, + "ĠBrothers": 33079, + "ĠBrow": 28377, + "ĠBrown": 10471, + "ĠBrowns": 45199, + "ĠBrowse": 41712, + "ĠBrowser": 16482, + "ĠBrowserAnimationsModule": 71436, + "ĠBrowserModule": 37371, + "ĠBrowserRouter": 62485, + "ĠBru": 18697, + "ĠBruce": 23845, + "ĠBruins": 64902, + "ĠBrun": 34461, + "ĠBruno": 51110, + "ĠBrunswick": 60150, + "ĠBrush": 32988, + "ĠBrushes": 66872, + "ĠBrussels": 37169, + "ĠBrut": 70725, + "ĠBry": 19248, + "ĠBryan": 36353, + "ĠBryant": 39592, + "ĠBryce": 83024, + "ĠBs": 82186, + "ĠBtn": 56308, + "ĠBu": 27227, + "ĠBubble": 53324, + "ĠBucc": 78247, + "ĠBuccane": 85770, + "ĠBuccaneers": 92068, + "ĠBuch": 33278, + "ĠBuchanan": 84190, + "ĠBuck": 26064, + "ĠBucket": 47768, + "ĠBuckingham": 94911, + "ĠBuckley": 93303, + "ĠBucks": 65048, + "ĠBud": 36135, + "ĠBudapest": 69595, + "ĠBudd": 18586, + "ĠBuddh": 49518, + "ĠBuddha": 44954, + "ĠBuddhism": 60224, + "ĠBuddhist": 46741, + "ĠBuddy": 59526, + "ĠBudget": 27269, + "ĠBuen": 53841, + "ĠBuenos": 68073, + "ĠBuf": 69013, + "ĠBuff": 18961, + "ĠBuffalo": 31389, + "ĠBuffer": 10312, + "ĠBuffered": 30702, + "ĠBufferedImage": 56170, + "ĠBufferedReader": 25568, + "ĠBufferedWriter": 63129, + "ĠBuffett": 85452, + "ĠBuffy": 70840, + "ĠBug": 30501, + "ĠBugs": 86501, + "ĠBugün": 138213, + "ĠBuild": 7854, + "ĠBuilder": 20626, + "ĠBuilders": 75748, + "ĠBuilding": 16858, + "ĠBuildings": 73148, + "ĠBuilds": 67118, + "ĠBuilt": 33054, + "ĠBuk": 69850, + "ĠBukkit": 94672, + "ĠBul": 13497, + "ĠBulg": 41250, + "ĠBulgaria": 58719, + "ĠBulgarian": 88624, + "ĠBulk": 60920, + "ĠBull": 21622, + "ĠBulld": 56654, + "ĠBulldogs": 75464, + "ĠBullet": 31812, + "ĠBulletin": 59963, + "ĠBulls": 43689, + "ĠBun": 65247, + "ĠBund": 29608, + "ĠBundes": 46307, + "ĠBundesliga": 84803, + "ĠBundle": 13182, + "ĠBundy": 95821, + "ĠBunifu": 67321, + "ĠBunny": 72062, + "ĠBur": 12369, + "ĠBurb": 91633, + "ĠBureau": 21802, + "ĠBurg": 40009, + "ĠBurger": 51771, + "ĠBurgess": 97467, + "ĠBurk": 70640, + "ĠBurke": 49623, + "ĠBurl": 61333, + "ĠBurlington": 72505, + "ĠBurma": 91571, + "ĠBurn": 18039, + "ĠBurnett": 98894, + "ĠBurning": 54692, + "ĠBurns": 49916, + "ĠBurr": 86184, + "ĠBurst": 77467, + "ĠBurton": 53655, + "ĠBus": 18598, + "ĠBush": 14079, + "ĠBusiness": 8025, + "ĠBusinessException": 96994, + "ĠBusinesses": 70090, + "ĠBust": 70333, + "ĠBuster": 95487, + "ĠBusty": 85823, + "ĠBusy": 94709, + "ĠBut": 1988, + "ĠButler": 37225, + "ĠButt": 69349, + "ĠButter": 30357, + "ĠButterKnife": 95217, + "ĠButterfly": 77720, + "ĠButton": 6601, + "ĠButtons": 52700, + "ĠBuy": 11292, + "ĠBuyer": 61078, + "ĠBuyers": 88707, + "ĠBuying": 54309, + "ĠBuzz": 37951, + "ĠBuzzFeed": 71822, + "ĠBy": 3216, + "ĠByVal": 27089, + "ĠBye": 89325, + "ĠByrne": 80321, + "ĠByron": 70884, + "ĠByte": 10906, + "ĠByteArray": 32920, + "ĠByteArrayInputStream": 85625, + "ĠByteArrayOutputStream": 56263, + "ĠByteBuffer": 50299, + "ĠByteString": 95302, + "ĠBytes": 30024, + "ĠByz": 81660, + "ĠByÅĤ": 139727, + "ĠBá": 132040, + "ĠBáo": 133602, + "ĠBên": 133802, + "ĠBình": 129336, + "ĠBöl": 138918, + "ĠBölge": 138919, + "ĠBöyle": 136064, + "ĠBü": 68557, + "ĠBücher": 140918, + "ĠBürger": 136458, + "ĠBüro": 141506, + "ĠBüyük": 127827, + "ĠBüyükÅŁehir": 140018, + "ĠBÃł": 132934, + "ĠBÃłi": 130590, + "ĠBÃŃ": 134082, + "ĠBạn": 129096, + "ĠBản": 130164, + "ĠBảo": 132847, + "ĠBắc": 130726, + "ĠBá»ĩnh": 135256, + "ĠBá»Ļ": 129029, + "ĠBợi": 139194, + "ĠC": 356, + "ĠCA": 9183, + "ĠCAB": 80117, + "ĠCACHE": 65656, + "ĠCAD": 47265, + "ĠCAL": 32907, + "ĠCALC": 88249, + "ĠCALL": 25752, + "ĠCALLBACK": 73293, + "ĠCAM": 28297, + "ĠCAMERA": 66798, + "ĠCAN": 19508, + "ĠCAP": 26101, + "ĠCAPITAL": 77287, + "ĠCAR": 27776, + "ĠCARD": 41700, + "ĠCARE": 62327, + "ĠCART": 78032, + "ĠCAS": 41790, + "ĠCASCADE": 97059, + "ĠCASE": 38319, + "ĠCAST": 61687, + "ĠCAT": 43981, + "ĠCATEGORY": 86047, + "ĠCAUSED": 30377, + "ĠCActive": 77704, + "ĠCB": 21318, + "ĠCBC": 45512, + "ĠCBD": 22059, + "ĠCBS": 24029, + "ĠCC": 13534, + "ĠCCC": 91474, + "ĠCCD": 92733, + "ĠCCP": 75921, + "ĠCCS": 73092, + "ĠCCT": 54997, + "ĠCCTV": 70238, + "ĠCD": 11078, + "ĠCDC": 39309, + "ĠCDDL": 97919, + "ĠCDN": 88365, + "ĠCDs": 61337, + "ĠCE": 26711, + "ĠCEL": 90421, + "ĠCELL": 59695, + "ĠCENT": 61963, + "ĠCENTER": 65992, + "ĠCEO": 12156, + "ĠCEOs": 74217, + "ĠCERT": 62357, + "ĠCES": 80413, + "ĠCET": 87320, + "ĠCF": 20795, + "ĠCFG": 46825, + "ĠCFL": 96316, + "ĠCFO": 93752, + "ĠCFR": 69821, + "ĠCG": 6159, + "ĠCGAffineTransform": 68789, + "ĠCGContext": 54243, + "ĠCGFloat": 15931, + "ĠCGI": 63427, + "ĠCGPoint": 26353, + "ĠCGPointMake": 59747, + "ĠCGRect": 16006, + "ĠCGRectGet": 62250, + "ĠCGRectMake": 32830, + "ĠCGSize": 31815, + "ĠCGSizeMake": 78687, + "ĠCH": 6826, + "ĠCHANGE": 43039, + "ĠCHANNEL": 58756, + "ĠCHAPTER": 95324, + "ĠCHAR": 23997, + "ĠCHARACTER": 70805, + "ĠCHARSET": 97701, + "ĠCHE": 49521, + "ĠCHECK": 13644, + "ĠCHILD": 72520, + "ĠCHIP": 76483, + "ĠCHK": 87276, + "ĠCHO": 82353, + "ĠCHR": 61576, + "ĠCHRIST": 83565, + "ĠCHUNK": 98617, + "ĠCHtml": 51336, + "ĠCI": 20694, + "ĠCIA": 21474, + "ĠCID": 68212, + "ĠCIF": 97590, + "ĠCIM": 82364, + "ĠCIS": 74780, + "ĠCIT": 97477, + "ĠCITY": 46552, + "ĠCJ": 60876, + "ĠCK": 30066, + "ĠCL": 6976, + "ĠCLAIM": 25825, + "ĠCLASS": 27939, + "ĠCLEAN": 77000, + "ĠCLEAR": 55219, + "ĠCLI": 39277, + "ĠCLICK": 57232, + "ĠCLIENT": 40297, + "ĠCLK": 61342, + "ĠCLL": 50296, + "ĠCLLocation": 52659, + "ĠCLLocationCoordinate": 97363, + "ĠCLOCK": 84043, + "ĠCLOSE": 49706, + "ĠCLOSED": 97313, + "ĠCLR": 60690, + "ĠCLUB": 92926, + "ĠCM": 18090, + "ĠCMD": 26913, + "ĠCMP": 72601, + "ĠCMS": 36241, + "ĠCN": 24872, + "ĠCNBC": 72259, + "ĠCNC": 59697, + "ĠCNN": 19769, + "ĠCNS": 92543, + "ĠCO": 7284, + "ĠCOD": 67401, + "ĠCODE": 13765, + "ĠCOL": 25578, + "ĠCOLL": 62508, + "ĠCOLLECTION": 94076, + "ĠCOLOR": 25419, + "ĠCOLORS": 86822, + "ĠCOLUMN": 39023, + "ĠCOM": 7682, + "ĠCOMM": 51285, + "ĠCOMMAND": 40218, + "ĠCOMMENT": 50505, + "ĠCOMMENTS": 94742, + "ĠCOMMIT": 83230, + "ĠCOMMON": 66677, + "ĠCOMP": 21968, + "ĠCOMPANY": 65448, + "ĠCOMPLETE": 77031, + "ĠCOMPONENT": 70303, + "ĠCOMPUT": 65798, + "ĠCOMPUTER": 88801, + "ĠCON": 3418, + "ĠCONCAT": 78143, + "ĠCONDITION": 10985, + "ĠCONDITIONS": 11342, + "ĠCONF": 45189, + "ĠCONFIG": 13202, + "ĠCONNECT": 63022, + "ĠCONNECTION": 24006, + "ĠCONS": 72906, + "ĠCONSEQUENTIAL": 27415, + "ĠCONST": 29602, + "ĠCONSTANT": 99051, + "ĠCONSTANTS": 93233, + "ĠCONSTRAINT": 93763, + "ĠCONT": 16120, + "ĠCONTACT": 62487, + "ĠCONTENT": 35768, + "ĠCONTEXT": 87336, + "ĠCONTR": 11570, + "ĠCONTRACT": 17126, + "ĠCONTRIBUT": 21076, + "ĠCONTRIBUTORS": 21740, + "ĠCONTROL": 53179, + "ĠCONVERT": 67287, + "ĠCOOKIE": 41565, + "ĠCOP": 67572, + "ĠCOPY": 12249, + "ĠCOPYING": 43263, + "ĠCOPYRIGHT": 14537, + "ĠCOR": 26465, + "ĠCORE": 50183, + "ĠCORPOR": 79383, + "ĠCORPORATION": 88740, + "ĠCORS": 90666, + "ĠCOS": 72944, + "ĠCOST": 63712, + "ĠCOUNT": 23989, + "ĠCOUNTY": 76076, + "ĠCOUR": 53448, + "ĠCOURT": 95700, + "ĠCOVER": 95039, + "ĠCOVID": 19966, + "ĠCP": 15284, + "ĠCPA": 86277, + "ĠCPC": 95146, + "ĠCPF": 45450, + "ĠCPI": 86539, + "ĠCPL": 89480, + "ĠCPP": 37671, + "ĠCPPUNIT": 51526, + "ĠCPR": 76798, + "ĠCPS": 71784, + "ĠCPU": 13940, + "ĠCPUs": 65475, + "ĠCR": 12617, + "ĠCRA": 93667, + "ĠCRC": 29395, + "ĠCRE": 49608, + "ĠCREATE": 30776, + "ĠCREATED": 92794, + "ĠCRM": 40341, + "ĠCROSS": 97773, + "ĠCRS": 94242, + "ĠCRT": 95034, + "ĠCRUD": 64133, + "ĠCS": 10006, + "ĠCSA": 93407, + "ĠCSC": 79676, + "ĠCSI": 87064, + "ĠCSL": 71656, + "ĠCSP": 90510, + "ĠCSR": 62976, + "ĠCSRF": 78595, + "ĠCSS": 15179, + "ĠCST": 58939, + "ĠCSV": 27445, + "ĠCString": 56956, + "ĠCT": 18572, + "ĠCTL": 86886, + "ĠCTRL": 52486, + "ĠCType": 18623, + "ĠCU": 40743, + "ĠCUDA": 54809, + "ĠCUR": 18548, + "ĠCURL": 70669, + "ĠCURLOPT": 22244, + "ĠCURRENT": 43107, + "ĠCUSTOM": 43299, + "ĠCUT": 72857, + "ĠCV": 14165, + "ĠCVE": 45769, + "ĠCVS": 77365, + "ĠCW": 40035, + "ĠCWE": 50860, + "ĠCX": 48483, + "ĠCY": 29569, + "ĠCZ": 69775, + "ĠCa": 14413, + "ĠCab": 26108, + "ĠCabin": 24764, + "ĠCabinet": 32946, + "ĠCabinets": 77389, + "ĠCable": 33485, + "ĠCabr": 98794, + "ĠCache": 19479, + "ĠCached": 74205, + "ĠCad": 32060, + "ĠCadillac": 77495, + "ĠCaesar": 53653, + "ĠCaf": 52641, + "ĠCafe": 42773, + "ĠCafé": 65671, + "ĠCage": 57889, + "ĠCah": 95010, + "ĠCain": 66088, + "ĠCair": 94841, + "ĠCairo": 52550, + "ĠCait": 75332, + "ĠCaj": 95779, + "ĠCake": 32760, + "ĠCal": 3319, + "ĠCalc": 34215, + "ĠCalcium": 95004, + "ĠCalcul": 31359, + "ĠCalculate": 20517, + "ĠCalculates": 59853, + "ĠCalculation": 74216, + "ĠCalculator": 36028, + "ĠCald": 71795, + "ĠCalder": 75300, + "ĠCaldwell": 81224, + "ĠCaleb": 83062, + "ĠCalendar": 16841, + "ĠCalgary": 37875, + "ĠCalibration": 96295, + "ĠCalif": 40789, + "ĠCaliforn": 78769, + "ĠCalifornia": 7043, + "ĠCall": 7143, + "ĠCallable": 53123, + "ĠCallback": 22668, + "ĠCalled": 21154, + "ĠCaller": 56068, + "ĠCalling": 32291, + "ĠCallingConvention": 58413, + "ĠCalls": 40127, + "ĠCalories": 82937, + "ĠCalvin": 50883, + "ĠCam": 8055, + "ĠCamb": 33796, + "ĠCambodia": 61038, + "ĠCambridge": 23641, + "ĠCamden": 82468, + "ĠCame": 75178, + "ĠCamel": 68154, + "ĠCamera": 14332, + "ĠCameras": 88770, + "ĠCameron": 26426, + "ĠCameroon": 94008, + "ĠCamp": 9512, + "ĠCampaign": 26545, + "ĠCampbell": 29424, + "ĠCamping": 72041, + "ĠCampo": 58153, + "ĠCampos": 88465, + "ĠCampus": 38580, + "ĠCan": 2980, + "ĠCanada": 6864, + "ĠCanadian": 11888, + "ĠCanadians": 35162, + "ĠCanadiens": 87593, + "ĠCanal": 52648, + "ĠCanary": 98956, + "ĠCanberra": 68790, + "ĠCanc": 80323, + "ĠCancel": 23542, + "ĠCancellationToken": 37237, + "ĠCancer": 25153, + "ĠCand": 69554, + "ĠCandid": 93816, + "ĠCandidate": 49583, + "ĠCandidates": 64438, + "ĠCandle": 72897, + "ĠCandy": 48428, + "ĠCann": 29468, + "ĠCannabis": 48565, + "ĠCannes": 83520, + "ĠCannon": 50723, + "ĠCannot": 34655, + "ĠCanon": 38116, + "ĠCanonical": 95177, + "ĠCant": 41831, + "ĠCanter": 74893, + "ĠCanterbury": 83500, + "ĠCanton": 72366, + "ĠCanucks": 80502, + "ĠCanvas": 25698, + "ĠCanyon": 40460, + "ĠCap": 8012, + "ĠCapability": 98551, + "ĠCapac": 97103, + "ĠCapacity": 40110, + "ĠCapcom": 93553, + "ĠCape": 28615, + "ĠCapital": 18374, + "ĠCapitals": 94284, + "ĠCapitol": 31533, + "ĠCaps": 43365, + "ĠCapt": 16603, + "ĠCaptain": 21316, + "ĠCaption": 37600, + "ĠCapture": 39885, + "ĠCar": 3261, + "ĠCara": 78913, + "ĠCarb": 68477, + "ĠCarbon": 21487, + "ĠCard": 6795, + "ĠCardiff": 60592, + "ĠCardinal": 45914, + "ĠCardinals": 43327, + "ĠCardio": 91876, + "ĠCards": 26766, + "ĠCare": 10627, + "ĠCareer": 40476, + "ĠCareers": 98355, + "ĠCarey": 65218, + "ĠCargo": 61288, + "ĠCaribbean": 34274, + "ĠCarl": 21998, + "ĠCarla": 93770, + "ĠCarlo": 57770, + "ĠCarlos": 29297, + "ĠCarlson": 69973, + "ĠCarlton": 89878, + "ĠCarly": 78091, + "ĠCarm": 34452, + "ĠCarmen": 69958, + "ĠCarn": 31649, + "ĠCarnegie": 63273, + "ĠCarnival": 77744, + "ĠCarol": 10252, + "ĠCarolina": 12740, + "ĠCaroline": 53391, + "ĠCarolyn": 82258, + "ĠCarousel": 66202, + "ĠCarp": 48777, + "ĠCarpenter": 67038, + "ĠCarpet": 71208, + "ĠCarr": 29374, + "ĠCarrie": 64451, + "ĠCarrier": 50861, + "ĠCarroll": 46109, + "ĠCarry": 62742, + "ĠCars": 35131, + "ĠCarson": 40176, + "ĠCart": 13115, + "ĠCarter": 24563, + "ĠCartesian": 80715, + "ĠCartoon": 68430, + "ĠCary": 91564, + "ĠCas": 11054, + "ĠCasa": 50071, + "ĠCasc": 95006, + "ĠCascade": 56770, + "ĠCascadeType": 90290, + "ĠCase": 11538, + "ĠCases": 46024, + "ĠCasey": 44524, + "ĠCash": 22338, + "ĠCasino": 20481, + "ĠCasinos": 76904, + "ĠCass": 25442, + "ĠCassandra": 81242, + "ĠCassidy": 91797, + "ĠCast": 11264, + "ĠCastillo": 95683, + "ĠCasting": 65748, + "ĠCastle": 26889, + "ĠCastro": 46524, + "ĠCasual": 48822, + "ĠCat": 17358, + "ĠCata": 88472, + "ĠCatal": 31444, + "ĠCatalan": 80844, + "ĠCatalog": 31221, + "ĠCatalonia": 95852, + "ĠCatalyst": 87857, + "ĠCatch": 20930, + "ĠCategoria": 72582, + "ĠCategories": 28212, + "ĠCategory": 10054, + "ĠCater": 60136, + "ĠCath": 12234, + "ĠCathedral": 56729, + "ĠCatherine": 41563, + "ĠCatholic": 16473, + "ĠCatholics": 50871, + "ĠCathy": 81348, + "ĠCats": 50749, + "ĠCaucas": 59432, + "ĠCaucasian": 84419, + "ĠCaucus": 79017, + "ĠCaught": 92314, + "ĠCaul": 89438, + "ĠCause": 44346, + "ĠCauses": 73005, + "ĠCav": 51066, + "ĠCaval": 49130, + "ĠCavaliers": 71856, + "ĠCave": 49392, + "ĠCavs": 95635, + "ĠCay": 52783, + "ĠCbd": 49350, + "ĠCd": 83990, + "ĠCe": 23769, + "ĠCec": 49145, + "ĠCecil": 89127, + "ĠCed": 78513, + "ĠCedar": 56648, + "ĠCeiling": 64425, + "ĠCel": 46543, + "ĠCele": 22687, + "ĠCelebr": 32192, + "ĠCelebration": 74100, + "ĠCelebrity": 64338, + "ĠCell": 13972, + "ĠCells": 58090, + "ĠCellular": 87107, + "ĠCelsius": 61347, + "ĠCelt": 35307, + "ĠCeltic": 54664, + "ĠCeltics": 57696, + "ĠCement": 61332, + "ĠCemetery": 58478, + "ĠCena": 88563, + "ĠCensus": 45527, + "ĠCent": 5719, + "ĠCentOS": 95513, + "ĠCenter": 5832, + "ĠCenters": 40891, + "ĠCentral": 10684, + "ĠCentre": 14481, + "ĠCentro": 56574, + "ĠCentury": 27101, + "ĠCer": 27664, + "ĠCeramic": 78250, + "ĠCeremony": 94652, + "ĠCert": 15304, + "ĠCertain": 34111, + "ĠCertainly": 52366, + "ĠCertificate": 31402, + "ĠCertification": 50210, + "ĠCertified": 35442, + "ĠCes": 61363, + "ĠCette": 61308, + "ĠCf": 89894, + "ĠCh": 910, + "ĠCha": 27721, + "ĠChad": 42030, + "ĠChain": 28525, + "ĠChains": 81769, + "ĠChair": 16090, + "ĠChairman": 26294, + "ĠChairs": 68124, + "ĠChall": 69385, + "ĠChallenge": 25257, + "ĠChallenger": 86483, + "ĠChallenges": 68678, + "ĠCham": 41783, + "ĠChamber": 31379, + "ĠChambers": 64176, + "ĠChamp": 55590, + "ĠChampagne": 87539, + "ĠChampion": 32497, + "ĠChampions": 28808, + "ĠChampionship": 18621, + "ĠChampionships": 47754, + "ĠChan": 41302, + "ĠChance": 38934, + "ĠChancellor": 46963, + "ĠChand": 66636, + "ĠChandler": 60348, + "ĠChanel": 90372, + "ĠChang": 43372, + "ĠChange": 10388, + "ĠChanged": 46294, + "ĠChanges": 28140, + "ĠChanging": 50169, + "ĠChannel": 13434, + "ĠChannels": 62900, + "ĠChaos": 43105, + "ĠChap": 26649, + "ĠChapel": 53362, + "ĠChapman": 59814, + "ĠChapter": 15585, + "ĠChapters": 68381, + "ĠChar": 4864, + "ĠCharSequence": 57457, + "ĠCharSet": 66991, + "ĠCharacter": 15633, + "ĠCharacteristics": 83984, + "ĠCharacters": 44516, + "ĠCharg": 27617, + "ĠCharge": 36795, + "ĠCharger": 72408, + "ĠChargers": 59593, + "ĠCharges": 77969, + "ĠCharging": 89798, + "ĠCharity": 65931, + "ĠCharl": 45366, + "ĠCharles": 14927, + "ĠCharleston": 52293, + "ĠCharlie": 24927, + "ĠCharlotte": 28373, + "ĠCharlottesville": 63638, + "ĠCharm": 57500, + "ĠCharset": 87871, + "ĠChart": 21266, + "ĠCharter": 48605, + "ĠCharts": 66276, + "ĠChase": 34574, + "ĠChat": 12853, + "ĠChatColor": 79320, + "ĠChatt": 78213, + "ĠChattanooga": 96843, + "ĠChavez": 76241, + "ĠChe": 8436, + "ĠCheap": 35934, + "ĠCheat": 98023, + "ĠCheck": 4248, + "ĠCheckBox": 45365, + "ĠCheckbox": 60041, + "ĠChecked": 93361, + "ĠChecker": 80678, + "ĠChecking": 46093, + "ĠChecklist": 94969, + "ĠCheckout": 56735, + "ĠChecks": 24843, + "ĠCheer": 92631, + "ĠCheers": 78989, + "ĠCheese": 44319, + "ĠChef": 35975, + "ĠChelsea": 26518, + "ĠChem": 18994, + "ĠChemical": 35324, + "ĠChemistry": 41746, + "ĠChen": 24495, + "ĠCheney": 80246, + "ĠCheng": 56707, + "ĠChennai": 63821, + "ĠCher": 55107, + "ĠChern": 90659, + "ĠCherokee": 70041, + "ĠCherry": 44705, + "ĠCheryl": 79837, + "ĠChes": 73654, + "ĠChess": 47387, + "ĠChest": 41771, + "ĠChester": 52412, + "ĠChevrolet": 42377, + "ĠChevron": 89993, + "ĠChevy": 57298, + "ĠChew": 89865, + "ĠChi": 33282, + "ĠChic": 70720, + "ĠChicago": 10557, + "ĠChick": 62126, + "ĠChicken": 33519, + "ĠChief": 14344, + "ĠChiefs": 44744, + "ĠChild": 9391, + "ĠChildhood": 71310, + "ĠChildren": 15044, + "ĠChile": 33000, + "ĠChili": 81415, + "ĠChill": 86489, + "ĠChim": 81448, + "ĠChin": 48235, + "ĠChina": 5616, + "ĠChinese": 8453, + "ĠChip": 30913, + "ĠChips": 86668, + "ĠChiến": 142226, + "ĠChloe": 59370, + "ĠChlor": 91379, + "ĠCho": 32580, + "ĠChocolate": 38420, + "ĠChoi": 86573, + "ĠChoice": 27107, + "ĠChoices": 78191, + "ĠChoir": 98332, + "ĠChoose": 22201, + "ĠChoosing": 76251, + "ĠChop": 63997, + "ĠChore": 95538, + "ĠChow": 73958, + "ĠChr": 52207, + "ĠChris": 11266, + "ĠChrist": 3686, + "ĠChristian": 8876, + "ĠChristianity": 31086, + "ĠChristians": 21410, + "ĠChristie": 43306, + "ĠChristina": 59671, + "ĠChristine": 45893, + "ĠChristmas": 10074, + "ĠChristoph": 94593, + "ĠChristopher": 24986, + "ĠChrom": 34218, + "ĠChrome": 17225, + "ĠChromium": 57165, + "ĠChron": 25481, + "ĠChronic": 72682, + "ĠChronicle": 41059, + "ĠChronicles": 65316, + "ĠChrysler": 63935, + "ĠChu": 61011, + "ĠChuck": 33249, + "ĠChun": 89354, + "ĠChung": 87325, + "ĠChunk": 45513, + "ĠChurch": 9257, + "ĠChurches": 93541, + "ĠChurchill": 57174, + "ĠChâu": 132690, + "ĠChúa": 132756, + "ĠChúng": 131272, + "ĠChÃŃ": 131868, + "ĠChÃŃnh": 129591, + "ĠChương": 131786, + "ĠChá»§": 130085, + "ĠChá»ī": 132026, + "ĠChá»ĭ": 141654, + "ĠCi": 31644, + "ĠCialis": 32302, + "ĠCic": 73951, + "ĠCin": 28911, + "ĠCincinnati": 35139, + "ĠCinder": 77330, + "ĠCindy": 69331, + "ĠCinema": 48218, + "ĠCipher": 63288, + "ĠCir": 40909, + "ĠCirc": 16338, + "ĠCircle": 21224, + "ĠCircuit": 27218, + "ĠCircular": 45761, + "ĠCircularProgress": 78197, + "ĠCircularProgressIndicator": 98479, + "ĠCircus": 86821, + "ĠCisco": 42134, + "ĠCit": 17543, + "ĠCitadel": 96226, + "ĠCitation": 69323, + "ĠCities": 37273, + "ĠCitizen": 46217, + "ĠCitizens": 39896, + "ĠCitizenship": 93736, + "ĠCity": 4311, + "ĠCiudad": 74145, + "ĠCiv": 79135, + "ĠCivic": 54829, + "ĠCivil": 16398, + "ĠCivilization": 90678, + "ĠCl": 2435, + "ĠCla": 56338, + "ĠClaim": 37502, + "ĠClaims": 50098, + "ĠClair": 92510, + "ĠClaire": 42575, + "ĠClamp": 92818, + "ĠClan": 54595, + "ĠClar": 30081, + "ĠClara": 50557, + "ĠClare": 61040, + "ĠClarence": 87730, + "ĠClark": 21304, + "ĠClarke": 42555, + "ĠClarkson": 93194, + "ĠClash": 84570, + "ĠClass": 3228, + "ĠClassName": 98457, + "ĠClassNotFoundException": 65641, + "ĠClasse": 91887, + "ĠClasses": 33869, + "ĠClassic": 21837, + "ĠClassical": 67433, + "ĠClassics": 72315, + "ĠClassification": 54882, + "ĠClassified": 79436, + "ĠClassifier": 89186, + "ĠClassroom": 76029, + "ĠClaud": 61312, + "ĠClaude": 74330, + "ĠClaudia": 80056, + "ĠClaus": 67019, + "ĠClause": 55797, + "ĠClaw": 87010, + "ĠClay": 31401, + "ĠClayton": 63444, + "ĠCle": 36002, + "ĠClean": 9590, + "ĠCleaner": 75219, + "ĠCleaning": 46655, + "ĠCleans": 85662, + "ĠCleanup": 53512, + "ĠClear": 12023, + "ĠClearance": 92130, + "ĠClearly": 53404, + "ĠClears": 78071, + "ĠClem": 45168, + "ĠClement": 52127, + "ĠClemson": 62031, + "ĠCler": 40776, + "ĠClerk": 62140, + "ĠCleveland": 23475, + "ĠCli": 96971, + "ĠClick": 9189, + "ĠClient": 8423, + "ĠCliente": 40831, + "ĠClients": 47408, + "ĠCliff": 47852, + "ĠClifford": 85683, + "ĠClim": 61830, + "ĠClimate": 30536, + "ĠClin": 18398, + "ĠClinic": 39224, + "ĠClinical": 32035, + "ĠClint": 55029, + "ĠClinton": 8120, + "ĠClintons": 92030, + "ĠClip": 29692, + "ĠClipboard": 81327, + "ĠClippers": 71012, + "ĠClips": 83373, + "ĠClo": 18679, + "ĠClock": 26142, + "ĠClone": 27913, + "ĠClose": 13032, + "ĠClosed": 44679, + "ĠCloset": 99140, + "ĠClosing": 61113, + "ĠClosure": 55567, + "ĠCloth": 83635, + "ĠClothes": 76524, + "ĠClothing": 53658, + "ĠCloud": 14817, + "ĠClover": 94801, + "ĠClown": 98323, + "ĠClub": 10140, + "ĠClubs": 69032, + "ĠCluster": 35380, + "ĠCly": 68288, + "ĠClyde": 90394, + "ĠCmd": 40210, + "ĠCo": 3539, + "ĠCoach": 27176, + "ĠCoaching": 74378, + "ĠCoal": 24813, + "ĠCoalition": 35792, + "ĠCoast": 15994, + "ĠCoastal": 71558, + "ĠCoat": 67767, + "ĠCob": 45951, + "ĠCobb": 67905, + "ĠCobra": 84475, + "ĠCoc": 76388, + "ĠCoca": 62854, + "ĠCoch": 94395, + "ĠCock": 33927, + "ĠCocktail": 92716, + "ĠCoco": 83382, + "ĠCocoa": 75651, + "ĠCoconut": 71279, + "ĠCod": 33677, + "ĠCodable": 80482, + "ĠCode": 6119, + "ĠCodec": 67077, + "ĠCodes": 40561, + "ĠCoding": 48075, + "ĠCody": 57282, + "ĠCoff": 75133, + "ĠCoffee": 26079, + "ĠCognitive": 72135, + "ĠCoh": 83575, + "ĠCohen": 31096, + "ĠCoil": 67665, + "ĠCoin": 26233, + "ĠCoinbase": 79808, + "ĠCoins": 61776, + "ĠCoke": 73262, + "ĠCol": 4254, + "ĠColbert": 78647, + "ĠCold": 23185, + "ĠCole": 23407, + "ĠColeman": 49432, + "ĠColin": 39879, + "ĠColl": 12905, + "ĠCollabor": 48581, + "ĠCollaboration": 86587, + "ĠCollapse": 75818, + "ĠCollect": 20513, + "ĠCollection": 11101, + "ĠCollections": 25326, + "ĠCollective": 58918, + "ĠCollector": 58548, + "ĠColleg": 56565, + "ĠCollege": 9126, + "ĠColleges": 78119, + "ĠCollider": 73821, + "ĠCollins": 28670, + "ĠCollision": 51548, + "ĠColo": 95075, + "ĠCologne": 78058, + "ĠColomb": 30802, + "ĠColombia": 38033, + "ĠColombian": 83553, + "ĠColon": 30942, + "ĠColonel": 51698, + "ĠColonial": 72748, + "ĠColony": 81846, + "ĠColor": 3478, + "ĠColorado": 15383, + "ĠColoring": 52794, + "ĠColors": 9526, + "ĠColour": 40862, + "ĠColt": 90384, + "ĠColts": 51984, + "ĠColum": 13901, + "ĠColumbia": 18796, + "ĠColumbus": 31817, + "ĠColumn": 9332, + "ĠColumns": 53573, + "ĠCom": 1198, + "ĠComVisible": 54315, + "ĠComb": 22454, + "ĠCombat": 35982, + "ĠCombination": 85856, + "ĠCombine": 46812, + "ĠCombined": 57652, + "ĠCombo": 53023, + "ĠComboBox": 54153, + "ĠComcast": 57004, + "ĠCome": 15565, + "ĠComedy": 43751, + "ĠComes": 53228, + "ĠComet": 98098, + "ĠComey": 34978, + "ĠComfort": 43742, + "ĠComic": 39963, + "ĠComics": 34056, + "ĠComing": 39891, + "ĠComm": 6804, + "ĠCommand": 7348, + "ĠCommandLine": 78218, + "ĠCommandType": 36927, + "ĠCommander": 34138, + "ĠCommands": 47183, + "ĠComment": 12255, + "ĠCommentary": 72444, + "ĠComments": 17681, + "ĠCommerce": 30380, + "ĠCommercial": 27843, + "ĠCommission": 9652, + "ĠCommissioner": 29354, + "ĠCommissioners": 96245, + "ĠCommit": 9205, + "ĠCommittee": 10341, + "ĠCommittees": 67917, + "ĠCommod": 78180, + "ĠCommon": 7718, + "ĠCommonModule": 37638, + "ĠCommons": 25585, + "ĠCommonwealth": 37198, + "ĠCommun": 56198, + "ĠCommunic": 16433, + "ĠCommunication": 30866, + "ĠCommunications": 25466, + "ĠCommunist": 36861, + "ĠCommunities": 57343, + "ĠCommunity": 12062, + "ĠComo": 45674, + "ĠComp": 13317, + "ĠCompact": 54782, + "ĠCompanies": 31786, + "ĠCompanion": 68975, + "ĠCompany": 8188, + "ĠCompar": 22771, + "ĠComparable": 58106, + "ĠComparative": 96180, + "ĠComparator": 50541, + "ĠCompare": 23768, + "ĠCompared": 58713, + "ĠComparison": 42451, + "ĠCompass": 59580, + "ĠCompatibility": 77995, + "ĠCompatible": 66265, + "ĠCompensation": 69296, + "ĠCompet": 25440, + "ĠCompetition": 41221, + "ĠCompetitive": 78050, + "ĠCompilation": 70081, + "ĠCompile": 54467, + "ĠCompiled": 96471, + "ĠCompiler": 45631, + "ĠComplaint": 67438, + "ĠComple": 42354, + "ĠCompletableFuture": 91535, + "ĠComplete": 18608, + "ĠCompleted": 45494, + "ĠCompletely": 85114, + "ĠCompletion": 56250, + "ĠComplex": 22096, + "ĠComplexity": 70618, + "ĠCompliance": 59170, + "ĠComponent": 5578, + "ĠComponentFixture": 28363, + "ĠComponents": 34085, + "ĠComposer": 78771, + "ĠComposite": 40639, + "ĠComposition": 67106, + "ĠCompound": 61572, + "ĠComprehensive": 66863, + "ĠCompression": 66161, + "ĠComput": 23144, + "ĠComputational": 91928, + "ĠCompute": 22603, + "ĠComputer": 17407, + "ĠComputers": 68449, + "ĠComputes": 64975, + "ĠComputing": 45779, + "ĠCon": 1200, + "ĠConan": 78812, + "ĠConc": 62698, + "ĠConcat": 78195, + "ĠConce": 78618, + "ĠConcent": 61161, + "ĠConcept": 34355, + "ĠConcepts": 75772, + "ĠConcern": 51247, + "ĠConcert": 46684, + "ĠConclusion": 73877, + "ĠConcord": 63093, + "ĠConcrete": 36532, + "ĠConcurrent": 42704, + "ĠConcurrentHashMap": 89075, + "ĠCond": 44826, + "ĠCondition": 15180, + "ĠConditional": 91973, + "ĠConditioning": 82392, + "ĠConditions": 31834, + "ĠConduct": 49835, + "ĠCone": 46412, + "ĠConexion": 72468, + "ĠConf": 14974, + "ĠConfeder": 38044, + "ĠConfederate": 56357, + "ĠConference": 14872, + "ĠConfidence": 86324, + "ĠConfidential": 73365, + "ĠConfig": 5532, + "ĠConfiguration": 12221, + "ĠConfigurationManager": 55831, + "ĠConfigure": 22169, + "ĠConfigureServices": 96608, + "ĠConfirm": 33563, + "ĠConfirmation": 68309, + "ĠConflict": 58585, + "ĠCong": 7261, + "ĠCongo": 53832, + "ĠCongratulations": 70541, + "ĠCongress": 7992, + "ĠCongressional": 38373, + "ĠCongressman": 70383, + "ĠConj": 97157, + "ĠConn": 18213, + "ĠConnect": 13015, + "ĠConnected": 49486, + "ĠConnecticut": 30361, + "ĠConnecting": 83938, + "ĠConnection": 11032, + "ĠConnectionState": 71693, + "ĠConnections": 65952, + "ĠConnectivity": 96054, + "ĠConnector": 54814, + "ĠConnie": 96063, + "ĠConnor": 57180, + "ĠConor": 84115, + "ĠConrad": 75989, + "ĠCons": 7292, + "ĠConscious": 91652, + "ĠConse": 66466, + "ĠConsent": 92387, + "ĠConsequently": 52023, + "ĠConserv": 20049, + "ĠConservation": 44335, + "ĠConservative": 28971, + "ĠConservatives": 48244, + "ĠConsider": 21144, + "ĠConsidering": 55777, + "ĠConsole": 5277, + "ĠConsoleColor": 58834, + "ĠConsolid": 78880, + "ĠConsort": 67603, + "ĠConsortium": 72447, + "ĠConspiracy": 97152, + "ĠConst": 24522, + "ĠConstant": 19305, + "ĠConstantin": 91187, + "ĠConstants": 16605, + "ĠConstit": 75604, + "ĠConstitution": 17579, + "ĠConstitutional": 62185, + "ĠConstr": 95584, + "ĠConstraint": 44264, + "ĠConstraints": 87473, + "ĠConstruct": 18678, + "ĠConstruction": 24026, + "ĠConstructor": 16786, + "ĠConstructors": 51768, + "ĠConstructs": 47213, + "ĠConsult": 19960, + "ĠConsultant": 55446, + "ĠConsultants": 90050, + "ĠConsulting": 46411, + "ĠConsum": 40962, + "ĠConsumer": 25197, + "ĠConsumers": 81991, + "ĠConsumption": 86166, + "ĠCont": 2093, + "ĠContact": 9180, + "ĠContacts": 46039, + "ĠContainer": 9678, + "ĠContainers": 98293, + "ĠContains": 30811, + "ĠConte": 93378, + "ĠContemporary": 47202, + "ĠContent": 8883, + "ĠContentType": 70431, + "ĠContentValues": 55929, + "ĠContentView": 93928, + "ĠContents": 35862, + "ĠContest": 46533, + "ĠContext": 9608, + "ĠContin": 20297, + "ĠContinent": 97745, + "ĠContinental": 60734, + "ĠContinue": 15003, + "ĠContinued": 50621, + "ĠContinuing": 81017, + "ĠContinuous": 68967, + "ĠContr": 34241, + "ĠContract": 19185, + "ĠContractor": 86518, + "ĠContractors": 97793, + "ĠContracts": 70559, + "ĠContrast": 90048, + "ĠContrib": 23177, + "ĠContribution": 93528, + "ĠContributions": 96004, + "ĠContributor": 64724, + "ĠContributors": 39536, + "ĠControl": 7779, + "ĠControlled": 81345, + "ĠController": 9771, + "ĠControllerBase": 79024, + "ĠControllers": 85476, + "ĠControls": 32070, + "ĠConv": 28988, + "ĠConvenience": 80648, + "ĠConvenient": 92950, + "ĠConvention": 25872, + "ĠConvers": 55396, + "ĠConversation": 50830, + "ĠConversely": 81571, + "ĠConversion": 42509, + "ĠConvert": 7169, + "ĠConverted": 93693, + "ĠConverter": 39328, + "ĠConverts": 41729, + "ĠConveyor": 99155, + "ĠConway": 59474, + "ĠCook": 12514, + "ĠCookbook": 88203, + "ĠCooke": 98739, + "ĠCookie": 24356, + "ĠCookies": 25994, + "ĠCooking": 56310, + "ĠCool": 23931, + "ĠCooler": 94143, + "ĠCooling": 82305, + "ĠCooper": 23517, + "ĠCooperation": 72708, + "ĠCooperative": 85705, + "ĠCoord": 47849, + "ĠCoordinate": 41035, + "ĠCoordinates": 62501, + "ĠCoordinator": 42324, + "ĠCop": 24946, + "ĠCopa": 84531, + "ĠCopenhagen": 63061, + "ĠCopies": 79568, + "ĠCopp": 80999, + "ĠCopper": 42540, + "ĠCopy": 14540, + "ĠCopyright": 2955, + "ĠCor": 4463, + "ĠCoral": 63816, + "ĠCorbyn": 41656, + "ĠCord": 36887, + "ĠCore": 9518, + "ĠCoreData": 91122, + "ĠCorey": 54652, + "ĠCorinth": 59793, + "ĠCorinthians": 72974, + "ĠCork": 51936, + "ĠCorm": 96189, + "ĠCorn": 21330, + "ĠCornel": 97945, + "ĠCornell": 55719, + "ĠCorner": 38248, + "ĠCornwall": 75950, + "ĠCoron": 47083, + "ĠCorona": 60229, + "ĠCoronavirus": 69820, + "ĠCoroutine": 87466, + "ĠCorp": 21863, + "ĠCorpor": 10858, + "ĠCorporate": 35465, + "ĠCorporation": 13034, + "ĠCorps": 30142, + "ĠCorpus": 75734, + "ĠCorrect": 39970, + "ĠCorrection": 63069, + "ĠCorrections": 90472, + "ĠCorrespond": 50942, + "ĠCorruption": 89794, + "ĠCors": 52518, + "ĠCort": 35675, + "ĠCortex": 89902, + "ĠCorvette": 85825, + "ĠCory": 63528, + "ĠCos": 18114, + "ĠCosby": 91951, + "ĠCosmetic": 76321, + "ĠCosmic": 94689, + "ĠCosmos": 83424, + "ĠCost": 11194, + "ĠCosta": 31280, + "ĠCostco": 80295, + "ĠCosts": 63041, + "ĠCostume": 63689, + "ĠCot": 67725, + "ĠCottage": 75395, + "ĠCotton": 37696, + "ĠCou": 18232, + "ĠCouch": 61128, + "ĠCougar": 89757, + "ĠCoul": 66848, + "ĠCould": 16503, + "ĠCouldn": 72688, + "ĠCoun": 31254, + "ĠCouncil": 9074, + "ĠCounsel": 36645, + "ĠCounseling": 88449, + "ĠCount": 4504, + "ĠCountdown": 86386, + "ĠCounter": 19735, + "ĠCounties": 92888, + "ĠCountries": 53737, + "ĠCountry": 14106, + "ĠCounts": 85659, + "ĠCounty": 6272, + "ĠCoupe": 80420, + "ĠCouple": 65825, + "ĠCouples": 89150, + "ĠCoupon": 44930, + "ĠCoupons": 67067, + "ĠCour": 6244, + "ĠCourage": 98530, + "ĠCourier": 74593, + "ĠCourse": 16615, + "ĠCourses": 47678, + "ĠCourt": 7154, + "ĠCourtesy": 53097, + "ĠCourtney": 70682, + "ĠCourts": 65463, + "ĠCous": 60469, + "ĠCousins": 77604, + "ĠCout": 71253, + "ĠCov": 62636, + "ĠCove": 64379, + "ĠCovenant": 62941, + "ĠCovent": 88717, + "ĠCover": 17757, + "ĠCoverage": 55669, + "ĠCovered": 72195, + "ĠCovers": 70425, + "ĠCovid": 36974, + "ĠCow": 21851, + "ĠCowboy": 95022, + "ĠCowboys": 40305, + "ĠCox": 38660, + "ĠCoy": 61973, + "ĠCp": 64256, + "ĠCr": 4553, + "ĠCra": 23243, + "ĠCrab": 74104, + "ĠCrack": 54945, + "ĠCraft": 24009, + "ĠCrafting": 97169, + "ĠCrafts": 64941, + "ĠCraig": 28417, + "ĠCraigslist": 72292, + "ĠCran": 66430, + "ĠCrane": 72899, + "ĠCrash": 50335, + "ĠCrate": 68390, + "ĠCraw": 43666, + "ĠCrawford": 53176, + "ĠCrazy": 50263, + "ĠCre": 7792, + "ĠCream": 29700, + "ĠCreat": 17667, + "ĠCreate": 4230, + "ĠCreateMap": 89924, + "ĠCreateTable": 88604, + "ĠCreateUser": 95021, + "ĠCreated": 4290, + "ĠCreates": 14684, + "ĠCreating": 31306, + "ĠCreation": 34286, + "ĠCreative": 24263, + "ĠCreator": 35678, + "ĠCreature": 52514, + "ĠCreatures": 97412, + "ĠCredential": 93379, + "ĠCredentials": 61260, + "ĠCredit": 16267, + "ĠCredits": 48700, + "ĠCree": 95788, + "ĠCreed": 60377, + "ĠCreek": 23198, + "ĠCres": 60910, + "ĠCrescent": 76469, + "ĠCrest": 70441, + "ĠCrew": 34937, + "ĠCri": 91140, + "ĠCricket": 57919, + "ĠCrime": 23900, + "ĠCrimea": 60711, + "ĠCrimes": 76908, + "ĠCriminal": 32736, + "ĠCrimson": 66077, + "ĠCrisis": 45150, + "ĠCrist": 34118, + "ĠCristiano": 99115, + "ĠCristina": 99114, + "ĠCrit": 33207, + "ĠCriteria": 14243, + "ĠCriterion": 68400, + "ĠCritical": 34661, + "ĠCritics": 76413, + "ĠCro": 24592, + "ĠCroat": 44082, + "ĠCroatia": 57211, + "ĠCroatian": 99070, + "ĠCrom": 91607, + "ĠCron": 55817, + "ĠCrop": 67320, + "ĠCros": 60719, + "ĠCrosby": 75736, + "ĠCross": 11261, + "ĠCrossAxisAlignment": 43289, + "ĠCrossRef": 65673, + "ĠCrossing": 64993, + "ĠCrow": 26893, + "ĠCrowd": 97984, + "ĠCrowley": 82008, + "ĠCrown": 28643, + "ĠCru": 14423, + "ĠCruc": 75004, + "ĠCrud": 75078, + "ĠCruise": 46377, + "ĠCruiser": 91899, + "ĠCrunch": 77213, + "ĠCrus": 50103, + "ĠCrush": 12048, + "ĠCrusher": 16200, + "ĠCrushers": 52197, + "ĠCrushing": 40394, + "ĠCruz": 20842, + "ĠCry": 42705, + "ĠCrypt": 37447, + "ĠCrypto": 32886, + "ĠCrystal": 27916, + "ĠCré": 135893, + "ĠCs": 67332, + "ĠCss": 79214, + "ĠCsv": 91671, + "ĠCt": 62279, + "ĠCtrl": 37014, + "ĠCu": 26462, + "ĠCuando": 87914, + "ĠCub": 18030, + "ĠCuba": 31677, + "ĠCuban": 43710, + "ĠCube": 42592, + "ĠCubs": 50043, + "ĠCue": 96894, + "ĠCuisine": 80861, + "ĠCul": 31694, + "ĠCult": 25594, + "ĠCultural": 40233, + "ĠCulture": 20397, + "ĠCultureInfo": 44467, + "ĠCum": 31578, + "ĠCumberland": 91682, + "ĠCummings": 87371, + "ĠCunning": 67745, + "ĠCunningham": 73828, + "ĠCuomo": 67207, + "ĠCup": 10861, + "ĠCupertino": 96735, + "ĠCups": 78106, + "ĠCur": 12886, + "ĠCure": 69719, + "ĠCurl": 78338, + "ĠCurrency": 28453, + "ĠCurrent": 9125, + "ĠCurrently": 24150, + "ĠCurriculum": 74206, + "ĠCurry": 45975, + "ĠCurse": 70627, + "ĠCurso": 92538, + "ĠCursor": 28067, + "ĠCursors": 93166, + "ĠCurt": 30423, + "ĠCurtain": 83557, + "ĠCurtis": 50716, + "ĠCurve": 53677, + "ĠCush": 83659, + "ĠCust": 72222, + "ĠCustom": 8406, + "ĠCustomer": 12277, + "ĠCustomers": 39589, + "ĠCustomize": 46334, + "ĠCustoms": 45671, + "ĠCut": 19649, + "ĠCute": 57539, + "ĠCutter": 89778, + "ĠCutting": 62425, + "ĠCuá»Ļc": 141380, + "ĠCv": 52742, + "ĠCy": 17748, + "ĠCyan": 92143, + "ĠCyber": 33611, + "ĠCyc": 80981, + "ĠCycl": 71445, + "ĠCycle": 41292, + "ĠCycling": 59970, + "ĠCylinder": 86662, + "ĠCynthia": 74571, + "ĠCypress": 79793, + "ĠCyprus": 56684, + "ĠCyr": 47315, + "ĠCyril": 94705, + "ĠCyrus": 70824, + "ĠCzech": 33150, + "ĠCá": 130494, + "ĠCác": 128678, + "ĠCách": 131577, + "ĠCâm": 138955, + "ĠCâmara": 138956, + "ĠCòn": 135466, + "ĠCó": 129016, + "ĠCódigo": 140003, + "ĠCô": 131249, + "ĠCông": 128534, + "ĠCùng": 139602, + "ĠCÅ©ng": 139014, + "ĠCÆ¡": 134695, + "ĠCảnh": 141555, + "ĠCần": 138310, + "ĠCá»Ļng": 138465, + "ĠD": 422, + "ĠDA": 24543, + "ĠDAC": 39196, + "ĠDACA": 79574, + "ĠDAG": 79987, + "ĠDAL": 49366, + "ĠDAM": 12360, + "ĠDAMAGE": 28204, + "ĠDAMAGES": 16501, + "ĠDAN": 83410, + "ĠDAO": 44368, + "ĠDAR": 96127, + "ĠDARK": 88652, + "ĠDAT": 39362, + "ĠDATA": 14112, + "ĠDATABASE": 42859, + "ĠDATE": 28543, + "ĠDAY": 37688, + "ĠDAYS": 82831, + "ĠDB": 5952, + "ĠDBG": 37392, + "ĠDBHelper": 93705, + "ĠDBNull": 54750, + "ĠDC": 10922, + "ĠDCHECK": 46224, + "ĠDD": 30904, + "ĠDDR": 43760, + "ĠDDS": 64972, + "ĠDE": 3385, + "ĠDEA": 74910, + "ĠDEAD": 78069, + "ĠDEAL": 25491, + "ĠDEALINGS": 28470, + "ĠDEBUG": 12657, + "ĠDEC": 42817, + "ĠDECL": 41960, + "ĠDECLARE": 70819, + "ĠDEF": 23865, + "ĠDEFAULT": 11955, + "ĠDEFIN": 60926, + "ĠDEFINE": 66914, + "ĠDEL": 23777, + "ĠDELETE": 17197, + "ĠDEM": 50814, + "ĠDEN": 66186, + "ĠDER": 46777, + "ĠDERP": 71029, + "ĠDES": 12732, + "ĠDESC": 16089, + "ĠDESCRIPTION": 32185, + "ĠDESIGN": 67243, + "ĠDEST": 71990, + "ĠDET": 40243, + "ĠDETAILS": 76996, + "ĠDEV": 50895, + "ĠDEVELO": 70261, + "ĠDEVICE": 44632, + "ĠDF": 43376, + "ĠDFA": 93631, + "ĠDFS": 56383, + "ĠDG": 50275, + "ĠDH": 32787, + "ĠDHCP": 57587, + "ĠDHS": 72603, + "ĠDI": 16536, + "ĠDIAG": 98167, + "ĠDIC": 97760, + "ĠDID": 59119, + "ĠDIE": 71406, + "ĠDIFF": 68549, + "ĠDIG": 55296, + "ĠDIM": 50859, + "ĠDIN": 77259, + "ĠDIR": 39052, + "ĠDIRECT": 20230, + "ĠDIRECTORY": 54140, + "ĠDIS": 11977, + "ĠDISABLE": 56671, + "ĠDISCLAIM": 24717, + "ĠDISCLAIMED": 31201, + "ĠDISCLAIMER": 97097, + "ĠDISCLAIMS": 83306, + "ĠDISP": 84040, + "ĠDISPATCH": 94581, + "ĠDISPLAY": 52264, + "ĠDIST": 36103, + "ĠDISTINCT": 59086, + "ĠDISTRIBUT": 97841, + "ĠDIV": 46260, + "ĠDIY": 31458, + "ĠDJ": 21387, + "ĠDJs": 92527, + "ĠDK": 55240, + "ĠDL": 32975, + "ĠDLC": 42919, + "ĠDLL": 47192, + "ĠDM": 20192, + "ĠDMA": 25450, + "ĠDN": 60656, + "ĠDNA": 15552, + "ĠDNC": 51965, + "ĠDNS": 27598, + "ĠDO": 9319, + "ĠDOC": 60355, + "ĠDOCUMENT": 57013, + "ĠDOE": 92057, + "ĠDOES": 57263, + "ĠDOI": 58570, + "ĠDOJ": 60810, + "ĠDOM": 18051, + "ĠDOMAIN": 84586, + "ĠDON": 44273, + "ĠDONE": 54685, + "ĠDOS": 58480, + "ĠDOT": 48326, + "ĠDOUBLE": 56272, + "ĠDOWN": 27494, + "ĠDOWNLOAD": 42471, + "ĠDP": 31757, + "ĠDPI": 97025, + "ĠDPR": 87901, + "ĠDPS": 58483, + "ĠDR": 14309, + "ĠDRAW": 71652, + "ĠDRIVE": 95787, + "ĠDRIVER": 48729, + "ĠDRM": 60397, + "ĠDROP": 56942, + "ĠDRV": 71332, + "ĠDS": 20578, + "ĠDSL": 45558, + "ĠDSM": 79167, + "ĠDSP": 56593, + "ĠDST": 77100, + "ĠDT": 25014, + "ĠDTO": 80325, + "ĠDU": 47055, + "ĠDUI": 72851, + "ĠDUP": 87524, + "ĠDV": 51244, + "ĠDVD": 18092, + "ĠDVDs": 63336, + "ĠDVR": 93886, + "ĠDW": 37752, + "ĠDWC": 70259, + "ĠDWORD": 20301, + "ĠDX": 30808, + "ĠDXGI": 74500, + "ĠDY": 95858, + "ĠDa": 14235, + "ĠDabei": 96943, + "ĠDad": 32521, + "ĠDaddy": 53446, + "ĠDaemon": 92236, + "ĠDag": 77027, + "ĠDagger": 82183, + "ĠDah": 75291, + "ĠDahl": 86867, + "ĠDai": 79123, + "ĠDaily": 13385, + "ĠDairy": 78022, + "ĠDaisy": 70164, + "ĠDak": 24648, + "ĠDakota": 27872, + "ĠDal": 28110, + "ĠDale": 36728, + "ĠDallas": 18542, + "ĠDalton": 71454, + "ĠDaly": 94435, + "ĠDam": 16354, + "ĠDamage": 20765, + "ĠDamascus": 57023, + "ĠDame": 40698, + "ĠDamen": 93084, + "ĠDamian": 90955, + "ĠDamien": 88912, + "ĠDamn": 82415, + "ĠDamon": 72249, + "ĠDan": 11563, + "ĠDana": 46613, + "ĠDance": 29604, + "ĠDancing": 69529, + "ĠDane": 79872, + "ĠDanger": 47299, + "ĠDangerous": 73032, + "ĠDani": 29539, + "ĠDaniel": 15118, + "ĠDanielle": 71616, + "ĠDaniels": 49535, + "ĠDanish": 43680, + "ĠDank": 70407, + "ĠDann": 72915, + "ĠDanny": 32599, + "ĠDans": 45606, + "ĠDante": 73300, + "ĠDao": 51581, + "ĠDar": 15017, + "ĠDare": 66831, + "ĠDark": 12258, + "ĠDarkness": 53696, + "ĠDarling": 94539, + "ĠDarren": 56990, + "ĠDart": 48097, + "ĠDarth": 80416, + "ĠDarwin": 44380, + "ĠDas": 19000, + "ĠDash": 36670, + "ĠDashboard": 26808, + "ĠDat": 21631, + "ĠData": 2885, + "ĠDataAccess": 68555, + "ĠDataBase": 66930, + "ĠDataColumn": 72515, + "ĠDataContext": 81577, + "ĠDataFrame": 45786, + "ĠDataGridView": 30253, + "ĠDataGridViewCellStyle": 75263, + "ĠDataLoader": 67188, + "ĠDataManager": 87007, + "ĠDataRow": 50184, + "ĠDataService": 76555, + "ĠDataSet": 27036, + "ĠDataSource": 43384, + "ĠDataTable": 16922, + "ĠDataType": 33172, + "ĠDataTypes": 35043, + "ĠDataView": 94570, + "ĠDatabase": 9994, + "ĠDatabaseReference": 80046, + "ĠDatagram": 96804, + "ĠDataset": 39183, + "ĠDate": 2631, + "ĠDateFormat": 67857, + "ĠDateFormatter": 97953, + "ĠDatePicker": 63426, + "ĠDateTime": 6520, + "ĠDateTimeKind": 96209, + "ĠDateTimeOffset": 78494, + "ĠDaten": 46954, + "ĠDatensch": 91741, + "ĠDates": 51624, + "ĠDating": 17332, + "ĠDatos": 72359, + "ĠDatum": 68459, + "ĠDaughter": 63294, + "ĠDav": 5347, + "ĠDave": 20238, + "ĠDavid": 6798, + "ĠDavidson": 53245, + "ĠDavies": 55772, + "ĠDavis": 16782, + "ĠDaw": 36751, + "ĠDawn": 34507, + "ĠDawson": 66212, + "ĠDay": 6059, + "ĠDays": 21192, + "ĠDayton": 48173, + "ĠDaytona": 86912, + "ĠDb": 11988, + "ĠDbContext": 60194, + "ĠDbSet": 31150, + "ĠDbType": 49231, + "ĠDe": 1581, + "ĠDeV": 72596, + "ĠDead": 15021, + "ĠDeadline": 61956, + "ĠDeadly": 94837, + "ĠDeadpool": 93224, + "ĠDeal": 26263, + "ĠDealer": 51896, + "ĠDealers": 78189, + "ĠDeals": 41182, + "ĠDean": 24061, + "ĠDear": 42822, + "ĠDeath": 15911, + "ĠDeaths": 84712, + "ĠDeb": 35461, + "ĠDebate": 82788, + "ĠDebbie": 63389, + "ĠDebian": 56607, + "ĠDeborah": 69525, + "ĠDebt": 57554, + "ĠDebug": 11091, + "ĠDebugger": 93646, + "ĠDec": 3714, + "ĠDecay": 97286, + "ĠDecember": 6652, + "ĠDecide": 98881, + "ĠDecimal": 26728, + "ĠDecimalFormat": 80796, + "ĠDecision": 40425, + "ĠDeck": 28416, + "ĠDecl": 31625, + "ĠDeclaration": 40921, + "ĠDeclarations": 88275, + "ĠDeclare": 61310, + "ĠDecode": 50194, + "ĠDecoder": 50472, + "ĠDecom": 96378, + "ĠDecomp": 62378, + "ĠDecompiled": 79417, + "ĠDecor": 26971, + "ĠDecorating": 97220, + "ĠDecoration": 64019, + "ĠDecre": 64101, + "ĠDecrypt": 89146, + "ĠDed": 50399, + "ĠDedicated": 95558, + "ĠDee": 54904, + "ĠDeep": 18183, + "ĠDeepCopy": 75931, + "ĠDeer": 63091, + "ĠDef": 3892, + "ĠDefault": 7899, + "ĠDefaultValue": 97780, + "ĠDefaults": 35990, + "ĠDefence": 38907, + "ĠDefendant": 87545, + "ĠDefendants": 98839, + "ĠDefender": 65014, + "ĠDefense": 16373, + "ĠDefensive": 70148, + "ĠDeferred": 75515, + "ĠDefine": 18614, + "ĠDefined": 68846, + "ĠDefines": 37826, + "ĠDefinitely": 59898, + "ĠDefinition": 19709, + "ĠDefinitions": 46513, + "ĠDeg": 57237, + "ĠDegree": 37082, + "ĠDegrees": 92901, + "ĠDek": 74463, + "ĠDel": 7313, + "ĠDelaware": 39738, + "ĠDelay": 39793, + "ĠDelegate": 40180, + "ĠDelete": 10428, + "ĠDeleted": 54117, + "ĠDeletes": 55524, + "ĠDeleting": 81850, + "ĠDelhi": 21996, + "ĠDelicious": 84688, + "ĠDeliver": 64652, + "ĠDelivery": 26211, + "ĠDell": 38324, + "ĠDelta": 24957, + "ĠDeluxe": 52961, + "ĠDem": 4724, + "ĠDemand": 54253, + "ĠDemo": 28523, + "ĠDemocr": 6251, + "ĠDemocracy": 47089, + "ĠDemocrat": 23901, + "ĠDemocratic": 11396, + "ĠDemocrats": 12363, + "ĠDemon": 32305, + "ĠDemonstr": 85401, + "ĠDemp": 97744, + "ĠDems": 69591, + "ĠDen": 9774, + "ĠDeng": 91729, + "ĠDenied": 87112, + "ĠDenis": 60603, + "ĠDenise": 80249, + "ĠDenmark": 34340, + "ĠDenn": 71161, + "ĠDennis": 34627, + "ĠDense": 42522, + "ĠDensity": 72610, + "ĠDent": 42811, + "ĠDental": 47704, + "ĠDenver": 22117, + "ĠDep": 4148, + "ĠDepart": 39755, + "ĠDepartment": 5887, + "ĠDepartments": 84508, + "ĠDepend": 98328, + "ĠDependencies": 80490, + "ĠDependency": 32977, + "ĠDependencyProperty": 56552, + "ĠDepending": 39630, + "ĠDepends": 94011, + "ĠDeploy": 70595, + "ĠDeployment": 66292, + "ĠDepos": 77050, + "ĠDeposit": 48471, + "ĠDepot": 46970, + "ĠDeprecated": 83916, + "ĠDepression": 45804, + "ĠDept": 50221, + "ĠDepth": 43920, + "ĠDepths": 95918, + "ĠDeputy": 31624, + "ĠDer": 12741, + "ĠDerby": 45773, + "ĠDere": 89678, + "ĠDerek": 42323, + "ĠDerived": 67849, + "ĠDerm": 75408, + "ĠDerne": 142213, + "ĠDerneÄŁi": 142214, + "ĠDerrick": 72089, + "ĠDes": 3874, + "ĠDesc": 31185, + "ĠDescribe": 60785, + "ĠDescription": 7662, + "ĠDescriptor": 55692, + "ĠDesde": 97943, + "ĠDeserialize": 48440, + "ĠDesert": 42186, + "ĠDesign": 6982, + "ĠDesigned": 47425, + "ĠDesigner": 15569, + "ĠDesigns": 43266, + "ĠDesire": 81424, + "ĠDesired": 88061, + "ĠDesk": 38694, + "ĠDesktop": 35300, + "ĠDesmond": 96523, + "ĠDespite": 17715, + "ĠDespués": 141005, + "ĠDess": 69496, + "ĠDest": 28531, + "ĠDestination": 41728, + "ĠDestiny": 50397, + "ĠDestroy": 22251, + "ĠDestruction": 79747, + "ĠDestructor": 59068, + "ĠDet": 16423, + "ĠDetail": 25771, + "ĠDetailed": 62665, + "ĠDetails": 12309, + "ĠDetay": 139461, + "ĠDetaylı": 139462, + "ĠDetect": 33287, + "ĠDetected": 87950, + "ĠDetection": 57353, + "ĠDetective": 55279, + "ĠDetector": 79019, + "ĠDetermin": 34907, + "ĠDetermine": 29901, + "ĠDetermines": 52712, + "ĠDetroit": 19761, + "ĠDeus": 71785, + "ĠDeutsch": 33315, + "ĠDeutsche": 57671, + "ĠDeutschland": 44979, + "ĠDev": 6040, + "ĠDevComponents": 41225, + "ĠDevExpress": 10056, + "ĠDevelop": 7843, + "ĠDeveloped": 79010, + "ĠDeveloper": 24879, + "ĠDevelopers": 46617, + "ĠDeveloping": 80645, + "ĠDevelopment": 10816, + "ĠDevice": 13903, + "ĠDevices": 40377, + "ĠDevil": 39201, + "ĠDevils": 58954, + "ĠDevin": 79992, + "ĠDevon": 59334, + "ĠDew": 49300, + "ĠDex": 72642, + "ĠDexter": 86364, + "ĠDez": 74882, + "ĠDeze": 87836, + "ĠDh": 43227, + "ĠDhabi": 72779, + "ĠDi": 7767, + "ĠDia": 65303, + "ĠDiabetes": 52589, + "ĠDiablo": 73237, + "ĠDiagnosis": 94352, + "ĠDiagnostic": 49988, + "ĠDiagram": 35261, + "ĠDial": 66155, + "ĠDialog": 13393, + "ĠDialogInterface": 46308, + "ĠDialogResult": 36510, + "ĠDialogue": 69489, + "ĠDiameter": 75561, + "ĠDiamond": 24334, + "ĠDiamonds": 90110, + "ĠDiana": 44409, + "ĠDiane": 53090, + "ĠDiary": 52025, + "ĠDias": 87339, + "ĠDiaz": 57031, + "ĠDice": 56645, + "ĠDich": 97703, + "ĠDick": 22557, + "ĠDickens": 92899, + "ĠDickinson": 97237, + "ĠDict": 29126, + "ĠDictionary": 10466, + "ĠDid": 14568, + "ĠDidn": 56972, + "ĠDie": 8408, + "ĠDied": 87887, + "ĠDiego": 18336, + "ĠDien": 60404, + "ĠDies": 23980, + "ĠDiese": 54068, + "ĠDiesel": 53794, + "ĠDiet": 26212, + "ĠDietary": 82708, + "ĠDiff": 28369, + "ĠDifference": 55080, + "ĠDifferences": 85797, + "ĠDifferent": 33396, + "ĠDifferential": 98736, + "ĠDifficulty": 54946, + "ĠDig": 37969, + "ĠDigest": 53289, + "ĠDigit": 71465, + "ĠDigital": 14102, + "ĠDil": 52767, + "ĠDillon": 82566, + "ĠDim": 8126, + "ĠDimension": 27923, + "ĠDimensions": 32023, + "ĠDimit": 87590, + "ĠDin": 44773, + "ĠDing": 79373, + "ĠDining": 38090, + "ĠDinner": 48091, + "ĠDio": 84942, + "ĠDion": 64057, + "ĠDios": 68098, + "ĠDip": 55047, + "ĠDipl": 45462, + "ĠDiploma": 76031, + "ĠDir": 30094, + "ĠDire": 44511, + "ĠDirect": 7139, + "ĠDirectX": 66764, + "ĠDirected": 77205, + "ĠDirection": 18904, + "ĠDirections": 53486, + "ĠDirective": 56752, + "ĠDirector": 10560, + "ĠDirectorate": 93366, + "ĠDirectors": 44985, + "ĠDirectory": 18033, + "ĠDirectoryInfo": 81688, + "ĠDirk": 83929, + "ĠDirt": 86319, + "ĠDirty": 54291, + "ĠDis": 4093, + "ĠDisabilities": 97449, + "ĠDisability": 74268, + "ĠDisable": 28027, + "ĠDisabled": 58206, + "ĠDisaster": 72278, + "ĠDisc": 11735, + "ĠDiscipline": 97546, + "ĠDisclaimer": 66829, + "ĠDisclosure": 77254, + "ĠDisco": 93629, + "ĠDisconnect": 69090, + "ĠDiscord": 34064, + "ĠDiscount": 32659, + "ĠDiscounts": 98374, + "ĠDiscover": 32939, + "ĠDiscovery": 38093, + "ĠDiscrim": 77300, + "ĠDiscuss": 65279, + "ĠDiscussion": 35513, + "ĠDiscussions": 97125, + "ĠDise": 24002, + "ĠDisease": 30874, + "ĠDiseases": 69574, + "ĠDish": 48168, + "ĠDisk": 38868, + "ĠDisney": 16390, + "ĠDisneyland": 74074, + "ĠDisorder": 62777, + "ĠDisorders": 72033, + "ĠDisp": 69730, + "ĠDispatch": 21133, + "ĠDispatchQueue": 47851, + "ĠDispatcher": 58969, + "ĠDisplay": 10623, + "ĠDisplayName": 81783, + "ĠDisplays": 49606, + "ĠDisposable": 51286, + "ĠDispose": 16469, + "ĠDisqus": 80751, + "ĠDiss": 51032, + "ĠDissertation": 98016, + "ĠDist": 27604, + "ĠDistance": 31135, + "ĠDistrib": 53591, + "ĠDistributed": 43955, + "ĠDistribution": 33909, + "ĠDistrict": 10942, + "ĠDit": 52418, + "ĠDiv": 8765, + "ĠDive": 88517, + "ĠDiversity": 64971, + "ĠDivide": 62902, + "ĠDivider": 47457, + "ĠDivine": 42261, + "ĠDivision": 14489, + "ĠDix": 75143, + "ĠDixon": 63695, + "ĠDiy": 94904, + "ĠDiÄŁer": 138556, + "ĠDiá»ħn": 141351, + "ĠDj": 51062, + "ĠDjango": 52604, + "ĠDmit": 66121, + "ĠDmitry": 87131, + "ĠDo": 3155, + "ĠDob": 64251, + "ĠDoc": 21709, + "ĠDoch": 78021, + "ĠDock": 44101, + "ĠDocker": 40549, + "ĠDocs": 60691, + "ĠDoctor": 18635, + "ĠDoctors": 53843, + "ĠDoctrine": 28632, + "ĠDocument": 11789, + "ĠDocumentary": 98787, + "ĠDocumentation": 44465, + "ĠDocuments": 44790, + "ĠDod": 35868, + "ĠDodd": 82993, + "ĠDodge": 42881, + "ĠDodgers": 55467, + "ĠDoe": 49628, + "ĠDoes": 12553, + "ĠDoesn": 48832, + "ĠDog": 14254, + "ĠDogs": 38425, + "ĠDoing": 43145, + "ĠDok": 66850, + "ĠDol": 24244, + "ĠDoll": 38757, + "ĠDollar": 42364, + "ĠDollars": 80189, + "ĠDolphin": 94996, + "ĠDolphins": 53712, + "ĠDom": 20751, + "ĠDomain": 21070, + "ĠDome": 68445, + "ĠDomestic": 51749, + "ĠDomin": 22474, + "ĠDoming": 77757, + "ĠDominic": 67034, + "ĠDominican": 66013, + "ĠDominion": 75421, + "ĠDomino": 93765, + "ĠDon": 4320, + "ĠDonald": 9453, + "ĠDonate": 96202, + "ĠDonation": 77230, + "ĠDone": 27357, + "ĠDonetsk": 97351, + "ĠDong": 52383, + "ĠDonna": 46763, + "ĠDonne": 69983, + "ĠDonovan": 70700, + "ĠDont": 61484, + "ĠDoom": 50234, + "ĠDoor": 24191, + "ĠDoors": 49617, + "ĠDop": 96574, + "ĠDor": 24242, + "ĠDorm": 97532, + "ĠDorothy": 72558, + "ĠDort": 58616, + "ĠDortmund": 71450, + "ĠDos": 56920, + "ĠDot": 31262, + "ĠDota": 85127, + "ĠDou": 17007, + "ĠDoub": 54864, + "ĠDouble": 7093, + "ĠDoug": 31508, + "ĠDough": 89295, + "ĠDouglas": 30064, + "ĠDove": 88984, + "ĠDover": 86826, + "ĠDow": 42432, + "ĠDown": 6285, + "ĠDowning": 83679, + "ĠDownload": 8577, + "ĠDownloads": 49161, + "ĠDowns": 84678, + "ĠDowntown": 43885, + "ĠDoyle": 65209, + "ĠDoÄŁu": 138709, + "ĠDr": 2926, + "ĠDra": 73102, + "ĠDraco": 71702, + "ĠDraft": 28564, + "ĠDrag": 24861, + "ĠDragon": 16145, + "ĠDragons": 48009, + "ĠDrain": 61268, + "ĠDrake": 41765, + "ĠDram": 79110, + "ĠDrama": 46469, + "ĠDraw": 11992, + "ĠDrawable": 58016, + "ĠDrawer": 48260, + "ĠDrawing": 37759, + "ĠDraws": 83432, + "ĠDre": 65010, + "ĠDread": 81204, + "ĠDream": 17829, + "ĠDreams": 52618, + "ĠDresden": 85445, + "ĠDress": 28218, + "ĠDresses": 93517, + "ĠDrew": 40108, + "ĠDrill": 76874, + "ĠDrink": 47473, + "ĠDrinking": 82581, + "ĠDrinks": 96998, + "ĠDrive": 16150, + "ĠDriver": 14577, + "ĠDriverManager": 63508, + "ĠDrivers": 45737, + "ĠDriving": 52253, + "ĠDro": 56867, + "ĠDrone": 73481, + "ĠDrop": 15733, + "ĠDropDownList": 98755, + "ĠDropIndex": 94176, + "ĠDropbox": 72584, + "ĠDropdown": 45594, + "ĠDropout": 81132, + "ĠDrops": 89636, + "ĠDrug": 25109, + "ĠDrugs": 50666, + "ĠDruid": 72635, + "ĠDrum": 45406, + "ĠDrupal": 35044, + "ĠDry": 30841, + "ĠDt": 94151, + "ĠDtype": 86627, + "ĠDu": 15687, + "ĠDual": 33659, + "ĠDub": 17095, + "ĠDubai": 34038, + "ĠDublin": 32877, + "ĠDuc": 79773, + "ĠDuch": 60022, + "ĠDuchess": 94939, + "ĠDuck": 45770, + "ĠDucks": 67782, + "ĠDud": 73167, + "ĠDude": 85209, + "ĠDudley": 86902, + "ĠDue": 23662, + "ĠDuel": 57194, + "ĠDuffy": 85862, + "ĠDuis": 96215, + "ĠDuke": 26356, + "ĠDul": 62615, + "ĠDum": 63982, + "ĠDumbledore": 80983, + "ĠDummy": 50567, + "ĠDump": 29693, + "ĠDumpster": 43264, + "ĠDun": 28738, + "ĠDuncan": 41309, + "ĠDund": 73915, + "ĠDungeon": 50847, + "ĠDungeons": 93001, + "ĠDunk": 68683, + "ĠDunn": 63745, + "ĠDuo": 73212, + "ĠDup": 50283, + "ĠDuplicate": 71071, + "ĠDur": 20138, + "ĠDurant": 66340, + "ĠDuration": 21045, + "ĠDurch": 60423, + "ĠDurham": 57714, + "ĠDuring": 11954, + "ĠDust": 31993, + "ĠDustin": 78318, + "ĠDut": 43536, + "ĠDutch": 23234, + "ĠDuterte": 63452, + "ĠDuty": 35589, + "ĠDw": 48707, + "ĠDwarf": 79730, + "ĠDwight": 71745, + "ĠDx": 88797, + "ĠDy": 41948, + "ĠDylan": 43358, + "ĠDyn": 43938, + "ĠDynam": 32493, + "ĠDynamic": 21886, + "ĠDynamics": 52611, + "ĠDynamo": 71813, + "ĠDynasty": 73395, + "ĠDys": 64042, + "ĠDysfunction": 87166, + "ĠDzi": 134288, + "ĠDziÄĻki": 134289, + "ĠDé": 49401, + "ĠDön": 140901, + "ĠDü": 131634, + "ĠDünya": 131635, + "ĠDÅ©ng": 142840, + "ĠDương": 130695, + "ĠDá»±": 133784, + "ĠDá»ĭch": 134586, + "ĠE": 468, + "ĠEA": 38362, + "ĠEACH": 94413, + "ĠEAR": 90296, + "ĠEAST": 77625, + "ĠEB": 49142, + "ĠEC": 20633, + "ĠECB": 79947, + "ĠECC": 77316, + "ĠECM": 79600, + "ĠECS": 87664, + "ĠED": 16036, + "ĠEDGE": 82806, + "ĠEDIT": 32257, + "ĠEDM": 98017, + "ĠEDT": 47923, + "ĠEDUC": 92977, + "ĠEE": 41352, + "ĠEEG": 75808, + "ĠEEPROM": 61195, + "ĠEF": 43990, + "ĠEFF": 95188, + "ĠEFFECT": 62053, + "ĠEFI": 39082, + "ĠEG": 56111, + "ĠEGL": 58655, + "ĠEH": 94686, + "ĠEI": 92811, + "ĠEIF": 27695, + "ĠEINA": 95382, + "ĠEINVAL": 76269, + "ĠEITHER": 93475, + "ĠEL": 17258, + "ĠELECT": 72477, + "ĠELEMENT": 62356, + "ĠELF": 72568, + "ĠELSE": 52926, + "ĠEM": 16902, + "ĠEMAIL": 51708, + "ĠEMC": 73704, + "ĠEMP": 51592, + "ĠEMPTY": 41314, + "ĠEMS": 70065, + "ĠEN": 5190, + "ĠENABLE": 33240, + "ĠENC": 72262, + "ĠEND": 11176, + "ĠENERGY": 83276, + "ĠENG": 70539, + "ĠENGINE": 48360, + "ĠENT": 27816, + "ĠENTER": 52905, + "ĠENTITY": 73871, + "ĠENTRY": 72792, + "ĠENUM": 68404, + "ĠENV": 32791, + "ĠEO": 92526, + "ĠEOF": 29457, + "ĠEOS": 48901, + "ĠEP": 19072, + "ĠEPA": 26158, + "ĠEPS": 40607, + "ĠEQ": 38994, + "ĠEQUAL": 91609, + "ĠEQUI": 85399, + "ĠER": 26492, + "ĠERA": 44020, + "ĠERC": 56052, + "ĠERP": 65179, + "ĠERR": 10165, + "ĠERROR": 12874, + "ĠES": 19287, + "ĠESA": 91643, + "ĠESC": 43274, + "ĠESL": 68322, + "ĠESP": 18117, + "ĠESPN": 26198, + "ĠEST": 25877, + "ĠESV": 83011, + "ĠET": 17768, + "ĠETA": 89014, + "ĠETF": 53063, + "ĠETH": 37742, + "ĠEU": 9812, + "ĠEUR": 37088, + "ĠEURO": 90434, + "ĠEV": 14892, + "ĠEVAL": 96392, + "ĠEVE": 86635, + "ĠEVEN": 26657, + "ĠEVENT": 12742, + "ĠEVENTS": 58619, + "ĠEVER": 73666, + "ĠEVERY": 44874, + "ĠEVP": 72758, + "ĠEVT": 93257, + "ĠEW": 62995, + "ĠEX": 4063, + "ĠEXAMPLE": 66246, + "ĠEXCEPTION": 80452, + "ĠEXEC": 62247, + "ĠEXEMPLARY": 30542, + "ĠEXISTS": 34839, + "ĠEXIT": 29839, + "ĠEXP": 13460, + "ĠEXPECT": 8921, + "ĠEXPER": 80081, + "ĠEXPORT": 74584, + "ĠEXPRESS": 16427, + "ĠEXT": 33757, + "ĠEXTI": 85018, + "ĠEXTRA": 65393, + "ĠEZ": 63286, + "ĠEa": 96088, + "ĠEach": 8886, + "ĠEag": 20916, + "ĠEagle": 35795, + "ĠEagles": 31609, + "ĠEar": 6809, + "ĠEarl": 47410, + "ĠEarlier": 46893, + "ĠEarly": 22752, + "ĠEarn": 47693, + "ĠEarth": 9237, + "ĠEase": 94021, + "ĠEasily": 45991, + "ĠEast": 6326, + "ĠEaster": 32400, + "ĠEastern": 18028, + "ĠEasy": 18609, + "ĠEat": 44514, + "ĠEating": 59538, + "ĠEaton": 95866, + "ĠEb": 25570, + "ĠEbay": 91316, + "ĠEbola": 46982, + "ĠEbony": 72114, + "ĠEbook": 88671, + "ĠEc": 36111, + "ĠEcc": 80660, + "ĠEcho": 37806, + "ĠEck": 81640, + "ĠEclipse": 36123, + "ĠEco": 50851, + "ĠEcology": 77275, + "ĠEcon": 12413, + "ĠEconom": 50083, + "ĠEconomic": 22546, + "ĠEconomics": 34200, + "ĠEconomist": 82938, + "ĠEconomy": 37561, + "ĠEcuador": 53991, + "ĠEd": 3199, + "ĠEddie": 40680, + "ĠEden": 50815, + "ĠEdgar": 58991, + "ĠEdge": 10349, + "ĠEdgeInsets": 15396, + "ĠEdinburgh": 40306, + "ĠEdison": 73776, + "ĠEdit": 8340, + "ĠEditText": 30483, + "ĠEdited": 65789, + "ĠEditing": 61053, + "ĠEdition": 14069, + "ĠEditor": 12580, + "ĠEditorGUI": 94408, + "ĠEditorGUILayout": 52822, + "ĠEditorial": 65133, + "ĠEditors": 81786, + "ĠEdmonton": 45287, + "ĠEdmund": 70811, + "ĠEdu": 46273, + "ĠEduardo": 76554, + "ĠEduc": 10145, + "ĠEducación": 142597, + "ĠEducation": 11668, + "ĠEducational": 45845, + "ĠEducação": 137911, + "ĠEdward": 21891, + "ĠEdwards": 36763, + "ĠEdwin": 75149, + "ĠEen": 56418, + "ĠEf": 94089, + "ĠEff": 30022, + "ĠEffect": 13449, + "ĠEffective": 46923, + "ĠEffects": 29146, + "ĠEfficiency": 66567, + "ĠEfficient": 89006, + "ĠEg": 38075, + "ĠEgg": 41213, + "ĠEggs": 75792, + "ĠEgypt": 14867, + "ĠEgyptian": 32489, + "ĠEgyptians": 81504, + "ĠEh": 60551, + "ĠEid": 85744, + "ĠEig": 90080, + "ĠEigen": 24092, + "ĠEight": 35844, + "ĠEighth": 96488, + "ĠEin": 18069, + "ĠEine": 51310, + "ĠEins": 60352, + "ĠEinsatz": 84247, + "ĠEinstein": 54052, + "ĠEis": 96880, + "ĠEisen": 54493, + "ĠEisenhower": 88081, + "ĠEither": 20988, + "ĠEj": 52758, + "ĠEk": 48988, + "ĠEl": 3984, + "ĠElaine": 82147, + "ĠElastic": 51910, + "ĠElasticsearch": 58887, + "ĠEld": 26867, + "ĠElder": 38569, + "ĠEle": 25949, + "ĠEleanor": 89430, + "ĠElect": 9884, + "ĠElection": 25754, + "ĠElections": 57501, + "ĠElectoral": 66576, + "ĠElectric": 20598, + "ĠElectrical": 43934, + "ĠElectricity": 73356, + "ĠElectro": 68290, + "ĠElectron": 76876, + "ĠElectronic": 34169, + "ĠElectronics": 37684, + "ĠElegant": 80210, + "ĠElem": 78768, + "ĠElement": 8543, + "ĠElementRef": 57690, + "ĠElementType": 80526, + "ĠElemental": 70638, + "ĠElementary": 40663, + "ĠElements": 34157, + "ĠElena": 68575, + "ĠElephant": 78089, + "ĠElev": 54564, + "ĠElevated": 96593, + "ĠEleven": 81393, + "ĠElf": 43600, + "ĠEli": 33262, + "ĠElias": 85656, + "ĠElig": 68024, + "ĠElijah": 82592, + "ĠElim": 42320, + "ĠElis": 90086, + "ĠElite": 33764, + "ĠElizabeth": 20731, + "ĠElk": 87606, + "ĠEll": 13542, + "ĠElla": 77796, + "ĠElle": 45308, + "ĠEllen": 46585, + "ĠElli": 46712, + "ĠEllie": 88261, + "ĠElliot": 68418, + "ĠElliott": 47544, + "ĠEllis": 43607, + "ĠEllison": 81647, + "ĠElm": 64229, + "ĠElo": 99069, + "ĠElon": 68539, + "ĠElsa": 47964, + "ĠElse": 18804, + "ĠElseIf": 62223, + "ĠElves": 97604, + "ĠElvis": 66206, + "ĠEm": 5748, + "ĠEmacs": 63476, + "ĠEmail": 8299, + "ĠEmails": 71866, + "ĠEmanuel": 76660, + "ĠEmb": 29127, + "ĠEmbassy": 52427, + "ĠEmbed": 37068, + "ĠEmbedded": 61383, + "ĠEmber": 48234, + "ĠEmblem": 93136, + "ĠEmer": 20543, + "ĠEmerald": 63096, + "ĠEmergency": 31608, + "ĠEmerging": 85048, + "ĠEmerson": 79454, + "ĠEmil": 87724, + "ĠEmily": 34166, + "ĠEmin": 80400, + "ĠEmirates": 48701, + "ĠEmit": 61537, + "ĠEmm": 65455, + "ĠEmma": 34935, + "ĠEmmanuel": 60444, + "ĠEmmy": 66012, + "ĠEmoji": 82367, + "ĠEmotional": 94200, + "ĠEmp": 13409, + "ĠEmperor": 34314, + "ĠEmpire": 20448, + "ĠEmploy": 20782, + "ĠEmployee": 16850, + "ĠEmployees": 43641, + "ĠEmployer": 74852, + "ĠEmployment": 40852, + "ĠEmpresa": 83673, + "ĠEmpty": 22228, + "ĠEn": 2925, + "ĠEnable": 18567, + "ĠEnabled": 29653, + "ĠEnables": 76316, + "ĠEnc": 10751, + "ĠEncode": 56562, + "ĠEncoder": 55115, + "ĠEncoding": 29330, + "ĠEncore": 99078, + "ĠEncounter": 91857, + "ĠEncrypt": 55459, + "ĠEncryption": 62055, + "ĠEncyclopedia": 67098, + "ĠEnd": 3972, + "ĠEnde": 51292, + "ĠEnded": 76071, + "ĠEnding": 73421, + "ĠEndpoint": 47269, + "ĠEnds": 60560, + "ĠEnemies": 60481, + "ĠEnemy": 30751, + "ĠEnerg": 74001, + "ĠEnergy": 12354, + "ĠEnforcement": 42956, + "ĠEng": 3285, + "ĠEngagement": 59363, + "ĠEngel": 90890, + "ĠEngine": 8200, + "ĠEngineer": 28383, + "ĠEngineering": 16595, + "ĠEngineers": 48696, + "ĠEngines": 94318, + "ĠEngl": 98630, + "ĠEngland": 9448, + "ĠEnglish": 6364, + "ĠEnh": 28657, + "ĠEnhanced": 61449, + "ĠEnhancement": 52517, + "ĠEnjoy": 22656, + "ĠEnlight": 74050, + "ĠEnlightenment": 91831, + "ĠEnough": 61592, + "ĠEnrique": 84030, + "ĠEnrollment": 87984, + "ĠEns": 57868, + "ĠEnsemble": 88357, + "ĠEnsure": 29279, + "ĠEnt": 4863, + "ĠEnter": 11252, + "ĠEntered": 74975, + "ĠEnterprise": 25472, + "ĠEnterprises": 65956, + "ĠEntertainment": 22518, + "ĠEntire": 79240, + "ĠEntities": 52249, + "ĠEntity": 10390, + "ĠEntityManager": 48767, + "ĠEntityState": 65187, + "ĠEntityType": 73117, + "ĠEntr": 47141, + "ĠEntrance": 83120, + "ĠEntre": 41659, + "ĠEntrepreneur": 68002, + "ĠEntries": 65193, + "ĠEntry": 15788, + "ĠEntryPoint": 72685, + "ĠEnts": 93403, + "ĠEntwicklung": 92979, + "ĠEntão": 139848, + "ĠEnum": 14086, + "ĠEnumerable": 44125, + "ĠEnumerableStream": 71916, + "ĠEnumeration": 77125, + "ĠEnumerator": 76511, + "ĠEnums": 82947, + "ĠEnv": 37039, + "ĠEnvironment": 11586, + "ĠEnvironmental": 24060, + "ĠEp": 11020, + "ĠEph": 68372, + "ĠEpic": 39367, + "ĠEpid": 68180, + "ĠEpidemi": 87853, + "ĠEpiscopal": 95954, + "ĠEpisode": 19832, + "ĠEpisodes": 89137, + "ĠEpoch": 87379, + "ĠEpstein": 80032, + "ĠEq": 33122, + "ĠEqu": 11702, + "ĠEqual": 38474, + "ĠEquality": 51037, + "ĠEquals": 38735, + "ĠEquation": 84670, + "ĠEquip": 69714, + "ĠEquipment": 20236, + "ĠEquity": 57488, + "ĠEquivalent": 91228, + "ĠEr": 9740, + "ĠEra": 47588, + "ĠErd": 44693, + "ĠErdogan": 56089, + "ĠErdoÄŁan": 133757, + "ĠEre": 95013, + "ĠErect": 79155, + "ĠErectile": 88789, + "ĠErf": 41663, + "ĠErg": 73202, + "ĠErgebn": 78796, + "ĠEric": 16247, + "ĠErica": 91835, + "ĠErick": 79915, + "ĠErie": 78487, + "ĠErik": 41346, + "ĠErin": 55814, + "ĠErl": 80834, + "ĠErnest": 60923, + "ĠErnst": 73110, + "ĠErot": 27964, + "ĠErotic": 81842, + "ĠErotik": 83162, + "ĠErotische": 65652, + "ĠErr": 15495, + "ĠError": 4600, + "ĠErrorCode": 60084, + "ĠErrorHandler": 88331, + "ĠErrorMessage": 41827, + "ĠErrorResponse": 95724, + "ĠErrors": 39256, + "ĠEs": 9236, + "ĠEsc": 15547, + "ĠEscape": 45643, + "ĠEscort": 26585, + "ĠEscorts": 49654, + "ĠEsk": 58369, + "ĠEsp": 26514, + "ĠEspañ": 88982, + "ĠEspaña": 56508, + "ĠEspañol": 142658, + "ĠEspecially": 35525, + "ĠEsper": 69528, + "ĠEspresso": 98490, + "ĠEss": 12838, + "ĠEssay": 25957, + "ĠEssays": 55614, + "ĠEssen": 98675, + "ĠEssence": 83770, + "ĠEssential": 47733, + "ĠEssentially": 70754, + "ĠEssentials": 75160, + "ĠEssex": 63058, + "ĠEst": 9403, + "ĠEsta": 55708, + "ĠEstablish": 39398, + "ĠEstablished": 86250, + "ĠEstablishment": 86765, + "ĠEstado": 40174, + "ĠEstados": 63893, + "ĠEstate": 24443, + "ĠEstates": 91600, + "ĠEste": 38676, + "ĠEsther": 83638, + "ĠEstimate": 72715, + "ĠEstimated": 81986, + "ĠEstimates": 94519, + "ĠEsto": 92025, + "ĠEston": 53323, + "ĠEstonia": 68362, + "ĠEstr": 81484, + "ĠEstá": 136473, + "ĠEt": 18888, + "ĠEternal": 66772, + "ĠEth": 14356, + "ĠEthan": 62164, + "ĠEther": 34878, + "ĠEthereum": 33946, + "ĠEthernet": 45408, + "ĠEthi": 44058, + "ĠEthics": 49181, + "ĠEthiopia": 59357, + "ĠEthiopian": 95534, + "ĠEthnic": 91418, + "ĠEtsy": 64529, + "ĠEu": 21963, + "ĠEug": 40059, + "ĠEugene": 50458, + "ĠEuler": 80018, + "ĠEur": 84377, + "ĠEuras": 87077, + "ĠEuro": 19461, + "ĠEurop": 67015, + "ĠEuropa": 38500, + "ĠEurope": 4505, + "ĠEuropean": 7513, + "ĠEuropeans": 50507, + "ĠEuros": 85466, + "ĠEv": 10424, + "ĠEva": 54392, + "ĠEval": 58239, + "ĠEvalu": 75098, + "ĠEvaluate": 54115, + "ĠEvaluation": 39288, + "ĠEvan": 43943, + "ĠEvangel": 54706, + "ĠEvans": 29463, + "ĠEve": 31360, + "ĠEvel": 80996, + "ĠEvelyn": 95325, + "ĠEven": 7418, + "ĠEvening": 56102, + "ĠEvent": 3665, + "ĠEventArgs": 8124, + "ĠEventBus": 89567, + "ĠEventEmitter": 30004, + "ĠEventHandler": 31957, + "ĠEventType": 63128, + "ĠEvents": 17627, + "ĠEventually": 37174, + "ĠEver": 17892, + "ĠEverest": 86478, + "ĠEverett": 77095, + "ĠEverton": 62439, + "ĠEvery": 7209, + "ĠEverybody": 50798, + "ĠEveryday": 78616, + "ĠEveryone": 21455, + "ĠEverything": 20094, + "ĠEverywhere": 97225, + "ĠEvidence": 43696, + "ĠEvil": 33719, + "ĠEvo": 97870, + "ĠEvolution": 37221, + "ĠEx": 1374, + "ĠExact": 74480, + "ĠExactly": 68490, + "ĠExam": 32310, + "ĠExamination": 71494, + "ĠExaminer": 78246, + "ĠExample": 13383, + "ĠExamples": 25311, + "ĠExc": 38895, + "ĠExcel": 21029, + "ĠExcell": 49907, + "ĠExcellence": 57140, + "ĠExcellent": 36766, + "ĠExcept": 37167, + "ĠException": 4112, + "ĠExceptions": 64629, + "ĠExchange": 18705, + "ĠExclude": 98477, + "ĠExclusive": 48280, + "ĠExec": 10290, + "ĠExecute": 20848, + "ĠExecutes": 55542, + "ĠExecution": 30928, + "ĠExecutionContext": 85451, + "ĠExecutive": 17880, + "ĠExecutor": 56032, + "ĠExecutors": 95093, + "ĠExercise": 32818, + "ĠExercises": 90454, + "ĠExhaust": 88234, + "ĠExhib": 39274, + "ĠExhibit": 76495, + "ĠExhibition": 66933, + "ĠExist": 61809, + "ĠExisting": 68471, + "ĠExists": 72426, + "ĠExit": 18995, + "ĠExiting": 86262, + "ĠExodus": 87842, + "ĠExp": 7787, + "ĠExpand": 50141, + "ĠExpanded": 39237, + "ĠExpansion": 54554, + "ĠExpect": 32085, + "ĠExpected": 31021, + "ĠExped": 63312, + "ĠExpedition": 83774, + "ĠExpense": 67531, + "ĠExpenses": 91423, + "ĠExperience": 20796, + "ĠExperienced": 99023, + "ĠExperiment": 46793, + "ĠExperimental": 56608, + "ĠExpert": 32157, + "ĠExperts": 50759, + "ĠExpires": 97410, + "ĠExpl": 18006, + "ĠExplain": 81917, + "ĠExplanation": 71287, + "ĠExplicit": 31330, + "ĠExploration": 74922, + "ĠExplore": 44052, + "ĠExplorer": 29152, + "ĠExplos": 84866, + "ĠExplosion": 92686, + "ĠExpo": 51323, + "ĠExport": 22191, + "ĠExposure": 70766, + "ĠExpr": 28819, + "ĠExpress": 17399, + "ĠExpression": 16378, + "ĠExt": 9447, + "ĠExtend": 70083, + "ĠExtended": 40565, + "ĠExtension": 26473, + "ĠExtensions": 55795, + "ĠExterior": 70681, + "ĠExternal": 30936, + "ĠExtr": 71234, + "ĠExtra": 24992, + "ĠExtract": 22826, + "ĠExtraction": 94506, + "ĠExtras": 87917, + "ĠExtreme": 49395, + "ĠExtremely": 92696, + "ĠExxon": 76027, + "ĠEy": 43411, + "ĠEye": 27829, + "ĠEyes": 41996, + "ĠEyl": 135120, + "ĠEylül": 135121, + "ĠEz": 38626, + "ĠEzek": 97379, + "ĠEzra": 84318, + "ĠEÄŁ": 132386, + "ĠEÄŁer": 135039, + "ĠEÄŁitim": 132387, + "ĠF": 434, + "ĠFA": 15008, + "ĠFAA": 65936, + "ĠFAC": 84303, + "ĠFACE": 58227, + "ĠFACT": 58543, + "ĠFAG": 86180, + "ĠFAIL": 33107, + "ĠFAILED": 49671, + "ĠFAILURE": 64024, + "ĠFALL": 87492, + "ĠFALSE": 7833, + "ĠFAMILY": 87469, + "ĠFAQ": 30972, + "ĠFAQs": 86584, + "ĠFAR": 56315, + "ĠFAST": 50733, + "ĠFAT": 67953, + "ĠFB": 31921, + "ĠFBI": 13839, + "ĠFC": 16013, + "ĠFCC": 37576, + "ĠFD": 35318, + "ĠFDA": 29373, + "ĠFE": 27931, + "ĠFEATURE": 46760, + "ĠFEATURES": 93262, + "ĠFEC": 92311, + "ĠFEMA": 91689, + "ĠFETCH": 32316, + "ĠFF": 29942, + "ĠFFT": 60036, + "ĠFG": 48297, + "ĠFH": 81198, + "ĠFHA": 77740, + "ĠFI": 50535, + "ĠFIELD": 40278, + "ĠFIFA": 43642, + "ĠFIFO": 47007, + "ĠFIG": 37599, + "ĠFIL": 63335, + "ĠFILE": 11837, + "ĠFILES": 78745, + "ĠFILTER": 35352, + "ĠFIN": 24344, + "ĠFINAL": 66959, + "ĠFIND": 63758, + "ĠFINSEQ": 70127, + "ĠFIR": 77608, + "ĠFIRE": 60463, + "ĠFIRST": 34813, + "ĠFIT": 7456, + "ĠFITNESS": 7721, + "ĠFIX": 20151, + "ĠFIXED": 91549, + "ĠFIXME": 27475, + "ĠFK": 84246, + "ĠFL": 12772, + "ĠFLAC": 89099, + "ĠFLAG": 30379, + "ĠFLAGS": 47239, + "ĠFLASH": 52326, + "ĠFLOAT": 50116, + "ĠFLT": 96061, + "ĠFM": 23447, + "ĠFML": 82927, + "ĠFN": 49487, + "ĠFName": 96524, + "ĠFO": 21763, + "ĠFOLLOW": 57672, + "ĠFONT": 61068, + "ĠFOOD": 87127, + "ĠFOOT": 80037, + "ĠFOR": 4613, + "ĠFORCE": 82589, + "ĠFORE": 43923, + "ĠFOREIGN": 81356, + "ĠFORM": 27824, + "ĠFORMAT": 52225, + "ĠFORWARD": 90079, + "ĠFOUND": 52559, + "ĠFOUR": 74091, + "ĠFOX": 44188, + "ĠFP": 33551, + "ĠFPGA": 89462, + "ĠFPS": 43628, + "ĠFR": 16654, + "ĠFRAME": 56416, + "ĠFRE": 62606, + "ĠFREE": 16257, + "ĠFRIEND": 79343, + "ĠFROM": 4295, + "ĠFRONT": 81547, + "ĠFS": 24289, + "ĠFSM": 70967, + "ĠFStar": 48150, + "ĠFString": 65236, + "ĠFT": 23301, + "ĠFTC": 78260, + "ĠFTP": 47550, + "ĠFU": 95512, + "ĠFUCK": 75021, + "ĠFULL": 39741, + "ĠFUN": 35449, + "ĠFUNC": 43352, + "ĠFUNCT": 62355, + "ĠFUNCTION": 24819, + "ĠFUNCTIONS": 52164, + "ĠFVector": 70284, + "ĠFW": 48029, + "ĠFX": 28901, + "ĠFXMLLoader": 90239, + "ĠFY": 46366, + "ĠFa": 17677, + "ĠFab": 19243, + "ĠFabric": 36307, + "ĠFac": 16945, + "ĠFace": 18596, + "ĠFacebook": 5573, + "ĠFaces": 51226, + "ĠFach": 78765, + "ĠFacial": 74522, + "ĠFacilities": 65289, + "ĠFacility": 46650, + "ĠFacing": 95318, + "ĠFact": 36712, + "ĠFaction": 90307, + "ĠFactor": 37729, + "ĠFactors": 67218, + "ĠFactory": 16937, + "ĠFactoryBot": 88469, + "ĠFactoryGirl": 57540, + "ĠFacts": 44983, + "ĠFaculty": 41804, + "ĠFade": 56660, + "ĠFah": 73507, + "ĠFahr": 77200, + "ĠFahrenheit": 68723, + "ĠFail": 39645, + "ĠFailed": 21379, + "ĠFailure": 32260, + "ĠFair": 14588, + "ĠFairfax": 76536, + "ĠFairfield": 87429, + "ĠFairy": 55076, + "ĠFaith": 32953, + "ĠFake": 36965, + "ĠFaker": 46027, + "ĠFal": 22869, + "ĠFalcon": 42861, + "ĠFalcons": 53339, + "ĠFalk": 71464, + "ĠFall": 14785, + "ĠFallen": 82988, + "ĠFalling": 84199, + "ĠFallon": 87068, + "ĠFallout": 58540, + "ĠFalls": 29643, + "ĠFalse": 3557, + "ĠFam": 33601, + "ĠFame": 38527, + "ĠFamil": 52159, + "ĠFamilie": 88537, + "ĠFamilies": 49456, + "ĠFamily": 12238, + "ĠFamous": 66788, + "ĠFan": 24714, + "ĠFancy": 81799, + "ĠFang": 57927, + "ĠFans": 41796, + "ĠFant": 19663, + "ĠFantastic": 56938, + "ĠFantasy": 26484, + "ĠFar": 13452, + "ĠFarage": 93906, + "ĠFare": 66471, + "ĠFargo": 57650, + "ĠFarm": 18438, + "ĠFarmer": 67464, + "ĠFarmers": 62322, + "ĠFarms": 67021, + "ĠFarr": 73906, + "ĠFarrell": 85345, + "ĠFasc": 69607, + "ĠFashion": 30600, + "ĠFast": 17288, + "ĠFaster": 84486, + "ĠFat": 25348, + "ĠFatal": 65629, + "ĠFate": 58925, + "ĠFather": 20322, + "ĠFathers": 75905, + "ĠFauc": 96362, + "ĠFaul": 96243, + "ĠFault": 59149, + "ĠFav": 80941, + "ĠFavor": 73787, + "ĠFavorite": 39722, + "ĠFavorites": 63218, + "ĠFax": 55364, + "ĠFay": 89305, + "ĠFayette": 98815, + "ĠFe": 3841, + "ĠFear": 41967, + "ĠFeast": 90928, + "ĠFeather": 78876, + "ĠFeature": 19998, + "ĠFeatured": 50219, + "ĠFeatures": 19710, + "ĠFeaturing": 51231, + "ĠFeb": 13498, + "ĠFebruary": 7400, + "ĠFecha": 60478, + "ĠFed": 23608, + "ĠFedEx": 85235, + "ĠFeder": 21099, + "ĠFederal": 12137, + "ĠFederation": 27232, + "ĠFedora": 79506, + "ĠFee": 40458, + "ĠFeed": 28870, + "ĠFeedback": 36857, + "ĠFeel": 31733, + "ĠFeeling": 73558, + "ĠFees": 58578, + "ĠFeet": 61189, + "ĠFehler": 91265, + "ĠFeinstein": 84789, + "ĠFel": 26154, + "ĠFeld": 49753, + "ĠFelipe": 81447, + "ĠFelix": 55923, + "ĠFell": 62605, + "ĠFellow": 36846, + "ĠFellowship": 64642, + "ĠFem": 34195, + "ĠFemale": 28638, + "ĠFemin": 56301, + "ĠFemme": 57525, + "ĠFen": 44306, + "ĠFence": 79963, + "ĠFeng": 42658, + "ĠFer": 28462, + "ĠFerd": 82327, + "ĠFerdinand": 92507, + "ĠFerguson": 37376, + "ĠFerm": 98262, + "ĠFern": 25987, + "ĠFernandez": 82626, + "ĠFernando": 50385, + "ĠFerr": 27942, + "ĠFerrari": 45753, + "ĠFerry": 78449, + "ĠFest": 38650, + "ĠFestival": 17322, + "ĠFet": 68037, + "ĠFetch": 22104, + "ĠFetchType": 77182, + "ĠFetish": 98919, + "ĠFever": 83901, + "ĠFew": 42915, + "ĠFey": 88220, + "ĠFi": 39871, + "ĠFiat": 69299, + "ĠFib": 42464, + "ĠFiber": 53627, + "ĠFibonacci": 79683, + "ĠFiction": 42654, + "ĠField": 8601, + "ĠFieldType": 84614, + "ĠFields": 24580, + "ĠFiesta": 93803, + "ĠFif": 18502, + "ĠFifth": 22843, + "ĠFifty": 74920, + "ĠFig": 23095, + "ĠFight": 18629, + "ĠFighter": 41567, + "ĠFighters": 77967, + "ĠFighting": 49317, + "ĠFigure": 19037, + "ĠFigures": 68256, + "ĠFiji": 85096, + "ĠFil": 16402, + "ĠFile": 2887, + "ĠFileAccess": 92843, + "ĠFileInfo": 56189, + "ĠFileInputStream": 40185, + "ĠFileManager": 79288, + "ĠFileMode": 59348, + "ĠFileName": 47157, + "ĠFileNotFoundError": 95841, + "ĠFileNotFoundException": 68299, + "ĠFileOutputStream": 42656, + "ĠFilePath": 62703, + "ĠFileReader": 40733, + "ĠFileStream": 50771, + "ĠFileSystem": 53461, + "ĠFileType": 93238, + "ĠFileUtils": 59974, + "ĠFileWriter": 55821, + "ĠFiled": 83001, + "ĠFilename": 49811, + "ĠFiles": 17378, + "ĠFilip": 41278, + "ĠFilipino": 62417, + "ĠFill": 21979, + "ĠFilm": 16631, + "ĠFilme": 96185, + "ĠFilms": 45464, + "ĠFilter": 12339, + "ĠFiltering": 81531, + "ĠFilters": 45012, + "ĠFin": 5649, + "ĠFinal": 13023, + "ĠFinally": 17375, + "ĠFinals": 45047, + "ĠFinance": 22453, + "ĠFinancial": 17503, + "ĠFinancing": 92548, + "ĠFinch": 95233, + "ĠFind": 7379, + "ĠFindObjectOfType": 91422, + "ĠFinder": 57170, + "ĠFinding": 42797, + "ĠFinds": 50630, + "ĠFine": 30153, + "ĠFinger": 62861, + "ĠFinish": 35533, + "ĠFinished": 52626, + "ĠFinite": 93619, + "ĠFinland": 36255, + "ĠFinn": 34062, + "ĠFinnish": 57853, + "ĠFiona": 88843, + "ĠFior": 99003, + "ĠFir": 79247, + "ĠFire": 6647, + "ĠFirearms": 82517, + "ĠFirebase": 19683, + "ĠFirebaseAuth": 51981, + "ĠFirebaseDatabase": 68474, + "ĠFirebaseFirestore": 92240, + "ĠFired": 74830, + "ĠFirefox": 25929, + "ĠFireplace": 83728, + "ĠFires": 82830, + "ĠFirestore": 94266, + "ĠFirewall": 96231, + "ĠFirm": 37625, + "ĠFirmware": 80830, + "ĠFirst": 5512, + "ĠFirstName": 49467, + "ĠFirstly": 76695, + "ĠFiscal": 62972, + "ĠFischer": 61916, + "ĠFish": 16608, + "ĠFisher": 35504, + "ĠFisheries": 93405, + "ĠFishing": 52308, + "ĠFist": 81419, + "ĠFit": 29890, + "ĠFitness": 35708, + "ĠFits": 78245, + "ĠFitz": 36678, + "ĠFitzgerald": 61214, + "ĠFive": 20924, + "ĠFix": 19716, + "ĠFixed": 20149, + "ĠFixedUpdate": 86602, + "ĠFixes": 85963, + "ĠFixture": 73252, + "ĠFl": 2988, + "ĠFla": 53197, + "ĠFlag": 22666, + "ĠFlags": 33205, + "ĠFlake": 95315, + "ĠFlam": 96857, + "ĠFlame": 48857, + "ĠFlames": 72197, + "ĠFlash": 17263, + "ĠFlask": 28173, + "ĠFlat": 23577, + "ĠFlatButton": 74617, + "ĠFlatten": 85638, + "ĠFlavor": 64999, + "ĠFle": 46690, + "ĠFleet": 43455, + "ĠFlem": 64022, + "ĠFleming": 85197, + "ĠFlesh": 98055, + "ĠFletcher": 68068, + "ĠFlex": 26141, + "ĠFlexible": 52311, + "ĠFlickr": 54358, + "ĠFlight": 26577, + "ĠFlint": 52836, + "ĠFlip": 40284, + "ĠFlo": 19580, + "ĠFloat": 13001, + "ĠFloating": 47512, + "ĠFloatingActionButton": 81248, + "ĠFlood": 56127, + "ĠFloor": 16581, + "ĠFlooring": 77958, + "ĠFloors": 96405, + "ĠFlor": 8780, + "ĠFloral": 90652, + "ĠFlorence": 47506, + "ĠFlores": 72952, + "ĠFlorian": 96056, + "ĠFlorida": 9589, + "ĠFlour": 89028, + "ĠFlow": 22452, + "ĠFlower": 42686, + "ĠFlowers": 51545, + "ĠFloyd": 45799, + "ĠFlu": 60526, + "ĠFluent": 59797, + "ĠFluid": 59596, + "ĠFlush": 57626, + "ĠFlutter": 50795, + "ĠFlux": 69278, + "ĠFly": 20182, + "ĠFlyers": 73734, + "ĠFlying": 46052, + "ĠFlynn": 39386, + "ĠFn": 50182, + "ĠFo": 31657, + "ĠFoam": 76456, + "ĠFocus": 25806, + "ĠFog": 58148, + "ĠFol": 39455, + "ĠFold": 60473, + "ĠFolder": 35843, + "ĠFolding": 91150, + "ĠFoley": 79806, + "ĠFolk": 64237, + "ĠFollow": 11112, + "ĠFollowers": 85550, + "ĠFollowing": 22713, + "ĠFon": 70493, + "ĠFond": 91541, + "ĠFont": 9562, + "ĠFontAwesome": 58911, + "ĠFontStyle": 76227, + "ĠFontWeight": 26706, + "ĠFonts": 61621, + "ĠFoo": 33428, + "ĠFood": 12098, + "ĠFoods": 39060, + "ĠFool": 71535, + "ĠFoot": 15455, + "ĠFootball": 20761, + "ĠFooter": 26477, + "ĠFor": 1752, + "ĠForCanBeConverted": 79269, + "ĠForCanBeConvertedToF": 79270, + "ĠForCanBeConvertedToForeach": 79271, + "ĠForbes": 47281, + "ĠForbidden": 66440, + "ĠForce": 11732, + "ĠForced": 83313, + "ĠForces": 29400, + "ĠFord": 14010, + "ĠFore": 8207, + "ĠForecast": 55675, + "ĠForeign": 19078, + "ĠForeignKey": 78915, + "ĠForest": 19988, + "ĠForever": 54606, + "ĠForex": 36862, + "ĠForg": 69575, + "ĠForge": 57261, + "ĠForget": 60228, + "ĠForgot": 66174, + "ĠForgotten": 94447, + "ĠFork": 47850, + "ĠForm": 3377, + "ĠFormBuilder": 43480, + "ĠFormControl": 27677, + "ĠFormData": 46593, + "ĠFormGroup": 30926, + "ĠFormal": 79367, + "ĠFormat": 15042, + "ĠFormation": 71366, + "ĠFormats": 79348, + "ĠFormatter": 81387, + "ĠFormatting": 89588, + "ĠFormer": 32500, + "ĠForms": 23572, + "ĠFormsModule": 36224, + "ĠFormula": 30822, + "ĠForrest": 83730, + "ĠFors": 49890, + "ĠFort": 11002, + "ĠFortnite": 86420, + "ĠFortress": 71435, + "ĠFortunately": 41436, + "ĠFortune": 45231, + "ĠForty": 84943, + "ĠForum": 17538, + "ĠForums": 44515, + "ĠForward": 22164, + "ĠFoster": 38174, + "ĠFot": 73211, + "ĠFoto": 49342, + "ĠFotos": 66419, + "ĠFou": 94210, + "ĠFound": 12315, + "ĠFoundation": 5007, + "ĠFoundations": 98512, + "ĠFounded": 77711, + "ĠFounder": 54528, + "ĠFountain": 77224, + "ĠFour": 13322, + "ĠFourier": 88954, + "ĠFourth": 35074, + "ĠFowler": 74143, + "ĠFox": 13282, + "ĠFr": 2869, + "ĠFra": 22700, + "ĠFraction": 51893, + "ĠFrag": 58089, + "ĠFrage": 67232, + "ĠFragen": 70777, + "ĠFragment": 19034, + "ĠFram": 65140, + "ĠFrame": 16321, + "ĠFrames": 64597, + "ĠFramework": 23752, + "ĠFran": 30825, + "ĠFranc": 9694, + "ĠFrance": 9625, + "ĠFrances": 42733, + "ĠFrancesco": 88599, + "ĠFrancis": 25127, + "ĠFrancisco": 12879, + "ĠFranco": 43843, + "ĠFrancois": 91371, + "ĠFrank": 9270, + "ĠFranken": 90715, + "ĠFrankfurt": 55184, + "ĠFrankie": 93445, + "ĠFranklin": 18841, + "ĠFranz": 65520, + "ĠFranç": 83839, + "ĠFrançais": 138861, + "ĠFrançois": 84707, + "ĠFraser": 55130, + "ĠFrau": 51743, + "ĠFraud": 65947, + "ĠFrauen": 47610, + "ĠFre": 7577, + "ĠFreak": 80125, + "ĠFred": 27488, + "ĠFreddie": 67262, + "ĠFreddy": 95782, + "ĠFreder": 35432, + "ĠFrederick": 50664, + "ĠFree": 3574, + "ĠFreeBSD": 74560, + "ĠFreed": 86876, + "ĠFreedom": 24327, + "ĠFreel": 79933, + "ĠFreem": 95086, + "ĠFreeman": 49564, + "ĠFreeze": 78592, + "ĠFreight": 90238, + "ĠFrem": 67210, + "ĠFrench": 8585, + "ĠFrequ": 34032, + "ĠFrequency": 42380, + "ĠFrequently": 92620, + "ĠFres": 59650, + "ĠFresh": 29160, + "ĠFresno": 81170, + "ĠFreud": 71534, + "ĠFreund": 77779, + "ĠFri": 30402, + "ĠFrid": 63004, + "ĠFriday": 6602, + "ĠFridays": 79423, + "ĠFried": 30251, + "ĠFriedman": 61358, + "ĠFriedrich": 79374, + "ĠFriend": 11587, + "ĠFriendly": 48417, + "ĠFriends": 22508, + "ĠFriendship": 72069, + "ĠFritz": 89983, + "ĠFrm": 91995, + "ĠFro": 23413, + "ĠFrog": 71419, + "ĠFrom": 5542, + "ĠFront": 14902, + "ĠFrontier": 57318, + "ĠFrost": 41220, + "ĠFrozen": 57746, + "ĠFruit": 43087, + "ĠFry": 52417, + "ĠFs": 84619, + "ĠFt": 44379, + "ĠFu": 29253, + "ĠFuck": 25090, + "ĠFucked": 63562, + "ĠFucking": 65860, + "ĠFuel": 36284, + "ĠFuj": 62270, + "ĠFuji": 87053, + "ĠFuk": 62659, + "ĠFukushima": 88739, + "ĠFul": 47450, + "ĠFull": 8627, + "ĠFullName": 97147, + "ĠFuller": 70469, + "ĠFully": 48156, + "ĠFulton": 88884, + "ĠFun": 16071, + "ĠFunc": 18016, + "ĠFunction": 5712, + "ĠFunctional": 54450, + "ĠFunctions": 23550, + "ĠFunctor": 87632, + "ĠFund": 13190, + "ĠFundamental": 91439, + "ĠFunding": 63715, + "ĠFunds": 60188, + "ĠFuneral": 58349, + "ĠFunk": 39610, + "ĠFunktion": 71065, + "ĠFunny": 52840, + "ĠFur": 49840, + "ĠFurious": 92331, + "ĠFurn": 64472, + "ĠFurniture": 29239, + "ĠFurther": 15533, + "ĠFurthermore": 23405, + "ĠFury": 49379, + "ĠFus": 93300, + "ĠFuse": 73515, + "ĠFusion": 42237, + "ĠFut": 42547, + "ĠFuture": 12498, + "ĠFutures": 76267, + "ĠFuÃŁ": 91062, + "ĠFuÃŁball": 143575, + "ĠFälle": 143053, + "ĠFé": 88150, + "ĠFör": 86467, + "ĠFörder": 142407, + "ĠFührung": 142332, + "ĠFür": 45481, + "ĠG": 479, + "ĠGA": 26341, + "ĠGAL": 83362, + "ĠGAM": 81811, + "ĠGAME": 31336, + "ĠGANG": 96477, + "ĠGAP": 89887, + "ĠGAR": 96591, + "ĠGAS": 89709, + "ĠGB": 18865, + "ĠGBP": 83261, + "ĠGC": 22381, + "ĠGCBO": 82746, + "ĠGCC": 44938, + "ĠGD": 39627, + "ĠGDK": 73886, + "ĠGDP": 29730, + "ĠGDPR": 90948, + "ĠGE": 29857, + "ĠGEN": 42674, + "ĠGENER": 28108, + "ĠGENERAL": 52312, + "ĠGENERATED": 72032, + "ĠGENERIC": 94360, + "ĠGEO": 85109, + "ĠGER": 89221, + "ĠGET": 7890, + "ĠGETGLOBAL": 90854, + "ĠGF": 40026, + "ĠGFP": 60070, + "ĠGFX": 76155, + "ĠGG": 56698, + "ĠGH": 47484, + "ĠGHC": 83716, + "ĠGHz": 51305, + "ĠGI": 48509, + "ĠGIF": 46766, + "ĠGIR": 85009, + "ĠGIS": 82371, + "ĠGIT": 90092, + "ĠGIVEN": 89836, + "ĠGK": 97409, + "ĠGL": 5588, + "ĠGLES": 78648, + "ĠGLFW": 41570, + "ĠGLOBAL": 41287, + "ĠGLUT": 69273, + "ĠGLenum": 52672, + "ĠGLfloat": 46482, + "ĠGLint": 48165, + "ĠGLsizei": 79703, + "ĠGLuint": 43252, + "ĠGM": 19172, + "ĠGMC": 91568, + "ĠGMO": 66220, + "ĠGMT": 24488, + "ĠGN": 41002, + "ĠGNOME": 97586, + "ĠGNU": 4253, + "ĠGNUNET": 99099, + "ĠGO": 12604, + "ĠGOD": 58378, + "ĠGOLD": 72763, + "ĠGOOD": 21722, + "ĠGOODMAN": 94946, + "ĠGOODS": 29666, + "ĠGOOGLE": 90216, + "ĠGOP": 18106, + "ĠGOT": 79909, + "ĠGObject": 79125, + "ĠGP": 23690, + "ĠGPA": 70055, + "ĠGPI": 96537, + "ĠGPIO": 10703, + "ĠGPL": 22942, + "ĠGPLv": 82602, + "ĠGPS": 23344, + "ĠGPU": 22670, + "ĠGPUs": 70403, + "ĠGR": 14773, + "ĠGRA": 64020, + "ĠGRAPH": 68636, + "ĠGRAT": 94241, + "ĠGRE": 61896, + "ĠGREAT": 60993, + "ĠGREEN": 53559, + "ĠGRID": 64927, + "ĠGROUP": 26870, + "ĠGS": 35270, + "ĠGSL": 97368, + "ĠGSM": 67555, + "ĠGST": 32834, + "ĠGT": 11911, + "ĠGTA": 58745, + "ĠGTK": 37379, + "ĠGTX": 33940, + "ĠGU": 45272, + "ĠGUI": 16435, + "ĠGUIContent": 82520, + "ĠGUID": 30221, + "ĠGUIDATA": 91388, + "ĠGUIDE": 94657, + "ĠGUILayout": 46429, + "ĠGUIStyle": 98550, + "ĠGURL": 93419, + "ĠGV": 80800, + "ĠGW": 41253, + "ĠGX": 62052, + "ĠGa": 18373, + "ĠGab": 23732, + "ĠGabri": 70714, + "ĠGabriel": 38743, + "ĠGad": 70942, + "ĠGael": 88638, + "ĠGaga": 80574, + "ĠGain": 49709, + "ĠGaines": 82679, + "ĠGal": 10620, + "ĠGala": 75441, + "ĠGalactic": 75217, + "ĠGalaxy": 19662, + "ĠGale": 78232, + "ĠGall": 24877, + "ĠGallagher": 77987, + "ĠGallery": 19295, + "ĠGallup": 79801, + "ĠGam": 29979, + "ĠGamb": 66789, + "ĠGambling": 85343, + "ĠGame": 4050, + "ĠGameController": 96896, + "ĠGameManager": 48714, + "ĠGameObject": 10705, + "ĠGameState": 47987, + "ĠGameplay": 85898, + "ĠGamer": 75659, + "ĠGames": 11610, + "ĠGaming": 30462, + "ĠGamma": 57682, + "ĠGan": 49081, + "ĠGand": 38321, + "ĠGandhi": 47894, + "ĠGang": 34417, + "ĠGap": 57915, + "ĠGar": 12193, + "ĠGarage": 38896, + "ĠGarc": 29314, + "ĠGarcia": 37710, + "ĠGarcÃŃa": 84241, + "ĠGard": 22443, + "ĠGarden": 19021, + "ĠGardens": 42043, + "ĠGardner": 56629, + "ĠGareth": 87047, + "ĠGarland": 78974, + "ĠGarlic": 94725, + "ĠGarmin": 88169, + "ĠGarn": 64032, + "ĠGarner": 85988, + "ĠGarr": 94157, + "ĠGarrett": 57616, + "ĠGarrison": 98486, + "ĠGarten": 93912, + "ĠGary": 23826, + "ĠGas": 20854, + "ĠGast": 62642, + "ĠGat": 72113, + "ĠGate": 29243, + "ĠGates": 35493, + "ĠGateway": 39378, + "ĠGather": 48995, + "ĠGathering": 75585, + "ĠGauge": 72060, + "ĠGaul": 94526, + "ĠGauss": 93216, + "ĠGaussian": 48568, + "ĠGaut": 95054, + "ĠGavin": 63395, + "ĠGaw": 85538, + "ĠGay": 20677, + "ĠGaz": 43292, + "ĠGaza": 28662, + "ĠGazette": 74582, + "ĠGda": 141844, + "ĠGdaÅĦsk": 141845, + "ĠGdk": 75540, + "ĠGe": 4229, + "ĠGeForce": 37520, + "ĠGear": 27503, + "ĠGeb": 44336, + "ĠGebä": 142417, + "ĠGebäude": 142418, + "ĠGecko": 42599, + "ĠGed": 64477, + "ĠGee": 95597, + "ĠGeek": 70365, + "ĠGef": 55857, + "ĠGefühl": 141176, + "ĠGeg": 98863, + "ĠGeh": 73580, + "ĠGeile": 76404, + "ĠGel": 44382, + "ĠGeld": 51239, + "ĠGeli": 143571, + "ĠGeliÅŁ": 143572, + "ĠGem": 31145, + "ĠGeme": 77775, + "ĠGemini": 85711, + "ĠGems": 81072, + "ĠGen": 9316, + "ĠGender": 28217, + "ĠGene": 24022, + "ĠGener": 2607, + "ĠGeneral": 3251, + "ĠGenerally": 43393, + "ĠGenerate": 19813, + "ĠGenerated": 30488, + "ĠGenerates": 52492, + "ĠGenerating": 96454, + "ĠGeneration": 23470, + "ĠGenerationType": 55700, + "ĠGenerator": 28358, + "ĠGeneric": 21281, + "ĠGenesis": 40788, + "ĠGenetic": 74126, + "ĠGenetics": 83286, + "ĠGeneva": 44245, + "ĠGenius": 77801, + "ĠGenome": 81817, + "ĠGenre": 40295, + "ĠGent": 35723, + "ĠGentle": 73469, + "ĠGenuine": 82882, + "ĠGenç": 138734, + "ĠGeo": 31910, + "ĠGeoff": 50430, + "ĠGeoffrey": 88139, + "ĠGeographic": 65442, + "ĠGeography": 83325, + "ĠGeological": 79750, + "ĠGeometry": 38918, + "ĠGeorg": 13326, + "ĠGeorge": 9857, + "ĠGeorges": 94259, + "ĠGeorgetown": 64939, + "ĠGeorgia": 15893, + "ĠGeorgian": 86442, + "ĠGer": 19929, + "ĠGerald": 54257, + "ĠGerard": 81118, + "ĠGerm": 36735, + "ĠGerman": 5938, + "ĠGermans": 40937, + "ĠGermany": 9856, + "ĠGerr": 85865, + "ĠGerry": 90987, + "ĠGerät": 136800, + "ĠGerçek": 143033, + "ĠGes": 20404, + "ĠGesch": 35346, + "ĠGeschichte": 77930, + "ĠGeschä": 137193, + "ĠGeschäfts": 137194, + "ĠGespr": 142126, + "ĠGespräch": 142127, + "ĠGest": 53899, + "ĠGesture": 49864, + "ĠGestureDetector": 63132, + "ĠGet": 2126, + "ĠGetAll": 45587, + "ĠGetById": 92085, + "ĠGetComponent": 27750, + "ĠGetCurrent": 94098, + "ĠGetData": 75016, + "ĠGetEnumerator": 46479, + "ĠGetHashCode": 81678, + "ĠGetLastError": 79362, + "ĠGetMessage": 94229, + "ĠGetName": 92788, + "ĠGetString": 97921, + "ĠGetType": 66859, + "ĠGetUser": 85937, + "ĠGetValue": 68921, + "ĠGets": 10310, + "ĠGetter": 56130, + "ĠGetting": 24515, + "ĠGetty": 20530, + "ĠGew": 42678, + "ĠGh": 23908, + "ĠGhana": 47568, + "ĠGhost": 25044, + "ĠGi": 15392, + "ĠGian": 78604, + "ĠGiant": 40741, + "ĠGiants": 29735, + "ĠGib": 28379, + "ĠGibbs": 83551, + "ĠGibraltar": 98803, + "ĠGibson": 43422, + "ĠGift": 22062, + "ĠGifts": 55960, + "ĠGig": 44134, + "ĠGil": 20793, + "ĠGilbert": 44992, + "ĠGiles": 84831, + "ĠGill": 32726, + "ĠGilles": 76124, + "ĠGim": 85671, + "ĠGin": 50583, + "ĠGina": 76212, + "ĠGing": 74746, + "ĠGinger": 67294, + "ĠGingrich": 86393, + "ĠGinny": 93154, + "ĠGins": 85729, + "ĠGio": 66665, + "ĠGiov": 60578, + "ĠGiovanni": 75852, + "ĠGir": 46935, + "ĠGiriÅŁ": 140666, + "ĠGirl": 11363, + "ĠGirlfriend": 98493, + "ĠGirls": 20065, + "ĠGit": 21120, + "ĠGitHub": 32095, + "ĠGithub": 48923, + "ĠGiul": 62551, + "ĠGiuliani": 88050, + "ĠGive": 20678, + "ĠGiven": 16246, + "ĠGives": 68913, + "ĠGiving": 57851, + "ĠGiz": 66288, + "ĠGiá": 131310, + "ĠGiám": 137295, + "ĠGiáo": 133218, + "ĠGiải": 135816, + "ĠGiấy": 140288, + "ĠGiỼi": 138290, + "ĠGl": 8280, + "ĠGlacier": 95386, + "ĠGlad": 51641, + "ĠGladiator": 95128, + "ĠGlam": 86713, + "ĠGlas": 37832, + "ĠGlasgow": 42897, + "ĠGlass": 20734, + "ĠGlasses": 98266, + "ĠGle": 71397, + "ĠGlen": 39961, + "ĠGlenn": 39108, + "ĠGlide": 56026, + "ĠGlo": 24373, + "ĠGlob": 62288, + "ĠGlobal": 7962, + "ĠGlobalKey": 85358, + "ĠGlobals": 50176, + "ĠGlobe": 40810, + "ĠGlock": 98980, + "ĠGloria": 74993, + "ĠGlory": 58161, + "ĠGloss": 66042, + "ĠGlouce": 86486, + "ĠGlover": 94083, + "ĠGloves": 67454, + "ĠGlow": 87348, + "ĠGluten": 94912, + "ĠGly": 78083, + "ĠGlyph": 87413, + "ĠGlück": 135972, + "ĠGmail": 60946, + "ĠGmbH": 31477, + "ĠGn": 95151, + "ĠGo": 5994, + "ĠGoa": 85522, + "ĠGoal": 39947, + "ĠGoals": 54193, + "ĠGoat": 92772, + "ĠGobierno": 84072, + "ĠGoblin": 80164, + "ĠGod": 4264, + "ĠGoddess": 60685, + "ĠGods": 43775, + "ĠGodzilla": 86996, + "ĠGoes": 60161, + "ĠGoing": 34871, + "ĠGoku": 76036, + "ĠGol": 54579, + "ĠGold": 7421, + "ĠGoldberg": 71575, + "ĠGolden": 17809, + "ĠGoldman": 47328, + "ĠGolf": 27033, + "ĠGomez": 62298, + "ĠGon": 79706, + "ĠGone": 55255, + "ĠGong": 97375, + "ĠGonz": 32455, + "ĠGonzalez": 51086, + "ĠGood": 7684, + "ĠGoodman": 70392, + "ĠGoods": 41595, + "ĠGoogle": 5085, + "ĠGoose": 81107, + "ĠGor": 46147, + "ĠGord": 98334, + "ĠGordon": 25867, + "ĠGore": 48525, + "ĠGorgeous": 86961, + "ĠGors": 97674, + "ĠGos": 62381, + "ĠGospel": 43368, + "ĠGot": 24528, + "ĠGoth": 31937, + "ĠGotham": 73882, + "ĠGothic": 43101, + "ĠGott": 68009, + "ĠGould": 84331, + "ĠGov": 24424, + "ĠGover": 93884, + "ĠGovern": 7955, + "ĠGovernance": 80589, + "ĠGovernment": 10212, + "ĠGovernments": 87699, + "ĠGovernor": 23358, + "ĠGovernors": 89989, + "ĠGow": 92471, + "ĠGr": 2825, + "ĠGra": 39013, + "ĠGrab": 36194, + "ĠGrace": 31071, + "ĠGrad": 21794, + "ĠGrade": 23812, + "ĠGrades": 95220, + "ĠGradient": 53107, + "ĠGraduate": 43725, + "ĠGraf": 74629, + "ĠGraham": 25124, + "ĠGrain": 74274, + "ĠGram": 28837, + "ĠGrammar": 61977, + "ĠGrammy": 73579, + "ĠGran": 26524, + "ĠGrand": 10304, + "ĠGrande": 36523, + "ĠGrandma": 97968, + "ĠGranite": 64494, + "ĠGranny": 88140, + "ĠGrant": 23736, + "ĠGranted": 72843, + "ĠGrants": 72801, + "ĠGrape": 79529, + "ĠGraph": 12165, + "ĠGraphQL": 35087, + "ĠGraphic": 43951, + "ĠGraphics": 20296, + "ĠGrass": 39946, + "ĠGrat": 42081, + "ĠGratis": 23775, + "ĠGratuit": 53509, + "ĠGrave": 73742, + "ĠGraves": 79058, + "ĠGravity": 47490, + "ĠGray": 23366, + "ĠGraz": 92930, + "ĠGre": 13532, + "ĠGreat": 8513, + "ĠGreater": 32281, + "ĠGreatest": 61812, + "ĠGree": 20422, + "ĠGreece": 24427, + "ĠGreek": 17860, + "ĠGreeks": 60680, + "ĠGreen": 7840, + "ĠGreene": 58074, + "ĠGreenland": 72678, + "ĠGreens": 41591, + "ĠGreenville": 92554, + "ĠGreenwich": 85838, + "ĠGreenwood": 89974, + "ĠGreg": 16046, + "ĠGregg": 80696, + "ĠGregory": 42969, + "ĠGren": 38124, + "ĠGret": 87552, + "ĠGrey": 25685, + "ĠGri": 62904, + "ĠGrid": 10587, + "ĠGridBagConstraints": 51923, + "ĠGridLayout": 80079, + "ĠGridView": 40935, + "ĠGriff": 27794, + "ĠGriffin": 40396, + "ĠGriffith": 63540, + "ĠGrill": 49848, + "ĠGrim": 73304, + "ĠGrimm": 84792, + "ĠGrinder": 68500, + "ĠGrinding": 38612, + "ĠGrip": 78821, + "ĠGro": 17888, + "ĠGrocery": 95581, + "ĠGron": 86230, + "ĠGros": 69852, + "ĠGross": 42807, + "ĠGround": 25413, + "ĠGroup": 5737, + "ĠGroupLayout": 91843, + "ĠGroups": 34580, + "ĠGrove": 40134, + "ĠGrow": 46560, + "ĠGrowing": 59680, + "ĠGrowth": 33771, + "ĠGroÃŁ": 133661, + "ĠGrund": 47227, + "ĠGrupo": 74907, + "ĠGry": 98448, + "ĠGrÃ¶ÃŁe": 137208, + "ĠGrü": 94063, + "ĠGründe": 140917, + "ĠGson": 38394, + "ĠGst": 80320, + "ĠGtk": 27296, + "ĠGtkWidget": 66108, + "ĠGu": 4570, + "ĠGuam": 91688, + "ĠGuang": 64437, + "ĠGuantanamo": 98300, + "ĠGuar": 39000, + "ĠGuarantee": 64676, + "ĠGuaranteed": 80437, + "ĠGuard": 12262, + "ĠGuardian": 24454, + "ĠGuardians": 61731, + "ĠGuards": 69285, + "ĠGuatemala": 61978, + "ĠGuerr": 79783, + "ĠGuerrero": 97736, + "ĠGuess": 54279, + "ĠGuest": 26215, + "ĠGuests": 61582, + "ĠGui": 49998, + "ĠGuid": 12157, + "ĠGuidance": 81561, + "ĠGuide": 12712, + "ĠGuidelines": 47428, + "ĠGuides": 59445, + "ĠGuil": 92568, + "ĠGuild": 32492, + "ĠGuill": 73638, + "ĠGuinea": 47803, + "ĠGuinness": 88459, + "ĠGuitar": 46659, + "ĠGujar": 55631, + "ĠGujarat": 61853, + "ĠGul": 58889, + "ĠGulf": 26847, + "ĠGum": 74489, + "ĠGun": 21707, + "ĠGund": 69380, + "ĠGundam": 87376, + "ĠGunn": 65149, + "ĠGuns": 59465, + "ĠGupta": 73698, + "ĠGur": 60933, + "ĠGuru": 61979, + "ĠGus": 81379, + "ĠGust": 48620, + "ĠGut": 51583, + "ĠGutenberg": 51586, + "ĠGuth": 91755, + "ĠGuy": 25273, + "ĠGuys": 50200, + "ĠGuzzle": 96833, + "ĠGwen": 94784, + "ĠGy": 54915, + "ĠGym": 45531, + "ĠGä": 94589, + "ĠGäste": 136088, + "ĠGòn": 136942, + "ĠGór": 137785, + "ĠGö": 82146, + "ĠGör": 132906, + "ĠGü": 80163, + "ĠGül": 135210, + "ĠGün": 130022, + "ĠGünc": 141677, + "ĠGüncelle": 141678, + "ĠGüncelleme": 141679, + "ĠGüzel": 141378, + "ĠGÅĤ": 141783, + "ĠGÅĤówn": 141784, + "ĠH": 472, + "ĠHA": 41316, + "ĠHACK": 95123, + "ĠHAL": 20533, + "ĠHALF": 95322, + "ĠHAND": 52469, + "ĠHANDLE": 45273, + "ĠHAPP": 86777, + "ĠHAR": 86488, + "ĠHARD": 69902, + "ĠHAS": 35748, + "ĠHASH": 58857, + "ĠHAVE": 18590, + "ĠHB": 39701, + "ĠHBO": 42878, + "ĠHC": 45828, + "ĠHCI": 84335, + "ĠHD": 12169, + "ĠHDC": 96317, + "ĠHDD": 68523, + "ĠHDF": 80240, + "ĠHDMI": 41050, + "ĠHDR": 54391, + "ĠHE": 11685, + "ĠHEAD": 33080, + "ĠHEADER": 39665, + "ĠHEALTH": 72625, + "ĠHEAP": 99091, + "ĠHEIGHT": 41900, + "ĠHEL": 37657, + "ĠHELP": 55471, + "ĠHER": 15645, + "ĠHERE": 19249, + "ĠHERO": 86617, + "ĠHEX": 72221, + "ĠHF": 50558, + "ĠHG": 69397, + "ĠHH": 24756, + "ĠHI": 33411, + "ĠHID": 61415, + "ĠHIGH": 37617, + "ĠHIM": 88894, + "ĠHIP": 80296, + "ĠHIS": 65126, + "ĠHISTORY": 61779, + "ĠHIT": 78403, + "ĠHIV": 22664, + "ĠHK": 42217, + "ĠHL": 52487, + "ĠHLS": 92177, + "ĠHM": 45414, + "ĠHMAC": 95927, + "ĠHMS": 96070, + "ĠHO": 30250, + "ĠHOL": 95622, + "ĠHOLD": 17542, + "ĠHOLDER": 60693, + "ĠHOLDERS": 20620, + "ĠHOME": 40567, + "ĠHOR": 83566, + "ĠHOST": 44578, + "ĠHOT": 53373, + "ĠHOUR": 95964, + "ĠHOUSE": 68361, + "ĠHOW": 23532, + "ĠHOWEVER": 29316, + "ĠHP": 12200, + "ĠHPV": 73369, + "ĠHQ": 45255, + "ĠHR": 22299, + "ĠHRESULT": 29603, + "ĠHS": 33414, + "ĠHSV": 80311, + "ĠHT": 5260, + "ĠHTC": 48267, + "ĠHTML": 9308, + "ĠHTMLElement": 43899, + "ĠHTTP": 10130, + "ĠHTTPS": 61044, + "ĠHUD": 48596, + "ĠHUGE": 72656, + "ĠHV": 48217, + "ĠHVAC": 65333, + "ĠHW": 26670, + "ĠHWND": 60384, + "ĠHY": 63252, + "ĠHa": 14101, + "ĠHab": 28876, + "ĠHabit": 94078, + "ĠHabitat": 98588, + "ĠHack": 34982, + "ĠHacker": 88065, + "ĠHad": 23864, + "ĠHag": 66539, + "ĠHague": 84926, + "ĠHai": 62682, + "ĠHair": 25697, + "ĠHairst": 88315, + "ĠHait": 42402, + "ĠHaiti": 55190, + "ĠHaj": 83426, + "ĠHak": 71298, + "ĠHal": 19851, + "ĠHale": 66151, + "ĠHaley": 67711, + "ĠHalf": 25839, + "ĠHalifax": 69669, + "ĠHall": 10926, + "ĠHalloween": 26613, + "ĠHalo": 45249, + "ĠHam": 9582, + "ĠHamas": 40641, + "ĠHamburg": 50462, + "ĠHamilton": 23176, + "ĠHamm": 73347, + "ĠHammer": 36894, + "ĠHammond": 64343, + "ĠHamp": 28414, + "ĠHampshire": 30897, + "ĠHampton": 63674, + "ĠHan": 20644, + "ĠHancock": 80267, + "ĠHand": 8536, + "ĠHandbook": 48824, + "ĠHandle": 13760, + "ĠHandler": 19954, + "ĠHandlers": 95563, + "ĠHandles": 21008, + "ĠHandling": 55713, + "ĠHands": 42296, + "ĠHandy": 71559, + "ĠHang": 40775, + "ĠHanging": 89555, + "ĠHank": 54661, + "ĠHann": 27729, + "ĠHanna": 84763, + "ĠHannah": 47261, + "ĠHannity": 89693, + "ĠHans": 24807, + "ĠHansen": 57153, + "ĠHanson": 75213, + "ĠHao": 93709, + "ĠHapp": 34700, + "ĠHappiness": 78144, + "ĠHappy": 23355, + "ĠHar": 5227, + "ĠHaram": 84542, + "ĠHarbor": 39182, + "ĠHarbour": 62007, + "ĠHard": 11232, + "ĠHardcore": 59550, + "ĠHardcover": 97527, + "ĠHarden": 89793, + "ĠHarding": 96493, + "ĠHardware": 36765, + "ĠHardy": 57274, + "ĠHari": 97445, + "ĠHarlem": 82752, + "ĠHarley": 51676, + "ĠHarm": 90971, + "ĠHarmon": 39659, + "ĠHarmony": 65380, + "ĠHarness": 93899, + "ĠHarold": 52115, + "ĠHarper": 32007, + "ĠHarr": 15234, + "ĠHarris": 21071, + "ĠHarrison": 35527, + "ĠHarry": 13928, + "ĠHart": 22900, + "ĠHartford": 72120, + "ĠHarvard": 24951, + "ĠHarvest": 55835, + "ĠHarvey": 33322, + "ĠHas": 11443, + "ĠHasan": 98972, + "ĠHash": 6531, + "ĠHashMap": 10528, + "ĠHashSet": 18931, + "ĠHashtable": 49224, + "ĠHaskell": 62176, + "ĠHass": 42022, + "ĠHassan": 59677, + "ĠHast": 59507, + "ĠHastings": 78744, + "ĠHat": 21341, + "ĠHatch": 58894, + "ĠHate": 65812, + "ĠHath": 90568, + "ĠHats": 90215, + "ĠHaupt": 89931, + "ĠHaus": 47375, + "ĠHaut": 83234, + "ĠHav": 55108, + "ĠHavana": 85875, + "ĠHave": 12243, + "ĠHaven": 41614, + "ĠHaving": 20035, + "ĠHaw": 12611, + "ĠHawai": 21437, + "ĠHawaii": 27521, + "ĠHawaiian": 58003, + "ĠHawk": 55323, + "ĠHawkins": 69587, + "ĠHawks": 56612, + "ĠHawth": 85783, + "ĠHay": 17798, + "ĠHayden": 76185, + "ĠHayes": 52422, + "ĠHayward": 92205, + "ĠHaz": 31495, + "ĠHazard": 68774, + "ĠHazel": 75379, + "ĠHd": 64916, + "ĠHe": 1260, + "ĠHead": 11203, + "ĠHeader": 12104, + "ĠHeaderComponent": 93112, + "ĠHeaders": 21426, + "ĠHeading": 51349, + "ĠHeadquarters": 69636, + "ĠHeads": 70507, + "ĠHeal": 81030, + "ĠHealing": 48641, + "ĠHealth": 6267, + "ĠHealthcare": 38335, + "ĠHealthy": 43354, + "ĠHeap": 47307, + "ĠHear": 56099, + "ĠHeard": 94108, + "ĠHearing": 65450, + "ĠHeart": 17965, + "ĠHearth": 84893, + "ĠHearts": 52776, + "ĠHeat": 26070, + "ĠHeater": 87428, + "ĠHeath": 47462, + "ĠHeather": 46263, + "ĠHeating": 67854, + "ĠHeaven": 31350, + "ĠHeavenly": 87050, + "ĠHeavy": 28101, + "ĠHeb": 84319, + "ĠHebrew": 36266, + "ĠHeck": 72717, + "ĠHector": 88040, + "ĠHed": 74163, + "ĠHedge": 79021, + "ĠHeg": 70117, + "ĠHeidi": 81836, + "ĠHeight": 21432, + "ĠHeights": 39403, + "ĠHein": 63682, + "ĠHel": 15806, + "ĠHeld": 82431, + "ĠHelen": 42781, + "ĠHelena": 71946, + "ĠHell": 23887, + "ĠHeller": 92793, + "ĠHello": 21927, + "ĠHelloWorld": 90071, + "ĠHelm": 62042, + "ĠHelmet": 66529, + "ĠHelp": 11479, + "ĠHelper": 12906, + "ĠHelpers": 48314, + "ĠHelpful": 46554, + "ĠHelping": 90701, + "ĠHelps": 88516, + "ĠHelsinki": 76334, + "ĠHelvetica": 59546, + "ĠHem": 32824, + "ĠHemisphere": 95778, + "ĠHemp": 56208, + "ĠHen": 13070, + "ĠHence": 31040, + "ĠHend": 29494, + "ĠHenderson": 44577, + "ĠHenri": 73210, + "ĠHenrik": 98087, + "ĠHenry": 17599, + "ĠHentai": 92042, + "ĠHep": 60145, + "ĠHer": 6252, + "ĠHera": 82557, + "ĠHerald": 41669, + "ĠHerb": 73856, + "ĠHerbal": 86872, + "ĠHerbert": 57363, + "ĠHerc": 77152, + "ĠHercules": 93534, + "ĠHere": 5692, + "ĠHeritage": 33143, + "ĠHerm": 31465, + "ĠHerman": 60238, + "ĠHermes": 93963, + "ĠHermione": 53499, + "ĠHern": 40456, + "ĠHernandez": 53421, + "ĠHero": 16498, + "ĠHeroes": 36999, + "ĠHerr": 42250, + "ĠHerrera": 96008, + "ĠHers": 64366, + "ĠHerz": 52639, + "ĠHess": 98705, + "ĠHet": 33166, + "ĠHew": 59250, + "ĠHex": 27228, + "ĠHexatrigesimal": 78640, + "ĠHey": 27553, + "ĠHezbollah": 61894, + "ĠHi": 21018, + "ĠHib": 91736, + "ĠHibernate": 63448, + "ĠHick": 77928, + "ĠHicks": 73505, + "ĠHidden": 34242, + "ĠHide": 25590, + "ĠHier": 38747, + "ĠHiggins": 82097, + "ĠHigh": 5124, + "ĠHigher": 34221, + "ĠHighest": 59613, + "ĠHighland": 64087, + "ĠHighlander": 48630, + "ĠHighlands": 82925, + "ĠHighlight": 55994, + "ĠHighlights": 52200, + "ĠHighly": 52385, + "ĠHighway": 28766, + "ĠHij": 67990, + "ĠHil": 38596, + "ĠHilfe": 92660, + "ĠHill": 8108, + "ĠHillary": 15033, + "ĠHills": 24919, + "ĠHilton": 52765, + "ĠHim": 20426, + "ĠHimal": 75338, + "ĠHimself": 65469, + "ĠHin": 49633, + "ĠHind": 19825, + "ĠHindered": 94389, + "ĠHinderedRotor": 95065, + "ĠHindi": 43980, + "ĠHindu": 35042, + "ĠHindus": 81629, + "ĠHint": 51830, + "ĠHip": 44528, + "ĠHipp": 74363, + "ĠHir": 79635, + "ĠHire": 61628, + "ĠHiring": 88197, + "ĠHiro": 56292, + "ĠHis": 5301, + "ĠHispan": 72647, + "ĠHispanic": 40885, + "ĠHispanics": 83832, + "ĠHist": 65905, + "ĠHistogram": 82138, + "ĠHistor": 21683, + "ĠHistoric": 50787, + "ĠHistorical": 40043, + "ĠHistory": 11099, + "ĠHit": 15882, + "ĠHitch": 70364, + "ĠHitler": 30554, + "ĠHits": 53651, + "ĠHive": 68178, + "ĠHiá»ĩn": 132743, + "ĠHiá»ĩp": 140923, + "ĠHmm": 88190, + "ĠHo": 17275, + "ĠHob": 36776, + "ĠHobby": 72592, + "ĠHoch": 67775, + "ĠHockey": 40242, + "ĠHod": 50934, + "ĠHodg": 82884, + "ĠHoe": 86369, + "ĠHof": 71712, + "ĠHoff": 42722, + "ĠHoffman": 61882, + "ĠHog": 49769, + "ĠHogan": 57271, + "ĠHogwarts": 77153, + "ĠHok": 95165, + "ĠHol": 15696, + "ĠHold": 22749, + "ĠHolden": 86385, + "ĠHolder": 40333, + "ĠHolding": 54677, + "ĠHoldings": 53542, + "ĠHolds": 65984, + "ĠHole": 50929, + "ĠHoliday": 31669, + "ĠHolidays": 67003, + "ĠHoll": 24504, + "ĠHolland": 37815, + "ĠHollande": 93331, + "ĠHollow": 48693, + "ĠHolly": 52325, + "ĠHollywood": 17236, + "ĠHolmes": 39301, + "ĠHolocaust": 43098, + "ĠHolt": 58906, + "ĠHoly": 18710, + "ĠHom": 13222, + "ĠHome": 5379, + "ĠHomeComponent": 56852, + "ĠHomeController": 60878, + "ĠHomePage": 44137, + "ĠHomeland": 43704, + "ĠHomemade": 84808, + "ĠHomepage": 67435, + "ĠHomer": 65705, + "ĠHomes": 36564, + "ĠHomework": 82713, + "ĠHomo": 83397, + "ĠHon": 16549, + "ĠHond": 58115, + "ĠHonda": 28887, + "ĠHonduras": 70315, + "ĠHonest": 53184, + "ĠHonestly": 70947, + "ĠHoney": 39615, + "ĠHong": 19180, + "ĠHonolulu": 81540, + "ĠHonor": 41944, + "ĠHonour": 95814, + "ĠHood": 35343, + "ĠHook": 28171, + "ĠHooks": 38430, + "ĠHoover": 72309, + "ĠHop": 25553, + "ĠHope": 17758, + "ĠHopefully": 37894, + "ĠHopkins": 44216, + "ĠHor": 14741, + "ĠHorde": 70382, + "ĠHorizon": 55451, + "ĠHorizontal": 39909, + "ĠHorm": 91108, + "ĠHorn": 26114, + "ĠHornets": 95327, + "ĠHorny": 68254, + "ĠHorror": 51712, + "ĠHorse": 33292, + "ĠHort": 96109, + "ĠHorton": 94595, + "ĠHos": 63990, + "ĠHose": 55584, + "ĠHosp": 56933, + "ĠHospital": 15079, + "ĠHospitality": 87605, + "ĠHospitals": 84297, + "ĠHost": 16102, + "ĠHosting": 68423, + "ĠHot": 8007, + "ĠHotel": 14552, + "ĠHotels": 44771, + "ĠHou": 59600, + "ĠHour": 30833, + "ĠHours": 29092, + "ĠHouse": 4678, + "ĠHousehold": 75973, + "ĠHouses": 57659, + "ĠHousing": 30333, + "ĠHouston": 16003, + "ĠHover": 85569, + "ĠHow": 2585, + "ĠHoward": 19870, + "ĠHowe": 85531, + "ĠHowell": 96179, + "ĠHowever": 4354, + "ĠHoy": 79692, + "ĠHoÃłng": 130597, + "ĠHp": 87833, + "ĠHtml": 19200, + "ĠHtmlWebpackPlugin": 92415, + "ĠHttp": 4823, + "ĠHttpClient": 18674, + "ĠHttpClientModule": 53649, + "ĠHttpContext": 43555, + "ĠHttpHeaders": 43572, + "ĠHttpMethod": 77765, + "ĠHttpNotFound": 73568, + "ĠHttpRequest": 44049, + "ĠHttpResponse": 17580, + "ĠHttpResponseMessage": 68704, + "ĠHttpResponseRedirect": 53030, + "ĠHttpServlet": 46219, + "ĠHttpServletRequest": 49594, + "ĠHttpServletResponse": 24988, + "ĠHttpSession": 75568, + "ĠHttpStatus": 39003, + "ĠHttpStatusCode": 49339, + "ĠHttpStatusCodeResult": 80202, + "ĠHttpURLConnection": 94327, + "ĠHu": 21793, + "ĠHuang": 58409, + "ĠHuawei": 50607, + "ĠHub": 26538, + "ĠHubb": 77593, + "ĠHubbard": 79900, + "ĠHuck": 78364, + "ĠHuckabee": 96084, + "ĠHud": 92905, + "ĠHudson": 38153, + "ĠHue": 65951, + "ĠHuff": 35041, + "ĠHuffPost": 93454, + "ĠHuffington": 63641, + "ĠHuffman": 88497, + "ĠHuge": 50783, + "ĠHugh": 29106, + "ĠHughes": 40311, + "ĠHugo": 51094, + "ĠHulk": 60846, + "ĠHull": 57386, + "ĠHulu": 77392, + "ĠHum": 19858, + "ĠHuman": 11097, + "ĠHumanities": 86245, + "ĠHumanity": 82595, + "ĠHumans": 64994, + "ĠHumb": 87309, + "ĠHumph": 70925, + "ĠHun": 14494, + "ĠHund": 97914, + "ĠHundred": 87270, + "ĠHundreds": 72336, + "ĠHung": 25523, + "ĠHungarian": 56769, + "ĠHungary": 42776, + "ĠHunger": 66536, + "ĠHunt": 26592, + "ĠHunter": 23133, + "ĠHunters": 81888, + "ĠHunting": 44254, + "ĠHuntington": 74751, + "ĠHunts": 70626, + "ĠHur": 20995, + "ĠHurricane": 37101, + "ĠHurricanes": 88503, + "ĠHurt": 86249, + "ĠHus": 27897, + "ĠHusband": 73359, + "ĠHuss": 78981, + "ĠHussein": 60528, + "ĠHust": 85450, + "ĠHut": 66313, + "ĠHutch": 59687, + "ĠHutchinson": 98971, + "ĠHv": 91987, + "ĠHwy": 78977, + "ĠHy": 10112, + "ĠHybrid": 49627, + "ĠHyde": 64339, + "ĠHyderabad": 68447, + "ĠHydra": 86861, + "ĠHydraulic": 80004, + "ĠHydro": 39502, + "ĠHyp": 38415, + "ĠHyper": 32732, + "ĠHyundai": 59840, + "ĠHz": 36092, + "ĠHá": 135611, + "ĠHã": 132546, + "ĠHãy": 132547, + "ĠHä": 90593, + "ĠHình": 137515, + "ĠHòa": 134315, + "ĠHóa": 140304, + "ĠHôm": 143958, + "ĠHö": 70264, + "ĠHöhe": 136142, + "ĠHü": 134883, + "ĠHÃł": 128521, + "ĠHÃłn": 130442, + "ĠHÃłng": 137327, + "ĠHÆ¡n": 142121, + "ĠHưng": 139203, + "ĠHương": 139177, + "ĠHÆ°á»Ľng": 136745, + "ĠHạ": 132922, + "ĠHải": 131039, + "ĠHữu": 143965, + "ĠHá»ĩ": 138907, + "ĠHá»į": 136935, + "ĠHá»įc": 133907, + "ĠHá»ĵ": 129785, + "ĠHá»ĵng": 135196, + "ĠHá»Ļi": 129935, + "ĠI": 358, + "ĠIA": 43090, + "ĠIAM": 88186, + "ĠIActionResult": 34514, + "ĠIB": 12102, + "ĠIBM": 27922, + "ĠIBOutlet": 22347, + "ĠIC": 19288, + "ĠICC": 58232, + "ĠICE": 40563, + "ĠICMP": 83988, + "ĠICO": 46536, + "ĠICON": 59324, + "ĠICT": 74202, + "ĠICU": 83915, + "ĠICollection": 36558, + "ĠICommand": 49966, + "ĠIConfiguration": 64490, + "ĠID": 3034, + "ĠIDC": 25000, + "ĠIDD": 80883, + "ĠIDE": 27975, + "ĠIDEA": 72045, + "ĠIDENT": 72754, + "ĠIDEOGRAPH": 96602, + "ĠIDF": 71555, + "ĠIDM": 81455, + "ĠIDR": 96654, + "ĠIDS": 39197, + "ĠIDX": 95937, + "ĠIData": 75876, + "ĠIDb": 85756, + "ĠIDictionary": 59331, + "ĠIDirect": 96632, + "ĠIDisposable": 53889, + "ĠIDs": 28360, + "ĠIE": 18160, + "ĠIEEE": 39035, + "ĠIEntity": 91458, + "ĠIEnumerable": 13050, + "ĠIEnumerator": 27952, + "ĠIF": 11551, + "ĠIG": 46180, + "ĠIGN": 49839, + "ĠIGNORE": 81290, + "ĠIH": 45311, + "ĠIHttp": 73519, + "ĠIHttpActionResult": 91457, + "ĠII": 7946, + "ĠIID": 62121, + "ĠIII": 14429, + "ĠIK": 47391, + "ĠIKE": 84364, + "ĠIL": 11344, + "ĠIList": 28714, + "ĠILogger": 51693, + "ĠIM": 6517, + "ĠIMAGE": 33669, + "ĠIMAGES": 84864, + "ĠIMD": 77263, + "ĠIMDb": 86401, + "ĠIMF": 59070, + "ĠIMG": 43583, + "ĠIMM": 68393, + "ĠIMO": 84071, + "ĠIMP": 39440, + "ĠIMPLEMENT": 80611, + "ĠIMPLIED": 12867, + "ĠIMPORT": 43603, + "ĠIMPORTANT": 67140, + "ĠIMS": 87277, + "ĠIMapper": 90499, + "ĠIMessage": 89937, + "ĠIN": 1964, + "ĠINA": 96299, + "ĠINC": 18118, + "ĠINCIDENT": 28167, + "ĠINCIDENTAL": 28553, + "ĠINCLUDE": 50393, + "ĠINCLUDED": 76451, + "ĠINCLUDING": 16092, + "ĠIND": 19317, + "ĠINDEX": 39300, + "ĠINDIRECT": 27072, + "ĠINF": 34972, + "ĠINFO": 30771, + "ĠINFORMATION": 44513, + "ĠINIT": 29964, + "ĠINITIAL": 56854, + "ĠINLINE": 65586, + "ĠINNER": 30348, + "ĠINPUT": 26149, + "ĠINS": 25692, + "ĠINSERT": 39518, + "ĠINST": 65344, + "ĠINSTALL": 79353, + "ĠINSTANCE": 61127, + "ĠINT": 9221, + "ĠINTEGER": 30381, + "ĠINTER": 29342, + "ĠINTERFACE": 89205, + "ĠINTERN": 61671, + "ĠINTERNAL": 72655, + "ĠINTERNATIONAL": 91411, + "ĠINTERRU": 29082, + "ĠINTERRUPTION": 30569, + "ĠINTERVAL": 91143, + "ĠINTO": 12496, + "ĠINV": 66924, + "ĠINVALID": 32269, + "ĠIO": 6429, + "ĠIOC": 96903, + "ĠIOCTL": 98544, + "ĠIOError": 60961, + "ĠIOException": 9346, + "ĠIOS": 64080, + "ĠIP": 6790, + "ĠIPA": 55747, + "ĠIPAddress": 83296, + "ĠIPC": 45833, + "ĠIPCC": 94561, + "ĠIPL": 76255, + "ĠIPO": 66892, + "ĠIPP": 63673, + "ĠIPPROTO": 86938, + "ĠIPS": 63157, + "ĠIPT": 45473, + "ĠIPV": 91498, + "ĠIPs": 96319, + "ĠIPv": 31560, + "ĠIQ": 36711, + "ĠIQueryable": 61789, + "ĠIR": 16248, + "ĠIRA": 58683, + "ĠIRC": 59328, + "ĠIRQ": 41668, + "ĠIRS": 33383, + "ĠIReadOnly": 82096, + "ĠIRepository": 66476, + "ĠIRequest": 81566, + "ĠIS": 3424, + "ĠISA": 84290, + "ĠISBN": 33426, + "ĠISC": 96181, + "ĠISI": 77201, + "ĠISIL": 82069, + "ĠISIS": 22647, + "ĠISO": 21940, + "ĠISP": 53433, + "ĠISPs": 80594, + "ĠISR": 77913, + "ĠISS": 36396, + "ĠISSN": 86686, + "ĠISSUE": 85901, + "ĠIST": 75458, + "ĠIService": 65725, + "ĠIT": 8700, + "ĠITE": 66956, + "ĠITEM": 35932, + "ĠITEMS": 77465, + "ĠITER": 87816, + "ĠITS": 46075, + "ĠITV": 93355, + "ĠIU": 80625, + "ĠIUser": 50894, + "ĠIV": 16824, + "ĠIW": 74536, + "ĠIX": 39712, + "ĠIan": 28235, + "ĠIb": 57497, + "ĠIbid": 75468, + "ĠIbn": 72286, + "ĠIbrahim": 65633, + "ĠIce": 19939, + "ĠIceland": 38396, + "ĠIcelandic": 99148, + "ĠIch": 25861, + "ĠIcon": 11226, + "ĠIconButton": 45586, + "ĠIconData": 65464, + "ĠIcons": 31187, + "ĠId": 5223, + "ĠIdaho": 39587, + "ĠIde": 13804, + "ĠIdea": 51001, + "ĠIdeal": 48427, + "ĠIdeally": 66701, + "ĠIdeas": 22898, + "ĠIdent": 22507, + "ĠIdentification": 58676, + "ĠIdentified": 96337, + "ĠIdentifier": 34621, + "ĠIdentify": 64547, + "ĠIdentity": 26540, + "ĠIdle": 70300, + "ĠIdol": 77864, + "ĠIf": 1416, + "ĠIg": 38451, + "ĠIgn": 33781, + "ĠIgnore": 38971, + "ĠIgor": 80998, + "ĠIh": 19051, + "ĠIhnen": 43860, + "ĠIhr": 35783, + "ĠIhre": 38472, + "ĠIhrem": 77065, + "ĠIhren": 70836, + "ĠIhrer": 75077, + "ĠIk": 41333, + "ĠIke": 98634, + "ĠIkea": 74439, + "ĠIl": 7543, + "ĠIll": 12516, + "ĠIllegal": 15038, + "ĠIllegalAccessException": 88780, + "ĠIllegalArgumentException": 21374, + "ĠIllegalStateException": 40827, + "ĠIllinois": 18657, + "ĠIllum": 60620, + "ĠIlluminate": 7518, + "ĠIllustr": 38054, + "ĠIllustrated": 70061, + "ĠIllustrator": 99060, + "ĠIls": 80145, + "ĠIm": 2362, + "ĠImGui": 38541, + "ĠImVec": 97132, + "ĠImag": 76433, + "ĠImage": 4654, + "ĠImageButton": 74098, + "ĠImageIcon": 48671, + "ĠImageView": 25677, + "ĠImages": 11779, + "ĠImagine": 37791, + "ĠImaging": 64506, + "ĠImam": 89812, + "ĠImg": 49930, + "ĠImm": 15335, + "ĠImmediate": 80310, + "ĠImmediately": 68252, + "ĠImmigration": 39682, + "ĠImmun": 66235, + "ĠImmutable": 39262, + "ĠImmutableList": 92255, + "ĠImp": 14390, + "ĠImpact": 28580, + "ĠImper": 24505, + "ĠImperial": 29913, + "ĠImpl": 88092, + "ĠImplement": 31075, + "ĠImplementation": 30813, + "ĠImplemented": 87013, + "ĠImplements": 49960, + "ĠImplicit": 97032, + "ĠImport": 13213, + "ĠImportError": 54065, + "ĠImportance": 93000, + "ĠImportant": 43821, + "ĠImported": 72266, + "ĠImports": 66098, + "ĠImpossible": 82652, + "ĠImpress": 96783, + "ĠImpro": 21961, + "ĠImprove": 64084, + "ĠImproved": 58123, + "ĠImprovement": 52651, + "ĠIn": 758, + "ĠInbox": 95579, + "ĠInc": 4848, + "ĠInch": 54592, + "ĠInches": 85929, + "ĠIncident": 68735, + "ĠInclude": 29734, + "ĠIncluded": 45964, + "ĠIncludes": 25954, + "ĠIncluding": 55121, + "ĠIncome": 32520, + "ĠIncoming": 96143, + "ĠIncontri": 54030, + "ĠIncorpor": 53704, + "ĠIncorporated": 66695, + "ĠIncorrect": 81857, + "ĠIncre": 31611, + "ĠIncrease": 39633, + "ĠIncreased": 61597, + "ĠIncreases": 67048, + "ĠIncreasing": 73440, + "ĠIncredible": 79853, + "ĠIncrement": 52835, + "ĠInd": 2263, + "ĠIndeed": 22348, + "ĠIndepend": 17127, + "ĠIndependence": 43077, + "ĠIndependent": 21994, + "ĠIndex": 8008, + "ĠIndexError": 60203, + "ĠIndexPath": 32896, + "ĠIndexed": 84404, + "ĠIndia": 6747, + "ĠIndian": 7748, + "ĠIndiana": 21591, + "ĠIndianapolis": 41351, + "ĠIndians": 29407, + "ĠIndicates": 44267, + "ĠIndicator": 63604, + "ĠIndices": 85769, + "ĠIndie": 76339, + "ĠIndies": 84218, + "ĠIndigenous": 46003, + "ĠIndividual": 29340, + "ĠIndividuals": 61425, + "ĠIndo": 75885, + "ĠIndones": 20263, + "ĠIndonesia": 23968, + "ĠIndonesian": 58829, + "ĠIndoor": 63268, + "ĠIndust": 17210, + "ĠIndustrial": 24545, + "ĠIndustries": 36428, + "ĠIndustry": 23840, + "ĠIndy": 56470, + "ĠInetAddress": 94890, + "ĠInf": 14921, + "ĠInfant": 82388, + "ĠInfantry": 64556, + "ĠInfect": 87811, + "ĠInfer": 62658, + "ĠInfinite": 54543, + "ĠInfinity": 39889, + "ĠInflate": 68738, + "ĠInflu": 87554, + "ĠInfluence": 74940, + "ĠInfo": 13074, + "ĠInform": 30601, + "ĠInformation": 8085, + "ĠInformationen": 48842, + "ĠInfos": 77070, + "ĠInfragistics": 54902, + "ĠInfrastructure": 44487, + "ĠIng": 24871, + "ĠIngen": 94468, + "ĠIngram": 94366, + "ĠIngredient": 54897, + "ĠIngredients": 51175, + "ĠInhal": 91912, + "ĠInherits": 72467, + "ĠIni": 73767, + "ĠInit": 15690, + "ĠIniti": 31397, + "ĠInitial": 4127, + "ĠInitialise": 71340, + "ĠInitialization": 39267, + "ĠInitialize": 9008, + "ĠInitializeComponent": 16233, + "ĠInitialized": 7590, + "ĠInitializes": 31882, + "ĠInitializing": 94312, + "ĠInitially": 58556, + "ĠInitiative": 37656, + "ĠInject": 21843, + "ĠInjectable": 29515, + "ĠInjection": 53811, + "ĠInjector": 91965, + "ĠInjury": 65076, + "ĠInk": 41083, + "ĠInkWell": 84840, + "ĠInline": 54235, + "ĠInn": 16952, + "ĠInner": 36356, + "ĠInnoc": 83493, + "ĠInnov": 54847, + "ĠInnovation": 37610, + "ĠInnovative": 97569, + "ĠInput": 5571, + "ĠInputDecoration": 54478, + "ĠInputStream": 24857, + "ĠInputStreamReader": 40287, + "ĠInputs": 46281, + "ĠInquiry": 65702, + "ĠIns": 9726, + "ĠInsert": 17101, + "ĠInserts": 84753, + "ĠInsets": 75367, + "ĠInside": 27368, + "ĠInsider": 44999, + "ĠInsight": 70971, + "ĠInsights": 72037, + "ĠInsp": 29916, + "ĠInspection": 65946, + "ĠInspector": 44495, + "ĠInspir": 88857, + "ĠInspiration": 78477, + "ĠInspired": 64825, + "ĠInst": 5984, + "ĠInstagram": 13991, + "ĠInstall": 19242, + "ĠInstallation": 39145, + "ĠInstalled": 98077, + "ĠInstaller": 88091, + "ĠInstalling": 79800, + "ĠInstance": 19283, + "ĠInstances": 89241, + "ĠInstant": 18058, + "ĠInstantiate": 32288, + "ĠInstead": 12090, + "ĠInstit": 31896, + "ĠInstitut": 95462, + "ĠInstitute": 9976, + "ĠInstitutes": 72728, + "ĠInstitution": 42864, + "ĠInstitutional": 97884, + "ĠInstitutions": 87780, + "ĠInstituto": 86716, + "ĠInstruction": 29051, + "ĠInstructions": 38297, + "ĠInstructor": 62362, + "ĠInstrument": 42305, + "ĠInstruments": 57413, + "ĠInsurance": 21674, + "ĠInt": 1333, + "ĠIntPtr": 23936, + "ĠInteger": 4440, + "ĠIntegr": 29001, + "ĠIntegral": 91660, + "ĠIntegrated": 49421, + "ĠIntegration": 40069, + "ĠIntegrity": 64807, + "ĠIntel": 15611, + "ĠIntellectual": 76856, + "ĠIntelli": 84549, + "ĠIntelliJ": 94082, + "ĠIntelligence": 21392, + "ĠIntelligent": 77088, + "ĠIntent": 8829, + "ĠInter": 5665, + "ĠInteraction": 42707, + "ĠInteractive": 41866, + "ĠIntercept": 98338, + "ĠInterest": 24106, + "ĠInterested": 76262, + "ĠInteresting": 70670, + "ĠInterestingly": 57503, + "ĠInterface": 20019, + "ĠInterfaces": 79857, + "ĠInterior": 28858, + "ĠIntermediate": 60648, + "ĠIntern": 4414, + "ĠInternacional": 95067, + "ĠInternal": 15412, + "ĠInternalEnumerator": 75609, + "ĠInternational": 7179, + "ĠInternet": 8031, + "ĠInterpret": 72077, + "ĠInterpreter": 82493, + "ĠInterr": 79380, + "ĠInterracial": 97063, + "ĠInterrupt": 22145, + "ĠInterruptedException": 35620, + "ĠIntersection": 79014, + "ĠInterstate": 65493, + "ĠInterval": 40584, + "ĠIntervention": 70578, + "ĠInterview": 31467, + "ĠInterviews": 88072, + "ĠIntl": 84578, + "ĠInto": 31645, + "ĠIntr": 60794, + "ĠIntro": 51986, + "ĠIntroduced": 41689, + "ĠIntroduction": 28338, + "ĠInv": 18807, + "ĠInvalid": 13882, + "ĠInvalidArgumentException": 80518, + "ĠInvalidOperationException": 52612, + "ĠInvalidate": 95246, + "ĠInvasion": 94404, + "ĠInvent": 92158, + "ĠInventory": 28126, + "ĠInvest": 12748, + "ĠInvestig": 32080, + "ĠInvestigation": 46897, + "ĠInvestigations": 88105, + "ĠInvestigators": 95752, + "ĠInvesting": 92596, + "ĠInvestment": 32250, + "ĠInvestments": 76890, + "ĠInvestor": 75712, + "ĠInvestors": 69533, + "ĠInvisible": 89038, + "ĠInvitation": 83242, + "ĠInvite": 84923, + "ĠInvocation": 78493, + "ĠInvoice": 40190, + "ĠInvoke": 39667, + "ĠIo": 29655, + "ĠIoT": 49080, + "ĠIon": 44805, + "ĠIonic": 39574, + "ĠIonicModule": 81182, + "ĠIonicPage": 93168, + "ĠIowa": 20700, + "ĠIp": 35033, + "ĠIps": 88962, + "ĠIpsum": 61362, + "ĠIr": 11960, + "ĠIran": 10259, + "ĠIranian": 27401, + "ĠIranians": 89331, + "ĠIraq": 11093, + "ĠIraqi": 30234, + "ĠIre": 13892, + "ĠIreland": 14648, + "ĠIrene": 87213, + "ĠIris": 65721, + "ĠIrish": 17623, + "ĠIrma": 98392, + "ĠIron": 16570, + "ĠIronically": 93880, + "ĠIrr": 78041, + "ĠIrvine": 71285, + "ĠIrving": 55210, + "ĠIs": 2160, + "ĠIsActive": 94600, + "ĠIsNot": 68243, + "ĠIsPlainOldData": 53623, + "ĠIsValid": 70647, + "ĠIsa": 28005, + "ĠIsaac": 41508, + "ĠIsabel": 86786, + "ĠIsaiah": 63430, + "ĠIsh": 56604, + "ĠIsis": 53967, + "ĠIsl": 9219, + "ĠIslam": 14910, + "ĠIslamabad": 91432, + "ĠIslamic": 15203, + "ĠIslamist": 51493, + "ĠIsland": 10720, + "ĠIslanders": 78453, + "ĠIslands": 22235, + "ĠIsle": 54265, + "ĠIsles": 85943, + "ĠIsn": 52196, + "ĠIso": 86943, + "ĠIsrael": 6778, + "ĠIsraeli": 15907, + "ĠIsraelis": 55623, + "ĠIss": 15935, + "ĠIssue": 25226, + "ĠIssues": 36443, + "ĠIst": 59490, + "ĠIstanbul": 53368, + "ĠIt": 1084, + "ĠItal": 83083, + "ĠItalia": 52122, + "ĠItalian": 14811, + "ĠItalians": 96233, + "ĠItaly": 15344, + "ĠItem": 5739, + "ĠItemStack": 38115, + "ĠItemType": 73081, + "ĠItems": 19410, + "ĠIter": 13704, + "ĠIterable": 38016, + "ĠIterate": 54340, + "ĠIterator": 23023, + "ĠIts": 11445, + "ĠIv": 33483, + "ĠIvan": 41421, + "ĠIvanka": 82332, + "ĠIve": 86650, + "ĠIvory": 76609, + "ĠIvy": 56879, + "ĠIz": 47823, + "ĠJ": 619, + "ĠJA": 53739, + "ĠJACK": 91042, + "ĠJADX": 88508, + "ĠJAN": 83487, + "ĠJAVA": 85560, + "ĠJAXB": 60640, + "ĠJAXBElement": 80140, + "ĠJB": 76279, + "ĠJButton": 27117, + "ĠJC": 60583, + "ĠJComboBox": 87649, + "ĠJD": 43937, + "ĠJDBC": 88320, + "ĠJDK": 97425, + "ĠJE": 70330, + "ĠJFK": 83299, + "ĠJFactory": 74304, + "ĠJFrame": 50363, + "ĠJIT": 96187, + "ĠJJ": 70040, + "ĠJK": 93226, + "ĠJL": 89435, + "ĠJLabel": 23466, + "ĠJM": 82052, + "ĠJMP": 88251, + "ĠJMenuItem": 73260, + "ĠJNI": 70829, + "ĠJNICALL": 60940, + "ĠJNIEnv": 85753, + "ĠJO": 10247, + "ĠJOB": 61659, + "ĠJOHN": 75175, + "ĠJOIN": 13069, + "ĠJObject": 65369, + "ĠJOptionPane": 27080, + "ĠJP": 48780, + "ĠJPEG": 54562, + "ĠJPG": 88956, + "ĠJPanel": 27728, + "ĠJQuery": 86645, + "ĠJR": 52566, + "ĠJS": 12162, + "ĠJSBracketAccess": 96684, + "ĠJSGlobal": 93735, + "ĠJSImport": 78752, + "ĠJSName": 96481, + "ĠJSON": 4718, + "ĠJSONArray": 40461, + "ĠJSONException": 84690, + "ĠJSONObject": 17883, + "ĠJSX": 55496, + "ĠJScrollPane": 81905, + "ĠJT": 87504, + "ĠJText": 38555, + "ĠJTextField": 37298, + "ĠJUL": 87060, + "ĠJUST": 47530, + "ĠJV": 95461, + "ĠJVM": 72379, + "ĠJW": 95476, + "ĠJWT": 47567, + "ĠJa": 22871, + "ĠJab": 52982, + "ĠJac": 15035, + "ĠJack": 7607, + "ĠJacket": 54770, + "ĠJackets": 85564, + "ĠJackie": 50865, + "ĠJackson": 13658, + "ĠJacksonville": 47946, + "ĠJacob": 24521, + "ĠJacobs": 51656, + "ĠJacqu": 79077, + "ĠJacqueline": 98664, + "ĠJacques": 63750, + "ĠJad": 72724, + "ĠJade": 57527, + "ĠJae": 95560, + "ĠJag": 29411, + "ĠJaguar": 73537, + "ĠJaguars": 69326, + "ĠJah": 34624, + "ĠJahr": 37419, + "ĠJahre": 48015, + "ĠJahren": 45506, + "ĠJahres": 97748, + "ĠJail": 56146, + "ĠJaime": 77723, + "ĠJain": 95117, + "ĠJak": 34834, + "ĠJakarta": 63033, + "ĠJake": 32072, + "ĠJal": 71895, + "ĠJam": 20016, + "ĠJama": 40159, + "ĠJamaica": 56175, + "ĠJamal": 95657, + "ĠJame": 92088, + "ĠJames": 7801, + "ĠJamie": 35757, + "ĠJan": 4350, + "ĠJane": 21475, + "ĠJaneiro": 60323, + "ĠJanet": 53665, + "ĠJanuary": 6058, + "ĠJapan": 6323, + "ĠJapanese": 10769, + "ĠJar": 30551, + "ĠJared": 43228, + "ĠJarvis": 98520, + "ĠJas": 85116, + "ĠJasmine": 81352, + "ĠJason": 18477, + "ĠJasper": 70484, + "ĠJaune": 86125, + "ĠJava": 7943, + "ĠJavaScript": 12914, + "ĠJavascript": 31852, + "ĠJavier": 70209, + "ĠJaw": 33194, + "ĠJay": 18919, + "ĠJays": 65735, + "ĠJazeera": 84924, + "ĠJazz": 35867, + "ĠJe": 14133, + "ĠJean": 19685, + "ĠJeans": 81407, + "ĠJeb": 70096, + "ĠJed": 65132, + "ĠJedi": 40395, + "ĠJeep": 49902, + "ĠJeff": 11885, + "ĠJefferson": 33544, + "ĠJeffrey": 41007, + "ĠJeg": 53117, + "ĠJeh": 61933, + "ĠJehovah": 78560, + "ĠJelly": 73139, + "ĠJen": 24893, + "ĠJenkins": 44136, + "ĠJenn": 22923, + "ĠJenna": 87254, + "ĠJenner": 76064, + "ĠJennifer": 29650, + "ĠJennings": 70190, + "ĠJenny": 51079, + "ĠJens": 95030, + "ĠJensen": 64009, + "ĠJer": 8403, + "ĠJeremiah": 82625, + "ĠJeremy": 27118, + "ĠJerome": 69774, + "ĠJerry": 28708, + "ĠJersey": 15849, + "ĠJerseys": 80542, + "ĠJerusalem": 25445, + "ĠJes": 9066, + "ĠJess": 23039, + "ĠJesse": 39171, + "ĠJessica": 32367, + "ĠJessie": 79117, + "ĠJest": 93469, + "ĠJesus": 10588, + "ĠJesús": 142115, + "ĠJet": 24366, + "ĠJetBrains": 75124, + "ĠJets": 36489, + "ĠJetzt": 77717, + "ĠJew": 7958, + "ĠJewel": 83204, + "ĠJewelry": 61868, + "ĠJewish": 13350, + "ĠJews": 16987, + "ĠJeżeli": 136851, + "ĠJeÅĽli": 130602, + "ĠJi": 54451, + "ĠJian": 87304, + "ĠJiang": 60822, + "ĠJihad": 90139, + "ĠJill": 47211, + "ĠJim": 11387, + "ĠJimmy": 27833, + "ĠJin": 38511, + "ĠJing": 85439, + "ĠJinping": 93386, + "ĠJo": 10946, + "ĠJoan": 50106, + "ĠJoanna": 90062, + "ĠJob": 12011, + "ĠJobs": 23076, + "ĠJoe": 12846, + "ĠJoel": 38347, + "ĠJoey": 54735, + "ĠJog": 86407, + "ĠJoh": 26176, + "ĠJohan": 88807, + "ĠJohann": 87864, + "ĠJohannes": 54105, + "ĠJohannesburg": 85541, + "ĠJohn": 3757, + "ĠJohnny": 31880, + "ĠJohns": 52140, + "ĠJohnson": 11351, + "ĠJohnston": 60482, + "ĠJoi": 41168, + "ĠJoin": 16471, + "ĠJoined": 40782, + "ĠJoint": 29733, + "ĠJoker": 60277, + "ĠJon": 12285, + "ĠJonah": 93628, + "ĠJonas": 68086, + "ĠJonathan": 23269, + "ĠJones": 11935, + "ĠJong": 39548, + "ĠJoomla": 72005, + "ĠJord": 80155, + "ĠJordan": 17089, + "ĠJorge": 55400, + "ĠJos": 27878, + "ĠJose": 10860, + "ĠJosef": 97441, + "ĠJoseph": 15115, + "ĠJosh": 18246, + "ĠJoshua": 39492, + "ĠJosé": 43265, + "ĠJou": 96048, + "ĠJournal": 9936, + "ĠJournalism": 68553, + "ĠJourney": 42580, + "ĠJoy": 27138, + "ĠJoyce": 53626, + "ĠJoão": 135970, + "ĠJpaRepository": 65618, + "ĠJr": 15640, + "ĠJs": 39122, + "ĠJson": 8308, + "ĠJsonConvert": 38867, + "ĠJsonObject": 49343, + "ĠJsonRequest": 68709, + "ĠJsonRequestBehavior": 69831, + "ĠJsonResponse": 38263, + "ĠJsonResult": 58178, + "ĠJsonSerializer": 70919, + "ĠJu": 21671, + "ĠJuan": 28504, + "ĠJub": 79280, + "ĠJud": 11213, + "ĠJuda": 57461, + "ĠJudaism": 63183, + "ĠJude": 61634, + "ĠJudge": 20207, + "ĠJudges": 85855, + "ĠJudgment": 88562, + "ĠJudicial": 67208, + "ĠJudiciary": 52029, + "ĠJudith": 78093, + "ĠJudy": 54058, + "ĠJug": 55319, + "ĠJugend": 90167, + "ĠJuice": 65751, + "ĠJul": 10057, + "ĠJuli": 72124, + "ĠJulia": 39294, + "ĠJulian": 37797, + "ĠJulie": 41187, + "ĠJulien": 87642, + "ĠJuliet": 71024, + "ĠJulio": 80825, + "ĠJulius": 69245, + "ĠJuly": 5768, + "ĠJump": 28788, + "ĠJun": 11782, + "ĠJunction": 67956, + "ĠJune": 5534, + "ĠJung": 49332, + "ĠJungle": 59663, + "ĠJuni": 80157, + "ĠJunior": 30770, + "ĠJunk": 66552, + "ĠJupiter": 49689, + "ĠJur": 38798, + "ĠJurassic": 83374, + "ĠJury": 95271, + "ĠJust": 4599, + "ĠJusti": 140232, + "ĠJustice": 11745, + "ĠJustin": 22467, + "ĠJustiça": 140233, + "ĠJuventus": 77121, + "ĠJwt": 54369, + "ĠJá": 134179, + "ĠK": 730, + "ĠKA": 88978, + "ĠKB": 25016, + "ĠKC": 21084, + "ĠKD": 62990, + "ĠKDE": 75083, + "ĠKE": 11090, + "ĠKEEP": 95809, + "ĠKERNEL": 79691, + "ĠKEY": 12013, + "ĠKG": 70087, + "ĠKH": 80355, + "ĠKIND": 9297, + "ĠKING": 73811, + "ĠKK": 96432, + "ĠKL": 67710, + "ĠKM": 55776, + "ĠKN": 31292, + "ĠKNOW": 58027, + "ĠKO": 37512, + "ĠKP": 77524, + "ĠKR": 65169, + "ĠKS": 46374, + "ĠKT": 77461, + "ĠKV": 84648, + "ĠKW": 71915, + "ĠKY": 49479, + "ĠKa": 22309, + "ĠKab": 40436, + "ĠKabul": 86545, + "ĠKad": 73676, + "ĠKaepernick": 78614, + "ĠKafka": 57025, + "ĠKag": 75991, + "ĠKah": 84007, + "ĠKahn": 98566, + "ĠKai": 50610, + "ĠKaiser": 66965, + "ĠKak": 74471, + "ĠKal": 26832, + "ĠKale": 96742, + "ĠKam": 28449, + "ĠKami": 98866, + "ĠKamp": 70944, + "ĠKan": 30563, + "ĠKane": 45556, + "ĠKang": 54276, + "ĠKann": 77211, + "ĠKansas": 20148, + "ĠKant": 62162, + "ĠKanye": 58717, + "ĠKap": 31665, + "ĠKaplan": 80137, + "ĠKapoor": 81369, + "ĠKar": 13225, + "ĠKara": 78350, + "ĠKarachi": 88608, + "ĠKardash": 61262, + "ĠKardashian": 68007, + "ĠKare": 95231, + "ĠKaren": 34645, + "ĠKarl": 34031, + "ĠKarma": 74880, + "ĠKarn": 60223, + "ĠKarnataka": 78049, + "ĠKart": 60097, + "ĠKas": 37542, + "ĠKash": 41608, + "ĠKashmir": 49180, + "ĠKasich": 74981, + "ĠKasım": 135335, + "ĠKat": 17363, + "ĠKata": 98838, + "ĠKate": 29201, + "ĠKath": 32895, + "ĠKatherine": 61326, + "ĠKathleen": 64063, + "ĠKathryn": 91068, + "ĠKathy": 64198, + "ĠKatie": 44463, + "ĠKatrina": 68656, + "ĠKaty": 72127, + "ĠKatz": 77109, + "ĠKauf": 58154, + "ĠKavanaugh": 37183, + "ĠKaw": 50516, + "ĠKawasaki": 97438, + "ĠKay": 30145, + "ĠKaz": 34974, + "ĠKazakhstan": 72437, + "ĠKaż": 142278, + "ĠKażdy": 142279, + "ĠKb": 98086, + "ĠKe": 6569, + "ĠKear": 99038, + "ĠKee": 95728, + "ĠKeep": 13655, + "ĠKeeper": 81627, + "ĠKeeping": 55378, + "ĠKeeps": 90820, + "ĠKeith": 31502, + "ĠKel": 27164, + "ĠKeller": 65503, + "ĠKelley": 73713, + "ĠKelly": 18661, + "ĠKelvin": 90973, + "ĠKem": 65966, + "ĠKemp": 80508, + "ĠKen": 14259, + "ĠKend": 38117, + "ĠKendall": 73076, + "ĠKendrick": 87613, + "ĠKenn": 15378, + "ĠKennedy": 23651, + "ĠKenneth": 48398, + "ĠKenny": 48319, + "ĠKens": 73023, + "ĠKensington": 95733, + "ĠKent": 17734, + "ĠKentucky": 24987, + "ĠKenya": 36666, + "ĠKepler": 97743, + "ĠKer": 35158, + "ĠKerala": 60407, + "ĠKern": 84687, + "ĠKernel": 36603, + "ĠKerr": 59195, + "ĠKerry": 39240, + "ĠKes": 82977, + "ĠKet": 67084, + "ĠKeto": 76345, + "ĠKevin": 16364, + "ĠKey": 5309, + "ĠKeyCode": 65152, + "ĠKeyError": 38094, + "ĠKeyEvent": 48275, + "ĠKeyValue": 98620, + "ĠKeyValuePair": 53604, + "ĠKeyboard": 25616, + "ĠKeyboardInterrupt": 84188, + "ĠKeynes": 81503, + "ĠKeys": 24133, + "ĠKeystone": 70573, + "ĠKeyword": 48970, + "ĠKeywords": 55695, + "ĠKg": 86722, + "ĠKh": 20165, + "ĠKhal": 51609, + "ĠKhan": 24287, + "ĠKhá": 141872, + "ĠKhách": 138053, + "ĠKhánh": 141425, + "ĠKhông": 129182, + "ĠKi": 29458, + "ĠKia": 80270, + "ĠKick": 26678, + "ĠKickstarter": 39980, + "ĠKid": 31566, + "ĠKidd": 97138, + "ĠKids": 22522, + "ĠKiev": 61625, + "ĠKil": 37888, + "ĠKill": 26835, + "ĠKilled": 43796, + "ĠKiller": 52953, + "ĠKilling": 69577, + "ĠKills": 60853, + "ĠKim": 13509, + "ĠKimber": 63807, + "ĠKimberly": 85653, + "ĠKin": 30891, + "ĠKind": 16840, + "ĠKinder": 43999, + "ĠKindle": 34312, + "ĠKinect": 94503, + "ĠKing": 6210, + "ĠKingdom": 15072, + "ĠKings": 24019, + "ĠKingston": 62469, + "ĠKir": 25527, + "ĠKirby": 57711, + "ĠKirk": 31346, + "ĠKirst": 80677, + "ĠKis": 85637, + "ĠKiss": 43630, + "ĠKit": 20951, + "ĠKitchen": 19354, + "ĠKits": 61893, + "ĠKitt": 80667, + "ĠKitty": 66181, + "ĠKiÅŁi": 141508, + "ĠKl": 27941, + "ĠKlan": 87341, + "ĠKlaus": 81097, + "ĠKle": 26349, + "ĠKlein": 42141, + "ĠKling": 79027, + "ĠKlo": 73641, + "ĠKlopp": 93831, + "ĠKm": 92482, + "ĠKn": 13620, + "ĠKnee": 97212, + "ĠKnicks": 61412, + "ĠKnife": 61202, + "ĠKnight": 22984, + "ĠKnights": 39576, + "ĠKnock": 63794, + "ĠKnot": 92144, + "ĠKnow": 14187, + "ĠKnowing": 57633, + "ĠKnowledge": 31925, + "ĠKnown": 48286, + "ĠKnox": 53350, + "ĠKnoxville": 94902, + "ĠKo": 29317, + "ĠKob": 59148, + "ĠKobe": 69311, + "ĠKoch": 39493, + "ĠKod": 67302, + "ĠKodi": 81312, + "ĠKoh": 33875, + "ĠKohana": 82673, + "ĠKok": 64715, + "ĠKol": 45515, + "ĠKolkata": 81534, + "ĠKom": 26692, + "ĠKomm": 87727, + "ĠKomment": 67489, + "ĠKon": 23388, + "ĠKong": 18211, + "ĠKont": 36866, + "ĠKontakt": 65608, + "ĠKop": 58009, + "ĠKor": 35070, + "ĠKore": 8062, + "ĠKorea": 11862, + "ĠKorean": 16134, + "ĠKoreans": 69711, + "ĠKos": 37108, + "ĠKosovo": 79897, + "ĠKosten": 40866, + "ĠKostenlos": 81442, + "ĠKostenlose": 86784, + "ĠKot": 61663, + "ĠKota": 92215, + "ĠKotlin": 92854, + "ĠKou": 75865, + "ĠKov": 92881, + "ĠKr": 16447, + "ĠKra": 61067, + "ĠKraft": 64560, + "ĠKrak": 88443, + "ĠKramer": 89477, + "ĠKrank": 91970, + "ĠKre": 29618, + "ĠKremlin": 50324, + "ĠKris": 52002, + "ĠKrish": 60964, + "ĠKrishna": 79309, + "ĠKrist": 26875, + "ĠKristen": 76114, + "ĠKro": 67817, + "ĠKron": 96560, + "ĠKsi": 143833, + "ĠKsiÄĻ": 143834, + "ĠKu": 32379, + "ĠKuala": 86333, + "ĠKub": 62549, + "ĠKubernetes": 66374, + "ĠKul": 84506, + "ĠKum": 85241, + "ĠKumar": 40140, + "ĠKun": 42783, + "ĠKund": 92887, + "ĠKunden": 61291, + "ĠKunst": 80331, + "ĠKur": 31275, + "ĠKurd": 33423, + "ĠKurdish": 41640, + "ĠKurdistan": 78944, + "ĠKurds": 67014, + "ĠKurt": 42923, + "ĠKush": 49175, + "ĠKushner": 61331, + "ĠKut": 95907, + "ĠKuwait": 56011, + "ĠKv": 88321, + "ĠKw": 64362, + "ĠKy": 22878, + "ĠKyle": 30259, + "ĠKylie": 97164, + "ĠKyoto": 81022, + "ĠKö": 64894, + "ĠKöln": 141377, + "ĠKön": 141774, + "ĠKönig": 141775, + "ĠKörper": 135720, + "ĠKü": 69252, + "ĠKüche": 136112, + "ĠKültür": 139223, + "ĠKı": 143804, + "ĠKısa": 143805, + "ĠKết": 136909, + "ĠKỳ": 133169, + "ĠL": 444, + "ĠLA": 12959, + "ĠLAB": 30974, + "ĠLABEL": 56874, + "ĠLAN": 46425, + "ĠLAND": 91856, + "ĠLANG": 65997, + "ĠLANGUAGE": 34800, + "ĠLAP": 71328, + "ĠLARGE": 76656, + "ĠLAS": 64131, + "ĠLAST": 47295, + "ĠLAT": 53187, + "ĠLATIN": 92244, + "ĠLAW": 66603, + "ĠLB": 40150, + "ĠLC": 30871, + "ĠLCD": 21976, + "ĠLCS": 83868, + "ĠLD": 28877, + "ĠLDAP": 56379, + "ĠLDL": 85443, + "ĠLDS": 71536, + "ĠLE": 11148, + "ĠLED": 13113, + "ĠLEDs": 55572, + "ĠLEFT": 21920, + "ĠLEG": 35426, + "ĠLEGO": 55801, + "ĠLEN": 83099, + "ĠLENG": 69635, + "ĠLENGTH": 71838, + "ĠLES": 96429, + "ĠLESS": 74948, + "ĠLET": 35657, + "ĠLETTER": 44428, + "ĠLEVEL": 42709, + "ĠLF": 57136, + "ĠLG": 23403, + "ĠLGBT": 23111, + "ĠLGBTQ": 49942, + "ĠLGPL": 47529, + "ĠLH": 82763, + "ĠLI": 7555, + "ĠLIABILITY": 14162, + "ĠLIABLE": 17386, + "ĠLIB": 30271, + "ĠLIC": 83238, + "ĠLICENSE": 12096, + "ĠLIFE": 59994, + "ĠLIGHT": 52203, + "ĠLIKE": 20529, + "ĠLIMIT": 10376, + "ĠLIMITED": 13104, + "ĠLIN": 64789, + "ĠLINE": 30851, + "ĠLINK": 40491, + "ĠLINUX": 91618, + "ĠLIS": 99026, + "ĠLIST": 26966, + "ĠLIVE": 44169, + "ĠLJ": 96688, + "ĠLL": 19504, + "ĠLLC": 15262, + "ĠLLP": 67965, + "ĠLLVM": 52339, + "ĠLM": 46414, + "ĠLN": 88182, + "ĠLNG": 87830, + "ĠLO": 5017, + "ĠLOAD": 38340, + "ĠLOC": 27096, + "ĠLOCAL": 42501, + "ĠLOCATION": 65583, + "ĠLOCK": 49463, + "ĠLOD": 86963, + "ĠLOG": 8081, + "ĠLOGGER": 28088, + "ĠLOGIN": 48131, + "ĠLOL": 55333, + "ĠLONG": 33942, + "ĠLOOK": 49869, + "ĠLOOP": 69530, + "ĠLORD": 57377, + "ĠLOS": 52687, + "ĠLOSS": 27353, + "ĠLOT": 53360, + "ĠLOVE": 39735, + "ĠLOW": 40505, + "ĠLOWER": 85457, + "ĠLP": 17102, + "ĠLPARAM": 66987, + "ĠLPC": 49075, + "ĠLR": 44550, + "ĠLS": 29202, + "ĠLSB": 93310, + "ĠLSD": 71070, + "ĠLSM": 90474, + "ĠLSTM": 85471, + "ĠLSU": 64412, + "ĠLT": 34469, + "ĠLTC": 95939, + "ĠLTD": 50185, + "ĠLTE": 51398, + "ĠLTS": 79028, + "ĠLU": 74950, + "ĠLUA": 70505, + "ĠLV": 39169, + "ĠLW": 55741, + "ĠLX": 75798, + "ĠLY": 86005, + "ĠLZ": 94480, + "ĠLa": 4929, + "ĠLaTeX": 97913, + "ĠLab": 11607, + "ĠLabel": 9402, + "ĠLabels": 60996, + "ĠLabor": 14930, + "ĠLaboratories": 77617, + "ĠLaboratory": 31084, + "ĠLabour": 18486, + "ĠLabrador": 79276, + "ĠLabs": 40640, + "ĠLac": 54041, + "ĠLace": 74139, + "ĠLack": 67402, + "ĠLad": 69030, + "ĠLaden": 51384, + "ĠLadies": 48013, + "ĠLady": 20621, + "ĠLaf": 60844, + "ĠLafayette": 77181, + "ĠLag": 32371, + "ĠLage": 95506, + "ĠLager": 84758, + "ĠLagos": 75996, + "ĠLaguna": 93098, + "ĠLah": 67442, + "ĠLahore": 98089, + "ĠLair": 57096, + "ĠLak": 47662, + "ĠLake": 11678, + "ĠLakers": 46496, + "ĠLakes": 41579, + "ĠLal": 83783, + "ĠLam": 32694, + "ĠLama": 82559, + "ĠLamar": 61184, + "ĠLamb": 31302, + "ĠLambda": 44521, + "ĠLambert": 69543, + "ĠLamp": 41177, + "ĠLan": 34782, + "ĠLana": 94787, + "ĠLanc": 38534, + "ĠLancaster": 65123, + "ĠLance": 48931, + "ĠLand": 11426, + "ĠLanding": 46424, + "ĠLands": 37423, + "ĠLandscape": 62541, + "ĠLane": 26018, + "ĠLang": 22463, + "ĠLange": 87722, + "ĠLanguage": 11434, + "ĠLanguages": 54964, + "ĠLanka": 42479, + "ĠLans": 67836, + "ĠLansing": 96484, + "ĠLantern": 67492, + "ĠLaos": 95744, + "ĠLap": 49055, + "ĠLaptop": 56125, + "ĠLar": 31304, + "ĠLara": 72157, + "ĠLaravel": 64206, + "ĠLarge": 20286, + "ĠLarger": 80808, + "ĠLargest": 83319, + "ĠLarry": 29290, + "ĠLars": 58176, + "ĠLarson": 85078, + "ĠLas": 15755, + "ĠLaser": 39608, + "ĠLast": 7996, + "ĠLastName": 48946, + "ĠLastly": 70709, + "ĠLat": 9926, + "ĠLatLng": 47165, + "ĠLate": 35831, + "ĠLater": 24885, + "ĠLatest": 28157, + "ĠLatin": 19458, + "ĠLatina": 58763, + "ĠLatino": 43762, + "ĠLatinos": 86799, + "ĠLatitude": 67340, + "ĠLatter": 89066, + "ĠLatvia": 78158, + "ĠLau": 84991, + "ĠLaud": 70036, + "ĠLauderdale": 77155, + "ĠLaugh": 83035, + "ĠLaunch": 23205, + "ĠLauncher": 52230, + "ĠLaundry": 81632, + "ĠLaur": 53200, + "ĠLaura": 29828, + "ĠLaure": 32667, + "ĠLaurel": 80464, + "ĠLauren": 42360, + "ĠLaurent": 63221, + "ĠLaurie": 79523, + "ĠLav": 42850, + "ĠLaw": 7506, + "ĠLawn": 68392, + "ĠLawrence": 27474, + "ĠLaws": 40989, + "ĠLawson": 74851, + "ĠLawyer": 70214, + "ĠLawyers": 70408, + "ĠLay": 43924, + "ĠLayer": 22735, + "ĠLayers": 83822, + "ĠLayout": 13821, + "ĠLayoutInflater": 28569, + "ĠLaz": 97803, + "ĠLazar": 86158, + "ĠLazy": 44263, + "ĠLe": 1967, + "ĠLeBron": 57235, + "ĠLead": 29882, + "ĠLeader": 23030, + "ĠLeaders": 27886, + "ĠLeadership": 36163, + "ĠLeading": 58043, + "ĠLeads": 81528, + "ĠLeaf": 26869, + "ĠLeafs": 64150, + "ĠLeague": 8953, + "ĠLeah": 66861, + "ĠLeak": 37140, + "ĠLeakage": 42026, + "ĠLean": 62106, + "ĠLeap": 71282, + "ĠLear": 59978, + "ĠLearn": 14934, + "ĠLearned": 89396, + "ĠLearning": 20909, + "ĠLease": 80651, + "ĠLeast": 90516, + "ĠLeather": 31078, + "ĠLeave": 25193, + "ĠLeaves": 77514, + "ĠLeaving": 72962, + "ĠLeban": 31224, + "ĠLebanese": 68245, + "ĠLebanon": 39271, + "ĠLeben": 55397, + "ĠLebens": 65314, + "ĠLect": 40943, + "ĠLecture": 66972, + "ĠLed": 31655, + "ĠLedger": 70306, + "ĠLee": 12066, + "ĠLeeds": 51529, + "ĠLeer": 92798, + "ĠLeft": 13727, + "ĠLeg": 7610, + "ĠLegacy": 37887, + "ĠLegal": 24678, + "ĠLegend": 26935, + "ĠLegendary": 59089, + "ĠLegends": 41886, + "ĠLegion": 39835, + "ĠLegisl": 32715, + "ĠLegislation": 71967, + "ĠLegislative": 67405, + "ĠLegislature": 46026, + "ĠLego": 64670, + "ĠLegs": 89934, + "ĠLeh": 62558, + "ĠLehr": 91853, + "ĠLei": 72210, + "ĠLeia": 87329, + "ĠLeicester": 57749, + "ĠLeigh": 73022, + "ĠLeipzig": 86547, + "ĠLeisure": 90399, + "ĠLem": 47406, + "ĠLemma": 85810, + "ĠLemon": 51210, + "ĠLen": 24250, + "ĠLena": 81062, + "ĠLength": 17287, + "ĠLenin": 62812, + "ĠLennon": 79740, + "ĠLenovo": 63699, + "ĠLens": 42671, + "ĠLent": 83370, + "ĠLeo": 36748, + "ĠLeon": 20932, + "ĠLeonard": 40854, + "ĠLeonardo": 65386, + "ĠLeone": 65391, + "ĠLeopard": 97538, + "ĠLer": 79840, + "ĠLes": 11615, + "ĠLesb": 59961, + "ĠLesbian": 52543, + "ĠLeslie": 53683, + "ĠLess": 9668, + "ĠLesser": 13907, + "ĠLesson": 48915, + "ĠLessons": 60466, + "ĠLester": 82658, + "ĠLet": 6771, + "ĠLets": 57066, + "ĠLetter": 26659, + "ĠLetters": 44297, + "ĠLev": 27022, + "ĠLevel": 9395, + "ĠLevels": 52693, + "ĠLever": 78579, + "ĠLevi": 55041, + "ĠLevin": 66877, + "ĠLevine": 80602, + "ĠLevitra": 69148, + "ĠLevy": 67190, + "ĠLew": 17195, + "ĠLewis": 20607, + "ĠLex": 26819, + "ĠLexer": 85082, + "ĠLexington": 70658, + "ĠLexus": 87869, + "ĠLey": 53425, + "ĠLi": 14511, + "ĠLia": 83657, + "ĠLiability": 90043, + "ĠLiam": 60642, + "ĠLib": 5688, + "ĠLiber": 94297, + "ĠLiberal": 30058, + "ĠLiberals": 51381, + "ĠLiberation": 71884, + "ĠLiberia": 83525, + "ĠLibert": 51081, + "ĠLibertarian": 84723, + "ĠLiberties": 91486, + "ĠLiberty": 31392, + "ĠLibraries": 54512, + "ĠLibrary": 11635, + "ĠLibre": 75168, + "ĠLibya": 36959, + "ĠLibyan": 78497, + "ĠLic": 32575, + "ĠLicence": 53275, + "ĠLicense": 1876, + "ĠLicensed": 10103, + "ĠLicensing": 72325, + "ĠLicht": 88937, + "ĠLid": 80609, + "ĠLie": 21491, + "ĠLiebe": 93719, + "ĠLies": 77086, + "ĠLieutenant": 54137, + "ĠLif": 64763, + "ĠLife": 9414, + "ĠLifecycle": 74392, + "ĠLifestyle": 68156, + "ĠLifetime": 64352, + "ĠLift": 52435, + "ĠLig": 98178, + "ĠLiga": 59027, + "ĠLight": 8658, + "ĠLighting": 44523, + "ĠLightning": 33844, + "ĠLights": 34170, + "ĠLightweight": 83267, + "ĠLik": 32929, + "ĠLike": 8909, + "ĠLikely": 93637, + "ĠLikes": 50914, + "ĠLikewise": 49387, + "ĠLil": 40478, + "ĠLilly": 81834, + "ĠLily": 47290, + "ĠLim": 13589, + "ĠLima": 61953, + "ĠLime": 70496, + "ĠLimit": 28008, + "ĠLimited": 18905, + "ĠLimits": 71855, + "ĠLin": 8564, + "ĠLincoln": 24379, + "ĠLind": 27219, + "ĠLinda": 38062, + "ĠLinden": 99065, + "ĠLindsay": 62784, + "ĠLindsey": 64279, + "ĠLine": 7083, + "ĠLinear": 28263, + "ĠLinearGradient": 96492, + "ĠLinearLayout": 27146, + "ĠLinearLayoutManager": 45851, + "ĠLines": 38233, + "ĠLing": 50858, + "ĠLingu": 88233, + "ĠLink": 5948, + "ĠLinked": 13114, + "ĠLinkedHashMap": 64333, + "ĠLinkedIn": 32767, + "ĠLinkedList": 22917, + "ĠLinkedin": 91939, + "ĠLinks": 25422, + "ĠLinux": 14340, + "ĠLion": 32099, + "ĠLionel": 83124, + "ĠLions": 32765, + "ĠLip": 41139, + "ĠLips": 95263, + "ĠLiqu": 68518, + "ĠLiquid": 41258, + "ĠLis": 43072, + "ĠLisa": 28556, + "ĠLisbon": 80701, + "ĠLisp": 54975, + "ĠList": 1759, + "ĠListBox": 98206, + "ĠListItem": 40924, + "ĠListNode": 20501, + "ĠListTile": 63421, + "ĠListView": 26636, + "ĠListViewItem": 95659, + "ĠLista": 53002, + "ĠListe": 73086, + "ĠListed": 68448, + "ĠListen": 32149, + "ĠListener": 47999, + "ĠListening": 76697, + "ĠListing": 38764, + "ĠListings": 83884, + "ĠLists": 37905, + "ĠLit": 38251, + "ĠLite": 40865, + "ĠLitecoin": 94092, + "ĠLiter": 25471, + "ĠLiteral": 49674, + "ĠLiterary": 73624, + "ĠLiterature": 46370, + "ĠLith": 40578, + "ĠLithuania": 70506, + "ĠLittle": 14671, + "ĠLiu": 37705, + "ĠLiv": 36211, + "ĠLive": 11158, + "ĠLiveData": 62291, + "ĠLiver": 21859, + "ĠLiverpool": 24688, + "ĠLives": 39395, + "ĠLiving": 18539, + "ĠLivingston": 93974, + "ĠLiz": 47199, + "ĠLiên": 130578, + "ĠLloyd": 45196, + "ĠLn": 97819, + "ĠLo": 6485, + "ĠLoad": 8893, + "ĠLoaded": 62639, + "ĠLoader": 27811, + "ĠLoading": 17937, + "ĠLoads": 51223, + "ĠLoan": 35081, + "ĠLoans": 47515, + "ĠLob": 67517, + "ĠLobby": 62966, + "ĠLoc": 13600, + "ĠLocal": 8774, + "ĠLocalDate": 40816, + "ĠLocalDateTime": 47465, + "ĠLocale": 34526, + "ĠLocalization": 80820, + "ĠLocate": 95210, + "ĠLocated": 41767, + "ĠLocation": 9866, + "ĠLocations": 41469, + "ĠLocator": 98653, + "ĠLoch": 82703, + "ĠLock": 15701, + "ĠLocke": 81922, + "ĠLocked": 92945, + "ĠLocker": 93978, + "ĠLockheed": 84136, + "ĠLod": 87940, + "ĠLodge": 43568, + "ĠLoft": 83154, + "ĠLog": 2835, + "ĠLogLevel": 63321, + "ĠLogManager": 69365, + "ĠLogan": 43046, + "ĠLogged": 57055, + "ĠLogger": 9514, + "ĠLoggerFactory": 35086, + "ĠLogging": 41706, + "ĠLogic": 36101, + "ĠLogical": 62069, + "ĠLogin": 8779, + "ĠLoginActivity": 98346, + "ĠLoginComponent": 59754, + "ĠLoginForm": 69995, + "ĠLoginPage": 63411, + "ĠLogistic": 78068, + "ĠLogistics": 70322, + "ĠLogo": 30052, + "ĠLogout": 46285, + "ĠLogs": 54570, + "ĠLois": 83946, + "ĠLok": 57401, + "ĠLoki": 83755, + "ĠLol": 79137, + "ĠLola": 96357, + "ĠLomb": 62818, + "ĠLon": 48181, + "ĠLond": 79593, + "ĠLondon": 7148, + "ĠLone": 75181, + "ĠLonely": 85723, + "ĠLong": 5724, + "ĠLonger": 93326, + "ĠLongitude": 79109, + "ĠLook": 9192, + "ĠLooking": 21131, + "ĠLooks": 41806, + "ĠLookup": 50311, + "ĠLoop": 21359, + "ĠLoose": 86002, + "ĠLoot": 83054, + "ĠLopez": 44215, + "ĠLor": 68730, + "ĠLord": 10214, + "ĠLords": 56048, + "ĠLore": 53233, + "ĠLorem": 46931, + "ĠLoren": 51190, + "ĠLorenzo": 73400, + "ĠLori": 81085, + "ĠLos": 9656, + "ĠLose": 60689, + "ĠLosing": 98139, + "ĠLoss": 24704, + "ĠLost": 27252, + "ĠLot": 21755, + "ĠLots": 40215, + "ĠLottery": 72237, + "ĠLotto": 97713, + "ĠLotus": 60169, + "ĠLou": 9729, + "ĠLoud": 79548, + "ĠLouis": 11876, + "ĠLouise": 55478, + "ĠLouisiana": 28874, + "ĠLouisville": 45034, + "ĠLounge": 49667, + "ĠLov": 34293, + "ĠLove": 10689, + "ĠLoved": 84027, + "ĠLovely": 64365, + "ĠLover": 89490, + "ĠLoves": 84023, + "ĠLoving": 89322, + "ĠLow": 12041, + "ĠLowe": 64587, + "ĠLowell": 95130, + "ĠLower": 27536, + "ĠLowest": 68530, + "ĠLoy": 67242, + "ĠLt": 41870, + "ĠLtd": 12324, + "ĠLu": 18360, + "ĠLua": 37662, + "ĠLub": 57651, + "ĠLuc": 13784, + "ĠLuca": 83178, + "ĠLucas": 31003, + "ĠLucia": 79174, + "ĠLucifer": 93394, + "ĠLuck": 28772, + "ĠLuckily": 47500, + "ĠLucky": 48567, + "ĠLucy": 44070, + "ĠLud": 45170, + "ĠLudwig": 84450, + "ĠLuft": 88049, + "ĠLug": 92490, + "ĠLuigi": 82083, + "ĠLuis": 33197, + "ĠLuk": 54657, + "ĠLuke": 24453, + "ĠLum": 42601, + "ĠLumia": 91351, + "ĠLump": 74293, + "ĠLumpur": 86244, + "ĠLun": 44349, + "ĠLuna": 49805, + "ĠLunar": 78095, + "ĠLunch": 55599, + "ĠLund": 68181, + "ĠLung": 77236, + "ĠLuo": 66383, + "ĠLup": 78703, + "ĠLust": 73255, + "ĠLuther": 35202, + "ĠLutheran": 81981, + "ĠLux": 26369, + "ĠLuxembourg": 64871, + "ĠLuxury": 51250, + "ĠLuz": 81639, + "ĠLuáºŃt": 136786, + "ĠLv": 62205, + "ĠLy": 15953, + "ĠLydia": 91206, + "ĠLyft": 84439, + "ĠLyme": 93116, + "ĠLyn": 21136, + "ĠLynch": 37106, + "ĠLynn": 44538, + "ĠLyon": 55201, + "ĠLyons": 90966, + "ĠLyrics": 56243, + "ĠLâm": 134925, + "ĠLänder": 140275, + "ĠLê": 131699, + "ĠLópez": 90975, + "ĠLö": 132678, + "ĠLös": 89452, + "ĠLúc": 138933, + "ĠLý": 134129, + "ĠLÃł": 127314, + "ĠLÃłm": 134330, + "ĠLưu": 138019, + "ĠLá»ĭch": 141034, + "ĠM": 386, + "ĠMA": 9718, + "ĠMAC": 22884, + "ĠMACHINE": 86011, + "ĠMAD": 75373, + "ĠMADE": 78699, + "ĠMAG": 49520, + "ĠMAGIC": 59249, + "ĠMAIL": 75384, + "ĠMAIN": 34339, + "ĠMAK": 71411, + "ĠMAKE": 36821, + "ĠMAL": 88039, + "ĠMAN": 25735, + "ĠMANAGEMENT": 93278, + "ĠMANUAL": 92423, + "ĠMANY": 98256, + "ĠMAP": 27223, + "ĠMAR": 37499, + "ĠMARK": 18018, + "ĠMART": 89565, + "ĠMAS": 87048, + "ĠMASK": 55094, + "ĠMASS": 92613, + "ĠMASTER": 71197, + "ĠMAT": 24795, + "ĠMATCH": 48171, + "ĠMATERIAL": 75123, + "ĠMATLAB": 49347, + "ĠMATRIX": 68559, + "ĠMAV": 80819, + "ĠMAX": 8334, + "ĠMAY": 39230, + "ĠMB": 13339, + "ĠMBA": 53454, + "ĠMC": 20869, + "ĠMCC": 98252, + "ĠMCP": 79148, + "ĠMCS": 93193, + "ĠMCU": 66227, + "ĠMD": 13979, + "ĠMDB": 86352, + "ĠMDMA": 98728, + "ĠME": 16292, + "ĠMED": 51599, + "ĠMEDIA": 60186, + "ĠMEDIATEK": 50709, + "ĠMEM": 24478, + "ĠMEMBER": 81824, + "ĠMEMORY": 80962, + "ĠMEN": 67928, + "ĠMENU": 35697, + "ĠMEP": 90836, + "ĠMER": 72373, + "ĠMERCHANTABILITY": 7935, + "ĠMES": 92318, + "ĠMESSAGE": 40758, + "ĠMET": 36822, + "ĠMETA": 79341, + "ĠMETH": 77824, + "ĠMETHOD": 32397, + "ĠMETHODS": 55229, + "ĠMF": 38254, + "ĠMG": 51192, + "ĠMGM": 98391, + "ĠMH": 45560, + "ĠMHz": 36494, + "ĠMI": 18476, + "ĠMIC": 44741, + "ĠMICRO": 86165, + "ĠMID": 69784, + "ĠMIDI": 47039, + "ĠMIL": 41809, + "ĠMILF": 73298, + "ĠMILL": 82485, + "ĠMIME": 57477, + "ĠMIN": 16701, + "ĠMIPS": 98662, + "ĠMIS": 67025, + "ĠMISS": 62747, + "ĠMISSING": 98779, + "ĠMIT": 14865, + "ĠMIX": 80209, + "ĠMJ": 60945, + "ĠMK": 26859, + "ĠML": 19614, + "ĠMLA": 79231, + "ĠMLB": 40117, + "ĠMLM": 90182, + "ĠMLP": 94891, + "ĠMLS": 28898, + "ĠMM": 21665, + "ĠMMA": 42567, + "ĠMMC": 67935, + "ĠMMI": 93087, + "ĠMMM": 56105, + "ĠMMO": 77185, + "ĠMMP": 89204, + "ĠMN": 34995, + "ĠMO": 11418, + "ĠMOCK": 83015, + "ĠMOD": 18669, + "ĠMODE": 22690, + "ĠMODEL": 47970, + "ĠMODIFY": 69665, + "ĠMODULE": 40374, + "ĠMOM": 92972, + "ĠMON": 28537, + "ĠMONEY": 93462, + "ĠMONTH": 57980, + "ĠMOR": 71313, + "ĠMORE": 18447, + "ĠMOS": 73074, + "ĠMOST": 79099, + "ĠMOT": 45847, + "ĠMOTOR": 86272, + "ĠMOUSE": 83190, + "ĠMOV": 37628, + "ĠMOVE": 34926, + "ĠMOZ": 63744, + "ĠMP": 9412, + "ĠMPC": 91598, + "ĠMPEG": 69989, + "ĠMPG": 90398, + "ĠMPH": 87656, + "ĠMPI": 17104, + "ĠMPL": 47356, + "ĠMPS": 92943, + "ĠMPU": 61323, + "ĠMPs": 36661, + "ĠMQ": 42983, + "ĠMQTT": 61424, + "ĠMR": 28333, + "ĠMRI": 51360, + "ĠMS": 10292, + "ĠMSC": 83513, + "ĠMSD": 71395, + "ĠMSE": 94599, + "ĠMSG": 23317, + "ĠMSI": 67398, + "ĠMSM": 67491, + "ĠMSNBC": 55815, + "ĠMSP": 68373, + "ĠMSR": 99068, + "ĠMSS": 91649, + "ĠMST": 84280, + "ĠMT": 19087, + "ĠMTV": 61099, + "ĠMU": 59786, + "ĠMUCH": 78333, + "ĠMULT": 50154, + "ĠMULTI": 82737, + "ĠMUSIC": 74973, + "ĠMUST": 27732, + "ĠMUT": 87420, + "ĠMV": 42271, + "ĠMVC": 67996, + "ĠMVP": 31866, + "ĠMW": 44482, + "ĠMX": 27585, + "ĠMY": 18224, + "ĠMYSQL": 54197, + "ĠMa": 11331, + "ĠMaar": 83967, + "ĠMac": 7401, + "ĠMacBook": 57579, + "ĠMacDonald": 80873, + "ĠMacOS": 89717, + "ĠMaced": 56452, + "ĠMacedonia": 76409, + "ĠMach": 18294, + "ĠMachine": 12960, + "ĠMachinery": 44106, + "ĠMachines": 43904, + "ĠMack": 32391, + "ĠMacro": 53317, + "ĠMacron": 62301, + "ĠMacros": 98447, + "ĠMacy": 96273, + "ĠMad": 9483, + "ĠMadagascar": 94628, + "ĠMadame": 83176, + "ĠMadd": 43537, + "ĠMadden": 83583, + "ĠMade": 18802, + "ĠMadison": 29915, + "ĠMadness": 70328, + "ĠMadonna": 80569, + "ĠMadrid": 24081, + "ĠMaduro": 77702, + "ĠMae": 64814, + "ĠMafia": 80895, + "ĠMag": 6879, + "ĠMagazine": 21451, + "ĠMage": 23371, + "ĠMagento": 35585, + "ĠMaggie": 61695, + "ĠMagic": 15485, + "ĠMagicMock": 89708, + "ĠMagical": 72710, + "ĠMagick": 87154, + "ĠMagn": 20300, + "ĠMagnet": 81228, + "ĠMagnetic": 62655, + "ĠMagnum": 93861, + "ĠMagnus": 73799, + "ĠMah": 16171, + "ĠMahar": 52590, + "ĠMaharashtra": 67419, + "ĠMaher": 96168, + "ĠMahm": 71611, + "ĠMahmoud": 93186, + "ĠMai": 42256, + "ĠMaid": 47085, + "ĠMaiden": 82842, + "ĠMail": 14874, + "ĠMain": 4697, + "ĠMainActivity": 27457, + "ĠMainAxisAlignment": 33214, + "ĠMainForm": 74677, + "ĠMainMenu": 77171, + "ĠMainPage": 70076, + "ĠMainWindow": 23846, + "ĠMaine": 29790, + "ĠMaint": 42887, + "ĠMaintain": 86377, + "ĠMaintenance": 38206, + "ĠMais": 33347, + "ĠMaison": 87883, + "ĠMaj": 30679, + "ĠMajesty": 66244, + "ĠMajor": 17121, + "ĠMajority": 54035, + "ĠMak": 39324, + "ĠMake": 7405, + "ĠMaker": 40528, + "ĠMakes": 36870, + "ĠMakeup": 72914, + "ĠMaking": 24288, + "ĠMal": 8394, + "ĠMalay": 79140, + "ĠMalays": 22914, + "ĠMalaysia": 27696, + "ĠMalaysian": 65431, + "ĠMalcolm": 46969, + "ĠMale": 19396, + "ĠMali": 63894, + "ĠMalik": 72220, + "ĠMalk": 97891, + "ĠMall": 31398, + "ĠMalone": 83071, + "ĠMalta": 60650, + "ĠMam": 47769, + "ĠMama": 60706, + "ĠMan": 2363, + "ĠMana": 49338, + "ĠManafort": 57813, + "ĠManage": 42650, + "ĠManaged": 60744, + "ĠManagement": 9551, + "ĠManager": 10567, + "ĠManagers": 61434, + "ĠManaging": 48307, + "ĠManchester": 19361, + "ĠMand": 23849, + "ĠMandal": 88879, + "ĠMandarin": 82771, + "ĠMandatory": 70782, + "ĠMandela": 87330, + "ĠMane": 78074, + "ĠMang": 59048, + "ĠManga": 69935, + "ĠMango": 90863, + "ĠManhattan": 28790, + "ĠManifest": 39361, + "ĠManila": 56564, + "ĠManip": 60911, + "ĠManit": 53235, + "ĠManitoba": 63240, + "ĠMann": 29860, + "ĠManning": 39523, + "ĠManny": 94155, + "ĠManor": 69951, + "ĠMans": 43135, + "ĠMansion": 82296, + "ĠManson": 98133, + "ĠMant": 50968, + "ĠManual": 14539, + "ĠManuals": 58841, + "ĠManuel": 43581, + "ĠManufact": 17657, + "ĠManufacturer": 34451, + "ĠManufacturers": 46098, + "ĠManufacturing": 41077, + "ĠManus": 95333, + "ĠMany": 8999, + "ĠMao": 56178, + "ĠMap": 5027, + "ĠMaple": 43470, + "ĠMapper": 45734, + "ĠMapping": 38446, + "ĠMaps": 27408, + "ĠMar": 2876, + "ĠMara": 85504, + "ĠMarathon": 50173, + "ĠMarble": 72521, + "ĠMarc": 18880, + "ĠMarcel": 48871, + "ĠMarch": 5470, + "ĠMarco": 34181, + "ĠMarcos": 65505, + "ĠMarcus": 35683, + "ĠMare": 63798, + "ĠMarg": 23201, + "ĠMargaret": 37549, + "ĠMargin": 71124, + "ĠMari": 28729, + "ĠMaria": 23016, + "ĠMarian": 95419, + "ĠMarie": 32016, + "ĠMarijuana": 64839, + "ĠMarilyn": 76360, + "ĠMarin": 63377, + "ĠMarina": 51536, + "ĠMarine": 22963, + "ĠMariners": 81862, + "ĠMarines": 50789, + "ĠMarino": 88266, + "ĠMario": 23381, + "ĠMarion": 55579, + "ĠMaritime": 76231, + "ĠMark": 4389, + "ĠMarkdown": 73192, + "ĠMarker": 39875, + "ĠMarket": 7993, + "ĠMarketable": 23058, + "ĠMarketing": 18228, + "ĠMarketplace": 57258, + "ĠMarkets": 46810, + "ĠMarks": 48095, + "ĠMarkt": 98317, + "ĠMarkup": 98561, + "ĠMarkus": 75840, + "ĠMarl": 57667, + "ĠMarlins": 88890, + "ĠMarr": 68598, + "ĠMarriage": 46099, + "ĠMarried": 64524, + "ĠMarriott": 75822, + "ĠMars": 21048, + "ĠMarseille": 87693, + "ĠMarsh": 39063, + "ĠMarshal": 35667, + "ĠMarshall": 29408, + "ĠMart": 8127, + "ĠMartha": 50870, + "ĠMartial": 71261, + "ĠMartian": 81028, + "ĠMartin": 11565, + "ĠMartinez": 42927, + "ĠMartins": 89699, + "ĠMarty": 55154, + "ĠMarvel": 22105, + "ĠMarvin": 73526, + "ĠMarx": 27088, + "ĠMarxism": 82715, + "ĠMarxist": 68950, + "ĠMary": 10244, + "ĠMaryland": 22652, + "ĠMarÃŃa": 82205, + "ĠMas": 19868, + "ĠMasc": 88898, + "ĠMash": 68179, + "ĠMask": 19924, + "ĠMasks": 61132, + "ĠMason": 28827, + "ĠMass": 9167, + "ĠMassachusetts": 21393, + "ĠMassage": 26385, + "ĠMassive": 81464, + "ĠMast": 36482, + "ĠMaster": 10824, + "ĠMasters": 33622, + "ĠMastery": 93043, + "ĠMat": 6867, + "ĠMatButtonModule": 93692, + "ĠMatDialog": 52475, + "ĠMatSnackBar": 98509, + "ĠMatTable": 84447, + "ĠMata": 88354, + "ĠMatch": 14152, + "ĠMatcher": 60632, + "ĠMatchers": 92709, + "ĠMatches": 61254, + "ĠMatching": 70691, + "ĠMate": 43570, + "ĠMater": 98308, + "ĠMaterial": 10230, + "ĠMaterialApp": 82548, + "ĠMaterialPageRoute": 55267, + "ĠMaterials": 30909, + "ĠMath": 4149, + "ĠMathematic": 71040, + "ĠMathematical": 91002, + "ĠMathematics": 49795, + "ĠMathf": 26196, + "ĠMaths": 92578, + "ĠMatlab": 97589, + "ĠMatrix": 11631, + "ĠMats": 67179, + "ĠMatt": 13374, + "ĠMatte": 64082, + "ĠMatter": 33364, + "ĠMatters": 57264, + "ĠMatth": 68198, + "ĠMatthew": 18939, + "ĠMatthews": 50684, + "ĠMatthias": 89487, + "ĠMattis": 86354, + "ĠMature": 42159, + "ĠMaui": 92450, + "ĠMaul": 93201, + "ĠMaur": 33392, + "ĠMaurice": 67429, + "ĠMaurit": 88213, + "ĠMaven": 98575, + "ĠMaver": 64112, + "ĠMavericks": 86156, + "ĠMax": 7487, + "ĠMaxim": 55525, + "ĠMaximum": 26599, + "ĠMaxwell": 58397, + "ĠMay": 3217, + "ĠMaya": 50344, + "ĠMaybe": 10696, + "ĠMayer": 73711, + "ĠMayo": 57057, + "ĠMayor": 22093, + "ĠMayweather": 97334, + "ĠMayıs": 134032, + "ĠMaz": 38991, + "ĠMazda": 61942, + "ĠMaze": 65091, + "ĠMb": 50265, + "ĠMbps": 79475, + "ĠMc": 4483, + "ĠMcA": 64713, + "ĠMcB": 98874, + "ĠMcC": 14249, + "ĠMcCabe": 84958, + "ĠMcCain": 35535, + "ĠMcCart": 86458, + "ĠMcCarthy": 43940, + "ĠMcCartney": 90393, + "ĠMcCl": 59640, + "ĠMcConnell": 46683, + "ĠMcCorm": 87971, + "ĠMcCoy": 70632, + "ĠMcD": 50253, + "ĠMcDon": 71574, + "ĠMcDonald": 30914, + "ĠMcG": 22927, + "ĠMcGill": 92976, + "ĠMcGr": 87750, + "ĠMcGregor": 68016, + "ĠMcGu": 83064, + "ĠMcInt": 84481, + "ĠMcK": 25636, + "ĠMcKay": 89870, + "ĠMcKenzie": 92623, + "ĠMcKin": 66393, + "ĠMcL": 45607, + "ĠMcLaren": 71364, + "ĠMcM": 81714, + "ĠMcMahon": 79828, + "ĠMcMaster": 99042, + "ĠMcN": 54651, + "ĠMd": 53110, + "ĠMe": 2157, + "ĠMeadow": 88742, + "ĠMeadows": 77042, + "ĠMeal": 56388, + "ĠMeals": 95250, + "ĠMean": 16327, + "ĠMeaning": 48103, + "ĠMeans": 59707, + "ĠMeanwhile": 25894, + "ĠMeasure": 34104, + "ĠMeasurement": 54240, + "ĠMeasurements": 76817, + "ĠMeasures": 72556, + "ĠMeat": 44865, + "ĠMech": 90498, + "ĠMechan": 27801, + "ĠMechanical": 50584, + "ĠMechanics": 76823, + "ĠMed": 3264, + "ĠMedal": 17411, + "ĠMedi": 91757, + "ĠMedia": 7816, + "ĠMediaPlayer": 71010, + "ĠMediaQuery": 42591, + "ĠMediaTek": 84902, + "ĠMediaType": 50423, + "ĠMedian": 62590, + "ĠMedic": 12566, + "ĠMedicaid": 33029, + "ĠMedical": 12939, + "ĠMedicare": 30722, + "ĠMedicine": 18637, + "ĠMedieval": 77148, + "ĠMedina": 84166, + "ĠMeditation": 76471, + "ĠMediterr": 36437, + "ĠMediterranean": 37685, + "ĠMedium": 24355, + "ĠMeer": 94716, + "ĠMeet": 25826, + "ĠMeeting": 29055, + "ĠMeetings": 85583, + "ĠMeg": 27343, + "ĠMega": 34256, + "ĠMegan": 54529, + "ĠMeghan": 79034, + "ĠMeh": 67224, + "ĠMehr": 54412, + "ĠMei": 90933, + "ĠMein": 59330, + "ĠMej": 86145, + "ĠMel": 10978, + "ĠMelania": 82170, + "ĠMelanie": 84250, + "ĠMelbourne": 26437, + "ĠMelee": 63239, + "ĠMelissa": 47048, + "ĠMell": 60185, + "ĠMellon": 87387, + "ĠMem": 13550, + "ĠMember": 12039, + "ĠMembers": 16954, + "ĠMembership": 42552, + "ĠMemo": 56201, + "ĠMemor": 72666, + "ĠMemorial": 26774, + "ĠMemories": 86616, + "ĠMemory": 13850, + "ĠMemoryStream": 36598, + "ĠMemphis": 42062, + "ĠMen": 11012, + "ĠMend": 45111, + "ĠMeng": 60054, + "ĠMens": 23276, + "ĠMenschen": 42786, + "ĠMent": 48593, + "ĠMental": 37795, + "ĠMention": 85148, + "ĠMentor": 91191, + "ĠMenu": 9738, + "ĠMenuItem": 31274, + "ĠMer": 8755, + "ĠMerc": 24171, + "ĠMercedes": 33228, + "ĠMercer": 80011, + "ĠMerch": 65025, + "ĠMerchant": 55399, + "ĠMercury": 43562, + "ĠMercy": 63694, + "ĠMeredith": 83677, + "ĠMerge": 40085, + "ĠMerkel": 46420, + "ĠMerlin": 75966, + "ĠMerr": 47861, + "ĠMerrill": 87683, + "ĠMerry": 76701, + "ĠMes": 35584, + "ĠMesa": 54865, + "ĠMesh": 25122, + "ĠMess": 18713, + "ĠMessage": 4856, + "ĠMessageBox": 10299, + "ĠMessageBoxButton": 97603, + "ĠMessageBoxButtons": 31407, + "ĠMessageBoxIcon": 34036, + "ĠMessageLookup": 98774, + "ĠMessageType": 62336, + "ĠMessages": 26729, + "ĠMessaging": 73170, + "ĠMessenger": 41519, + "ĠMessi": 70005, + "ĠMessiah": 80341, + "ĠMet": 6212, + "ĠMeta": 15819, + "ĠMetadata": 33589, + "ĠMetal": 19206, + "ĠMetallic": 86871, + "ĠMetals": 92715, + "ĠMeteor": 39632, + "ĠMeter": 52539, + "ĠMeth": 56075, + "ĠMethod": 6730, + "ĠMethodInfo": 23837, + "ĠMethodInvocation": 66672, + "ĠMethodist": 67501, + "ĠMethods": 18801, + "ĠMetodo": 94656, + "ĠMetric": 52458, + "ĠMetrics": 54190, + "ĠMetro": 20957, + "ĠMetroFramework": 51355, + "ĠMetropolitan": 44778, + "ĠMets": 56435, + "ĠMex": 9369, + "ĠMexican": 23279, + "ĠMexicans": 96762, + "ĠMexico": 12270, + "ĠMey": 83862, + "ĠMeyer": 47190, + "ĠMg": 72593, + "ĠMi": 20740, + "ĠMia": 60597, + "ĠMiami": 17584, + "ĠMic": 26997, + "ĠMich": 5271, + "ĠMicha": 69920, + "ĠMichael": 7937, + "ĠMichaels": 86756, + "ĠMiche": 62700, + "ĠMichel": 44406, + "ĠMichele": 73306, + "ĠMichelle": 32026, + "ĠMichigan": 14630, + "ĠMick": 59233, + "ĠMickey": 56559, + "ĠMicro": 18157, + "ĠMicrosoft": 5100, + "ĠMicrosystems": 94661, + "ĠMicrowave": 97541, + "ĠMid": 13699, + "ĠMiddle": 12592, + "ĠMiddleton": 91614, + "ĠMiddleware": 72012, + "ĠMidi": 83076, + "ĠMidlands": 80151, + "ĠMidnight": 63580, + "ĠMidwest": 51157, + "ĠMig": 37551, + "ĠMight": 33251, + "ĠMighty": 74168, + "ĠMigration": 21248, + "ĠMiguel": 45891, + "ĠMik": 42242, + "ĠMike": 11268, + "ĠMikhail": 91451, + "ĠMil": 10147, + "ĠMilan": 38755, + "ĠMilano": 74867, + "ĠMild": 60189, + "ĠMile": 38597, + "ĠMiles": 35203, + "ĠMilf": 51558, + "ĠMilit": 67185, + "ĠMilitary": 26960, + "ĠMilk": 43825, + "ĠMilky": 88719, + "ĠMill": 8220, + "ĠMillenn": 51328, + "ĠMillennials": 96643, + "ĠMillennium": 70937, + "ĠMiller": 17035, + "ĠMillion": 33529, + "ĠMillionen": 98435, + "ĠMillions": 73318, + "ĠMills": 33102, + "ĠMilo": 84446, + "ĠMilton": 57347, + "ĠMilwaukee": 35818, + "ĠMim": 91846, + "ĠMime": 82430, + "ĠMin": 3386, + "ĠMind": 22904, + "ĠMinds": 60310, + "ĠMine": 30683, + "ĠMinecraft": 40230, + "ĠMiner": 89912, + "ĠMineral": 49316, + "ĠMinerals": 83786, + "ĠMines": 65782, + "ĠMing": 55883, + "ĠMinh": 92155, + "ĠMini": 19643, + "ĠMinimal": 75112, + "ĠMinimum": 30925, + "ĠMining": 25832, + "ĠMinist": 64892, + "ĠMinister": 9486, + "ĠMinisters": 77973, + "ĠMinistry": 19640, + "ĠMinistério": 139506, + "ĠMinneapolis": 41858, + "ĠMinnesota": 18925, + "ĠMinor": 29793, + "ĠMinority": 70058, + "ĠMint": 41310, + "ĠMinute": 55167, + "ĠMinuten": 80624, + "ĠMinutes": 40999, + "ĠMir": 14268, + "ĠMiracle": 75964, + "ĠMirage": 98142, + "ĠMiranda": 58314, + "ĠMirror": 33854, + "ĠMis": 32559, + "ĠMisc": 39051, + "ĠMiscellaneous": 78277, + "ĠMish": 77476, + "ĠMiss": 8906, + "ĠMissile": 67770, + "ĠMissing": 35264, + "ĠMission": 23220, + "ĠMississippi": 28438, + "ĠMissouri": 24378, + "ĠMist": 34098, + "ĠMister": 91608, + "ĠMistress": 96029, + "ĠMit": 21766, + "ĠMitar": 78722, + "ĠMitarbeiter": 96428, + "ĠMitch": 22586, + "ĠMitchell": 29938, + "ĠMitgli": 77094, + "ĠMits": 59576, + "ĠMitsubishi": 78553, + "ĠMitt": 32618, + "ĠMix": 19219, + "ĠMixed": 50168, + "ĠMixer": 71292, + "ĠMixing": 96599, + "ĠMiy": 66459, + "ĠMiz": 71595, + "ĠMk": 52848, + "ĠMn": 56731, + "ĠMo": 6050, + "ĠMob": 34906, + "ĠMobil": 69830, + "ĠMobile": 13411, + "ĠMobility": 62310, + "ĠMock": 14563, + "ĠMockito": 47109, + "ĠMod": 5650, + "ĠModal": 21311, + "ĠMode": 14562, + "ĠModel": 4903, + "ĠModelAndView": 35068, + "ĠModelRenderer": 81010, + "ĠModelState": 70048, + "ĠModeling": 76249, + "ĠModelo": 90518, + "ĠModels": 26874, + "ĠModer": 43427, + "ĠModerate": 94975, + "ĠModerator": 91442, + "ĠModern": 18265, + "ĠModes": 81689, + "ĠModi": 34389, + "ĠModification": 73855, + "ĠModifications": 95616, + "ĠModified": 29053, + "ĠModifier": 45208, + "ĠModify": 49370, + "ĠMods": 85064, + "ĠModular": 84848, + "ĠModule": 13711, + "ĠModules": 43565, + "ĠMoff": 96381, + "ĠMog": 59921, + "ĠMoh": 19822, + "ĠMohamed": 59659, + "ĠMohammad": 64452, + "ĠMohammed": 45292, + "ĠMoines": 77058, + "ĠMoist": 81704, + "ĠMoj": 89337, + "ĠMojo": 95649, + "ĠMol": 32418, + "ĠMold": 54163, + "ĠMolecular": 59725, + "ĠMolly": 57400, + "ĠMom": 22180, + "ĠMoment": 38996, + "ĠMoments": 82021, + "ĠMomentum": 95110, + "ĠMoms": 83295, + "ĠMon": 3128, + "ĠMona": 98783, + "ĠMonaco": 61551, + "ĠMonad": 59685, + "ĠMond": 50872, + "ĠMonday": 7014, + "ĠMondays": 90171, + "ĠMonetary": 73114, + "ĠMoney": 17633, + "ĠMong": 43439, + "ĠMongo": 24415, + "ĠMongoClient": 60432, + "ĠMongoDB": 45328, + "ĠMongolia": 90750, + "ĠMonica": 45744, + "ĠMonitor": 23519, + "ĠMonitoring": 50703, + "ĠMonk": 75681, + "ĠMonkey": 57837, + "ĠMono": 12691, + "ĠMonoBehaviour": 15714, + "ĠMonroe": 49787, + "ĠMons": 49501, + "ĠMonsanto": 72926, + "ĠMonster": 27973, + "ĠMonsters": 62708, + "ĠMont": 9795, + "ĠMontana": 36005, + "ĠMonte": 45767, + "ĠMonter": 64998, + "ĠMonterey": 97395, + "ĠMontgomery": 43025, + "ĠMonth": 19397, + "ĠMonthly": 44334, + "ĠMonths": 51346, + "ĠMontreal": 29513, + "ĠMonument": 71277, + "ĠMoo": 81996, + "ĠMood": 93567, + "ĠMoodle": 59443, + "ĠMoody": 71301, + "ĠMoon": 17330, + "ĠMoor": 83448, + "ĠMoore": 20219, + "ĠMoose": 68128, + "ĠMor": 8446, + "ĠMoral": 88506, + "ĠMorales": 75502, + "ĠMoran": 92235, + "ĠMore": 4398, + "ĠMoreno": 78175, + "ĠMoreover": 22827, + "ĠMorg": 78060, + "ĠMorgan": 22954, + "ĠMori": 94197, + "ĠMorm": 31294, + "ĠMormon": 36597, + "ĠMormons": 91873, + "ĠMorning": 27984, + "ĠMoroccan": 93967, + "ĠMorocco": 58130, + "ĠMorph": 74640, + "ĠMorr": 44857, + "ĠMorris": 29183, + "ĠMorrison": 48954, + "ĠMorrow": 86274, + "ĠMorse": 86239, + "ĠMort": 22031, + "ĠMortgage": 47680, + "ĠMorton": 71936, + "ĠMos": 12562, + "ĠMoscow": 22415, + "ĠMoses": 40948, + "ĠMosque": 98931, + "ĠMoss": 43510, + "ĠMost": 7496, + "ĠMostly": 63185, + "ĠMosul": 61813, + "ĠMot": 18977, + "ĠMother": 21043, + "ĠMothers": 84935, + "ĠMotion": 26562, + "ĠMotionEvent": 56957, + "ĠMoto": 47498, + "ĠMotor": 17614, + "ĠMotorcycle": 69662, + "ĠMotorola": 62490, + "ĠMotors": 36692, + "ĠMou": 72834, + "ĠMount": 10423, + "ĠMountain": 18634, + "ĠMountains": 40014, + "ĠMounted": 92741, + "ĠMour": 50548, + "ĠMourinho": 65318, + "ĠMouse": 17721, + "ĠMouseButton": 91203, + "ĠMouseEvent": 66217, + "ĠMouseEventArgs": 81904, + "ĠMouth": 65155, + "ĠMov": 53285, + "ĠMove": 14561, + "ĠMoved": 89697, + "ĠMovement": 27998, + "ĠMoves": 56599, + "ĠMovie": 13944, + "ĠMovies": 25930, + "ĠMoving": 37627, + "ĠMoy": 76538, + "ĠMoz": 51443, + "ĠMozart": 94669, + "ĠMozilla": 34733, + "ĠMoż": 141416, + "ĠMoże": 133590, + "ĠMożna": 141417, + "ĠMp": 49965, + "ĠMr": 4392, + "ĠMrs": 17618, + "ĠMs": 16064, + "ĠMsg": 24205, + "ĠMsgBox": 51566, + "ĠMt": 35508, + "ĠMu": 14787, + "ĠMuch": 23309, + "ĠMud": 68340, + "ĠMueller": 31667, + "ĠMug": 63474, + "ĠMuhammad": 35731, + "ĠMuj": 57236, + "ĠMuk": 69219, + "ĠMul": 29351, + "ĠMull": 51527, + "ĠMult": 22162, + "ĠMulti": 17439, + "ĠMultimedia": 82222, + "ĠMultip": 58712, + "ĠMultiple": 28811, + "ĠMultiply": 71059, + "ĠMum": 66837, + "ĠMumbai": 34712, + "ĠMun": 27882, + "ĠMund": 66561, + "ĠMundo": 89823, + "ĠMunich": 46204, + "ĠMunicip": 35703, + "ĠMunicipal": 44240, + "ĠMur": 15006, + "ĠMurder": 51383, + "ĠMurdoch": 78609, + "ĠMurphy": 29953, + "ĠMurray": 27154, + "ĠMus": 5331, + "ĠMuscle": 67327, + "ĠMuse": 50008, + "ĠMuseum": 16328, + "ĠMush": 58105, + "ĠMushroom": 90197, + "ĠMusic": 10717, + "ĠMusical": 56257, + "ĠMusik": 75410, + "ĠMusk": 39538, + "ĠMuslim": 10240, + "ĠMuslims": 19503, + "ĠMuss": 89714, + "ĠMust": 15465, + "ĠMustang": 52779, + "ĠMut": 31228, + "ĠMutable": 31143, + "ĠMutableList": 78024, + "ĠMutableLiveData": 55281, + "ĠMutation": 67203, + "ĠMutex": 66215, + "ĠMutual": 74042, + "ĠMvc": 97585, + "ĠMy": 3017, + "ĠMyApp": 54016, + "ĠMyBase": 60133, + "ĠMyClass": 83826, + "ĠMySQL": 26339, + "ĠMySql": 25725, + "ĠMySqlCommand": 58307, + "ĠMySqlConnection": 81987, + "ĠMyanmar": 52355, + "ĠMyers": 50223, + "ĠMyst": 29627, + "ĠMystery": 48005, + "ĠMystic": 88525, + "ĠMyth": 46495, + "ĠMá": 127428, + "ĠMás": 83875, + "ĠMáy": 133151, + "ĠMã": 137017, + "ĠMädchen": 78596, + "ĠMänner": 98376, + "ĠMär": 84132, + "ĠMärz": 135777, + "ĠMé": 42931, + "ĠMéd": 97390, + "ĠMét": 61258, + "ĠMétodo": 80376, + "ĠMéxico": 62171, + "ĠMö": 48682, + "ĠMöglich": 66461, + "ĠMöglichkeit": 90817, + "ĠMü": 63630, + "ĠMüd": 139913, + "ĠMüdü": 137463, + "ĠMüdürl": 137464, + "ĠMüdürlÃ¼ÄŁÃ¼": 137465, + "ĠMüdürü": 139914, + "ĠMüller": 98918, + "ĠMünchen": 89266, + "ĠMüsl": 141413, + "ĠMüslü": 141414, + "ĠMüslüman": 141415, + "ĠMÃł": 137701, + "ĠMặc": 139333, + "ĠMặt": 140366, + "ĠMẹ": 137864, + "ĠMỹ": 128819, + "ĠMá»įi": 141206, + "ĠMá»Ĺi": 138824, + "ĠMá»Ļt": 129005, + "ĠN": 451, + "ĠNA": 20662, + "ĠNAFTA": 84719, + "ĠNAME": 19122, + "ĠNAMES": 90910, + "ĠNAN": 92129, + "ĠNAND": 76457, + "ĠNAS": 16086, + "ĠNASA": 21431, + "ĠNASCAR": 64590, + "ĠNAT": 18248, + "ĠNATIONAL": 93681, + "ĠNATO": 29234, + "ĠNAV": 53303, + "ĠNB": 34979, + "ĠNBA": 17390, + "ĠNBC": 23521, + "ĠNC": 20059, + "ĠNCAA": 33160, + "ĠND": 38444, + "ĠNDEBUG": 98529, + "ĠNDP": 53650, + "ĠNE": 7856, + "ĠNEC": 82268, + "ĠNEED": 48133, + "ĠNEG": 84065, + "ĠNEGLIGENCE": 28548, + "ĠNEO": 86583, + "ĠNES": 73213, + "ĠNET": 24977, + "ĠNETWORK": 57436, + "ĠNEVER": 55025, + "ĠNEW": 16165, + "ĠNEWS": 40126, + "ĠNEXT": 45011, + "ĠNF": 44727, + "ĠNFC": 47205, + "ĠNFL": 12588, + "ĠNFS": 87815, + "ĠNG": 20018, + "ĠNGO": 46786, + "ĠNGOs": 71514, + "ĠNGX": 89105, + "ĠNH": 34831, + "ĠNHL": 23677, + "ĠNHS": 36281, + "ĠNI": 41501, + "ĠNIC": 64215, + "ĠNIGHT": 75031, + "ĠNIH": 83270, + "ĠNIL": 64147, + "ĠNIR": 98754, + "ĠNJ": 26718, + "ĠNK": 70063, + "ĠNL": 32160, + "ĠNM": 39714, + "ĠNN": 44678, + "ĠNO": 5664, + "ĠNOAA": 85648, + "ĠNODE": 34817, + "ĠNOI": 24767, + "ĠNOM": 97711, + "ĠNON": 20575, + "ĠNONE": 42869, + "ĠNONINFRINGEMENT": 27986, + "ĠNOP": 86376, + "ĠNOR": 69088, + "ĠNORMAL": 52431, + "ĠNORTH": 74759, + "ĠNOT": 4183, + "ĠNOTE": 16743, + "ĠNOTES": 69946, + "ĠNOTHING": 86081, + "ĠNOTICE": 29229, + "ĠNOTIFY": 83102, + "ĠNOW": 22407, + "ĠNP": 43240, + "ĠNPC": 29717, + "ĠNPCs": 73570, + "ĠNPR": 44200, + "ĠNR": 39295, + "ĠNRA": 52409, + "ĠNRF": 72010, + "ĠNRL": 94154, + "ĠNS": 3043, + "ĠNSA": 30783, + "ĠNSArray": 25504, + "ĠNSAttributedString": 93237, + "ĠNSCoder": 68962, + "ĠNSData": 49770, + "ĠNSDate": 45267, + "ĠNSDictionary": 28896, + "ĠNSError": 31238, + "ĠNSF": 79528, + "ĠNSIndexPath": 65168, + "ĠNSInteger": 37962, + "ĠNSK": 82142, + "ĠNSLayoutConstraint": 59582, + "ĠNSLocalizedString": 60272, + "ĠNSLog": 18322, + "ĠNSMutable": 21092, + "ĠNSMutableArray": 33268, + "ĠNSMutableDictionary": 51905, + "ĠNSNumber": 48473, + "ĠNSObject": 26209, + "ĠNSRange": 92096, + "ĠNSS": 56008, + "ĠNSString": 9335, + "ĠNSStringFromClass": 86361, + "ĠNST": 38542, + "ĠNSTextAlignment": 74234, + "ĠNSUInteger": 60132, + "ĠNSURL": 39863, + "ĠNSW": 37908, + "ĠNT": 17658, + "ĠNTN": 78815, + "ĠNU": 60725, + "ĠNUIT": 73976, + "ĠNUITKA": 74420, + "ĠNULL": 1770, + "ĠNUM": 15943, + "ĠNUMBER": 36836, + "ĠNUnit": 52450, + "ĠNV": 24458, + "ĠNVIC": 85932, + "ĠNVIDIA": 33561, + "ĠNW": 41074, + "ĠNX": 57185, + "ĠNXT": 90401, + "ĠNY": 12271, + "ĠNYC": 33285, + "ĠNYPD": 73155, + "ĠNYT": 98712, + "ĠNZ": 41757, + "ĠNa": 12812, + "ĠNaN": 32178, + "ĠNab": 87306, + "ĠNach": 28275, + "ĠNachricht": 80072, + "ĠNacht": 98359, + "ĠNacional": 44589, + "ĠNack": 95062, + "ĠNad": 34662, + "ĠNadu": 74805, + "ĠNag": 29062, + "ĠNagar": 87194, + "ĠNah": 71988, + "ĠNail": 74428, + "ĠNairobi": 96525, + "ĠNaj": 66353, + "ĠNak": 43229, + "ĠNaked": 53738, + "ĠNam": 29974, + "ĠName": 3988, + "ĠNamed": 40459, + "ĠNamen": 84100, + "ĠNames": 34875, + "ĠNamespace": 41962, + "ĠNaming": 96475, + "ĠNan": 32142, + "ĠNancy": 34236, + "ĠNano": 62951, + "ĠNaomi": 77722, + "ĠNap": 28476, + "ĠNaples": 82621, + "ĠNapoleon": 69427, + "ĠNapoli": 85676, + "ĠNar": 28753, + "ĠNarc": 78320, + "ĠNarendra": 66332, + "ĠNarr": 56869, + "ĠNarrative": 95721, + "ĠNarrow": 82311, + "ĠNaruto": 48994, + "ĠNas": 38222, + "ĠNash": 27412, + "ĠNashville": 36540, + "ĠNass": 71902, + "ĠNasıl": 137038, + "ĠNat": 23833, + "ĠNatal": 41601, + "ĠNatalie": 63020, + "ĠNatasha": 86666, + "ĠNate": 54669, + "ĠNath": 60809, + "ĠNathan": 36737, + "ĠNation": 17227, + "ĠNational": 5055, + "ĠNationals": 61091, + "ĠNations": 19140, + "ĠNationwide": 89654, + "ĠNative": 16703, + "ĠNatur": 39449, + "ĠNatural": 18448, + "ĠNaturally": 59742, + "ĠNature": 21331, + "ĠNatürlich": 142839, + "ĠNaughty": 86741, + "ĠNav": 11966, + "ĠNavBar": 78666, + "ĠNavController": 55762, + "ĠNavLink": 94216, + "ĠNavParams": 75900, + "ĠNaval": 45729, + "ĠNavbar": 38309, + "ĠNavigate": 81739, + "ĠNavigation": 17980, + "ĠNavigator": 26542, + "ĠNavy": 19036, + "ĠNaw": 89710, + "ĠNay": 98073, + "ĠNaz": 28774, + "ĠNazi": 31427, + "ĠNazis": 48433, + "ĠNb": 77483, + "ĠNd": 87368, + "ĠNe": 4182, + "ĠNeal": 65449, + "ĠNear": 30394, + "ĠNearby": 95548, + "ĠNearly": 48569, + "ĠNeb": 51709, + "ĠNebraska": 37279, + "ĠNec": 63124, + "ĠNecessary": 94533, + "ĠNeck": 42955, + "ĠNecklace": 85360, + "ĠNed": 69999, + "ĠNederland": 65392, + "ĠNeed": 14656, + "ĠNeeded": 56561, + "ĠNeedle": 88800, + "ĠNeedless": 82283, + "ĠNeeds": 43140, + "ĠNeg": 23995, + "ĠNegative": 50857, + "ĠNegot": 74211, + "ĠNegro": 56047, + "ĠNeh": 72436, + "ĠNeighbor": 97163, + "ĠNeighborhood": 61512, + "ĠNeil": 33121, + "ĠNeither": 24233, + "ĠNel": 88361, + "ĠNelson": 26464, + "ĠNem": 62168, + "ĠNeo": 40644, + "ĠNeon": 76964, + "ĠNep": 36569, + "ĠNepal": 48964, + "ĠNeptune": 79624, + "ĠNer": 65804, + "ĠNero": 84078, + "ĠNes": 91408, + "ĠNess": 87502, + "ĠNest": 38905, + "ĠNested": 71742, + "ĠNet": 9374, + "ĠNetanyahu": 44642, + "ĠNetflix": 22642, + "ĠNetherlands": 25662, + "ĠNets": 58387, + "ĠNetwork": 8141, + "ĠNetworking": 59463, + "ĠNetworks": 38710, + "ĠNetz": 92283, + "ĠNeu": 44850, + "ĠNeue": 75563, + "ĠNeural": 60477, + "ĠNeuro": 31259, + "ĠNeuroscience": 84779, + "ĠNeutral": 58694, + "ĠNev": 24324, + "ĠNevada": 26868, + "ĠNever": 14695, + "ĠNevertheless": 33953, + "ĠNeville": 81235, + "ĠNew": 1532, + "ĠNewark": 64499, + "ĠNewcastle": 43126, + "ĠNewest": 78697, + "ĠNewfoundland": 88296, + "ĠNewly": 98835, + "ĠNewman": 55621, + "ĠNewport": 58800, + "ĠNews": 5398, + "ĠNewsletter": 38593, + "ĠNewsp": 60003, + "ĠNewspaper": 86739, + "ĠNewton": 20668, + "ĠNewtonsoft": 31579, + "ĠNewtown": 91829, + "ĠNex": 87829, + "ĠNext": 9295, + "ĠNexus": 40022, + "ĠNey": 95048, + "ĠNg": 16619, + "ĠNgModule": 27823, + "ĠNgb": 66289, + "ĠNgh": 128018, + "ĠNghá»ĩ": 138770, + "ĠNghá»ĭ": 141127, + "ĠNgo": 126949, + "ĠNgoÃłi": 130676, + "ĠNgoại": 142769, + "ĠNguy": 128983, + "ĠNguyen": 63161, + "ĠNguyên": 132019, + "ĠNguyá»ħn": 128984, + "ĠNgân": 138737, + "ĠNgô": 137192, + "ĠNgÃłi": 142562, + "ĠNgÃły": 131160, + "ĠNgưá»Ŀi": 129639, + "ĠNgá»įc": 132896, + "ĠNh": 90453, + "ĠNhi": 135996, + "ĠNhiá»ģu": 135997, + "ĠNhân": 133404, + "ĠNhÃł": 129645, + "ĠNhư": 133958, + "ĠNhưng": 129840, + "ĠNhất": 140285, + "ĠNháºŃn": 142435, + "ĠNháºŃt": 129898, + "ĠNhững": 128965, + "ĠNi": 21890, + "ĠNiagara": 77877, + "ĠNic": 17552, + "ĠNicar": 66038, + "ĠNicaragua": 76324, + "ĠNice": 28859, + "ĠNich": 25671, + "ĠNicholas": 39696, + "ĠNichols": 80150, + "ĠNicholson": 88883, + "ĠNicht": 86417, + "ĠNick": 14991, + "ĠNickel": 66842, + "ĠNico": 74035, + "ĠNicol": 82340, + "ĠNicola": 78441, + "ĠNicolas": 51736, + "ĠNicole": 44030, + "ĠNie": 52283, + "ĠNielsen": 63451, + "ĠNiet": 59298, + "ĠNietzsche": 93106, + "ĠNigel": 65816, + "ĠNiger": 20369, + "ĠNigeria": 28158, + "ĠNigerian": 54333, + "ĠNight": 12826, + "ĠNightmare": 74125, + "ĠNights": 63543, + "ĠNik": 21573, + "ĠNike": 33144, + "ĠNikki": 69107, + "ĠNikol": 95557, + "ĠNikola": 96315, + "ĠNikon": 62152, + "ĠNil": 32274, + "ĠNile": 76190, + "ĠNim": 46583, + "ĠNimbus": 71493, + "ĠNin": 26727, + "ĠNina": 63888, + "ĠNine": 37066, + "ĠNing": 85083, + "ĠNinja": 46985, + "ĠNintendo": 22799, + "ĠNinth": 85758, + "ĠNir": 77743, + "ĠNissan": 41881, + "ĠNit": 49516, + "ĠNixon": 41626, + "ĠNo": 2308, + "ĠNoSuch": 47977, + "ĠNoSuchElementException": 94449, + "ĠNoah": 41984, + "ĠNob": 19017, + "ĠNobel": 46978, + "ĠNoble": 48587, + "ĠNobody": 36458, + "ĠNode": 6018, + "ĠNodeList": 98711, + "ĠNodeType": 82749, + "ĠNodes": 52501, + "ĠNodo": 94129, + "ĠNoel": 69561, + "ĠNoir": 72714, + "ĠNoise": 50523, + "ĠNokia": 35706, + "ĠNolan": 61632, + "ĠNom": 36900, + "ĠNombre": 47695, + "ĠNome": 58275, + "ĠNon": 11581, + "ĠNone": 2240, + "ĠNonetheless": 55633, + "ĠNoon": 82856, + "ĠNope": 85719, + "ĠNor": 8011, + "ĠNora": 72545, + "ĠNord": 29181, + "ĠNordic": 63366, + "ĠNorfolk": 58692, + "ĠNorm": 20317, + "ĠNormal": 18437, + "ĠNormalize": 68092, + "ĠNormally": 51683, + "ĠNorman": 40068, + "ĠNorris": 70765, + "ĠNorse": 93514, + "ĠNorte": 82293, + "ĠNorth": 4787, + "ĠNortheast": 44580, + "ĠNorthern": 16926, + "ĠNorthwest": 39405, + "ĠNorthwestern": 66263, + "ĠNorton": 59264, + "ĠNorway": 31503, + "ĠNorwegian": 44621, + "ĠNorwich": 83123, + "ĠNos": 49997, + "ĠNose": 92123, + "ĠNot": 2806, + "ĠNotFound": 23420, + "ĠNotFoundException": 66646, + "ĠNotImplemented": 21548, + "ĠNotImplementedError": 38199, + "ĠNotImplementedException": 33146, + "ĠNotSupportedException": 92990, + "ĠNote": 7036, + "ĠNotebook": 68655, + "ĠNotes": 18068, + "ĠNothing": 12064, + "ĠNotice": 24742, + "ĠNoticed": 70134, + "ĠNotices": 95253, + "ĠNotification": 16571, + "ĠNotificationCenter": 83335, + "ĠNotifications": 52938, + "ĠNotify": 34970, + "ĠNotre": 43464, + "ĠNottingham": 66559, + "ĠNotÃŃc": 140352, + "ĠNotÃŃcias": 140353, + "ĠNou": 59187, + "ĠNous": 47009, + "ĠNov": 4620, + "ĠNova": 34641, + "ĠNovel": 52876, + "ĠNovember": 6702, + "ĠNow": 4695, + "ĠNowadays": 86481, + "ĠNpgsql": 78879, + "ĠNr": 69093, + "ĠNu": 32324, + "ĠNuclear": 37444, + "ĠNude": 54668, + "ĠNuevo": 87758, + "ĠNug": 58245, + "ĠNuggets": 84477, + "ĠNull": 18084, + "ĠNullPointerException": 56850, + "ĠNullable": 52655, + "ĠNum": 16212, + "ĠNumber": 5624, + "ĠNumberOf": 98349, + "ĠNumbers": 34713, + "ĠNumer": 47124, + "ĠNumeric": 50367, + "ĠNumero": 90985, + "ĠNumerous": 85815, + "ĠNun": 63678, + "ĠNunes": 85112, + "ĠNur": 56231, + "ĠNurs": 32341, + "ĠNurse": 48776, + "ĠNursery": 92185, + "ĠNurses": 95236, + "ĠNursing": 45253, + "ĠNut": 18372, + "ĠNutrition": 38600, + "ĠNutzung": 84605, + "ĠNvidia": 61367, + "ĠNx": 81047, + "ĠNy": 36732, + "ĠNylon": 90662, + "ĠNz": 98692, + "ĠNº": 140124, + "ĠNão": 57649, + "ĠNä": 84974, + "ĠNähe": 137900, + "ĠNó": 131031, + "ĠNÄĥm": 133670, + "ĠNẵ": 135240, + "ĠNẵng": 135241, + "ĠNếu": 129024, + "ĠNữ": 140870, + "ĠNá»Ļi": 128685, + "ĠO": 506, + "ĠOA": 80442, + "ĠOAuth": 38316, + "ĠOB": 43173, + "ĠOBJ": 42583, + "ĠOBJECT": 39786, + "ĠOBS": 75860, + "ĠOC": 31867, + "ĠOCC": 88013, + "ĠOCD": 97924, + "ĠOCI": 92087, + "ĠOCR": 80577, + "ĠOCT": 66177, + "ĠOD": 52732, + "ĠOE": 55530, + "ĠOECD": 78417, + "ĠOEM": 40762, + "ĠOF": 3008, + "ĠOFF": 17611, + "ĠOFFSET": 48104, + "ĠOG": 66359, + "ĠOH": 28722, + "ĠOID": 76358, + "ĠOK": 10402, + "ĠOL": 55270, + "ĠOLD": 64891, + "ĠOLED": 70952, + "ĠOM": 47337, + "ĠOMAP": 94536, + "ĠOMG": 97545, + "ĠOMIT": 69400, + "ĠOMX": 59830, + "ĠON": 6197, + "ĠONE": 24038, + "ĠONLINE": 52274, + "ĠONLY": 26687, + "ĠOO": 86818, + "ĠOP": 13134, + "ĠOPC": 86371, + "ĠOPEN": 29841, + "ĠOPER": 41376, + "ĠOPERATION": 98684, + "ĠOPP": 94181, + "ĠOPS": 82855, + "ĠOPT": 40034, + "ĠOPTION": 29895, + "ĠOPTIONAL": 76232, + "ĠOPTIONS": 51280, + "ĠOR": 2726, + "ĠORD": 69269, + "ĠORDER": 15520, + "ĠORIGINAL": 86782, + "ĠORM": 53590, + "ĠOS": 10085, + "ĠOSC": 74509, + "ĠOSD": 93522, + "ĠOSError": 74242, + "ĠOSI": 88078, + "ĠOSP": 96539, + "ĠOSS": 83233, + "ĠOST": 93301, + "ĠOSX": 87746, + "ĠOT": 8605, + "ĠOTA": 96747, + "ĠOTHER": 10065, + "ĠOTHERWISE": 18071, + "ĠOTP": 80868, + "ĠOU": 62835, + "ĠOUR": 44913, + "ĠOUT": 9808, + "ĠOUTER": 89675, + "ĠOUTPUT": 30990, + "ĠOV": 74302, + "ĠOVER": 35464, + "ĠOVERRIDE": 87301, + "ĠOW": 28869, + "ĠOWN": 68469, + "ĠOWNER": 50178, + "ĠOak": 18283, + "ĠOakland": 29681, + "ĠOaks": 74256, + "ĠOasis": 87291, + "ĠOb": 4657, + "ĠObama": 7105, + "ĠObamacare": 35996, + "ĠOber": 51145, + "ĠObesity": 94071, + "ĠObj": 26639, + "ĠObject": 3002, + "ĠObjectId": 50090, + "ĠObjectMapper": 54350, + "ĠObjectOutputStream": 97786, + "ĠObjectType": 78886, + "ĠObjective": 54289, + "ĠObjects": 24114, + "ĠObl": 55775, + "ĠObs": 50441, + "ĠObserv": 30843, + "ĠObservable": 11231, + "ĠObservableCollection": 44464, + "ĠObservation": 86429, + "ĠObservatory": 57874, + "ĠObserver": 34041, + "ĠObst": 94844, + "ĠObt": 56647, + "ĠObtain": 63424, + "ĠObviously": 35430, + "ĠOc": 82452, + "ĠOcc": 19927, + "ĠOccasionally": 88326, + "ĠOccup": 32412, + "ĠOccupation": 98352, + "ĠOccupational": 84155, + "ĠOccupy": 73935, + "ĠOcean": 21575, + "ĠOct": 4915, + "ĠOctober": 6527, + "ĠOculus": 55384, + "ĠOczy": 141295, + "ĠOczywiÅĽcie": 141296, + "ĠOd": 24560, + "ĠOdd": 55036, + "ĠOdds": 83357, + "ĠOdin": 91073, + "ĠOdyssey": 65781, + "ĠOf": 4940, + "ĠOff": 4114, + "ĠOffensive": 75489, + "ĠOffer": 24498, + "ĠOffering": 66018, + "ĠOffers": 51318, + "ĠOffice": 8246, + "ĠOfficer": 19577, + "ĠOfficers": 48847, + "ĠOffices": 86538, + "ĠOfficial": 22651, + "ĠOfficials": 63139, + "ĠOffline": 66370, + "ĠOffset": 16861, + "ĠOften": 34916, + "ĠOg": 40837, + "ĠOgre": 63822, + "ĠOh": 8670, + "ĠOhio": 14352, + "ĠOil": 15526, + "ĠOilers": 89227, + "ĠOk": 7622, + "ĠOkHttpClient": 78920, + "ĠOkay": 35439, + "ĠOklahoma": 22797, + "ĠOktober": 73208, + "ĠOl": 11959, + "ĠOld": 10621, + "ĠOlder": 53925, + "ĠOle": 32697, + "ĠOleDb": 56845, + "ĠOliv": 64284, + "ĠOlive": 46714, + "ĠOliveira": 95314, + "ĠOliver": 30246, + "ĠOlivia": 59287, + "ĠOlivier": 77018, + "ĠOlsen": 81781, + "ĠOlson": 74255, + "ĠOlymp": 14792, + "ĠOlympia": 95729, + "ĠOlympic": 24900, + "ĠOlympics": 31963, + "ĠOlympus": 96791, + "ĠOm": 18603, + "ĠOmaha": 67205, + "ĠOman": 81125, + "ĠOmar": 53558, + "ĠOmega": 43309, + "ĠOmn": 89834, + "ĠOmni": 85225, + "ĠOn": 1913, + "ĠOnClickListener": 67307, + "ĠOnCollision": 92511, + "ĠOnDestroy": 45889, + "ĠOnInit": 10972, + "ĠOnPropertyChanged": 49281, + "ĠOnTrigger": 50576, + "ĠOnTriggerEnter": 64104, + "ĠOnc": 76754, + "ĠOnce": 9646, + "ĠOne": 3776, + "ĠOnePlus": 80965, + "ĠOnes": 73833, + "ĠOnion": 85935, + "ĠOnline": 8105, + "ĠOnly": 8278, + "ĠOnt": 17819, + "ĠOntario": 20551, + "ĠOok": 94555, + "ĠOops": 95252, + "ĠOp": 10672, + "ĠOpCode": 97779, + "ĠOpcode": 78899, + "ĠOpen": 5264, + "ĠOpenFileDialog": 94086, + "ĠOpenGL": 36703, + "ĠOpenSSL": 65617, + "ĠOpening": 40037, + "ĠOpens": 66179, + "ĠOper": 7461, + "ĠOpera": 38579, + "ĠOperand": 61085, + "ĠOperating": 35133, + "ĠOperation": 16730, + "ĠOperational": 89296, + "ĠOperations": 24883, + "ĠOperator": 28498, + "ĠOperators": 64420, + "ĠOpinion": 59667, + "ĠOpp": 28610, + "ĠOpport": 34278, + "ĠOpportunities": 65785, + "ĠOpportunity": 47279, + "ĠOpposition": 65559, + "ĠOpr": 77518, + "ĠOprah": 91158, + "ĠOps": 55158, + "ĠOpt": 16554, + "ĠOptical": 74839, + "ĠOptim": 30097, + "ĠOptimization": 57739, + "ĠOption": 6959, + "ĠOptional": 12256, + "ĠOptionally": 96195, + "ĠOptions": 14566, + "ĠOr": 2521, + "ĠOracle": 21422, + "ĠOral": 53401, + "ĠOrange": 21958, + "ĠOrb": 42988, + "ĠOrbit": 84457, + "ĠOrc": 56586, + "ĠOrch": 54305, + "ĠOrchard": 83152, + "ĠOrchestra": 53517, + "ĠOrd": 30037, + "ĠOrden": 96845, + "ĠOrder": 7217, + "ĠOrdered": 39581, + "ĠOrderedDict": 53213, + "ĠOrdering": 68621, + "ĠOrders": 31836, + "ĠOrdinary": 97911, + "ĠOre": 33443, + "ĠOregon": 18784, + "ĠOrg": 33706, + "ĠOrgan": 10762, + "ĠOrganic": 42937, + "ĠOrganisation": 46743, + "ĠOrganization": 20395, + "ĠOrganizations": 66692, + "ĠOrganizer": 85837, + "ĠOri": 37898, + "ĠOrient": 61474, + "ĠOriental": 70951, + "ĠOrientation": 54134, + "ĠOrig": 45498, + "ĠOrigin": 17116, + "ĠOriginal": 17230, + "ĠOriginally": 24805, + "ĠOrigins": 69491, + "ĠOrioles": 79260, + "ĠOrion": 68673, + "ĠOrlando": 27844, + "ĠOrleans": 25919, + "ĠOrn": 87839, + "ĠOro": 59729, + "ĠOrr": 86019, + "ĠOrt": 44340, + "ĠOrth": 31110, + "ĠOrthodox": 45173, + "ĠOrtiz": 78550, + "ĠOrwell": 82753, + "ĠOs": 15433, + "ĠOsaka": 86985, + "ĠOsama": 82242, + "ĠOsborne": 70129, + "ĠOsc": 55636, + "ĠOscar": 30697, + "ĠOscars": 88796, + "ĠOslo": 57858, + "ĠOsman": 140411, + "ĠOsmanlı": 140412, + "ĠOst": 65435, + "ĠOsw": 65221, + "ĠOswald": 89350, + "ĠOt": 30576, + "ĠOther": 6944, + "ĠOthers": 25028, + "ĠOtherwise": 18214, + "ĠOtt": 23930, + "ĠOttawa": 32166, + "ĠOtto": 65891, + "ĠOttoman": 69010, + "ĠOu": 68227, + "ĠOur": 5633, + "ĠOut": 4371, + "ĠOutcome": 94609, + "ĠOutdoor": 32682, + "ĠOuter": 55197, + "ĠOutlet": 75649, + "ĠOutline": 51909, + "ĠOutlineInputBorder": 82537, + "ĠOutlook": 41058, + "ĠOutput": 9258, + "ĠOutputStream": 52058, + "ĠOutputs": 58414, + "ĠOutreach": 96450, + "ĠOuts": 85400, + "ĠOutside": 41151, + "ĠOutstanding": 75341, + "ĠOval": 55689, + "ĠOven": 86113, + "ĠOver": 6065, + "ĠOverall": 27893, + "ĠOverflow": 70447, + "ĠOverlay": 64867, + "ĠOvernight": 92468, + "ĠOverride": 38258, + "ĠOverrides": 46083, + "ĠOvers": 61561, + "ĠOverse": 85303, + "ĠOversight": 89851, + "ĠOverview": 34807, + "ĠOverwatch": 72338, + "ĠOw": 40796, + "ĠOwen": 46709, + "ĠOwens": 63357, + "ĠOwl": 73695, + "ĠOwn": 20547, + "ĠOwned": 85093, + "ĠOwner": 25833, + "ĠOwners": 39466, + "ĠOwnership": 84936, + "ĠOx": 50615, + "ĠOxford": 25210, + "ĠOxygen": 86569, + "ĠOz": 35604, + "ĠP": 393, + "ĠPA": 12878, + "ĠPAC": 39864, + "ĠPACK": 33332, + "ĠPACKAGE": 66258, + "ĠPACKET": 73205, + "ĠPAD": 61754, + "ĠPAGE": 25040, + "ĠPAL": 51469, + "ĠPAN": 72281, + "ĠPANEL": 80024, + "ĠPAR": 26081, + "ĠPARA": 50400, + "ĠPARAM": 28608, + "ĠPARAMETERS": 68079, + "ĠPARK": 66284, + "ĠPART": 6766, + "ĠPARTIC": 7550, + "ĠPARTICULAR": 7652, + "ĠPARTY": 63227, + "ĠPAS": 59917, + "ĠPASS": 50835, + "ĠPASSWORD": 50671, + "ĠPAT": 44370, + "ĠPATCH": 59345, + "ĠPATH": 26773, + "ĠPAY": 49377, + "ĠPB": 30934, + "ĠPBS": 59045, + "ĠPC": 6673, + "ĠPCA": 60929, + "ĠPCB": 50077, + "ĠPCI": 27789, + "ĠPCIe": 90690, + "ĠPCM": 75007, + "ĠPCR": 66691, + "ĠPCS": 91097, + "ĠPCs": 46831, + "ĠPD": 26474, + "ĠPDF": 11358, + "ĠPDO": 24964, + "ĠPDT": 45457, + "ĠPE": 21804, + "ĠPED": 93753, + "ĠPEM": 96734, + "ĠPEN": 80670, + "ĠPEOPLE": 66725, + "ĠPER": 17854, + "ĠPERF": 78377, + "ĠPERFORMANCE": 80946, + "ĠPERMISSION": 96540, + "ĠPERSON": 56377, + "ĠPET": 49259, + "ĠPF": 28415, + "ĠPG": 30044, + "ĠPGA": 79785, + "ĠPH": 14659, + "ĠPHI": 84835, + "ĠPHONE": 91083, + "ĠPHOTO": 64777, + "ĠPHP": 13119, + "ĠPHPUnit": 62186, + "ĠPHY": 57241, + "ĠPHYS": 83987, + "ĠPI": 22578, + "ĠPIC": 60606, + "ĠPICK": 83537, + "ĠPID": 35771, + "ĠPIE": 79455, + "ĠPIL": 42075, + "ĠPIN": 27129, + "ĠPIO": 86521, + "ĠPIPE": 71674, + "ĠPIT": 92919, + "ĠPIX": 61601, + "ĠPIXEL": 95822, + "ĠPIXI": 93570, + "ĠPJ": 64980, + "ĠPK": 24826, + "ĠPKK": 97958, + "ĠPL": 10315, + "ĠPLA": 96966, + "ĠPLACE": 81882, + "ĠPLAN": 72566, + "ĠPLATFORM": 68670, + "ĠPLAY": 41988, + "ĠPLAYER": 44717, + "ĠPLC": 69352, + "ĠPLEASE": 53133, + "ĠPLL": 54345, + "ĠPLUGIN": 82507, + "ĠPLUS": 58453, + "ĠPM": 5851, + "ĠPMC": 87032, + "ĠPMID": 60941, + "ĠPN": 61901, + "ĠPNG": 41639, + "ĠPO": 12932, + "ĠPOD": 90501, + "ĠPOINT": 46595, + "ĠPOINTER": 78031, + "ĠPOL": 31640, + "ĠPOLIT": 83220, + "ĠPOLITICO": 98439, + "ĠPOLL": 93115, + "ĠPOP": 45728, + "ĠPOR": 57870, + "ĠPORT": 19075, + "ĠPOS": 26494, + "ĠPOSITION": 73723, + "ĠPOSIX": 72147, + "ĠPOSS": 27556, + "ĠPOSSIBILITY": 29382, + "ĠPOST": 12869, + "ĠPOT": 61502, + "ĠPOV": 74521, + "ĠPOW": 69793, + "ĠPOWER": 42637, + "ĠPP": 29850, + "ĠPPC": 69727, + "ĠPPP": 74038, + "ĠPQ": 61717, + "ĠPR": 8575, + "ĠPRE": 20952, + "ĠPREF": 98002, + "ĠPREFIX": 69978, + "ĠPRES": 78764, + "ĠPRESS": 73819, + "ĠPRI": 50167, + "ĠPRICE": 59052, + "ĠPRIMARY": 37467, + "ĠPRINT": 35893, + "ĠPRIV": 89345, + "ĠPRIVATE": 57793, + "ĠPRO": 5308, + "ĠPROC": 24363, + "ĠPROCESS": 51437, + "ĠPROCUREMENT": 30332, + "ĠPRODUCT": 38083, + "ĠPRODUCTS": 87290, + "ĠPROF": 25518, + "ĠPROFILE": 70950, + "ĠPROFITS": 28224, + "ĠPROGMEM": 89960, + "ĠPROGRAM": 45661, + "ĠPROJECT": 39892, + "ĠPROM": 67688, + "ĠPROP": 41977, + "ĠPROPERTY": 55640, + "ĠPROT": 90828, + "ĠPROVID": 15476, + "ĠPROVIDED": 16525, + "ĠPS": 11405, + "ĠPSA": 69486, + "ĠPSD": 73634, + "ĠPSG": 92677, + "ĠPSI": 88178, + "ĠPSP": 68814, + "ĠPST": 40984, + "ĠPSU": 87993, + "ĠPT": 19370, + "ĠPTR": 55196, + "ĠPTS": 96261, + "ĠPTSD": 61319, + "ĠPU": 52088, + "ĠPUB": 76744, + "ĠPUBLIC": 31416, + "ĠPUR": 7330, + "ĠPURE": 83488, + "ĠPURPOSE": 7515, + "ĠPUSH": 70509, + "ĠPUT": 34380, + "ĠPV": 37864, + "ĠPVC": 49866, + "ĠPVOID": 81576, + "ĠPW": 58293, + "ĠPWM": 36034, + "ĠPX": 55484, + "ĠPY": 42760, + "ĠPYTHON": 76678, + "ĠPa": 15681, + "ĠPablo": 52763, + "ĠPac": 12637, + "ĠPace": 84519, + "ĠPacers": 84537, + "ĠPacific": 16462, + "ĠPack": 13795, + "ĠPackage": 16906, + "ĠPackageManager": 56395, + "ĠPackages": 63384, + "ĠPackaging": 69788, + "ĠPackers": 43528, + "ĠPacket": 28889, + "ĠPacking": 97818, + "ĠPacks": 89532, + "ĠPact": 98373, + "ĠPad": 25299, + "ĠPadding": 23024, + "ĠPadres": 90645, + "ĠPag": 42355, + "ĠPage": 5755, + "ĠPageInfo": 96293, + "ĠPageSize": 94832, + "ĠPager": 87308, + "ĠPages": 21772, + "ĠPagination": 44898, + "ĠPaginator": 93064, + "ĠPaging": 96038, + "ĠPai": 87514, + "ĠPaid": 54082, + "ĠPaige": 86411, + "ĠPain": 34467, + "ĠPaint": 17203, + "ĠPainter": 96764, + "ĠPainting": 54043, + "ĠPair": 25995, + "ĠPais": 81787, + "ĠPak": 13358, + "ĠPakistan": 16663, + "ĠPakistani": 44452, + "ĠPal": 10925, + "ĠPalace": 30296, + "ĠPaladin": 96500, + "ĠPale": 12349, + "ĠPaleo": 87407, + "ĠPalest": 13930, + "ĠPalestin": 16331, + "ĠPalestine": 40934, + "ĠPalestinian": 21841, + "ĠPalestinians": 33134, + "ĠPalette": 66409, + "ĠPalin": 77321, + "ĠPall": 71360, + "ĠPalm": 32478, + "ĠPalmer": 41116, + "ĠPalo": 84151, + "ĠPam": 43413, + "ĠPamela": 82143, + "ĠPan": 10990, + "ĠPanama": 48036, + "ĠPanasonic": 79642, + "ĠPanc": 96214, + "ĠPand": 33506, + "ĠPanda": 72410, + "ĠPandora": 71981, + "ĠPane": 98542, + "ĠPanel": 18946, + "ĠPanels": 80492, + "ĠPang": 99136, + "ĠPanic": 83740, + "ĠPant": 53122, + "ĠPanther": 58512, + "ĠPanthers": 44167, + "ĠPants": 66453, + "ĠPaolo": 83369, + "ĠPap": 31643, + "ĠPapa": 64573, + "ĠPaper": 17862, + "ĠPaperback": 69132, + "ĠPapers": 44131, + "ĠPapua": 89081, + "ĠPar": 4270, + "ĠPara": 22989, + "ĠParad": 39272, + "ĠParade": 62554, + "ĠParadise": 47591, + "ĠParagraph": 49351, + "ĠParallel": 49272, + "ĠParam": 6991, + "ĠParameter": 15168, + "ĠParameterDirection": 65927, + "ĠParameters": 13522, + "ĠParamount": 81810, + "ĠParams": 34352, + "ĠParan": 97284, + "ĠParas": 93037, + "ĠParcel": 42994, + "ĠParcelable": 82992, + "ĠPare": 74185, + "ĠParen": 45686, + "ĠParent": 17022, + "ĠParenthood": 49165, + "ĠParents": 42926, + "ĠParis": 12095, + "ĠParish": 59247, + "ĠPark": 5540, + "ĠParker": 28206, + "ĠParking": 40182, + "ĠParkinson": 61045, + "ĠParks": 38172, + "ĠParkway": 67153, + "ĠParl": 94887, + "ĠParliament": 19723, + "ĠParliamentary": 86315, + "ĠParm": 55644, + "ĠParr": 80530, + "ĠPars": 52711, + "ĠParse": 14775, + "ĠParseException": 61849, + "ĠParser": 21102, + "ĠParses": 73430, + "ĠParsing": 74789, + "ĠParsons": 77244, + "ĠPart": 3660, + "ĠPartial": 24552, + "ĠPartialEq": 55039, + "ĠPartialView": 90081, + "ĠParticip": 43481, + "ĠParticipant": 46376, + "ĠParticipants": 51778, + "ĠParticipation": 63403, + "ĠParticle": 35854, + "ĠParticularly": 96385, + "ĠParties": 59905, + "ĠPartition": 54626, + "ĠPartner": 31313, + "ĠPartners": 22817, + "ĠPartnership": 46262, + "ĠParts": 26303, + "ĠParty": 8554, + "ĠPas": 24010, + "ĠPasadena": 82450, + "ĠPascal": 57359, + "ĠPaso": 66329, + "ĠPass": 9970, + "ĠPassage": 98475, + "ĠPassed": 49120, + "ĠPassenger": 67634, + "ĠPassing": 76292, + "ĠPassion": 63788, + "ĠPassive": 71589, + "ĠPassport": 66800, + "ĠPassword": 12362, + "ĠPast": 23640, + "ĠPasta": 97346, + "ĠPaste": 57225, + "ĠPastor": 58666, + "ĠPat": 7135, + "ĠPatch": 30412, + "ĠPatel": 62755, + "ĠPatent": 57990, + "ĠPath": 7933, + "ĠPathfinder": 84181, + "ĠPaths": 55606, + "ĠPatient": 28924, + "ĠPatients": 43330, + "ĠPatio": 63986, + "ĠPatreon": 65112, + "ĠPatri": 22999, + "ĠPatricia": 53934, + "ĠPatrick": 19626, + "ĠPatriot": 66931, + "ĠPatriots": 32517, + "ĠPatrol": 44398, + "ĠPatron": 89400, + "ĠPatt": 44303, + "ĠPattern": 18834, + "ĠPatterns": 62723, + "ĠPatterson": 57634, + "ĠPatton": 95122, + "ĠPatty": 88245, + "ĠPaul": 6898, + "ĠPaula": 61153, + "ĠPaulo": 36926, + "ĠPause": 49656, + "ĠPav": 42756, + "ĠPavel": 78269, + "ĠPavilion": 74479, + "ĠPaw": 59971, + "ĠPawn": 82871, + "ĠPax": 70321, + "ĠPay": 11471, + "ĠPayPal": 35331, + "ĠPayload": 52916, + "ĠPayment": 20453, + "ĠPayments": 57876, + "ĠPayne": 68495, + "ĠPaypal": 68021, + "ĠPays": 93608, + "ĠPaz": 87882, + "ĠPaÃŃs": 141754, + "ĠPaÅĦst": 142009, + "ĠPaÅĦstwo": 142010, + "ĠPb": 97354, + "ĠPc": 88833, + "ĠPdf": 28150, + "ĠPdfP": 76516, + "ĠPdfPCell": 90562, + "ĠPe": 5139, + "ĠPeace": 25803, + "ĠPeach": 63595, + "ĠPeak": 43162, + "ĠPeaks": 88270, + "ĠPeanut": 92944, + "ĠPear": 23041, + "ĠPearce": 93056, + "ĠPearl": 36243, + "ĠPearson": 58542, + "ĠPeb": 88905, + "ĠPed": 19318, + "ĠPediatric": 94836, + "ĠPediatrics": 98771, + "ĠPedido": 97686, + "ĠPedro": 42482, + "ĠPee": 95308, + "ĠPeek": 96587, + "ĠPeel": 88594, + "ĠPeer": 45147, + "ĠPeg": 49666, + "ĠPeggy": 85552, + "ĠPel": 23663, + "ĠPell": 87908, + "ĠPelosi": 58980, + "ĠPemb": 68283, + "ĠPen": 13294, + "ĠPenal": 90658, + "ĠPenalty": 95242, + "ĠPence": 45960, + "ĠPend": 62400, + "ĠPendant": 80401, + "ĠPending": 41840, + "ĠPendingIntent": 71420, + "ĠPeng": 51050, + "ĠPenguin": 70144, + "ĠPenguins": 74844, + "ĠPeninsula": 49614, + "ĠPenis": 53318, + "ĠPenn": 13505, + "ĠPennsylvania": 19771, + "ĠPenny": 52915, + "ĠPens": 63380, + "ĠPension": 80457, + "ĠPent": 22633, + "ĠPentagon": 34536, + "ĠPeople": 8853, + "ĠPeoples": 80298, + "ĠPep": 52986, + "ĠPepper": 51782, + "ĠPepsi": 83078, + "ĠPer": 3616, + "ĠPerc": 51626, + "ĠPercent": 55348, + "ĠPercentage": 63241, + "ĠPerception": 85850, + "ĠPercy": 74686, + "ĠPere": 72478, + "ĠPerez": 49365, + "ĠPerf": 71448, + "ĠPerfect": 23239, + "ĠPerform": 25001, + "ĠPerformance": 20651, + "ĠPerforming": 69648, + "ĠPerforms": 57716, + "ĠPerhaps": 18765, + "ĠPeriod": 25492, + "ĠPeripheral": 83000, + "ĠPerkins": 64911, + "ĠPerl": 44432, + "ĠPerm": 90721, + "ĠPermanent": 56216, + "ĠPermission": 18135, + "ĠPermissions": 53357, + "ĠPermit": 80639, + "ĠPero": 55432, + "ĠPerr": 85748, + "ĠPerry": 30321, + "ĠPers": 20465, + "ĠPersian": 49861, + "ĠPersist": 93343, + "ĠPersistence": 64236, + "ĠPersistent": 66544, + "ĠPerson": 7357, + "ĠPersona": 53890, + "ĠPersonal": 19207, + "ĠPersonality": 64815, + "ĠPersonally": 62547, + "ĠPersonen": 87040, + "ĠPersonnel": 69977, + "ĠPersons": 61930, + "ĠPerspective": 69188, + "ĠPerspectives": 93885, + "ĠPert": 81784, + "ĠPerth": 49444, + "ĠPeru": 47747, + "ĠPes": 56856, + "ĠPessoa": 79576, + "ĠPest": 88145, + "ĠPet": 11333, + "ĠPete": 36273, + "ĠPeter": 11044, + "ĠPeters": 31184, + "ĠPetersburg": 53948, + "ĠPeterson": 39791, + "ĠPetit": 91817, + "ĠPetite": 86983, + "ĠPetr": 95776, + "ĠPetra": 83937, + "ĠPetro": 61479, + "ĠPetroleum": 71926, + "ĠPets": 35886, + "ĠPetsc": 52951, + "ĠPett": 82496, + "ĠPetty": 94698, + "ĠPew": 56617, + "ĠPey": 72880, + "ĠPeyton": 85456, + "ĠPf": 42068, + "ĠPfizer": 96854, + "ĠPg": 97265, + "ĠPh": 2350, + "ĠPhD": 29561, + "ĠPhantom": 46197, + "ĠPhar": 87571, + "ĠPharm": 24584, + "ĠPharma": 78781, + "ĠPharmac": 70781, + "ĠPharmaceutical": 58273, + "ĠPharmaceuticals": 90671, + "ĠPharmacy": 47075, + "ĠPhase": 27573, + "ĠPhaser": 63668, + "ĠPhelps": 97515, + "ĠPhen": 68783, + "ĠPhi": 63813, + "ĠPhil": 7155, + "ĠPhiladelphia": 19335, + "ĠPhilip": 25179, + "ĠPhilipp": 19945, + "ĠPhilippe": 66854, + "ĠPhilippine": 56181, + "ĠPhilippines": 25296, + "ĠPhilips": 73503, + "ĠPhill": 24517, + "ĠPhillies": 82986, + "ĠPhillip": 58709, + "ĠPhillips": 36414, + "ĠPhilly": 65154, + "ĠPhilosoph": 37256, + "ĠPhilosophy": 45869, + "ĠPhoenix": 22672, + "ĠPhon": 70324, + "ĠPhone": 14307, + "ĠPhoneNumber": 73674, + "ĠPhones": 84575, + "ĠPhonetic": 95734, + "ĠPhot": 17762, + "ĠPhoto": 10828, + "ĠPhotograph": 32155, + "ĠPhotographer": 74739, + "ĠPhotography": 40047, + "ĠPhoton": 62138, + "ĠPhotos": 16505, + "ĠPhotoshop": 44879, + "ĠPhp": 31134, + "ĠPhpStorm": 43881, + "ĠPhrase": 68000, + "ĠPhy": 92582, + "ĠPhys": 12809, + "ĠPhysical": 27379, + "ĠPhysician": 87198, + "ĠPhysicians": 87499, + "ĠPhysics": 27315, + "ĠPhysiology": 94846, + "ĠPháp": 132034, + "ĠPhát": 135322, + "ĠPhòng": 132850, + "ĠPhó": 133761, + "ĠPhú": 130751, + "ĠPhương": 132352, + "ĠPhạm": 136290, + "ĠPhần": 137572, + "ĠPháºŃt": 133319, + "ĠPhụ": 137265, + "ĠPi": 20636, + "ĠPiano": 55391, + "ĠPic": 25898, + "ĠPicasso": 66183, + "ĠPick": 19725, + "ĠPicker": 85383, + "ĠPicks": 65190, + "ĠPickup": 74460, + "ĠPics": 59835, + "ĠPicture": 24568, + "ĠPictureBox": 79649, + "ĠPictures": 28385, + "ĠPie": 39997, + "ĠPiece": 28930, + "ĠPieces": 67540, + "ĠPied": 83330, + "ĠPier": 20718, + "ĠPierce": 49830, + "ĠPierre": 36977, + "ĠPiet": 85188, + "ĠPig": 48554, + "ĠPik": 61162, + "ĠPikachu": 96919, + "ĠPike": 68729, + "ĠPil": 36351, + "ĠPill": 34136, + "ĠPillow": 80735, + "ĠPills": 47865, + "ĠPilot": 44499, + "ĠPin": 17471, + "ĠPine": 41509, + "ĠPing": 48657, + "ĠPink": 26119, + "ĠPins": 69250, + "ĠPinterest": 25051, + "ĠPioneer": 65350, + "ĠPip": 77382, + "ĠPipe": 27805, + "ĠPipeline": 40907, + "ĠPiper": 69807, + "ĠPipes": 78723, + "ĠPir": 38493, + "ĠPirate": 61148, + "ĠPirates": 52405, + "ĠPis": 90942, + "ĠPist": 44496, + "ĠPistol": 83814, + "ĠPistons": 94708, + "ĠPit": 38979, + "ĠPitch": 45576, + "ĠPitt": 21139, + "ĠPittsburgh": 27527, + "ĠPivot": 97893, + "ĠPix": 49466, + "ĠPixar": 99123, + "ĠPixel": 27469, + "ĠPixels": 85446, + "ĠPizza": 34818, + "ĠPl": 1818, + "ĠPlace": 10771, + "ĠPlaceholder": 86721, + "ĠPlacement": 77416, + "ĠPlaces": 44736, + "ĠPlain": 43199, + "ĠPlains": 62811, + "ĠPlaint": 56311, + "ĠPlaintiff": 77282, + "ĠPlan": 9680, + "ĠPlane": 43044, + "ĠPlanet": 28835, + "ĠPlanned": 46140, + "ĠPlanner": 74765, + "ĠPlanning": 27680, + "ĠPlans": 34595, + "ĠPlant": 17838, + "ĠPlants": 49198, + "ĠPlasma": 64532, + "ĠPlastic": 36008, + "ĠPlat": 86642, + "ĠPlate": 30303, + "ĠPlates": 79293, + "ĠPlatform": 17458, + "ĠPlatforms": 95109, + "ĠPlatinum": 43992, + "ĠPlato": 68061, + "ĠPlatt": 94678, + "ĠPlatz": 72055, + "ĠPlay": 7054, + "ĠPlayStation": 31265, + "ĠPlayback": 95301, + "ĠPlayboy": 89979, + "ĠPlayed": 74006, + "ĠPlayer": 7312, + "ĠPlayerPrefs": 59283, + "ĠPlayers": 24618, + "ĠPlayground": 88592, + "ĠPlaying": 39451, + "ĠPlaylist": 54269, + "ĠPlayoff": 94113, + "ĠPlays": 76165, + "ĠPlaystation": 92767, + "ĠPlaza": 40307, + "ĠPle": 32440, + "ĠPleasant": 63812, + "ĠPlease": 5209, + "ĠPlenty": 68389, + "ĠPlex": 95732, + "ĠPlot": 26033, + "ĠPlug": 52809, + "ĠPlugin": 21245, + "ĠPlugins": 62498, + "ĠPlum": 83309, + "ĠPlumbing": 92600, + "ĠPlus": 12343, + "ĠPluto": 77581, + "ĠPly": 53170, + "ĠPlymouth": 70998, + "ĠPo": 13808, + "ĠPocket": 45724, + "ĠPod": 16821, + "ĠPodcast": 29769, + "ĠPodesta": 87128, + "ĠPods": 52232, + "ĠPodsDummy": 70290, + "ĠPoe": 81595, + "ĠPoetry": 67687, + "ĠPoint": 5126, + "ĠPointF": 71174, + "ĠPointer": 21635, + "ĠPoints": 20725, + "ĠPoison": 51112, + "ĠPok": 22364, + "ĠPoke": 97626, + "ĠPokemon": 27731, + "ĠPoker": 37568, + "ĠPokémon": 26621, + "ĠPol": 3651, + "ĠPoland": 27602, + "ĠPolar": 55896, + "ĠPole": 52627, + "ĠPolic": 98112, + "ĠPolice": 10082, + "ĠPolicies": 62248, + "ĠPolicy": 10974, + "ĠPolish": 31984, + "ĠPolit": 15928, + "ĠPolitical": 30497, + "ĠPolitico": 70733, + "ĠPolitics": 34879, + "ĠPoll": 24385, + "ĠPollution": 95101, + "ĠPolly": 93417, + "ĠPolo": 72771, + "ĠPols": 57719, + "ĠPoly": 18767, + "ĠPolyester": 74849, + "ĠPolygon": 48159, + "ĠPolymer": 96670, + "ĠPolynomial": 75153, + "ĠPolÃŃcia": 142252, + "ĠPolÃŃtica": 139366, + "ĠPom": 38312, + "ĠPompe": 61411, + "ĠPompeo": 89634, + "ĠPon": 65315, + "ĠPond": 70865, + "ĠPont": 39770, + "ĠPony": 68035, + "ĠPool": 22728, + "ĠPoor": 44673, + "ĠPop": 10254, + "ĠPope": 29360, + "ĠPopular": 31395, + "ĠPopulate": 70938, + "ĠPopulation": 39529, + "ĠPopup": 46190, + "ĠPor": 19801, + "ĠPork": 83934, + "ĠPorn": 16472, + "ĠPornhub": 76907, + "ĠPorno": 20687, + "ĠPorsche": 50021, + "ĠPort": 5776, + "ĠPortable": 41485, + "ĠPortal": 33731, + "ĠPorter": 37937, + "ĠPortfolio": 46192, + "ĠPortions": 77797, + "ĠPortland": 23077, + "ĠPorto": 69698, + "ĠPortrait": 78582, + "ĠPorts": 69866, + "ĠPortsmouth": 86570, + "ĠPortug": 22234, + "ĠPortugal": 33311, + "ĠPortuguese": 42188, + "ĠPos": 18876, + "ĠPose": 50373, + "ĠPosition": 12380, + "ĠPositioned": 89981, + "ĠPositions": 75383, + "ĠPositive": 43903, + "ĠPoss": 41045, + "ĠPossible": 36107, + "ĠPossibly": 79191, + "ĠPost": 3877, + "ĠPostal": 54217, + "ĠPosted": 15275, + "ĠPoster": 62386, + "ĠPostgreSQL": 73601, + "ĠPosting": 78673, + "ĠPosts": 15419, + "ĠPosté": 136530, + "ĠPot": 13706, + "ĠPotato": 77603, + "ĠPotential": 53957, + "ĠPotion": 74841, + "ĠPotter": 29327, + "ĠPou": 89023, + "ĠPound": 78252, + "ĠPour": 25968, + "ĠPoverty": 69174, + "ĠPow": 23605, + "ĠPowder": 43656, + "ĠPowell": 43009, + "ĠPower": 7420, + "ĠPowerPoint": 53500, + "ĠPowerShell": 73962, + "ĠPowered": 59620, + "ĠPowerful": 74358, + "ĠPowers": 45037, + "ĠPoz": 78985, + "ĠPr": 2340, + "ĠPra": 40365, + "ĠPract": 17891, + "ĠPractical": 65636, + "ĠPractice": 26984, + "ĠPractices": 63713, + "ĠPradesh": 42358, + "ĠPrague": 67289, + "ĠPrairie": 70949, + "ĠPratt": 73987, + "ĠPrayer": 54960, + "ĠPre": 4968, + "ĠPrec": 41670, + "ĠPrecision": 51359, + "ĠPreconditions": 85916, + "ĠPred": 29824, + "ĠPredator": 87187, + "ĠPredicate": 49827, + "ĠPredict": 32710, + "ĠPrediction": 61865, + "ĠPredictor": 60418, + "ĠPref": 42588, + "ĠPrefer": 84468, + "ĠPreference": 48805, + "ĠPreferences": 47870, + "ĠPreferred": 62503, + "ĠPrefix": 56483, + "ĠPregn": 59647, + "ĠPregnancy": 82527, + "ĠPreis": 60939, + "ĠPrel": 40088, + "ĠPrelude": 52562, + "ĠPrem": 11767, + "ĠPremier": 19637, + "ĠPremiere": 68824, + "ĠPremiership": 96888, + "ĠPremium": 25661, + "ĠPrep": 48069, + "ĠPreparation": 73335, + "ĠPrepare": 31166, + "ĠPrepared": 87167, + "ĠPreparedStatement": 42058, + "ĠPres": 4111, + "ĠPresbyterian": 85364, + "ĠPrescott": 90779, + "ĠPrescription": 76075, + "ĠPresence": 78623, + "ĠPresent": 26642, + "ĠPresentation": 50868, + "ĠPresented": 87021, + "ĠPresenter": 96223, + "ĠPresents": 82736, + "ĠPreservation": 90941, + "ĠPreserve": 81306, + "ĠPresidency": 86648, + "ĠPresident": 4795, + "ĠPresidential": 41755, + "ĠPresidents": 78639, + "ĠPress": 8445, + "ĠPressure": 38990, + "ĠPrest": 34902, + "ĠPreston": 62225, + "ĠPret": 61939, + "ĠPretty": 35316, + "ĠPrev": 59001, + "ĠPrevent": 38068, + "ĠPrevention": 35847, + "ĠPreview": 31241, + "ĠPrevious": 28913, + "ĠPreviously": 58687, + "ĠPri": 27785, + "ĠPrice": 8483, + "ĠPrices": 36033, + "ĠPricing": 59326, + "ĠPride": 42146, + "ĠPriest": 48903, + "ĠPrim": 35183, + "ĠPrimary": 25094, + "ĠPrime": 12518, + "ĠPrimer": 87025, + "ĠPrimitive": 51460, + "ĠPrince": 19351, + "ĠPrincess": 29289, + "ĠPrinceton": 49321, + "ĠPrincip": 44529, + "ĠPrincipal": 36309, + "ĠPrinciple": 80986, + "ĠPrinciples": 56914, + "ĠPrint": 8213, + "ĠPrintWriter": 48011, + "ĠPrintable": 64753, + "ĠPrinted": 69965, + "ĠPrinter": 45933, + "ĠPrintf": 90908, + "ĠPrinting": 50769, + "ĠPrints": 49285, + "ĠPrior": 31399, + "ĠPriority": 31920, + "ĠPriorityQueue": 84102, + "ĠPrism": 71931, + "ĠPrison": 41518, + "ĠPriv": 15438, + "ĠPrivacy": 18874, + "ĠPrivate": 9679, + "ĠPrix": 43294, + "ĠPrize": 31193, + "ĠPro": 1298, + "ĠProb": 25627, + "ĠProbability": 86639, + "ĠProbably": 37154, + "ĠProbe": 77278, + "ĠProble": 74890, + "ĠProblem": 22079, + "ĠProblems": 44535, + "ĠProc": 41087, + "ĠProcedure": 44446, + "ĠProcedures": 74011, + "ĠProceed": 40770, + "ĠProceedings": 54127, + "ĠProcess": 8603, + "ĠProcesses": 62303, + "ĠProcessing": 28125, + "ĠProcessor": 43377, + "ĠProd": 71272, + "ĠProdu": 12042, + "ĠProduce": 86888, + "ĠProduced": 87265, + "ĠProducer": 43359, + "ĠProduct": 5643, + "ĠProductService": 93198, + "ĠProduction": 24039, + "ĠProductions": 51864, + "ĠProducto": 63864, + "ĠProducts": 15530, + "ĠProduk": 83600, + "ĠProdukt": 69342, + "ĠProdukte": 87479, + "ĠProduto": 67254, + "ĠProf": 8459, + "ĠProfession": 49211, + "ĠProfessional": 21237, + "ĠProfessionals": 70402, + "ĠProfessor": 16642, + "ĠProfile": 12329, + "ĠProfiles": 71727, + "ĠProfit": 57782, + "ĠProgram": 6687, + "ĠProgramm": 73910, + "ĠProgramme": 46416, + "ĠProgrammer": 88024, + "ĠProgramming": 38424, + "ĠPrograms": 37616, + "ĠProgress": 16033, + "ĠProgressBar": 62044, + "ĠProgressDialog": 61831, + "ĠProgressive": 51770, + "ĠProj": 42952, + "ĠProject": 5787, + "ĠProjectile": 83602, + "ĠProjection": 69315, + "ĠProjects": 31223, + "ĠProjekt": 77742, + "ĠProm": 17582, + "ĠPrometheus": 97137, + "ĠPromise": 7303, + "ĠPromo": 57403, + "ĠPromotion": 56104, + "ĠPrompt": 59501, + "ĠPron": 86915, + "ĠProof": 36991, + "ĠProp": 3911, + "ĠPropTypes": 9602, + "ĠPropel": 43859, + "ĠPropelException": 85293, + "ĠProper": 64558, + "ĠProperties": 11831, + "ĠProperty": 8655, + "ĠPropertyChanged": 43189, + "ĠPropertyChangedEventArgs": 92116, + "ĠPropertyInfo": 94423, + "ĠPropertyValue": 94338, + "ĠProphet": 36628, + "ĠProposal": 65890, + "ĠProposed": 98883, + "ĠProposition": 86755, + "ĠProps": 29306, + "ĠPros": 31034, + "ĠProsec": 51170, + "ĠProsecutor": 87328, + "ĠProspect": 77815, + "ĠProsper": 88713, + "ĠProstit": 52817, + "ĠProstitutas": 98226, + "ĠProt": 11708, + "ĠProte": 15187, + "ĠProtect": 33663, + "ĠProtected": 36995, + "ĠProtection": 19171, + "ĠProtective": 71142, + "ĠProtector": 81119, + "ĠProtein": 48375, + "ĠProtest": 36659, + "ĠProtestant": 59653, + "ĠProto": 57677, + "ĠProtocol": 24572, + "ĠPrototype": 76033, + "ĠProud": 83249, + "ĠProv": 58329, + "ĠProvid": 20516, + "ĠProvide": 39565, + "ĠProvided": 53874, + "ĠProvidence": 57841, + "ĠProvider": 22916, + "ĠProviders": 69929, + "ĠProvides": 36617, + "ĠProviding": 80100, + "ĠProvince": 37794, + "ĠProvincial": 79695, + "ĠProvision": 71573, + "ĠProxy": 32778, + "ĠProzent": 78233, + "ĠPru": 73308, + "ĠPruitt": 77806, + "ĠPry": 73414, + "ĠPrä": 138971, + "ĠPré": 134282, + "ĠPs": 11802, + "ĠPsalm": 81876, + "ĠPsi": 53451, + "ĠPsr": 89422, + "ĠPsy": 40018, + "ĠPsych": 17235, + "ĠPsychiat": 60664, + "ĠPsychiatry": 68144, + "ĠPsychic": 92875, + "ĠPsycho": 69176, + "ĠPsychological": 68969, + "ĠPsychology": 35573, + "ĠPt": 51070, + "ĠPtr": 46409, + "ĠPty": 78992, + "ĠPu": 22199, + "ĠPub": 22611, + "ĠPubMed": 52668, + "ĠPubli": 138190, + "ĠPublic": 3066, + "ĠPublicKey": 70280, + "ĠPublication": 56894, + "ĠPublications": 61670, + "ĠPublish": 23499, + "ĠPublished": 29014, + "ĠPublisher": 39133, + "ĠPublishers": 71614, + "ĠPublishing": 36833, + "ĠPublié": 138191, + "ĠPuerto": 30219, + "ĠPul": 34862, + "ĠPulitzer": 95715, + "ĠPull": 31828, + "ĠPulse": 49249, + "ĠPump": 33726, + "ĠPumpkin": 75257, + "ĠPun": 30436, + "ĠPunch": 53175, + "ĠPune": 75238, + "ĠPunjab": 59317, + "ĠPunk": 51333, + "ĠPuppet": 73880, + "ĠPuppy": 95533, + "ĠPur": 14532, + "ĠPurch": 52692, + "ĠPurchase": 30988, + "ĠPurdue": 72234, + "ĠPure": 29588, + "ĠPureComponent": 89088, + "ĠPurple": 40389, + "ĠPurpose": 29045, + "ĠPurs": 70660, + "ĠPush": 22950, + "ĠPussy": 46845, + "ĠPut": 10224, + "ĠPutin": 21126, + "ĠPutting": 77890, + "ĠPuzzle": 60012, + "ĠPv": 50339, + "ĠPvP": 65852, + "ĠPvt": 60675, + "ĠPx": 95493, + "ĠPy": 5355, + "ĠPyErr": 75667, + "ĠPyObject": 15891, + "ĠPyQt": 48677, + "ĠPyTuple": 80824, + "ĠPyongyang": 60059, + "ĠPyramid": 82395, + "ĠPython": 13027, + "ĠPÃ¥": 93025, + "ĠPé": 79867, + "ĠPúb": 140039, + "ĠPública": 140006, + "ĠPúblico": 140040, + "ĠQ": 1207, + "ĠQA": 65908, + "ĠQAction": 55062, + "ĠQApplication": 62449, + "ĠQB": 36570, + "ĠQByteArray": 49399, + "ĠQC": 42607, + "ĠQCOMPARE": 59145, + "ĠQColor": 55758, + "ĠQDateTime": 98724, + "ĠQDialog": 69077, + "ĠQDir": 83842, + "ĠQDom": 98523, + "ĠQE": 99016, + "ĠQFile": 78970, + "ĠQFont": 97600, + "ĠQGraphics": 78303, + "ĠQHBoxLayout": 81291, + "ĠQIcon": 94401, + "ĠQImage": 77427, + "ĠQLD": 92567, + "ĠQLabel": 43458, + "ĠQLatin": 74800, + "ĠQLineEdit": 78649, + "ĠQList": 37251, + "ĠQMainWindow": 92922, + "ĠQMap": 98261, + "ĠQMessageBox": 41218, + "ĠQModelIndex": 48599, + "ĠQName": 76381, + "ĠQObject": 35050, + "ĠQPainter": 97337, + "ĠQPixmap": 80430, + "ĠQPoint": 47743, + "ĠQPointF": 76385, + "ĠQPushButton": 52522, + "ĠQQ": 73115, + "ĠQR": 38699, + "ĠQRect": 65049, + "ĠQS": 77746, + "ĠQSize": 86257, + "ĠQSql": 74988, + "ĠQString": 7879, + "ĠQStringList": 42862, + "ĠQStringLiteral": 70697, + "ĠQT": 47305, + "ĠQTableWidgetItem": 90483, + "ĠQTest": 93656, + "ĠQText": 65818, + "ĠQTimer": 96669, + "ĠQU": 25999, + "ĠQUAL": 70275, + "ĠQUE": 29676, + "ĠQUERY": 67135, + "ĠQUEST": 56728, + "ĠQUESTION": 89922, + "ĠQUEUE": 97239, + "ĠQUI": 88530, + "ĠQUICK": 93012, + "ĠQVBoxLayout": 75155, + "ĠQVERIFY": 66034, + "ĠQVariant": 40066, + "ĠQVector": 54255, + "ĠQWidget": 36278, + "ĠQaeda": 53018, + "ĠQatar": 41918, + "ĠQed": 60588, + "ĠQgs": 61008, + "ĠQi": 56994, + "ĠQin": 73351, + "ĠQing": 61912, + "ĠQt": 10594, + "ĠQtAws": 92805, + "ĠQtCore": 53347, + "ĠQtGui": 39640, + "ĠQtWidgets": 32536, + "ĠQty": 76557, + "ĠQu": 3406, + "ĠQuad": 38018, + "ĠQuadr": 63948, + "ĠQual": 19572, + "ĠQualcomm": 61402, + "ĠQualified": 60306, + "ĠQuality": 17927, + "ĠQualität": 98060, + "ĠQuando": 91666, + "ĠQuant": 31441, + "ĠQuantity": 33523, + "ĠQuantum": 55313, + "ĠQuar": 92745, + "ĠQuarry": 71083, + "ĠQuart": 69084, + "ĠQuarter": 34394, + "ĠQuarterly": 74383, + "ĠQuartz": 65088, + "ĠQuaternion": 24801, + "ĠQue": 9418, + "ĠQuebec": 34095, + "ĠQueen": 16259, + "ĠQueens": 27063, + "ĠQueensland": 40206, + "ĠQuentin": 90051, + "ĠQueries": 85873, + "ĠQuery": 11361, + "ĠQuest": 15053, + "ĠQuestion": 15846, + "ĠQuestions": 23382, + "ĠQueue": 18745, + "ĠQui": 41115, + "ĠQuick": 17251, + "ĠQuickly": 96817, + "ĠQuiet": 67487, + "ĠQuincy": 97398, + "ĠQuinn": 40523, + "ĠQuint": 50699, + "ĠQuit": 53552, + "ĠQuite": 57695, + "ĠQuiz": 41148, + "ĠQuote": 24535, + "ĠQuotes": 35613, + "ĠQur": 67111, + "ĠQuran": 65660, + "ĠQuân": 137833, + "ĠQué": 65603, + "ĠQuébec": 79188, + "ĠQuý": 135595, + "ĠQuản": 140428, + "ĠQuảng": 133314, + "ĠQuáºŃn": 135642, + "ĠQuá»ijc": 128494, + "ĠR": 431, + "ĠRA": 28456, + "ĠRAD": 74150, + "ĠRADIO": 82160, + "ĠRAF": 71052, + "ĠRAID": 77632, + "ĠRAM": 22038, + "ĠRAND": 72854, + "ĠRANDOM": 82711, + "ĠRANGE": 86506, + "ĠRAT": 97880, + "ĠRATE": 92470, + "ĠRAW": 52851, + "ĠRB": 27974, + "ĠRBI": 52941, + "ĠRC": 21594, + "ĠRCA": 98331, + "ĠRCC": 31917, + "ĠRCMP": 74624, + "ĠRCS": 79156, + "ĠRD": 37992, + "ĠRDD": 77255, + "ĠRDF": 68399, + "ĠRE": 3596, + "ĠREAD": 18666, + "ĠREADME": 61945, + "ĠREADY": 85332, + "ĠREAL": 25272, + "ĠREALLY": 65700, + "ĠREC": 74136, + "ĠRECE": 42285, + "ĠRECEIVE": 88875, + "ĠRECEIVER": 76176, + "ĠRECORD": 55993, + "ĠRECT": 75803, + "ĠRED": 25810, + "ĠREF": 38029, + "ĠREFER": 68685, + "ĠREFERENCES": 78656, + "ĠREG": 13676, + "ĠREGARD": 96567, + "ĠREGION": 64160, + "ĠREGISTER": 52637, + "ĠREL": 31512, + "ĠRELATED": 84347, + "ĠRELEASE": 53034, + "ĠREM": 25587, + "ĠREMOVE": 48783, + "ĠRENDER": 91788, + "ĠREP": 49260, + "ĠREPL": 92580, + "ĠREPLACE": 84371, + "ĠREPORT": 44076, + "ĠREPRESENT": 97122, + "ĠREQ": 78696, + "ĠREQUEST": 33416, + "ĠREQUIRE": 27972, + "ĠREQUIRED": 66577, + "ĠRES": 13170, + "ĠRESERVED": 77739, + "ĠRESET": 53215, + "ĠRESOURCE": 76065, + "ĠRESP": 45477, + "ĠRESPONS": 88979, + "ĠRESPONSE": 76173, + "ĠREST": 25414, + "ĠRESULT": 28536, + "ĠRESULTS": 53028, + "ĠRET": 18496, + "ĠRETURN": 30880, + "ĠRETURNS": 68039, + "ĠREUTERS": 51791, + "ĠREV": 65823, + "ĠREVIEW": 80051, + "ĠRF": 22060, + "ĠRFC": 39233, + "ĠRFID": 85879, + "ĠRG": 50388, + "ĠRGB": 20978, + "ĠRGBA": 79954, + "ĠRH": 56156, + "ĠRHS": 73832, + "ĠRI": 47497, + "ĠRID": 27326, + "ĠRIGHT": 27477, + "ĠRIGHTS": 77436, + "ĠRIP": 90864, + "ĠRJ": 76304, + "ĠRK": 67137, + "ĠRL": 47496, + "ĠRM": 30815, + "ĠRMS": 77178, + "ĠRN": 45816, + "ĠRNA": 40114, + "ĠRNG": 72606, + "ĠRO": 11813, + "ĠROAD": 82193, + "ĠROC": 97826, + "ĠROCK": 78520, + "ĠROI": 50652, + "ĠROLE": 69676, + "ĠROM": 30686, + "ĠROOM": 57948, + "ĠROOT": 35736, + "ĠROS": 37417, + "ĠROT": 58211, + "ĠROUND": 76138, + "ĠROUT": 53181, + "ĠROW": 45715, + "ĠRP": 41461, + "ĠRPC": 34008, + "ĠRPG": 33502, + "ĠRPM": 53774, + "ĠRR": 43398, + "ĠRS": 23229, + "ĠRSA": 45641, + "ĠRSS": 29274, + "ĠRSVP": 81889, + "ĠRT": 10635, + "ĠRTAL": 76160, + "ĠRTBU": 93707, + "ĠRTC": 35047, + "ĠRTE": 81855, + "ĠRTL": 63735, + "ĠRTP": 85627, + "ĠRTS": 84948, + "ĠRTVF": 87578, + "ĠRTWF": 62685, + "ĠRU": 68825, + "ĠRULE": 43797, + "ĠRUN": 30760, + "ĠRUNNING": 97677, + "ĠRV": 32470, + "ĠRVA": 44596, + "ĠRW": 46206, + "ĠRX": 28170, + "ĠRa": 18482, + "ĠRab": 35724, + "ĠRabbi": 70869, + "ĠRabbit": 48331, + "ĠRac": 55706, + "ĠRace": 23659, + "ĠRaces": 97100, + "ĠRach": 83385, + "ĠRachel": 31432, + "ĠRacing": 32282, + "ĠRack": 50039, + "ĠRad": 20605, + "ĠRadar": 67786, + "ĠRadeon": 37429, + "ĠRadi": 34869, + "ĠRadiation": 89226, + "ĠRadical": 87881, + "ĠRadio": 13484, + "ĠRadioButton": 64511, + "ĠRadius": 40453, + "ĠRae": 89499, + "ĠRaf": 39577, + "ĠRafael": 54400, + "ĠRag": 50259, + "ĠRage": 65467, + "ĠRagnar": 94262, + "ĠRah": 32248, + "ĠRahman": 97276, + "ĠRahmen": 89507, + "ĠRahul": 85136, + "ĠRai": 86434, + "ĠRaid": 32700, + "ĠRaider": 80354, + "ĠRaiders": 42036, + "ĠRail": 23786, + "ĠRailroad": 69017, + "ĠRails": 36405, + "ĠRailway": 54456, + "ĠRain": 21911, + "ĠRainbow": 46645, + "ĠRaise": 42546, + "ĠRaisePropertyChanged": 89330, + "ĠRaised": 53731, + "ĠRaisedButton": 79701, + "ĠRaises": 60311, + "ĠRaj": 25225, + "ĠRajasthan": 86340, + "ĠRak": 68792, + "ĠRaleigh": 64083, + "ĠRally": 58700, + "ĠRalph": 39804, + "ĠRam": 15152, + "ĠRamadan": 81808, + "ĠRamirez": 70993, + "ĠRamos": 68813, + "ĠRamp": 74570, + "ĠRams": 37503, + "ĠRamsey": 64546, + "ĠRan": 38331, + "ĠRanch": 41882, + "ĠRand": 23963, + "ĠRandall": 67344, + "ĠRandolph": 85752, + "ĠRandom": 10612, + "ĠRandomForest": 90714, + "ĠRandy": 45215, + "ĠRange": 16437, + "ĠRanger": 48322, + "ĠRangers": 33187, + "ĠRank": 19298, + "ĠRanked": 96948, + "ĠRanking": 64075, + "ĠRankings": 82799, + "ĠRao": 87900, + "ĠRap": 22300, + "ĠRape": 52605, + "ĠRaphael": 89886, + "ĠRapid": 46990, + "ĠRapids": 54186, + "ĠRaptors": 73578, + "ĠRaq": 96409, + "ĠRaqqa": 97860, + "ĠRare": 34959, + "ĠRarity": 65724, + "ĠRas": 58030, + "ĠRash": 59383, + "ĠRaspberry": 47565, + "ĠRat": 40577, + "ĠRate": 19775, + "ĠRated": 37460, + "ĠRates": 46976, + "ĠRath": 85455, + "ĠRather": 25764, + "ĠRating": 19216, + "ĠRatings": 56965, + "ĠRatio": 50748, + "ĠRational": 54525, + "ĠRaum": 95602, + "ĠRav": 33974, + "ĠRaven": 45107, + "ĠRavens": 47480, + "ĠRaw": 23022, + "ĠRay": 13255, + "ĠRaycastHit": 95869, + "ĠRaymond": 42570, + "ĠRays": 79675, + "ĠRaz": 76819, + "ĠRazor": 73868, + "ĠRc": 81463, + "ĠRd": 27467, + "ĠRe": 1032, + "ĠReSharper": 52155, + "ĠReach": 49842, + "ĠReached": 57325, + "ĠReact": 3592, + "ĠReactDOM": 39117, + "ĠReaction": 66841, + "ĠReactive": 39822, + "ĠReactiveFormsModule": 52671, + "ĠRead": 4457, + "ĠReadOnly": 43406, + "ĠReader": 25166, + "ĠReaders": 64414, + "ĠReading": 17769, + "ĠReadonly": 93359, + "ĠReads": 43721, + "ĠReady": 30982, + "ĠReagan": 34126, + "ĠReal": 8800, + "ĠReality": 43774, + "ĠReally": 28208, + "ĠRealm": 41106, + "ĠRealt": 87872, + "ĠRealty": 68014, + "ĠReaper": 83422, + "ĠRear": 45902, + "ĠReason": 26759, + "ĠReasons": 65031, + "ĠReb": 51643, + "ĠRebecca": 44457, + "ĠRebel": 63164, + "ĠRebellion": 93979, + "ĠRebels": 89335, + "ĠRec": 4067, + "ĠRecall": 79540, + "ĠRecap": 98034, + "ĠRece": 60296, + "ĠReceipt": 72896, + "ĠReceive": 37422, + "ĠReceived": 38417, + "ĠReceiver": 50952, + "ĠRecent": 34290, + "ĠRecently": 40996, + "ĠReception": 75951, + "ĠRecipe": 25303, + "ĠRecipes": 44202, + "ĠReco": 31462, + "ĠRecogn": 66875, + "ĠRecognition": 47598, + "ĠRecognitionException": 79316, + "ĠRecomm": 29801, + "ĠRecommend": 46606, + "ĠRecommendation": 96531, + "ĠRecommendations": 88420, + "ĠRecommended": 50662, + "ĠRecon": 82981, + "ĠReconstruction": 94694, + "ĠRecord": 13583, + "ĠRecorded": 96364, + "ĠRecorder": 85519, + "ĠRecording": 60547, + "ĠRecords": 21566, + "ĠRecover": 76112, + "ĠRecovery": 33911, + "ĠRecreation": 56757, + "ĠRecru": 82640, + "ĠRecruitment": 56444, + "ĠRect": 11920, + "ĠRectTransform": 96697, + "ĠRectangle": 19280, + "ĠRecursive": 85896, + "ĠRecyclerView": 22562, + "ĠRecycling": 80585, + "ĠRed": 3731, + "ĠReddit": 26630, + "ĠRede": 72131, + "ĠRedemption": 88641, + "ĠRedirect": 16029, + "ĠRedirectTo": 26641, + "ĠRedirectToAction": 28177, + "ĠRedis": 34158, + "ĠRedistribution": 31943, + "ĠRedistributions": 19605, + "ĠReds": 54808, + "ĠRedskins": 59278, + "ĠReduce": 52153, + "ĠReduced": 79469, + "ĠReduction": 58100, + "ĠRedux": 68011, + "ĠRee": 56821, + "ĠReed": 35421, + "ĠReef": 75936, + "ĠReese": 81539, + "ĠReeves": 81752, + "ĠRef": 8550, + "ĠRefer": 28634, + "ĠReference": 17207, + "ĠReferences": 44078, + "ĠReferentialAction": 58770, + "ĠReflect": 34598, + "ĠReflection": 42876, + "ĠReflex": 93393, + "ĠReform": 39265, + "ĠRefresh": 29007, + "ĠRefriger": 74158, + "ĠRefuge": 51119, + "ĠRefugee": 97072, + "ĠReg": 3184, + "ĠRegExp": 28487, + "ĠRegarding": 72673, + "ĠRegardless": 43740, + "ĠRegards": 91079, + "ĠRegel": 97504, + "ĠRegents": 77677, + "ĠRegex": 26146, + "ĠRegexOptions": 84598, + "ĠReggie": 88882, + "ĠRegiment": 72686, + "ĠRegina": 69680, + "ĠRegion": 17152, + "ĠRegional": 25294, + "ĠRegions": 77347, + "ĠRegister": 8451, + "ĠRegistered": 35539, + "ĠRegisters": 54774, + "ĠRegistr": 74559, + "ĠRegistrar": 80226, + "ĠRegistration": 24516, + "ĠRegistro": 83637, + "ĠRegistry": 32112, + "ĠRegression": 47470, + "ĠRegular": 28800, + "ĠRegulation": 47238, + "ĠRegulations": 48257, + "ĠRegulatory": 68722, + "ĠRehab": 60379, + "ĠRehabilitation": 80218, + "ĠRei": 83141, + "ĠReich": 50559, + "ĠReid": 37674, + "ĠReign": 80871, + "ĠRein": 46069, + "ĠReject": 87293, + "ĠRel": 6512, + "ĠRelated": 24369, + "ĠRelation": 50024, + "ĠRelations": 31367, + "ĠRelationship": 32807, + "ĠRelationships": 68568, + "ĠRelative": 39402, + "ĠRelativeLayout": 69228, + "ĠRelatives": 75477, + "ĠRelax": 67585, + "ĠRelay": 46172, + "ĠRelayCommand": 76843, + "ĠRelease": 17381, + "ĠReleased": 44794, + "ĠReleases": 61015, + "ĠRelevant": 87416, + "ĠReliable": 91563, + "ĠRelief": 52395, + "ĠReligion": 43093, + "ĠReligious": 52076, + "ĠReload": 56341, + "ĠRem": 4926, + "ĠRemain": 91592, + "ĠRemaining": 89630, + "ĠRemark": 82733, + "ĠRemarks": 84582, + "ĠRemed": 96870, + "ĠRemember": 19881, + "ĠReminder": 95902, + "ĠRemix": 50027, + "ĠRemote": 20738, + "ĠRemoteException": 60209, + "ĠRemoval": 56717, + "ĠRemove": 10783, + "ĠRemoved": 51083, + "ĠRemoves": 38775, + "ĠRemoving": 66703, + "ĠRen": 13775, + "ĠRena": 48794, + "ĠRenaissance": 54283, + "ĠRename": 63708, + "ĠRenault": 66331, + "ĠRencontre": 44585, + "ĠRencontres": 93373, + "ĠRender": 13848, + "ĠRenderer": 41367, + "ĠRendering": 64637, + "ĠRenders": 93733, + "ĠRene": 66427, + "ĠRenew": 48986, + "ĠRenewable": 92338, + "ĠReno": 63418, + "ĠRent": 29737, + "ĠRental": 39198, + "ĠRentals": 83117, + "ĠRep": 3321, + "ĠRepair": 27167, + "ĠRepeat": 44801, + "ĠReplace": 29558, + "ĠReplacement": 46212, + "ĠReplay": 59061, + "ĠReplica": 94036, + "ĠReplies": 63339, + "ĠReply": 17841, + "ĠRepo": 71509, + "ĠReport": 8259, + "ĠReported": 78611, + "ĠReporter": 46981, + "ĠReporting": 46693, + "ĠReports": 28040, + "ĠRepos": 24286, + "ĠRepository": 26377, + "ĠRepresent": 21950, + "ĠRepresentation": 78046, + "ĠRepresentative": 37266, + "ĠRepresentatives": 39745, + "ĠRepresents": 38893, + "ĠRepublic": 5429, + "ĠRepublican": 9356, + "ĠRepublicans": 12773, + "ĠReputation": 88080, + "ĠRepública": 136962, + "ĠReq": 82458, + "ĠRequest": 6145, + "ĠRequestContext": 90057, + "ĠRequestMethod": 29557, + "ĠRequestOptions": 82349, + "ĠRequests": 50882, + "ĠRequire": 24046, + "ĠRequired": 12659, + "ĠRequirement": 75938, + "ĠRequirements": 33784, + "ĠRequires": 44089, + "ĠRes": 1800, + "ĠRescue": 44403, + "ĠResearch": 8319, + "ĠResearchers": 58150, + "ĠReservation": 53312, + "ĠReserve": 24785, + "ĠReserved": 15163, + "ĠReset": 16932, + "ĠResets": 96997, + "ĠResidence": 71760, + "ĠResident": 59676, + "ĠResidential": 54455, + "ĠResidents": 63248, + "ĠResist": 78868, + "ĠResistance": 39880, + "ĠResize": 63343, + "ĠResolution": 37116, + "ĠResolve": 47302, + "ĠResolver": 81854, + "ĠResort": 33837, + "ĠResorts": 85271, + "ĠResource": 11765, + "ĠResourceBundle": 95361, + "ĠResourceManager": 87269, + "ĠResourceType": 97801, + "ĠResources": 16209, + "ĠResp": 79786, + "ĠRespect": 77038, + "ĠRespir": 76834, + "ĠRespond": 39533, + "ĠRespons": 32071, + "ĠResponse": 5949, + "ĠResponseEntity": 19803, + "ĠResponses": 80467, + "ĠResponsibilities": 73012, + "ĠResponsibility": 77612, + "ĠResponsible": 67628, + "ĠResponsive": 70465, + "ĠRest": 9063, + "ĠRestClient": 96536, + "ĠRestart": 56801, + "ĠRestaurant": 25488, + "ĠRestaurants": 58093, + "ĠRestoration": 66825, + "ĠRestore": 43820, + "ĠRestr": 41327, + "ĠRestricted": 93860, + "ĠRestrictions": 91975, + "ĠResult": 5714, + "ĠResultSet": 33582, + "ĠResults": 18099, + "ĠResume": 33398, + "ĠRet": 10392, + "ĠRetail": 34039, + "ĠRetirement": 69189, + "ĠRetorna": 94111, + "ĠRetreat": 87394, + "ĠRetrie": 19470, + "ĠRetrieve": 31562, + "ĠRetrieved": 57791, + "ĠRetrieves": 48661, + "ĠRetro": 36526, + "ĠRetrofit": 57398, + "ĠRetry": 78870, + "ĠReturn": 3411, + "ĠReturnType": 88156, + "ĠReturnValue": 96584, + "ĠReturned": 75264, + "ĠReturning": 73247, + "ĠReturns": 5184, + "ĠReuters": 22841, + "ĠRev": 10107, + "ĠReve": 66553, + "ĠRevel": 54743, + "ĠRevelation": 75507, + "ĠRevenge": 80549, + "ĠRevenue": 37393, + "ĠRever": 82179, + "ĠReverse": 24277, + "ĠReview": 10294, + "ĠReviewed": 77751, + "ĠReviews": 19276, + "ĠRevised": 66730, + "ĠRevision": 26795, + "ĠRevolution": 22126, + "ĠRevolutionary": 66579, + "ĠRew": 31691, + "ĠReward": 55800, + "ĠRewards": 49768, + "ĠRewrite": 93213, + "ĠRex": 41807, + "ĠRey": 48212, + "ĠReyes": 76159, + "ĠReyn": 39083, + "ĠReynolds": 45072, + "ĠRh": 17968, + "ĠRhe": 70536, + "ĠRhino": 79465, + "ĠRhode": 49881, + "ĠRhodes": 56611, + "ĠRi": 51730, + "ĠRib": 63105, + "ĠRibbon": 76467, + "ĠRic": 32552, + "ĠRica": 51949, + "ĠRican": 95664, + "ĠRicardo": 65950, + "ĠRice": 29516, + "ĠRich": 8107, + "ĠRichard": 11867, + "ĠRichards": 52972, + "ĠRichardson": 45723, + "ĠRichie": 98974, + "ĠRichmond": 34248, + "ĠRicht": 57123, + "ĠRick": 22388, + "ĠRicky": 65793, + "ĠRico": 33148, + "ĠRid": 44895, + "ĠRide": 40913, + "ĠRider": 53842, + "ĠRiders": 86070, + "ĠRidge": 34205, + "ĠRiding": 76319, + "ĠRidley": 92301, + "ĠRif": 89731, + "ĠRifle": 47038, + "ĠRift": 56327, + "ĠRig": 51532, + "ĠRight": 10083, + "ĠRights": 10512, + "ĠRigidbody": 64608, + "ĠRihanna": 95427, + "ĠRiley": 46835, + "ĠRim": 51379, + "ĠRin": 57368, + "ĠRing": 21525, + "ĠRings": 46280, + "ĠRio": 26961, + "ĠRiot": 64706, + "ĠRip": 35437, + "ĠRipple": 69707, + "ĠRis": 53363, + "ĠRise": 42677, + "ĠRising": 48887, + "ĠRisk": 31288, + "ĠRit": 70122, + "ĠRita": 70968, + "ĠRitch": 96882, + "ĠRitual": 87788, + "ĠRiv": 50668, + "ĠRiver": 10948, + "ĠRivera": 63110, + "ĠRivers": 35639, + "ĠRiverside": 62841, + "ĠRiy": 78790, + "ĠRiyadh": 89790, + "ĠRo": 11830, + "ĠRoad": 9536, + "ĠRoads": 79253, + "ĠRob": 4892, + "ĠRobbie": 73132, + "ĠRobbins": 87215, + "ĠRobert": 8397, + "ĠRoberto": 54083, + "ĠRoberts": 30148, + "ĠRobertson": 52496, + "ĠRobin": 17143, + "ĠRobinson": 27181, + "ĠRobot": 25860, + "ĠRobotics": 76464, + "ĠRobots": 96046, + "ĠRoc": 90826, + "ĠRoch": 35869, + "ĠRochester": 52368, + "ĠRock": 9127, + "ĠRockefeller": 84638, + "ĠRocket": 39218, + "ĠRockets": 60900, + "ĠRockies": 79051, + "ĠRocks": 68031, + "ĠRocky": 47541, + "ĠRod": 13308, + "ĠRodgers": 49017, + "ĠRodney": 74943, + "ĠRodr": 56725, + "ĠRodrig": 36385, + "ĠRodrigo": 77768, + "ĠRodriguez": 40632, + "ĠRoe": 78452, + "ĠRog": 15259, + "ĠRoger": 28507, + "ĠRogers": 33367, + "ĠRogue": 49784, + "ĠRoh": 40987, + "ĠRohing": 91160, + "ĠRohingya": 96751, + "ĠRoku": 95992, + "ĠRol": 30976, + "ĠRoland": 57806, + "ĠRole": 15404, + "ĠRoles": 50907, + "ĠRolex": 96532, + "ĠRoll": 14686, + "ĠRolled": 93391, + "ĠRoller": 44542, + "ĠRolling": 44948, + "ĠRolls": 69610, + "ĠRom": 11774, + "ĠRoma": 45501, + "ĠRoman": 12751, + "ĠRomance": 51156, + "ĠRomania": 46049, + "ĠRomanian": 73597, + "ĠRomans": 48717, + "ĠRomantic": 75730, + "ĠRome": 21718, + "ĠRomeo": 70192, + "ĠRomero": 77387, + "ĠRomney": 25318, + "ĠRon": 14325, + "ĠRonald": 38648, + "ĠRonaldo": 64415, + "ĠRonnie": 83705, + "ĠRoo": 66714, + "ĠRoof": 46349, + "ĠRookie": 82125, + "ĠRoom": 10420, + "ĠRooms": 47303, + "ĠRooney": 79630, + "ĠRoose": 44318, + "ĠRoosevelt": 45942, + "ĠRoot": 18854, + "ĠRootState": 97312, + "ĠRoots": 74587, + "ĠRope": 97896, + "ĠRory": 82321, + "ĠRos": 16465, + "ĠRosa": 46830, + "ĠRose": 15964, + "ĠRosen": 34020, + "ĠRosenberg": 96607, + "ĠRosenstein": 90018, + "ĠRoses": 80370, + "ĠRosie": 96767, + "ĠRoss": 20482, + "ĠRossi": 76069, + "ĠRot": 27360, + "ĠRotary": 74856, + "ĠRotate": 49720, + "ĠRotation": 47528, + "ĠRoth": 36412, + "ĠRoths": 95915, + "ĠRotterdam": 97777, + "ĠRou": 28522, + "ĠRouge": 55647, + "ĠRough": 57321, + "ĠRoulette": 79504, + "ĠRound": 17097, + "ĠRounded": 57450, + "ĠRoundedRectangle": 77715, + "ĠRoundedRectangleBorder": 82957, + "ĠRounds": 97823, + "ĠRousse": 97889, + "ĠRout": 92022, + "ĠRoute": 9572, + "ĠRouteServiceProvider": 85702, + "ĠRoutedEventArgs": 31066, + "ĠRouter": 10554, + "ĠRouterModule": 34787, + "ĠRoutes": 22356, + "ĠRoutine": 71062, + "ĠRouting": 65707, + "ĠRover": 52176, + "ĠRow": 10801, + "ĠRowe": 94840, + "ĠRowling": 95507, + "ĠRows": 61706, + "ĠRox": 71423, + "ĠRoy": 11038, + "ĠRoyal": 16194, + "ĠRoyale": 86199, + "ĠRoyals": 69531, + "ĠRoz": 90480, + "ĠRp": 68108, + "ĠRpc": 79961, + "ĠRs": 19215, + "ĠRt": 75483, + "ĠRu": 27914, + "ĠRub": 12839, + "ĠRubber": 52306, + "ĠRubin": 81926, + "ĠRubio": 43914, + "ĠRuby": 23726, + "ĠRud": 47438, + "ĠRudd": 85363, + "ĠRudy": 79119, + "ĠRue": 78051, + "ĠRuf": 98157, + "ĠRug": 36350, + "ĠRugby": 50902, + "ĠRuiz": 89883, + "ĠRule": 18100, + "ĠRules": 22847, + "ĠRum": 45232, + "ĠRun": 6452, + "ĠRune": 70778, + "ĠRunnable": 22109, + "ĠRunner": 44946, + "ĠRunning": 28025, + "ĠRuns": 49990, + "ĠRuntime": 10954, + "ĠRuntimeError": 44543, + "ĠRuntimeException": 21247, + "ĠRuntimeMethod": 38778, + "ĠRuntimeObject": 53659, + "ĠRupert": 86169, + "ĠRural": 57220, + "ĠRus": 40976, + "ĠRush": 34075, + "ĠRuss": 4507, + "ĠRussell": 24909, + "ĠRussia": 8359, + "ĠRussian": 8522, + "ĠRussians": 33966, + "ĠRusso": 92891, + "ĠRust": 33789, + "ĠRut": 47121, + "ĠRutgers": 70991, + "ĠRuth": 41793, + "ĠRw": 55294, + "ĠRwanda": 78104, + "ĠRx": 35376, + "ĠRy": 25691, + "ĠRyan": 13646, + "ĠRyder": 88023, + "ĠRyu": 98231, + "ĠRyzen": 92736, + "ĠRé": 50123, + "ĠRück": 75074, + "ĠS": 328, + "ĠSA": 16589, + "ĠSAC": 89933, + "ĠSAF": 84104, + "ĠSAFE": 65869, + "ĠSAL": 53109, + "ĠSALE": 63880, + "ĠSAM": 46587, + "ĠSAME": 83490, + "ĠSAMPLE": 62420, + "ĠSAN": 59700, + "ĠSAND": 97241, + "ĠSAP": 36221, + "ĠSAR": 68778, + "ĠSAS": 50726, + "ĠSAT": 42254, + "ĠSATA": 72178, + "ĠSAVE": 50622, + "ĠSAX": 74226, + "ĠSAY": 84629, + "ĠSB": 25247, + "ĠSC": 7531, + "ĠSCALE": 78123, + "ĠSCAN": 98315, + "ĠSCC": 79299, + "ĠSCE": 93081, + "ĠSCH": 74838, + "ĠSCHOOL": 70401, + "ĠSCI": 80463, + "ĠSCIP": 82250, + "ĠSCM": 68548, + "ĠSCN": 92525, + "ĠSCO": 63134, + "ĠSCORE": 69254, + "ĠSCP": 48523, + "ĠSCR": 76700, + "ĠSCREEN": 35073, + "ĠSCRIPT": 53679, + "ĠSCSI": 81626, + "ĠSCT": 94285, + "ĠSD": 8030, + "ĠSDK": 26623, + "ĠSDL": 14653, + "ĠSDLK": 61906, + "ĠSDS": 95612, + "ĠSE": 5052, + "ĠSEA": 77811, + "ĠSEAL": 87579, + "ĠSEARCH": 49766, + "ĠSEC": 21638, + "ĠSECOND": 68695, + "ĠSECRET": 69956, + "ĠSECTION": 42333, + "ĠSECURITY": 70610, + "ĠSEE": 26103, + "ĠSEEK": 46240, + "ĠSEG": 65022, + "ĠSEL": 40838, + "ĠSELECT": 19094, + "ĠSELF": 66225, + "ĠSEM": 45444, + "ĠSEN": 84807, + "ĠSEND": 55663, + "ĠSENSOR": 67282, + "ĠSENT": 95751, + "ĠSEO": 24980, + "ĠSEP": 73769, + "ĠSEQ": 71408, + "ĠSER": 13984, + "ĠSERIAL": 70718, + "ĠSERIES": 95844, + "ĠSERVER": 39664, + "ĠSERVICE": 38086, + "ĠSERVICES": 25633, + "ĠSES": 91750, + "ĠSESSION": 65342, + "ĠSET": 9019, + "ĠSETTINGS": 73991, + "ĠSETUP": 90857, + "ĠSEX": 73487, + "ĠSEXP": 86040, + "ĠSF": 23463, + "ĠSG": 29626, + "ĠSGD": 93983, + "ĠSH": 6434, + "ĠSHA": 21721, + "ĠSHALL": 16681, + "ĠSHARE": 53670, + "ĠSHARES": 90794, + "ĠSHE": 53595, + "ĠSHIFT": 82503, + "ĠSHIPPING": 87318, + "ĠSHOP": 75795, + "ĠSHORT": 64924, + "ĠSHOULD": 60875, + "ĠSHOW": 38481, + "ĠSHR": 77048, + "ĠSI": 30548, + "ĠSID": 56021, + "ĠSIDE": 79324, + "ĠSIG": 25038, + "ĠSIGN": 43172, + "ĠSIGNAL": 24502, + "ĠSIL": 59723, + "ĠSIM": 22889, + "ĠSIMD": 66631, + "ĠSIMPLE": 91484, + "ĠSIN": 70002, + "ĠSINGLE": 66759, + "ĠSIP": 65441, + "ĠSITE": 54671, + "ĠSIX": 97638, + "ĠSIZE": 25341, + "ĠSJ": 68223, + "ĠSK": 12072, + "ĠSKF": 70207, + "ĠSKIP": 81404, + "ĠSKU": 77986, + "ĠSL": 16797, + "ĠSLOT": 25485, + "ĠSM": 13716, + "ĠSMA": 95896, + "ĠSMALL": 49197, + "ĠSMART": 79608, + "ĠSMB": 63658, + "ĠSME": 78767, + "ĠSMP": 92800, + "ĠSMS": 29909, + "ĠSMTP": 51819, + "ĠSN": 17924, + "ĠSNAP": 92533, + "ĠSND": 83536, + "ĠSNMP": 88133, + "ĠSNP": 59318, + "ĠSO": 5627, + "ĠSOAP": 63232, + "ĠSOC": 37650, + "ĠSOCIAL": 97080, + "ĠSOCK": 34551, + "ĠSOCKET": 68978, + "ĠSOFTWARE": 8562, + "ĠSOL": 35923, + "ĠSOLD": 83151, + "ĠSOLUTION": 97897, + "ĠSOM": 73290, + "ĠSOME": 65555, + "ĠSON": 76412, + "ĠSOP": 86507, + "ĠSORT": 57376, + "ĠSOS": 81203, + "ĠSOUND": 84273, + "ĠSOUR": 48763, + "ĠSOURCE": 53331, + "ĠSOUTH": 79615, + "ĠSP": 9256, + "ĠSPA": 76408, + "ĠSPACE": 52060, + "ĠSPD": 62724, + "ĠSPDX": 35486, + "ĠSPE": 67735, + "ĠSPEC": 47104, + "ĠSPECIAL": 23027, + "ĠSPEED": 73843, + "ĠSPELL": 47844, + "ĠSPF": 86748, + "ĠSPI": 20429, + "ĠSPL": 64700, + "ĠSPORT": 72495, + "ĠSPR": 51267, + "ĠSQ": 51618, + "ĠSQL": 7870, + "ĠSQLAlchemy": 84729, + "ĠSQLException": 24393, + "ĠSQLITE": 55252, + "ĠSQLite": 28334, + "ĠSQLiteDatabase": 62436, + "ĠSQUARE": 92282, + "ĠSR": 20880, + "ĠSRC": 49510, + "ĠSS": 18180, + "ĠSSA": 93853, + "ĠSSC": 97365, + "ĠSSD": 36362, + "ĠSSE": 73473, + "ĠSSH": 40463, + "ĠSSL": 25316, + "ĠSSP": 84989, + "ĠSSR": 95304, + "ĠSST": 91416, + "ĠST": 3928, + "ĠSTA": 88672, + "ĠSTACK": 69967, + "ĠSTANDARD": 76357, + "ĠSTAR": 56745, + "ĠSTART": 20998, + "ĠSTAT": 24981, + "ĠSTATE": 22425, + "ĠSTATES": 72900, + "ĠSTATIC": 63549, + "ĠSTATS": 75982, + "ĠSTATUS": 24014, + "ĠSTD": 29859, + "ĠSTDCALL": 69500, + "ĠSTDERR": 68637, + "ĠSTDMETHOD": 39765, + "ĠSTDMETHODCALLTYPE": 50125, + "ĠSTDOUT": 87907, + "ĠSTE": 26499, + "ĠSTEM": 63082, + "ĠSTEP": 48356, + "ĠSTILL": 92250, + "ĠSTL": 69849, + "ĠSTM": 47173, + "ĠSTOCK": 82651, + "ĠSTOP": 45537, + "ĠSTORAGE": 90484, + "ĠSTORE": 53771, + "ĠSTORY": 82840, + "ĠSTR": 12152, + "ĠSTREAM": 69811, + "ĠSTREET": 84561, + "ĠSTRICT": 29010, + "ĠSTRING": 35255, + "ĠSTRUCT": 59578, + "ĠSTUD": 57516, + "ĠSTYLE": 62947, + "ĠSU": 15490, + "ĠSUB": 16140, + "ĠSUBJECT": 95880, + "ĠSUBSTITUTE": 30530, + "ĠSUCCESS": 33941, + "ĠSUCH": 26094, + "ĠSUM": 30735, + "ĠSUMMARY": 95785, + "ĠSUN": 56228, + "ĠSUP": 53622, + "ĠSUPER": 54802, + "ĠSUPPORT": 50736, + "ĠSUR": 51983, + "ĠSUS": 89029, + "ĠSUV": 38673, + "ĠSV": 17481, + "ĠSVC": 88212, + "ĠSVG": 39800, + "ĠSVM": 90009, + "ĠSVN": 75714, + "ĠSW": 13387, + "ĠSWAT": 97054, + "ĠSWITCH": 72543, + "ĠSWT": 46700, + "ĠSX": 54768, + "ĠSY": 16079, + "ĠSYMBOL": 74538, + "ĠSYN": 61694, + "ĠSYNC": 86677, + "ĠSYS": 32560, + "ĠSYSTEM": 34752, + "ĠSZ": 69244, + "ĠSa": 15854, + "ĠSab": 20603, + "ĠSabb": 69246, + "ĠSabbath": 80176, + "ĠSabha": 86464, + "ĠSac": 23047, + "ĠSach": 37337, + "ĠSachs": 59927, + "ĠSacr": 97839, + "ĠSacramento": 40234, + "ĠSacred": 60018, + "ĠSad": 30681, + "ĠSadd": 57683, + "ĠSaddam": 66347, + "ĠSadly": 56143, + "ĠSaf": 12582, + "ĠSafari": 28761, + "ĠSafe": 22291, + "ĠSafety": 18702, + "ĠSag": 54261, + "ĠSaga": 61712, + "ĠSage": 53284, + "ĠSah": 41959, + "ĠSahara": 87978, + "ĠSai": 70741, + "ĠSaid": 50124, + "ĠSail": 56995, + "ĠSailor": 97902, + "ĠSaint": 14205, + "ĠSaints": 32219, + "ĠSak": 38767, + "ĠSakura": 78307, + "ĠSal": 8211, + "ĠSala": 79048, + "ĠSalad": 63546, + "ĠSalah": 87385, + "ĠSalary": 41758, + "ĠSale": 13315, + "ĠSalem": 59381, + "ĠSales": 15829, + "ĠSalesforce": 80935, + "ĠSalisbury": 91955, + "ĠSally": 52291, + "ĠSalman": 68643, + "ĠSalmon": 61726, + "ĠSalon": 55173, + "ĠSalt": 27040, + "ĠSalv": 30966, + "ĠSalvador": 48359, + "ĠSalvation": 85685, + "ĠSam": 8224, + "ĠSamantha": 62808, + "ĠSamar": 85256, + "ĠSame": 25739, + "ĠSammy": 90029, + "ĠSamoa": 87020, + "ĠSamp": 90174, + "ĠSample": 19143, + "ĠSampler": 98604, + "ĠSamples": 58350, + "ĠSampling": 95309, + "ĠSamsung": 18400, + "ĠSamuel": 31907, + "ĠSamurai": 88018, + "ĠSan": 5836, + "ĠSanchez": 47154, + "ĠSanct": 58002, + "ĠSanctuary": 74233, + "ĠSand": 8677, + "ĠSandbox": 96860, + "ĠSanders": 16859, + "ĠSandra": 55686, + "ĠSands": 78170, + "ĠSandwich": 66736, + "ĠSandy": 38385, + "ĠSanford": 94041, + "ĠSang": 50922, + "ĠSanity": 96302, + "ĠSans": 14747, + "ĠSant": 22896, + "ĠSanta": 15993, + "ĠSantana": 91695, + "ĠSantiago": 54578, + "ĠSanto": 76637, + "ĠSantos": 47623, + "ĠSao": 89872, + "ĠSap": 80175, + "ĠSapphire": 81999, + "ĠSar": 13637, + "ĠSara": 42920, + "ĠSarah": 20445, + "ĠSaras": 94814, + "ĠSark": 78645, + "ĠSas": 61089, + "ĠSasha": 82178, + "ĠSask": 48646, + "ĠSaskatchewan": 65153, + "ĠSass": 84862, + "ĠSat": 13177, + "ĠSatan": 43351, + "ĠSatellite": 60557, + "ĠSatisfaction": 93857, + "ĠSatoshi": 92152, + "ĠSaturday": 7728, + "ĠSaturdays": 75214, + "ĠSaturn": 49153, + "ĠSau": 41016, + "ĠSauce": 60176, + "ĠSaud": 63568, + "ĠSaudi": 17904, + "ĠSaudis": 87835, + "ĠSaul": 67524, + "ĠSaunders": 77217, + "ĠSav": 20079, + "ĠSavage": 52936, + "ĠSavannah": 72076, + "ĠSave": 10255, + "ĠSaved": 49162, + "ĠSaves": 65567, + "ĠSaving": 48438, + "ĠSavings": 57679, + "ĠSavior": 75946, + "ĠSaw": 37007, + "ĠSawyer": 91119, + "ĠSax": 66927, + "ĠSay": 24917, + "ĠSaying": 95998, + "ĠSays": 46459, + "ĠSaúde": 136360, + "ĠSaÄŁ": 136373, + "ĠSaÄŁlık": 136374, + "ĠSc": 2463, + "ĠScaffold": 34267, + "ĠScal": 68834, + "ĠScala": 65250, + "ĠScalar": 35176, + "ĠScalars": 80537, + "ĠScale": 24613, + "ĠScalia": 93649, + "ĠScaling": 88001, + "ĠScan": 31108, + "ĠScandin": 59180, + "ĠScandinavian": 84574, + "ĠScanner": 17170, + "ĠScar": 29660, + "ĠScarborough": 85432, + "ĠScarlet": 80718, + "ĠScarlett": 95472, + "ĠScatter": 94359, + "ĠScenario": 58663, + "ĠScene": 17494, + "ĠSceneManager": 53592, + "ĠScenes": 83183, + "ĠSch": 5016, + "ĠScha": 91718, + "ĠSche": 53672, + "ĠSchedule": 23512, + "ĠScheduled": 77096, + "ĠScheduler": 44759, + "ĠSchema": 12539, + "ĠScheme": 43781, + "ĠSchiff": 87583, + "ĠSchl": 49279, + "ĠSchmidt": 51065, + "ĠSchn": 60074, + "ĠSchneider": 53787, + "ĠScho": 56184, + "ĠScholar": 24525, + "ĠScholars": 98294, + "ĠScholarship": 57884, + "ĠSchool": 6022, + "ĠSchools": 30383, + "ĠSchro": 94582, + "ĠSchul": 52990, + "ĠSchultz": 66805, + "ĠSchumer": 64671, + "ĠSchw": 29505, + "ĠSchwar": 65764, + "ĠSchwartz": 64785, + "ĠSchwarz": 86028, + "ĠSchwe": 55734, + "ĠSchön": 139349, + "ĠSchüler": 137845, + "ĠSci": 40372, + "ĠScience": 9965, + "ĠSciences": 22393, + "ĠScient": 20649, + "ĠScientific": 37030, + "ĠScientist": 67309, + "ĠScientists": 56016, + "ĠScientology": 69985, + "ĠSco": 49059, + "ĠScope": 34920, + "ĠScoped": 98687, + "ĠScor": 79335, + "ĠScore": 18115, + "ĠScores": 69994, + "ĠScot": 15468, + "ĠScotch": 91950, + "ĠScotia": 62866, + "ĠScotland": 19085, + "ĠScots": 74267, + "ĠScott": 9815, + "ĠScottish": 26096, + "ĠScout": 47535, + "ĠScouts": 69307, + "ĠScr": 32134, + "ĠScrap": 79085, + "ĠScratch": 79974, + "ĠScre": 81698, + "ĠScreen": 13948, + "ĠScreening": 69089, + "ĠScreens": 68681, + "ĠScrew": 72175, + "ĠScri": 75680, + "ĠScript": 13710, + "ĠScripts": 63621, + "ĠScripture": 53150, + "ĠScriptures": 89196, + "ĠScroll": 22392, + "ĠScrollView": 68540, + "ĠScrolls": 78229, + "ĠSculpt": 82697, + "ĠSe": 1345, + "ĠSea": 15029, + "ĠSeah": 42169, + "ĠSeahawks": 44926, + "ĠSeal": 52525, + "ĠSeam": 77966, + "ĠSeamless": 96962, + "ĠSean": 24995, + "ĠSearch": 7542, + "ĠSearchResult": 96597, + "ĠSearches": 81169, + "ĠSearching": 79893, + "ĠSears": 73234, + "ĠSeas": 92396, + "ĠSeason": 14263, + "ĠSeasons": 67200, + "ĠSeat": 39223, + "ĠSeats": 74435, + "ĠSeattle": 16355, + "ĠSebast": 93713, + "ĠSebastian": 43509, + "ĠSec": 4520, + "ĠSecond": 10440, + "ĠSecondary": 43534, + "ĠSecondly": 71031, + "ĠSeconds": 60983, + "ĠSecret": 8599, + "ĠSecretary": 12386, + "ĠSecrets": 52079, + "ĠSect": 94766, + "ĠSection": 11113, + "ĠSections": 59037, + "ĠSector": 40724, + "ĠSecure": 34502, + "ĠSecurities": 45172, + "ĠSecurity": 8234, + "ĠSed": 35278, + "ĠSee": 3496, + "ĠSeed": 35822, + "ĠSeeder": 70738, + "ĠSeeds": 72989, + "ĠSeeing": 55024, + "ĠSeek": 29739, + "ĠSeeking": 58705, + "ĠSeems": 63957, + "ĠSeen": 73512, + "ĠSeg": 17209, + "ĠSega": 79849, + "ĠSegment": 37103, + "ĠSegún": 139015, + "ĠSeit": 96520, + "ĠSeite": 47404, + "ĠSeiten": 82276, + "ĠSek": 49241, + "ĠSeks": 76394, + "ĠSel": 23204, + "ĠSelbst": 91452, + "ĠSele": 94093, + "ĠSelect": 8427, + "ĠSelectList": 52902, + "ĠSelectListItem": 95883, + "ĠSelected": 29302, + "ĠSelection": 24145, + "ĠSelector": 49610, + "ĠSelenium": 85717, + "ĠSelf": 10115, + "ĠSell": 42063, + "ĠSeller": 48232, + "ĠSellers": 81088, + "ĠSelling": 53720, + "ĠSem": 14248, + "ĠSemantic": 74333, + "ĠSemaphore": 87920, + "ĠSemester": 86848, + "ĠSemi": 54638, + "ĠSemiconductor": 78434, + "ĠSeminar": 82182, + "ĠSen": 5363, + "ĠSenate": 9891, + "ĠSenator": 22276, + "ĠSenators": 53718, + "ĠSend": 11000, + "ĠSendMessage": 62292, + "ĠSender": 54610, + "ĠSending": 57646, + "ĠSends": 64402, + "ĠSenior": 19342, + "ĠSens": 31152, + "ĠSense": 46229, + "ĠSensor": 30045, + "ĠSensors": 94420, + "ĠSent": 23361, + "ĠSentence": 79542, + "ĠSentinel": 56793, + "ĠSentry": 94060, + "ĠSeoul": 50189, + "ĠSep": 17449, + "ĠSepar": 33099, + "ĠSeparate": 76140, + "ĠSeparator": 78289, + "ĠSept": 5375, + "ĠSeptember": 6122, + "ĠSeq": 23744, + "ĠSequ": 24811, + "ĠSequelize": 32040, + "ĠSequence": 28871, + "ĠSequential": 59009, + "ĠSer": 8245, + "ĠSerbia": 58714, + "ĠSerbian": 87164, + "ĠSerena": 90990, + "ĠSerg": 54247, + "ĠSerge": 31922, + "ĠSergeant": 64544, + "ĠSergei": 83998, + "ĠSergey": 73429, + "ĠSergio": 65194, + "ĠSerial": 11215, + "ĠSerializable": 24859, + "ĠSerialization": 72034, + "ĠSerialize": 39900, + "ĠSerialized": 91472, + "ĠSerializer": 58822, + "ĠSerie": 56740, + "ĠSeries": 11131, + "ĠSerif": 20003, + "ĠSerious": 73496, + "ĠSeriously": 56506, + "ĠSerum": 93491, + "ĠServ": 26224, + "ĠServe": 52932, + "ĠServer": 8422, + "ĠServers": 71797, + "ĠService": 5362, + "ĠServiceException": 72107, + "ĠServiceProvider": 87617, + "ĠServices": 8307, + "ĠServing": 71643, + "ĠServlet": 23124, + "ĠServletException": 30091, + "ĠSes": 62927, + "ĠSession": 9164, + "ĠSessions": 30800, + "ĠSesso": 53547, + "ĠSet": 2573, + "ĠSetLastError": 99002, + "ĠSetProperty": 87046, + "ĠSetUp": 92101, + "ĠSetValue": 61293, + "ĠSeth": 44218, + "ĠSets": 12525, + "ĠSetter": 69264, + "ĠSetting": 20037, + "ĠSettings": 11296, + "ĠSettlement": 69048, + "ĠSetup": 18626, + "ĠSeven": 29948, + "ĠSevent": 77218, + "ĠSeventh": 74625, + "ĠSever": 58216, + "ĠSeveral": 25694, + "ĠSeverity": 51139, + "ĠSevilla": 94894, + "ĠSew": 49213, + "ĠSex": 6695, + "ĠSexe": 45492, + "ĠSexo": 56814, + "ĠSext": 66949, + "ĠSexual": 38667, + "ĠSexy": 28279, + "ĠSey": 57880, + "ĠSeymour": 93511, + "ĠSgt": 55660, + "ĠSh": 1417, + "ĠSha": 27970, + "ĠShack": 85423, + "ĠShade": 79818, + "ĠShader": 34788, + "ĠShades": 80147, + "ĠShadow": 24298, + "ĠShadows": 66449, + "ĠShaft": 95759, + "ĠShah": 36517, + "ĠShak": 86565, + "ĠShake": 72323, + "ĠShakespeare": 41382, + "ĠShall": 85461, + "ĠSham": 72091, + "ĠShaman": 91961, + "ĠShame": 87535, + "ĠShan": 59767, + "ĠShane": 49975, + "ĠShanghai": 37047, + "ĠShank": 82810, + "ĠShannon": 53663, + "ĠShape": 22526, + "ĠShapes": 61531, + "ĠShapiro": 79731, + "ĠShar": 35116, + "ĠShard": 95366, + "ĠShare": 11775, + "ĠSharePoint": 76146, + "ĠShared": 16990, + "ĠSharedModule": 83639, + "ĠSharedPreferences": 35084, + "ĠShares": 44185, + "ĠSharia": 89995, + "ĠSharing": 56035, + "ĠShark": 60436, + "ĠSharks": 70823, + "ĠSharma": 60015, + "ĠSharon": 51852, + "ĠSharp": 39261, + "ĠShaun": 76271, + "ĠShaw": 35185, + "ĠShawn": 55537, + "ĠShay": 73884, + "ĠShe": 2932, + "ĠShea": 84968, + "ĠShed": 94452, + "ĠSheep": 82908, + "ĠSheet": 27741, + "ĠSheets": 60472, + "ĠSheffield": 60025, + "ĠSheikh": 71529, + "ĠSheila": 94075, + "ĠShel": 36840, + "ĠShelby": 69408, + "ĠSheldon": 77610, + "ĠShelf": 70346, + "ĠShell": 29402, + "ĠShelley": 90509, + "ĠShelter": 74821, + "ĠShemale": 73447, + "ĠShen": 57780, + "ĠShepard": 71819, + "ĠShepherd": 58546, + "ĠSher": 17092, + "ĠSheridan": 94123, + "ĠSheriff": 28683, + "ĠSherlock": 60626, + "ĠSherman": 51883, + "ĠShi": 40704, + "ĠShia": 77409, + "ĠShib": 95337, + "ĠShield": 29743, + "ĠShields": 74797, + "ĠShift": 26510, + "ĠShiite": 93506, + "ĠShim": 53485, + "ĠShin": 34449, + "ĠShine": 86155, + "ĠShip": 26803, + "ĠShipping": 23954, + "ĠShips": 73532, + "ĠShir": 49256, + "ĠShirley": 75237, + "ĠShirt": 54707, + "ĠShiv": 89813, + "ĠSho": 62940, + "ĠShock": 46983, + "ĠShoe": 67081, + "ĠShoes": 38629, + "ĠShoot": 49266, + "ĠShooter": 98307, + "ĠShooting": 60781, + "ĠShop": 14028, + "ĠShopify": 95025, + "ĠShopping": 28964, + "ĠShoppingCart": 90907, + "ĠShops": 74544, + "ĠShore": 44719, + "ĠShort": 10698, + "ĠShortcut": 95465, + "ĠShortly": 66487, + "ĠShorts": 90682, + "ĠShot": 32347, + "ĠShotgun": 97107, + "ĠShots": 91032, + "ĠShould": 12260, + "ĠShoulder": 79772, + "ĠShow": 6928, + "ĠShowcase": 87063, + "ĠShower": 47371, + "ĠShowing": 70930, + "ĠShows": 36280, + "ĠShr": 94738, + "ĠShrine": 93474, + "ĠShuffle": 88925, + "ĠShut": 48274, + "ĠShutdown": 65379, + "ĠShutterstock": 84042, + "ĠShuttle": 77687, + "ĠSi": 11832, + "ĠSiber": 67500, + "ĠSic": 55854, + "ĠSicher": 74303, + "ĠSick": 81311, + "ĠSid": 33889, + "ĠSidd": 84713, + "ĠSide": 16659, + "ĠSidebar": 38700, + "ĠSidney": 75896, + "ĠSie": 8495, + "ĠSiege": 86990, + "ĠSiemens": 90224, + "ĠSierra": 35803, + "ĠSig": 41560, + "ĠSight": 65337, + "ĠSigma": 49537, + "ĠSign": 7075, + "ĠSignIn": 55189, + "ĠSignUp": 57077, + "ĠSignal": 27230, + "ĠSignals": 82499, + "ĠSignature": 32232, + "ĠSigned": 52453, + "ĠSignificant": 89362, + "ĠSigning": 87424, + "ĠSigns": 58024, + "ĠSignup": 54726, + "ĠSik": 77349, + "ĠSikh": 82562, + "ĠSil": 8051, + "ĠSilence": 68088, + "ĠSilent": 59455, + "ĠSilicon": 37150, + "ĠSilicone": 98800, + "ĠSilk": 51036, + "ĠSilva": 41041, + "ĠSilver": 14997, + "ĠSim": 4467, + "ĠSimilar": 21476, + "ĠSimilarly": 34239, + "ĠSimmons": 50096, + "ĠSimon": 21315, + "ĠSimone": 79432, + "ĠSimpl": 61242, + "ĠSimple": 8993, + "ĠSimpleDateFormat": 26180, + "ĠSimpleName": 40799, + "ĠSimply": 28424, + "ĠSimpson": 34666, + "ĠSimpsons": 84633, + "ĠSims": 51623, + "ĠSimulation": 43120, + "ĠSimulator": 52238, + "ĠSin": 17632, + "ĠSinai": 78885, + "ĠSinatra": 92739, + "ĠSince": 8704, + "ĠSinclair": 73189, + "ĠSind": 96681, + "ĠSing": 10833, + "ĠSingapore": 20539, + "ĠSinger": 54670, + "ĠSingh": 26894, + "ĠSingle": 11327, + "ĠSingleChildScrollView": 88554, + "ĠSingles": 46322, + "ĠSingleton": 38685, + "ĠSingular": 94930, + "ĠSink": 56451, + "ĠSinn": 84579, + "ĠSioux": 81593, + "ĠSir": 16759, + "ĠSiri": 74974, + "ĠSirius": 60034, + "ĠSistema": 76254, + "ĠSister": 47268, + "ĠSisters": 60988, + "ĠSit": 28408, + "ĠSite": 12911, + "ĠSites": 31938, + "ĠSith": 71492, + "ĠSitting": 83754, + "ĠSituation": 68378, + "ĠSix": 18680, + "ĠSixth": 66977, + "ĠSize": 8478, + "ĠSized": 21967, + "ĠSizedBox": 22905, + "ĠSizes": 70182, + "ĠSk": 4818, + "ĠSkate": 96696, + "ĠSke": 62181, + "ĠSkeleton": 65933, + "ĠSketch": 38401, + "ĠSki": 60495, + "ĠSkill": 27482, + "ĠSkills": 30240, + "ĠSkin": 26951, + "ĠSkinner": 97042, + "ĠSkinny": 94224, + "ĠSkip": 25784, + "ĠSkipping": 96118, + "ĠSkull": 60791, + "ĠSky": 14722, + "ĠSkyl": 93011, + "ĠSkype": 41209, + "ĠSkyrim": 64929, + "ĠSkywalker": 76662, + "ĠSl": 6852, + "ĠSlack": 57244, + "ĠSlam": 73623, + "ĠSlash": 72057, + "ĠSlate": 68319, + "ĠSlater": 97003, + "ĠSlave": 59368, + "ĠSlayer": 74623, + "ĠSle": 88458, + "ĠSlee": 46463, + "ĠSleep": 23774, + "ĠSleeping": 74147, + "ĠSleeve": 60262, + "ĠSlice": 56476, + "ĠSlide": 43364, + "ĠSlider": 38129, + "ĠSlim": 44391, + "ĠSlip": 58532, + "ĠSlo": 25603, + "ĠSloan": 93093, + "ĠSlot": 31316, + "ĠSlots": 52348, + "ĠSlovak": 61264, + "ĠSlovakia": 80685, + "ĠSloven": 59395, + "ĠSlovenia": 78217, + "ĠSlow": 38147, + "ĠSlug": 93613, + "ĠSlut": 96344, + "ĠSly": 90736, + "ĠSm": 4388, + "ĠSmall": 14994, + "ĠSmart": 15770, + "ĠSmartphone": 73937, + "ĠSmarty": 50314, + "ĠSmash": 47294, + "ĠSmile": 78305, + "ĠSmith": 9082, + "ĠSmithsonian": 88189, + "ĠSmoke": 53204, + "ĠSmoking": 72599, + "ĠSmooth": 50725, + "ĠSms": 95389, + "ĠSmy": 93402, + "ĠSn": 13059, + "ĠSnackbar": 56254, + "ĠSnake": 39656, + "ĠSnap": 28232, + "ĠSnapchat": 55179, + "ĠSnapdragon": 82055, + "ĠSnape": 85684, + "ĠSnapshot": 68697, + "ĠSne": 50013, + "ĠSniper": 75054, + "ĠSno": 92671, + "ĠSnow": 18901, + "ĠSnowden": 46360, + "ĠSnyder": 53131, + "ĠSo": 2055, + "ĠSoap": 60631, + "ĠSob": 66437, + "ĠSoc": 26572, + "ĠSoccer": 37533, + "ĠSoci": 41409, + "ĠSocial": 9784, + "ĠSocialist": 56110, + "ĠSociety": 13278, + "ĠSociology": 96057, + "ĠSocket": 20954, + "ĠSoda": 96072, + "ĠSodium": 82441, + "ĠSof": 35547, + "ĠSofa": 60848, + "ĠSofia": 78563, + "ĠSoft": 24079, + "ĠSoftware": 4377, + "ĠSoil": 75519, + "ĠSok": 84114, + "ĠSol": 11473, + "ĠSolar": 24446, + "ĠSold": 27140, + "ĠSoldier": 52429, + "ĠSoldiers": 74431, + "ĠSole": 64515, + "ĠSolic": 76194, + "ĠSolid": 22138, + "ĠSolidColorBrush": 91251, + "ĠSolo": 35123, + "ĠSolomon": 49799, + "ĠSolution": 12478, + "ĠSolutions": 22676, + "ĠSolve": 63284, + "ĠSolver": 68295, + "ĠSom": 17564, + "ĠSomali": 84694, + "ĠSomalia": 59751, + "ĠSome": 4329, + "ĠSomebody": 93092, + "ĠSomehow": 79047, + "ĠSomeone": 34172, + "ĠSomerset": 71132, + "ĠSomething": 24656, + "ĠSometimes": 17688, + "ĠSommer": 79509, + "ĠSon": 11840, + "ĠSonata": 94351, + "ĠSong": 18971, + "ĠSongs": 39100, + "ĠSonia": 89391, + "ĠSonic": 43217, + "ĠSons": 50738, + "ĠSony": 20864, + "ĠSoon": 31762, + "ĠSoph": 33840, + "ĠSophia": 70371, + "ĠSophie": 59310, + "ĠSor": 29531, + "ĠSoros": 70187, + "ĠSorry": 32286, + "ĠSort": 15967, + "ĠSorted": 52688, + "ĠSorting": 77143, + "ĠSou": 9240, + "ĠSoul": 29142, + "ĠSouls": 57670, + "ĠSound": 14594, + "ĠSounds": 44141, + "ĠSoup": 60334, + "ĠSour": 94488, + "ĠSource": 8748, + "ĠSources": 47032, + "ĠSous": 88478, + "ĠSouth": 4882, + "ĠSouthampton": 61151, + "ĠSoutheast": 35564, + "ĠSouthern": 16244, + "ĠSouthwest": 45685, + "ĠSov": 17867, + "ĠSovere": 88221, + "ĠSoviet": 19390, + "ĠSoviets": 84289, + "ĠSox": 38545, + "ĠSoy": 62003, + "ĠSp": 3089, + "ĠSpa": 34814, + "ĠSpace": 11487, + "ĠSpaceX": 61924, + "ĠSpacer": 70173, + "ĠSpaces": 68056, + "ĠSpain": 17689, + "ĠSpam": 81667, + "ĠSpan": 11903, + "ĠSpanish": 15154, + "ĠSpar": 69806, + "ĠSpare": 95205, + "ĠSpark": 26467, + "ĠSparks": 93794, + "ĠSparse": 71794, + "ĠSpart": 48351, + "ĠSpartan": 76186, + "ĠSpatial": 74697, + "ĠSpawn": 47361, + "ĠSpaÃŁ": 140748, + "ĠSpe": 12307, + "ĠSpeak": 67201, + "ĠSpeaker": 29073, + "ĠSpeakers": 80096, + "ĠSpeaking": 43972, + "ĠSpear": 77437, + "ĠSpears": 93955, + "ĠSpec": 10956, + "ĠSpecial": 9785, + "ĠSpecialist": 39320, + "ĠSpecialists": 91694, + "ĠSpecialty": 81514, + "ĠSpecies": 50467, + "ĠSpecific": 28262, + "ĠSpecifically": 44763, + "ĠSpecification": 51277, + "ĠSpecifications": 62588, + "ĠSpecifies": 44980, + "ĠSpecify": 47395, + "ĠSpecs": 79892, + "ĠSpect": 26628, + "ĠSpectrum": 51839, + "ĠSpeech": 38741, + "ĠSpeed": 15181, + "ĠSpeedway": 85649, + "ĠSpell": 29080, + "ĠSpells": 78053, + "ĠSpencer": 35121, + "ĠSpend": 77037, + "ĠSpending": 86653, + "ĠSphere": 54499, + "ĠSphinx": 94193, + "ĠSpi": 19751, + "ĠSpice": 67821, + "ĠSpicer": 79666, + "ĠSpider": 28390, + "ĠSpiel": 31380, + "ĠSpielberg": 93616, + "ĠSpiele": 69462, + "ĠSpieler": 81574, + "ĠSpike": 60252, + "ĠSpin": 40685, + "ĠSpinner": 47545, + "ĠSpir": 65047, + "ĠSpiral": 88788, + "ĠSpirit": 16899, + "ĠSpirits": 81777, + "ĠSpiritual": 62749, + "ĠSpit": 81741, + "ĠSpl": 51198, + "ĠSplash": 45442, + "ĠSplashScreen": 90324, + "ĠSplit": 27810, + "ĠSpo": 55949, + "ĠSpokane": 88180, + "ĠSponge": 80743, + "ĠSponsor": 47561, + "ĠSponsored": 73747, + "ĠSpoon": 93513, + "ĠSport": 18207, + "ĠSporting": 64303, + "ĠSports": 13180, + "ĠSpot": 23297, + "ĠSpotify": 40537, + "ĠSpotlight": 80299, + "ĠSpr": 15515, + "ĠSpray": 60561, + "ĠSpread": 47716, + "ĠSpreadsheet": 98501, + "ĠSpring": 12252, + "ĠSpringApplication": 91751, + "ĠSpringer": 66815, + "ĠSpringfield": 52424, + "ĠSprings": 29204, + "ĠSprint": 44812, + "ĠSprite": 21556, + "ĠSpurs": 45042, + "ĠSpy": 46314, + "ĠSq": 64571, + "ĠSql": 7224, + "ĠSqlCommand": 21354, + "ĠSqlConnection": 27981, + "ĠSqlDataAdapter": 46843, + "ĠSqlDataReader": 65511, + "ĠSqlDbType": 27567, + "ĠSqlParameter": 21923, + "ĠSqu": 20084, + "ĠSquad": 34707, + "ĠSquadron": 70072, + "ĠSquare": 15619, + "ĠSr": 33588, + "ĠSrc": 67684, + "ĠSri": 33345, + "ĠSt": 794, + "ĠStObject": 29439, + "ĠSta": 59820, + "ĠStability": 80138, + "ĠStable": 83341, + "ĠStack": 14284, + "ĠStacy": 97325, + "ĠStadium": 22636, + "ĠStadt": 54092, + "ĠStaff": 16951, + "ĠStafford": 62474, + "ĠStage": 22112, + "ĠStainless": 32490, + "ĠStake": 98246, + "ĠStalin": 46665, + "ĠStall": 71870, + "ĠStam": 68080, + "ĠStamford": 98287, + "ĠStamina": 94272, + "ĠStamp": 54787, + "ĠStan": 15205, + "ĠStand": 15577, + "ĠStandard": 11766, + "ĠStandards": 34553, + "ĠStanding": 45227, + "ĠStanford": 30688, + "ĠStanley": 30452, + "ĠStanton": 80664, + "ĠStap": 62506, + "ĠStaples": 91398, + "ĠStar": 7679, + "ĠStarbucks": 48784, + "ĠStard": 97828, + "ĠStark": 46749, + "ĠStarr": 75022, + "ĠStars": 24651, + "ĠStart": 5145, + "ĠStartCoroutine": 45212, + "ĠStartTime": 96798, + "ĠStarted": 35812, + "ĠStarter": 59787, + "ĠStarting": 27657, + "ĠStarts": 47728, + "ĠStartup": 38414, + "ĠStat": 12166, + "ĠState": 3234, + "ĠStatefulWidget": 37642, + "ĠStateless": 35857, + "ĠStatelessWidget": 36798, + "ĠStatement": 21756, + "ĠStatements": 69716, + "ĠStaten": 95306, + "ĠStates": 4180, + "ĠStatic": 23105, + "ĠStation": 16629, + "ĠStations": 86916, + "ĠStatistical": 65694, + "ĠStatistics": 24624, + "ĠStats": 29927, + "ĠStatue": 88575, + "ĠStatus": 8104, + "ĠStatusBar": 68023, + "ĠStatusCode": 53403, + "ĠStay": 28737, + "ĠStd": 42517, + "ĠSte": 3360, + "ĠSteak": 84243, + "ĠStealth": 73467, + "ĠSteam": 21825, + "ĠSteel": 12500, + "ĠSteele": 56393, + "ĠSteelers": 50455, + "ĠSteering": 78721, + "ĠStef": 68320, + "ĠStefan": 48902, + "ĠStein": 34506, + "ĠStella": 74598, + "ĠStellar": 94542, + "ĠStem": 80848, + "ĠStep": 14822, + "ĠSteph": 40559, + "ĠStephan": 38364, + "ĠStephanie": 48143, + "ĠStephen": 18095, + "ĠStephens": 75923, + "ĠSteps": 39861, + "ĠSter": 26063, + "ĠStereo": 71975, + "ĠSterling": 43735, + "ĠStern": 50750, + "ĠSteve": 13809, + "ĠSteven": 23644, + "ĠStevens": 35077, + "ĠStevenson": 81948, + "ĠStewart": 28768, + "ĠStick": 46461, + "ĠSticky": 92643, + "ĠStill": 16377, + "ĠStim": 70772, + "ĠSting": 82870, + "ĠStir": 64192, + "ĠStitch": 67923, + "ĠStmt": 97023, + "ĠSto": 70783, + "ĠStock": 12648, + "ĠStockholm": 52082, + "ĠStocks": 79236, + "ĠStoke": 83597, + "ĠStokes": 97520, + "ĠSton": 35119, + "ĠStone": 14302, + "ĠStones": 56100, + "ĠStop": 14215, + "ĠStops": 88973, + "ĠStopwatch": 70887, + "ĠStorage": 14693, + "ĠStore": 9129, + "ĠStored": 93785, + "ĠStores": 38302, + "ĠStories": 29029, + "ĠStorm": 21862, + "ĠStory": 15106, + "ĠStout": 91030, + "ĠStr": 4509, + "ĠStra": 26647, + "ĠStraight": 45810, + "ĠStrait": 82063, + "ĠStrand": 77302, + "ĠStrange": 42593, + "ĠStranger": 72744, + "ĠStrap": 84406, + "ĠStrat": 52935, + "ĠStrateg": 28223, + "ĠStrategic": 45561, + "ĠStrategies": 55519, + "ĠStrategy": 27745, + "ĠStrauss": 93411, + "ĠStraw": 56844, + "ĠStrawberry": 87977, + "ĠStraÃŁen": 143089, + "ĠStre": 35672, + "ĠStream": 9203, + "ĠStreamLazy": 71918, + "ĠStreamReader": 46811, + "ĠStreamWriter": 57257, + "ĠStreaming": 44810, + "ĠStreams": 94522, + "ĠStreet": 6686, + "ĠStreets": 65518, + "ĠStrength": 31837, + "ĠStress": 50468, + "ĠStretch": 60105, + "ĠStrict": 52881, + "ĠStrike": 35378, + "ĠStrikes": 97026, + "ĠString": 923, + "ĠStringBuffer": 29891, + "ĠStringBuilder": 11410, + "ĠStringComparison": 47687, + "ĠStringField": 59002, + "ĠStringIO": 66935, + "ĠStringSplitOptions": 79876, + "ĠStringTokenizer": 71601, + "ĠStringType": 93635, + "ĠStringUtil": 69280, + "ĠStringUtils": 42938, + "ĠStringWriter": 84787, + "ĠStrings": 41651, + "ĠStrip": 34899, + "ĠStripe": 59566, + "ĠStroke": 68934, + "ĠStrom": 92603, + "ĠStrong": 26099, + "ĠStruct": 16139, + "ĠStructural": 72700, + "ĠStructure": 28596, + "ĠStructures": 70050, + "ĠStuart": 43532, + "ĠStub": 66611, + "ĠStud": 7659, + "ĠStudent": 11726, + "ĠStudents": 20173, + "ĠStudies": 18720, + "ĠStudio": 18563, + "ĠStudios": 30262, + "ĠStudy": 19173, + "ĠStuff": 45486, + "ĠStunden": 80544, + "ĠStunning": 92353, + "ĠStuttgart": 82139, + "ĠSty": 46565, + "ĠStyle": 11913, + "ĠStyleSheet": 24757, + "ĠStyled": 57299, + "ĠStyles": 37370, + "ĠStück": 137296, + "ĠSu": 16931, + "ĠSuarez": 93663, + "ĠSub": 3719, + "ĠSubLObject": 79057, + "ĠSubaru": 59906, + "ĠSubcommittee": 98156, + "ĠSubject": 17450, + "ĠSubjects": 64718, + "ĠSubmission": 58734, + "ĠSubmit": 29170, + "ĠSubmitted": 71889, + "ĠSubscribe": 27270, + "ĠSubscriber": 69653, + "ĠSubscription": 29617, + "ĠSubset": 73223, + "ĠSubstance": 94287, + "ĠSubstitute": 87149, + "ĠSubtract": 93210, + "ĠSubway": 85081, + "ĠSuc": 67890, + "ĠSucc": 88179, + "ĠSuccess": 13047, + "ĠSuccessful": 45526, + "ĠSuccessfully": 34591, + "ĠSuch": 15132, + "ĠSuche": 79041, + "ĠSud": 26333, + "ĠSudan": 42454, + "ĠSuddenly": 57701, + "ĠSudoku": 94254, + "ĠSue": 47649, + "ĠSuff": 59263, + "ĠSuffolk": 80783, + "ĠSug": 83767, + "ĠSugar": 30919, + "ĠSuggestions": 87615, + "ĠSuicide": 64800, + "ĠSuit": 32611, + "ĠSuitable": 85246, + "ĠSuite": 20977, + "ĠSuites": 76012, + "ĠSuk": 73774, + "ĠSul": 45862, + "ĠSullivan": 41989, + "ĠSultan": 74986, + "ĠSum": 8116, + "ĠSummary": 21517, + "ĠSummer": 18836, + "ĠSummers": 79648, + "ĠSummit": 34669, + "ĠSummon": 53906, + "ĠSun": 8059, + "ĠSund": 36802, + "ĠSunday": 7270, + "ĠSundays": 61080, + "ĠSunder": 72061, + "ĠSunderland": 82665, + "ĠSung": 59571, + "ĠSunni": 55010, + "ĠSunny": 55038, + "ĠSunrise": 87188, + "ĠSuns": 72968, + "ĠSunset": 62690, + "ĠSunshine": 61294, + "ĠSup": 6299, + "ĠSuper": 7297, + "ĠSuperintendent": 67884, + "ĠSuperior": 40941, + "ĠSuperman": 37709, + "ĠSupern": 95948, + "ĠSuperv": 61824, + "ĠSupervisor": 57107, + "ĠSupplement": 42391, + "ĠSupplementary": 98271, + "ĠSupplements": 91310, + "ĠSupplier": 31704, + "ĠSuppliers": 46989, + "ĠSupplies": 50252, + "ĠSupply": 29809, + "ĠSupport": 9186, + "ĠSupported": 48980, + "ĠSupporters": 90263, + "ĠSupporting": 73996, + "ĠSupports": 66446, + "ĠSuppose": 82610, + "ĠSuppress": 86798, + "ĠSupreme": 13506, + "ĠSur": 8082, + "ĠSure": 22555, + "ĠSurely": 64188, + "ĠSurf": 64097, + "ĠSurface": 26963, + "ĠSurg": 33276, + "ĠSurge": 56157, + "ĠSurgery": 47090, + "ĠSurgical": 70112, + "ĠSurprise": 84484, + "ĠSurre": 54066, + "ĠSurrey": 66964, + "ĠSurround": 98960, + "ĠSurv": 28778, + "ĠSurveillance": 75358, + "ĠSurvey": 23591, + "ĠSurvival": 53351, + "ĠSurvivor": 64480, + "ĠSus": 16288, + "ĠSusan": 29933, + "ĠSusp": 52206, + "ĠSuspension": 90010, + "ĠSussex": 69508, + "ĠSustainability": 88712, + "ĠSustainable": 60493, + "ĠSutton": 73013, + "ĠSuz": 34387, + "ĠSuzanne": 84152, + "ĠSuzuki": 55746, + "ĠSv": 63605, + "ĠSvens": 86173, + "ĠSvg": 81055, + "ĠSw": 4492, + "ĠSwagger": 79631, + "ĠSwal": 59073, + "ĠSwamp": 82811, + "ĠSwan": 43401, + "ĠSwansea": 77803, + "ĠSwap": 43849, + "ĠSwarm": 96470, + "ĠSwe": 36226, + "ĠSweat": 85960, + "ĠSweden": 23190, + "ĠSwedish": 30109, + "ĠSweep": 85022, + "ĠSweet": 26589, + "ĠSwift": 23670, + "ĠSwiftUI": 74881, + "ĠSwim": 68233, + "ĠSwimming": 74700, + "ĠSwing": 52086, + "ĠSwinger": 60165, + "ĠSwipe": 63800, + "ĠSwiss": 29691, + "ĠSwitch": 15586, + "ĠSwitzerland": 29121, + "ĠSword": 35651, + "ĠSwords": 88351, + "ĠSy": 5718, + "ĠSydney": 21273, + "ĠSylv": 46037, + "ĠSylvia": 88306, + "ĠSym": 11375, + "ĠSymbol": 19612, + "ĠSymbols": 83580, + "ĠSymfony": 16015, + "ĠSymphony": 62206, + "ĠSymposium": 73838, + "ĠSymptoms": 76317, + "ĠSyn": 23153, + "ĠSync": 28937, + "ĠSynd": 41219, + "ĠSyndrome": 56396, + "ĠSynopsis": 87998, + "ĠSyntax": 32117, + "ĠSynthetic": 74828, + "ĠSyracuse": 61596, + "ĠSyria": 12623, + "ĠSyrian": 18141, + "ĠSyrians": 78956, + "ĠSys": 28909, + "ĠSystem": 739, + "ĠSystems": 14917, + "ĠSz": 44190, + "ĠSão": 47690, + "ĠSé": 72387, + "ĠSöz": 135191, + "ĠSü": 77549, + "ĠSüd": 137256, + "ĠSÃłi": 136313, + "ĠSÅĤow": 143407, + "ĠSÆ¡n": 131279, + "ĠSản": 134360, + "ĠSá»±": 133357, + "ĠSá»ij": 134175, + "ĠSợ": 133827, + "ĠSá»Ń": 136091, + "ĠT": 350, + "ĠTA": 38891, + "ĠTAB": 49176, + "ĠTABLE": 14363, + "ĠTAG": 21494, + "ĠTAKE": 85139, + "ĠTAM": 91620, + "ĠTAR": 74574, + "ĠTARGET": 41033, + "ĠTAS": 91288, + "ĠTASK": 42979, + "ĠTAX": 90773, + "ĠTArray": 95273, + "ĠTB": 30080, + "ĠTBD": 72264, + "ĠTBranch": 92126, + "ĠTC": 24591, + "ĠTCHAR": 67069, + "ĠTCL": 65231, + "ĠTCP": 26656, + "ĠTD": 27716, + "ĠTE": 18289, + "ĠTEAM": 58103, + "ĠTECH": 53428, + "ĠTECHNO": 87534, + "ĠTED": 83196, + "ĠTELE": 89240, + "ĠTEM": 75670, + "ĠTEMP": 57431, + "ĠTEMPLATE": 67387, + "ĠTEN": 74266, + "ĠTER": 48848, + "ĠTERM": 88552, + "ĠTERMIN": 93760, + "ĠTERMS": 98928, + "ĠTEST": 13602, + "ĠTESTING": 95535, + "ĠTEX": 76224, + "ĠTEXT": 15762, + "ĠTEntity": 90800, + "ĠTF": 29145, + "ĠTFT": 98601, + "ĠTForm": 93617, + "ĠTG": 55872, + "ĠTH": 4434, + "ĠTHAN": 90006, + "ĠTHANK": 80857, + "ĠTHAT": 25269, + "ĠTHC": 48490, + "ĠTHE": 3168, + "ĠTHEIR": 83604, + "ĠTHEM": 73703, + "ĠTHEME": 86655, + "ĠTHEN": 24117, + "ĠTHEORY": 29657, + "ĠTHERE": 61107, + "ĠTHESE": 82335, + "ĠTHEY": 62493, + "ĠTHINK": 92119, + "ĠTHIRD": 68157, + "ĠTHIS": 10039, + "ĠTHR": 90586, + "ĠTHREAD": 56515, + "ĠTHREE": 14040, + "ĠTHROUGH": 93105, + "ĠTHROW": 77534, + "ĠTI": 38137, + "ĠTICK": 75711, + "ĠTIFF": 74121, + "ĠTILE": 52849, + "ĠTIM": 17742, + "ĠTIME": 22236, + "ĠTIMEOUT": 90846, + "ĠTIMER": 49931, + "ĠTIMES": 86930, + "ĠTIMESTAMP": 92592, + "ĠTITLE": 48573, + "ĠTInt": 71238, + "ĠTJ": 72957, + "ĠTK": 44299, + "ĠTKey": 88750, + "ĠTL": 29615, + "ĠTLC": 90482, + "ĠTLS": 41654, + "ĠTLabel": 80997, + "ĠTM": 23975, + "ĠTMP": 66253, + "ĠTMPro": 87926, + "ĠTMZ": 98955, + "ĠTN": 30923, + "ĠTNT": 88131, + "ĠTO": 5146, + "ĠTOD": 34174, + "ĠTODAY": 40262, + "ĠTODO": 5343, + "ĠTOK": 79501, + "ĠTOKEN": 43574, + "ĠTOM": 82465, + "ĠTOO": 92576, + "ĠTOOL": 92813, + "ĠTOP": 25012, + "ĠTOR": 69961, + "ĠTORT": 17791, + "ĠTOTAL": 46962, + "ĠTOUCH": 88779, + "ĠTOUR": 80301, + "ĠTObject": 31711, + "ĠTP": 29070, + "ĠTPM": 63644, + "ĠTPP": 69011, + "ĠTR": 4984, + "ĠTRACE": 37519, + "ĠTRACK": 73502, + "ĠTRADE": 96861, + "ĠTRAIN": 67509, + "ĠTRAN": 40347, + "ĠTRANS": 47533, + "ĠTRANSACTION": 93923, + "ĠTRE": 76886, + "ĠTREE": 71761, + "ĠTRI": 54563, + "ĠTRUE": 8214, + "ĠTRY": 78481, + "ĠTResult": 46940, + "ĠTS": 22965, + "ĠTSA": 78904, + "ĠTSR": 74671, + "ĠTSRMLS": 86966, + "ĠTString": 52226, + "ĠTT": 32744, + "ĠTTC": 87800, + "ĠTTL": 78532, + "ĠTU": 69414, + "ĠTURN": 82369, + "ĠTV": 5883, + "ĠTValue": 56374, + "ĠTVs": 59766, + "ĠTW": 35555, + "ĠTWO": 46258, + "ĠTX": 17031, + "ĠTXT": 72429, + "ĠTY": 48862, + "ĠTYPE": 12454, + "ĠTYPES": 82325, + "ĠTYPO": 77274, + "ĠTZ": 87105, + "ĠTa": 23291, + "ĠTab": 15139, + "ĠTable": 6633, + "ĠTableCell": 84370, + "ĠTableColumn": 72322, + "ĠTableName": 81790, + "ĠTableRow": 80217, + "ĠTableView": 95497, + "ĠTables": 42152, + "ĠTablet": 57303, + "ĠTablets": 84382, + "ĠTabs": 52614, + "ĠTac": 55509, + "ĠTaco": 78897, + "ĠTacoma": 84528, + "ĠTactical": 70558, + "ĠTactics": 93838, + "ĠTag": 12353, + "ĠTage": 71177, + "ĠTags": 27683, + "ĠTah": 56987, + "ĠTahoe": 95886, + "ĠTai": 21705, + "ĠTail": 36079, + "ĠTaipei": 96345, + "ĠTaiwan": 28289, + "ĠTaiwanese": 92937, + "ĠTaj": 66699, + "ĠTak": 33290, + "ĠTake": 11778, + "ĠTaken": 55974, + "ĠTakes": 37607, + "ĠTaking": 35825, + "ĠTal": 17589, + "ĠTale": 54348, + "ĠTalent": 54507, + "ĠTales": 53855, + "ĠTaliban": 40670, + "ĠTalk": 18976, + "ĠTalking": 60922, + "ĠTalks": 80447, + "ĠTall": 41102, + "ĠTam": 28888, + "ĠTamb": 58126, + "ĠTambién": 90130, + "ĠTambém": 141016, + "ĠTamil": 43783, + "ĠTampa": 32125, + "ĠTan": 24548, + "ĠTang": 40362, + "ĠTango": 98348, + "ĠTank": 31394, + "ĠTanks": 82781, + "ĠTanner": 89796, + "ĠTantra": 88211, + "ĠTanz": 57680, + "ĠTanzania": 68794, + "ĠTao": 59508, + "ĠTap": 36134, + "ĠTape": 57486, + "ĠTar": 23959, + "ĠTara": 69697, + "ĠTarget": 13483, + "ĠTargets": 84947, + "ĠTart": 77541, + "ĠTas": 44687, + "ĠTask": 5430, + "ĠTasks": 46471, + "ĠTasmania": 92428, + "ĠTaste": 57981, + "ĠTat": 62467, + "ĠTata": 87145, + "ĠTate": 67008, + "ĠTatto": 60092, + "ĠTattoo": 75534, + "ĠTau": 80085, + "ĠTavern": 80300, + "ĠTax": 15190, + "ĠTaxes": 71737, + "ĠTaxi": 78275, + "ĠTay": 57528, + "ĠTaylor": 16439, + "ĠTb": 98782, + "ĠTcl": 56491, + "ĠTcp": 64876, + "ĠTe": 2655, + "ĠTea": 30025, + "ĠTeach": 69277, + "ĠTeacher": 29069, + "ĠTeachers": 46702, + "ĠTeaching": 44277, + "ĠTeam": 7909, + "ĠTeams": 39613, + "ĠTear": 76982, + "ĠTears": 90914, + "ĠTec": 92487, + "ĠTech": 17374, + "ĠTechn": 7001, + "ĠTechnical": 26668, + "ĠTechnician": 59056, + "ĠTechnique": 77619, + "ĠTechniques": 64944, + "ĠTechnologies": 24561, + "ĠTechnology": 11791, + "ĠTecn": 87220, + "ĠTed": 23115, + "ĠTeddy": 70066, + "ĠTee": 66999, + "ĠTeen": 23257, + "ĠTeens": 69053, + "ĠTeeth": 97878, + "ĠTeh": 44889, + "ĠTehran": 52695, + "ĠTeil": 47355, + "ĠTek": 55434, + "ĠTel": 22836, + "ĠTele": 13563, + "ĠTelecom": 70864, + "ĠTelefon": 68987, + "ĠTelegram": 42963, + "ĠTelegraph": 48806, + "ĠTelephone": 56930, + "ĠTelerik": 38846, + "ĠTelescope": 85962, + "ĠTelevision": 40345, + "ĠTell": 24647, + "ĠTells": 98890, + "ĠTem": 8647, + "ĠTemp": 19944, + "ĠTempData": 80967, + "ĠTemper": 74683, + "ĠTemperature": 37022, + "ĠTempl": 78810, + "ĠTemplate": 14355, + "ĠTemplates": 19264, + "ĠTemple": 18659, + "ĠTempo": 87503, + "ĠTemporary": 54448, + "ĠTen": 17695, + "ĠTenant": 73724, + "ĠTencent": 69921, + "ĠTender": 93867, + "ĠTenn": 75526, + "ĠTennessee": 23219, + "ĠTennis": 57151, + "ĠTensor": 26036, + "ĠTensorFlow": 94986, + "ĠTent": 66580, + "ĠTer": 10127, + "ĠTeresa": 63440, + "ĠTerm": 17519, + "ĠTermin": 45945, + "ĠTerminal": 34090, + "ĠTerminator": 97361, + "ĠTerms": 19591, + "ĠTerr": 17655, + "ĠTerra": 49426, + "ĠTerrace": 79170, + "ĠTerrain": 62018, + "ĠTerraria": 85592, + "ĠTerritories": 86343, + "ĠTerritory": 48087, + "ĠTerror": 37156, + "ĠTerrorism": 79664, + "ĠTerry": 31518, + "ĠTes": 23099, + "ĠTesla": 27199, + "ĠTess": 91045, + "ĠTest": 3393, + "ĠTestBed": 17722, + "ĠTestCase": 30573, + "ĠTestData": 93200, + "ĠTestUtils": 84745, + "ĠTestament": 39114, + "ĠTested": 88505, + "ĠTester": 76816, + "ĠTesting": 26768, + "ĠTests": 20150, + "ĠTet": 49762, + "ĠTeuchos": 85089, + "ĠTex": 31468, + "ĠTexans": 52691, + "ĠTexas": 8257, + "ĠText": 2918, + "ĠTextAlign": 54410, + "ĠTextArea": 88728, + "ĠTextBox": 29694, + "ĠTextEditingController": 45998, + "ĠTextField": 28435, + "ĠTextFormField": 78111, + "ĠTextInput": 42232, + "ĠTextInputType": 92551, + "ĠTextStyle": 15981, + "ĠTextView": 13865, + "ĠTexture": 18757, + "ĠTf": 91456, + "ĠTh": 663, + "ĠThai": 26392, + "ĠThailand": 29467, + "ĠThames": 81188, + "ĠThan": 33676, + "ĠThank": 9731, + "ĠThankfully": 62454, + "ĠThanks": 11114, + "ĠThanksgiving": 36150, + "ĠThat": 2938, + "ĠThatcher": 80985, + "ĠThe": 576, + "ĠTheater": 37471, + "ĠTheatre": 26221, + "ĠTheft": 60999, + "ĠTheir": 10964, + "ĠThem": 47938, + "ĠThema": 61150, + "ĠTheme": 16442, + "ĠThemeData": 95550, + "ĠThemes": 62087, + "ĠThen": 5005, + "ĠTheo": 84137, + "ĠTheodore": 76349, + "ĠTheory": 30435, + "ĠTher": 22450, + "ĠTherapy": 39043, + "ĠThere": 2619, + "ĠTherefore": 15277, + "ĠTheresa": 42712, + "ĠTherm": 67272, + "ĠThermal": 65626, + "ĠThese": 4220, + "ĠThesis": 73997, + "ĠTheta": 68710, + "ĠThey": 2379, + "ĠThi": 59123, + "ĠThick": 81277, + "ĠThickness": 69401, + "ĠThief": 85598, + "ĠThin": 69622, + "ĠThing": 37200, + "ĠThings": 20093, + "ĠThink": 21149, + "ĠThinking": 52289, + "ĠThird": 20861, + "ĠThirty": 66620, + "ĠThis": 1096, + "ĠThiên": 131460, + "ĠThiết": 133086, + "ĠThom": 8715, + "ĠThomas": 11108, + "ĠThompson": 25164, + "ĠThomson": 64232, + "ĠThor": 28450, + "ĠThorn": 76924, + "ĠThornton": 78234, + "ĠThose": 12969, + "ĠThou": 85371, + "ĠThough": 17594, + "ĠThought": 35187, + "ĠThoughts": 60299, + "ĠThousand": 74353, + "ĠThousands": 62194, + "ĠThr": 29565, + "ĠThread": 8752, + "ĠThreadPool": 72716, + "ĠThreads": 74742, + "ĠThreat": 52335, + "ĠThree": 14513, + "ĠThreshold": 58079, + "ĠThrone": 90601, + "ĠThrones": 46394, + "ĠThrough": 16904, + "ĠThroughout": 45882, + "ĠThrow": 31345, + "ĠThrowable": 22124, + "ĠThrows": 66847, + "ĠThu": 35892, + "ĠThumb": 92256, + "ĠThumbnail": 87792, + "ĠThumbnails": 93722, + "ĠThunder": 26691, + "ĠThur": 75055, + "ĠThurs": 82685, + "ĠThursday": 7794, + "ĠThus": 14301, + "ĠThy": 66575, + "ĠThái": 131053, + "ĠTháng": 132493, + "ĠThánh": 139329, + "ĠThông": 130303, + "ĠThÃłnh": 129932, + "ĠThư": 138532, + "ĠThương": 137048, + "ĠThưá»Ŀng": 141948, + "ĠThảo": 142235, + "ĠThần": 140255, + "ĠThế": 131346, + "ĠThá»§": 130390, + "ĠThứ": 132496, + "ĠThá»±c": 136138, + "ĠThá»ĥ": 141770, + "ĠThá»ĭ": 130934, + "ĠThá»Ŀi": 132809, + "ĠTi": 22325, + "ĠTian": 47343, + "ĠTib": 36816, + "ĠTibet": 44161, + "ĠTibetan": 71974, + "ĠTic": 78426, + "ĠTick": 53092, + "ĠTicket": 28397, + "ĠTickets": 35615, + "ĠTide": 71016, + "ĠTie": 58725, + "ĠTier": 35232, + "ĠTiffany": 67473, + "ĠTig": 93879, + "ĠTiger": 35745, + "ĠTigers": 39716, + "ĠTight": 74346, + "ĠTik": 72742, + "ĠTil": 57669, + "ĠTile": 20744, + "ĠTiles": 71580, + "ĠTill": 39688, + "ĠTillerson": 74180, + "ĠTim": 9354, + "ĠTimber": 44148, + "ĠTime": 4120, + "ĠTimeInterval": 98548, + "ĠTimeSpan": 29158, + "ĠTimeUnit": 36767, + "ĠTimeZone": 88709, + "ĠTimeline": 49528, + "ĠTimeout": 39380, + "ĠTimer": 17759, + "ĠTimes": 8523, + "ĠTimestamp": 32758, + "ĠTiming": 70813, + "ĠTimothy": 44468, + "ĠTin": 56255, + "ĠTina": 63687, + "ĠTinder": 65356, + "ĠTing": 98532, + "ĠTiny": 47974, + "ĠTip": 29873, + "ĠTipo": 45020, + "ĠTips": 25704, + "ĠTir": 82555, + "ĠTire": 67855, + "ĠTit": 24016, + "ĠTitan": 27447, + "ĠTitanic": 88762, + "ĠTitanium": 79354, + "ĠTitans": 46611, + "ĠTitle": 10869, + "ĠTitles": 84387, + "ĠTits": 68986, + "ĠTiên": 137713, + "ĠTiến": 133778, + "ĠTiếp": 141937, + "ĠTiá»ģn": 141630, + "ĠTiá»ĥu": 136746, + "ĠTk": 49160, + "ĠTmax": 65774, + "ĠTmin": 67423, + "ĠTo": 2014, + "ĠToDo": 67053, + "ĠToString": 32128, + "ĠToast": 9609, + "ĠToastr": 90419, + "ĠTob": 31756, + "ĠTobacco": 65110, + "ĠTobias": 79002, + "ĠToby": 78379, + "ĠTod": 58583, + "ĠToday": 11201, + "ĠTodd": 29434, + "ĠTodo": 25404, + "ĠTodos": 63231, + "ĠToe": 89804, + "ĠTogether": 31155, + "ĠToggle": 37105, + "ĠToilet": 81894, + "ĠTok": 21447, + "ĠToken": 9660, + "ĠTokenType": 56582, + "ĠTokens": 58166, + "ĠTokyo": 26194, + "ĠTol": 58502, + "ĠToledo": 72277, + "ĠTolkien": 79303, + "ĠToll": 85294, + "ĠTom": 8364, + "ĠTomas": 87936, + "ĠTomato": 92991, + "ĠTomb": 56172, + "ĠTome": 93866, + "ĠTommy": 39039, + "ĠTomorrow": 54996, + "ĠTon": 30716, + "ĠTone": 66904, + "ĠTong": 50391, + "ĠToni": 89302, + "ĠTonight": 54214, + "ĠTony": 18528, + "ĠToo": 24599, + "ĠTook": 89696, + "ĠTool": 13474, + "ĠToolStrip": 61591, + "ĠToolbar": 44284, + "ĠToolbox": 95962, + "ĠToolkit": 54776, + "ĠTools": 13852, + "ĠTooltip": 59223, + "ĠTooth": 82399, + "ĠTop": 6909, + "ĠTopic": 32911, + "ĠTopics": 40894, + "ĠTops": 92405, + "ĠTor": 8444, + "ĠTorah": 71950, + "ĠTorch": 78743, + "ĠTories": 65969, + "ĠToro": 83980, + "ĠToronto": 14632, + "ĠTorrent": 63786, + "ĠTorres": 54407, + "ĠTort": 55827, + "ĠTory": 46150, + "ĠTos": 69303, + "ĠToshiba": 95441, + "ĠTot": 38020, + "ĠTotal": 10657, + "ĠTotally": 81965, + "ĠTottenham": 55548, + "ĠTou": 56795, + "ĠTouch": 19338, + "ĠTouchableOpacity": 63743, + "ĠTough": 77054, + "ĠTour": 14644, + "ĠTourism": 51552, + "ĠTournament": 18371, + "ĠTours": 56049, + "ĠTout": 79805, + "ĠTow": 40393, + "ĠTowards": 93865, + "ĠTower": 21938, + "ĠTowers": 67357, + "ĠTown": 13971, + "ĠTownsend": 95682, + "ĠTownship": 52667, + "ĠToxic": 66402, + "ĠToy": 21603, + "ĠToyota": 29200, + "ĠToys": 48381, + "ĠTr": 1163, + "ĠTra": 17298, + "ĠTrace": 27163, + "ĠTrack": 19785, + "ĠTracker": 40179, + "ĠTracking": 46070, + "ĠTracks": 64740, + "ĠTracy": 59242, + "ĠTrad": 15065, + "ĠTradable": 23070, + "ĠTrade": 17214, + "ĠTrader": 66431, + "ĠTrades": 95780, + "ĠTrading": 31604, + "ĠTradition": 80478, + "ĠTraditional": 45460, + "ĠTraff": 66748, + "ĠTraffic": 36981, + "ĠTrafford": 94646, + "ĠTrail": 18660, + "ĠTrailer": 36942, + "ĠTrails": 93741, + "ĠTrain": 26125, + "ĠTrainer": 46055, + "ĠTraining": 16151, + "ĠTrait": 68301, + "ĠTraits": 67026, + "ĠTran": 57228, + "ĠTrans": 4058, + "ĠTransaction": 17869, + "ĠTransactions": 55285, + "ĠTranscript": 90303, + "ĠTransfer": 23200, + "ĠTransform": 15226, + "ĠTransformation": 53652, + "ĠTransformer": 62379, + "ĠTransformers": 80532, + "ĠTransit": 45855, + "ĠTransition": 37721, + "ĠTransitional": 66216, + "ĠTranslate": 37740, + "ĠTranslation": 38041, + "ĠTranslator": 76191, + "ĠTransmission": 47025, + "ĠTransmit": 97657, + "ĠTransparency": 94131, + "ĠTransparent": 89868, + "ĠTransport": 16742, + "ĠTransportation": 29878, + "ĠTrap": 52820, + "ĠTrash": 70623, + "ĠTrav": 42259, + "ĠTravel": 18097, + "ĠTraverse": 86733, + "ĠTravis": 40710, + "ĠTray": 64616, + "ĠTre": 12475, + "ĠTreasure": 57148, + "ĠTreasurer": 87875, + "ĠTreasury": 31891, + "ĠTreat": 18877, + "ĠTreatment": 30869, + "ĠTreaty": 51031, + "ĠTree": 8942, + "ĠTreeMap": 76883, + "ĠTreeNode": 17111, + "ĠTreeSet": 90427, + "ĠTrees": 58884, + "ĠTrek": 30471, + "ĠTrem": 75163, + "ĠTrend": 30653, + "ĠTrends": 49630, + "ĠTrent": 47520, + "ĠTrev": 45001, + "ĠTrevor": 53551, + "ĠTrey": 81460, + "ĠTri": 12359, + "ĠTrial": 40474, + "ĠTrials": 69444, + "ĠTriangle": 38031, + "ĠTrib": 20523, + "ĠTribal": 90112, + "ĠTribe": 62704, + "ĠTribunal": 56971, + "ĠTribune": 40339, + "ĠTribute": 95202, + "ĠTrick": 69225, + "ĠTricks": 84176, + "ĠTrident": 92066, + "ĠTrie": 62783, + "ĠTried": 84083, + "ĠTrigger": 30291, + "ĠTrilogy": 96213, + "ĠTrim": 44376, + "ĠTrinidad": 85130, + "ĠTrinity": 44559, + "ĠTrio": 84995, + "ĠTrip": 26754, + "ĠTripAdvisor": 83966, + "ĠTriple": 36649, + "ĠTristan": 96590, + "ĠTrit": 84183, + "ĠTriumph": 95894, + "ĠTriá»ģu": 143656, + "ĠTro": 19321, + "ĠTrojan": 95515, + "ĠTroll": 92637, + "ĠTrom": 93367, + "ĠTrophy": 55104, + "ĠTropical": 70243, + "ĠTrot": 89230, + "ĠTrotsky": 90893, + "ĠTrou": 41682, + "ĠTrouble": 80263, + "ĠTrout": 87538, + "ĠTroy": 43399, + "ĠTruck": 33685, + "ĠTrucks": 85335, + "ĠTrudeau": 43709, + "ĠTrue": 3007, + "ĠTruly": 87148, + "ĠTruman": 79836, + "ĠTrump": 3339, + "ĠTrust": 16816, + "ĠTrusted": 95425, + "ĠTrustees": 84096, + "ĠTruth": 29098, + "ĠTruyá»ģn": 141901, + "ĠTry": 9735, + "ĠTrying": 49858, + "ĠTrên": 136067, + "ĠTrÆ°á»Ľc": 133884, + "ĠTrưá»Ŀng": 131431, + "ĠTrần": 131479, + "ĠTrá»įng": 142741, + "ĠTs": 25076, + "ĠTu": 28649, + "ĠTub": 39540, + "ĠTube": 29024, + "ĠTubes": 94370, + "ĠTucker": 55156, + "ĠTucson": 66852, + "ĠTud": 90151, + "ĠTue": 39167, + "ĠTues": 82683, + "ĠTuesday": 7589, + "ĠTul": 42216, + "ĠTulsa": 74562, + "ĠTumblr": 43599, + "ĠTun": 27456, + "ĠTune": 74174, + "ĠTunis": 94677, + "ĠTunisia": 76835, + "ĠTunnel": 65601, + "ĠTup": 96097, + "ĠTuple": 24622, + "ĠTur": 8705, + "ĠTurbo": 47175, + "ĠTuring": 94430, + "ĠTurk": 20545, + "ĠTurkey": 17009, + "ĠTurkish": 23734, + "ĠTurks": 71757, + "ĠTurn": 11999, + "ĠTurnbull": 67846, + "ĠTurner": 33031, + "ĠTurning": 71533, + "ĠTurns": 57234, + "ĠTurtle": 66806, + "ĠTus": 70904, + "ĠTut": 65963, + "ĠTutor": 80619, + "ĠTutorial": 48007, + "ĠTuấn": 140593, + "ĠTv": 49553, + "ĠTw": 11753, + "ĠTwe": 61698, + "ĠTween": 66335, + "ĠTweet": 25155, + "ĠTweets": 77701, + "ĠTwelve": 69232, + "ĠTwenty": 43856, + "ĠTwice": 93566, + "ĠTwig": 41631, + "ĠTwilight": 51045, + "ĠTwin": 34947, + "ĠTwins": 65268, + "ĠTwist": 73374, + "ĠTwitch": 50801, + "ĠTwitter": 6271, + "ĠTwo": 9043, + "ĠTx": 39850, + "ĠTxt": 79104, + "ĠTy": 13897, + "ĠTyler": 31220, + "ĠTyp": 17518, + "ĠType": 3990, + "ĠTypeError": 25030, + "ĠTypeInfo": 72452, + "ĠTypeName": 97725, + "ĠTypeScript": 87457, + "ĠTyped": 50554, + "ĠTypeface": 95184, + "ĠTypes": 20768, + "ĠTypical": 72925, + "ĠTypically": 45302, + "ĠTypography": 46863, + "ĠTyr": 49495, + "ĠTyson": 62814, + "ĠTâm": 137257, + "ĠTân": 133508, + "ĠTây": 130951, + "ĠTélé": 94088, + "ĠTìm": 137328, + "ĠTình": 138458, + "ĠTôi": 129718, + "ĠTüm": 134091, + "ĠTür": 57358, + "ĠTürk": 126441, + "ĠTürkiye": 128504, + "ĠTürkçe": 136891, + "ĠTÃłi": 134352, + "ĠTư": 132590, + "ĠTại": 131677, + "ĠTất": 137388, + "ĠTết": 136832, + "ĠTáºŃp": 137408, + "ĠTừ": 130232, + "ĠTá»ķ": 126778, + "ĠTá»ķng": 130017, + "ĠTá»Ń": 136690, + "ĠU": 547, + "ĠUA": 45570, + "ĠUAE": 46849, + "ĠUART": 23267, + "ĠUAV": 94496, + "ĠUB": 73061, + "ĠUC": 30513, + "ĠUCHAR": 76989, + "ĠUCLA": 49651, + "ĠUCS": 79246, + "ĠUClass": 84844, + "ĠUD": 78551, + "ĠUDP": 42394, + "ĠUE": 29988, + "ĠUEFA": 78715, + "ĠUF": 68592, + "ĠUFC": 27295, + "ĠUFO": 47944, + "ĠUI": 3689, + "ĠUIAlert": 30049, + "ĠUIAlertAction": 55610, + "ĠUIAlertController": 53255, + "ĠUIAlertView": 84860, + "ĠUIApplication": 34904, + "ĠUIB": 40547, + "ĠUIBar": 50785, + "ĠUIBarButtonItem": 53502, + "ĠUIButton": 21098, + "ĠUICollectionView": 28577, + "ĠUICollectionViewCell": 79613, + "ĠUIColor": 18628, + "ĠUIControl": 77904, + "ĠUID": 37428, + "ĠUIEdgeInsets": 69257, + "ĠUIF": 44404, + "ĠUIFont": 46007, + "ĠUIG": 49058, + "ĠUIGraphics": 61669, + "ĠUIImage": 20681, + "ĠUIImagePickerController": 77733, + "ĠUIImageView": 28968, + "ĠUIKit": 18151, + "ĠUIL": 15737, + "ĠUILabel": 16238, + "ĠUIManager": 92765, + "ĠUINT": 13661, + "ĠUINavigationController": 66067, + "ĠUIP": 57142, + "ĠUIPickerView": 92231, + "ĠUIResponder": 68821, + "ĠUIS": 28797, + "ĠUIScreen": 88007, + "ĠUIScrollView": 73652, + "ĠUIStoryboard": 91942, + "ĠUIStoryboardSegue": 69160, + "ĠUIT": 19923, + "ĠUITableView": 13797, + "ĠUITableViewCell": 27330, + "ĠUITableViewController": 96726, + "ĠUITableViewDataSource": 79768, + "ĠUITableViewDelegate": 82057, + "ĠUITapGestureRecognizer": 79885, + "ĠUITextField": 32063, + "ĠUITextView": 88199, + "ĠUIView": 17620, + "ĠUIViewController": 20784, + "ĠUIWindow": 57700, + "ĠUInt": 22275, + "ĠUK": 6424, + "ĠUL": 22143, + "ĠULONG": 30676, + "ĠUM": 57122, + "ĠUN": 6643, + "ĠUNC": 75255, + "ĠUNDER": 56218, + "ĠUNESCO": 80776, + "ĠUNICODE": 98608, + "ĠUNION": 77553, + "ĠUNIQUE": 72350, + "ĠUNIT": 38202, + "ĠUNITED": 84510, + "ĠUNITY": 56084, + "ĠUNIVERS": 64799, + "ĠUNIVERSITY": 81273, + "ĠUNIX": 54810, + "ĠUNKNOWN": 61395, + "ĠUNS": 45983, + "ĠUNSIGNED": 94178, + "ĠUNU": 66925, + "ĠUNUSED": 68001, + "ĠUObject": 55112, + "ĠUP": 11982, + "ĠUPC": 95129, + "ĠUPDATE": 22893, + "ĠUPDATED": 77861, + "ĠUPLOAD": 83818, + "ĠUPPER": 96312, + "ĠUPS": 57046, + "ĠUR": 34414, + "ĠURI": 16020, + "ĠURL": 5548, + "ĠURLRequest": 95833, + "ĠURLSession": 85820, + "ĠURLWithString": 59579, + "ĠURLs": 35006, + "ĠUS": 2274, + "ĠUSA": 7279, + "ĠUSAGE": 89740, + "ĠUSART": 38885, + "ĠUSB": 11348, + "ĠUSC": 43279, + "ĠUSD": 19551, + "ĠUSDA": 59991, + "ĠUSE": 9457, + "ĠUSED": 65217, + "ĠUSER": 13872, + "ĠUSERNAME": 92115, + "ĠUSERS": 90248, + "ĠUSHORT": 97694, + "ĠUSING": 62051, + "ĠUSPS": 84876, + "ĠUSS": 38508, + "ĠUSSR": 72215, + "ĠUT": 25558, + "ĠUTC": 27403, + "ĠUTF": 20076, + "ĠUTIL": 68107, + "ĠUUID": 23698, + "ĠUV": 29036, + "ĠUW": 65616, + "ĠUX": 61493, + "ĠUb": 26461, + "ĠUber": 27612, + "ĠUbisoft": 86897, + "ĠUbuntu": 34960, + "ĠUd": 66880, + "ĠUg": 46330, + "ĠUganda": 57049, + "ĠUh": 68049, + "ĠUhr": 38183, + "ĠUi": 31634, + "ĠUint": 27883, + "ĠUk": 59313, + "ĠUkr": 14299, + "ĠUkrain": 30604, + "ĠUkraine": 18753, + "ĠUkrainian": 33625, + "ĠUl": 16582, + "ĠUlt": 20604, + "ĠUltimate": 28850, + "ĠUltimately": 54006, + "ĠUltr": 66857, + "ĠUltra": 28213, + "ĠUm": 23334, + "ĠUma": 85768, + "ĠUmb": 65959, + "ĠUn": 1230, + "ĠUna": 42971, + "ĠUnable": 41444, + "ĠUnary": 86951, + "ĠUnauthorized": 63102, + "ĠUnblock": 76288, + "ĠUnc": 28779, + "ĠUncategorized": 80928, + "ĠUncle": 50421, + "ĠUncomment": 63858, + "ĠUnd": 16887, + "ĠUndefined": 62370, + "ĠUnder": 9449, + "ĠUnderground": 53887, + "ĠUnderstand": 70894, + "ĠUnderstanding": 45451, + "ĠUndert": 98673, + "ĠUndo": 70809, + "ĠUne": 38683, + "ĠUnexpected": 70400, + "ĠUnfortunately": 18656, + "ĠUng": 91717, + "ĠUni": 48866, + "ĠUnicode": 35897, + "ĠUnicorn": 94970, + "ĠUnidos": 68419, + "ĠUnified": 72434, + "ĠUniform": 47889, + "ĠUnion": 9145, + "ĠUnique": 28650, + "ĠUnit": 7954, + "ĠUnited": 3639, + "ĠUnits": 35181, + "ĠUnity": 21717, + "ĠUnityEditor": 75417, + "ĠUnityEngine": 7207, + "ĠUniv": 66512, + "ĠUnivers": 15545, + "ĠUniversal": 25500, + "ĠUniverse": 28749, + "ĠUniversidad": 66513, + "ĠUniversities": 74798, + "ĠUniversity": 3822, + "ĠUniversität": 142485, + "ĠUnix": 46995, + "ĠUnknown": 21693, + "ĠUnless": 10878, + "ĠUnlike": 26048, + "ĠUnlimited": 48505, + "ĠUnlock": 57430, + "ĠUno": 83101, + "ĠUnreal": 79919, + "ĠUns": 61043, + "ĠUnsafe": 73067, + "ĠUnsigned": 97747, + "ĠUnsupported": 33081, + "ĠUnsupportedOperationException": 42978, + "ĠUnsure": 77831, + "ĠUnt": 15098, + "ĠUnter": 33312, + "ĠUnternehmen": 61197, + "ĠUnters": 91513, + "ĠUnterstüt": 141785, + "ĠUnterstützung": 141786, + "ĠUntil": 28970, + "ĠUnused": 86473, + "ĠUp": 3138, + "ĠUpdate": 5549, + "ĠUpdated": 16073, + "ĠUpdates": 27500, + "ĠUpdating": 78118, + "ĠUpgrade": 40713, + "ĠUph": 98905, + "ĠUpload": 24996, + "ĠUploaded": 68204, + "ĠUpon": 29438, + "ĠUpper": 30614, + "ĠUps": 71316, + "ĠUr": 16809, + "ĠUran": 79670, + "ĠUrb": 71213, + "ĠUrban": 28322, + "ĠUrdu": 93335, + "ĠUrg": 85486, + "ĠUri": 17226, + "ĠUrl": 22840, + "ĠUrs": 70038, + "ĠUruguay": 72290, + "ĠUs": 3985, + "ĠUsa": 95938, + "ĠUsage": 24567, + "ĠUse": 5443, + "ĠUsed": 12199, + "ĠUseful": 50512, + "ĠUser": 2657, + "ĠUserControl": 72665, + "ĠUserController": 58704, + "ĠUserDao": 97104, + "ĠUserData": 71294, + "ĠUserDefaults": 61792, + "ĠUserDetails": 88286, + "ĠUserID": 68567, + "ĠUserId": 40883, + "ĠUserInfo": 47757, + "ĠUserManager": 56234, + "ĠUserModel": 52784, + "ĠUserName": 47398, + "ĠUserProfile": 55099, + "ĠUserRepository": 60233, + "ĠUserRole": 95019, + "ĠUserService": 30119, + "ĠUserType": 93372, + "ĠUsername": 29545, + "ĠUsers": 14627, + "ĠUsersController": 96427, + "ĠUses": 38823, + "ĠUsing": 12091, + "ĠUsually": 32967, + "ĠUsuario": 28816, + "ĠUsuarios": 94993, + "ĠUt": 17139, + "ĠUtah": 22389, + "ĠUtf": 43432, + "ĠUtil": 10167, + "ĠUtilities": 34994, + "ĠUtility": 25169, + "ĠUtils": 17954, + "ĠUtt": 66962, + "ĠUttar": 85673, + "ĠUzbek": 89859, + "ĠV": 647, + "ĠVA": 20901, + "ĠVAL": 9714, + "ĠVALID": 43291, + "ĠVALUE": 26827, + "ĠVALUES": 14710, + "ĠVAN": 96653, + "ĠVAR": 43508, + "ĠVARCHAR": 37589, + "ĠVARIABLE": 54293, + "ĠVARIABLES": 76753, + "ĠVARIANT": 82346, + "ĠVAT": 47592, + "ĠVB": 63439, + "ĠVBox": 97992, + "ĠVC": 43495, + "ĠVE": 60888, + "ĠVECTOR": 71459, + "ĠVER": 29261, + "ĠVERBOSE": 87041, + "ĠVERIFY": 78204, + "ĠVERSION": 33792, + "ĠVERY": 47074, + "ĠVF": 72892, + "ĠVG": 57818, + "ĠVGA": 75356, + "ĠVH": 98022, + "ĠVI": 29668, + "ĠVIA": 98427, + "ĠVIC": 70103, + "ĠVID": 83952, + "ĠVIDEO": 41769, + "ĠVIEW": 40656, + "ĠVII": 44969, + "ĠVIII": 57233, + "ĠVIN": 83030, + "ĠVIP": 35069, + "ĠVIR": 74695, + "ĠVIS": 59491, + "ĠVK": 21559, + "ĠVL": 60222, + "ĠVLAN": 75880, + "ĠVLC": 93505, + "ĠVM": 17792, + "ĠVMware": 64912, + "ĠVN": 97507, + "ĠVO": 43652, + "ĠVOC": 92521, + "ĠVOID": 59257, + "ĠVOL": 68226, + "ĠVP": 36010, + "ĠVPN": 30747, + "ĠVR": 19168, + "ĠVS": 30650, + "ĠVStack": 94659, + "ĠVT": 32951, + "ĠVW": 52185, + "ĠVX": 78177, + "ĠVa": 26615, + "ĠVac": 30526, + "ĠVacation": 69151, + "ĠVacc": 58688, + "ĠVaccine": 98958, + "ĠVacuum": 75010, + "ĠVad": 92112, + "ĠVader": 68409, + "ĠVal": 4104, + "ĠVale": 53969, + "ĠValencia": 60249, + "ĠValent": 76940, + "ĠValentine": 38769, + "ĠValerie": 89257, + "ĠValid": 7818, + "ĠValidate": 23282, + "ĠValidates": 97843, + "ĠValidation": 18954, + "ĠValidationError": 53029, + "ĠValidationResult": 98594, + "ĠValidator": 32566, + "ĠValidators": 24659, + "ĠValk": 85162, + "ĠVall": 78663, + "ĠValle": 77494, + "ĠValley": 13046, + "ĠValor": 61728, + "ĠValue": 5162, + "ĠValueError": 15402, + "ĠValueEventListener": 98145, + "ĠValueType": 55545, + "ĠValues": 24979, + "ĠValve": 44293, + "ĠVampire": 60114, + "ĠVan": 12710, + "ĠVance": 91268, + "ĠVancouver": 22575, + "ĠVand": 96847, + "ĠVander": 58042, + "ĠVanderbilt": 83026, + "ĠVanessa": 70519, + "ĠVanguard": 75821, + "ĠVanilla": 65582, + "ĠVanity": 68394, + "ĠVapor": 88654, + "ĠVar": 8735, + "ĠVari": 27868, + "ĠVariable": 12407, + "ĠVariables": 21419, + "ĠVariant": 39292, + "ĠVariation": 88485, + "ĠVariety": 70732, + "ĠVarious": 39641, + "ĠVas": 52834, + "ĠVatican": 46547, + "ĠVaugh": 62955, + "ĠVaughan": 95148, + "ĠVault": 41397, + "ĠVaults": 93555, + "ĠVe": 22584, + "ĠVec": 11312, + "ĠVect": 90156, + "ĠVector": 4196, + "ĠVed": 74787, + "ĠVeg": 89913, + "ĠVega": 64606, + "ĠVegan": 62115, + "ĠVegas": 17596, + "ĠVeget": 42700, + "ĠVegetable": 94292, + "ĠVehicle": 24707, + "ĠVehicles": 55020, + "ĠVel": 30198, + "ĠVelocity": 54434, + "ĠVelvet": 73027, + "ĠVen": 18231, + "ĠVend": 84553, + "ĠVendor": 45136, + "ĠVenez": 27983, + "ĠVenezuel": 70814, + "ĠVenezuela": 35903, + "ĠVenezuelan": 82649, + "ĠVenice": 55650, + "ĠVenom": 89586, + "ĠVent": 26407, + "ĠVentura": 84654, + "ĠVenture": 70374, + "ĠVentures": 74796, + "ĠVenue": 68360, + "ĠVenus": 48976, + "ĠVer": 6250, + "ĠVera": 67100, + "ĠVeranst": 92610, + "ĠVerb": 59214, + "ĠVerd": 92707, + "ĠVerde": 72848, + "ĠVere": 68832, + "ĠVerfüg": 83279, + "ĠVerfügung": 90769, + "ĠVerg": 51309, + "ĠVerification": 55473, + "ĠVerified": 63169, + "ĠVerify": 25429, + "ĠVerizon": 42314, + "ĠVerm": 29247, + "ĠVermont": 34639, + "ĠVern": 48745, + "ĠVernon": 59137, + "ĠVeronica": 82521, + "ĠVers": 24209, + "ĠVerse": 84068, + "ĠVersion": 6079, + "ĠVersions": 85174, + "ĠVert": 15058, + "ĠVertex": 23225, + "ĠVertical": 35463, + "ĠVery": 15308, + "ĠVes": 78462, + "ĠVest": 55888, + "ĠVet": 80340, + "ĠVeter": 25745, + "ĠVeteran": 67526, + "ĠVeterans": 39332, + "ĠVeterinary": 77003, + "ĠVi": 11401, + "ĠVia": 44799, + "ĠViagra": 30478, + "ĠVib": 96564, + "ĠVibr": 59188, + "ĠVic": 43747, + "ĠVice": 22461, + "ĠVick": 70262, + "ĠVict": 11760, + "ĠVictim": 93024, + "ĠVictims": 88126, + "ĠVictor": 32312, + "ĠVictoria": 22417, + "ĠVictorian": 43177, + "ĠVictory": 48327, + "ĠVid": 47554, + "ĠVide": 81501, + "ĠVideo": 8354, + "ĠVideos": 19544, + "ĠVie": 59302, + "ĠVienna": 46287, + "ĠViet": 17731, + "ĠVietnam": 22500, + "ĠVietnamese": 48477, + "ĠView": 2738, + "ĠViewBag": 27044, + "ĠViewChild": 68507, + "ĠViewController": 25295, + "ĠViewData": 62066, + "ĠViewGroup": 23801, + "ĠViewHolder": 36410, + "ĠViewModel": 33614, + "ĠViewPager": 90185, + "ĠViewState": 72414, + "ĠViewer": 58503, + "ĠViewing": 94057, + "ĠViews": 24942, + "ĠVig": 57101, + "ĠVij": 78027, + "ĠVik": 28376, + "ĠViking": 64350, + "ĠVikings": 40348, + "ĠViktor": 76016, + "ĠVil": 63649, + "ĠVill": 16550, + "ĠVilla": 36812, + "ĠVillage": 24069, + "ĠVille": 89130, + "ĠVim": 94484, + "ĠVimeo": 98825, + "ĠVin": 20391, + "ĠVince": 55132, + "ĠVincent": 34307, + "ĠVinci": 96766, + "ĠVine": 49695, + "ĠVintage": 35845, + "ĠVinyl": 53534, + "ĠViol": 29455, + "ĠViolence": 46632, + "ĠViolet": 73474, + "ĠVir": 9542, + "ĠVirgin": 11214, + "ĠVirginia": 12989, + "ĠVirt": 47775, + "ĠVirtual": 20721, + "ĠVirus": 63992, + "ĠVis": 7656, + "ĠVisa": 43334, + "ĠVish": 87152, + "ĠVisibility": 33735, + "ĠVisible": 70296, + "ĠVision": 30441, + "ĠVisit": 19008, + "ĠVisitor": 55882, + "ĠVisitors": 68840, + "ĠVista": 34612, + "ĠVisual": 20185, + "ĠVisualization": 90760, + "ĠVit": 28460, + "ĠVita": 54624, + "ĠVital": 54271, + "ĠVitamin": 44674, + "ĠViv": 41036, + "ĠVive": 65854, + "ĠVivo": 88168, + "ĠViá»ĩc": 133817, + "ĠViá»ĩn": 141004, + "ĠViá»ĩt": 128324, + "ĠVk": 32656, + "ĠVlad": 29634, + "ĠVladimir": 34911, + "ĠVo": 28079, + "ĠVoc": 86045, + "ĠVocabulary": 98172, + "ĠVocal": 97303, + "ĠVocê": 80797, + "ĠVog": 65911, + "ĠVogue": 89972, + "ĠVoice": 27930, + "ĠVoices": 82326, + "ĠVoid": 19193, + "ĠVol": 10882, + "ĠVoldemort": 80691, + "ĠVolk": 42514, + "ĠVolkswagen": 50839, + "ĠVoll": 96382, + "ĠVolley": 77179, + "ĠVolt": 65389, + "ĠVoltage": 60938, + "ĠVolume": 20265, + "ĠVolunteer": 64435, + "ĠVolunteers": 85170, + "ĠVolvo": 67252, + "ĠVon": 42079, + "ĠVoor": 68395, + "ĠVor": 33328, + "ĠVote": 34034, + "ĠVoter": 94024, + "ĠVoters": 82722, + "ĠVotes": 65293, + "ĠVoting": 63131, + "ĠVous": 40521, + "ĠVox": 79612, + "ĠVoy": 43623, + "ĠVoyage": 95272, + "ĠVoyager": 85160, + "ĠVs": 42982, + "ĠVu": 56602, + "ĠVue": 22256, + "ĠVuex": 88339, + "ĠVul": 37406, + "ĠVulcan": 93055, + "ĠVulkan": 85864, + "ĠVulner": 80111, + "ĠVy": 95000, + "ĠVân": 136269, + "ĠVì": 130629, + "ĠVÃł": 129758, + "ĠVÄ©nh": 142193, + "ĠVÄĥn": 129126, + "ĠVÅ©": 132644, + "ĠVương": 138219, + "ĠVáºŃy": 136241, + "ĠVá»ģ": 135433, + "ĠVỼi": 129496, + "ĠW": 467, + "ĠWA": 28566, + "ĠWAIT": 54390, + "ĠWAL": 73792, + "ĠWALL": 78375, + "ĠWAN": 98802, + "ĠWANT": 71206, + "ĠWAR": 57983, + "ĠWARN": 68509, + "ĠWARNING": 26479, + "ĠWARRANT": 4750, + "ĠWARRANTIES": 7427, + "ĠWARRANTY": 8236, + "ĠWAS": 37776, + "ĠWATCH": 47407, + "ĠWATER": 75247, + "ĠWAV": 88330, + "ĠWAY": 24400, + "ĠWB": 52503, + "ĠWC": 36646, + "ĠWCHAR": 73264, + "ĠWCS": 98351, + "ĠWD": 63612, + "ĠWE": 19677, + "ĠWEB": 43956, + "ĠWEEK": 57429, + "ĠWEIGHT": 95820, + "ĠWELL": 90152, + "ĠWEST": 65090, + "ĠWF": 72324, + "ĠWG": 98075, + "ĠWH": 8494, + "ĠWHAT": 37435, + "ĠWHATSOEVER": 91892, + "ĠWHEN": 33633, + "ĠWHERE": 5288, + "ĠWHETHER": 17647, + "ĠWHICH": 78164, + "ĠWHILE": 92134, + "ĠWHITE": 42767, + "ĠWHO": 39212, + "ĠWHY": 81694, + "ĠWI": 41923, + "ĠWIDTH": 42810, + "ĠWIFI": 69576, + "ĠWILL": 27915, + "ĠWIN": 22064, + "ĠWINAPI": 53910, + "ĠWINDOW": 39259, + "ĠWINDOWS": 75165, + "ĠWITH": 4769, + "ĠWITHOUT": 6007, + "ĠWK": 67350, + "ĠWL": 59514, + "ĠWLAN": 70847, + "ĠWM": 23155, + "ĠWN": 77264, + "ĠWOM": 80275, + "ĠWON": 97367, + "ĠWORD": 36891, + "ĠWORK": 31804, + "ĠWORLD": 50891, + "ĠWOW": 61300, + "ĠWP": 23253, + "ĠWPARAM": 84428, + "ĠWR": 24328, + "ĠWRITE": 38552, + "ĠWRONG": 98660, + "ĠWS": 24906, + "ĠWT": 58099, + "ĠWTF": 77170, + "ĠWTO": 99133, + "ĠWV": 74425, + "ĠWW": 28044, + "ĠWWE": 33438, + "ĠWWII": 60030, + "ĠWWW": 67945, + "ĠWX": 77441, + "ĠWa": 28514, + "ĠWade": 42882, + "ĠWag": 85893, + "ĠWage": 94437, + "ĠWagner": 51375, + "ĠWah": 69590, + "ĠWahl": 74857, + "ĠWaist": 94685, + "ĠWait": 13824, + "ĠWaitFor": 37601, + "ĠWaitForSeconds": 44878, + "ĠWaiting": 52186, + "ĠWak": 85537, + "ĠWake": 40574, + "ĠWal": 14574, + "ĠWald": 50568, + "ĠWales": 22931, + "ĠWalk": 12554, + "ĠWalker": 22278, + "ĠWalking": 39055, + "ĠWall": 9736, + "ĠWallace": 38856, + "ĠWallet": 36483, + "ĠWallpaper": 70796, + "ĠWalls": 71178, + "ĠWalmart": 35767, + "ĠWalnut": 85202, + "ĠWalsh": 51194, + "ĠWalt": 35267, + "ĠWalter": 32205, + "ĠWalters": 90536, + "ĠWalton": 70409, + "ĠWan": 71018, + "ĠWand": 68929, + "ĠWander": 68885, + "ĠWang": 28246, + "ĠWanna": 95708, + "ĠWant": 23252, + "ĠWanted": 66610, + "ĠWants": 77478, + "ĠWar": 5004, + "ĠWarcraft": 63652, + "ĠWard": 26640, + "ĠWare": 68734, + "ĠWarehouse": 51366, + "ĠWarfare": 66594, + "ĠWarm": 45763, + "ĠWarn": 67746, + "ĠWarner": 31680, + "ĠWarning": 26858, + "ĠWarp": 77810, + "ĠWarranty": 56529, + "ĠWarren": 25631, + "ĠWarrior": 39222, + "ĠWarriors": 31887, + "ĠWars": 14968, + "ĠWarsaw": 72176, + "ĠWarwick": 77102, + "ĠWas": 14804, + "ĠWash": 5985, + "ĠWasher": 96082, + "ĠWashing": 81138, + "ĠWashington": 6515, + "ĠWass": 74519, + "ĠWasser": 73794, + "ĠWaste": 48422, + "ĠWat": 20693, + "ĠWatch": 10357, + "ĠWatches": 98260, + "ĠWatching": 68576, + "ĠWater": 9959, + "ĠWaterloo": 76375, + "ĠWaterproof": 78889, + "ĠWaters": 46878, + "ĠWatkins": 78288, + "ĠWatson": 31480, + "ĠWatt": 58975, + "ĠWatts": 58236, + "ĠWave": 31318, + "ĠWaves": 80458, + "ĠWax": 75742, + "ĠWay": 12149, + "ĠWayne": 27540, + "ĠWays": 41319, + "ĠWe": 1205, + "ĠWeak": 41164, + "ĠWealth": 59741, + "ĠWeapon": 27777, + "ĠWeapons": 46564, + "ĠWear": 41062, + "ĠWeather": 22629, + "ĠWeaver": 66272, + "ĠWeb": 4895, + "ĠWebClient": 93354, + "ĠWebDriver": 41896, + "ĠWebDriverWait": 63646, + "ĠWebElement": 39882, + "ĠWebGL": 85841, + "ĠWebSocket": 47042, + "ĠWebView": 69551, + "ĠWebb": 54388, + "ĠWebcam": 46573, + "ĠWeber": 55520, + "ĠWebseite": 94376, + "ĠWebsite": 16023, + "ĠWebsites": 63637, + "ĠWebster": 63277, + "ĠWed": 6521, + "ĠWedding": 30817, + "ĠWednesday": 7920, + "ĠWeed": 93309, + "ĠWeek": 10348, + "ĠWeekend": 47434, + "ĠWeekly": 33014, + "ĠWeeks": 63037, + "ĠWeg": 58534, + "ĠWei": 52448, + "ĠWeight": 16516, + "ĠWein": 70513, + "ĠWeiner": 93666, + "ĠWeinstein": 55582, + "ĠWeird": 76153, + "ĠWeiss": 41155, + "ĠWeiter": 72998, + "ĠWel": 25006, + "ĠWelch": 84163, + "ĠWelcome": 20166, + "ĠWeld": 60486, + "ĠWelfare": 65087, + "ĠWell": 8325, + "ĠWellington": 56004, + "ĠWellness": 60183, + "ĠWells": 36858, + "ĠWelsh": 45781, + "ĠWelt": 44966, + "ĠWen": 61223, + "ĠWend": 41268, + "ĠWendy": 54576, + "ĠWenger": 76927, + "ĠWenn": 43997, + "ĠWent": 53759, + "ĠWer": 35780, + "ĠWere": 38970, + "ĠWerk": 89423, + "ĠWerner": 79735, + "ĠWert": 78445, + "ĠWes": 36376, + "ĠWesley": 57606, + "ĠWest": 4312, + "ĠWestbrook": 88218, + "ĠWestern": 10867, + "ĠWestminster": 46943, + "ĠWeston": 90019, + "ĠWet": 44856, + "ĠWh": 1232, + "ĠWhale": 96002, + "ĠWhat": 3555, + "ĠWhatever": 32639, + "ĠWhats": 29756, + "ĠWhatsApp": 40013, + "ĠWhatsapp": 71901, + "ĠWhe": 32004, + "ĠWheat": 55638, + "ĠWheel": 32342, + "ĠWheeler": 60346, + "ĠWheels": 51449, + "ĠWhen": 3197, + "ĠWhenever": 42533, + "ĠWhere": 10967, + "ĠWhereas": 60595, + "ĠWhether": 13139, + "ĠWhich": 15920, + "ĠWhile": 5976, + "ĠWhilst": 70728, + "ĠWhip": 94302, + "ĠWhisper": 96974, + "ĠWhit": 26386, + "ĠWhite": 5807, + "ĠWhites": 72462, + "ĠWhitney": 66097, + "ĠWho": 10479, + "ĠWhoever": 90295, + "ĠWhole": 40493, + "ĠWholesale": 55416, + "ĠWhy": 8429, + "ĠWi": 17220, + "ĠWiFi": 33307, + "ĠWich": 76589, + "ĠWichita": 87618, + "ĠWick": 74106, + "ĠWid": 47640, + "ĠWide": 32745, + "ĠWideString": 98115, + "ĠWidget": 12980, + "ĠWidgets": 77176, + "ĠWidow": 88020, + "ĠWidth": 24627, + "ĠWie": 42616, + "ĠWien": 98886, + "ĠWife": 42408, + "ĠWifi": 59871, + "ĠWig": 97156, + "ĠWii": 39192, + "ĠWik": 42105, + "ĠWiki": 29228, + "ĠWikiLeaks": 58547, + "ĠWikimedia": 89840, + "ĠWikipedia": 26587, + "ĠWil": 10562, + "ĠWild": 13630, + "ĠWildcats": 96850, + "ĠWilde": 88923, + "ĠWilderness": 84092, + "ĠWildlife": 41549, + "ĠWiley": 74802, + "ĠWilhelm": 92437, + "ĠWilkinson": 94566, + "ĠWill": 4841, + "ĠWilliam": 12375, + "ĠWilliams": 13612, + "ĠWilliamson": 68531, + "ĠWillie": 61885, + "ĠWillis": 66642, + "ĠWillow": 64349, + "ĠWilmington": 84505, + "ĠWilson": 17426, + "ĠWimbledon": 90157, + "ĠWin": 12190, + "ĠWinchester": 79471, + "ĠWind": 22087, + "ĠWindow": 13642, + "ĠWindowManager": 77755, + "ĠWindows": 5515, + "ĠWinds": 43283, + "ĠWindsor": 56455, + "ĠWine": 31826, + "ĠWing": 34522, + "ĠWings": 45097, + "ĠWinn": 40531, + "ĠWinner": 45920, + "ĠWinners": 76312, + "ĠWinning": 72965, + "ĠWinnipeg": 51882, + "ĠWins": 48444, + "ĠWinston": 47108, + "ĠWinter": 20101, + "ĠWir": 34903, + "ĠWire": 19378, + "ĠWired": 86285, + "ĠWireless": 33642, + "ĠWiring": 47853, + "ĠWis": 15830, + "ĠWisconsin": 20441, + "ĠWisdom": 56101, + "ĠWise": 61001, + "ĠWish": 37647, + "ĠWishlist": 84757, + "ĠWit": 71859, + "ĠWitch": 38350, + "ĠWith": 3085, + "ĠWithEvents": 16371, + "ĠWithdraw": 67381, + "ĠWithin": 24236, + "ĠWithout": 17147, + "ĠWitness": 48899, + "ĠWitnesses": 86339, + "ĠWitt": 82176, + "ĠWizard": 38387, + "ĠWizards": 57594, + "ĠWo": 27258, + "ĠWoW": 87645, + "ĠWoche": 81440, + "ĠWochen": 69713, + "ĠWohn": 52795, + "ĠWohnung": 98344, + "ĠWoj": 96160, + "ĠWol": 29543, + "ĠWolf": 25230, + "ĠWolfe": 79552, + "ĠWolff": 95145, + "ĠWolfgang": 86498, + "ĠWolver": 58518, + "ĠWolverine": 83179, + "ĠWolves": 63479, + "ĠWoman": 24511, + "ĠWomen": 10973, + "ĠWomens": 68619, + "ĠWon": 42449, + "ĠWonder": 26113, + "ĠWonderful": 67863, + "ĠWonderland": 88924, + "ĠWong": 55628, + "ĠWoo": 55846, + "ĠWooCommerce": 88206, + "ĠWood": 12131, + "ĠWooden": 62492, + "ĠWoods": 34748, + "ĠWoodward": 86118, + "ĠWoody": 74959, + "ĠWool": 46300, + "ĠWor": 35536, + "ĠWorce": 73188, + "ĠWorcester": 88369, + "ĠWord": 9322, + "ĠWordPress": 19807, + "ĠWordpress": 88069, + "ĠWords": 27630, + "ĠWork": 5547, + "ĠWorkbook": 83462, + "ĠWorker": 33086, + "ĠWorkers": 35698, + "ĠWorkflow": 60173, + "ĠWorking": 22150, + "ĠWorkout": 69142, + "ĠWorkplace": 93187, + "ĠWorks": 21104, + "ĠWorksheet": 82631, + "ĠWorksheets": 76179, + "ĠWorkshop": 35102, + "ĠWorkspace": 70934, + "ĠWorld": 4337, + "ĠWorlds": 51847, + "ĠWorldwide": 51935, + "ĠWorm": 73030, + "ĠWorse": 86804, + "ĠWorship": 86122, + "ĠWorst": 69092, + "ĠWort": 93504, + "ĠWorth": 36146, + "ĠWould": 18885, + "ĠWouldn": 89864, + "ĠWow": 45717, + "ĠWr": 31630, + "ĠWrap": 42187, + "ĠWrapped": 94715, + "ĠWrapper": 35488, + "ĠWrath": 89515, + "ĠWrest": 44937, + "ĠWrestle": 98299, + "ĠWrestling": 61568, + "ĠWright": 26741, + "ĠWrite": 9645, + "ĠWriteLine": 83926, + "ĠWriter": 29404, + "ĠWriters": 58698, + "ĠWrites": 53288, + "ĠWriting": 23893, + "ĠWritten": 33891, + "ĠWroc": 139221, + "ĠWrocÅĤaw": 139222, + "ĠWrong": 40756, + "ĠWu": 36130, + "ĠWww": 81210, + "ĠWy": 23222, + "ĠWyatt": 87462, + "ĠWyn": 76085, + "ĠWyoming": 48732, + "ĠWäh": 142072, + "ĠWährend": 142073, + "ĠX": 1599, + "ĠXB": 74660, + "ĠXBOOLE": 71645, + "ĠXC": 60342, + "ĠXCT": 15929, + "ĠXCTAssert": 20272, + "ĠXCTAssertEqual": 32550, + "ĠXCTAssertTrue": 79462, + "ĠXCTest": 41240, + "ĠXCTestCase": 62810, + "ĠXD": 71190, + "ĠXElement": 58165, + "ĠXF": 95690, + "ĠXHTML": 64852, + "ĠXI": 53243, + "ĠXII": 86815, + "ĠXIII": 95760, + "ĠXIV": 93415, + "ĠXK": 65723, + "ĠXL": 29881, + "ĠXM": 51020, + "ĠXML": 11874, + "ĠXMLHttpRequest": 45838, + "ĠXO": 98080, + "ĠXOR": 69887, + "ĠXP": 24799, + "ĠXPAR": 91154, + "ĠXPath": 65142, + "ĠXR": 67589, + "ĠXS": 57866, + "ĠXSS": 88502, + "ĠXT": 61866, + "ĠXV": 52594, + "ĠXVI": 93837, + "ĠXX": 29288, + "ĠXXX": 19975, + "ĠXY": 57319, + "ĠXYZ": 71089, + "ĠXamarin": 44361, + "ĠXavier": 61760, + "ĠXbox": 20577, + "ĠXen": 48229, + "ĠXi": 42426, + "ĠXia": 40135, + "ĠXiao": 65590, + "ĠXiaomi": 71449, + "ĠXin": 69661, + "ĠXm": 67559, + "ĠXml": 19201, + "ĠXmlDocument": 75431, + "ĠXmlNode": 75533, + "ĠXperia": 85889, + "ĠXt": 45453, + "ĠXu": 56414, + "ĠXunit": 59361, + "ĠXuân": 133361, + "ĠXxx": 49028, + "ĠY": 809, + "ĠYA": 78150, + "ĠYAML": 53127, + "ĠYE": 76948, + "ĠYEAR": 50905, + "ĠYEARS": 88813, + "ĠYELLOW": 81276, + "ĠYES": 14080, + "ĠYM": 84445, + "ĠYORK": 46304, + "ĠYOU": 14985, + "ĠYOUR": 20922, + "ĠYT": 81992, + "ĠYY": 16866, + "ĠYYS": 96787, + "ĠYYSTACK": 86814, + "ĠYYSTYPE": 97588, + "ĠYYYY": 60591, + "ĠYa": 39384, + "ĠYad": 84983, + "ĠYah": 82462, + "ĠYahoo": 30689, + "ĠYak": 70868, + "ĠYale": 43452, + "ĠYam": 33554, + "ĠYamaha": 58435, + "ĠYan": 24213, + "ĠYang": 24474, + "ĠYankee": 85576, + "ĠYankees": 44283, + "ĠYao": 89930, + "ĠYap": 136668, + "ĠYapı": 136669, + "ĠYard": 38237, + "ĠYas": 69928, + "ĠYates": 83713, + "ĠYay": 138496, + "ĠYayın": 138497, + "ĠYaz": 88570, + "ĠYazı": 138055, + "ĠYaÅŁ": 135255, + "ĠYe": 31169, + "ĠYeah": 21607, + "ĠYear": 9742, + "ĠYears": 22315, + "ĠYellow": 25462, + "ĠYellowstone": 95079, + "ĠYelp": 85314, + "ĠYemen": 32690, + "ĠYep": 84194, + "ĠYes": 7414, + "ĠYesterday": 60033, + "ĠYet": 14626, + "ĠYi": 47834, + "ĠYield": 77378, + "ĠYii": 18258, + "ĠYin": 82858, + "ĠYing": 87670, + "ĠYo": 43088, + "ĠYog": 94571, + "ĠYoga": 37573, + "ĠYok": 93536, + "ĠYong": 67954, + "ĠYork": 4261, + "ĠYorker": 63774, + "ĠYorkers": 88173, + "ĠYorkshire": 50227, + "ĠYosemite": 94827, + "ĠYosh": 73008, + "ĠYoshi": 89642, + "ĠYou": 1446, + "ĠYouTube": 13370, + "ĠYoung": 13263, + "ĠYour": 4615, + "ĠYours": 96918, + "ĠYourself": 58905, + "ĠYouth": 29060, + "ĠYoutube": 37303, + "ĠYu": 27273, + "ĠYuan": 68884, + "ĠYue": 81031, + "ĠYug": 69725, + "ĠYugosl": 74451, + "ĠYugoslavia": 96777, + "ĠYuk": 57663, + "ĠYun": 53992, + "ĠYup": 53853, + "ĠYuri": 85773, + "ĠYus": 93348, + "ĠYên": 138902, + "ĠYö": 137177, + "ĠYönet": 137178, + "ĠYönetim": 137179, + "ĠYük": 139257, + "ĠYüksek": 139258, + "ĠYıl": 138477, + "ĠZ": 1863, + "ĠZEND": 92370, + "ĠZERO": 54593, + "ĠZIP": 56674, + "ĠZX": 71669, + "ĠZZ": 90181, + "ĠZa": 64708, + "ĠZac": 91646, + "ĠZach": 38215, + "ĠZack": 70466, + "ĠZag": 88618, + "ĠZah": 96855, + "ĠZahl": 82745, + "ĠZak": 67453, + "ĠZam": 73889, + "ĠZambia": 94004, + "ĠZap": 69231, + "ĠZar": 61431, + "ĠZe": 9919, + "ĠZealand": 16911, + "ĠZeit": 28831, + "ĠZelda": 54631, + "ĠZen": 31789, + "ĠZend": 31117, + "ĠZero": 18306, + "ĠZeroConstructor": 54714, + "ĠZeus": 74357, + "ĠZh": 33349, + "ĠZhang": 36020, + "ĠZhao": 69281, + "ĠZheng": 76447, + "ĠZhou": 66827, + "ĠZhu": 67744, + "ĠZi": 93284, + "ĠZiel": 77356, + "ĠZig": 73761, + "ĠZika": 80926, + "ĠZimbabwe": 51220, + "ĠZimmer": 40407, + "ĠZimmerman": 65658, + "ĠZinc": 95399, + "ĠZion": 40335, + "ĠZionist": 77473, + "ĠZip": 29232, + "ĠZo": 44803, + "ĠZodiac": 97079, + "ĠZoe": 74641, + "ĠZombie": 56991, + "ĠZombies": 88837, + "ĠZone": 22178, + "ĠZones": 94586, + "ĠZoo": 40860, + "ĠZoom": 37529, + "ĠZot": 88824, + "ĠZu": 60082, + "ĠZub": 95714, + "ĠZucker": 54504, + "ĠZuckerberg": 65934, + "ĠZug": 76716, + "ĠZuk": 88019, + "ĠZukunft": 99122, + "ĠZum": 71474, + "ĠZur": 49192, + "ĠZurich": 71726, + "ĠZus": 58485, + "ĠZusammen": 76604, + "ĠZust": 93716, + "ĠZw": 63165, + "ĠZwe": 83126, + "ĠZy": 86554, + "Ġ[": 508, + "Ġ[\"": 4383, + "Ġ[#": 83703, + "Ġ[$": 24264, + "Ġ[%": 23610, + "Ġ[&": 34336, + "Ġ[&](": 61338, + "Ġ['": 2509, + "Ġ['#": 90084, + "Ġ['$": 44509, + "Ġ['',": 43660, + "Ġ['-": 87641, + "Ġ['.": 95225, + "Ġ['./": 18201, + "Ġ['/": 73805, + "Ġ[(": 17826, + "Ġ[(\"": 84019, + "Ġ[('": 51128, + "Ġ[*": 60974, + "Ġ[+": 78801, + "Ġ[,": 78168, + "Ġ[-": 10055, + "Ġ[--": 73817, + "Ġ[-]:": 23086, + "Ġ[.": 48339, + "Ġ[...": 16303, + "Ġ[...]": 48179, + "Ġ[...]ĊĊ": 36979, + "Ġ[/": 64928, + "Ġ[:": 20865, + "Ġ[<": 67226, + "Ġ[=": 87076, + "Ġ[=[": 90297, + "Ġ[@": 73879, + "Ġ[[": 4318, + "Ġ[[\"": 78686, + "Ġ[['": 31827, + "Ġ[[[": 46727, + "Ġ[[]": 84507, + "Ġ[\\": 93815, + "Ġ[]": 3056, + "Ġ[](": 39444, + "Ġ[])": 42182, + "Ġ[]),Ċ": 96033, + "Ġ[]).": 61556, + "Ġ[]);Ċ": 42197, + "Ġ[]);ĊĊ": 41401, + "Ġ[])Ċ": 27614, + "Ġ[])ĊĊ": 61969, + "Ġ[]*": 29838, + "Ġ[],": 10071, + "Ġ[],Ċ": 10239, + "Ġ[],čĊ": 93227, + "Ġ[].": 72375, + "Ġ[]:Ċ": 96021, + "Ġ[];": 39371, + "Ġ[];Ċ": 5907, + "Ġ[];ĊĊ": 15436, + "Ġ[];čĊ": 34327, + "Ġ[][]": 52931, + "Ġ[]Ċ": 4167, + "Ġ[]ĊĊ": 14599, + "Ġ[]ĊĊĊ": 89238, + "Ġ[]čĊ": 23147, + "Ġ[]čĊčĊ": 79993, + "Ġ[_": 15133, + "Ġ[`": 40518, + "Ġ[{": 18466, + "Ġ[{\"": 61753, + "Ġ[{'": 61108, + "Ġ[{Ċ": 18396, + "Ġ[|": 81537, + "Ġ[~,": 93877, + "Ġ[â̦": 10205, + "Ġ[â̦]": 27524, + "Ġ[â̦]ĊĊ": 12706, + "Ġ[Ċ": 2278, + "Ġ[ĊĊ": 51308, + "Ġ[čĊ": 23637, + "Ġ\\": 1124, + "Ġ\\\"": 7245, + "Ġ\\\"\"": 34768, + "Ġ\\\"$": 65486, + "Ġ\\\"%": 32328, + "Ġ\\\"{": 77782, + "Ġ\\$": 32882, + "Ġ\\%": 96292, + "Ġ\\'": 27152, + "Ġ\\''": 86659, + "Ġ\\(": 17767, + "Ġ\\/": 36815, + "Ġ\\<": 4444, + "Ġ\\<^": 76889, + "Ġ\\@": 74080, + "Ġ\\\\": 24984, + "Ġ\\\\Ċ": 90155, + "Ġ\\`": 95547, + "Ġ\\|": 97155, + "Ġ\\Ċ": 3044, + "Ġ\\ĊĊ": 69911, + "Ġ\\čĊ": 21672, + "Ġ]": 2279, + "Ġ])": 41573, + "Ġ]),Ċ": 38408, + "Ġ])->": 91413, + "Ġ]).": 96711, + "Ġ]);": 76768, + "Ġ]);Ċ": 13201, + "Ġ]);ĊĊ": 22712, + "Ġ])Ċ": 23474, + "Ġ])ĊĊ": 56993, + "Ġ],": 10654, + "Ġ],Ċ": 3211, + "Ġ],ĊĊ": 39774, + "Ġ],čĊ": 34341, + "Ġ].": 20455, + "Ġ];": 13084, + "Ġ];Ċ": 9747, + "Ġ];ĊĊ": 15424, + "Ġ];čĊ": 65383, + "Ġ][": 35640, + "Ġ]]": 40596, + "Ġ]];": 27481, + "Ġ]]Ċ": 63040, + "Ġ]}Ċ": 68165, + "Ġ]Ċ": 5133, + "Ġ]ĊĊ": 10444, + "Ġ]ĊĊĊ": 83007, + "Ġ]čĊ": 48064, + "Ġ^": 6306, + "Ġ^(": 75097, + "Ġ^.": 97291, + "Ġ^=": 31602, + "Ġ^^": 57583, + "Ġ^{Ċ": 47423, + "Ġ^Ċ": 75396, + "Ġ_": 716, + "Ġ_$": 32186, + "Ġ_(": 52617, + "Ġ_(\"": 24816, + "Ġ_('": 35755, + "Ġ_)": 27439, + "Ġ_,": 8358, + "Ġ_.": 13194, + "Ġ_:": 58536, + "Ġ__": 1304, + "Ġ__(": 15617, + "Ġ__(\"": 55561, + "Ġ__('": 14414, + "Ġ___": 7436, + "Ġ____": 30743, + "Ġ_____": 65892, + "Ġ______": 32671, + "Ġ_______,": 56537, + "Ġ__________________": 43841, + "Ġ__________________________________": 81008, + "Ġ__________________ĊĊ": 79364, + "Ġ_|": 85137, + "Ġ_Ċ": 22983, + "Ġ_ĊĊ": 87951, + "Ġ_čĊ": 76564, + "Ġ`": 1565, + "Ġ`\"": 53305, + "Ġ`$": 55693, + "Ġ`${": 11518, + "Ġ`%": 61400, + "Ġ`'": 72911, + "Ġ`(": 48651, + "Ġ`,": 90190, + "Ġ`,Ċ": 63545, + "Ġ`-": 93797, + "Ġ`.": 74090, + "Ġ`/": 37301, + "Ġ`;Ċ": 64417, + "Ġ`<": 30586, + "Ġ`[": 77644, + "Ġ`\\": 91405, + "Ġ`_": 91621, + "Ġ``": 9902, + "Ġ``(": 39595, + "Ġ```": 54275, + "Ġ```Ċ": 41233, + "Ġ`{": 53692, + "Ġ`}Ċ": 91479, + "Ġ`Ċ": 22074, + "Ġa": 264, + "ĠaDecoder": 66050, + "ĠaValue": 89911, + "ĠaVar": 93662, + "Ġaa": 29009, + "Ġaaa": 83465, + "Ġaalborg": 75844, + "Ġaan": 18455, + "Ġaantal": 96511, + "Ġab": 668, + "Ġaba": 74791, + "Ġabaixo": 88792, + "Ġabandon": 15313, + "Ġabandoned": 22978, + "Ġabandoning": 77239, + "Ġabandonment": 90302, + "Ġabb": 29095, + "Ġabbiamo": 94296, + "Ġabbrev": 39515, + "Ġabbreviated": 94775, + "Ġabbreviation": 71478, + "Ġabc": 39022, + "Ġabdom": 35984, + "Ġabdomen": 63672, + "Ġabdominal": 55956, + "Ġabducted": 83445, + "Ġabduction": 96124, + "Ġaber": 19104, + "Ġaberr": 81002, + "Ġabi": 61050, + "Ġabide": 57199, + "Ġabilities": 17541, + "Ġability": 5726, + "Ġabl": 97392, + "Ġable": 2952, + "Ġabnormal": 34563, + "Ġabnormalities": 74715, + "Ġaboard": 36506, + "Ġabol": 36176, + "Ġabolish": 89276, + "Ġabolished": 79981, + "Ġabolition": 75345, + "Ġabort": 11326, + "Ġaborted": 58892, + "Ġabortion": 20107, + "Ġabortions": 53273, + "Ġabound": 86590, + "Ġabout": 911, + "Ġabove": 3403, + "Ġabras": 58581, + "Ġabrasive": 93704, + "Ġabre": 66341, + "Ġabril": 87287, + "Ġabrir": 72563, + "Ġabroad": 22131, + "Ġabrupt": 42977, + "Ġabruptly": 59745, + "Ġabs": 3647, + "Ġabsence": 19265, + "Ġabsent": 27211, + "Ġabsentee": 93090, + "Ġabsl": 88302, + "Ġabsol": 87561, + "Ġabsolut": 63162, + "Ġabsolute": 10740, + "Ġabsolutely": 10875, + "Ġabsor": 15567, + "Ġabsorb": 34306, + "Ġabsorbed": 41001, + "Ġabsorbing": 69175, + "Ġabsorbs": 90011, + "Ġabsorption": 43125, + "Ġabst": 62844, + "Ġabstract": 8115, + "Ġabstraction": 58751, + "Ġabsurd": 31577, + "Ġabund": 22509, + "Ġabundance": 36392, + "Ġabundant": 43511, + "Ġabuse": 11480, + "Ġabused": 35565, + "Ġabuses": 49062, + "Ġabusing": 68479, + "Ġabusive": 42558, + "Ġaby": 42715, + "Ġabyss": 87214, + "Ġac": 1613, + "Ġacab": 66062, + "Ġacad": 10406, + "Ġacademia": 66571, + "Ġacademic": 14250, + "Ġacademics": 47609, + "Ġacademy": 43345, + "Ġacc": 1029, + "Ġaccel": 66511, + "Ġacceler": 14177, + "Ġaccelerate": 42780, + "Ġaccelerated": 48758, + "Ġaccelerating": 68641, + "Ġacceleration": 30803, + "Ġaccelerator": 64356, + "Ġaccelerometer": 96748, + "Ġaccent": 29100, + "Ġaccents": 58470, + "Ġaccept": 4193, + "Ġacceptable": 21555, + "Ġacceptance": 25505, + "Ġaccepted": 11666, + "Ġaccepting": 24667, + "Ġaccepts": 26344, + "Ġacces": 84095, + "Ġacceso": 68529, + "Ġaccess": 2615, + "ĠaccessToken": 37725, + "Ġaccessed": 24758, + "Ġaccesses": 77648, + "Ġaccessibility": 39700, + "Ġaccessible": 15614, + "Ġaccessing": 31788, + "Ġaccession": 83945, + "Ġaccessor": 44384, + "Ġaccessories": 22293, + "Ġaccessory": 41981, + "Ġaccident": 11423, + "Ġaccidental": 48916, + "Ġaccidentally": 32384, + "Ġaccidents": 32688, + "Ġacciones": 93385, + "Ġacción": 89064, + "Ġacclaim": 96863, + "Ġacclaimed": 48982, + "Ġaccol": 85543, + "Ġaccom": 5725, + "Ġaccommod": 13927, + "Ġaccommodate": 28276, + "Ġaccommodating": 76041, + "Ġaccommodation": 27278, + "Ġaccommodations": 54613, + "Ġaccomp": 57333, + "Ġaccompagn": 87678, + "Ġaccompanied": 23943, + "Ġaccompanies": 48056, + "Ġaccompany": 19226, + "Ġaccompanying": 23534, + "Ġaccompl": 13089, + "Ġaccomplish": 22054, + "Ġaccomplished": 26237, + "Ġaccomplishment": 60138, + "Ġaccomplishments": 52781, + "Ġaccord": 14560, + "Ġaccordance": 18353, + "Ġaccording": 4092, + "Ġaccordingly": 27079, + "Ġaccordion": 86999, + "Ġaccount": 2692, + "ĠaccountId": 60443, + "Ġaccountability": 38142, + "Ġaccountable": 35877, + "Ġaccountant": 74921, + "Ġaccounted": 40753, + "Ġaccounting": 23168, + "Ġaccounts": 9618, + "Ġaccr": 85559, + "Ġaccred": 40471, + "Ġaccreditation": 83384, + "Ġaccredited": 54225, + "Ġaccru": 82732, + "Ġaccrued": 98116, + "Ġacct": 88567, + "Ġaccum": 15421, + "Ġaccumulate": 46276, + "Ġaccumulated": 40065, + "Ġaccumulating": 91585, + "Ġaccumulation": 45735, + "Ġaccumulator": 56312, + "Ġaccur": 7419, + "Ġaccuracy": 13403, + "Ġaccurate": 13382, + "Ġaccurately": 29257, + "Ġaccus": 25508, + "Ġaccusation": 64406, + "Ġaccusations": 35469, + "Ġaccuse": 61642, + "Ġaccused": 13185, + "Ġaccuses": 86056, + "Ġaccusing": 48479, + "Ġaccustomed": 52084, + "Ġace": 26747, + "Ġacept": 54888, + "Ġacess": 84970, + "Ġacesso": 83707, + "Ġacest": 58609, + "Ġacet": 64702, + "Ġach": 33261, + "Ġache": 77731, + "Ġacheter": 98390, + "Ġachie": 6087, + "Ġachievable": 88153, + "Ġachieve": 11075, + "Ġachieved": 16994, + "Ġachievement": 25426, + "Ġachievements": 32897, + "Ġachieves": 82591, + "Ġachieving": 31045, + "Ġachter": 86808, + "Ġacid": 13621, + "Ġacidic": 83803, + "Ġacidity": 96926, + "Ġacids": 32869, + "Ġack": 10725, + "Ġacknow": 11861, + "Ġacknowled": 15823, + "Ġacknowledge": 24645, + "Ġacknowledged": 25498, + "Ġacknowledgement": 66923, + "Ġacknowledges": 54930, + "Ġacknowledging": 60608, + "Ġacknowledgment": 77470, + "Ġacl": 57120, + "Ġacne": 45805, + "Ġacomp": 85701, + "Ġacompan": 51410, + "Ġacompanh": 60797, + "Ġacompañ": 92871, + "Ġacompaña": 137174, + "Ġaconte": 67050, + "Ġacordo": 76926, + "Ġacos": 80211, + "Ġacoustic": 44066, + "Ġacqu": 19274, + "Ġacquaint": 53527, + "Ġacquaintance": 81307, + "Ġacquainted": 72634, + "Ġacquire": 21256, + "Ġacquired": 18892, + "Ġacquiring": 41890, + "Ġacquisition": 23390, + "Ġacquisitions": 62848, + "Ġacquitted": 90433, + "Ġacre": 45223, + "Ġacres": 26367, + "Ġacronym": 74514, + "Ġacross": 3941, + "Ġacrylic": 58641, + "Ġact": 1160, + "Ġacted": 30432, + "Ġacting": 15358, + "Ġaction": 1917, + "ĠactionBar": 89136, + "ĠactionPerformed": 24059, + "ĠactionTypes": 67134, + "Ġactionable": 91078, + "Ġactions": 6168, + "Ġactiv": 4105, + "Ġactivate": 20276, + "Ġactivated": 21985, + "Ġactivates": 73942, + "Ġactivating": 71092, + "Ġactivation": 15099, + "Ġactivations": 92495, + "Ġactive": 4541, + "ĠactiveClassName": 98122, + "Ġactively": 22040, + "Ġactividad": 68740, + "Ġactividades": 72979, + "Ġactivism": 54180, + "Ġactivist": 27841, + "Ġactivists": 20902, + "Ġactivities": 7488, + "Ġactivity": 5702, + "Ġactivités": 137596, + "Ġactor": 12089, + "Ġactors": 19571, + "Ġactress": 23654, + "Ġactresses": 90028, + "Ġacts": 14056, + "Ġactu": 80453, + "Ġactual": 5042, + "Ġactualizar": 78629, + "Ġactually": 3520, + "Ġacuerdo": 62860, + "Ġacum": 68093, + "Ġacupuncture": 88438, + "Ġacute": 29783, + "Ġad": 993, + "Ġada": 34234, + "Ġadalah": 37560, + "Ġadam": 48966, + "Ġadamant": 83288, + "Ġadap": 38632, + "Ġadapt": 10515, + "Ġadaptable": 93748, + "Ġadaptation": 33085, + "Ġadaptations": 76665, + "Ġadapted": 29364, + "Ġadapter": 12956, + "Ġadapters": 69025, + "Ġadapting": 69717, + "Ġadaptive": 47132, + "Ġadaptor": 91941, + "Ġadapté": 140322, + "Ġadb": 82412, + "Ġadc": 47445, + "Ġadd": 912, + "ĠaddAction": 86010, + "ĠaddButton": 88196, + "ĠaddChild": 62959, + "ĠaddCriterion": 18724, + "ĠaddItem": 63928, + "ĠaddObject": 28723, + "ĠaddObserver": 76939, + "ĠaddSubview": 18171, + "ĠaddTarget": 49927, + "ĠaddTo": 53417, + "ĠaddUser": 77475, + "Ġadded": 3694, + "Ġaddict": 38226, + "Ġaddicted": 56627, + "Ġaddiction": 26373, + "Ġaddictive": 56307, + "Ġaddicts": 77474, + "Ġadding": 7842, + "Ġaddition": 5256, + "Ġadditional": 5107, + "Ġadditionally": 36838, + "Ġadditions": 37214, + "Ġadditive": 63238, + "Ġadditives": 84186, + "Ġaddon": 47997, + "Ġaddons": 89145, + "Ġaddr": 10789, + "Ġaddress": 2621, + "Ġaddressed": 20068, + "Ġaddresses": 14230, + "Ġaddressing": 27020, + "Ġadds": 11367, + "Ġaddslashes": 77307, + "Ġade": 17108, + "Ġadec": 86718, + "Ġadel": 71969, + "Ġademás": 62330, + "Ġaden": 99113, + "Ġadept": 75488, + "Ġadequ": 40128, + "Ġadequate": 25532, + "Ġadequately": 48572, + "Ġadher": 34951, + "Ġadhere": 48453, + "Ġadherence": 70528, + "Ġadhesive": 64458, + "Ġadicion": 80367, + "Ġadidas": 75190, + "Ġadip": 29266, + "Ġadipiscing": 57924, + "Ġadipisicing": 56681, + "Ġadj": 12469, + "Ġadjacency": 73849, + "Ġadjacent": 23942, + "Ġadjective": 84492, + "Ġadjoining": 90758, + "Ġadjud": 69180, + "Ġadjunct": 89595, + "Ġadjust": 7500, + "Ġadjustable": 35930, + "Ġadjusted": 23368, + "Ġadjusting": 42368, + "Ġadjustment": 27075, + "Ġadjustments": 32974, + "Ġadjusts": 83331, + "Ġadm": 14591, + "Ġadmin": 3986, + "Ġadminister": 26093, + "Ġadministered": 36918, + "Ġadministering": 85561, + "Ġadministr": 11150, + "Ġadministration": 8567, + "Ġadministrations": 77073, + "Ġadministrative": 22707, + "Ġadministrator": 28093, + "Ġadministrators": 37112, + "Ġadmins": 57095, + "Ġadmir": 62598, + "Ġadmirable": 84584, + "Ġadmiration": 65839, + "Ġadmire": 49763, + "Ġadmired": 63647, + "Ġadmission": 25293, + "Ġadmissions": 47811, + "Ġadmit": 16698, + "Ġadmits": 37139, + "Ġadmitted": 16188, + "Ġadmittedly": 78713, + "Ġadmitting": 54589, + "Ġadmon": 81985, + "Ġado": 55185, + "Ġadoles": 23109, + "Ġadolescence": 82809, + "Ġadolescent": 34988, + "Ġadolescente": 81935, + "Ġadolescents": 50416, + "Ġadopt": 10902, + "Ġadopted": 17827, + "Ġadopting": 47710, + "Ġadoption": 24376, + "Ġadopts": 85957, + "Ġadorable": 40608, + "Ġadore": 60635, + "Ġadorn": 57946, + "Ġadorned": 83173, + "Ġadr": 93133, + "Ġadrenal": 59464, + "Ġadrenaline": 78392, + "Ġadres": 65761, + "Ġadress": 62782, + "Ġadresse": 62174, + "Ġads": 13740, + "Ġadul": 39594, + "Ġadult": 6683, + "Ġadulte": 55330, + "Ġadultery": 95596, + "Ġadultes": 86925, + "Ġadulthood": 62933, + "Ġadulti": 77250, + "Ġadultos": 88631, + "Ġadults": 12598, + "Ġadv": 3358, + "Ġadvance": 11912, + "Ġadvanced": 10847, + "Ġadvancement": 49825, + "Ġadvancements": 82687, + "Ġadvances": 29903, + "Ġadvancing": 43069, + "Ġadvant": 7871, + "Ġadvantage": 9423, + "Ġadvantageous": 75724, + "Ġadvantages": 22146, + "Ġadvent": 11345, + "Ġadventure": 17943, + "Ġadventurer": 99135, + "Ġadventurers": 87713, + "Ġadventures": 30978, + "Ġadventurous": 67088, + "Ġadvers": 28411, + "Ġadversaries": 87391, + "Ġadversary": 81399, + "Ġadverse": 30859, + "Ġadversely": 68114, + "Ġadversity": 89830, + "Ġadvert": 7690, + "Ġadvertis": 19220, + "Ġadvertise": 50836, + "Ġadvertised": 45741, + "Ġadvertisement": 32689, + "Ġadvertisements": 42649, + "Ġadvertiser": 86286, + "Ġadvertisers": 49170, + "Ġadvertising": 12876, + "Ġadverts": 87725, + "Ġadvice": 9462, + "Ġadvis": 9622, + "Ġadvisable": 68819, + "Ġadvise": 31013, + "Ġadvised": 25104, + "Ġadviser": 31753, + "Ġadvisers": 50975, + "Ġadvises": 68974, + "Ġadvising": 62679, + "Ġadvisor": 36613, + "Ġadvisors": 57684, + "Ġadvisory": 40437, + "Ġadvoc": 12237, + "Ġadvocacy": 34854, + "Ġadvocate": 28015, + "Ġadvocated": 63754, + "Ġadvocates": 27324, + "Ġadvocating": 58316, + "Ġadım": 135081, + "Ġadına": 136880, + "Ġae": 41856, + "Ġaer": 33448, + "Ġaerial": 40784, + "Ġaerobic": 90390, + "Ġaeros": 48725, + "Ġaerospace": 77064, + "Ġaes": 56575, + "Ġaest": 26816, + "Ġaesthetic": 36537, + "Ġaesthetics": 66223, + "Ġaf": 7977, + "Ġafar": 98278, + "Ġafect": 67264, + "Ġaff": 3527, + "Ġaffair": 29917, + "Ġaffairs": 21978, + "Ġaffect": 7802, + "Ġaffected": 11495, + "Ġaffecting": 27887, + "Ġaffection": 36705, + "Ġaffects": 21501, + "Ġaffid": 49782, + "Ġaffidavit": 59935, + "Ġaffili": 11938, + "Ġaffiliate": 21597, + "Ġaffiliated": 36406, + "Ġaffiliates": 31033, + "Ġaffiliation": 52791, + "Ġaffine": 68809, + "Ġaffinity": 50452, + "Ġaffirm": 32349, + "Ġaffirmation": 95863, + "Ġaffirmative": 67641, + "Ġaffirmed": 73449, + "Ġafflict": 94523, + "Ġafflicted": 91168, + "Ġaffluent": 80779, + "Ġafford": 9946, + "Ġaffordability": 77910, + "Ġaffordable": 16637, + "Ġafforded": 65599, + "Ġaficion": 93805, + "Ġafin": 53301, + "Ġafirm": 85438, + "Ġafl": 98057, + "Ġafore": 43773, + "Ġaforementioned": 45652, + "Ġafr": 84964, + "Ġafraid": 16575, + "Ġafrica": 44281, + "Ġafrican": 74400, + "Ġafs": 66713, + "Ġaft": 59701, + "Ġafter": 1283, + "ĠafterEach": 73491, + "Ġaftermarket": 89892, + "Ġaftermath": 39596, + "Ġafternoon": 13354, + "Ġafterward": 47943, + "Ġafterwards": 26807, + "Ġag": 933, + "Ġagain": 1549, + "Ġagainst": 2348, + "Ġagar": 69379, + "Ġage": 4231, + "Ġaged": 19749, + "Ġageing": 78943, + "Ġagencies": 13304, + "Ġagency": 9088, + "Ġagenda": 18402, + "Ġagendas": 82929, + "Ġagent": 8315, + "Ġagents": 13009, + "Ġages": 16639, + "Ġagg": 50020, + "Ġaggi": 68369, + "Ġaggrav": 44941, + "Ġaggravated": 62268, + "Ġaggreg": 25198, + "Ġaggregate": 23192, + "Ġaggregated": 70822, + "Ġaggregates": 70543, + "Ġaggregation": 51629, + "Ġaggregator": 96680, + "Ġaggress": 13329, + "Ġaggression": 39207, + "Ġaggressive": 19188, + "Ġaggressively": 48039, + "Ġagile": 61465, + "Ġagility": 66641, + "Ġaging": 28984, + "Ġagitation": 90217, + "Ġagli": 98682, + "Ġago": 4134, + "Ġagon": 56670, + "Ġagony": 78830, + "Ġagora": 65783, + "Ġagosto": 82189, + "Ġagr": 39474, + "Ġagre": 4294, + "Ġagree": 7503, + "Ġagreed": 7230, + "Ġagreeing": 38327, + "Ġagreement": 9128, + "Ġagreements": 19473, + "Ġagrees": 32908, + "Ġagreg": 78194, + "Ġagregar": 72106, + "Ġagricult": 17295, + "Ġagricultural": 28049, + "Ġagriculture": 28929, + "Ġagré": 141521, + "Ġagréable": 141522, + "Ġagua": 55462, + "Ġah": 22327, + "Ġahead": 8305, + "Ġahora": 46317, + "Ġai": 16391, + "Ġaid": 12296, + "Ġaide": 40575, + "Ġaided": 62857, + "Ġaider": 90778, + "Ġaides": 52889, + "Ġaiding": 85287, + "Ġaids": 51697, + "Ġail": 59143, + "Ġailments": 81807, + "Ġaim": 9213, + "Ġaime": 85707, + "Ġaimed": 19469, + "Ġaiming": 37078, + "Ġaims": 21538, + "Ġain": 36102, + "Ġainda": 43909, + "Ġainsi": 43251, + "Ġaio": 83943, + "Ġair": 3720, + "Ġairborne": 69763, + "Ġaircraft": 14135, + "Ġaire": 84108, + "Ġaired": 41987, + "Ġairflow": 73020, + "Ġairing": 70253, + "Ġairl": 82657, + "Ġairline": 32475, + "Ġairlines": 47358, + "Ġairplane": 42924, + "Ġairplanes": 75381, + "Ġairport": 16733, + "Ġairports": 45063, + "Ġairs": 78891, + "Ġairspace": 87182, + "Ġairst": 58310, + "Ġairstrikes": 79273, + "Ġairy": 96963, + "Ġais": 88695, + "Ġaisle": 59679, + "Ġaj": 25437, + "Ġajax": 26281, + "Ġajout": 88996, + "Ġajud": 86059, + "Ġajust": 73134, + "Ġak": 17324, + "Ġaka": 37141, + "Ġakan": 32670, + "Ġakin": 51907, + "Ġakka": 51342, + "Ġako": 78714, + "Ġakt": 22318, + "Ġaktiv": 49183, + "Ġaktual": 84301, + "Ġaktuellen": 97268, + "Ġaku": 86195, + "ĠakÅŁ": 140245, + "ĠakÅŁam": 140246, + "Ġal": 452, + "Ġalan": 72562, + "Ġalanı": 139111, + "Ġalanında": 140379, + "Ġalarm": 16624, + "Ġalarmed": 81040, + "Ġalarming": 56276, + "Ġalarms": 62311, + "Ġalas": 88098, + "Ġalb": 81598, + "Ġalbeit": 42069, + "Ġalbum": 8017, + "Ġalbums": 27685, + "Ġalc": 41624, + "Ġalcan": 75771, + "Ġalcanç": 142913, + "Ġalcohol": 12904, + "Ġalcoholic": 55361, + "Ġalcuni": 96356, + "Ġald": 80304, + "Ġaldı": 132984, + "Ġaldıģı": 136390, + "Ġale": 21463, + "Ġalert": 5115, + "ĠalertController": 72681, + "ĠalertDialog": 62557, + "Ġalerted": 66669, + "Ġalerts": 29250, + "Ġalex": 56478, + "Ġalf": 84666, + "Ġalg": 17345, + "Ġalgae": 67851, + "Ġalgebra": 46876, + "Ġalgo": 27928, + "Ġalgorithm": 12111, + "Ġalgorithms": 25185, + "Ġalgu": 140030, + "Ġalguien": 88319, + "Ġalgum": 96675, + "Ġalgumas": 92012, + "Ġalgun": 36471, + "Ġalguna": 72027, + "Ġalgunas": 73754, + "Ġalgunos": 57210, + "Ġalguns": 77269, + "Ġalguém": 140031, + "Ġalgún": 84309, + "Ġali": 18991, + "Ġalias": 15534, + "Ġaliases": 40386, + "Ġalice": 70433, + "Ġalien": 19594, + "Ġaliens": 36119, + "Ġalign": 5285, + "ĠalignItems": 21716, + "ĠalignSelf": 67578, + "Ġaligned": 26118, + "Ġalignment": 17189, + "Ġalignments": 92816, + "Ġalike": 25992, + "Ġaliment": 37350, + "Ġalimentos": 89491, + "Ġaliqu": 42291, + "Ġaliqua": 85927, + "Ġalist": 95695, + "Ġalive": 13675, + "Ġalk": 57898, + "Ġalkal": 65687, + "Ġall": 678, + "Ġalla": 22702, + "Ġalle": 12304, + "Ġalleen": 71173, + "Ġalleg": 7034, + "Ġallegation": 57394, + "Ġallegations": 18367, + "Ġallege": 82617, + "Ġalleged": 10491, + "Ġallegedly": 19204, + "Ġalleges": 58697, + "Ġallegiance": 66978, + "Ġalleging": 61446, + "Ġallele": 69410, + "Ġalleles": 97160, + "Ġallem": 60297, + "Ġallen": 50041, + "Ġaller": 47808, + "Ġallerdings": 85725, + "Ġallerg": 24401, + "Ġallergic": 56496, + "Ġallergies": 57642, + "Ġallergy": 59654, + "Ġalles": 35951, + "Ġallev": 45549, + "Ġalleviate": 60605, + "Ġalley": 52950, + "Ġalliance": 29664, + "Ġalliances": 69700, + "Ġallied": 53402, + "Ġallies": 20120, + "Ġallo": 62647, + "Ġalloc": 5574, + "Ġallocate": 22089, + "Ġallocated": 20204, + "Ġallocating": 72762, + "Ġallocation": 23757, + "Ġallocations": 69642, + "Ġallocator": 43655, + "Ġallot": 84174, + "Ġallotted": 91543, + "Ġallow": 2138, + "ĠallowNull": 41440, + "Ġallowable": 84752, + "Ġallowance": 40358, + "Ġallowances": 89020, + "Ġallowed": 5420, + "Ġallowing": 10693, + "Ġallows": 6147, + "Ġalloy": 45864, + "Ġalloys": 94751, + "Ġallure": 87403, + "Ġally": 30853, + "Ġalm": 89484, + "Ġalma": 81480, + "Ġalmacen": 83227, + "Ġalmond": 62068, + "Ġalmonds": 87822, + "Ġalmost": 4558, + "Ġalo": 53166, + "Ġaload": 49620, + "Ġalone": 7484, + "Ġalong": 3156, + "Ġalongside": 16263, + "Ġalors": 44475, + "Ġalot": 56238, + "Ġaloud": 70411, + "Ġalph": 64595, + "Ġalpha": 8287, + "Ġalphabet": 27790, + "Ġalphabetical": 83298, + "Ġalphanumeric": 98979, + "Ġalready": 2669, + "Ġalright": 50117, + "Ġals": 10712, + "Ġalso": 1083, + "Ġalt": 4797, + "Ġalta": 45362, + "Ġaltar": 50843, + "Ġalte": 43908, + "Ġalter": 11596, + "Ġalteration": 72581, + "Ġalterations": 59986, + "Ġaltercation": 97687, + "Ġaltered": 28752, + "Ġaltering": 59823, + "Ġaltern": 6919, + "Ġalternate": 24609, + "Ġalternating": 72362, + "Ġalternative": 10555, + "Ġalternatively": 68387, + "Ġalternatives": 26450, + "Ġalters": 87587, + "Ġalthough": 7892, + "Ġaltijd": 87769, + "Ġaltitude": 35858, + "Ġalto": 23299, + "Ġaltogether": 30055, + "Ġaltre": 93560, + "Ġaltri": 66168, + "Ġaltro": 91917, + "Ġaltru": 92492, + "Ġaltura": 59252, + "Ġaltın": 133518, + "Ġaltında": 131564, + "Ġalum": 32960, + "Ġaluminium": 54893, + "Ġaluminum": 24649, + "Ġalumni": 51021, + "Ġalumno": 98095, + "Ġalumnos": 97859, + "Ġaluno": 76260, + "Ġalunos": 93144, + "Ġalways": 2677, + "Ġalém": 83669, + "Ġalın": 128925, + "ĠalÄ±ÅŁ": 141863, + "ĠalÄ±ÅŁver": 141864, + "ĠalÄ±ÅŁveriÅŁ": 141865, + "Ġam": 1079, + "Ġama": 70762, + "Ġamacı": 128092, + "Ġamacıyla": 139854, + "Ġamalg": 89765, + "Ġamassed": 92150, + "Ġamat": 42133, + "Ġamateur": 18566, + "Ġamateurs": 75550, + "Ġamatør": 72528, + "Ġamazed": 45204, + "Ġamazing": 7897, + "Ġamazingly": 62515, + "Ġamazon": 38416, + "Ġamb": 8873, + "Ġambassador": 36656, + "Ġambassadors": 95562, + "Ġamber": 67605, + "Ġambiance": 94573, + "Ġambient": 34188, + "Ġambiente": 66413, + "Ġambigu": 61297, + "Ġambiguity": 71768, + "Ġambiguous": 54761, + "Ġambit": 29398, + "Ġambition": 44615, + "Ġambitions": 50466, + "Ġambitious": 31755, + "Ġambos": 90693, + "Ġambulance": 50532, + "Ġambush": 78965, + "Ġamd": 75562, + "Ġamen": 29119, + "Ġamend": 29469, + "Ġamended": 29911, + "Ġamendment": 27139, + "Ġamendments": 40593, + "Ġamenities": 35383, + "Ġamer": 26720, + "Ġameric": 65779, + "Ġamerica": 63666, + "Ġamerican": 38442, + "Ġamet": 27212, + "Ġami": 51727, + "Ġamid": 22617, + "Ġamidst": 64804, + "Ġamigo": 70211, + "Ġamigos": 56557, + "Ġamino": 41400, + "Ġamis": 86792, + "Ġammo": 37904, + "Ġammon": 68811, + "Ġammonia": 90903, + "Ġammunition": 36668, + "Ġamnesty": 85135, + "Ġamo": 86100, + "Ġamong": 4221, + "Ġamongst": 23183, + "Ġamor": 44165, + "Ġamore": 71275, + "Ġamort": 92264, + "Ġamount": 3311, + "Ġamounted": 73219, + "Ġamounts": 14713, + "Ġamour": 98153, + "Ġamp": 22758, + "Ġamph": 64992, + "Ġamphib": 87922, + "Ġampl": 22395, + "Ġample": 41753, + "Ġamplified": 82498, + "Ġamplifier": 60066, + "Ġamplify": 96068, + "Ġamplitude": 44109, + "Ġamps": 79962, + "Ġamsterdam": 55939, + "Ġamt": 61169, + "Ġamused": 85224, + "Ġamusement": 61368, + "Ġamusing": 59886, + "Ġamy": 63283, + "Ġaméric": 137391, + "Ġaméricain": 137392, + "Ġan": 458, + "Ġana": 40290, + "Ġanak": 83255, + "Ġanal": 3181, + "Ġanale": 78237, + "Ġanalog": 23400, + "Ġanalogous": 78183, + "Ġanalogue": 90243, + "Ġanalogy": 55103, + "Ġanalsex": 89832, + "Ġanaly": 8510, + "Ġanalys": 21488, + "Ġanalyse": 48486, + "Ġanalysed": 66358, + "Ġanalyses": 27960, + "Ġanalysis": 6358, + "Ġanalyst": 18237, + "Ġanalysts": 30399, + "Ġanalytic": 78036, + "Ġanalytical": 42964, + "Ġanalytics": 27875, + "Ġanalyze": 23643, + "Ġanalyzed": 29139, + "Ġanalyzer": 54660, + "Ġanalyzes": 85181, + "Ġanalyzing": 41018, + "Ġanarch": 43465, + "Ġanarchist": 77331, + "Ġanarchists": 92034, + "Ġanatom": 74793, + "Ġanatomy": 61590, + "Ġanc": 45745, + "Ġance": 43160, + "Ġancest": 25918, + "Ġancestor": 45731, + "Ġancestors": 37518, + "Ġancestral": 77671, + "Ġancestry": 64904, + "Ġanch": 33230, + "Ġanche": 28057, + "Ġanchor": 17105, + "Ġanchored": 77119, + "Ġanchors": 55439, + "Ġancient": 13833, + "Ġancor": 60891, + "Ġancora": 63414, + "Ġand": 323, + "Ġanda": 44001, + "Ġander": 61805, + "Ġandere": 32936, + "Ġanderen": 38684, + "Ġanders": 66653, + "Ġandra": 89814, + "Ġandre": 59672, + "Ġandroid": 2103, + "Ġandroidx": 9527, + "Ġanecd": 49650, + "Ġanecdotes": 92966, + "Ġanesthesia": 90806, + "Ġanew": 92292, + "Ġang": 6454, + "Ġange": 32175, + "Ġangel": 33326, + "Ġangels": 45868, + "Ġanger": 19234, + "Ġangered": 86494, + "Ġangi": 91862, + "Ġangl": 55304, + "Ġanglais": 75085, + "Ġangle": 9210, + "Ġangled": 98310, + "Ġangles": 25941, + "Ġangrily": 86525, + "Ġangry": 18514, + "Ġangst": 90319, + "Ġanguish": 94922, + "Ġangular": 20314, + "Ġani": 51988, + "Ġanim": 3952, + "Ġanimal": 9864, + "Ġanimals": 9898, + "Ġanimate": 25063, + "ĠanimateWithDuration": 81456, + "Ġanimated": 11371, + "Ġanimation": 10355, + "Ġanimations": 26053, + "Ġanimator": 39044, + "Ġanime": 22809, + "Ġank": 70472, + "Ġankle": 38348, + "Ġankles": 89613, + "Ġanlam": 137511, + "Ġanlamı": 137512, + "ĠanlaÅŁ": 141596, + "ĠanlaÅŁma": 141168, + "ĠanlaÅŁÄ±l": 141597, + "Ġanmeld": 61061, + "Ġann": 2935, + "Ġanne": 85897, + "Ġannex": 53268, + "Ġanni": 44574, + "Ġannihil": 97345, + "Ġanniversary": 21582, + "Ġanno": 18344, + "Ġannon": 62891, + "Ġannonce": 45220, + "Ġannonces": 75739, + "Ġannot": 36042, + "Ġannotate": 88398, + "Ġannotated": 62851, + "Ġannotation": 21223, + "Ġannotations": 32207, + "Ġannounc": 5353, + "Ġannounce": 21483, + "Ġannounced": 7228, + "Ġannouncement": 17043, + "Ġannouncements": 44876, + "Ġannounces": 47682, + "Ġannouncing": 37687, + "Ġannoy": 37192, + "Ġannoyance": 90804, + "Ġannoyed": 56030, + "Ġannoying": 29831, + "Ġannual": 9775, + "Ġannually": 29071, + "Ġannum": 86657, + "Ġannunci": 28885, + "Ġannée": 75032, + "Ġannées": 64738, + "Ġano": 38024, + "Ġanom": 36682, + "Ġanomal": 69573, + "Ġanomalies": 74459, + "Ġanomaly": 62948, + "Ġanon": 74812, + "Ġanonym": 31148, + "Ġanonymity": 48512, + "Ġanonymous": 22151, + "Ġanonymously": 72151, + "Ġanos": 37001, + "Ġanother": 2441, + "Ġans": 8099, + "Ġansch": 93644, + "Ġansible": 90719, + "Ġanswer": 4226, + "Ġanswered": 18577, + "Ġanswering": 35764, + "Ġanswers": 11253, + "Ġant": 3196, + "Ġantagon": 42115, + "Ġantagonist": 81059, + "Ġante": 39424, + "Ġantenn": 71477, + "Ġantenna": 39932, + "Ġantennas": 86751, + "Ġanterior": 36129, + "Ġantes": 33335, + "Ġanth": 22268, + "Ġanthem": 55564, + "Ġanthology": 83008, + "Ġanthrop": 40316, + "Ġanthropology": 93944, + "Ġanti": 7147, + "Ġantib": 20554, + "Ġantibiot": 32109, + "Ġantibiotic": 59495, + "Ġantibiotics": 45750, + "Ġantibodies": 58754, + "Ġantibody": 61952, + "Ġantic": 87043, + "Ġanticip": 16552, + "Ġanticipate": 47148, + "Ġanticipated": 29099, + "Ġanticipating": 85259, + "Ġanticipation": 49819, + "Ġantics": 87123, + "Ġantid": 89587, + "Ġantidepress": 64111, + "Ġantig": 67837, + "Ġantigen": 81989, + "Ġantim": 78212, + "Ġantioxid": 45712, + "Ġantioxidant": 69239, + "Ġantioxidants": 80015, + "Ġantiqu": 60286, + "Ġantique": 46591, + "Ġantis": 64517, + "Ġantivirus": 79549, + "Ġantlr": 85269, + "Ġants": 60395, + "Ġanunci": 90833, + "Ġanus": 81308, + "Ġanv": 59481, + "Ġanvänd": 83948, + "Ġanx": 73784, + "Ġanxiety": 18056, + "Ġanxious": 37000, + "Ġany": 894, + "Ġanybody": 21061, + "Ġanyhow": 88964, + "Ġanymore": 14584, + "Ġanyone": 5489, + "Ġanything": 4113, + "Ġanytime": 29094, + "Ġanyway": 13657, + "Ġanyways": 62345, + "Ġanywhere": 12379, + "Ġanz": 77868, + "Ġanzeigen": 83797, + "Ġanál": 71290, + "Ġanálise": 138482, + "Ġanálisis": 141108, + "Ġao": 14845, + "Ġaos": 42814, + "Ġaoût": 137228, + "Ġap": 1443, + "Ġapa": 60593, + "Ġapache": 73406, + "Ġapar": 57216, + "Ġapare": 55170, + "Ġapart": 10747, + "Ġapartheid": 91959, + "Ġapartment": 13154, + "Ġapartments": 31629, + "Ġape": 84880, + "Ġapellido": 63885, + "Ġapenas": 46667, + "Ġaperture": 57001, + "Ġapex": 71527, + "Ġaph": 88011, + "Ġapi": 6330, + "ĠapiKey": 44924, + "ĠapiUrl": 88231, + "Ġapis": 97723, + "Ġapk": 58033, + "Ġapl": 52920, + "Ġaplic": 29440, + "Ġaplicación": 74327, + "Ġaplik": 76245, + "Ġapo": 72890, + "Ġapocalypse": 87814, + "Ġapolog": 20419, + "Ġapologies": 72173, + "Ġapologise": 81385, + "Ġapologize": 36879, + "Ġapologized": 56977, + "Ġapology": 40158, + "Ġapopt": 82946, + "Ġapoptosis": 94774, + "Ġapost": 38471, + "Ġapp": 906, + "ĠappBar": 42287, + "ĠappDelegate": 96255, + "ĠappId": 50556, + "ĠappName": 67819, + "Ġappalling": 84805, + "Ġappar": 33288, + "Ġapparatus": 40605, + "Ġappare": 96162, + "Ġapparel": 54325, + "Ġapparent": 9981, + "Ġapparently": 13812, + "Ġappart": 73656, + "Ġappe": 2642, + "Ġappeal": 14303, + "Ġappealed": 52282, + "Ġappealing": 32252, + "Ġappeals": 34630, + "Ġappear": 4994, + "Ġappearance": 11094, + "Ġappearances": 26255, + "Ġappeared": 9723, + "Ġappearing": 25377, + "Ġappears": 7952, + "Ġappel": 75090, + "Ġappell": 51951, + "Ġappellant": 89100, + "Ġappellate": 97487, + "Ġappend": 8737, + "ĠappendString": 96340, + "Ġappended": 51187, + "Ġappending": 93283, + "Ġappendix": 87577, + "Ġappet": 29068, + "Ġappetite": 37475, + "Ġappl": 17099, + "Ġapplaud": 51390, + "Ġapplauded": 95628, + "Ġapplause": 63526, + "Ġapple": 23268, + "Ġapples": 40676, + "Ġappliance": 55354, + "Ġappliances": 33702, + "Ġapplic": 4563, + "Ġapplicable": 8415, + "Ġapplicant": 31268, + "Ġapplicants": 31077, + "Ġapplication": 3766, + "ĠapplicationContext": 85299, + "ĠapplicationWill": 92367, + "Ġapplications": 8357, + "Ġapplied": 9251, + "Ġapplies": 16790, + "Ġapply": 3796, + "ĠapplyMiddleware": 83364, + "Ġapplying": 18950, + "Ġappoint": 9540, + "Ġappointed": 20822, + "Ġappointment": 17635, + "Ġappointments": 36156, + "Ġappraisal": 78292, + "Ġapprec": 9790, + "Ġappreciate": 15401, + "Ġappreciated": 25808, + "Ġappreciation": 34896, + "Ġappreh": 46191, + "Ġapprent": 45330, + "Ġapprentice": 82199, + "Ġapprentices": 71251, + "Ġappro": 2984, + "Ġapproach": 5486, + "Ġapproached": 24706, + "Ġapproaches": 19827, + "Ġapproaching": 29947, + "Ġappropri": 7185, + "Ġappropriate": 8311, + "Ġappropriated": 86413, + "Ġappropriately": 34901, + "Ġappropriation": 85640, + "Ġappropriations": 92566, + "Ġapproval": 14427, + "Ġapprovals": 82823, + "Ġapprove": 28151, + "Ġapproved": 11792, + "Ġapproves": 93830, + "Ġapproving": 76158, + "Ġapprox": 9848, + "Ġapproximate": 44868, + "Ġapproximately": 13187, + "Ġapproximation": 56204, + "Ġappré": 141758, + "Ġappréci": 141759, + "Ġapps": 10500, + "Ġapr": 21049, + "Ġaprend": 67346, + "Ġaprender": 92717, + "Ġapresent": 54762, + "Ġapril": 73700, + "Ġapro": 75855, + "Ġaprove": 87801, + "Ġaprox": 72233, + "Ġaproxim": 94424, + "Ġaprès": 41038, + "Ġapt": 20095, + "Ġapós": 75081, + "Ġaqu": 15355, + "Ġaquarium": 61183, + "Ġaquatic": 71391, + "Ġaque": 65200, + "Ġaquel": 78276, + "Ġaqui": 43576, + "ĠaquÃŃ": 53089, + "Ġar": 796, + "Ġara": 127238, + "Ġarab": 51312, + "Ġaracı": 136296, + "Ġaras": 126448, + "Ġarasında": 129010, + "Ġarasındaki": 138451, + "Ġaraç": 132048, + "ĠaraÅŁtırma": 134349, + "Ġarb": 58488, + "Ġarbe": 80044, + "Ġarbit": 16699, + "Ġarbitr": 57957, + "Ġarbitrarily": 85657, + "Ġarbitrary": 24168, + "Ġarbitration": 53608, + "Ġarc": 15580, + "Ġarcade": 47212, + "Ġarcane": 89353, + "Ġarch": 5325, + "Ġarchae": 38111, + "Ġarchaeological": 68537, + "Ġarchetype": 85157, + "Ġarchitect": 11470, + "Ġarchitects": 54935, + "Ġarchitectural": 42463, + "Ġarchitecture": 17646, + "Ġarchitectures": 77235, + "Ġarchival": 93947, + "Ġarchive": 18132, + "Ġarchived": 51993, + "Ġarchives": 38004, + "Ġarchivo": 29718, + "Ġarchivos": 87395, + "Ġarcpy": 91893, + "Ġarcs": 66750, + "Ġard": 83909, + "Ġardından": 131985, + "Ġare": 525, + "Ġarea": 3082, + "Ġareas": 5671, + "Ġaren": 7629, + "Ġarena": 24902, + "Ġarenas": 97667, + "Ġarg": 1392, + "Ġargc": 11844, + "Ġargent": 88144, + "Ġargparse": 22321, + "Ġargs": 2827, + "Ġarguably": 35559, + "Ġargue": 17585, + "Ġargued": 18280, + "Ġargues": 28533, + "Ġarguing": 29574, + "Ġargument": 5693, + "Ġarguments": 5977, + "Ġargv": 10213, + "Ġaria": 7131, + "Ġarise": 30789, + "Ġarisen": 89854, + "Ġarises": 47182, + "Ġarising": 39886, + "Ġarist": 63738, + "Ġarithmetic": 34784, + "Ġarity": 79932, + "Ġark": 55217, + "ĠarkadaÅŁ": 129255, + "Ġarm": 6773, + "Ġarma": 51605, + "Ġarmed": 17445, + "Ġarmies": 46883, + "Ġarmor": 20033, + "Ġarmored": 64645, + "Ġarmour": 46677, + "Ġarms": 11715, + "Ġarmy": 13390, + "Ġarom": 31269, + "Ġaroma": 54179, + "Ġaromatic": 81588, + "Ġarose": 49963, + "Ġarou": 65108, + "Ġaround": 2163, + "Ġarous": 74023, + "Ġarousal": 86817, + "Ġaroused": 91524, + "Ġarp": 98598, + "Ġarquivo": 43405, + "Ġarr": 2890, + "Ġarrang": 16700, + "Ġarrange": 30893, + "Ġarranged": 27802, + "Ġarrangement": 26112, + "Ġarrangements": 27804, + "Ġarranging": 76118, + "Ġarray": 1334, + "ĠarrayList": 47400, + "ĠarrayOf": 43254, + "ĠarrayWith": 61152, + "Ġarrays": 18386, + "Ġarreglo": 80581, + "Ġarrest": 8004, + "Ġarrested": 12517, + "Ġarresting": 83766, + "Ġarrests": 37711, + "Ġarriv": 67903, + "Ġarrival": 18647, + "Ġarrivals": 69047, + "Ġarrive": 17331, + "Ġarrived": 11465, + "Ġarrives": 29682, + "Ġarriving": 32011, + "Ġarrog": 43736, + "Ġarrogance": 82743, + "Ġarrogant": 65368, + "Ġarrow": 17921, + "Ġarrows": 36957, + "Ġarsch": 93236, + "Ġarsen": 86264, + "Ġarsenal": 56931, + "Ġarson": 86750, + "Ġart": 1947, + "Ġarte": 51348, + "Ġarter": 46257, + "Ġarterial": 79835, + "Ġarteries": 84548, + "Ġartery": 64315, + "Ġarthritis": 54552, + "Ġartic": 27564, + "Ġarticle": 4549, + "Ġarticles": 9709, + "Ġarticulate": 68218, + "Ġarticulated": 82180, + "Ġartifact": 36639, + "Ġartifacts": 35036, + "Ġartificial": 20443, + "Ġartificially": 77120, + "Ġartikel": 88849, + "Ġartillery": 53135, + "Ġartisan": 66119, + "Ġartisans": 99030, + "Ġartist": 10049, + "Ġartistic": 31592, + "Ġartists": 13511, + "Ġarts": 18560, + "Ġarttır": 141653, + "Ġartwork": 28309, + "Ġartworks": 96449, + "Ġarty": 135024, + "Ġartyku": 135025, + "ĠartykuÅĤ": 135026, + "ĠartÃŃculo": 79589, + "Ġartık": 131818, + "Ġartır": 137080, + "ĠartÄ±ÅŁ": 140542, + "Ġary": 72921, + "Ġas": 438, + "Ġasbestos": 76135, + "Ġasc": 14601, + "Ġascend": 75312, + "Ġascending": 35388, + "Ġascent": 74989, + "Ġascertain": 76177, + "Ġascii": 47120, + "Ġaseg": 64744, + "Ġash": 23828, + "Ġashamed": 50875, + "Ġashes": 66001, + "Ġasi": 46631, + "Ġasia": 83499, + "Ġasian": 35625, + "Ġasiat": 73848, + "Ġaside": 15663, + "Ġasign": 60079, + "Ġask": 2548, + "Ġasked": 4588, + "Ġasking": 10161, + "Ġasks": 17064, + "Ġasleep": 32073, + "Ġaslı": 138714, + "Ġaslında": 138715, + "Ġasm": 32828, + "Ġasn": 75252, + "Ġasoci": 68996, + "Ġaspect": 12893, + "Ġaspects": 13566, + "Ġasphalt": 45262, + "Ġaspir": 24663, + "Ġaspiration": 97641, + "Ġaspirations": 57422, + "Ġaspire": 83511, + "Ġaspiring": 59113, + "Ġass": 1071, + "Ġassail": 78688, + "Ġassass": 25732, + "Ġassassin": 67669, + "Ġassassination": 48317, + "Ġassault": 11703, + "Ġassaulted": 45571, + "Ġassaulting": 78226, + "Ġassaults": 56819, + "Ġassay": 63933, + "Ġassays": 98492, + "Ġassemble": 41740, + "Ġassembled": 34005, + "Ġassembler": 88524, + "Ġassemblies": 61307, + "Ġassembling": 80659, + "Ġassembly": 14614, + "Ġassert": 2060, + "ĠassertEquals": 10037, + "ĠassertFalse": 44397, + "ĠassertNotNull": 44328, + "ĠassertNull": 72609, + "ĠassertThat": 16981, + "ĠassertTrue": 21627, + "Ġasserted": 49597, + "Ġasserting": 42425, + "Ġassertion": 27419, + "Ġassertions": 54836, + "Ġasserts": 56776, + "Ġasses": 79040, + "Ġassess": 8552, + "Ġassessed": 31348, + "Ġassessing": 46514, + "Ġassessment": 15449, + "Ġassessments": 40200, + "Ġasset": 9329, + "Ġassets": 11770, + "Ġassez": 78280, + "Ġassh": 48552, + "Ġasshole": 58863, + "Ġassign": 9793, + "Ġassignable": 89447, + "Ġassigned": 12607, + "Ġassigning": 60753, + "Ġassignment": 16319, + "Ġassignments": 31172, + "Ġassigns": 49912, + "Ġassim": 38954, + "Ġassist": 7789, + "Ġassistance": 12994, + "Ġassistant": 17847, + "Ġassistants": 56519, + "Ġassisted": 37961, + "Ġassisting": 45827, + "Ġassistir": 88515, + "Ġassists": 28844, + "Ġassms": 41071, + "Ġassoc": 59607, + "Ġassoci": 4097, + "Ġassociate": 21946, + "Ġassociated": 5815, + "Ġassociates": 39431, + "Ġassociation": 15010, + "Ġassociations": 29157, + "Ġassociative": 68903, + "Ġassort": 49934, + "Ġassorted": 83442, + "Ġassortment": 55295, + "Ġassum": 7736, + "Ġassume": 9658, + "Ġassumed": 19110, + "Ġassumes": 21484, + "Ġassuming": 25538, + "Ġassumption": 24335, + "Ġassumptions": 31846, + "Ġassurance": 43601, + "Ġassurances": 84045, + "Ġassure": 34915, + "Ġassured": 27378, + "Ġassures": 92987, + "Ġast": 11763, + "Ġaster": 33937, + "Ġasteroid": 54379, + "Ġasteroids": 84222, + "Ġasthma": 50543, + "Ġaston": 36373, + "Ġastonished": 99074, + "Ġastonishing": 48913, + "Ġastore": 91807, + "Ġastounding": 73054, + "Ġastr": 22311, + "Ġastro": 46704, + "Ġastrology": 90847, + "Ġastronaut": 46633, + "Ġastronauts": 68422, + "Ġastronom": 45282, + "Ġastronomers": 86787, + "Ġastronomical": 86183, + "Ġastronomy": 81113, + "Ġasylum": 32170, + "Ġasym": 53295, + "Ġasymmetric": 96829, + "Ġasympt": 96254, + "Ġasync": 3312, + "Ġasynchronous": 39007, + "Ġasynchronously": 67781, + "Ġasyncio": 44918, + "ĠasÃŃ": 36460, + "Ġat": 518, + "ĠatIndex": 92749, + "Ġata": 73476, + "Ġatan": 57846, + "Ġatas": 87418, + "Ġatau": 38372, + "Ġate": 29812, + "Ġatención": 80541, + "Ġatenção": 138314, + "ĠateÅŁ": 140367, + "Ġath": 11742, + "Ġathe": 37814, + "Ġatheist": 66364, + "Ġatheists": 78525, + "Ġathlete": 33780, + "Ġathletes": 22742, + "Ġathletic": 28603, + "Ġathleticism": 92074, + "Ġathletics": 73075, + "Ġativ": 68332, + "Ġatl": 70454, + "Ġatlas": 60932, + "Ġatleast": 89928, + "Ġatm": 69787, + "Ġatmos": 13720, + "Ġatmosphere": 16566, + "Ġatmospheric": 44375, + "Ġatof": 64957, + "Ġatoi": 25227, + "Ġatol": 88655, + "Ġatom": 19124, + "Ġatomic": 24510, + "Ġatoms": 32199, + "Ġatop": 45988, + "Ġatr": 29570, + "Ġatrav": 70406, + "Ġatravés": 78116, + "Ġatrib": 52012, + "Ġatroc": 57728, + "Ġatrocities": 71236, + "Ġatrás": 135486, + "Ġatt": 1619, + "Ġattach": 15498, + "Ġattached": 12392, + "Ġattaches": 84503, + "Ġattaching": 71808, + "Ġattachment": 19984, + "Ġattachments": 33679, + "Ġattack": 3359, + "Ġattacked": 18349, + "Ġattacker": 30710, + "Ġattackers": 40965, + "Ġattacking": 22818, + "Ġattacks": 8776, + "Ġattain": 35761, + "Ġattained": 62160, + "Ġattainment": 92865, + "Ġatte": 92281, + "Ġattempt": 4774, + "Ġattempted": 17201, + "Ġattempting": 19405, + "Ġattempts": 13553, + "Ġattend": 9417, + "Ġattendance": 27018, + "Ġattendant": 63922, + "Ġattended": 18178, + "Ġattendee": 93565, + "Ġattendees": 39185, + "Ġattending": 23218, + "Ġattends": 74012, + "Ġattent": 51209, + "Ġattention": 6529, + "Ġattentive": 72798, + "Ġattenu": 56632, + "Ġattenuation": 94118, + "Ġattest": 86917, + "Ġattic": 73621, + "Ġattire": 68160, + "Ġattitude": 18915, + "Ġattitudes": 32626, + "ĠattivitÃł": 131201, + "Ġattn": 97817, + "Ġattorney": 13747, + "Ġattorneys": 30334, + "Ġattr": 6376, + "Ġattract": 9320, + "Ġattracted": 28023, + "Ġattracting": 50547, + "Ġattraction": 32364, + "Ġattractions": 38491, + "Ġattractive": 18879, + "Ġattractiveness": 98945, + "Ġattracts": 60091, + "Ġattravers": 97307, + "Ġattrib": 17329, + "Ġattributable": 70426, + "Ġattribute": 7035, + "ĠattributeName": 93259, + "Ġattributed": 29606, + "Ġattributes": 8201, + "Ġattribution": 62024, + "Ġattrs": 16204, + "Ġatual": 40875, + "Ġaté": 38483, + "Ġau": 7906, + "Ġauc": 45953, + "Ġauch": 10928, + "Ġauction": 21165, + "Ġauctions": 70132, + "Ġaucun": 84524, + "Ġaud": 6136, + "Ġaudi": 46368, + "Ġaudible": 79315, + "Ġaudience": 10650, + "Ġaudiences": 29894, + "Ġaudio": 7699, + "Ġaudit": 24275, + "Ġauditing": 92496, + "Ġaudition": 61888, + "Ġauditor": 53306, + "Ġauditory": 82529, + "Ġaudits": 74520, + "Ġauf": 7219, + "Ġaug": 19303, + "Ġaugment": 48706, + "Ġaugmentation": 78785, + "Ġaugmented": 55988, + "Ġaugust": 70619, + "Ġaujourd": 74704, + "Ġaument": 40780, + "Ġaumento": 81636, + "Ġaun": 89168, + "Ġaunque": 58366, + "Ġaunt": 50066, + "Ġaup": 97444, + "Ġauprès": 138807, + "Ġaur": 43421, + "Ġaura": 39042, + "Ġaure": 85668, + "Ġaus": 9421, + "Ġausge": 65078, + "Ġausp": 89096, + "Ġauss": 24131, + "Ġaussi": 27363, + "Ġaust": 99106, + "Ġausterity": 60530, + "Ġaustralia": 43232, + "Ġaut": 3078, + "Ġaute": 88037, + "Ġauth": 4166, + "ĠauthDomain": 97906, + "ĠauthService": 64696, + "ĠauthToken": 86022, + "Ġauthentic": 13210, + "Ġauthenticate": 33189, + "Ġauthenticated": 37260, + "Ġauthentication": 16653, + "Ġauthenticity": 53248, + "Ġauthor": 3150, + "Ġauthored": 66113, + "Ġauthorised": 69384, + "Ġauthoritarian": 57921, + "Ġauthoritative": 64593, + "Ġauthorities": 11276, + "Ġauthority": 11198, + "Ġauthorization": 23715, + "Ġauthorize": 36826, + "Ġauthorized": 18630, + "Ġauthors": 12014, + "Ġautism": 37181, + "Ġautistic": 80291, + "Ġauto": 3233, + "ĠautoComplete": 72486, + "ĠautoFocus": 94481, + "Ġautob": 98999, + "Ġautobi": 67065, + "Ġautobiography": 90437, + "Ġautoc": 56153, + "Ġautocomplete": 31705, + "Ġautof": 44445, + "Ġautofocus": 53032, + "Ġautogenerated": 81833, + "Ġautoimmune": 87091, + "Ġautoload": 80497, + "Ġautom": 5006, + "Ġautomat": 60752, + "Ġautomate": 68611, + "Ġautomated": 27498, + "Ġautomatic": 16962, + "Ġautomatically": 9463, + "Ġautomation": 32662, + "Ġautomobile": 34428, + "Ġautomobiles": 78548, + "Ġautomotive": 39408, + "Ġautomát": 79347, + "Ġautomáticamente": 83587, + "Ġautonom": 94003, + "Ġautonomous": 38193, + "Ġautonomy": 50260, + "Ġautop": 71158, + "Ġautoplay": 57310, + "Ġautopsy": 75801, + "Ġautor": 26767, + "Ġautorelease": 74566, + "Ġautos": 46872, + "Ġautour": 88953, + "Ġautre": 46738, + "Ġautres": 45006, + "Ġautumn": 41674, + "Ġaux": 10047, + "Ġauxiliary": 53458, + "Ġav": 1822, + "Ġava": 83664, + "Ġavail": 37911, + "Ġavailability": 18048, + "Ġavailable": 2500, + "Ġavait": 49490, + "Ġaval": 40535, + "Ġavalanche": 95594, + "Ġavalia": 139270, + "Ġavaliação": 138044, + "Ġavaliações": 139271, + "Ġavant": 32570, + "Ġavanz": 98292, + "Ġavatar": 20701, + "Ġave": 41638, + "Ġavec": 9753, + "Ġavent": 90706, + "Ġavenue": 61703, + "Ġavenues": 72134, + "Ġaver": 17302, + "Ġaverage": 5461, + "Ġaveraged": 36856, + "Ġaverages": 48820, + "Ġaveraging": 43764, + "Ġavere": 82825, + "Ġavez": 46492, + "Ġavg": 19712, + "Ġavi": 98188, + "Ġaviation": 45530, + "Ġavid": 55553, + "Ġavis": 54747, + "Ġavocado": 67246, + "Ġavoid": 5648, + "Ġavoidance": 64311, + "Ġavoided": 30790, + "Ġavoiding": 30426, + "Ġavoids": 54852, + "Ġavoir": 38829, + "Ġavons": 55938, + "Ġavril": 99036, + "Ġaw": 1360, + "Ġawait": 2535, + "Ġawaited": 78478, + "Ġawaiting": 43988, + "Ġawaits": 70804, + "Ġawake": 34347, + "ĠawakeFromNib": 74949, + "Ġawaken": 41418, + "Ġawakened": 89247, + "Ġawakening": 74049, + "Ġaward": 10084, + "Ġawarded": 21328, + "Ġawards": 22344, + "Ġaware": 7853, + "Ġawareness": 17526, + "Ġaway": 3123, + "Ġawe": 50417, + "Ġawesome": 12456, + "Ġawful": 24607, + "Ġawhile": 50469, + "Ġawk": 19976, + "Ġawkward": 28759, + "Ġaws": 31521, + "Ġax": 3859, + "Ġaxe": 50464, + "Ġaxes": 24745, + "Ġaxial": 97180, + "Ġaxiom": 88177, + "Ġaxios": 12974, + "Ġaxis": 8024, + "Ġaxle": 72100, + "Ġaxs": 75162, + "Ġay": 18898, + "Ġayant": 85771, + "Ġaynı": 129301, + "Ġayr": 135502, + "Ġayrı": 133353, + "Ġayrıca": 137241, + "Ġayrıl": 135503, + "Ġayud": 58137, + "Ġayuda": 73659, + "Ġayında": 139830, + "Ġayır": 141523, + "Ġaz": 12376, + "Ġazi": 74062, + "Ġazimuth": 91960, + "Ġazt": 97894, + "Ġazure": 76530, + "Ġaç": 62910, + "Ġaçık": 129245, + "Ġaçıkl": 130165, + "Ġaçıklama": 130166, + "Ġaçıl": 131358, + "Ġaçıs": 136697, + "Ġaçısından": 136698, + "Ġañ": 72186, + "Ġaño": 35816, + "Ġaños": 26748, + "Ġaún": 77558, + "ĠaÄŁ": 127485, + "ĠaÄŁrı": 140757, + "Ġaģır": 133504, + "ĠaÅŁ": 125409, + "ĠaÅŁama": 139898, + "ĠaÅŁaģı": 128120, + "ĠaÅŁaģıd": 142961, + "ĠaÅŁaģıdaki": 142962, + "ĠaÅŁk": 138124, + "Ġb": 293, + "Ġba": 12789, + "Ġbab": 16584, + "Ġbabe": 40246, + "Ġbabel": 79826, + "Ġbabes": 59500, + "Ġbabies": 23920, + "Ġbaby": 8770, + "Ġbabys": 70583, + "Ġbac": 79880, + "Ġbach": 47019, + "Ġbacheca": 61545, + "Ġbachelor": 48583, + "Ġback": 1182, + "ĠbackButton": 89726, + "Ġbackbone": 55427, + "Ġbackdrop": 38477, + "Ġbacked": 21411, + "Ġbackend": 19163, + "Ġbackers": 53630, + "Ġbackground": 4004, + "ĠbackgroundColor": 12694, + "ĠbackgroundImage": 67029, + "Ġbackgrounds": 35476, + "Ġbacking": 24668, + "Ġbacklash": 47056, + "Ġbacklight": 94322, + "Ġbacklog": 70722, + "Ġbackpack": 33136, + "Ġbackpage": 94959, + "Ġbacks": 27578, + "Ġbackstage": 89725, + "Ġbackstory": 92271, + "Ġbacktrack": 98921, + "Ġbackup": 15725, + "Ġbackups": 59666, + "Ġbackward": 27555, + "Ġbackwards": 28412, + "Ġbackyard": 35660, + "Ġbacon": 40352, + "Ġbacter": 17398, + "Ġbacteria": 23157, + "Ġbacterial": 44864, + "Ġbad": 3873, + "Ġbada": 143378, + "Ġbadass": 86302, + "ĠbadaÅĦ": 143379, + "Ġbadge": 23961, + "Ġbadges": 60434, + "Ġbadly": 24569, + "Ġbaff": 64416, + "Ġbag": 8968, + "Ġbaggage": 60549, + "Ġbagi": 85650, + "Ġbags": 17899, + "Ġbah": 41929, + "Ġbahwa": 77336, + "Ġbahçe": 143827, + "Ġbaik": 79187, + "Ġbail": 24479, + "Ġbailout": 74664, + "Ġbais": 96904, + "Ġbaise": 34816, + "Ġbaiser": 78682, + "Ġbait": 48060, + "Ġbaj": 90741, + "Ġbaja": 87648, + "Ġbajo": 62412, + "Ġbak": 36203, + "Ġbake": 22544, + "Ġbakeca": 39957, + "Ġbaked": 40678, + "Ġbakeka": 81943, + "Ġbaker": 75828, + "Ġbakery": 65144, + "Ġbaking": 27815, + "Ġbakım": 134364, + "Ġbal": 9642, + "Ġbalance": 8172, + "Ġbalanced": 23831, + "Ġbalances": 38854, + "Ġbalancing": 43363, + "Ġbalcon": 37305, + "Ġbalcony": 43852, + "Ġbald": 47553, + "Ġbalk": 91982, + "Ġball": 4935, + "Ġballet": 70210, + "Ġballistic": 59533, + "Ġballo": 56853, + "Ġballoon": 46756, + "Ġballoons": 69480, + "Ġballot": 25853, + "Ġballots": 48636, + "Ġballpark": 96741, + "Ġballs": 20332, + "Ġbam": 41304, + "Ġbamb": 75993, + "Ġbamboo": 58882, + "Ġban": 8943, + "Ġbanana": 43096, + "Ġbananas": 67342, + "Ġbanc": 68307, + "Ġbanco": 52465, + "Ġband": 7055, + "Ġbanda": 96111, + "Ġbande": 94038, + "Ġbands": 20892, + "Ġbandwidth": 33394, + "Ġbang": 25441, + "Ġbanged": 93599, + "Ġbanging": 86111, + "Ġbank": 6073, + "Ġbanker": 71659, + "Ġbankers": 63143, + "Ġbanking": 22798, + "Ġbankrupt": 28220, + "Ġbankruptcy": 35607, + "Ġbanks": 13959, + "Ġbanned": 20834, + "Ġbanner": 23323, + "Ġbanners": 58372, + "Ġbanning": 48536, + "Ġbanquet": 86738, + "Ġbans": 47509, + "Ġbanyak": 61892, + "Ġbapt": 42624, + "Ġbaptism": 73827, + "Ġbaptized": 89125, + "Ġbar": 3619, + "Ġbara": 81087, + "Ġbarang": 95704, + "Ġbarbar": 53390, + "Ġbarbecue": 61212, + "Ġbarber": 95994, + "Ġbarcelona": 81414, + "Ġbarcode": 49407, + "Ġbard": 41810, + "Ġbardzo": 56858, + "Ġbare": 12461, + "Ġbarely": 19460, + "Ġbarg": 25881, + "Ġbargain": 44563, + "Ġbargaining": 54209, + "Ġbark": 53742, + "Ġbarley": 80717, + "Ġbarn": 32319, + "Ġbarr": 30939, + "Ġbarracks": 94764, + "Ġbarrage": 80590, + "Ġbarred": 58944, + "Ġbarrel": 20311, + "Ġbarrels": 41875, + "Ġbarren": 93988, + "Ġbarric": 77580, + "Ġbarrier": 22103, + "Ġbarriers": 29640, + "Ġbarring": 86699, + "Ġbars": 15904, + "Ġbart": 61518, + "Ġbartender": 90405, + "Ġbaru": 68343, + "Ġbas": 3046, + "Ġbasal": 79610, + "Ġbase": 2331, + "ĠbasePath": 71577, + "ĠbaseURL": 56741, + "ĠbaseUrl": 39775, + "Ġbaseball": 19507, + "Ġbased": 3118, + "Ġbaseline": 25869, + "Ġbaseman": 83303, + "Ġbasement": 30641, + "Ġbasename": 38196, + "Ġbases": 23092, + "Ġbash": 27023, + "Ġbasic": 6770, + "Ġbasically": 13221, + "Ġbasics": 31774, + "Ġbasil": 67086, + "Ġbasin": 57209, + "Ġbasis": 8037, + "Ġbask": 138832, + "Ġbasket": 14024, + "Ġbasketball": 19240, + "Ġbaskets": 68344, + "Ġbaskı": 138833, + "Ġbass": 21529, + "Ġbast": 31977, + "Ġbastante": 70562, + "Ġbastard": 73577, + "Ġbasın": 138010, + "Ġbat": 15743, + "Ġbatch": 7162, + "ĠbatchSize": 97400, + "Ġbatches": 44792, + "Ġbatching": 84256, + "Ġbath": 8885, + "Ġbathing": 72409, + "Ġbathroom": 14852, + "Ġbathrooms": 39883, + "Ġbaths": 58200, + "Ġbathtub": 80088, + "Ġbats": 42208, + "Ġbatt": 14754, + "Ġbattalion": 90416, + "Ġbatter": 8745, + "Ġbattered": 79656, + "Ġbatteries": 26264, + "Ġbattery": 11602, + "Ġbatting": 50810, + "Ġbattle": 8049, + "Ġbattled": 74340, + "Ġbattlefield": 37386, + "Ġbattleground": 90566, + "Ġbattles": 24554, + "Ġbattling": 46236, + "Ġbaud": 56822, + "Ġbay": 22708, + "Ġbaz": 50247, + "Ġbazı": 130330, + "Ġbaño": 140521, + "ĠbaÄŁ": 125791, + "ĠbaÄŁl": 139159, + "ĠbaÄŁlant": 139160, + "ĠbaÄŁlantı": 139161, + "ĠbaÄŁlı": 132155, + "ĠbaÅŁ": 46555, + "ĠbaÅŁar": 137133, + "ĠbaÅŁarı": 132962, + "ĠbaÅŁarılı": 137134, + "ĠbaÅŁka": 129149, + "ĠbaÅŁladı": 131727, + "ĠbaÅŁlat": 137188, + "ĠbaÅŁlayan": 139324, + "ĠbaÅŁv": 135667, + "ĠbaÅŁvuru": 135668, + "ĠbaÅŁÄ±na": 139358, + "ĠbaÅŁÄ±nda": 137835, + "Ġbb": 16520, + "Ġbbc": 90626, + "Ġbbox": 29749, + "Ġbbw": 40760, + "Ġbc": 17916, + "Ġbcm": 56768, + "Ġbcrypt": 33520, + "Ġbd": 21486, + "Ġbdsm": 41697, + "Ġbe": 387, + "Ġbeach": 11321, + "Ġbeaches": 34809, + "Ġbeacon": 51302, + "Ġbead": 74507, + "Ġbeads": 54208, + "Ġbeam": 23418, + "Ġbeams": 49945, + "Ġbean": 20427, + "Ġbeans": 26896, + "Ġbear": 11722, + "Ġbeard": 47688, + "Ġbearer": 69771, + "Ġbearing": 18043, + "Ġbearings": 34453, + "Ġbears": 29724, + "Ġbeast": 31136, + "Ġbeasts": 56436, + "Ġbeat": 9382, + "Ġbeaten": 30294, + "Ġbeating": 26150, + "Ġbeats": 33327, + "Ġbeau": 70906, + "Ġbeaucoup": 47388, + "Ġbeaut": 4912, + "Ġbeautiful": 6233, + "Ġbeautifully": 31619, + "Ġbeauty": 13143, + "Ġbeb": 67046, + "Ġbeberapa": 84938, + "Ġbecame": 6116, + "Ġbecause": 1576, + "Ġbeck": 91086, + "Ġbecome": 3635, + "Ġbecomes": 9044, + "Ġbecoming": 10454, + "Ġbed": 4845, + "Ġbedding": 60153, + "Ġbedeut": 81831, + "Ġbedroom": 13829, + "Ġbedrooms": 27589, + "Ġbeds": 26938, + "Ġbedside": 95703, + "Ġbedtime": 88507, + "Ġbee": 37228, + "Ġbeef": 24317, + "Ġbeen": 1012, + "Ġbeep": 90954, + "Ġbeer": 12883, + "Ġbeers": 36966, + "Ġbees": 39458, + "Ġbeet": 48012, + "Ġbeetle": 97386, + "Ġbef": 56946, + "Ġbefind": 81616, + "Ġbefore": 1573, + "ĠbeforeEach": 17723, + "ĠbeforeSend": 91512, + "Ġbeforehand": 51059, + "Ġbeg": 2148, + "Ġbegan": 6009, + "Ġbegged": 79292, + "Ġbegging": 59106, + "Ġbegin": 3161, + "Ġbeginner": 48948, + "Ġbeginners": 46850, + "Ġbeginning": 7167, + "Ġbeginnings": 66382, + "Ġbegins": 12033, + "Ġbegr": 87541, + "Ġbegs": 85201, + "Ġbegun": 21376, + "Ġbeh": 2755, + "Ġbehalf": 17522, + "Ġbehand": 72202, + "Ġbehave": 35692, + "Ġbehaved": 88731, + "Ġbehaves": 73057, + "Ġbehaving": 86557, + "Ġbehavior": 7709, + "Ġbehavioral": 35595, + "Ġbehaviors": 27099, + "Ġbehaviour": 16999, + "Ġbehavioural": 87879, + "Ġbehaviours": 70077, + "Ġbehind": 4815, + "Ġbehold": 56115, + "Ġbei": 13279, + "Ġbeide": 97144, + "Ġbeiden": 69791, + "Ġbeige": 78265, + "Ġbeim": 28789, + "Ġbeing": 1660, + "Ġbeings": 22977, + "Ġbek": 49933, + "Ġbekan": 61606, + "Ġbekannt": 81331, + "Ġbekom": 93077, + "Ġbekommen": 74675, + "Ġbel": 1656, + "Ġbelang": 83246, + "Ġbele": 70056, + "Ġbelie": 2818, + "Ġbelief": 16396, + "Ġbeliefs": 20799, + "Ġbelievable": 91395, + "Ġbelieve": 4411, + "Ġbelieved": 11585, + "Ġbeliever": 61279, + "Ġbelievers": 50913, + "Ġbelieves": 13605, + "Ġbelieving": 33990, + "Ġbell": 28419, + "Ġbella": 96460, + "Ġbelle": 38792, + "Ġbells": 60694, + "Ġbelly": 35417, + "Ġbelong": 9173, + "Ġbelonged": 45859, + "Ġbelonging": 32052, + "Ġbelongings": 62928, + "Ġbelongs": 17180, + "Ġbeloved": 27430, + "Ġbelow": 3685, + "Ġbelt": 19125, + "Ġbelts": 52679, + "Ġbelum": 90757, + "Ġbem": 31915, + "Ġben": 3318, + "Ġbench": 13425, + "Ġbenches": 78409, + "Ġbenchmark": 28431, + "Ġbenchmarks": 62019, + "Ġbend": 36820, + "Ġbending": 57118, + "Ġbends": 92797, + "Ġbene": 19750, + "Ġbeneath": 23969, + "Ġbenef": 4177, + "Ġbenefici": 36945, + "Ġbeneficial": 23699, + "Ġbeneficiaries": 60568, + "Ġbeneficiary": 62832, + "Ġbeneficiation": 78219, + "Ġbenefit": 8760, + "Ġbenefited": 55028, + "Ġbenefiting": 82915, + "Ġbenefits": 7567, + "Ġbenign": 64209, + "Ġbent": 29180, + "Ġbenz": 70946, + "Ġbenöt": 91304, + "Ġber": 10207, + "Ġbere": 26933, + "Ġbereits": 47143, + "Ġberg": 45885, + "Ġbergen": 77325, + "Ġberhasil": 75582, + "Ġberk": 96583, + "Ġberlin": 44581, + "Ġbern": 89263, + "Ġberries": 62394, + "Ġberry": 85088, + "Ġbers": 58554, + "Ġbert": 67099, + "Ġberth": 95447, + "Ġbes": 9057, + "Ġbesar": 84885, + "Ġbesch": 41852, + "Ġbeside": 29388, + "Ġbesides": 27758, + "Ġbesie": 91628, + "Ġbesoin": 62569, + "Ġbesonders": 75590, + "Ġbespoke": 75549, + "Ġbesser": 70336, + "Ġbest": 1850, + "Ġbeste": 28822, + "Ġbesteht": 98421, + "Ġbestellen": 97791, + "Ġbesten": 52747, + "Ġbestimm": 94718, + "Ġbestowed": 94729, + "Ġbestselling": 70902, + "Ġbet": 1274, + "Ġbeta": 13440, + "Ġbetr": 81314, + "Ġbetray": 33468, + "Ġbetrayal": 67276, + "Ġbetrayed": 64297, + "Ġbets": 42077, + "Ġbetter": 2664, + "Ġbetting": 25181, + "Ġbetween": 1948, + "Ġbeurette": 81712, + "Ġbev": 94808, + "Ġbeverage": 42350, + "Ġbeverages": 46397, + "Ġbevor": 94505, + "Ġbew": 26956, + "Ġbeware": 83805, + "Ġbewild": 98674, + "Ġbey": 7732, + "Ġbeyond": 7797, + "Ġbez": 22996, + "Ġbezier": 83066, + "ĠbezpieczeÅĦst": 142329, + "ĠbezpieczeÅĦstwa": 142330, + "Ġbezpo": 138350, + "ĠbezpoÅĽ": 138351, + "ĠbezpoÅĽredni": 138352, + "ĠbeÄŁ": 139264, + "ĠbeÄŁen": 139265, + "Ġbf": 39093, + "Ġbfd": 94047, + "Ġbfs": 66564, + "Ġbg": 8951, + "ĠbgColor": 73130, + "Ġbgcolor": 30195, + "Ġbh": 42989, + "Ġbi": 6032, + "Ġbian": 96489, + "Ġbias": 15470, + "Ġbiased": 47661, + "Ġbiases": 49083, + "Ġbib": 23829, + "Ġbible": 60201, + "Ġbibli": 43515, + "Ġbiblical": 52540, + "Ġbibliography": 93698, + "Ġbic": 59731, + "Ġbicy": 51417, + "Ġbicycle": 34986, + "Ġbicycles": 72861, + "Ġbid": 14103, + "Ġbidder": 77531, + "Ġbidding": 48400, + "Ġbidi": 79922, + "Ġbids": 43499, + "Ġbied": 97476, + "Ġbien": 14370, + "Ġbieten": 66238, + "Ġbietet": 60729, + "Ġbif": 91540, + "Ġbig": 2409, + "Ġbigger": 11243, + "Ġbiggest": 8538, + "Ġbigint": 79663, + "Ġbigot": 65703, + "Ġbigotry": 76746, + "Ġbij": 22424, + "Ġbik": 37545, + "Ġbike": 12963, + "Ġbikes": 30453, + "Ġbiking": 66080, + "Ġbikini": 54205, + "Ġbil": 20316, + "Ġbilateral": 51203, + "Ġbild": 71211, + "Ġbilder": 33700, + "Ġbile": 72294, + "Ġbilingual": 90289, + "Ġbill": 4031, + "Ġbillboard": 85206, + "Ġbilled": 49446, + "Ġbilleder": 95440, + "Ġbillig": 95139, + "Ġbilling": 33531, + "Ġbillion": 7094, + "Ġbillionaire": 46694, + "Ġbillionaires": 89286, + "Ġbillions": 32051, + "Ġbills": 18610, + "Ġbin": 9544, + "Ġbinaries": 69737, + "Ġbinary": 7868, + "Ġbind": 10719, + "ĠbindActionCreators": 78220, + "Ġbinder": 54014, + "Ġbinding": 10970, + "Ġbindings": 35700, + "Ġbinds": 57485, + "Ġbinge": 72822, + "Ġbingo": 77024, + "Ġbinnen": 82985, + "Ġbins": 28518, + "Ġbio": 16905, + "Ġbiochemical": 92432, + "Ġbiod": 55494, + "Ġbiodiversity": 72019, + "Ġbiography": 47245, + "Ġbiological": 23275, + "Ġbiologist": 87604, + "Ġbiology": 33358, + "Ġbiom": 38438, + "Ġbiomass": 57658, + "Ġbiome": 83610, + "Ġbiomedical": 93512, + "Ġbiopsy": 98547, + "Ġbios": 48454, + "Ġbip": 28878, + "Ġbipartisan": 50908, + "Ġbipolar": 64819, + "Ġbir": 15248, + "Ġbird": 11958, + "Ġbirds": 19654, + "Ġbirka": 137653, + "Ġbirkaç": 137654, + "Ġbirth": 7194, + "Ġbirthday": 15198, + "Ġbirthdays": 91850, + "Ġbirths": 65232, + "Ġbirç": 131653, + "Ġbirçok": 131654, + "Ġbis": 14840, + "Ġbisa": 39988, + "Ġbisc": 59459, + "Ġbiscuits": 85976, + "Ġbisexual": 55732, + "Ġbisher": 96042, + "Ġbishop": 53206, + "Ġbishops": 65623, + "Ġbist": 77528, + "Ġbit": 2699, + "Ġbitch": 39430, + "Ġbitcoin": 17873, + "Ġbitcoins": 57002, + "Ġbite": 22721, + "Ġbites": 48333, + "Ġbiting": 67096, + "Ġbitmap": 19746, + "Ġbitmask": 90961, + "Ġbitrate": 82643, + "Ġbits": 9472, + "Ġbitte": 67382, + "Ġbitten": 77449, + "Ġbitter": 25180, + "Ġbitterly": 98845, + "Ġbitterness": 78996, + "Ġbitwise": 97970, + "Ġbiz": 42378, + "Ġbizarre": 32188, + "Ġbiç": 137454, + "Ġbiçim": 137455, + "Ġbiên": 135547, + "Ġbiến": 128871, + "Ġbiết": 128345, + "Ġbiá»ĥn": 130027, + "Ġbiá»ĥu": 129626, + "Ġbiá»ĩn": 137261, + "Ġbiá»ĩt": 128589, + "Ġbj": 91188, + "Ġbk": 66101, + "Ġbl": 1501, + "Ġbla": 84541, + "Ġblack": 3691, + "ĠblackColor": 75554, + "Ġblackjack": 57262, + "Ġblacklist": 66049, + "Ġblackmail": 87163, + "Ġblackout": 96404, + "Ġblacks": 38977, + "Ġbladder": 61464, + "Ġblade": 24838, + "Ġblades": 41642, + "Ġblah": 52561, + "Ġblame": 18555, + "Ġblamed": 35097, + "Ġblames": 85790, + "Ġblaming": 58671, + "Ġblanc": 63247, + "Ġblanco": 89547, + "Ġbland": 49431, + "Ġblank": 10113, + "Ġblanket": 38039, + "Ġblankets": 64612, + "Ġblanks": 78540, + "Ġblas": 48577, + "Ġblasph": 98059, + "Ġblast": 20671, + "Ġblasted": 62321, + "Ġblasting": 76152, + "Ġblasts": 74041, + "Ġblat": 50949, + "Ġblatant": 71447, + "Ġblatantly": 97154, + "Ġblaze": 62473, + "Ġblazing": 85250, + "Ġble": 12422, + "Ġbleach": 82351, + "Ġbleak": 75267, + "Ġbleed": 66006, + "Ġbleeding": 34663, + "Ġbleiben": 90556, + "Ġbleibt": 93985, + "Ġblend": 20334, + "Ġblended": 54545, + "Ġblender": 61795, + "Ġblending": 55841, + "Ġblends": 57843, + "Ġbless": 16072, + "Ġblessed": 32844, + "Ġblessing": 39049, + "Ġblessings": 55550, + "Ġblev": 71999, + "Ġblew": 41323, + "Ġbli": 59289, + "Ġblij": 73280, + "Ġblind": 18020, + "Ġblinded": 89116, + "Ġblindly": 88347, + "Ġblindness": 84415, + "Ġblinds": 66061, + "Ġblink": 33131, + "Ġblinked": 99032, + "Ġblinking": 84022, + "Ġblir": 57437, + "Ġbliss": 57945, + "Ġblister": 80692, + "Ġblitz": 71434, + "Ġblk": 40643, + "Ġblo": 14211, + "Ġblob": 23404, + "Ġblobs": 76899, + "Ġbloc": 40740, + "Ġblock": 2504, + "ĠblockDim": 80993, + "ĠblockIdx": 84279, + "ĠblockSize": 64178, + "Ġblockade": 76137, + "Ġblockbuster": 70997, + "Ġblockchain": 17944, + "Ġblocked": 19299, + "Ġblocker": 50910, + "Ġblockers": 83135, + "Ġblocking": 22188, + "Ġblocks": 10010, + "Ġblog": 5010, + "Ġblogger": 39181, + "Ġbloggers": 49243, + "Ġblogging": 41992, + "Ġbloginfo": 59681, + "Ġblogs": 25659, + "Ġblond": 44369, + "Ġblonde": 26026, + "Ġblood": 6543, + "Ġbloodstream": 99116, + "Ġbloody": 35177, + "Ġbloom": 51454, + "Ġblooms": 91789, + "Ġbloque": 70726, + "Ġbloss": 49723, + "Ġblossom": 88758, + "Ġblot": 80882, + "Ġblouse": 95945, + "Ġblow": 12244, + "Ġblowing": 40342, + "Ġblowjob": 37781, + "Ġblown": 35474, + "Ġblows": 45203, + "Ġblue": 6303, + "Ġblueprint": 52929, + "Ġblues": 43595, + "Ġbluetooth": 55601, + "Ġbluff": 96288, + "Ġblunt": 48670, + "Ġblur": 28513, + "Ġblurred": 72400, + "Ġblurry": 99055, + "Ġblush": 69226, + "Ġbm": 34868, + "Ġbmi": 91897, + "Ġbmp": 43301, + "Ġbn": 40597, + "Ġbo": 708, + "Ġboa": 81932, + "Ġboard": 4479, + "Ġboarded": 83469, + "Ġboarding": 48969, + "Ġboards": 20490, + "Ġboast": 58127, + "Ġboasted": 83212, + "Ġboasting": 83644, + "Ġboasts": 37019, + "Ġboat": 15328, + "Ġboats": 31631, + "Ġbob": 35192, + "Ġboca": 96206, + "Ġbod": 29011, + "Ġbodies": 12866, + "Ġbodily": 47031, + "Ġbody": 2487, + "ĠbodyParser": 49936, + "Ġbog": 34419, + "Ġbogus": 59920, + "Ġboil": 43690, + "Ġboiled": 65085, + "Ġboiler": 27625, + "Ġboilers": 68949, + "Ġboiling": 49837, + "Ġboils": 88955, + "Ġbois": 65204, + "Ġbol": 20771, + "Ġbola": 97187, + "Ġbold": 13939, + "Ġboldly": 94929, + "Ġbols": 88216, + "Ġbolster": 47127, + "Ġbolt": 31842, + "Ġbolts": 48839, + "Ġbom": 18167, + "Ġbomb": 12764, + "Ġbombard": 85462, + "Ġbombed": 88838, + "Ġbomber": 57526, + "Ġbombers": 68820, + "Ġbombing": 36331, + "Ġbombings": 73379, + "Ġbombs": 32506, + "Ġbon": 7814, + "Ġbona": 93069, + "Ġbond": 10815, + "Ġbondage": 42494, + "Ġbonded": 69141, + "Ġbonding": 63086, + "Ġbonds": 26363, + "Ġbone": 17240, + "Ġbones": 24854, + "Ġbonne": 50551, + "Ġbons": 81034, + "Ġbonus": 12037, + "Ġbonuses": 32221, + "Ġboo": 46270, + "Ġboob": 78142, + "Ġboobs": 39792, + "Ġbook": 2311, + "Ġbooked": 32970, + "Ġbooking": 21857, + "Ġbookings": 58520, + "Ġbooklet": 73243, + "Ġbookmark": 35149, + "Ġbookmarks": 60919, + "Ġbooks": 6467, + "Ġbookstore": 78661, + "Ġbool": 1807, + "Ġboolean": 2710, + "Ġboom": 29745, + "Ġbooming": 67690, + "Ġboon": 86001, + "Ġboost": 7760, + "Ġboosted": 64108, + "Ġbooster": 56432, + "Ġboosting": 54928, + "Ġboosts": 66132, + "Ġboot": 10459, + "Ġbooth": 35406, + "Ġbooths": 91159, + "Ġbootloader": 88164, + "Ġboots": 25236, + "Ġbootstrap": 26925, + "Ġbooty": 61602, + "Ġbooze": 88936, + "Ġbor": 31632, + "Ġbord": 37902, + "Ġbordel": 93697, + "Ġborder": 3886, + "ĠborderBottom": 55206, + "ĠborderColor": 35826, + "ĠborderRadius": 19208, + "ĠborderSide": 97655, + "ĠborderTop": 85983, + "ĠborderWidth": 44105, + "Ġbordered": 76217, + "Ġborderline": 88514, + "Ġborders": 23806, + "Ġbore": 38686, + "Ġbored": 33286, + "Ġboredom": 89826, + "Ġboring": 27759, + "Ġborn": 9223, + "Ġborne": 64062, + "Ġborough": 65741, + "Ġborr": 96552, + "Ġborrow": 17193, + "Ġborrowed": 45564, + "Ġborrower": 69619, + "Ġborrowers": 63388, + "Ġborrowing": 52575, + "Ġborç": 143222, + "Ġbos": 42646, + "Ġboss": 13392, + "Ġbosses": 41747, + "Ġbot": 10924, + "Ġbotanical": 97202, + "Ġboth": 2176, + "Ġbother": 20147, + "Ġbothered": 45276, + "Ġbothering": 90159, + "Ġbothers": 81354, + "Ġboto": 60779, + "Ġboton": 79467, + "Ġbots": 49529, + "Ġbott": 10936, + "Ġbottle": 16486, + "Ġbottled": 71078, + "Ġbottleneck": 87838, + "Ġbottles": 26376, + "Ġbottom": 5622, + "Ġbottoms": 92550, + "Ġbou": 29770, + "Ġbought": 10788, + "Ġbounce": 33682, + "Ġbounced": 69804, + "Ġbouncing": 64028, + "Ġbound": 6822, + "Ġboundaries": 22711, + "Ġboundary": 18732, + "Ġbounded": 61115, + "Ġbounding": 30618, + "Ġbounds": 14262, + "Ġbounty": 52423, + "Ġbouquet": 91613, + "Ġbour": 35579, + "Ġbourbon": 92748, + "Ġbourgeois": 52316, + "Ġbourgeoisie": 93786, + "Ġbout": 24623, + "Ġboutique": 52085, + "Ġbouts": 95171, + "Ġbow": 15273, + "Ġbowed": 84373, + "Ġbowel": 65258, + "Ġbowl": 19212, + "Ġbowling": 58196, + "Ġbowls": 59980, + "Ġbows": 81465, + "Ġbox": 3745, + "ĠboxShadow": 68788, + "Ġboxed": 73664, + "Ġboxer": 78457, + "Ġboxes": 14697, + "Ġboxing": 45229, + "Ġboy": 8171, + "Ġboyc": 46234, + "Ġboycott": 52545, + "Ġboyfriend": 25838, + "Ġboys": 13007, + "ĠboÅŁ": 134195, + "Ġbp": 26685, + "Ġbpm": 97724, + "Ġbpp": 90533, + "Ġbpy": 51061, + "Ġbq": 94765, + "Ġbr": 1411, + "Ġbra": 20112, + "Ġbrace": 32864, + "Ġbracelet": 58419, + "Ġbracelets": 88865, + "Ġbraces": 59191, + "Ġbracket": 31642, + "Ġbrackets": 38929, + "Ġbrag": 56694, + "Ġbrain": 8109, + "Ġbrains": 34102, + "Ġbrainstorm": 86781, + "Ġbrake": 34618, + "Ġbrakes": 44564, + "Ġbraking": 72488, + "Ġbran": 52258, + "Ġbranch": 8870, + "Ġbranches": 23091, + "Ġbranching": 85467, + "Ġbrand": 6741, + "Ġbranded": 45835, + "Ġbranding": 43717, + "Ġbrands": 15721, + "Ġbras": 45396, + "Ġbrasile": 67973, + "Ġbrass": 36038, + "Ġbrav": 75713, + "Ġbrave": 33200, + "Ġbravery": 85066, + "Ġbrawl": 99067, + "Ġbraz": 97986, + "Ġbrazil": 96646, + "Ġbre": 5282, + "Ġbreach": 30371, + "Ġbreached": 81066, + "Ġbreaches": 68040, + "Ġbread": 16002, + "Ġbreadcrumb": 76815, + "Ġbreadcrumbs": 85566, + "Ġbreadth": 57221, + "Ġbreak": 1438, + "Ġbreakdown": 29985, + "Ġbreaker": 66767, + "Ġbreakfast": 17496, + "Ġbreaking": 14719, + "Ġbreakout": 62688, + "Ġbreakpoint": 52745, + "Ġbreakpoints": 91530, + "Ġbreaks": 18303, + "Ġbreakthrough": 41345, + "Ġbreakup": 84498, + "Ġbreast": 17216, + "Ġbreastfeeding": 64764, + "Ġbreasts": 36349, + "Ġbreat": 27755, + "Ġbreath": 11486, + "Ġbreathable": 89519, + "Ġbreathe": 36297, + "Ġbreathed": 91166, + "Ġbreathing": 25938, + "Ġbreathtaking": 56092, + "Ġbred": 54087, + "Ġbree": 67299, + "Ġbreed": 27775, + "Ġbreeding": 39208, + "Ġbreeds": 57145, + "Ġbreeze": 45285, + "Ġbrethren": 72803, + "Ġbreve": 97419, + "Ġbrew": 16695, + "Ġbrewed": 88095, + "Ġbreweries": 72702, + "Ġbrewers": 92870, + "Ġbrewery": 49195, + "Ġbrewing": 49159, + "Ġbri": 39468, + "Ġbrib": 75612, + "Ġbribery": 92037, + "Ġbrick": 24837, + "Ġbricks": 49037, + "Ġbrid": 76747, + "Ġbridal": 80412, + "Ġbride": 41347, + "Ġbrides": 60265, + "Ġbridge": 14164, + "Ġbridges": 38973, + "Ġbrief": 9814, + "Ġbriefed": 85765, + "Ġbriefing": 42536, + "Ġbriefly": 26753, + "Ġbrig": 54812, + "Ġbrigade": 83872, + "Ġbright": 9906, + "Ġbrighter": 52557, + "Ġbrightest": 70921, + "Ġbrightly": 75289, + "Ġbrightness": 32206, + "Ġbrill": 17581, + "Ġbrilliance": 82480, + "Ġbrilliant": 19752, + "Ġbrilliantly": 78249, + "Ġbring": 4446, + "Ġbringen": 90845, + "Ġbringing": 12678, + "Ġbrings": 12434, + "Ġbrink": 67115, + "Ġbrisk": 74391, + "Ġbrit": 37330, + "Ġbritish": 93927, + "Ġbrittle": 94649, + "Ġbro": 2896, + "Ġbroad": 7205, + "Ġbroadband": 40825, + "Ġbroadcast": 12899, + "Ġbroadcaster": 59883, + "Ġbroadcasters": 92914, + "Ġbroadcasting": 48827, + "Ġbroadcasts": 65098, + "Ġbroaden": 83613, + "Ġbroader": 26829, + "Ġbroadly": 42929, + "Ġbroccoli": 78176, + "Ġbrochure": 78803, + "Ġbroke": 14422, + "Ġbroken": 10865, + "Ġbroker": 22316, + "Ġbrokerage": 74855, + "Ġbrokers": 45336, + "Ġbrom": 93471, + "Ġbron": 29971, + "Ġbronze": 39807, + "Ġbroth": 44893, + "Ġbrother": 10641, + "Ġbrothers": 20208, + "Ġbrought": 7117, + "Ġbrow": 59275, + "Ġbrown": 13876, + "Ġbrows": 18833, + "Ġbrowse": 26009, + "Ġbrowser": 6929, + "Ġbrowsers": 32857, + "Ġbrowsing": 31321, + "Ġbru": 42000, + "Ġbruises": 97575, + "Ġbruk": 80771, + "Ġbrun": 41607, + "Ġbrunch": 69817, + "Ġbrunette": 48117, + "Ġbrush": 15625, + "Ġbrushed": 61539, + "Ġbrushes": 61626, + "Ġbrushing": 76726, + "Ġbrut": 19404, + "Ġbrutal": 27760, + "Ġbrutality": 62032, + "Ġbrutally": 69187, + "Ġbrute": 64098, + "Ġbryster": 46074, + "Ġbs": 17065, + "Ġbson": 50980, + "Ġbst": 70231, + "Ġbt": 19592, + "Ġbtc": 86037, + "Ġbtn": 3206, + "ĠbtnCancel": 87756, + "ĠbtnSave": 87364, + "Ġbtw": 89857, + "Ġbu": 1031, + "Ġbub": 77769, + "Ġbubb": 89612, + "Ġbubble": 23611, + "Ġbubbles": 43683, + "Ġbuc": 98044, + "Ġbuck": 22012, + "Ġbucket": 15621, + "Ġbuckets": 42112, + "Ġbuckle": 79995, + "Ġbucks": 47334, + "Ġbud": 36708, + "Ġbudd": 23485, + "Ġbuddies": 60350, + "Ġbudding": 85633, + "Ġbuddy": 36672, + "Ġbudget": 8039, + "Ġbudgets": 41384, + "Ġbuds": 67443, + "Ġbuen": 53713, + "Ġbuena": 64796, + "Ġbueno": 92541, + "Ġbuf": 6607, + "Ġbuff": 11522, + "Ġbuffalo": 81355, + "Ġbuffer": 4147, + "ĠbufferSize": 65158, + "Ġbuffered": 53057, + "Ġbuffering": 87139, + "Ġbuffers": 27389, + "Ġbuffet": 60786, + "Ġbuffs": 71897, + "Ġbufio": 96917, + "Ġbuflen": 88364, + "Ġbufsize": 97658, + "Ġbug": 9876, + "Ġbuggy": 79311, + "Ġbugs": 22551, + "Ġbugün": 131819, + "Ġbuild": 1936, + "Ġbuilder": 7363, + "Ġbuilders": 48086, + "Ġbuilding": 4752, + "Ġbuildings": 13702, + "Ġbuilds": 22111, + "Ġbuildup": 85665, + "Ġbuilt": 5798, + "Ġbuiltin": 59626, + "Ġbuiten": 98322, + "Ġbul": 7028, + "Ġbulb": 45812, + "Ġbulbs": 53220, + "Ġbulk": 19583, + "Ġbulky": 77821, + "Ġbull": 16811, + "Ġbulld": 78842, + "Ġbullet": 17432, + "Ġbulletin": 74729, + "Ġbullets": 33064, + "Ġbullied": 77813, + "Ġbullish": 89121, + "Ġbullpen": 81655, + "Ġbulls": 79681, + "Ġbullshit": 51301, + "Ġbully": 63076, + "Ġbullying": 44547, + "Ġbulun": 140373, + "Ġbulund": 135100, + "ĠbulunduÄŁu": 135101, + "Ġbulunmaktadır": 140374, + "ĠbuluÅŁ": 138895, + "Ġbum": 72874, + "Ġbump": 27575, + "Ġbumped": 78832, + "Ġbumper": 58885, + "Ġbumps": 64722, + "Ġbun": 44821, + "Ġbunch": 15493, + "Ġbund": 10365, + "Ġbundle": 12894, + "Ġbundled": 51450, + "Ġbundles": 48607, + "Ġbunk": 45469, + "Ġbunker": 83672, + "Ġbunny": 75514, + "Ġbuoy": 74620, + "Ġbur": 7795, + "Ġburden": 22305, + "Ġburdens": 62974, + "Ġbure": 30761, + "Ġbureau": 43302, + "Ġbureauc": 33746, + "Ġbureaucr": 36970, + "Ġbureaucracy": 63831, + "Ġbureaucratic": 79395, + "Ġbureaucrats": 88640, + "Ġburg": 29775, + "Ġburge": 86969, + "Ġburgeoning": 89569, + "Ġburger": 44623, + "Ġburgers": 62352, + "Ġburgl": 59171, + "Ġburglary": 76943, + "Ġburial": 54663, + "Ġburied": 26918, + "Ġburn": 8231, + "Ġburned": 26626, + "Ġburner": 64719, + "Ġburning": 19675, + "Ġburns": 43054, + "Ġburnt": 49340, + "Ġburst": 20524, + "Ġbursting": 76750, + "Ġbursts": 65701, + "Ġbury": 55344, + "Ġbus": 5828, + "Ġbusc": 63182, + "Ġbusca": 49814, + "Ġbuscar": 47374, + "Ġbuses": 33380, + "Ġbush": 29673, + "Ġbushes": 77206, + "Ġbusiest": 83508, + "Ġbusiness": 2562, + "Ġbusinesses": 9675, + "Ġbusinessman": 46278, + "Ġbusinessmen": 82550, + "Ġbust": 20781, + "Ġbusted": 76390, + "Ġbustling": 89156, + "Ġbusty": 65352, + "Ġbusy": 13028, + "Ġbut": 714, + "Ġbutcher": 75094, + "Ġbutt": 29956, + "Ġbutter": 14100, + "Ġbutterflies": 80676, + "Ġbutterfly": 55169, + "Ġbutterknife": 49939, + "Ġbutto": 96935, + "Ġbutton": 3137, + "ĠbuttonText": 86995, + "ĠbuttonWithType": 83824, + "Ġbuttons": 12424, + "Ġbuurt": 59017, + "Ġbuy": 3695, + "Ġbuyer": 25114, + "Ġbuyers": 22630, + "Ġbuying": 11833, + "Ġbuys": 49531, + "Ġbuz": 92154, + "Ġbuzz": 30427, + "Ġbuzzing": 85324, + "Ġbuá»ĵn": 136100, + "Ġbuá»ķi": 132214, + "Ġbuá»Ļc": 135571, + "Ġbv": 56937, + "Ġbw": 34375, + "Ġbx": 44241, + "Ġby": 553, + "Ġbye": 53041, + "Ġbypass": 30718, + "Ġbystand": 82383, + "Ġbyte": 4922, + "ĠbyteArray": 75003, + "Ġbytearray": 95603, + "Ġbytecode": 75129, + "Ġbytes": 5820, + "ĠbytesRead": 87173, + "ĠbyÄĩ": 59505, + "ĠbyÅĤ": 81635, + "ĠbyÅĤa": 130100, + "ĠbyÅĤo": 86517, + "ĠbyÅĤy": 131387, + "Ġbz": 80167, + "Ġbzw": 66809, + "Ġbác": 130103, + "Ġbán": 128627, + "Ġbánh": 132710, + "Ġbáo": 128603, + "Ġbás": 80351, + "Ġbásica": 140629, + "Ġbásico": 140807, + "Ġbât": 142620, + "Ġbâtiment": 142621, + "Ġbây": 136814, + "Ġbãi": 141109, + "Ġbä": 86553, + "ĠbÃ¥de": 79742, + "Ġbè": 136078, + "Ġbé": 89451, + "Ġbéné": 83133, + "Ġbénéfic": 141360, + "Ġbénéficie": 141361, + "Ġbên": 128608, + "Ġbình": 128823, + "Ġbó": 132195, + "Ġbóng": 129590, + "Ġbö": 98171, + "Ġböl": 125235, + "Ġbölge": 130209, + "Ġbölges": 142581, + "Ġbölgesinde": 142582, + "Ġbölüm": 127725, + "Ġbölümü": 138394, + "Ġbör": 91293, + "Ġböyle": 130378, + "Ġbú": 137495, + "Ġbúsqueda": 98499, + "Ġbü": 78475, + "Ġbüt": 130624, + "Ġbütün": 130625, + "Ġbüyü": 132288, + "Ġbüyük": 128565, + "ĠbÃł": 129302, + "ĠbÃłi": 128601, + "ĠbÃłn": 129030, + "ĠbÃło": 135774, + "ĠbÃły": 135247, + "ĠbÃŃ": 132642, + "Ġbır": 130147, + "Ġbırak": 130148, + "ĠbÄĥng": 139829, + "ĠbÄħd": 140861, + "ĠbÄħdź": 140862, + "ĠbÄĻd": 40808, + "ĠbÄĻdzie": 58231, + "ĠbÄĻdÄħ": 92494, + "ĠbÄĻdÄĻ": 142272, + "ĠbÅĤ": 142503, + "ĠbÅĤÄĻd": 142504, + "ĠbÆ°á»Ľc": 130203, + "Ġbạc": 134424, + "Ġbại": 139141, + "Ġbạn": 89437, + "Ġbản": 128450, + "Ġbảng": 134597, + "Ġbảo": 128417, + "Ġbất": 128633, + "Ġbầu": 134311, + "Ġbắt": 128802, + "Ġbằng": 128411, + "Ġbếp": 133651, + "ĠbáºŃ": 138269, + "ĠbáºŃc": 138270, + "ĠbáºŃt": 132345, + "Ġbụ": 127910, + "Ġbụi": 142522, + "Ġbụng": 137268, + "Ġbức": 133954, + "Ġbữa": 137790, + "Ġbá»ģ": 140836, + "Ġbá»ģn": 135268, + "Ġbá»ĩnh": 128453, + "Ġbá»ĭ": 128279, + "Ġbá»įn": 136497, + "Ġbá»ı": 129017, + "Ġbá»ij": 128888, + "Ġbá»ķ": 130623, + "Ġbá»Ļ": 128366, + "Ġbợi": 128944, + "Ġc": 272, + "ĠcJSON": 70580, + "Ġca": 2162, + "Ġcab": 21516, + "Ġcabbage": 73773, + "Ġcabe": 57779, + "Ġcabel": 95599, + "Ġcabeza": 94870, + "Ġcabeça": 140654, + "Ġcabin": 13386, + "Ġcabinet": 21921, + "Ġcabinets": 36228, + "Ġcabins": 88301, + "Ġcable": 14652, + "Ġcables": 36072, + "Ġcabo": 88384, + "Ġcac": 82881, + "Ġcach": 60287, + "Ġcache": 6500, + "Ġcached": 20579, + "Ġcaches": 53588, + "Ġcaching": 47430, + "Ġcad": 19409, + "Ġcada": 18862, + "Ġcadastr": 86177, + "Ġcade": 77318, + "Ġcadena": 54992, + "Ġcadre": 59597, + "Ġcaf": 29103, + "Ġcafe": 40930, + "Ġcafes": 73571, + "Ġcafeteria": 93848, + "Ġcaffe": 38570, + "Ġcaffeine": 48650, + "Ġcafé": 51950, + "Ġcage": 35360, + "Ġcages": 80847, + "Ġcairo": 53648, + "Ġcake": 19145, + "Ġcakes": 47569, + "Ġcal": 1620, + "Ġcalam": 79833, + "Ġcalc": 10035, + "Ġcalcium": 34619, + "Ġcalcul": 5812, + "Ġcalcular": 96527, + "Ġcalculate": 11047, + "Ġcalculated": 16588, + "Ġcalculates": 47417, + "Ġcalculating": 37614, + "Ġcalculation": 21937, + "Ġcalculations": 28117, + "Ġcalculator": 29952, + "Ġcalculus": 82668, + "Ġcalend": 49970, + "Ġcalendar": 13168, + "Ġcalendars": 70995, + "Ġcalf": 64543, + "Ġcaliber": 60768, + "Ġcalibrated": 84985, + "Ġcalibration": 37611, + "Ġcalidad": 55820, + "Ġcaliente": 97315, + "Ġcalifornia": 78857, + "Ġcall": 1618, + "ĠcallBack": 70897, + "Ġcallable": 40922, + "Ġcallback": 4822, + "Ġcallbacks": 26679, + "Ġcalle": 67835, + "Ġcalled": 2598, + "Ġcallee": 94800, + "Ġcaller": 19865, + "Ġcallers": 81903, + "Ġcalling": 8098, + "Ġcalloc": 46013, + "Ġcalls": 6738, + "Ġcalm": 19300, + "Ġcalming": 77629, + "Ġcalmly": 72070, + "Ġcalor": 85212, + "Ġcalorie": 51603, + "Ġcalories": 24262, + "Ġcalves": 96210, + "Ġcam": 6592, + "Ġcamar": 96586, + "Ġcamb": 24141, + "Ġcambi": 76603, + "Ġcambiar": 69720, + "Ġcambio": 57311, + "Ġcambios": 82262, + "Ġcame": 3697, + "Ġcamel": 49152, + "Ġcameo": 94056, + "Ġcamer": 92925, + "Ġcamera": 6249, + "Ġcameras": 18139, + "Ġcamino": 83444, + "Ġcamouflage": 86968, + "Ġcamp": 3113, + "Ġcampaign": 4796, + "Ġcampaigned": 86196, + "Ġcampaigners": 93293, + "Ġcampaigning": 54827, + "Ġcampaigns": 20686, + "Ġcampaña": 139024, + "Ġcamper": 91053, + "Ġcampground": 89135, + "Ġcamping": 32773, + "Ġcampo": 26575, + "Ġcampos": 50269, + "Ġcamps": 28035, + "Ġcampus": 15319, + "Ġcampuses": 51908, + "Ġcams": 58284, + "Ġcan": 646, + "ĠcanActivate": 58468, + "Ġcanada": 31763, + "Ġcanadian": 87598, + "Ġcanal": 38921, + "Ġcanc": 62309, + "Ġcancel": 9121, + "ĠcancelButton": 44026, + "ĠcancelButtonTitle": 76902, + "Ġcanceled": 33446, + "Ġcancell": 18467, + "Ġcancellation": 35835, + "ĠcancellationToken": 33133, + "Ġcancelled": 25681, + "Ġcancelling": 96915, + "Ġcancer": 9387, + "Ġcancers": 50323, + "Ġcand": 38638, + "Ġcandid": 5685, + "Ġcandidacy": 66160, + "Ġcandidate": 9144, + "Ġcandidates": 11178, + "Ġcandies": 97397, + "Ġcandle": 37799, + "Ġcandles": 51205, + "Ġcandy": 31556, + "Ġcane": 62235, + "Ġcanine": 74982, + "Ġcanlı": 134865, + "Ġcann": 13750, + "Ġcannabin": 64128, + "Ġcannabinoids": 92886, + "Ġcannabis": 18903, + "Ġcanned": 63014, + "Ġcannon": 48399, + "Ġcannons": 96914, + "Ġcannot": 4157, + "Ġcanoe": 83489, + "Ġcanon": 43816, + "Ġcanonical": 42453, + "Ġcanopy": 62971, + "Ġcans": 42632, + "Ġcant": 16464, + "Ġcantidad": 37276, + "Ġcanv": 72378, + "Ġcanvas": 10041, + "Ġcanyon": 83200, + "Ġcap": 2062, + "Ġcapabilities": 16928, + "Ġcapability": 22302, + "Ġcapable": 12875, + "Ġcapac": 23015, + "Ġcapacidad": 82292, + "Ġcapacit": 98658, + "Ġcapacities": 58439, + "Ġcapacitor": 84425, + "Ġcapacity": 8654, + "Ġcapacité": 139536, + "ĠcapacitÃł": 139336, + "Ġcape": 64292, + "Ġcapit": 60410, + "Ġcapita": 52055, + "Ġcapital": 6722, + "Ġcapitalism": 31582, + "Ġcapitalist": 39967, + "Ġcapitalists": 86308, + "Ġcapitalize": 52725, + "Ġcapitalized": 97321, + "Ġcapitals": 92999, + "Ġcapped": 60146, + "Ġcaps": 18568, + "Ġcapsule": 47639, + "Ġcapsules": 64779, + "Ġcapt": 6427, + "Ġcaptain": 23915, + "Ġcaptains": 89884, + "Ġcaptcha": 65089, + "Ġcaption": 17256, + "Ġcaptions": 77788, + "Ġcaptivating": 85182, + "Ġcaptive": 59851, + "Ġcaptivity": 82919, + "Ġcapture": 12322, + "Ġcaptured": 17006, + "Ġcaptures": 40155, + "Ġcapturing": 39780, + "ĠcapÃŃtulo": 143093, + "Ġcar": 1803, + "Ġcara": 46934, + "Ġcaract": 56605, + "Ġcaracter": 32229, + "Ġcaracteres": 59255, + "ĠcaracterÃŃsticas": 92454, + "Ġcaractère": 143704, + "Ġcaramel": 58711, + "Ġcaratter": 90650, + "Ġcaravan": 73660, + "Ġcarb": 34772, + "Ġcarbohydr": 40850, + "Ġcarbohydrate": 66632, + "Ġcarbohydrates": 61245, + "Ġcarbon": 12499, + "Ġcarbonate": 83343, + "Ġcarbs": 52509, + "Ġcarc": 74466, + "Ġcarcin": 51688, + "Ġcarcinoma": 88368, + "Ġcard": 3701, + "Ġcardboard": 53943, + "Ġcardi": 72051, + "Ġcardiac": 46245, + "Ġcardinal": 55880, + "Ġcardio": 31744, + "Ġcardiovascular": 40613, + "Ġcards": 7411, + "Ġcare": 2453, + "Ġcared": 41677, + "Ġcareer": 6931, + "Ġcareers": 30033, + "Ġcareful": 16585, + "Ġcarefully": 15516, + "Ġcareg": 49271, + "Ġcaregiver": 92461, + "Ġcaregivers": 73640, + "Ġcareless": 83734, + "Ġcares": 33572, + "Ġcaret": 52301, + "Ġcarg": 61662, + "Ġcarga": 75036, + "Ġcargar": 81911, + "Ġcargo": 25652, + "Ġcaric": 75971, + "Ġcaring": 29498, + "Ġcarn": 38338, + "Ġcarne": 94854, + "Ġcarniv": 96633, + "Ġcarnival": 93396, + "Ġcarousel": 52281, + "Ġcarp": 38639, + "Ġcarpet": 27854, + "Ġcarpets": 88241, + "Ġcarr": 47495, + "Ġcarrera": 89876, + "Ġcarriage": 51447, + "Ġcarried": 11691, + "Ġcarrier": 18602, + "Ġcarriers": 34891, + "Ġcarries": 23377, + "Ġcarro": 43900, + "Ġcarrot": 74194, + "Ġcarrots": 61417, + "Ġcarry": 6777, + "Ġcarrying": 15331, + "Ġcars": 9331, + "Ġcart": 7406, + "Ġcarta": 75547, + "Ġcarte": 47354, + "Ġcartel": 79640, + "Ġcartesian": 94381, + "Ġcartoon": 30349, + "Ġcartoons": 62023, + "Ġcartridge": 50862, + "Ġcartridges": 71733, + "Ġcarts": 65404, + "Ġcarve": 79637, + "Ġcarved": 48167, + "Ġcarving": 76691, + "Ġcará": 143158, + "Ġcarácter": 143159, + "Ġcas": 4760, + "Ġcasa": 24249, + "Ġcasc": 74957, + "Ġcascade": 42018, + "Ġcase": 1142, + "Ġcases": 5048, + "Ġcash": 8350, + "Ġcashier": 90355, + "Ġcasi": 61253, + "Ġcasing": 73344, + "Ġcasino": 11846, + "Ġcasinos": 33371, + "Ġcaso": 23020, + "Ġcasos": 57533, + "Ġcass": 44224, + "Ġcassette": 81258, + "Ġcast": 6311, + "Ġcaste": 68424, + "Ġcaster": 47982, + "Ġcasting": 24172, + "Ġcastle": 32584, + "Ġcasts": 56033, + "Ġcasual": 16334, + "Ġcasually": 64614, + "Ġcasualties": 47888, + "Ġcasualty": 91497, + "Ġcat": 8251, + "Ġcata": 30079, + "Ġcatal": 33354, + "Ġcatalog": 16403, + "Ġcatalogs": 90285, + "Ġcatalogue": 48539, + "Ġcatalyst": 50383, + "Ġcatapult": 95091, + "Ġcatast": 34105, + "Ġcatastrophe": 64296, + "Ġcatastrophic": 53577, + "Ġcatch": 2287, + "ĠcatchError": 63934, + "Ġcatcher": 67798, + "Ġcatches": 37834, + "Ġcatching": 33068, + "Ġcatchy": 85505, + "Ġcate": 58139, + "Ġcategor": 22049, + "Ġcategoria": 47018, + "Ġcategorical": 69536, + "Ġcategorie": 95819, + "Ġcategories": 11059, + "Ġcategorized": 70874, + "Ġcategory": 5582, + "ĠcategoryId": 58993, + "ĠcategoryName": 89610, + "ĠcategorÃŃa": 140647, + "Ġcater": 27968, + "Ġcatering": 53829, + "Ġcath": 30647, + "Ġcathedral": 79150, + "Ġcatholic": 82091, + "Ġcats": 19423, + "Ġcattle": 35852, + "Ġcatégorie": 142130, + "Ġcaucus": 65443, + "Ġcaught": 10568, + "Ġcaul": 85164, + "Ġcauliflower": 95870, + "Ġcaus": 24524, + "Ġcausa": 61999, + "Ġcausal": 58457, + "Ġcause": 5240, + "Ġcaused": 8881, + "Ġcauses": 11137, + "Ġcausing": 14381, + "Ġcaut": 35104, + "Ġcaution": 27830, + "Ġcautioned": 84332, + "Ġcautious": 45778, + "Ġcautiously": 91385, + "Ġcav": 56609, + "Ġcaval": 59290, + "Ġcavalry": 78557, + "Ġcave": 25385, + "Ġcaveat": 86051, + "Ġcavern": 75204, + "Ġcaves": 65564, + "Ġcavity": 55329, + "Ġcaz": 95085, + "Ġcazzo": 89533, + "ĠcaÅĤ": 90061, + "ĠcaÅĤego": 142602, + "ĠcaÅĤej": 140559, + "ĠcaÅĤy": 132186, + "Ġcb": 9858, + "Ġcbd": 58527, + "Ġcbo": 93677, + "Ġcc": 12527, + "Ġccp": 98897, + "Ġcd": 15307, + "Ġcdecl": 87174, + "Ġcdr": 88304, + "Ġcds": 93501, + "Ġce": 3761, + "Ġcease": 31516, + "Ġceased": 58377, + "Ġceasefire": 81773, + "Ġcedar": 75296, + "Ġceil": 28481, + "Ġceiling": 22170, + "Ġceilings": 55714, + "Ġcel": 19093, + "Ġcela": 44502, + "Ġcele": 6494, + "Ġcelebr": 7345, + "Ġcelebrate": 18383, + "Ġcelebrated": 27185, + "Ġcelebrates": 48093, + "Ġcelebrating": 31589, + "Ġcelebration": 24916, + "Ġcelebrations": 46574, + "Ġcelebrities": 39401, + "Ġcelebrity": 30444, + "Ġcelery": 69021, + "Ġcelestial": 76380, + "Ġcelib": 77533, + "Ġcell": 2779, + "ĠcellFor": 40604, + "ĠcellForRowAt": 82655, + "ĠcellForRowAtIndexPath": 65889, + "Ġcellar": 75775, + "Ġcelle": 72003, + "Ġcellpadding": 29104, + "Ġcellphone": 57309, + "Ġcells": 7761, + "Ġcellspacing": 27081, + "Ġcellul": 68902, + "Ġcellular": 34593, + "Ġcelui": 64990, + "Ġcelular": 87126, + "Ġcement": 23613, + "Ġcemetery": 58304, + "Ġcen": 64931, + "Ġcena": 77906, + "Ġcensor": 38965, + "Ġcensorship": 51888, + "Ġcensus": 43602, + "Ġcent": 2889, + "Ġcenter": 4126, + "ĠcenterX": 89801, + "ĠcenterY": 80230, + "Ġcentered": 30188, + "Ġcenterpiece": 90753, + "Ġcenters": 18652, + "Ġcentr": 38139, + "Ġcentral": 8622, + "Ġcentralized": 57883, + "Ġcentrally": 84313, + "Ġcentre": 12261, + "Ġcentres": 35182, + "Ġcentrif": 73082, + "Ġcentro": 40887, + "Ġcentroid": 58119, + "Ġcentroids": 80043, + "Ġcents": 30191, + "Ġcenturies": 23631, + "Ġcentury": 9294, + "Ġcep": 62090, + "Ġcer": 10152, + "Ġceramic": 42024, + "Ġceramics": 95706, + "Ġcerc": 46462, + "Ġcerca": 27449, + "Ġcerco": 70917, + "Ġcere": 26993, + "Ġcereal": 61133, + "Ġcerebral": 59645, + "Ġceremon": 42512, + "Ġceremonial": 92542, + "Ġceremonies": 59183, + "Ġceremony": 21536, + "Ġcerr": 26878, + "Ġcert": 2777, + "Ġcertain": 3654, + "Ġcertainly": 7838, + "Ġcertains": 74976, + "Ġcertainty": 43643, + "Ġcertificate": 15748, + "Ġcertificates": 34437, + "Ġcertification": 27606, + "Ġcertifications": 81471, + "Ġcertified": 22909, + "Ġcertify": 92230, + "Ġcerto": 87915, + "Ġcerts": 74406, + "Ġcerv": 51564, + "Ġcervical": 66727, + "Ġces": 26652, + "Ġcess": 64047, + "Ġcessation": 82360, + "Ġcet": 40967, + "Ġcette": 20061, + "Ġceux": 64956, + "Ġcf": 24111, + "Ġcfg": 13286, + "Ġcg": 35740, + "Ġcgi": 98832, + "Ġch": 521, + "Ġcha": 13517, + "Ġchac": 94817, + "Ġchai": 44081, + "Ġchain": 8781, + "Ġchained": 66969, + "Ġchaining": 73662, + "Ġchains": 26179, + "Ġchair": 10496, + "Ġchaired": 98106, + "Ġchairman": 21201, + "Ġchairs": 25904, + "Ġchalk": 42359, + "Ġchall": 4799, + "Ġchalleng": 7301, + "Ġchallenge": 8645, + "Ġchallenged": 28891, + "Ġchallenger": 69150, + "Ġchallenges": 11513, + "Ġchallenging": 17003, + "Ġcham": 33709, + "Ġchamber": 24220, + "Ġchambers": 52179, + "Ġchambre": 86838, + "Ġchamp": 34541, + "Ġchampagne": 64293, + "Ġchampion": 18319, + "Ġchampions": 33738, + "Ġchampionship": 21553, + "Ġchampionships": 58184, + "Ġchamps": 70481, + "Ġchan": 26023, + "Ġchance": 6012, + "Ġchancellor": 76175, + "Ġchances": 16963, + "Ġchang": 2547, + "Ġchange": 2297, + "Ġchanged": 5497, + "Ġchanger": 61339, + "Ġchanges": 4344, + "Ġchanging": 10018, + "Ġchannel": 5496, + "ĠchannelId": 81725, + "Ġchannels": 11744, + "Ġchant": 43353, + "Ġchanting": 76574, + "Ġchants": 87487, + "Ġchaos": 26915, + "Ġchaotic": 51987, + "Ġchap": 37738, + "Ġchapel": 82149, + "Ġchapter": 12453, + "Ġchapters": 29632, + "Ġchaque": 45677, + "Ġchar": 1161, + "Ġcharacter": 3668, + "Ġcharacteristic": 28583, + "Ġcharacteristics": 17452, + "Ġcharacterization": 59893, + "Ġcharacterize": 69655, + "Ġcharacterized": 31871, + "Ġcharacters": 5766, + "Ġcharcoal": 60103, + "Ġcharg": 5597, + "Ġcharge": 6757, + "Ġcharged": 11430, + "Ġcharger": 44569, + "Ġcharges": 10195, + "Ġcharging": 22641, + "Ġcharisma": 98649, + "Ġcharismatic": 78223, + "Ġcharitable": 47701, + "Ġcharities": 50271, + "Ġcharity": 22846, + "Ġcharm": 31253, + "Ġcharming": 34409, + "Ġcharms": 80650, + "Ġchars": 23000, + "Ġcharset": 11617, + "Ġchart": 9487, + "Ġcharter": 37024, + "Ġcharts": 26131, + "Ġchase": 32486, + "Ġchased": 61820, + "Ġchasing": 42831, + "Ġchassis": 44909, + "Ġchast": 74107, + "Ġchat": 6236, + "Ġchats": 48526, + "Ġchatt": 80242, + "Ġchatte": 53851, + "Ġchatter": 69357, + "Ġchatting": 50967, + "Ġchaud": 69312, + "Ġchauff": 73717, + "Ġchave": 97635, + "Ġchc": 141636, + "Ġchcia": 134685, + "ĠchciaÅĤ": 134686, + "ĠchcÄħ": 141637, + "Ġche": 3016, + "Ġcheap": 11872, + "Ġcheaper": 23048, + "Ġcheapest": 42049, + "Ġcheat": 40768, + "Ġcheated": 77678, + "Ġcheating": 41723, + "Ġcheats": 75388, + "Ġcheck": 1779, + "ĠcheckBox": 52558, + "Ġcheckbox": 31300, + "Ġcheckboxes": 90392, + "Ġchecked": 10067, + "Ġchecker": 40915, + "Ġchecking": 13295, + "Ġchecklist": 52573, + "Ġcheckout": 27264, + "Ġcheckpoint": 29295, + "Ġcheckpoints": 67209, + "Ġchecks": 12341, + "Ġchecksum": 32529, + "Ġcheek": 40703, + "Ġcheeks": 53788, + "Ġcheer": 25032, + "Ġcheered": 84523, + "Ġcheerful": 70314, + "Ġcheering": 67902, + "Ġcheers": 73883, + "Ġchees": 42832, + "Ġcheese": 17163, + "Ġcheeses": 97464, + "Ġcheesy": 86747, + "Ġchef": 29706, + "Ġchefs": 59127, + "Ġcheg": 96940, + "Ġchem": 8424, + "Ġchemical": 11483, + "Ġchemicals": 25266, + "Ġchemin": 87668, + "Ġchemistry": 29803, + "Ġchemotherapy": 61630, + "Ġcheque": 77010, + "Ġcher": 22216, + "Ġcherche": 51471, + "Ġcherish": 86685, + "Ġcherished": 80071, + "Ġcherry": 40880, + "Ġchess": 32719, + "Ġchest": 15138, + "Ġchests": 83318, + "Ġchevy": 96910, + "Ġchew": 36333, + "Ġchewing": 74377, + "Ġchez": 45106, + "Ġchi": 25798, + "Ġchiar": 98099, + "Ġchic": 46894, + "Ġchica": 79588, + "Ġchicago": 92208, + "Ġchicas": 56567, + "Ġchick": 30763, + "Ġchicken": 16158, + "Ġchickens": 53523, + "Ġchicks": 67989, + "Ġchief": 10178, + "Ġchiefly": 88282, + "Ġchiefs": 67702, + "Ġchiff": 82964, + "Ġchild": 1682, + "Ġchildbirth": 96194, + "Ġchildcare": 79171, + "Ġchildhood": 19990, + "Ġchildish": 93171, + "Ġchildren": 2841, + "Ġchilds": 88000, + "Ġchili": 58617, + "Ġchill": 36301, + "Ġchilled": 83942, + "Ġchilling": 61624, + "Ġchilly": 80875, + "Ġchim": 36994, + "Ġchimney": 89051, + "Ġchimpan": 98325, + "Ġchin": 45075, + "Ġchina": 32525, + "Ġchinese": 56387, + "Ġchip": 16392, + "Ġchips": 23595, + "Ġchipset": 91714, + "Ġchir": 70446, + "Ġchiropr": 76201, + "Ġchiếc": 129141, + "Ġchiếm": 136514, + "Ġchiến": 129009, + "Ġchiếu": 135714, + "Ġchiá»ģu": 131149, + "Ġchk": 39242, + "Ġchlor": 36733, + "Ġchloride": 81782, + "Ġchlorine": 84106, + "Ġchmod": 59615, + "Ġcho": 2600, + "Ġchoc": 141608, + "Ġchocia": 141609, + "Ġchociaż": 141610, + "Ġchocol": 71376, + "Ġchocolate": 17931, + "Ġchocolates": 98949, + "Ġchoice": 5754, + "Ġchoices": 11454, + "Ġchoir": 70608, + "Ġchois": 72561, + "Ġchoisir": 89094, + "Ġchoix": 58920, + "Ġchoke": 64036, + "Ġchoked": 93643, + "Ġchoking": 90520, + "Ġchol": 98640, + "Ġcholesterol": 37986, + "Ġchoose": 5157, + "Ġchooser": 94253, + "Ġchooses": 39911, + "Ġchoosing": 18774, + "Ġchop": 46994, + "Ġchopped": 37425, + "Ġchopping": 96287, + "Ġchops": 90078, + "Ġchor": 69256, + "Ġchord": 43221, + "Ġchords": 55659, + "Ġchore": 49571, + "Ġchores": 81469, + "Ġchorus": 55810, + "Ġchose": 14554, + "Ġchosen": 11882, + "Ġchoses": 78065, + "ĠchoÄĩ": 134811, + "Ġchr": 28131, + "Ġchrist": 25769, + "Ġchristian": 51047, + "Ġchristmas": 53202, + "Ġchrom": 21372, + "Ġchromat": 97184, + "Ġchrome": 26429, + "Ġchromium": 96862, + "Ġchromosome": 50715, + "Ġchromosomes": 82081, + "Ġchron": 25986, + "Ġchronic": 20601, + "Ġchrono": 80372, + "Ġchronological": 86534, + "Ġchu": 95799, + "Ġchubby": 93778, + "Ġchuck": 42460, + "Ġchuckled": 96611, + "Ġchunk": 11879, + "Ġchunks": 26757, + "Ġchurch": 8817, + "Ġchurches": 29912, + "Ġchurn": 80869, + "Ġchute": 94380, + "Ġchuyên": 128744, + "Ġchuyến": 135418, + "Ġchuyá»ĥn": 128830, + "Ġchuyá»ĩn": 128963, + "Ġchuẩn": 129311, + "Ġchá": 127832, + "Ġcháu": 136291, + "Ġcháy": 137297, + "Ġchân": 129213, + "Ġchâu": 133539, + "Ġchó": 137665, + "Ġchóng": 131827, + "Ġchù": 142937, + "Ġchùa": 142938, + "Ġchú": 129407, + "Ġchúng": 128347, + "Ġchút": 130575, + "ĠchÃłng": 136862, + "ĠchÃło": 139548, + "ĠchÃŃ": 128997, + "ĠchÃŃnh": 128311, + "ĠchÄĥm": 131619, + "ĠchÆ¡i": 128971, + "Ġchưa": 128651, + "Ġchương": 129784, + "Ġchạm": 141619, + "Ġchạy": 130702, + "Ġchảy": 138886, + "Ġchấp": 132677, + "Ġchất": 128380, + "Ġchắc": 130052, + "Ġchắn": 132457, + "Ġchẳng": 130685, + "Ġchặ": 139809, + "Ġchặn": 139810, + "Ġchặt": 137277, + "Ġchế": 128751, + "Ġchết": 130276, + "ĠcháºŃm": 138117, + "Ġchợ": 136625, + "Ġchụ": 135901, + "Ġchụp": 135902, + "Ġchá»§": 128427, + "Ġchứ": 127306, + "Ġchứa": 133456, + "Ġchức": 128502, + "Ġchứng": 128977, + "Ġchữ": 132870, + "Ġchữa": 130489, + "Ġchá»ī": 98127, + "Ġchá»īnh": 132314, + "Ġchá»ĭ": 126555, + "Ġchá»ĭu": 129949, + "Ġchá»įn": 128509, + "Ġchá»ijng": 129455, + "Ġchá»ĵng": 130627, + "Ġchá»Ĺ": 131501, + "Ġchá»Ŀ": 133459, + "Ġci": 11825, + "Ġcialis": 27793, + "ĠciaÅĤa": 139949, + "Ġcic": 53777, + "Ġciclo": 89570, + "Ġcid": 32141, + "Ġcidade": 55084, + "Ġcider": 65055, + "Ġcient": 87997, + "ĠcientÃŃf": 143249, + "ĠcientÃŃfico": 143250, + "Ġcif": 65472, + "Ġcig": 16934, + "Ġcigar": 53674, + "Ġcigaret": 21592, + "Ġcigarette": 35113, + "Ġcigarettes": 33476, + "Ġcigars": 81115, + "Ġcil": 61344, + "Ġcin": 8461, + "Ġcinco": 70327, + "Ġcine": 79752, + "Ġcinema": 33192, + "Ġcinemas": 97607, + "Ġcinemat": 87022, + "Ġcinematic": 64665, + "Ġcinnamon": 50892, + "Ġciné": 143376, + "Ġcinéma": 143377, + "Ġcio": 142972, + "Ġcioè": 142973, + "Ġcipher": 31088, + "Ġciphertext": 69789, + "Ġcir": 29470, + "Ġcirc": 4225, + "Ġcirca": 53507, + "Ġcircle": 12671, + "Ġcircles": 25362, + "Ġcircuit": 16224, + "Ġcircuits": 45021, + "Ġcircular": 26931, + "Ġcirculated": 62873, + "Ġcirculating": 53728, + "Ġcirculation": 34755, + "Ġcircum": 10198, + "Ġcircumcision": 73509, + "Ġcircumference": 74926, + "Ġcircumstance": 52137, + "Ġcircumstances": 13161, + "Ġcircus": 67024, + "Ġcis": 66404, + "Ġcit": 6544, + "Ġcita": 87057, + "Ġcitas": 54535, + "Ġcitation": 44377, + "Ġcitations": 51846, + "Ġcite": 38296, + "Ġcited": 21870, + "Ġcites": 57173, + "Ġcities": 9720, + "Ġciting": 32164, + "Ġcitiz": 8435, + "Ġcitizen": 21860, + "Ġcitizens": 10283, + "Ġcitizenship": 32679, + "Ġcitrus": 59190, + "Ġcitt": 71176, + "ĠcittÃł": 96939, + "Ġcity": 3283, + "ĠcityName": 96701, + "Ġciudad": 39357, + "Ġciv": 18462, + "Ġcivic": 43873, + "Ġcivil": 8267, + "Ġcivilian": 28928, + "Ġcivilians": 28436, + "Ġcivilization": 34917, + "Ġcivilizations": 85469, + "Ġcivilized": 84979, + "Ġciò": 133010, + "ĠciÄħ": 137922, + "ĠciÄħgu": 137923, + "Ġcj": 90997, + "Ġck": 38613, + "Ġcl": 1185, + "Ġcla": 86704, + "Ġclad": 82758, + "Ġclaim": 3717, + "Ġclaimed": 11660, + "Ġclaiming": 20409, + "Ġclaims": 8186, + "Ġclam": 59510, + "Ġclamp": 41904, + "Ġclan": 38462, + "Ġclandest": 92557, + "Ġclang": 18260, + "Ġclans": 85981, + "Ġclap": 89577, + "Ġclar": 19497, + "Ġclarification": 63684, + "Ġclarified": 64776, + "Ġclarify": 37163, + "Ġclarity": 31273, + "Ġclaro": 75221, + "Ġclas": 56120, + "Ġclase": 46380, + "Ġclases": 95573, + "Ġclash": 39903, + "Ġclashed": 97406, + "Ġclashes": 51510, + "Ġclass": 536, + "ĠclassName": 1980, + "ĠclassNames": 55465, + "Ġclasse": 36317, + "Ġclasses": 6846, + "Ġclassic": 11416, + "Ġclassical": 28824, + "Ġclassics": 52705, + "Ġclassification": 23850, + "Ġclassifications": 95671, + "Ġclassified": 21091, + "Ġclassifier": 33365, + "Ġclassifiers": 71291, + "Ġclassify": 48129, + "Ġclassmates": 60090, + "Ġclassname": 70650, + "Ġclassroom": 24017, + "Ġclassrooms": 57116, + "Ġclassy": 70258, + "Ġclause": 21647, + "Ġclauses": 49098, + "Ġclave": 60842, + "Ġclaw": 56490, + "Ġclaws": 67450, + "Ġclay": 36048, + "Ġclazz": 32472, + "Ġcle": 11299, + "Ġclean": 4240, + "Ġcleaned": 27722, + "Ġcleaner": 31881, + "Ġcleaners": 74865, + "Ġcleaning": 15826, + "Ġcleanliness": 95291, + "Ġcleanly": 96875, + "Ġcleans": 34194, + "Ġcleanse": 80055, + "Ġcleansing": 58952, + "Ġcleanup": 21290, + "Ġclear": 2797, + "ĠclearColor": 57978, + "ĠclearInterval": 45585, + "ĠclearTimeout": 53225, + "Ġclearance": 35554, + "Ġcleared": 22949, + "Ġclearer": 48379, + "Ġclearfix": 36627, + "Ġclearing": 32750, + "Ġclearly": 9355, + "Ġclears": 56598, + "Ġclen": 84934, + "Ġcler": 56360, + "Ġclergy": 62484, + "Ġcleric": 91558, + "Ġclerk": 44257, + "Ġclever": 27699, + "Ġclf": 46079, + "Ġcli": 21348, + "Ġclic": 47276, + "Ġclich": 62592, + "Ġclick": 4205, + "Ġclickable": 85563, + "Ġclicked": 19910, + "Ġclicking": 18458, + "Ġclicks": 27749, + "Ġclient": 2943, + "ĠclientId": 47983, + "Ġcliente": 26629, + "Ġclientele": 96172, + "Ġclientes": 44645, + "Ġclients": 8239, + "Ġcliff": 43006, + "Ġcliffs": 85849, + "Ġclim": 11076, + "Ġclimate": 9977, + "Ġclimates": 91299, + "Ġclimax": 70195, + "Ġclimb": 25367, + "Ġclimbed": 44419, + "Ġclimbers": 93651, + "Ġclimbing": 29508, + "Ġclimbs": 76411, + "Ġclin": 9218, + "Ġcling": 77164, + "Ġclinging": 96358, + "Ġclinic": 27813, + "Ġclinical": 14490, + "Ġclinically": 69332, + "Ġclinicians": 77445, + "Ġclinics": 43235, + "Ġclip": 12327, + "Ġclipboard": 46034, + "Ġclipped": 73059, + "Ġclipping": 61856, + "Ġclips": 26111, + "Ġclique": 80379, + "Ġclit": 91762, + "Ġcljs": 26284, + "Ġclk": 45798, + "Ġclo": 5292, + "Ġcloak": 61394, + "Ġclock": 8866, + "Ġclocks": 50337, + "Ġclockwise": 65670, + "Ġclone": 14715, + "Ġcloned": 54119, + "Ġclones": 65907, + "Ġcloning": 77866, + "Ġclos": 4496, + "Ġclose": 3265, + "ĠcloseButton": 88314, + "ĠcloseModal": 94248, + "Ġclosed": 7877, + "Ġclosely": 15148, + "Ġcloser": 12128, + "Ġcloses": 33250, + "Ġclosest": 18093, + "Ġcloset": 31944, + "Ġclosets": 94740, + "Ġclosing": 15316, + "Ġclosure": 21955, + "Ġclosures": 59900, + "Ġclot": 93163, + "Ġcloth": 27292, + "Ġclothes": 15097, + "Ġclothing": 17438, + "Ġcloud": 9437, + "Ġclouds": 29514, + "Ġcloudy": 73549, + "Ġcloves": 84288, + "Ġclown": 68468, + "Ġclr": 62127, + "Ġcls": 12790, + "Ġclub": 6335, + "Ġclubhouse": 93433, + "Ġclubs": 18890, + "Ġclue": 29989, + "Ġclues": 42675, + "Ġclums": 76957, + "Ġclumsy": 93218, + "Ġcluster": 10652, + "Ġclustered": 92758, + "Ġclustering": 58354, + "Ġclusters": 26968, + "Ġclutch": 42689, + "Ġclutter": 53816, + "Ġcm": 9961, + "Ġcmake": 92996, + "Ġcmap": 40778, + "Ġcmb": 60893, + "Ġcmd": 5439, + "Ġcmdline": 94106, + "Ġcmds": 47657, + "Ġcmp": 26089, + "Ġcms": 26270, + "Ġcn": 13665, + "Ġcname": 78514, + "Ġcnn": 60458, + "Ġcnt": 13195, + "Ġco": 1062, + "Ġcoach": 7247, + "Ġcoached": 69922, + "Ġcoaches": 23956, + "Ġcoaching": 23883, + "Ġcoal": 11497, + "Ġcoalition": 25217, + "Ġcoarse": 49247, + "Ġcoast": 13648, + "Ġcoastal": 34235, + "Ġcoaster": 81068, + "Ġcoastline": 79844, + "Ġcoat": 22875, + "Ġcoated": 48382, + "Ġcoating": 40294, + "Ġcoatings": 90925, + "Ġcoats": 55637, + "Ġcoax": 90866, + "Ġcob": 33828, + "Ġcoc": 21690, + "Ġcocaine": 38023, + "Ġcocci": 94909, + "Ġcoch": 79081, + "Ġcocina": 89606, + "Ġcock": 11272, + "Ġcockpit": 59624, + "Ġcocks": 68440, + "Ġcocktail": 39910, + "Ġcocktails": 56976, + "Ġcoco": 82350, + "Ġcocoa": 61788, + "Ġcoconut": 33457, + "Ġcocos": 78092, + "Ġcod": 20329, + "Ġcode": 2038, + "Ġcodec": 34647, + "Ġcodecs": 56152, + "Ġcoded": 46673, + "Ġcoder": 83542, + "Ġcodes": 13912, + "Ġcodigo": 34778, + "Ġcoding": 10822, + "Ġcoef": 55533, + "Ġcoeff": 17593, + "Ġcoefficient": 35606, + "Ġcoefficients": 36829, + "Ġcoeffs": 79172, + "Ġcoer": 93185, + "Ġcoerc": 54263, + "Ġcoerce": 83025, + "Ġcoercion": 77142, + "Ġcof": 68803, + "Ġcoff": 9894, + "Ġcoffee": 10799, + "Ġcoffin": 77666, + "Ġcog": 61566, + "Ġcogn": 18799, + "Ġcognition": 74210, + "Ġcognitive": 24675, + "Ġcoh": 70085, + "Ġcoherence": 77825, + "Ġcoherent": 55787, + "Ġcohesion": 95293, + "Ġcohesive": 86416, + "Ġcohort": 40844, + "Ġcohorts": 89288, + "Ġcoil": 39660, + "Ġcoils": 70760, + "Ġcoin": 16254, + "Ġcoinc": 22970, + "Ġcoincide": 71259, + "Ġcoincidence": 49178, + "Ġcoined": 77618, + "Ġcoins": 18762, + "Ġcoisa": 96700, + "Ġcoke": 83323, + "Ġcol": 1375, + "Ġcola": 83667, + "Ġcolabor": 69872, + "Ġcold": 9255, + "Ġcolder": 75114, + "Ġcole": 72941, + "Ġcoleg": 89055, + "Ġcoli": 73010, + "Ġcoll": 4530, + "Ġcollabor": 11182, + "Ġcollaborate": 50596, + "Ġcollaborated": 77074, + "Ġcollaborating": 72201, + "Ġcollaboration": 20031, + "Ġcollaborations": 82563, + "Ġcollaborative": 39706, + "Ġcollaborators": 78019, + "Ġcollage": 96902, + "Ġcollagen": 70213, + "Ġcollaps": 42023, + "Ġcollapse": 18179, + "Ġcollapsed": 28268, + "Ġcollapses": 86358, + "Ġcollapsing": 71518, + "Ġcollar": 36104, + "Ġcollateral": 51408, + "Ġcolle": 14148, + "Ġcolleague": 36344, + "Ġcolleagues": 17639, + "Ġcollect": 6530, + "Ġcollected": 14548, + "Ġcollecting": 25895, + "Ġcollection": 4426, + "ĠcollectionView": 26535, + "Ġcollections": 15302, + "Ġcollective": 21750, + "Ġcollectively": 44825, + "Ġcollector": 31953, + "Ġcollectors": 51448, + "Ġcollects": 51207, + "Ġcolleg": 24499, + "Ġcollege": 7770, + "Ġcolleges": 30152, + "Ġcollegiate": 94697, + "Ġcollide": 66090, + "Ġcollided": 78890, + "Ġcollider": 65544, + "Ġcollision": 18752, + "Ġcollisions": 47353, + "Ġcollo": 80948, + "Ġcollusion": 68337, + "Ġcoloc": 58740, + "Ġcolomb": 74224, + "Ġcolon": 14889, + "Ġcolonial": 36583, + "Ġcolonies": 47928, + "Ġcolonization": 95453, + "Ġcolony": 40936, + "Ġcolor": 1894, + "ĠcolorWith": 41511, + "ĠcolorWithRed": 61180, + "Ġcolore": 78787, + "Ġcolored": 27197, + "Ġcolorful": 33866, + "Ġcoloring": 34887, + "Ġcolormap": 76010, + "Ġcolors": 7987, + "Ġcolossal": 96837, + "Ġcolour": 12463, + "Ġcoloured": 57819, + "Ġcolourful": 69367, + "Ġcolours": 26138, + "Ġcols": 17000, + "Ġcolspan": 16848, + "Ġcolum": 93225, + "Ġcolumn": 3250, + "ĠcolumnHeader": 95624, + "ĠcolumnIndex": 66664, + "ĠcolumnName": 65746, + "Ġcolumna": 75299, + "Ġcolumnist": 53087, + "Ġcolumns": 8147, + "Ġcom": 469, + "Ġcoma": 69298, + "Ġcomando": 49868, + "Ġcomb": 3614, + "Ġcombat": 12610, + "Ġcombating": 95078, + "Ġcombin": 67125, + "Ġcombination": 10601, + "Ġcombinations": 27459, + "Ġcombine": 15963, + "ĠcombineReducers": 57943, + "Ġcombined": 10856, + "Ġcombines": 32411, + "Ġcombining": 34171, + "Ġcombo": 22734, + "ĠcomboBox": 42711, + "Ġcombos": 70133, + "Ġcombust": 43445, + "Ġcombustion": 58233, + "Ġcome": 2525, + "Ġcomeback": 46537, + "Ġcomed": 62013, + "Ġcomedian": 50812, + "Ġcomedic": 94371, + "Ġcomedy": 22358, + "Ġcoment": 32357, + "Ġcomentario": 82717, + "Ġcomentarios": 79412, + "Ġcomentário": 132365, + "Ġcomenz": 74109, + "Ġcomer": 90327, + "Ġcomerc": 70909, + "Ġcomercial": 73861, + "Ġcomes": 4041, + "Ġcomet": 69986, + "Ġcomeç": 81147, + "Ġcomeça": 136997, + "Ġcomeçar": 139143, + "Ġcomeçou": 142891, + "Ġcomfort": 6838, + "Ġcomfortable": 10655, + "Ġcomfortably": 44669, + "Ġcomforting": 68030, + "Ġcomforts": 96081, + "Ġcomfy": 59321, + "Ġcomic": 19724, + "Ġcomics": 28059, + "Ġcomida": 97778, + "Ġcoming": 5001, + "Ġcomm": 1063, + "Ġcomma": 31683, + "Ġcommand": 3210, + "Ġcommande": 67751, + "Ġcommanded": 53657, + "Ġcommander": 27994, + "Ġcommanders": 60327, + "Ġcommanding": 64340, + "Ġcommands": 11293, + "Ġcommas": 76602, + "Ġcomme": 21572, + "Ġcommem": 41217, + "Ġcommemor": 80054, + "Ġcommemorate": 82523, + "Ġcommenc": 57962, + "Ġcommence": 55345, + "Ġcommenced": 64262, + "Ġcommencement": 68444, + "Ġcommend": 73112, + "Ġcomment": 3980, + "Ġcommentaire": 86989, + "Ġcommentary": 30610, + "Ġcommentator": 68736, + "Ġcommentators": 62494, + "Ġcommented": 28686, + "Ġcommenter": 96066, + "Ġcommenting": 40265, + "Ġcomments": 6042, + "Ġcommerc": 94094, + "Ġcommerce": 35654, + "Ġcommercial": 8353, + "Ġcommercially": 53353, + "Ġcommercials": 62495, + "Ġcommission": 12123, + "Ġcommissioned": 43124, + "Ġcommissioner": 40343, + "Ġcommissioners": 73555, + "Ġcommissions": 53664, + "Ġcommit": 5266, + "Ġcommitment": 15155, + "Ġcommitments": 41256, + "Ġcommits": 41535, + "Ġcommitted": 11163, + "Ġcommittee": 12801, + "Ġcommittees": 41447, + "Ġcommitting": 39483, + "Ġcommod": 25550, + "Ġcommodities": 50867, + "Ġcommodity": 37883, + "Ġcommodo": 78324, + "Ġcommon": 4185, + "Ġcommonly": 16626, + "Ġcommonplace": 76914, + "Ġcommons": 64962, + "Ġcommun": 33461, + "Ġcommunal": 56837, + "Ġcommunauté": 141032, + "Ġcommune": 78145, + "Ġcommunic": 5932, + "Ġcommunicate": 19032, + "Ġcommunicated": 57935, + "Ġcommunicates": 91506, + "Ġcommunicating": 41147, + "Ġcommunication": 10535, + "Ġcommunications": 16893, + "Ġcommunicator": 86671, + "Ġcommunion": 93032, + "Ġcommunism": 70089, + "Ġcommunist": 49215, + "Ġcommunities": 10744, + "Ġcommunity": 3942, + "Ġcommute": 58163, + "Ġcommuter": 88003, + "Ġcommuters": 81721, + "Ġcommuting": 93850, + "Ġcomo": 7953, + "Ġcomp": 1367, + "Ġcompact": 16830, + "Ġcompagn": 82391, + "Ġcompan": 4536, + "Ġcompanies": 5110, + "Ġcompanion": 21741, + "Ġcompanions": 40857, + "Ġcompany": 2813, + "ĠcompanyId": 67175, + "ĠcompanyName": 92915, + "Ġcompar": 7654, + "Ġcomparable": 29039, + "Ġcomparative": 54294, + "Ġcomparatively": 70461, + "Ġcomparator": 52040, + "Ġcompare": 9429, + "ĠcompareTo": 79298, + "Ġcompared": 7707, + "Ġcomparer": 92355, + "Ġcompares": 42465, + "Ġcomparing": 26297, + "Ġcomparison": 12313, + "Ġcomparisons": 35495, + "Ġcompart": 38567, + "Ġcompartir": 81746, + "Ġcompartment": 43498, + "Ġcompartments": 86252, + "Ġcompass": 23721, + "Ġcompassion": 29381, + "Ġcompassionate": 59861, + "Ġcompat": 12139, + "Ġcompatibility": 24748, + "Ġcompatible": 18146, + "Ġcompañ": 63303, + "Ġcompañero": 142548, + "ĠcompañÃŃa": 136453, + "Ġcompel": 82073, + "Ġcompelled": 47884, + "Ġcompelling": 28622, + "Ġcompens": 14239, + "Ġcompensate": 45694, + "Ġcompensated": 65882, + "Ġcompensation": 19857, + "Ġcompet": 4533, + "Ġcompete": 20259, + "Ġcompeted": 66135, + "Ġcompetence": 57166, + "Ġcompetency": 92804, + "Ġcompetent": 39783, + "Ġcompeting": 26168, + "Ġcompetit": 51204, + "Ġcompetition": 10707, + "Ġcompetitions": 42582, + "Ġcompetitive": 14680, + "Ġcompetitiveness": 76551, + "Ġcompetitor": 42840, + "Ġcompetitors": 27550, + "Ġcompil": 96931, + "Ġcompilation": 28672, + "Ġcompile": 19192, + "Ġcompiled": 19697, + "Ġcompiler": 19415, + "Ġcompilers": 87075, + "Ġcompiling": 54220, + "Ġcompl": 4581, + "Ġcomplain": 27911, + "Ġcomplained": 33970, + "Ġcomplaining": 41026, + "Ġcomplains": 88958, + "Ġcomplaint": 12181, + "Ġcomplaints": 21171, + "Ġcomple": 1907, + "Ġcomplement": 22766, + "Ġcomplementary": 57435, + "Ġcomplet": 3459, + "Ġcompleta": 70201, + "Ġcompletamente": 84624, + "Ġcomplete": 4583, + "Ġcompleted": 8145, + "Ġcompletely": 6587, + "Ġcompleteness": 79314, + "Ġcompletes": 44595, + "Ġcompleting": 26568, + "Ġcompletion": 9755, + "ĠcompletionHandler": 45780, + "Ġcompleto": 55411, + "Ġcomplex": 6351, + "Ġcomplexes": 68025, + "Ġcomplexion": 90579, + "Ġcomplexities": 83040, + "Ġcomplexity": 23094, + "Ġcompliance": 8733, + "Ġcompliant": 48698, + "Ġcomplic": 68126, + "Ġcomplicated": 16965, + "Ġcomplication": 85819, + "Ġcomplications": 35405, + "Ġcomplied": 83764, + "Ġcompliment": 31141, + "Ġcomplimentary": 53328, + "Ġcompliments": 71639, + "Ġcomply": 25017, + "Ġcomplying": 88406, + "Ġcomplète": 138180, + "Ġcomponent": 3692, + "ĠcomponentDid": 23807, + "ĠcomponentDidMount": 26740, + "ĠcomponentDidUpdate": 85220, + "ĠcomponentName": 94142, + "ĠcomponentWill": 38845, + "ĠcomponentWillMount": 78138, + "ĠcomponentWillUnmount": 71576, + "Ġcomponente": 95689, + "Ġcomponents": 6813, + "Ġcomport": 51862, + "Ġcompos": 39221, + "Ġcompose": 30335, + "Ġcomposed": 23415, + "Ġcomposer": 39450, + "Ġcomposers": 94735, + "Ġcomposing": 75878, + "Ġcomposite": 27714, + "Ġcomposition": 18037, + "Ġcompositions": 61544, + "Ġcompost": 56723, + "Ġcompound": 23628, + "Ġcompounded": 87324, + "Ġcompounds": 31146, + "Ġcompr": 13247, + "Ġcompra": 57629, + "Ġcomprar": 53409, + "Ġcompreh": 12674, + "Ġcomprehend": 57289, + "Ġcomprehension": 61094, + "Ġcomprehensive": 15817, + "Ġcomprend": 59846, + "Ġcompress": 24611, + "Ġcompressed": 30649, + "Ġcompression": 25111, + "Ġcompressor": 58877, + "Ġcomprise": 53250, + "Ġcomprised": 38956, + "Ġcomprises": 39995, + "Ġcomprising": 45238, + "Ġcomprom": 25519, + "Ġcompromise": 29385, + "Ġcompromised": 43400, + "Ġcompromises": 91446, + "Ġcompromising": 75000, + "Ġcomps": 88273, + "Ġcompt": 53956, + "Ġcompte": 41363, + "Ġcompuls": 47106, + "Ġcompulsory": 64768, + "Ġcomput": 3716, + "Ġcomputation": 34447, + "Ġcomputational": 54480, + "Ġcomputations": 82599, + "Ġcompute": 12564, + "Ġcomputed": 24182, + "Ġcomputer": 6366, + "Ġcomputers": 18495, + "Ġcomputes": 57203, + "Ġcomputing": 24231, + "Ġcompét": 74232, + "Ġcomrades": 66028, + "Ġcomun": 42935, + "Ġcomunic": 45815, + "Ġcomunicación": 138332, + "Ġcomunidad": 93890, + "ĠcomunitÃł": 142560, + "Ġcomún": 140934, + "Ġcon": 390, + "Ġconc": 3529, + "Ġconcat": 33720, + "Ġconcaten": 39972, + "Ġconcatenate": 77784, + "Ġconcatenated": 97534, + "Ġconce": 19669, + "Ġconceal": 53162, + "Ġconcealed": 43033, + "Ġconced": 37730, + "Ġconcede": 80560, + "Ġconceded": 61587, + "Ġconceivable": 94783, + "Ġconceive": 76051, + "Ġconceived": 49078, + "Ġconcent": 10014, + "Ġconcentr": 17137, + "Ġconcentrate": 36355, + "Ġconcentrated": 37526, + "Ġconcentrates": 97905, + "Ġconcentrating": 75287, + "Ġconcentration": 19950, + "Ġconcentrations": 31366, + "Ġconcept": 7286, + "Ġconception": 42556, + "Ġconcepts": 18940, + "Ġconceptual": 43801, + "Ġconcern": 4643, + "Ġconcerned": 11658, + "Ġconcerning": 18310, + "Ġconcerns": 10520, + "Ġconcert": 20830, + "Ġconcerted": 96566, + "Ġconcerts": 46579, + "Ġconcess": 56826, + "Ġconcession": 68473, + "Ġconcessions": 61508, + "Ġconcise": 63594, + "Ġconcl": 92124, + "Ġconclude": 31094, + "Ġconcluded": 19941, + "Ġconcludes": 44437, + "Ġconcluding": 71026, + "Ġconclus": 82982, + "Ġconclusion": 16688, + "Ġconclusions": 30242, + "Ġconclusive": 97486, + "Ġconco": 75539, + "Ġconcrete": 14175, + "Ġconcurrency": 78010, + "Ġconcurrent": 34035, + "Ġconcurrently": 78026, + "Ġconcussion": 67906, + "Ġcond": 9756, + "Ġcondem": 21069, + "Ġcondemn": 38974, + "Ġcondemnation": 71006, + "Ġcondemned": 37925, + "Ġcondemning": 81134, + "Ġcondensed": 74726, + "Ġcondi": 86456, + "Ġcondiciones": 77658, + "Ġcondition": 2971, + "Ġconditional": 34687, + "Ġconditioned": 65631, + "Ġconditioner": 64324, + "Ġconditioning": 34021, + "Ġconditions": 4682, + "Ġcondições": 137578, + "Ġcondo": 43100, + "Ġcondol": 69460, + "Ġcondolences": 79193, + "Ġcondom": 36716, + "Ġcondominium": 84976, + "Ġcondoms": 78875, + "Ġcondos": 87375, + "Ġcondu": 74969, + "Ġconduc": 74296, + "Ġconducive": 94461, + "Ġconduct": 6786, + "Ġconducted": 13075, + "Ġconducting": 30374, + "Ġconductivity": 97871, + "Ġconductor": 60756, + "Ġconducts": 73098, + "Ġconduit": 76535, + "Ġcone": 22161, + "Ġconect": 69344, + "Ġconectar": 94754, + "Ġcones": 86892, + "Ġconex": 54204, + "Ġconexao": 97877, + "Ġconexion": 55426, + "Ġconexión": 89858, + "Ġconf": 2335, + "Ġconfer": 48743, + "Ġconference": 9816, + "Ġconferences": 37514, + "Ġconferred": 90570, + "Ġconfess": 47366, + "Ġconfessed": 62317, + "Ġconfession": 57373, + "Ġconfidence": 12136, + "Ġconfident": 16506, + "Ġconfidential": 26193, + "Ġconfidentiality": 64195, + "Ġconfidently": 76976, + "Ġconfig": 2193, + "ĠconfigFile": 92770, + "Ġconfigparser": 97437, + "Ġconfigs": 42309, + "Ġconfigur": 46971, + "Ġconfigurable": 43394, + "Ġconfiguration": 6546, + "Ġconfigurations": 32383, + "Ġconfigure": 14411, + "Ġconfigured": 19755, + "Ġconfiguring": 71783, + "Ġconfined": 44308, + "Ġconfinement": 65578, + "Ġconfines": 90020, + "Ġconfirm": 7683, + "ĠconfirmPassword": 88118, + "Ġconfirmation": 19539, + "Ġconfirmed": 10774, + "Ġconfirming": 48996, + "Ġconfirms": 42396, + "Ġconfisc": 56356, + "Ġconfiscated": 86270, + "Ġconfl": 82877, + "Ġconflic": 9159, + "Ġconflict": 12055, + "Ġconflicting": 51033, + "Ġconflicts": 25800, + "Ġconform": 25879, + "Ġconforme": 97542, + "Ġconformity": 89201, + "Ġconforms": 95164, + "Ġconfort": 65909, + "Ġconfront": 16877, + "Ġconfrontation": 47753, + "Ġconfronted": 40682, + "Ġconfronting": 74767, + "Ġconfuse": 58117, + "Ġconfused": 21815, + "Ġconfusing": 30615, + "Ġconfusion": 21340, + "Ġcong": 98798, + "Ġcongen": 81966, + "Ġcongest": 43285, + "Ġcongestion": 54046, + "Ġconglomer": 87651, + "Ġcongr": 30169, + "Ġcongrat": 39488, + "Ġcongratulate": 88146, + "Ġcongratulations": 81544, + "Ġcongreg": 76315, + "Ġcongregation": 58535, + "Ġcongress": 21040, + "Ġcongressional": 30619, + "Ġcongressman": 73488, + "Ġconhe": 91410, + "Ġconhec": 70483, + "Ġconj": 16491, + "Ġconject": 85915, + "Ġconjug": 63280, + "Ġconjunction": 31446, + "Ġconjunto": 67493, + "Ġconn": 4534, + "Ġconna": 73182, + "Ġconnaît": 139807, + "Ġconnect": 4564, + "Ġconnected": 8433, + "Ġconnecting": 20913, + "Ġconnection": 3633, + "ĠconnectionString": 62084, + "Ġconnections": 13234, + "Ġconnectivity": 30257, + "Ġconnector": 26989, + "Ġconnectors": 57070, + "Ġconnects": 33061, + "Ġconnexion": 72776, + "Ġconoc": 32846, + "Ġconocer": 59240, + "Ġconosc": 83069, + "Ġconqu": 38779, + "Ġconquer": 50530, + "Ġconquered": 68812, + "Ġconquest": 61266, + "Ġcons": 1585, + "Ġconsc": 13588, + "Ġconscience": 41463, + "Ġconscient": 73265, + "Ġconscious": 16941, + "Ġconsciously": 69803, + "Ġconsciousness": 24875, + "Ġconse": 29925, + "Ġconsec": 19908, + "Ġconsect": 33835, + "Ġconsectetur": 35140, + "Ġconsecutive": 23921, + "Ġconseg": 98051, + "Ġconsegu": 55788, + "Ġconseguir": 76859, + "Ġconseils": 91880, + "Ġconsensus": 23869, + "Ġconsent": 14433, + "Ġconsenting": 92401, + "Ġconsequ": 12494, + "Ġconsequat": 84887, + "Ġconsequence": 28674, + "Ġconsequences": 15917, + "Ġconsequential": 90083, + "Ġconsequently": 51294, + "Ġconserv": 11110, + "Ġconservation": 28611, + "Ġconservatism": 89574, + "Ġconservative": 15332, + "Ġconservatives": 33275, + "Ġconserve": 77448, + "Ġconsid": 76454, + "Ġconsider": 2908, + "Ġconsiderable": 23839, + "Ġconsiderably": 32352, + "Ġconsideration": 17879, + "Ġconsiderations": 37764, + "Ġconsidered": 6509, + "Ġconsidering": 12831, + "Ġconsiders": 31138, + "Ġconsist": 6685, + "Ġconsisted": 43560, + "Ġconsistency": 28137, + "Ġconsistent": 12966, + "Ġconsistently": 20699, + "Ġconsisting": 30606, + "Ġconsists": 17167, + "Ġconsolation": 98858, + "Ġconsole": 2339, + "Ġconsoles": 50093, + "Ġconsolid": 27131, + "Ġconsolidate": 73321, + "Ġconsolidated": 59291, + "Ġconsolidation": 59632, + "Ġconson": 77505, + "Ġconsort": 62829, + "Ġconsortium": 73994, + "Ġconspic": 74215, + "Ġconspicuous": 96885, + "Ġconspir": 21513, + "Ġconspiracy": 25292, + "Ġconst": 733, + "Ġconstant": 6783, + "Ġconstantly": 14971, + "Ġconstants": 18021, + "Ġconstellation": 82386, + "Ġconstexpr": 19512, + "Ġconstit": 8952, + "Ġconstitu": 30505, + "Ġconstituency": 64639, + "Ġconstituent": 74064, + "Ġconstituents": 50620, + "Ġconstitute": 34156, + "Ġconstituted": 64491, + "Ġconstitutes": 41575, + "Ġconstitution": 16407, + "Ġconstitutional": 24526, + "Ġconstr": 18941, + "Ġconstrain": 79699, + "Ġconstrained": 53752, + "Ġconstraint": 21568, + "Ġconstraints": 16982, + "Ġconstrucción": 140155, + "Ġconstruct": 9245, + "Ġconstructed": 20346, + "Ġconstructing": 49353, + "Ġconstruction": 8086, + "Ġconstructions": 95839, + "Ġconstructive": 53484, + "Ġconstructor": 4692, + "Ġconstructors": 54717, + "Ġconstructs": 56227, + "Ġconstrued": 70159, + "Ġconstrução": 138978, + "Ġconsts": 95432, + "Ġconsul": 74189, + "Ġconsulate": 98417, + "Ġconsult": 8498, + "Ġconsulta": 47562, + "Ġconsultancy": 73279, + "Ġconsultant": 30660, + "Ġconsultants": 49084, + "Ġconsultar": 98337, + "Ġconsultation": 28073, + "Ġconsultations": 74437, + "Ġconsulted": 60202, + "Ġconsulting": 30731, + "Ġconsum": 4662, + "Ġconsume": 24057, + "Ġconsumed": 25983, + "Ġconsumer": 11502, + "Ġconsumers": 13417, + "Ġconsumes": 59506, + "Ġconsuming": 34108, + "Ġconsumo": 79890, + "Ġconsumption": 15293, + "Ġcont": 683, + "Ġconta": 44766, + "Ġcontact": 3645, + "Ġcontacted": 24542, + "Ġcontacting": 46108, + "Ġcontacto": 67152, + "Ġcontacts": 18508, + "Ġcontador": 58568, + "Ġcontag": 63815, + "Ġcontagious": 88434, + "Ġcontain": 6644, + "Ġcontained": 12985, + "Ġcontainer": 5476, + "ĠcontainerView": 96961, + "Ġcontainers": 23853, + "Ġcontaining": 8482, + "Ġcontainment": 45211, + "Ġcontains": 5610, + "Ġcontamin": 34422, + "Ġcontaminants": 87859, + "Ġcontaminated": 51573, + "Ġcontamination": 46710, + "Ġcontar": 88705, + "Ġcontato": 74613, + "Ġconte": 59616, + "Ġcontempl": 38371, + "Ġcontemplate": 79807, + "Ġcontemplated": 92339, + "Ġcontemplating": 78242, + "Ġcontempor": 58404, + "Ġcontemporary": 18706, + "Ġcontempt": 45354, + "Ġconten": 54011, + "Ġcontend": 55095, + "Ġcontender": 56159, + "Ġcontenders": 67639, + "Ġcontends": 80933, + "Ġcontenido": 54254, + "Ġcontent": 2213, + "ĠcontentType": 32103, + "ĠcontentValues": 81418, + "ĠcontentView": 63253, + "Ġcontention": 51140, + "Ġcontentious": 66500, + "Ġcontents": 8794, + "Ġcontenu": 75254, + "Ġcontest": 13810, + "Ġcontestant": 87222, + "Ġcontestants": 74424, + "Ġcontested": 56210, + "Ġcontests": 46503, + "Ġcontext": 2266, + "Ġcontexto": 76743, + "Ġcontexts": 37597, + "Ġcontextual": 65151, + "Ġconteúdo": 136836, + "Ġcontiene": 80433, + "Ġcontiguous": 66503, + "Ġcontin": 1976, + "Ġcontinent": 31743, + "Ġcontinental": 57536, + "Ġcontinents": 65859, + "Ġconting": 38539, + "Ġcontingency": 82449, + "Ġcontingent": 58475, + "Ġcontinu": 16122, + "Ġcontinua": 90984, + "Ġcontinual": 76875, + "Ġcontinually": 34511, + "Ġcontinuar": 71007, + "Ġcontinuation": 41171, + "Ġcontinue": 3060, + "Ġcontinued": 8570, + "Ġcontinues": 9539, + "Ġcontinuing": 14354, + "Ġcontinuity": 49006, + "Ġcontinuous": 19259, + "Ġcontinuously": 30878, + "Ġcontinuum": 85801, + "Ġcontour": 41433, + "Ġcontours": 49031, + "Ġcontr": 6027, + "Ġcontra": 31050, + "Ġcontrace": 44836, + "Ġcontraception": 71254, + "Ġcontraceptive": 98290, + "Ġcontract": 5116, + "Ġcontracted": 49968, + "Ġcontracting": 59406, + "Ġcontraction": 70795, + "Ġcontractor": 29972, + "Ġcontractors": 32740, + "Ġcontracts": 17080, + "Ġcontractual": 75443, + "Ġcontrad": 22296, + "Ġcontradict": 42461, + "Ġcontradiction": 49759, + "Ġcontradictions": 80423, + "Ġcontradictory": 70140, + "Ġcontrario": 82813, + "Ġcontrary": 25047, + "Ġcontraseña": 65437, + "Ġcontrast": 12872, + "Ġcontrasting": 73955, + "Ġcontrasts": 82279, + "Ġcontrat": 87003, + "Ġcontrato": 81651, + "Ġcontre": 44873, + "Ġcontrib": 5919, + "Ġcontribute": 16792, + "Ġcontributed": 19590, + "Ġcontributes": 42972, + "Ġcontributing": 28720, + "Ġcontribution": 18527, + "Ġcontributions": 19026, + "Ġcontributor": 25305, + "Ġcontributors": 20343, + "Ġcontro": 82164, + "Ġcontrol": 2524, + "ĠcontrolId": 88532, + "Ġcontrole": 80113, + "Ġcontrolled": 14071, + "Ġcontroller": 6461, + "Ġcontrollers": 26225, + "Ġcontrolling": 25902, + "Ġcontrols": 11574, + "Ġcontrovers": 13365, + "Ġcontroversial": 20129, + "Ġcontroversies": 85793, + "Ġcontroversy": 25573, + "Ġcontrôle": 136577, + "Ġconv": 5686, + "Ġconven": 10138, + "Ġconvened": 94798, + "Ġconvenience": 19132, + "Ġconvenient": 16710, + "Ġconveniently": 48070, + "Ġconvent": 16705, + "Ġconvention": 21277, + "Ġconventional": 20692, + "Ġconventions": 44493, + "Ġconver": 19313, + "Ġconverge": 79767, + "Ġconverged": 96775, + "Ġconvergence": 55441, + "Ġconvers": 7517, + "Ġconversation": 10435, + "Ġconversations": 20959, + "Ġconverse": 94240, + "Ġconversion": 14409, + "Ġconversions": 48722, + "Ġconvert": 5508, + "ĠconvertView": 34544, + "Ġconverted": 16099, + "Ġconverter": 27058, + "Ġconverters": 88888, + "Ġconvertible": 67613, + "Ġconverting": 33437, + "Ġconverts": 32722, + "Ġconvex": 65930, + "Ġconvey": 20001, + "Ġconveyed": 72797, + "Ġconveying": 93065, + "Ġconveyor": 49141, + "Ġconvict": 90034, + "Ġconvicted": 23088, + "Ġconviction": 28091, + "Ġconvictions": 46915, + "Ġconvin": 21090, + "Ġconvinc": 35861, + "Ġconvince": 26910, + "Ġconvinced": 22166, + "Ġconvincing": 39561, + "Ġconvo": 97088, + "Ġconvolution": 55712, + "Ġconvoy": 79461, + "Ġcook": 4296, + "Ġcookbook": 76104, + "Ġcooked": 29105, + "Ġcooker": 62083, + "Ġcookie": 12544, + "Ġcookies": 8279, + "Ġcooking": 17233, + "Ġcooks": 72427, + "Ġcool": 7010, + "Ġcoolant": 96770, + "Ġcooldown": 43250, + "Ġcooled": 64310, + "Ġcooler": 35821, + "Ġcoolest": 73434, + "Ġcooling": 26917, + "Ġcoop": 82070, + "Ġcooper": 21676, + "Ġcooperate": 46803, + "Ġcooperating": 81489, + "Ġcooperation": 23046, + "Ġcooperative": 47466, + "Ġcoord": 16489, + "Ġcoorden": 96384, + "Ġcoordin": 29378, + "Ġcoordinate": 16184, + "Ġcoordinated": 46572, + "Ġcoordinates": 13934, + "Ġcoordinating": 65415, + "Ġcoordination": 37693, + "Ġcoordinator": 30284, + "Ġcoords": 23897, + "Ġcop": 6162, + "Ġcope": 36486, + "Ġcopied": 21774, + "Ġcopies": 10993, + "Ġcoping": 61982, + "Ġcopp": 56000, + "Ġcopper": 23285, + "Ġcoppia": 82246, + "Ġcops": 34217, + "Ġcopy": 2975, + "Ġcopying": 31039, + "Ġcopyright": 6920, + "Ġcopyrighted": 58613, + "Ġcopyrights": 85818, + "Ġcoquine": 64158, + "Ġcor": 1829, + "Ġcoral": 52003, + "Ġcoraz": 82002, + "Ġcorazón": 140766, + "Ġcoração": 140158, + "Ġcord": 22324, + "Ġcords": 79617, + "Ġcore": 6200, + "Ġcores": 35704, + "Ġcork": 89324, + "Ġcorn": 13776, + "Ġcorner": 9131, + "Ġcornerback": 67000, + "Ġcorners": 23462, + "Ġcornerstone": 81475, + "Ġcoron": 21989, + "Ġcoronary": 65198, + "Ġcoronavirus": 32233, + "Ġcoroutine": 77799, + "Ġcorp": 36466, + "Ġcorpo": 65634, + "Ġcorpor": 8368, + "Ġcorporate": 12870, + "Ġcorporation": 26669, + "Ġcorporations": 24346, + "Ġcorps": 33822, + "Ġcorpse": 51544, + "Ġcorpses": 72117, + "Ġcorpus": 42094, + "Ġcorr": 44353, + "Ġcorre": 12323, + "Ġcorrect": 4396, + "Ġcorrectamente": 58507, + "Ġcorrected": 35965, + "Ġcorrecting": 81375, + "Ġcorrection": 26262, + "Ġcorrections": 50379, + "Ġcorrective": 90752, + "Ġcorrectly": 12440, + "Ġcorrectness": 57323, + "Ġcorrel": 34883, + "Ġcorrelate": 80484, + "Ġcorrelated": 48293, + "Ġcorrelates": 96203, + "Ġcorrelation": 25588, + "Ġcorrelations": 68816, + "Ġcorreo": 45921, + "Ġcorres": 84064, + "Ġcorrespond": 7866, + "Ġcorrespondence": 43718, + "Ġcorrespondent": 48564, + "Ġcorresponding": 12159, + "Ġcorresponds": 33210, + "Ġcorrid": 33158, + "Ġcorridor": 45879, + "Ġcorridors": 74652, + "Ġcorro": 63356, + "Ġcorrobor": 78719, + "Ġcorros": 45557, + "Ġcorrosion": 55388, + "Ġcorrupt": 21917, + "Ġcorrupted": 49400, + "Ġcorruption": 21252, + "Ġcors": 43911, + "Ġcort": 22303, + "Ġcortex": 48270, + "Ġcortical": 82519, + "Ġcortisol": 88842, + "Ġcos": 7960, + "Ġcosa": 47513, + "Ġcosas": 55523, + "Ġcose": 85013, + "Ġcosine": 75259, + "Ġcosm": 55654, + "Ġcosmetic": 45452, + "Ġcosmetics": 64965, + "Ġcosmic": 54701, + "Ġcosmos": 82545, + "Ġcosplay": 88684, + "Ġcost": 2783, + "Ġcosta": 98702, + "Ġcosting": 53724, + "Ġcostly": 33248, + "Ġcosto": 77849, + "Ġcosts": 7049, + "Ġcostume": 31419, + "Ġcostumes": 44088, + "Ġcosy": 84879, + "Ġcosì": 68662, + "Ġcot": 47581, + "Ġcott": 96529, + "Ġcottage": 45622, + "Ġcotton": 23522, + "Ġcou": 3933, + "Ġcouch": 26148, + "Ġcougar": 47091, + "Ġcough": 39600, + "Ġcould": 1410, + "Ġcouldn": 7691, + "Ġcouleur": 75551, + "Ġcoun": 2294, + "Ġcouncil": 14833, + "Ġcouncill": 60045, + "Ġcouncillor": 90361, + "Ġcouncillors": 97405, + "Ġcouncils": 60683, + "Ġcounsel": 16080, + "Ġcounseling": 41216, + "Ġcounselling": 82681, + "Ġcounselor": 61375, + "Ġcounselors": 85247, + "Ġcount": 1760, + "Ġcountdown": 48028, + "Ġcounted": 29994, + "Ġcounter": 5546, + "Ġcountered": 85067, + "Ġcounterfeit": 82634, + "Ġcounterpart": 44016, + "Ġcounterparts": 37395, + "Ġcounters": 31532, + "Ġcountert": 60110, + "Ġcountertops": 84213, + "Ġcounties": 30176, + "Ġcounting": 25009, + "Ġcountless": 27601, + "Ġcountries": 5837, + "Ġcountry": 3146, + "ĠcountryCode": 90202, + "Ġcountryside": 46867, + "Ġcounts": 14579, + "Ġcounty": 13867, + "Ġcoup": 15706, + "Ġcoupe": 87509, + "Ġcouple": 5625, + "Ġcoupled": 33256, + "Ġcouples": 21264, + "Ġcoupling": 57986, + "Ġcoupon": 23821, + "Ġcoupons": 35298, + "Ġcour": 2081, + "Ġcourage": 24744, + "Ġcourageous": 74648, + "Ġcourier": 74517, + "Ġcours": 44269, + "Ġcourse": 3308, + "ĠcourseId": 87224, + "Ġcourses": 13980, + "Ġcoursework": 74433, + "Ġcourt": 5473, + "Ġcourte": 73899, + "Ġcourteous": 88188, + "Ġcourtesy": 26013, + "Ġcourthouse": 81032, + "Ġcourtroom": 58120, + "Ġcourts": 18828, + "Ġcourtyard": 65285, + "Ġcous": 26714, + "Ġcousin": 35171, + "Ġcousins": 59907, + "Ġcout": 5034, + "Ġcov": 21836, + "Ġcovariance": 59308, + "Ġcovenant": 63904, + "Ġcover": 3421, + "Ġcoverage": 10191, + "Ġcovered": 9761, + "Ġcovering": 18202, + "Ġcovers": 14521, + "Ġcovert": 56493, + "Ġcoveted": 69565, + "Ġcovid": 83017, + "Ġcow": 19362, + "Ġcoward": 65175, + "Ġcowboy": 80172, + "Ġcowork": 59968, + "Ġcoworkers": 82955, + "Ġcows": 45166, + "Ġcoy": 74609, + "Ġcoz": 84820, + "Ġcozy": 42435, + "Ġcoût": 136491, + "ĠcoÅĽ": 130743, + "Ġcp": 12490, + "Ġcpf": 63976, + "Ġcpp": 46824, + "Ġcps": 33265, + "Ġcpt": 94467, + "Ġcpu": 17319, + "Ġcq": 85822, + "Ġcr": 1560, + "Ġcra": 45041, + "Ġcrab": 59412, + "Ġcrack": 17486, + "Ġcrackdown": 58133, + "Ġcracked": 42839, + "Ġcrackers": 93811, + "Ġcracking": 51729, + "Ġcracks": 44863, + "Ġcraft": 10770, + "Ġcrafted": 33458, + "Ġcrafting": 44067, + "Ġcrafts": 43848, + "Ġcraftsm": 64781, + "Ġcraftsmanship": 83960, + "Ġcram": 99081, + "Ġcramped": 98732, + "Ġcran": 69537, + "Ġcrane": 72947, + "Ġcrank": 50258, + "Ġcrap": 35550, + "Ġcrappy": 82897, + "Ġcrash": 9920, + "Ġcrashed": 33753, + "Ġcrashes": 36137, + "Ġcrashing": 45474, + "Ġcrate": 17717, + "Ġcrater": 74707, + "Ġcrates": 81720, + "Ġcrave": 80313, + "Ġcraving": 63704, + "Ġcravings": 88828, + "Ġcraw": 23926, + "Ġcrawl": 45664, + "Ġcrawled": 92170, + "Ġcrawler": 73094, + "Ġcrawling": 71079, + "Ġcray": 88916, + "Ġcraz": 68591, + "Ġcrazy": 14264, + "Ġcrc": 37180, + "Ġcre": 1884, + "Ġcrea": 80770, + "Ġcread": 61486, + "Ġcreado": 93401, + "Ġcream": 12644, + "Ġcreampie": 61077, + "Ġcreams": 80216, + "Ġcreamy": 46885, + "Ġcrear": 47737, + "Ġcreat": 6056, + "Ġcreate": 1855, + "ĠcreateAction": 76636, + "ĠcreateContext": 83674, + "ĠcreateDate": 86485, + "ĠcreateElement": 86105, + "ĠcreateSelector": 82372, + "ĠcreateStackNavigator": 98820, + "ĠcreateState": 36717, + "ĠcreateStore": 53439, + "ĠcreateTime": 53149, + "ĠcreateUser": 60754, + "Ġcreated": 3465, + "ĠcreatedAt": 33051, + "ĠcreatedBy": 83773, + "Ġcreates": 11450, + "Ġcreating": 6825, + "Ġcreation": 9688, + "Ġcreations": 52762, + "Ġcreative": 11521, + "Ġcreatively": 92296, + "Ġcreativity": 27597, + "Ġcreator": 19919, + "Ġcreators": 33799, + "Ġcreature": 17218, + "Ġcreatures": 19970, + "Ġcred": 4187, + "Ġcredential": 40207, + "Ġcredentials": 16387, + "Ġcredibility": 37669, + "Ġcredible": 41910, + "Ġcredit": 6668, + "Ġcredited": 40757, + "Ġcreditor": 97389, + "Ġcreditors": 65499, + "Ġcredits": 20141, + "Ġcreds": 73177, + "Ġcree": 86439, + "Ġcreed": 91007, + "Ġcreek": 64405, + "Ġcreep": 27083, + "Ġcreeping": 87592, + "Ġcreepy": 50840, + "Ġcref": 11145, + "Ġcrem": 70196, + "Ġcreo": 88224, + "Ġcres": 45758, + "Ġcresc": 77988, + "Ġcrest": 71924, + "Ġcrew": 13627, + "Ġcrews": 42060, + "Ġcri": 27558, + "Ġcria": 83908, + "Ġcrian": 87138, + "Ġcriança": 141500, + "Ġcrianças": 134171, + "Ġcriar": 76410, + "Ġcrib": 79011, + "Ġcricket": 35999, + "Ġcried": 38069, + "Ġcries": 57745, + "Ġcrim": 84932, + "Ġcrime": 9778, + "Ġcrimes": 16660, + "Ġcriminal": 9158, + "Ġcriminals": 31538, + "Ġcrimson": 96019, + "Ġcripp": 70008, + "Ġcrippled": 97952, + "Ġcrippling": 96006, + "Ġcris": 58953, + "Ġcrises": 57087, + "Ġcrisis": 11251, + "Ġcrisp": 41854, + "Ġcrispy": 72524, + "Ġcrist": 67984, + "Ġcrit": 4638, + "Ġcriter": 80339, + "Ġcriteria": 12890, + "Ġcriterion": 35957, + "Ġcritic": 9741, + "Ġcritical": 9023, + "Ġcritically": 40340, + "Ġcriticised": 59379, + "Ġcriticism": 18817, + "Ġcriticisms": 62736, + "Ġcriticize": 60630, + "Ġcriticized": 31514, + "Ġcriticizing": 67988, + "Ġcritics": 22698, + "Ġcritique": 42565, + "Ġcritiques": 86213, + "Ġcro": 14093, + "Ġcrochet": 74019, + "Ġcron": 46582, + "Ġcrop": 19097, + "Ġcropped": 60143, + "Ġcropping": 98937, + "Ġcrops": 30565, + "Ġcrore": 44743, + "Ġcross": 5312, + "ĠcrossAxisAlignment": 43695, + "Ġcrossed": 27031, + "Ġcrosses": 49435, + "Ġcrossing": 26638, + "Ġcrossings": 86099, + "Ġcrossorigin": 30901, + "Ġcrossover": 48380, + "Ġcrossword": 91396, + "Ġcrow": 9276, + "Ġcrowd": 13428, + "Ġcrowded": 38213, + "Ġcrowdfunding": 76644, + "Ġcrowds": 34751, + "Ġcrown": 26533, + "Ġcrowned": 78643, + "Ġcrt": 97967, + "Ġcru": 22218, + "Ġcruc": 14788, + "Ġcrucial": 16587, + "Ġcrud": 75404, + "Ġcrude": 34650, + "Ġcruel": 27030, + "Ġcruelty": 53580, + "Ġcruis": 59097, + "Ġcruise": 30451, + "Ġcruiser": 75524, + "Ġcruising": 72227, + "Ġcrumbling": 88317, + "Ġcrumbs": 96577, + "Ġcrunch": 41734, + "Ġcrunchy": 94233, + "Ġcrus": 61718, + "Ġcrush": 5030, + "Ġcrushed": 32645, + "Ġcrusher": 7809, + "Ġcrushers": 25231, + "Ġcrushing": 14432, + "Ġcrust": 39612, + "Ġcruz": 86151, + "Ġcry": 15729, + "Ġcrying": 30199, + "Ġcrypt": 14436, + "Ġcrypto": 19028, + "Ġcryptoc": 20220, + "Ġcryptocurrencies": 43974, + "Ġcryptocurrency": 26037, + "Ġcryptographic": 89129, + "Ġcryptography": 86837, + "Ġcryst": 34342, + "Ġcrystal": 25055, + "Ġcrystall": 63468, + "Ġcrystals": 47373, + "Ġcré": 33180, + "Ġcréation": 80839, + "Ġcréd": 67192, + "Ġcrédit": 142263, + "Ġcrédito": 134517, + "Ġcréer": 74731, + "Ġcréé": 140613, + "ĠcrÃŃtica": 138841, + "Ġcs": 10532, + "Ġcsak": 80277, + "Ġcsr": 68932, + "Ġcsrf": 35522, + "Ġcss": 15877, + "Ġcst": 93804, + "Ġcsv": 13147, + "Ġcsvfile": 97895, + "Ġct": 20251, + "Ġctl": 64935, + "Ġctor": 56453, + "Ġctr": 50638, + "Ġctrl": 23743, + "Ġctx": 5635, + "Ġctxt": 59162, + "Ġctype": 85507, + "Ġctypes": 48219, + "Ġcu": 6871, + "Ġcual": 27090, + "Ġcuales": 94749, + "Ġcualquier": 45909, + "Ġcuando": 26346, + "Ġcuanto": 86487, + "Ġcuatro": 76008, + "Ġcub": 18728, + "Ġcube": 23739, + "Ġcubes": 54104, + "Ġcubic": 40899, + "Ġcuc": 74291, + "Ġcuck": 62179, + "Ġcuckold": 81252, + "Ġcucumber": 64932, + "Ġcud": 78587, + "Ġcuda": 24408, + "ĠcudaMemcpy": 61284, + "Ġcudd": 64145, + "Ġcue": 49398, + "Ġcuent": 87238, + "Ġcuenta": 31095, + "Ġcuer": 56295, + "Ġcuerpo": 62513, + "Ġcues": 55916, + "Ġcuff": 74423, + "Ġcuffs": 89469, + "Ġcui": 37944, + "Ġcuid": 56676, + "Ġcuis": 30447, + "Ġcuisine": 35005, + "Ġcuk": 92743, + "Ġcul": 11695, + "Ġculinary": 57341, + "Ġculmination": 92201, + "Ġculo": 64398, + "Ġculp": 96925, + "Ġculpa": 83285, + "Ġculprit": 74183, + "Ġcult": 4511, + "Ġcultiv": 26883, + "Ġcultivate": 66532, + "Ġcultivated": 66066, + "Ġcultivating": 98039, + "Ġcultivation": 51029, + "Ġcultura": 73635, + "Ġcultural": 12752, + "Ġculturally": 65863, + "Ġculture": 7674, + "Ġcultured": 88848, + "Ġcultures": 26735, + "Ġcum": 12177, + "Ġcumbersome": 95090, + "Ġcumpl": 73355, + "Ġcumshot": 94627, + "Ġcumulative": 39844, + "Ġcunning": 80679, + "Ġcunt": 68701, + "Ġcup": 10525, + "Ġcupboard": 85941, + "Ġcupcakes": 87893, + "Ġcupid": 92443, + "Ġcups": 25374, + "Ġcur": 2847, + "Ġcurated": 57632, + "Ġcurator": 86705, + "Ġcurb": 43862, + "Ġcure": 26116, + "Ġcured": 63588, + "Ġcuring": 96230, + "Ġcuriosity": 40228, + "Ġcurious": 22208, + "Ġcurl": 13957, + "Ġcurled": 91526, + "Ġcurls": 88949, + "Ġcurly": 68103, + "Ġcurr": 9804, + "Ġcurrencies": 35602, + "Ġcurrency": 11413, + "Ġcurrent": 1482, + "ĠcurrentDate": 59813, + "ĠcurrentIndex": 50040, + "ĠcurrentItem": 91627, + "ĠcurrentNode": 46299, + "ĠcurrentPage": 39314, + "ĠcurrentPlayer": 90068, + "ĠcurrentPosition": 84418, + "ĠcurrentState": 57878, + "ĠcurrentTime": 39199, + "ĠcurrentUser": 31443, + "ĠcurrentValue": 66248, + "Ġcurrently": 5023, + "Ġcurrents": 59601, + "Ġcurriculum": 29576, + "Ġcurry": 54078, + "Ġcurs": 23143, + "Ġcurse": 40000, + "Ġcursed": 62871, + "Ġcurses": 67147, + "Ġcurso": 47092, + "Ġcursor": 8128, + "Ġcursos": 86491, + "Ġcurt": 43178, + "Ġcurtain": 45766, + "Ġcurtains": 53399, + "Ġcurvature": 82133, + "Ġcurve": 15655, + "Ġcurved": 49164, + "Ġcurves": 35933, + "Ġcus": 79970, + "Ġcush": 32768, + "Ġcushion": 44638, + "Ġcushions": 67141, + "Ġcust": 16564, + "Ġcustody": 24492, + "Ġcustom": 2526, + "Ġcustomary": 72248, + "Ġcustomer": 6002, + "ĠcustomerId": 67505, + "Ġcustomers": 6310, + "Ġcustomizable": 62074, + "Ġcustomization": 48041, + "Ġcustomize": 31087, + "Ġcustomized": 31689, + "Ġcustoms": 34769, + "Ġcut": 3931, + "Ġcute": 18838, + "Ġcutoff": 44279, + "Ġcuts": 15104, + "Ġcutter": 61605, + "Ġcutting": 14376, + "Ġcuz": 90815, + "ĠcuÌ": 143455, + "ĠcuÌī": 143456, + "ĠcuÌīa": 143457, + "Ġcuá»iji": 129742, + "Ġcuá»ijn": 135570, + "Ġcuá»Ļc": 128564, + "Ġcv": 5544, + "Ġcvs": 93142, + "Ġcw": 53091, + "Ġcwd": 46938, + "Ġcx": 20716, + "Ġcy": 9503, + "Ġcyan": 57888, + "Ġcyber": 20847, + "Ġcybersecurity": 61442, + "Ġcyc": 43348, + "Ġcycl": 31243, + "Ġcycle": 10775, + "Ġcycles": 24484, + "Ġcyclic": 76002, + "Ġcycling": 32062, + "Ġcyclist": 80031, + "Ġcyclists": 55083, + "Ġcyl": 41921, + "Ġcylinder": 34564, + "Ġcylinders": 74321, + "Ġcylindrical": 78510, + "Ġcyn": 47264, + "Ġcynical": 69679, + "Ġcyst": 62481, + "Ġcyt": 78809, + "Ġcytok": 82085, + "Ġcz": 17988, + "Ġczas": 47419, + "Ġczy": 35950, + "ĠczÄĻ": 69294, + "ĠczÄĻsto": 131830, + "ĠczÄĻÅĽci": 130399, + "ĠczÄĻÅĽÄĩ": 135994, + "ĠczÅĤ": 133089, + "ĠczÅĤowie": 133090, + "ĠczÅĤowiek": 133091, + "Ġcá": 58701, + "Ġcác": 78228, + "Ġcách": 128328, + "Ġcái": 128530, + "Ġcámara": 142552, + "Ġcán": 132356, + "Ġcánh": 134374, + "Ġcáo": 129410, + "Ġcâ": 47824, + "Ġcân": 130660, + "Ġcâu": 129260, + "Ġcây": 129368, + "Ġcé": 63834, + "Ġcél": 83750, + "Ġcélib": 94336, + "Ġcòn": 128320, + "Ġcó": 28776, + "Ġcódigo": 44059, + "Ġcómo": 54896, + "Ġcô": 128434, + "Ġcông": 62029, + "Ġcôt": 70970, + "Ġcôté": 80237, + "Ġcùng": 128340, + "Ġcú": 132880, + "Ġcüm": 143989, + "Ġcümle": 143990, + "ĠcÃł": 130849, + "ĠcÃłng": 129357, + "ĠcÄĥ": 39257, + "ĠcÄĥn": 129209, + "ĠcÄĥng": 139529, + "ĠcÅ©": 133138, + "ĠcÅ©ng": 128268, + "ĠcÅĵur": 138295, + "ĠcÆ¡": 128322, + "Ġcư": 129447, + "ĠcÆ°á»Ľi": 136273, + "Ġcưá»Ŀi": 130692, + "Ġcưá»Ŀng": 132300, + "Ġcạnh": 129623, + "Ġcả": 125440, + "Ġcải": 131867, + "Ġcảm": 128467, + "Ġcảnh": 129111, + "Ġcấp": 128376, + "Ġcấu": 133795, + "Ġcầm": 134650, + "Ġcần": 128360, + "Ġcầu": 128451, + "Ġcắt": 134362, + "Ġcặp": 139519, + "ĠcáºŃn": 136855, + "ĠcáºŃp": 130569, + "ĠcáºŃu": 131873, + "Ġcụ": 129705, + "Ġcục": 138831, + "Ġcá»§": 140845, + "Ġcá»§a": 59735, + "Ġcứ": 126500, + "Ġcứng": 137142, + "Ġcứu": 129156, + "Ġcá»±c": 129541, + "Ġcá»ij": 129234, + "Ġcá»ķ": 129057, + "Ġcá»Ļng": 130700, + "Ġcá»Ń": 134603, + "Ġcá»Ńa": 129137, + "Ġd": 294, + "ĠdB": 43768, + "Ġda": 2994, + "Ġdaar": 59347, + "Ġdab": 82768, + "Ġdabei": 49536, + "Ġdac": 76892, + "Ġdad": 17760, + "Ġdaddy": 56542, + "Ġdado": 57792, + "Ġdados": 27945, + "Ġdads": 88761, + "Ġdaemon": 39293, + "Ġdafür": 67952, + "Ġdag": 28069, + "Ġdagen": 89951, + "Ġdagger": 39231, + "Ġdah": 52488, + "Ġdaha": 61494, + "Ġdaher": 93328, + "Ġdai": 62392, + "Ġdaily": 7298, + "Ġdairy": 33379, + "Ġdak": 97053, + "Ġdal": 24738, + "Ġdalam": 29792, + "Ġdall": 62993, + "Ġdalla": 48208, + "Ġdalle": 85244, + "Ġdam": 3739, + "Ġdamage": 5557, + "Ġdamaged": 20123, + "Ġdamages": 25129, + "Ġdamaging": 33346, + "Ġdame": 64767, + "Ġdamer": 41951, + "Ġdames": 84221, + "Ġdamit": 40870, + "Ġdamn": 26762, + "Ġdamned": 67365, + "Ġdamning": 95303, + "Ġdamp": 40269, + "Ġdamping": 83514, + "Ġdams": 82525, + "Ġdan": 9101, + "Ġdance": 15254, + "Ġdanced": 83752, + "Ġdancer": 63582, + "Ġdancers": 60483, + "Ġdances": 72706, + "Ġdancing": 27966, + "Ġdando": 92645, + "Ġdane": 88595, + "Ġdang": 57512, + "Ġdanger": 7978, + "Ġdangerous": 11406, + "Ġdangerously": 58494, + "Ġdangers": 35964, + "Ġdangling": 91623, + "Ġdanh": 98896, + "Ġdank": 80122, + "Ġdanmark": 89192, + "Ġdann": 24929, + "Ġdans": 6866, + "Ġdansk": 49320, + "Ġdanske": 55753, + "Ġdanych": 82217, + "Ġdao": 24775, + "Ġdap": 81521, + "Ġdapat": 44117, + "Ġdar": 15269, + "Ġdara": 64069, + "Ġdarauf": 74054, + "Ġdare": 32204, + "Ġdared": 74935, + "Ġdarf": 87231, + "Ġdari": 24200, + "Ġdaring": 58672, + "Ġdark": 6319, + "Ġdarken": 94688, + "Ġdarker": 39030, + "Ġdarkest": 75969, + "Ġdarkness": 26298, + "Ġdarling": 75645, + "Ġdarm": 83302, + "Ġdarn": 73677, + "Ġdart": 55967, + "Ġdarüber": 95038, + "Ġdas": 6616, + "Ġdash": 23911, + "Ġdashboard": 26967, + "Ġdashed": 66722, + "Ġdashes": 87546, + "Ġdass": 14948, + "Ġdat": 3258, + "Ġdata": 821, + "ĠdataArray": 83712, + "ĠdataGridView": 18038, + "ĠdataGridViewCellStyle": 26441, + "ĠdataGridViewTextBoxColumn": 94425, + "ĠdataIndex": 41994, + "ĠdataList": 58595, + "ĠdataSet": 56053, + "ĠdataSize": 89961, + "ĠdataSnapshot": 64193, + "ĠdataSource": 29554, + "ĠdataTable": 62766, + "ĠdataType": 22285, + "Ġdatab": 71240, + "Ġdatabase": 4625, + "Ġdatabases": 31806, + "Ġdataframe": 38228, + "Ġdatap": 61723, + "Ġdatas": 16879, + "Ġdataset": 10337, + "Ġdatasets": 29425, + "Ġdatasource": 69156, + "Ġdatastore": 64986, + "Ġdatatable": 93006, + "Ġdatatype": 46634, + "Ġdate": 2400, + "ĠdateFormat": 49267, + "ĠdateFormatter": 69776, + "ĠdatePicker": 92151, + "ĠdateString": 85498, + "ĠdateTime": 34608, + "Ġdated": 29005, + "Ġdates": 12713, + "Ġdatetime": 8874, + "Ġdati": 71427, + "Ġdating": 4924, + "Ġdatings": 42270, + "Ġdatingside": 76035, + "Ġdatingsider": 84865, + "Ġdato": 46292, + "Ġdatos": 17751, + "Ġdatum": 41556, + "Ġdau": 88280, + "Ġdaughter": 9803, + "Ġdaughters": 29868, + "Ġdaunting": 56597, + "Ġdav": 40769, + "Ġdavid": 53946, + "Ġdavidjl": 98844, + "Ġdavon": 74565, + "Ġdavran": 142801, + "ĠdavranÄ±ÅŁ": 142802, + "Ġdawn": 38393, + "Ġday": 1899, + "Ġdaycare": 99154, + "Ġdaylight": 52021, + "Ġdays": 2849, + "Ġdaytime": 61082, + "Ġdazu": 53797, + "Ġdazz": 59918, + "Ġdazzling": 76988, + "ĠdaÃŁ": 94567, + "Ġdaģı": 141670, + "Ġdaģıt": 141671, + "Ġdb": 2927, + "ĠdbContext": 55495, + "ĠdbHelper": 95980, + "ĠdbName": 75564, + "Ġdbc": 79645, + "Ġdbg": 47904, + "Ġdbl": 88995, + "Ġdbname": 83492, + "Ġdbo": 55859, + "Ġdbs": 84581, + "Ġdbus": 73696, + "Ġdc": 19402, + "Ġdcc": 91279, + "Ġdct": 90164, + "Ġdd": 19647, + "Ġddl": 55780, + "Ġde": 409, + "Ġdeactivate": 66245, + "Ġdeactivated": 83616, + "Ġdead": 5593, + "Ġdeadliest": 98369, + "Ġdeadline": 21428, + "Ġdeadlines": 57882, + "Ġdeadlock": 93345, + "Ġdeadly": 24142, + "Ġdeaf": 46742, + "Ġdeal": 3484, + "Ġdealer": 23321, + "Ġdealers": 26199, + "Ġdealership": 61196, + "Ġdealing": 14550, + "Ġdealings": 65929, + "Ġdealloc": 77670, + "Ġdeals": 12506, + "Ġdealt": 25934, + "Ġdean": 72862, + "Ġdear": 24253, + "Ġdearly": 97563, + "Ġdeath": 4545, + "Ġdeaths": 16375, + "Ġdeb": 4222, + "Ġdebacle": 98912, + "Ġdebate": 11004, + "Ġdebated": 58574, + "Ġdebates": 36535, + "Ġdebating": 70413, + "Ġdebe": 34374, + "Ġdeben": 73613, + "Ġdeber": 58205, + "Ġdeberá": 140551, + "ĠdeberÃŃa": 139605, + "Ġdebian": 96078, + "Ġdebido": 90892, + "Ġdebilitating": 91790, + "Ġdebit": 45353, + "Ġdebounce": 76839, + "Ġdebris": 33998, + "Ġdebt": 11636, + "Ġdebtor": 93762, + "Ġdebts": 44205, + "Ġdebug": 7390, + "Ġdebugger": 44473, + "Ġdebugging": 27703, + "Ġdebunk": 77193, + "Ġdebut": 17306, + "Ġdebuted": 57085, + "Ġdec": 1622, + "Ġdecad": 87033, + "Ġdecade": 13212, + "Ġdecades": 10793, + "Ġdecals": 97685, + "Ġdecay": 30715, + "Ġdece": 30852, + "Ġdeceased": 42620, + "Ġdeceit": 85485, + "Ġdeceive": 92828, + "Ġdeceived": 97784, + "Ġdecember": 89011, + "Ġdecency": 98270, + "Ġdecent": 14977, + "Ġdecentral": 38105, + "Ġdecentralized": 47963, + "Ġdeception": 63475, + "Ġdeceptive": 80274, + "Ġdecid": 69393, + "Ġdecide": 10279, + "Ġdecided": 6635, + "Ġdecidedly": 76020, + "Ġdecides": 27627, + "Ġdeciding": 29130, + "Ġdecimal": 12122, + "Ġdecimals": 58328, + "Ġdecipher": 74177, + "Ġdecir": 48918, + "Ġdecis": 33022, + "Ġdecision": 5480, + "Ġdecisions": 11181, + "Ġdecisive": 50291, + "Ġdecisión": 141760, + "Ġdecisão": 139374, + "Ġdeck": 9530, + "Ġdecking": 89994, + "Ġdecks": 29781, + "Ġdecl": 3963, + "Ġdeclar": 9445, + "Ġdeclaración": 90654, + "Ġdeclaration": 18004, + "Ġdeclarations": 15693, + "Ġdeclare": 15913, + "Ġdeclared": 14275, + "Ġdeclares": 49788, + "Ġdeclaring": 36531, + "Ġdecline": 17704, + "Ġdeclined": 18758, + "Ġdeclines": 56954, + "Ġdeclining": 42748, + "Ġdecltype": 75853, + "Ġdeco": 67552, + "Ġdecode": 16895, + "Ġdecoded": 29213, + "Ġdecoder": 24551, + "Ġdecoding": 47116, + "Ġdecom": 28502, + "Ġdecomposition": 65166, + "Ġdecor": 10576, + "Ġdecorate": 56581, + "Ġdecorated": 36009, + "Ġdecorating": 48582, + "Ġdecoration": 20228, + "Ġdecorations": 47579, + "Ġdecorative": 45436, + "Ġdecorator": 50678, + "Ġdecorators": 62678, + "Ġdecre": 9728, + "Ġdecrease": 18472, + "Ġdecreased": 24938, + "Ġdecreases": 42054, + "Ġdecreasing": 43549, + "Ġdecree": 60534, + "Ġdecrement": 60832, + "Ġdecrypt": 38126, + "Ġdecrypted": 62961, + "Ġdecryption": 89374, + "Ġded": 7681, + "Ġdedic": 56666, + "Ġdedicate": 73604, + "Ġdedicated": 12235, + "Ġdedication": 38855, + "Ġdeduct": 36810, + "Ġdeducted": 95504, + "Ġdeductible": 77812, + "Ġdeduction": 38843, + "Ġdeductions": 70637, + "Ġdee": 44733, + "Ġdeed": 55308, + "Ġdeeds": 53711, + "Ġdeem": 81477, + "Ġdeemed": 24636, + "Ġdeep": 5538, + "Ġdeepcopy": 55558, + "Ġdeepen": 91692, + "Ġdeeper": 19117, + "Ġdeepest": 50521, + "Ġdeeply": 17247, + "Ġdeer": 38049, + "Ġdef": 707, + "ĠdefStyle": 64875, + "ĠdefStyleAttr": 86090, + "Ġdefamation": 88101, + "Ġdefault": 1638, + "ĠdefaultCenter": 55646, + "ĠdefaultManager": 98920, + "ĠdefaultMessage": 78193, + "ĠdefaultProps": 79297, + "ĠdefaultValue": 20163, + "Ġdefaultdict": 42908, + "Ġdefaulted": 92989, + "Ġdefaults": 16674, + "Ġdefaultstate": 54458, + "Ġdefe": 10969, + "Ġdefeat": 18019, + "Ġdefeated": 23283, + "Ġdefeating": 53116, + "Ġdefeats": 68769, + "Ġdefect": 22219, + "Ġdefective": 63803, + "Ġdefects": 41555, + "Ġdefence": 22835, + "Ġdefend": 10505, + "Ġdefendant": 25799, + "Ġdefendants": 39146, + "Ġdefended": 34744, + "Ġdefender": 28215, + "Ġdefenders": 40031, + "Ġdefending": 28169, + "Ġdefends": 80960, + "Ġdefense": 9055, + "Ġdefenseman": 82056, + "Ġdefenses": 45516, + "Ġdefensive": 16016, + "Ġdefensively": 81980, + "Ġdefer": 17956, + "Ġdeferred": 26239, + "Ġdefiance": 89783, + "Ġdefiant": 86271, + "Ġdeficiencies": 71846, + "Ġdeficiency": 47194, + "Ġdeficient": 86444, + "Ġdeficit": 28187, + "Ġdeficits": 55996, + "Ġdefin": 35122, + "Ġdefine": 6979, + "Ġdefined": 4512, + "Ġdefines": 18653, + "Ġdefining": 26312, + "Ġdefinit": 30288, + "Ġdefinite": 43770, + "Ġdefinitely": 8491, + "Ġdefinition": 7271, + "Ġdefinitions": 17473, + "Ġdefinitive": 44713, + "Ġdeflate": 92689, + "Ġdeflect": 82940, + "Ġdeform": 78112, + "Ġdeformation": 90521, + "Ġdefs": 75127, + "Ġdefy": 89572, + "Ġdeg": 5254, + "Ġdegli": 46569, + "Ġdegradation": 52468, + "Ġdegrade": 95530, + "Ġdegraded": 90878, + "Ġdegree": 8381, + "Ġdegrees": 12348, + "Ġdehydration": 98757, + "Ġdei": 21805, + "Ġdein": 94919, + "Ġdeine": 72701, + "Ġdeity": 73390, + "Ġdeix": 64756, + "Ġdej": 32963, + "Ġdeja": 75633, + "Ġdejar": 80399, + "Ġdejting": 94981, + "Ġdejtings": 71004, + "Ġdejtingsaj": 80605, + "Ġdel": 1594, + "Ġdela": 89539, + "Ġdelay": 7626, + "Ġdelayed": 22706, + "Ġdelaying": 79330, + "Ġdelays": 31074, + "Ġdele": 59293, + "Ġdeleg": 19854, + "Ġdelegate": 13380, + "Ġdelegated": 90657, + "Ġdelegates": 35059, + "Ġdelegation": 45261, + "Ġdelet": 19269, + "Ġdelete": 3698, + "ĠdeleteUser": 81056, + "Ġdeleted": 11062, + "Ġdeletes": 54170, + "Ġdeleting": 33011, + "Ġdeletion": 36066, + "Ġdeliber": 21150, + "Ġdeliberate": 44972, + "Ġdeliberately": 35092, + "Ġdelic": 60253, + "Ġdelicate": 35201, + "Ġdelicious": 17923, + "Ġdelight": 17970, + "Ġdelighted": 33972, + "Ġdelightful": 49089, + "Ġdelights": 95675, + "Ġdelim": 63209, + "Ġdelimited": 85328, + "Ġdelimiter": 29020, + "Ġdeline": 90684, + "Ġdeliver": 6359, + "Ġdelivered": 12600, + "Ġdeliveries": 54559, + "Ġdelivering": 23988, + "Ġdelivers": 27321, + "Ġdelivery": 9691, + "Ġdell": 24237, + "Ġdella": 15230, + "Ġdelle": 26973, + "Ġdello": 81601, + "Ġdelt": 96250, + "Ġdelta": 9477, + "ĠdeltaTime": 58479, + "ĠdeltaX": 93591, + "ĠdeltaY": 89404, + "Ġdeltas": 90587, + "Ġdeluxe": 78424, + "Ġdelve": 81745, + "Ġdem": 2429, + "Ġdemand": 7479, + "Ġdemande": 61063, + "Ġdemanded": 29426, + "Ġdemanding": 25135, + "Ġdemands": 18154, + "Ġdemasi": 95452, + "Ġdemean": 60452, + "Ġdemeanor": 93015, + "Ġdementia": 51757, + "Ġdemi": 92028, + "Ġdemise": 57864, + "Ġdemo": 16661, + "Ġdemocr": 12074, + "Ġdemocracy": 19525, + "Ġdemocrat": 96067, + "Ġdemocratic": 25542, + "Ġdemographic": 37362, + "Ġdemographics": 62234, + "Ġdemol": 35161, + "Ġdemolished": 75893, + "Ġdemolition": 60598, + "Ġdemon": 6952, + "Ġdemonic": 85124, + "Ġdemons": 43975, + "Ġdemonstr": 7756, + "Ġdemonstrate": 19869, + "Ġdemonstrated": 20459, + "Ġdemonstrates": 31116, + "Ġdemonstrating": 44196, + "Ġdemonstration": 29716, + "Ġdemonstrations": 43795, + "Ġdemonstrators": 66408, + "Ġdemos": 67253, + "Ġdemás": 137492, + "Ġden": 3371, + "Ġdend": 88952, + "Ġdenen": 69507, + "Ġdengan": 20414, + "Ġdenial": 33913, + "Ġdenied": 14820, + "Ġdenies": 46491, + "Ġdenim": 63805, + "Ġdenn": 51939, + "Ġdenne": 64336, + "Ġdenom": 49744, + "Ġdenomin": 61165, + "Ġdenomination": 94936, + "Ġdenominator": 46912, + "Ġdenote": 78064, + "Ġdenotes": 71114, + "Ġdenounced": 70517, + "Ġdens": 38850, + "Ġdense": 27950, + "Ġdensely": 96517, + "Ġdensities": 89716, + "Ġdensity": 17457, + "Ġdent": 18156, + "Ġdental": 28006, + "Ġdentist": 49251, + "Ġdentro": 35128, + "Ġdeny": 23101, + "Ġdenying": 40466, + "Ġdep": 2170, + "Ġdepart": 11515, + "Ġdepartamento": 95889, + "Ġdeparted": 55596, + "Ġdeparting": 80473, + "Ġdepartment": 9292, + "Ġdepartments": 25215, + "Ġdeparture": 24915, + "Ġdepend": 6761, + "Ġdependable": 73336, + "Ġdepended": 78001, + "Ġdependence": 43293, + "Ġdependencies": 19543, + "Ġdependency": 24036, + "Ġdependent": 17749, + "Ġdepending": 11649, + "Ġdepends": 13798, + "Ġdepict": 42404, + "Ġdepicted": 43794, + "Ġdepicting": 59694, + "Ġdepiction": 72664, + "Ġdepicts": 61891, + "Ġdepleted": 78166, + "Ġdepletion": 91848, + "Ġdeploy": 10517, + "Ġdeployed": 26075, + "Ġdeploying": 60317, + "Ġdeployment": 23172, + "Ġdeployments": 71542, + "Ġdepois": 65388, + "Ġdeport": 27239, + "Ġdeportation": 58271, + "Ġdeported": 67110, + "Ġdepos": 25297, + "Ġdeposit": 16539, + "Ġdeposited": 53468, + "Ġdeposition": 64274, + "Ġdeposits": 33651, + "Ġdepot": 44619, + "Ġdepr": 35880, + "Ġdeprecated": 31590, + "Ġdepreci": 71958, + "Ġdepreciation": 93089, + "Ġdepress": 38200, + "Ġdepressed": 41542, + "Ġdepressing": 70710, + "Ġdepression": 18210, + "Ġdepressive": 83464, + "Ġdeprivation": 68664, + "Ġdeprived": 61583, + "Ġdeps": 48178, + "Ġdept": 44593, + "Ġdepth": 7990, + "Ġdepths": 42857, + "Ġdepuis": 40099, + "Ġdeputies": 52828, + "Ġdeputy": 26066, + "Ġdeque": 41950, + "Ġdequeue": 83873, + "ĠdequeueReusableCell": 65382, + "ĠdequeueReusableCellWithIdentifier": 72805, + "Ġder": 2694, + "Ġderail": 82702, + "Ġderby": 90670, + "Ġdere": 25277, + "Ġderecho": 75689, + "Ġderechos": 85535, + "Ġdereg": 71815, + "Ġderen": 87862, + "Ġderiv": 18763, + "Ġderivation": 95254, + "Ġderivative": 31805, + "Ġderivatives": 42545, + "Ġderive": 42430, + "Ġderived": 14257, + "Ġderives": 74449, + "Ġderiving": 48089, + "Ġdermat": 60385, + "Ġdern": 35752, + "Ġdernier": 61791, + "Ġdernière": 87718, + "Ġderog": 91533, + "Ġderp": 66391, + "Ġdes": 939, + "Ġdesar": 39759, + "Ġdesarroll": 63696, + "Ġdesarrollo": 62616, + "Ġdesc": 6560, + "Ġdescargar": 89658, + "Ġdescend": 54568, + "Ġdescendant": 75851, + "Ġdescendants": 48346, + "Ġdescended": 57742, + "Ġdescending": 43084, + "Ġdescent": 36952, + "Ġdescon": 86630, + "Ġdescr": 56217, + "Ġdescri": 5018, + "Ġdescribe": 7512, + "Ġdescribed": 7481, + "Ġdescribes": 16555, + "Ġdescribing": 22692, + "Ġdescricao": 66720, + "Ġdescripcion": 55690, + "Ġdescription": 4008, + "Ġdescriptions": 27787, + "Ġdescriptive": 52844, + "Ġdescriptor": 16414, + "Ġdescriptors": 47517, + "Ġdesde": 22718, + "Ġdese": 50288, + "Ġdesea": 83561, + "Ġdeselect": 97956, + "Ġdesenv": 60244, + "Ġdeser": 96955, + "Ġdeserialize": 35240, + "Ġdesert": 23603, + "Ġdeserted": 82310, + "Ġdeserve": 22695, + "Ġdeserved": 44447, + "Ġdeserves": 29575, + "Ġdeserving": 75454, + "Ġdesi": 85488, + "Ġdesign": 2884, + "Ġdesignate": 74124, + "Ġdesignated": 23195, + "Ġdesignation": 42751, + "Ġdesigned": 6188, + "Ġdesigner": 14692, + "Ġdesigners": 25812, + "Ġdesigning": 29729, + "Ġdesigns": 14431, + "Ġdesirable": 34846, + "Ġdesire": 12591, + "Ġdesired": 12685, + "Ġdesires": 30997, + "Ġdesk": 18010, + "Ġdesks": 75815, + "Ġdesktop": 17505, + "Ġdesn": 65280, + "Ġdesp": 38960, + "Ġdespair": 45896, + "Ġdesper": 17955, + "Ġdesperate": 27395, + "Ġdesperately": 37543, + "Ġdesperation": 73324, + "Ġdespite": 8818, + "Ġdespre": 86625, + "Ġdespués": 49549, + "Ġdess": 20624, + "Ġdessa": 85479, + "Ġdessert": 42749, + "Ġdesserts": 71413, + "Ġdest": 3201, + "Ġdesta": 84778, + "Ġdestabil": 71157, + "Ġdestac": 72367, + "Ġdeste": 76136, + "ĠdesteÄŁi": 140384, + "Ġdestin": 50987, + "Ġdestination": 9106, + "ĠdestinationViewController": 91471, + "Ġdestinations": 33105, + "Ġdestined": 50587, + "Ġdestino": 68097, + "Ġdestiny": 46076, + "Ġdestroy": 6921, + "Ġdestroyed": 14425, + "Ġdestroyer": 95189, + "Ġdestroying": 32712, + "Ġdestroys": 59732, + "Ġdestruct": 20780, + "Ġdestruction": 19258, + "Ġdestructive": 39552, + "Ġdestructor": 45999, + "Ġdet": 3392, + "Ġdetach": 46781, + "Ġdetached": 43917, + "Ġdetachment": 99077, + "Ġdetail": 7716, + "Ġdetailed": 11682, + "Ġdetailing": 44193, + "Ġdetails": 3565, + "Ġdetain": 45582, + "Ġdetained": 36893, + "Ġdetainees": 66934, + "Ġdetal": 52152, + "Ġdetalle": 79720, + "Ġdetalles": 91329, + "Ġdetect": 11140, + "Ġdetected": 16507, + "Ġdetecting": 53526, + "Ġdetection": 17984, + "Ġdetections": 88627, + "Ġdetective": 44159, + "Ġdetectives": 58759, + "Ġdetector": 31214, + "Ġdetectors": 67987, + "Ġdetects": 66478, + "Ġdetention": 32173, + "Ġdeter": 4040, + "Ġdetergent": 90134, + "Ġdeterior": 38336, + "Ġdeterioration": 81089, + "Ġdetermin": 6315, + "Ġdeterminant": 86960, + "Ġdetermination": 25248, + "Ġdetermine": 8253, + "Ġdetermined": 10838, + "Ġdetermines": 26569, + "Ġdetermining": 25597, + "Ġdeterministic": 72349, + "Ġdeterrent": 82620, + "Ġdeton": 54674, + "Ġdetox": 60887, + "Ġdetr": 34353, + "Ġdetriment": 48908, + "Ġdetrimental": 63969, + "Ġdetta": 97197, + "Ġdette": 58928, + "Ġdeut": 83630, + "Ġdeutsch": 57184, + "Ġdeutsche": 68517, + "Ġdeutschen": 92191, + "Ġdeutschland": 92516, + "Ġdeux": 25552, + "Ġdeuxième": 139346, + "Ġdev": 3483, + "Ġdevant": 76458, + "Ġdevast": 22440, + "Ġdevastated": 57997, + "Ġdevastating": 32218, + "Ġdevastation": 81915, + "Ġdeve": 36144, + "Ġdevel": 89770, + "Ġdevelop": 2225, + "Ġdeveloped": 7881, + "Ġdeveloper": 15754, + "Ġdevelopers": 13402, + "Ġdeveloping": 11220, + "Ġdevelopment": 4401, + "Ġdevelopmental": 46906, + "Ġdevelopments": 24961, + "Ġdevelops": 38571, + "Ġdever": 87381, + "Ġdevez": 85940, + "Ġdeviation": 37564, + "Ġdeviations": 85265, + "Ġdevice": 3671, + "ĠdeviceId": 64757, + "Ġdevices": 7611, + "Ġdevil": 37636, + "Ġdevis": 97133, + "Ġdevise": 52065, + "Ġdevised": 68020, + "Ġdevoid": 75839, + "Ġdevote": 47672, + "Ġdevoted": 28229, + "Ġdevotion": 55257, + "Ġdevour": 87189, + "Ġdevout": 94562, + "Ġdevs": 80069, + "Ġdevuelve": 95567, + "Ġdew": 66432, + "Ġdex": 79901, + "Ġdez": 70768, + "Ġdeze": 34223, + "ĠdeÄŁ": 60987, + "ĠdeÄŁer": 125724, + "ĠdeÄŁerl": 127866, + "ĠdeÄŁerlendir": 127867, + "ĠdeÄŁerlendirme": 139057, + "ĠdeÄŁi": 126571, + "ĠdeÄŁil": 128721, + "ĠdeÄŁild": 133594, + "ĠdeÄŁildir": 133595, + "ĠdeÄŁiÅŁ": 129399, + "ĠdeÄŁiÅŁtir": 133382, + "Ġdf": 6764, + "Ġdfs": 25792, + "Ġdg": 40286, + "Ġdgv": 50971, + "Ġdh": 34096, + "Ġdhcp": 85787, + "Ġdhe": 76905, + "Ġdi": 1853, + "Ġdia": 17733, + "Ġdiabetes": 19754, + "Ġdiabetic": 70875, + "Ġdiag": 39717, + "Ġdiagn": 13191, + "Ġdiagnose": 57581, + "Ġdiagnosed": 28604, + "Ġdiagnoses": 84388, + "Ġdiagnosis": 22982, + "Ġdiagnostic": 15089, + "Ġdiagnostics": 49418, + "Ġdiagon": 84018, + "Ġdiagonal": 37236, + "Ġdiagram": 13549, + "Ġdiagrams": 46187, + "Ġdial": 27860, + "Ġdialect": 42279, + "Ġdialog": 7254, + "ĠdialogRef": 79777, + "Ġdialogs": 86236, + "Ġdialogue": 21276, + "Ġdiam": 12757, + "Ġdiameter": 23033, + "Ġdiamond": 22205, + "Ġdiamonds": 48051, + "Ġdiaper": 75753, + "Ġdiapers": 82862, + "Ġdiarr": 57605, + "Ġdiarrhea": 68293, + "Ġdiary": 46142, + "Ġdias": 40370, + "Ġdib": 41068, + "Ġdic": 21249, + "Ġdice": 22120, + "Ġdiced": 88864, + "Ġdich": 28853, + "Ġdicho": 92860, + "Ġdici": 79819, + "Ġdick": 25985, + "Ġdicks": 95498, + "Ġdict": 6451, + "Ġdictate": 61874, + "Ġdictated": 80398, + "Ġdictates": 79230, + "Ġdictator": 41184, + "Ġdictatorship": 68848, + "Ġdictionaries": 57514, + "Ġdictionary": 10997, + "ĠdictionaryWith": 84748, + "Ġdicts": 97599, + "Ġdid": 1521, + "ĠdidFinish": 76150, + "ĠdidReceiveMemoryWarning": 38141, + "ĠdidSelect": 52379, + "ĠdidSelectRowAtIndexPath": 95607, + "ĠdidSet": 62109, + "Ġdidn": 3207, + "Ġdidnt": 47607, + "Ġdie": 2746, + "Ġdied": 8469, + "Ġdies": 8725, + "Ġdiese": 26769, + "Ġdiesel": 32780, + "Ġdiesem": 36744, + "Ġdiesen": 50435, + "Ġdieser": 31008, + "Ġdieses": 46668, + "Ġdiet": 9968, + "Ġdieta": 66856, + "Ġdietary": 33525, + "Ġdiets": 40843, + "Ġdif": 1396, + "Ġdifer": 26929, + "Ġdiferen": 139741, + "Ġdiferencia": 94427, + "Ġdiferente": 77173, + "Ġdiferentes": 45318, + "Ġdiferença": 139742, + "Ġdiff": 3638, + "Ġdiffer": 1745, + "Ġdiffered": 87975, + "Ġdifference": 6672, + "Ġdifferences": 11799, + "Ġdifferent": 2155, + "Ġdifferential": 40164, + "Ġdifferentiate": 53163, + "Ġdifferentiated": 88042, + "Ġdifferentiation": 58938, + "Ġdifferently": 21303, + "Ġdiffering": 60369, + "Ġdiffers": 43542, + "Ġdiffic": 60807, + "Ġdifficile": 76311, + "Ġdifficolt": 141990, + "ĠdifficoltÃł": 141991, + "Ġdifficult": 5000, + "Ġdifficulties": 26038, + "Ġdifficulty": 16829, + "Ġdifficulté": 142370, + "Ġdiffs": 86947, + "Ġdiffuse": 54417, + "Ġdiffusion": 57330, + "Ġdiffé": 43078, + "Ġdifférence": 143108, + "Ġdifférent": 50099, + "Ġdifférentes": 95542, + "Ġdifférents": 84014, + "Ġdific": 77880, + "ĠdifÃŃc": 78865, + "ĠdifÃŃcil": 85480, + "Ġdig": 4078, + "Ġdigest": 20882, + "Ġdigestion": 73402, + "Ġdigestive": 58971, + "Ġdigging": 41100, + "Ġdigit": 15723, + "Ġdigital": 7377, + "ĠdigitalWrite": 64797, + "Ġdigitally": 67778, + "Ġdigits": 18509, + "Ġdign": 27577, + "Ġdignity": 37829, + "Ġdigs": 88957, + "Ġdijo": 64910, + "Ġdik": 46199, + "Ġdikke": 75038, + "Ġdil": 18840, + "Ġdilation": 74278, + "Ġdildo": 36503, + "Ġdile": 43161, + "Ġdilemma": 54767, + "Ġdilig": 38287, + "Ġdiligence": 62482, + "Ġdiligent": 96553, + "Ġdiligently": 90605, + "Ġdiluted": 79603, + "Ġdim": 5103, + "Ġdime": 73853, + "Ġdimension": 12871, + "Ġdimensional": 55887, + "Ġdimensions": 15336, + "Ġdimin": 47316, + "Ġdiminish": 56060, + "Ġdiminished": 53082, + "Ġdiminishing": 83053, + "Ġdims": 40888, + "Ġdin": 11623, + "Ġdine": 65107, + "Ġdiner": 88106, + "Ġdinero": 65808, + "Ġding": 66442, + "Ġdinheiro": 96487, + "Ġdining": 17914, + "Ġdinner": 13856, + "Ġdinners": 73313, + "Ġdinosaur": 62889, + "Ġdinosaurs": 64275, + "Ġdint": 84489, + "Ġdio": 64470, + "Ġdioxide": 39489, + "Ġdip": 23604, + "Ġdipl": 16918, + "Ġdiplom": 29521, + "Ġdiploma": 60712, + "Ġdiplomacy": 61240, + "Ġdiplomat": 67950, + "Ġdiplomatic": 33516, + "Ġdiplomats": 64079, + "Ġdipped": 72207, + "Ġdipping": 85821, + "Ġdips": 96504, + "Ġdir": 5419, + "Ġdire": 13207, + "Ġdireccion": 68807, + "Ġdirección": 78113, + "Ġdirect": 2118, + "Ġdirected": 15540, + "Ġdirecting": 44990, + "Ġdirection": 5106, + "Ġdirectional": 72845, + "Ġdirections": 17961, + "Ġdirective": 33859, + "Ġdirectives": 49253, + "Ġdirectly": 5961, + "Ġdirector": 7538, + "Ġdirectories": 28625, + "Ġdirectors": 27354, + "Ġdirectory": 6220, + "Ġdirects": 66046, + "Ġdirekt": 54831, + "Ġdirent": 95957, + "Ġdiret": 80485, + "Ġdirig": 63856, + "Ġdirname": 29908, + "Ġdirs": 42248, + "Ġdirt": 25284, + "Ġdirty": 18595, + "Ġdis": 827, + "Ġdisabilities": 35701, + "Ġdisability": 27254, + "Ġdisable": 11156, + "Ġdisabled": 8386, + "Ġdisables": 72996, + "Ġdisabling": 60484, + "Ġdisadv": 29847, + "Ġdisadvantage": 48736, + "Ġdisadvantaged": 79847, + "Ġdisadvantages": 63625, + "Ġdisag": 18288, + "Ġdisagree": 28295, + "Ġdisagreed": 70703, + "Ġdisagreement": 61546, + "Ġdisagreements": 90872, + "Ġdisagrees": 94983, + "Ġdisap": 13950, + "Ġdisappe": 20837, + "Ġdisappear": 31053, + "Ġdisappearance": 51879, + "Ġdisappeared": 28396, + "Ġdisappearing": 66403, + "Ġdisappears": 60270, + "Ġdisappoint": 14097, + "Ġdisappointed": 24402, + "Ġdisappointing": 38123, + "Ġdisappointment": 40598, + "Ġdisarm": 86776, + "Ġdisaster": 20763, + "Ġdisasters": 50651, + "Ġdisastr": 50153, + "Ġdisastrous": 51957, + "Ġdisb": 93444, + "Ġdisbelief": 73254, + "Ġdisc": 2560, + "Ġdiscard": 37867, + "Ġdiscarded": 43210, + "Ġdiscern": 41545, + "Ġdischarge": 31543, + "Ġdischarged": 56091, + "Ġdiscipl": 15501, + "Ġdisciple": 83695, + "Ġdisciples": 48160, + "Ġdisciplinary": 56215, + "Ġdiscipline": 25364, + "Ġdisciplined": 69066, + "Ġdisciplines": 48155, + "Ġdisclaimer": 17673, + "Ġdisclose": 35233, + "Ġdisclosed": 35389, + "Ġdisclosing": 94835, + "Ġdisclosure": 27857, + "Ġdisclosures": 74892, + "Ġdisco": 53869, + "Ġdiscomfort": 43676, + "Ġdisconnect": 28969, + "Ġdisconnected": 39086, + "Ġdiscontent": 90122, + "Ġdiscontin": 44880, + "Ġdiscontinued": 64159, + "Ġdiscord": 31041, + "Ġdiscount": 11089, + "Ġdiscounted": 47674, + "Ġdiscounts": 31062, + "Ġdiscour": 33193, + "Ġdiscourage": 64987, + "Ġdiscouraged": 63670, + "Ġdiscourse": 40502, + "Ġdiscover": 6997, + "Ġdiscovered": 11105, + "Ġdiscoveries": 52998, + "Ġdiscovering": 41587, + "Ġdiscovers": 51014, + "Ġdiscovery": 18335, + "Ġdiscre": 33938, + "Ġdiscredit": 87019, + "Ġdiscreet": 61404, + "Ġdiscrepan": 53026, + "Ġdiscrepancies": 90267, + "Ġdiscrepancy": 78005, + "Ġdiscret": 27458, + "Ġdiscrete": 43179, + "Ġdiscretion": 30922, + "Ġdiscretionary": 95794, + "Ġdiscrim": 14238, + "Ġdiscrimin": 49319, + "Ġdiscriminate": 59441, + "Ġdiscrimination": 21240, + "Ġdiscriminator": 81738, + "Ġdiscriminatory": 67663, + "Ġdiscs": 56695, + "Ġdiscuss": 4263, + "Ġdiscussed": 14078, + "Ġdiscusses": 34334, + "Ġdiscussing": 24392, + "Ġdiscussion": 10219, + "Ġdiscussions": 20333, + "Ġdisdain": 78398, + "Ġdise": 6621, + "Ġdisease": 8457, + "Ġdiseases": 18808, + "Ġdisemb": 97544, + "Ġdisen": 90810, + "Ġdiseñ": 66481, + "Ġdiseñador": 93556, + "Ġdiseño": 83808, + "Ġdisfr": 66057, + "Ġdisg": 32183, + "Ġdisgr": 44186, + "Ġdisgrace": 63835, + "Ġdisgu": 40777, + "Ġdisguise": 63981, + "Ġdisguised": 78087, + "Ġdisgust": 67062, + "Ġdisgusted": 90074, + "Ġdisgusting": 53666, + "Ġdish": 12000, + "Ġdishes": 25779, + "Ġdishonest": 58916, + "Ġdishwasher": 60333, + "Ġdisillusion": 96909, + "Ġdisin": 63742, + "Ġdisjoint": 84784, + "Ġdisk": 13364, + "Ġdisks": 53758, + "Ġdislike": 47869, + "Ġdisliked": 98633, + "Ġdislikes": 98885, + "Ġdismal": 93453, + "Ġdismant": 48112, + "Ġdismantle": 94058, + "Ġdismay": 71982, + "Ġdismiss": 13432, + "Ġdismissal": 58223, + "Ġdismissed": 26200, + "Ġdismissing": 91495, + "Ġdisob": 67635, + "Ġdisobed": 91583, + "Ġdisorder": 19267, + "Ġdisorders": 23741, + "Ġdisp": 12965, + "Ġdispar": 28804, + "Ġdisparate": 84029, + "Ġdisparities": 81957, + "Ġdisparity": 65849, + "Ġdispatch": 6845, + "Ġdispatched": 49151, + "Ġdispatcher": 38799, + "Ġdispens": 35593, + "Ġdispensaries": 99139, + "Ġdispenser": 99127, + "Ġdispers": 78735, + "Ġdispersed": 76710, + "Ġdispersion": 85612, + "Ġdispl": 13462, + "Ġdisplaced": 48146, + "Ġdisplacement": 43053, + "Ġdisplay": 3037, + "ĠdisplayName": 34770, + "Ġdisplayed": 12596, + "Ġdisplaying": 27940, + "Ġdisplays": 18689, + "Ġdisple": 84234, + "Ġdispon": 25604, + "Ġdisponible": 54752, + "Ġdisponibles": 69620, + "ĠdisponÃŃveis": 142462, + "ĠdisponÃŃvel": 142672, + "Ġdispos": 53140, + "Ġdisposable": 43790, + "Ġdisposal": 33445, + "Ġdispose": 27390, + "Ġdisposed": 25574, + "Ġdisposing": 23212, + "Ġdisposit": 57451, + "Ġdisposition": 44996, + "Ġdispositivo": 94999, + "Ġdispro": 41503, + "Ġdisproportion": 46535, + "Ġdisproportionate": 79053, + "Ġdisproportionately": 72052, + "Ġdisput": 34060, + "Ġdispute": 25033, + "Ġdisputed": 53926, + "Ġdisputes": 41154, + "Ġdisqualified": 93576, + "Ġdisreg": 42803, + "Ġdisregard": 57717, + "Ġdisrespect": 56209, + "Ġdisrespectful": 98522, + "Ġdisrupt": 23973, + "Ġdisrupted": 68527, + "Ġdisrupting": 86743, + "Ġdisruption": 43119, + "Ġdisruptions": 96957, + "Ġdisruptive": 61542, + "Ġdiss": 13772, + "Ġdissatisfaction": 96794, + "Ġdisse": 32900, + "Ġdissect": 73950, + "Ġdissemination": 86664, + "Ġdissent": 46862, + "Ġdissert": 33649, + "Ġdissertation": 36345, + "Ġdissip": 92422, + "Ġdissoci": 90242, + "Ġdissolution": 80259, + "Ġdissolve": 72639, + "Ġdissolved": 55667, + "Ġdist": 1582, + "Ġdistance": 6010, + "Ġdistances": 26552, + "Ġdistancia": 82309, + "Ġdistancing": 72569, + "Ġdistant": 28727, + "Ġdistilled": 88767, + "Ġdistinct": 12460, + "Ġdistinction": 29196, + "Ġdistinctions": 86159, + "Ġdistinctive": 34847, + "Ġdistinctly": 70149, + "Ġdistingu": 18975, + "Ġdistinguish": 32037, + "Ġdistinguished": 38475, + "Ġdistinguishing": 84955, + "Ġdistint": 73378, + "Ġdistort": 69484, + "Ġdistorted": 61136, + "Ġdistortion": 49871, + "Ġdistr": 7905, + "Ġdistra": 57471, + "Ġdistract": 63817, + "Ġdistracted": 48704, + "Ġdistracting": 81298, + "Ġdistraction": 53616, + "Ġdistractions": 76449, + "Ġdistress": 34004, + "Ġdistressed": 70390, + "Ġdistrib": 2846, + "Ġdistribute": 16417, + "Ġdistributed": 4237, + "Ġdistributes": 95735, + "Ġdistributing": 49712, + "Ġdistribution": 7982, + "Ġdistributions": 42685, + "Ġdistributor": 43068, + "Ġdistributors": 55594, + "Ġdistrict": 9290, + "Ġdistricts": 26438, + "Ġdistrust": 75986, + "Ġdistur": 25343, + "Ġdisturb": 37223, + "Ġdisturbance": 64758, + "Ġdisturbances": 84060, + "Ġdisturbed": 55407, + "Ġdisturbing": 33873, + "Ġdistância": 143384, + "Ġdit": 21305, + "Ġditch": 49004, + "Ġdiv": 3429, + "Ġdive": 29863, + "Ġdiver": 36341, + "Ġdivergence": 81158, + "Ġdivers": 21113, + "Ġdiversas": 98283, + "Ġdiverse": 16807, + "Ġdiversified": 84857, + "Ġdiversion": 76264, + "Ġdiversity": 19492, + "Ġdiversos": 94551, + "Ġdivert": 35998, + "Ġdiverted": 85394, + "Ġdives": 93421, + "Ġdivid": 28832, + "Ġdivide": 21749, + "Ġdivided": 17779, + "Ġdividend": 45880, + "Ġdividends": 55264, + "Ġdivider": 48695, + "Ġdivides": 64828, + "Ġdividing": 49702, + "Ġdivine": 29367, + "Ġdiving": 42415, + "Ġdivis": 49368, + "Ġdivisible": 74916, + "Ġdivision": 12804, + "Ġdivisions": 36501, + "Ġdivisive": 74135, + "Ġdivisor": 49109, + "Ġdivor": 39979, + "Ġdivorce": 24532, + "Ġdivorced": 49492, + "Ġdivul": 81691, + "Ġdiy": 51338, + "Ġdiz": 42286, + "Ġdizzy": 84084, + "Ġdiá": 143565, + "Ġdiálogo": 143566, + "ĠdiÄŁer": 129189, + "Ġdiá»ħn": 128960, + "Ġdiá»ĩn": 129054, + "Ġdj": 47756, + "Ġdjango": 8262, + "Ġdk": 40204, + "Ġdl": 27513, + "Ġdla": 28976, + "Ġdlg": 40871, + "Ġdll": 58155, + "Ġdm": 28676, + "Ġdma": 50250, + "Ġdmg": 62153, + "Ġdn": 47488, + "Ġdna": 75334, + "Ġdni": 60759, + "Ġdns": 44077, + "Ġdo": 653, + "ĠdoGet": 56704, + "ĠdoInBackground": 71840, + "ĠdoPost": 65156, + "Ġdob": 45877, + "Ġdobr": 96568, + "Ġdoc": 4629, + "Ġdoch": 40157, + "Ġdock": 27549, + "Ġdocker": 26588, + "Ġdocking": 88210, + "Ġdocks": 96886, + "Ġdocs": 26340, + "Ġdoctor": 10668, + "Ġdoctoral": 73557, + "Ġdoctors": 16026, + "Ġdoctr": 60890, + "Ġdoctrine": 32135, + "Ġdoctrines": 82158, + "Ġdocument": 2197, + "Ġdocumentaries": 87366, + "Ġdocumentary": 24954, + "Ġdocumentation": 9705, + "Ġdocumented": 26372, + "Ġdocumenting": 66971, + "Ġdocumento": 53426, + "Ġdocumentos": 99027, + "Ġdocuments": 9293, + "Ġdod": 31089, + "Ġdodge": 47873, + "Ġdoe": 98453, + "Ġdoen": 53902, + "Ġdoença": 138716, + "Ġdoes": 1558, + "Ġdoesn": 3171, + "Ġdoesnt": 43104, + "Ġdof": 91760, + "Ġdog": 5562, + "Ġdogs": 12590, + "Ġdoi": 35895, + "Ġdoing": 3730, + "Ġdois": 50341, + "Ġdoit": 41082, + "Ġdoivent": 96469, + "Ġdojo": 84859, + "Ġdok": 41061, + "Ġdokument": 85961, + "ĠdokÅĤad": 140343, + "ĠdokÅĤadnie": 140344, + "Ġdol": 21270, + "Ġdolayı": 132595, + "Ġdoll": 7649, + "Ġdollar": 17692, + "Ġdollars": 11192, + "Ġdolls": 63781, + "Ġdolor": 23655, + "Ġdolore": 57296, + "Ġdolphin": 98169, + "Ġdolphins": 82406, + "Ġdom": 4719, + "Ġdomain": 7947, + "Ġdomaine": 95739, + "Ġdomains": 30476, + "Ġdome": 57338, + "Ġdomest": 73422, + "Ġdomestic": 12728, + "Ġdomestically": 97790, + "Ġdomic": 53885, + "Ġdomicile": 96739, + "Ġdomin": 11111, + "Ġdomina": 89037, + "Ġdominance": 43492, + "Ġdominant": 24456, + "Ġdominate": 40736, + "Ġdominated": 29701, + "Ġdominates": 82878, + "Ġdominating": 67266, + "Ġdomination": 54849, + "Ġdomingo": 89440, + "Ġdon": 1513, + "Ġdonate": 31909, + "Ġdonated": 33368, + "Ġdonating": 60859, + "Ġdonation": 24923, + "Ġdonations": 23957, + "Ġdonc": 38487, + "Ġdonde": 29040, + "Ġdone": 2814, + "Ġdong": 72736, + "Ġdonn": 32437, + "Ġdonna": 37756, + "Ġdonne": 20889, + "Ġdonner": 70651, + "Ġdonné": 136383, + "Ġdonnées": 45581, + "Ġdonor": 34458, + "Ġdonors": 32049, + "Ġdont": 15522, + "Ġdood": 87708, + "Ġdoom": 58614, + "Ġdoomed": 57637, + "Ġdoor": 6006, + "Ġdoors": 14038, + "Ġdoorstep": 75877, + "Ġdoorway": 69358, + "Ġdop": 26323, + "Ġdopamine": 65028, + "Ġdope": 91312, + "Ġdoping": 96828, + "Ġdopo": 59194, + "Ġdor": 52303, + "Ġdorm": 29109, + "Ġdormant": 88360, + "Ġdors": 76289, + "Ġdorsal": 95046, + "Ġdort": 52434, + "Ġdos": 8750, + "Ġdosage": 45940, + "Ġdose": 19115, + "Ġdoses": 34030, + "Ġdossier": 52476, + "Ġdost": 48315, + "ĠdostÄĻp": 127028, + "ĠdostÄĻpny": 142326, + "Ġdot": 12756, + "Ġdotenv": 91286, + "Ġdots": 30994, + "Ġdotted": 58101, + "Ġdotyc": 138089, + "ĠdotyczÄħ": 138090, + "ĠdotyczÄħce": 138091, + "Ġdou": 24576, + "Ġdoub": 7655, + "Ġdouble": 1990, + "Ġdoubled": 34617, + "Ġdoubles": 39296, + "Ġdoubling": 59015, + "Ġdoubly": 93889, + "Ġdoubt": 10492, + "Ġdoubted": 92662, + "Ġdoubtful": 74599, + "Ġdoubts": 38917, + "Ġdouche": 84353, + "Ġdough": 30352, + "Ġdout": 92520, + "Ġdov": 61270, + "Ġdove": 47222, + "Ġdow": 56487, + "Ġdown": 1495, + "Ġdownfall": 87767, + "Ġdowngrade": 90357, + "Ġdownhill": 72381, + "Ġdownload": 4139, + "Ġdownloadable": 69652, + "Ġdownloaded": 23293, + "Ġdownloader": 86774, + "Ġdownloading": 32477, + "Ġdownloads": 30472, + "Ġdownright": 69952, + "Ġdowns": 39191, + "Ġdownside": 51397, + "Ġdownstairs": 62791, + "Ġdownstream": 51352, + "Ġdownt": 40557, + "Ġdowntime": 74854, + "Ġdowntown": 18907, + "Ġdownturn": 75406, + "Ġdownward": 44478, + "Ġdownwards": 91270, + "Ġdozen": 20403, + "Ġdozens": 21935, + "ĠdoÄŁ": 93282, + "ĠdoÄŁal": 136317, + "ĠdoÄŁru": 129204, + "ĠdoÄŁum": 141663, + "Ġdp": 11329, + "Ġdpi": 50742, + "Ġdps": 94029, + "Ġdq": 71613, + "Ġdr": 1353, + "Ġdra": 44096, + "Ġdracon": 92703, + "Ġdraft": 9960, + "Ġdrafted": 37956, + "Ġdrafting": 57900, + "Ġdrafts": 63592, + "Ġdrag": 10921, + "Ġdraggable": 74556, + "Ġdragged": 37147, + "Ġdragging": 42376, + "Ġdragon": 25105, + "Ġdragons": 50436, + "Ġdrain": 23727, + "Ġdrainage": 57492, + "Ġdrained": 60408, + "Ġdraining": 73009, + "Ġdrains": 78650, + "Ġdram": 13548, + "Ġdrama": 19584, + "Ġdramas": 87726, + "Ġdramatic": 21771, + "Ġdramatically": 27957, + "Ġdrank": 53144, + "Ġdraped": 89880, + "Ġdrastic": 59783, + "Ġdrastically": 47763, + "Ġdraw": 4038, + "Ġdrawable": 52071, + "Ġdrawback": 87520, + "Ġdrawbacks": 88073, + "Ġdrawer": 26482, + "Ġdrawers": 61607, + "Ġdrawing": 13330, + "Ġdrawings": 37840, + "Ġdrawn": 14764, + "Ġdraws": 26643, + "Ġdre": 32101, + "Ġdread": 34975, + "Ġdreaded": 82490, + "Ġdreadful": 80852, + "Ġdream": 7904, + "Ġdreamed": 56536, + "Ġdreaming": 55674, + "Ġdreams": 18707, + "Ġdred": 90362, + "Ġdrei": 57519, + "Ġdress": 8511, + "Ġdressed": 25365, + "Ġdresser": 95537, + "Ġdresses": 36762, + "Ġdressing": 31523, + "Ġdrew": 23554, + "Ġdri": 49411, + "Ġdrib": 78546, + "Ġdried": 31620, + "Ġdrift": 33638, + "Ġdrifted": 91562, + "Ġdrifting": 84253, + "Ġdrill": 30546, + "Ġdrilled": 67765, + "Ġdrilling": 38562, + "Ġdrills": 56548, + "Ġdrink": 7027, + "Ġdrinkers": 97876, + "Ġdrinking": 16163, + "Ġdrinks": 20987, + "Ġdrip": 72499, + "Ġdripping": 76215, + "Ġdriv": 39255, + "Ġdrive": 6541, + "Ġdriven": 16227, + "Ġdriver": 5579, + "Ġdrivers": 11788, + "Ġdrives": 20118, + "Ġdriveway": 52142, + "Ġdriving": 9842, + "Ġdrm": 53086, + "Ġdro": 6973, + "Ġdroit": 62979, + "Ġdroits": 96122, + "Ġdrone": 26713, + "Ġdrones": 37232, + "Ġdrop": 5943, + "Ġdropdown": 20389, + "Ġdropout": 43281, + "Ġdropped": 12226, + "Ġdropping": 25100, + "Ġdrops": 21025, + "Ġdrought": 36746, + "Ġdrove": 23108, + "Ġdrown": 86868, + "Ġdrowned": 73895, + "Ġdrowning": 73437, + "Ġdru": 142231, + "Ġdrug": 5506, + "Ġdrugs": 10975, + "Ġdrum": 23196, + "Ġdrummer": 67946, + "Ġdrums": 46289, + "Ġdrunk": 28750, + "Ġdrunken": 62363, + "Ġdruż": 142232, + "Ġdrużyn": 142233, + "Ġdrv": 68770, + "Ġdry": 9058, + "Ġdryer": 46622, + "Ġdrying": 45379, + "Ġds": 11472, + "Ġdsp": 78615, + "Ġdst": 10648, + "Ġdt": 7594, + "Ġdto": 28335, + "Ġdtype": 13231, + "Ġdu": 3845, + "Ġdua": 96821, + "Ġdual": 18579, + "Ġduas": 71243, + "Ġdub": 22113, + "Ġdubbed": 41060, + "Ġdubious": 62089, + "Ġduck": 35985, + "Ġducks": 77778, + "Ġduct": 44239, + "Ġdude": 35057, + "Ġdudes": 72122, + "Ġdue": 4152, + "Ġduel": 65307, + "Ġdues": 79673, + "Ġdug": 43020, + "Ġduk": 87477, + "Ġduke": 95823, + "Ġdul": 81527, + "Ġdull": 40530, + "Ġduly": 76803, + "Ġdum": 62711, + "Ġdumb": 29255, + "Ġdummy": 17292, + "Ġdump": 10276, + "Ġdumped": 48273, + "Ġdumping": 60582, + "Ġdumps": 60381, + "Ġdumpster": 27450, + "Ġdumpsters": 98476, + "Ġdun": 49016, + "Ġdung": 85536, + "Ġdungeon": 42439, + "Ġdungeons": 82385, + "Ġdunk": 63472, + "Ġduo": 33721, + "Ġdup": 22737, + "Ġduplex": 80168, + "Ġduplic": 26347, + "Ġduplicate": 22513, + "Ġduplicated": 54903, + "Ġduplicates": 42328, + "Ġduplication": 66533, + "Ġdur": 10651, + "Ġdurability": 38565, + "Ġdurable": 26128, + "Ġdurante": 29231, + "Ġduration": 8090, + "Ġdurations": 89104, + "Ġdurch": 19767, + "Ġduring": 2337, + "Ġduro": 91096, + "Ġdurée": 141007, + "Ġdus": 52432, + "Ġdusk": 92885, + "Ġdust": 15797, + "Ġdusty": 76873, + "Ġdut": 10043, + "Ġduties": 21300, + "Ġduty": 14189, + "Ġduy": 141839, + "Ġduyá»ĩt": 141840, + "Ġduż": 88738, + "Ġduże": 135340, + "Ġdużo": 133768, + "Ġdv": 32776, + "Ġdvd": 49908, + "Ġdw": 13835, + "Ġdwar": 70043, + "Ġdwarf": 49461, + "Ġdwell": 43835, + "Ġdwelling": 50588, + "Ġdwind": 82122, + "Ġdword": 74750, + "Ġdwó": 134451, + "Ġdwóch": 134452, + "Ġdx": 13822, + "Ġdy": 13955, + "Ġdye": 53531, + "Ġdying": 22273, + "Ġdyn": 31070, + "Ġdynam": 17544, + "Ġdynamic": 8741, + "Ġdynamically": 42011, + "Ġdynamics": 29195, + "Ġdynasty": 66846, + "Ġdys": 21943, + "Ġdysfunction": 31303, + "Ġdysfunctional": 87704, + "Ġdyst": 68010, + "Ġdz": 25718, + "Ġdzi": 51026, + "ĠdziaÅĤ": 74281, + "ĠdziaÅĤa": 129558, + "ĠdziaÅĤal": 137683, + "ĠdziaÅĤalnoÅĽci": 137684, + "ĠdziaÅĤania": 135516, + "Ġdzie": 49673, + "Ġdzieci": 84883, + "Ġdziew": 83504, + "Ġdziewcz": 89841, + "ĠdzieÅĦ": 133223, + "ĠdziÄĻki": 131141, + "ĠdziÅĽ": 135481, + "Ġdá": 130691, + "Ġdáng": 133806, + "Ġdân": 128394, + "Ġdây": 132625, + "Ġdär": 77028, + "ĠdÃ¥": 74345, + "Ġdès": 86351, + "Ġdé": 7439, + "Ġdébut": 64285, + "Ġdéc": 34781, + "Ġdécada": 137411, + "Ġdécembre": 134829, + "Ġdécid": 142440, + "Ġdécidé": 142441, + "Ġdécision": 138828, + "Ġdéco": 134189, + "Ġdécor": 74111, + "Ġdécou": 58048, + "Ġdécouvert": 139318, + "Ġdécouvrir": 90627, + "Ġdéf": 46271, + "Ġdéfin": 135315, + "Ġdéfini": 135316, + "ĠdéjÃł": 45839, + "Ġdél": 143359, + "Ġdélai": 143360, + "Ġdém": 78541, + "Ġdémarch": 143764, + "Ġdémarche": 143765, + "Ġdép": 86785, + "Ġdépart": 75135, + "Ġdépartement": 141055, + "Ġdés": 45738, + "Ġdésir": 142429, + "Ġdésorm": 141803, + "Ġdésormais": 141804, + "Ġdét": 84777, + "Ġdétail": 135976, + "Ġdévelop": 51795, + "Ġdévelopp": 67766, + "Ġdéveloppe": 138681, + "Ġdéveloppement": 81520, + "Ġdòng": 130423, + "Ġdó": 76414, + "Ġdóla": 139042, + "Ġdólares": 139043, + "Ġdõ": 136752, + "Ġdõi": 136753, + "Ġdön": 124932, + "Ġdönem": 127061, + "Ġdönemde": 139016, + "Ġdönemin": 142391, + "Ġdöneminde": 142392, + "ĠdönÃ¼ÅŁ": 132768, + "Ġdù": 129655, + "Ġdùng": 128469, + "Ġdú": 137939, + "Ġdúvida": 137940, + "Ġdü": 51019, + "Ġdüny": 139704, + "Ġdünya": 129463, + "Ġdünyan": 139705, + "Ġdünyanın": 139706, + "Ġdür": 84952, + "Ġdürfen": 141489, + "Ġdüz": 138183, + "Ġdüzen": 125376, + "Ġdüzenle": 127545, + "Ġdüzenleme": 140392, + "Ġdüzenlen": 140323, + "Ġdüzenlenen": 140324, + "Ġdüzey": 138184, + "ĠdÃ¼ÅŁ": 125151, + "ĠdÃ¼ÅŁÃ¼k": 138145, + "ĠdÃ¼ÅŁÃ¼n": 126535, + "ĠdÃ¼ÅŁÃ¼nce": 136743, + "ĠdÃłi": 129121, + "ĠdÃłng": 131151, + "ĠdÃłnh": 130176, + "ĠdÃły": 136861, + "ĠdÃŃa": 34863, + "ĠdÃŃas": 45080, + "ĠdÄ±ÅŁ": 133220, + "ĠdÄ±ÅŁÄ±": 138792, + "ĠdÄ±ÅŁÄ±nda": 133221, + "ĠdÅĤ": 127507, + "ĠdÅĤugi": 139908, + "ĠdÅĤugo": 133671, + "Ġdư": 136525, + "Ġdương": 135676, + "Ġdưỡng": 129852, + "ĠdÆ°á»Ľi": 129227, + "Ġdạ": 135648, + "Ġdạng": 130249, + "Ġdạy": 132856, + "Ġdấu": 131575, + "Ġdần": 134724, + "Ġdầu": 131318, + "Ġdẫn": 128653, + "ĠdáºŃy": 139861, + "Ġdụ": 126200, + "Ġdục": 129677, + "Ġdụng": 128284, + "Ġdừng": 138251, + "Ġdữ": 134897, + "Ġdá»±": 126258, + "Ġdá»±ng": 128855, + "Ġdá»ħ": 128926, + "Ġdá»ĭ": 127657, + "Ġdá»ĭch": 128507, + "Ġdá»ĭp": 137789, + "Ġe": 384, + "ĠeBay": 29906, + "ĠeBook": 57510, + "ĠeBooks": 94621, + "ĠeCommerce": 93559, + "ĠePub": 88617, + "ĠeSports": 88696, + "Ġea": 36546, + "Ġeach": 1817, + "Ġeag": 78240, + "Ġeager": 23541, + "Ġeagerly": 62373, + "Ġeagle": 59889, + "Ġeapply": 53959, + "Ġear": 2430, + "Ġearlier": 6788, + "Ġearliest": 29658, + "Ġearly": 4124, + "Ġearm": 95108, + "Ġearn": 7232, + "Ġearned": 15303, + "Ġearners": 95963, + "Ġearnest": 54249, + "Ġearning": 27644, + "Ġearnings": 23681, + "Ġearns": 63759, + "Ġearrings": 66201, + "Ġears": 24230, + "Ġearth": 9393, + "Ġearthly": 89395, + "Ġearthqu": 29750, + "Ġearthquake": 37313, + "Ġearthquakes": 65479, + "Ġeas": 2531, + "Ġease": 14219, + "Ġeased": 94882, + "Ġeasier": 8661, + "Ġeasiest": 29589, + "Ġeasily": 6707, + "Ġeasing": 44304, + "Ġeast": 10984, + "Ġeastern": 23149, + "Ġeasy": 4135, + "Ġeat": 8180, + "Ġeaten": 34561, + "Ġeater": 86035, + "Ġeating": 12182, + "Ġeats": 49677, + "Ġeauto": 50273, + "Ġeax": 36252, + "Ġeb": 24463, + "Ġebay": 59850, + "Ġeben": 50697, + "Ġebenfalls": 95524, + "Ġebony": 53096, + "Ġebook": 33997, + "Ġebooks": 87902, + "Ġebp": 87307, + "Ġebx": 79296, + "Ġec": 11942, + "Ġecc": 32914, + "Ġeccentric": 54820, + "Ġech": 30872, + "Ġechang": 88509, + "Ġecho": 1687, + "Ġechoed": 54112, + "Ġechoes": 69957, + "Ġechoing": 93724, + "Ġecht": 66793, + "Ġecl": 59958, + "Ġeclectic": 77713, + "Ġeclips": 92434, + "Ġeclipse": 55428, + "Ġeco": 41588, + "Ġecological": 49853, + "Ġecology": 71446, + "Ġecommerce": 84143, + "Ġecon": 3813, + "Ġeconom": 11467, + "Ġeconomic": 6955, + "Ġeconomical": 59518, + "Ġeconomically": 46279, + "Ġeconomics": 27889, + "Ġeconomies": 36571, + "Ġeconomist": 45115, + "Ġeconomists": 44315, + "Ġeconomy": 8584, + "ĠeconomÃŃa": 142036, + "Ġeconóm": 72541, + "Ġeconómica": 137628, + "Ġeconómico": 138045, + "Ġecosystem": 24982, + "Ġecosystems": 60851, + "Ġecs": 78422, + "Ġecstasy": 92563, + "Ġecstatic": 97292, + "Ġect": 76494, + "Ġecx": 65229, + "Ġed": 1578, + "Ġedad": 52682, + "ĠedeceÄŁi": 141598, + "Ġedge": 6821, + "Ġedged": 89563, + "Ġedges": 12822, + "Ġedi": 75092, + "Ġedible": 66640, + "Ġedición": 140449, + "Ġedit": 4499, + "ĠeditText": 64045, + "Ġeditable": 37409, + "Ġeditar": 79217, + "Ġedited": 19138, + "Ġediting": 15664, + "Ġedition": 13688, + "Ġeditions": 46534, + "Ġeditor": 6440, + "Ġeditorial": 27046, + "Ġeditors": 28746, + "Ġedits": 50844, + "Ġedição": 138226, + "Ġedm": 65340, + "Ġeds": 97099, + "Ġedt": 80714, + "Ġedu": 21443, + "Ġeduc": 4729, + "Ġeducación": 138963, + "Ġeducate": 38070, + "Ġeducated": 32042, + "Ġeducating": 66705, + "Ġeducation": 6731, + "Ġeducational": 16229, + "Ġeducator": 71046, + "Ġeducators": 49694, + "Ġeducação": 138894, + "Ġedx": 63645, + "Ġee": 36343, + "Ġeen": 8352, + "Ġeens": 98666, + "Ġeer": 43272, + "Ġeerie": 96283, + "Ġeerste": 67570, + "Ġef": 30369, + "Ġefect": 42946, + "Ġeff": 3289, + "Ġeffect": 2456, + "Ġeffected": 88489, + "Ġeffective": 7373, + "Ġeffectively": 13444, + "Ġeffectiveness": 26279, + "Ġeffects": 6239, + "Ġeffet": 66681, + "Ġeffic": 30814, + "Ġefficacy": 40165, + "Ġefficiencies": 91026, + "Ġefficiency": 15024, + "Ġefficient": 11050, + "Ġefficiently": 29720, + "Ġeffort": 5041, + "Ġeffortless": 79092, + "Ġeffortlessly": 68604, + "Ġefforts": 8869, + "Ġefter": 46480, + "Ġeg": 8695, + "Ġegal": 76171, + "Ġeget": 69663, + "Ġegg": 18636, + "Ġeggs": 18805, + "Ġego": 36274, + "Ġegreg": 82064, + "Ġegregious": 89068, + "Ġegret": 91192, + "Ġegt": 71067, + "Ġegy": 30708, + "Ġeh": 35246, + "Ġeher": 93395, + "Ġei": 29429, + "Ġeid": 68331, + "Ġeig": 34734, + "Ġeigen": 28724, + "Ġeigenen": 72705, + "Ġeigentlich": 83880, + "Ġeight": 8063, + "Ġeighteen": 62034, + "Ġeighth": 36377, + "Ġeighty": 79579, + "Ġein": 4368, + "Ġeine": 9820, + "Ġeinem": 17443, + "Ġeinen": 15462, + "Ġeiner": 17058, + "Ġeines": 36108, + "Ġeinf": 31287, + "Ġeinfach": 35323, + "Ġeing": 48376, + "Ġeinige": 63620, + "Ġeinmal": 54459, + "Ġeins": 88748, + "Ġeinz": 74227, + "Ġeinzel": 94788, + "Ġeither": 2987, + "Ġeius": 68421, + "Ġeiusmod": 79122, + "Ġej": 19279, + "Ġejac": 63614, + "Ġejaculation": 98202, + "Ġejec": 49171, + "Ġeject": 89474, + "Ġejected": 94403, + "Ġejemplo": 57200, + "Ġejercicio": 95921, + "Ġek": 26857, + "Ġeks": 89662, + "Ġel": 655, + "Ġela": 50307, + "Ġelabor": 24940, + "Ġelaborate": 35967, + "Ġelapsed": 25333, + "ĠelapsedTime": 80855, + "Ġelast": 91085, + "Ġelastic": 35473, + "Ġelasticity": 94816, + "Ġelbow": 45711, + "Ġelbows": 80912, + "Ġelder": 22130, + "Ġelderly": 28820, + "Ġelders": 60676, + "Ġeldest": 73565, + "Ġeldre": 93360, + "Ġele": 10510, + "Ġelect": 4045, + "Ġelected": 16290, + "Ġelection": 6223, + "Ġelections": 15991, + "Ġelective": 97490, + "Ġelectoral": 33841, + "Ġelectorate": 61468, + "Ġelectr": 42806, + "Ġelectric": 9072, + "Ġelectrical": 19734, + "Ġelectricity": 17728, + "Ġelectro": 24394, + "Ġelectrode": 70948, + "Ġelectrodes": 79290, + "Ġelectroly": 72296, + "Ġelectrom": 53607, + "Ġelectromagnetic": 65569, + "Ġelectron": 16715, + "Ġelectronic": 14346, + "Ġelectronically": 70770, + "Ġelectronics": 30491, + "Ġelectrons": 56578, + "Ġelectrónico": 71466, + "Ġeleg": 17720, + "Ġelegance": 63443, + "Ġelegant": 25777, + "Ġelek": 62788, + "Ġelem": 11750, + "Ġelement": 2392, + "ĠelementType": 89573, + "Ġelemental": 61135, + "Ġelementary": 35156, + "Ġelemento": 40618, + "Ġelementos": 48147, + "Ġelements": 5424, + "Ġelems": 55968, + "Ġelephant": 45740, + "Ġelephants": 60766, + "Ġeles": 66441, + "Ġelev": 11964, + "Ġelevate": 68630, + "Ġelevated": 31289, + "Ġelevation": 26163, + "Ġelevator": 38636, + "Ġeleven": 44214, + "Ġelf": 40745, + "Ġelgg": 95653, + "Ġelic": 94260, + "Ġelif": 4409, + "Ġelig": 14176, + "Ġeligibility": 40634, + "Ġeligible": 17013, + "Ġelim": 9762, + "Ġelimin": 37316, + "Ġeliminar": 48404, + "Ġeliminate": 21725, + "Ġeliminated": 33273, + "Ġeliminates": 59844, + "Ġeliminating": 39499, + "Ġelimination": 42932, + "Ġelit": 30060, + "Ġelite": 20685, + "Ġelites": 55703, + "Ġelk": 76842, + "Ġelkaar": 94598, + "Ġell": 25607, + "Ġella": 53501, + "Ġellas": 97660, + "Ġelle": 26451, + "Ġeller": 25939, + "Ġelles": 83468, + "Ġellipse": 57397, + "Ġellipt": 77783, + "Ġello": 66315, + "Ġellos": 53330, + "Ġelm": 42205, + "Ġelo": 63135, + "Ġelong": 73495, + "Ġels": 20712, + "Ġelse": 770, + "Ġelseif": 11755, + "Ġelsewhere": 17920, + "Ġelsif": 24482, + "Ġelt": 54139, + "Ġelucid": 96198, + "Ġelusive": 65584, + "Ġelves": 76792, + "Ġelé": 85606, + "ĠelÅij": 88811, + "Ġem": 976, + "Ġemacs": 86683, + "Ġemail": 2551, + "ĠemailAddress": 89713, + "Ġemailed": 48514, + "Ġemailing": 91701, + "Ġemails": 14298, + "Ġeman": 68139, + "Ġemanc": 90125, + "Ġemb": 7967, + "Ġembar": 79601, + "Ġembargo": 47602, + "Ġembark": 69431, + "Ġembarked": 78020, + "Ġembarrass": 22281, + "Ġembarrassed": 48130, + "Ġembarrassing": 44005, + "Ġembarrassment": 58930, + "Ġembassy": 45467, + "Ġembed": 11579, + "Ġembedded": 22864, + "Ġembedding": 39088, + "Ġembeddings": 70547, + "Ġembell": 71414, + "Ġember": 98284, + "Ġemblem": 66274, + "Ġembod": 31023, + "Ġembodied": 77543, + "Ġembodies": 94022, + "Ġembodiment": 48397, + "Ġembodiments": 74553, + "Ġembody": 80558, + "Ġembr": 16618, + "Ġembrace": 26732, + "Ġembraced": 42503, + "Ġembraces": 82387, + "Ġembracing": 55401, + "Ġembroid": 53578, + "Ġembroidered": 92659, + "Ġembroidery": 84827, + "Ġembry": 43381, + "Ġembryo": 86601, + "Ġembryos": 88773, + "Ġemerg": 7277, + "Ġemerge": 32944, + "Ġemerged": 21992, + "Ġemergence": 48079, + "Ġemergencies": 67974, + "Ġemergency": 12851, + "Ġemerges": 58596, + "Ġemerging": 23206, + "Ġeminent": 83529, + "Ġemiss": 98663, + "Ġemission": 40253, + "Ġemissions": 20143, + "Ġemit": 16691, + "Ġemits": 72780, + "Ġemitted": 46942, + "Ġemitter": 47263, + "Ġemitting": 89456, + "Ġemlrt": 96212, + "Ġemo": 93097, + "Ġemoc": 84907, + "Ġemoji": 42365, + "Ġemojis": 99066, + "Ġemot": 10172, + "Ġemotion": 19772, + "Ġemotional": 14269, + "Ġemotionally": 37583, + "Ġemotions": 21261, + "Ġemp": 8486, + "Ġempath": 35581, + "Ġempathy": 47351, + "Ġemperor": 56783, + "Ġempez": 81849, + "Ġemph": 13155, + "Ġemphas": 20053, + "Ġemphasis": 24654, + "Ġemphasize": 45932, + "Ġemphasized": 45628, + "Ġemphasizes": 65059, + "Ġemphasizing": 80903, + "Ġempir": 37869, + "Ġempire": 31347, + "Ġempirical": 45663, + "Ġemple": 61026, + "Ġempleado": 86940, + "Ġemploi": 91700, + "Ġemploy": 3455, + "Ġemployed": 19446, + "Ġemployee": 9364, + "Ġemployees": 8256, + "Ġemployer": 19136, + "Ġemployers": 22426, + "Ġemploying": 50197, + "Ġemployment": 14402, + "Ġemploys": 50142, + "Ġempower": 25219, + "Ġempowered": 61835, + "Ġempowering": 65288, + "Ġempowerment": 73377, + "Ġempres": 89853, + "Ġempresa": 32764, + "Ġempresas": 50304, + "Ġempt": 53069, + "Ġemptied": 93872, + "Ġempty": 4287, + "Ġemulate": 65485, + "Ġemulation": 78841, + "Ġemulator": 58516, + "Ġen": 662, + "Ġenable": 7283, + "Ġenabled": 8970, + "Ġenables": 20081, + "Ġenabling": 27362, + "Ġenact": 46947, + "Ġenacted": 43877, + "Ġenactment": 77552, + "Ġenam": 81852, + "Ġenamel": 91748, + "Ġenc": 3209, + "Ġencaps": 42569, + "Ġench": 65712, + "Ġenchant": 46925, + "Ġenchanted": 97311, + "Ġenclave": 97712, + "Ġenclosed": 43810, + "Ġenclosing": 84928, + "Ġenclosure": 57812, + "Ġencode": 16164, + "ĠencodeURIComponent": 56737, + "Ġencoded": 20498, + "Ġencoder": 23668, + "Ġencoding": 11170, + "Ġencompass": 37532, + "Ġencompasses": 69910, + "Ġencontr": 22547, + "Ġencontrado": 55227, + "Ġencontrar": 45623, + "Ġencore": 34497, + "Ġencount": 29797, + "Ġencounter": 12828, + "Ġencountered": 23057, + "Ġencountering": 91272, + "Ġencounters": 33906, + "Ġencour": 7878, + "Ġencourage": 14907, + "Ġencouraged": 20548, + "Ġencouragement": 50375, + "Ġencourages": 36067, + "Ġencouraging": 25836, + "Ġencrypt": 29625, + "Ġencrypted": 24455, + "Ġencryption": 25463, + "Ġenctype": 37415, + "Ġencuent": 35737, + "Ġencuentra": 56168, + "Ġencyclopedia": 82608, + "Ġend": 835, + "ĠendDate": 30445, + "ĠendIndex": 61103, + "ĠendPoint": 69474, + "ĠendTime": 32061, + "Ġendanger": 36937, + "Ġendangered": 51256, + "Ġende": 28605, + "Ġendeavor": 53814, + "Ġendeavors": 78545, + "Ġendeavour": 88931, + "Ġended": 9482, + "Ġendemic": 94913, + "Ġendereco": 89153, + "Ġendforeach": 31154, + "Ġendian": 64688, + "Ġendif": 12330, + "Ġending": 13391, + "Ġendings": 76597, + "Ġendl": 5229, + "Ġendless": 25678, + "Ġendlessly": 78737, + "Ġendors": 33965, + "Ġendorse": 18970, + "Ġendorsed": 39628, + "Ġendorsement": 41043, + "Ġendorsements": 93894, + "Ġendorsing": 97759, + "Ġendoth": 92229, + "Ġendowed": 97628, + "Ġendpoint": 14887, + "Ġendpoints": 36342, + "Ġendregion": 85017, + "Ġends": 10335, + "Ġendurance": 48186, + "Ġendure": 45653, + "Ġendured": 55124, + "Ġenduring": 51776, + "Ġendwhile": 65914, + "Ġenemies": 13883, + "Ġenemy": 9175, + "Ġener": 37456, + "Ġenerg": 4501, + "Ġenergetic": 44855, + "Ġenergia": 95058, + "Ġenergies": 48869, + "Ġenergy": 4802, + "ĠenergÃŃa": 137064, + "Ġenf": 32532, + "Ġenfants": 59304, + "Ġenfer": 60997, + "Ġenfermed": 87017, + "Ġenforce": 28162, + "Ġenforced": 44321, + "Ġenforcement": 13324, + "Ġenforcing": 61808, + "Ġenfrent": 88499, + "Ġeng": 2922, + "Ġengage": 16579, + "Ġengaged": 16634, + "Ġengagement": 19805, + "Ġengagements": 78871, + "Ġengages": 68489, + "Ġengaging": 22570, + "Ġengine": 4712, + "Ġengineer": 23576, + "Ġengineered": 44936, + "Ġengineering": 14667, + "Ġengineers": 24198, + "Ġengines": 21106, + "Ġenglish": 28963, + "Ġengr": 50055, + "Ġengraved": 79710, + "Ġengulf": 76736, + "Ġenh": 8663, + "Ġenhance": 18379, + "Ġenhanced": 23922, + "Ġenhancement": 26788, + "Ġenhancements": 58529, + "Ġenhances": 56824, + "Ġenhancing": 46494, + "Ġenim": 59502, + "Ġenjo": 30579, + "Ġenjoy": 4669, + "Ġenjoyable": 31080, + "Ġenjoyed": 14006, + "Ġenjoying": 21413, + "Ġenjoyment": 44178, + "Ġenjoys": 31738, + "Ġenlarg": 39939, + "Ġenlarge": 52644, + "Ġenlarged": 73017, + "Ġenlargement": 65407, + "Ġenlight": 40216, + "Ġenlightened": 81215, + "Ġenlightenment": 80769, + "Ġenlist": 87624, + "Ġenlisted": 68476, + "Ġenn": 60524, + "Ġenorm": 17299, + "Ġenorme": 67806, + "Ġenormous": 22399, + "Ġenormously": 83463, + "Ġenough": 3322, + "Ġenqu": 76707, + "Ġenquanto": 94641, + "Ġenqueue": 53757, + "Ġenquiries": 83322, + "Ġenquiry": 76512, + "Ġenr": 80552, + "Ġenraged": 97195, + "Ġenrich": 30418, + "Ġenriched": 68571, + "Ġenrichment": 69172, + "Ġenrol": 65366, + "Ġenroll": 51780, + "Ġenrolled": 36091, + "Ġenrollment": 38048, + "Ġens": 4572, + "Ġense": 66961, + "Ġensemble": 39026, + "Ġensl": 66448, + "Ġenslaved": 94289, + "Ġensued": 95663, + "Ġensuing": 71658, + "Ġensuite": 75726, + "Ġensure": 5978, + "Ġensured": 58654, + "Ġensures": 25351, + "Ġensuring": 22573, + "Ġent": 1197, + "Ġentail": 85992, + "Ġentails": 70104, + "Ġentend": 68279, + "Ġentender": 95437, + "Ġenter": 3725, + "Ġentered": 10636, + "Ġentering": 16262, + "Ġenterprise": 20179, + "Ġenterprises": 39819, + "Ġenters": 28833, + "Ġentert": 22887, + "Ġentertain": 45176, + "Ġentertained": 64450, + "Ġentertaining": 29211, + "Ġentertainment": 16517, + "Ġentfer": 82195, + "Ġenth": 64208, + "Ġenthus": 16277, + "Ġenthusi": 21142, + "Ġenthusiasm": 35132, + "Ġenthusiast": 60812, + "Ġenthusiastic": 41602, + "Ġenthusiastically": 98225, + "Ġenthusiasts": 42348, + "Ġenticing": 87507, + "Ġentidad": 85808, + "Ġentire": 4453, + "Ġentirely": 11368, + "Ġentirety": 47917, + "Ġentities": 14744, + "Ġentitled": 19867, + "Ġentitlement": 53901, + "Ġentity": 5387, + "ĠentityId": 93042, + "ĠentityManager": 59775, + "ĠentityType": 96801, + "Ġentonces": 71399, + "Ġentr": 11492, + "Ġentra": 82686, + "Ġentrada": 42563, + "Ġentrance": 19809, + "Ġentrances": 94579, + "Ġentrar": 82313, + "Ġentre": 9281, + "Ġentreg": 93281, + "Ġentrega": 78418, + "Ġentren": 93821, + "Ġentrenched": 81044, + "Ġentreprene": 19414, + "Ġentrepreneur": 28249, + "Ġentrepreneurial": 69306, + "Ġentrepreneurs": 34390, + "Ġentrepreneurship": 74405, + "Ġentreprise": 92386, + "Ġentreprises": 96163, + "Ġentrev": 89032, + "Ġentries": 10695, + "Ġentropy": 47502, + "Ġentrusted": 85282, + "Ġentry": 4343, + "Ġents": 36852, + "Ġentsprech": 87488, + "Ġentwick": 80129, + "Ġentão": 84813, + "Ġenum": 7618, + "Ġenumer": 10153, + "Ġenumerable": 43942, + "Ġenumerate": 13252, + "Ġenumerated": 84869, + "Ġenumeration": 51578, + "Ġenumerator": 74631, + "Ġenums": 70717, + "Ġenv": 6105, + "Ġenvelop": 53185, + "Ġenvelope": 34398, + "Ġenvelopes": 86606, + "Ġenvi": 49264, + "Ġenviado": 98572, + "Ġenviar": 62778, + "Ġenviron": 48926, + "Ġenvironment": 4573, + "Ġenvironmental": 12158, + "Ġenvironmentally": 56240, + "Ġenvironments": 21737, + "Ġenvis": 84675, + "Ġenvision": 47046, + "Ġenvisioned": 73289, + "Ġenvoy": 59530, + "Ġenvy": 65433, + "Ġenzym": 30911, + "Ġenzyme": 48142, + "Ġenzymes": 54967, + "Ġeo": 94603, + "Ġeof": 76760, + "Ġeos": 61794, + "Ġep": 4155, + "Ġephem": 82134, + "Ġepic": 24679, + "Ġepid": 27717, + "Ġepidemi": 60957, + "Ġepidemic": 41520, + "Ġepile": 57552, + "Ġepilepsy": 69555, + "Ġepis": 66494, + "Ġepisode": 9234, + "Ġepisodes": 17770, + "Ġepit": 66322, + "Ġepith": 63679, + "Ġepoch": 16342, + "Ġepochs": 39346, + "Ġepoll": 95861, + "Ġepoxy": 96644, + "Ġeps": 17986, + "Ġepsilon": 31204, + "Ġepub": 63949, + "Ġeq": 8939, + "Ġeql": 67376, + "Ġequ": 3232, + "Ġequal": 6144, + "ĠequalTo": 60545, + "Ġequality": 21777, + "Ġequally": 18308, + "Ġequals": 16819, + "Ġequation": 23606, + "Ġequations": 37906, + "Ġequilibrium": 55311, + "Ġequip": 25386, + "Ġequipe": 96205, + "Ġequipment": 7096, + "Ġequipments": 71130, + "Ġequipo": 47912, + "Ġequipos": 91852, + "Ġequipped": 18650, + "Ġequitable": 76009, + "Ġequity": 24448, + "Ġequiv": 68458, + "Ġequival": 43454, + "Ġequivalence": 84162, + "Ġequivalent": 13578, + "Ġequivalents": 85006, + "Ġer": 2714, + "Ġera": 11385, + "Ġerad": 55437, + "Ġeradicate": 88414, + "Ġerase": 35988, + "Ġerased": 63700, + "Ġerb": 67648, + "Ġere": 38257, + "Ġerect": 37632, + "Ġerected": 65806, + "Ġerectile": 56117, + "Ġerection": 64046, + "Ġerf": 36408, + "Ġerfahren": 95402, + "Ġerfol": 53211, + "Ġerfolgre": 73796, + "Ġerfolgreich": 93094, + "Ġerg": 35276, + "Ġergonomic": 96691, + "Ġerh": 59801, + "Ġerhalten": 57984, + "Ġerk": 88766, + "Ġerklä": 86107, + "Ġerklärt": 138936, + "Ġerle": 76120, + "Ġerm": 76134, + "Ġermög": 97035, + "Ġern": 91929, + "Ġero": 71232, + "Ġeros": 42119, + "Ġerosion": 56997, + "Ġerot": 8054, + "Ġerotic": 33316, + "Ġerotica": 88389, + "Ġerotici": 39472, + "Ġerotico": 73138, + "Ġerotik": 27562, + "Ġerotique": 52827, + "Ġerotisch": 96563, + "Ġerotische": 22210, + "Ġerotisk": 38988, + "Ġerotiske": 65653, + "Ġerr": 1848, + "ĠerrMsg": 60078, + "Ġerratic": 94575, + "Ġerre": 55449, + "Ġerreur": 90920, + "Ġerrmsg": 67045, + "Ġerrno": 26955, + "Ġerro": 36310, + "Ġerrone": 57789, + "Ġerroneous": 76672, + "Ġerror": 1465, + "ĠerrorCallback": 89080, + "ĠerrorCode": 40442, + "ĠerrorHandler": 71073, + "ĠerrorMessage": 24943, + "ĠerrorMsg": 52247, + "ĠerrorThrown": 81699, + "Ġerrores": 90700, + "Ġerrors": 5975, + "Ġerrs": 70817, + "Ġers": 54395, + "Ġerst": 34368, + "Ġerste": 61217, + "Ġersten": 48158, + "Ġeru": 60254, + "Ġerupt": 43679, + "Ġerupted": 60174, + "Ġeruption": 79402, + "Ġerv": 71075, + "Ġerw": 66386, + "Ġery": 88257, + "Ġerót": 79443, + "Ġes": 1531, + "Ġesa": 49880, + "Ġesac": 81844, + "Ġesc": 3835, + "Ġescal": 26435, + "Ġescalate": 88590, + "Ġescalated": 80600, + "Ġescalating": 76057, + "Ġescalation": 70289, + "Ġescap": 88784, + "Ġescape": 12449, + "Ġescaped": 27783, + "Ġescapes": 65109, + "Ġescaping": 52654, + "Ġescol": 39654, + "Ġescort": 6155, + "Ġescorte": 50070, + "Ġescorted": 76556, + "Ġescorts": 26932, + "Ġescre": 96945, + "Ġescri": 65771, + "Ġescrit": 57444, + "Ġese": 41818, + "Ġesi": 77308, + "Ġesk": 23978, + "Ġeskort": 40541, + "Ġeskorte": 38167, + "Ġesl": 97846, + "Ġeslint": 21155, + "Ġeso": 43741, + "Ġesos": 93074, + "Ġesp": 16541, + "Ġespa": 35611, + "Ġespacio": 65484, + "Ġespan": 79555, + "Ġespaço": 98872, + "Ġespañ": 49375, + "Ġespañol": 69888, + "Ġespec": 42838, + "Ġespecial": 32297, + "Ġespecially": 5310, + "Ġespecialmente": 94002, + "Ġespecific": 68594, + "Ġespect": 94270, + "ĠespecÃŃf": 77920, + "ĠespecÃŃfica": 141033, + "ĠespecÃŃfico": 140947, + "Ġesper": 30057, + "Ġespera": 81744, + "Ġespionage": 77639, + "Ġesports": 80450, + "Ġespos": 70227, + "Ġesposa": 84890, + "Ġespresso": 65421, + "Ġess": 3956, + "Ġessa": 54629, + "Ġessay": 8895, + "Ġessays": 22844, + "Ġesse": 41379, + "Ġessen": 94655, + "Ġessence": 27491, + "Ġessential": 7565, + "Ġessentially": 15791, + "Ġessentials": 58786, + "Ġessere": 34266, + "Ġest": 1788, + "Ġesta": 15140, + "Ġestaba": 63224, + "Ġestable": 51664, + "Ġestablish": 5695, + "Ġestablished": 9555, + "Ġestablishes": 63564, + "Ġestablishing": 30592, + "Ġestablishment": 21269, + "Ġestablishments": 62237, + "Ġestad": 77681, + "Ġestado": 24062, + "Ġestados": 93524, + "Ġestamos": 67060, + "Ġestar": 34257, + "Ġestará": 136185, + "Ġestas": 47491, + "Ġestate": 12394, + "Ġestates": 82975, + "Ġestava": 77800, + "Ġeste": 10351, + "Ġesteem": 84497, + "Ġesteemed": 96464, + "Ġestilo": 69283, + "Ġestim": 14932, + "Ġestimate": 16045, + "Ġestimated": 12943, + "Ġestimates": 17530, + "Ġestimating": 76372, + "Ġestimation": 41204, + "Ġestimator": 67298, + "Ġesto": 37793, + "Ġestos": 44786, + "Ġestoy": 81284, + "Ġestr": 29507, + "Ġestratég": 140791, + "Ġestratégia": 140792, + "Ġestrogen": 69080, + "Ġestruct": 75409, + "Ġestud": 57463, + "Ġestudiantes": 97722, + "Ġestudio": 78936, + "Ġestá": 15466, + "Ġestán": 41839, + "Ġestão": 56454, + "Ġet": 1842, + "Ġeta": 38997, + "Ġetc": 4992, + "Ġeternal": 34725, + "Ġeternity": 63382, + "Ġeth": 8372, + "Ġethanol": 61672, + "Ġether": 50050, + "Ġethereum": 84522, + "Ġethernet": 80534, + "Ġethers": 95120, + "Ġethic": 64847, + "Ġethical": 30208, + "Ġethics": 30908, + "Ġethn": 78080, + "Ġethnic": 21551, + "Ġethnicity": 56878, + "Ġethos": 84239, + "Ġetiqu": 60944, + "Ġetiquette": 93205, + "Ġetree": 66655, + "Ġett": 26724, + "Ġetter": 55816, + "ĠettiÄŁi": 131192, + "Ġettä": 72351, + "Ġetwa": 77556, + "Ġetwas": 39771, + "Ġeu": 15555, + "Ġeuch": 65295, + "Ġeuler": 90970, + "Ġeup": 63943, + "Ġeuro": 17672, + "Ġeurop": 61921, + "Ġeurope": 37534, + "Ġeuropean": 85919, + "Ġeuropé": 94099, + "Ġeuropéen": 143446, + "Ġeuropéenne": 140927, + "Ġeuros": 32488, + "Ġeuth": 80580, + "Ġeux": 83340, + "Ġev": 3637, + "Ġevac": 29420, + "Ġevacuate": 88989, + "Ġevacuated": 62258, + "Ġevacuation": 59673, + "Ġevade": 76653, + "Ġeval": 5603, + "Ġevalu": 19326, + "Ġevaluate": 15442, + "Ġevaluated": 25070, + "Ġevaluates": 66249, + "Ġevaluating": 37563, + "Ġevaluation": 16460, + "Ġevaluations": 55081, + "Ġevaluator": 69810, + "Ġevangel": 38053, + "Ġevangelical": 66585, + "Ġevapor": 59050, + "Ġevasion": 76732, + "Ġeve": 48896, + "Ġeven": 1496, + "Ġevening": 11458, + "Ġevenings": 58838, + "Ġevenly": 41047, + "Ġevent": 1538, + "ĠeventData": 67867, + "ĠeventId": 70000, + "ĠeventName": 60177, + "ĠeventType": 58551, + "Ġeventdata": 44008, + "Ġevento": 44442, + "Ġeventos": 72394, + "Ġevents": 4357, + "Ġeventual": 41735, + "Ġeventually": 9583, + "Ġever": 3512, + "Ġeverlasting": 94949, + "Ġevery": 1449, + "Ġeverybody": 16083, + "Ġeveryday": 17778, + "Ġeveryone": 5019, + "Ġeverything": 4297, + "Ġeverytime": 89356, + "Ġeverywhere": 16852, + "Ġeviction": 78236, + "Ġevid": 5226, + "Ġevidence": 5904, + "Ġevidenced": 68357, + "Ġevident": 29476, + "Ġevidently": 66070, + "Ġevil": 13962, + "Ġevils": 85807, + "Ġevitar": 70808, + "Ġevoke": 94163, + "Ġevolution": 15379, + "Ġevolutionary": 40893, + "Ġevolve": 37580, + "Ġevolved": 27895, + "Ġevolves": 91954, + "Ġevolving": 40928, + "Ġevt": 12358, + "Ġew": 36890, + "Ġex": 505, + "Ġexacerb": 51775, + "Ġexacerbated": 91441, + "Ġexact": 4734, + "Ġexactly": 6896, + "Ġexagger": 37930, + "Ġexaggerated": 61158, + "Ġexam": 7006, + "Ġexamination": 23568, + "Ġexaminations": 68470, + "Ġexamine": 20961, + "Ġexamined": 24109, + "Ġexaminer": 83632, + "Ġexamines": 47995, + "Ġexamining": 37836, + "Ġexample": 3110, + "Ġexamples": 10295, + "Ġexams": 39686, + "Ġexc": 3438, + "Ġexcav": 44115, + "Ġexcavation": 96439, + "Ġexce": 90284, + "Ġexceed": 12488, + "Ġexceeded": 33808, + "Ġexceeding": 47905, + "Ġexceedingly": 75721, + "Ġexceeds": 35275, + "Ġexcel": 24538, + "Ġexcelente": 95407, + "Ġexcell": 8250, + "Ġexcellence": 37556, + "Ġexcellent": 9073, + "Ġexcept": 3650, + "Ġexception": 4683, + "Ġexceptional": 24364, + "Ġexceptionally": 47198, + "Ġexceptions": 19585, + "Ġexcer": 39077, + "Ġexcerpt": 49465, + "Ġexcerpts": 83390, + "Ġexcess": 13623, + "Ġexcessive": 26541, + "Ġexcessively": 86046, + "Ġexchange": 9289, + "Ġexchanged": 47401, + "Ġexchanges": 28998, + "Ġexchanging": 78431, + "Ġexcit": 24431, + "Ġexcited": 12035, + "Ġexcitement": 27262, + "Ġexciting": 13245, + "Ġexcl": 80284, + "Ġexclaimed": 79508, + "Ġexclude": 21687, + "Ġexcluded": 27444, + "Ġexcludes": 63368, + "Ġexcluding": 43778, + "Ġexclus": 18034, + "Ġexclusion": 41208, + "Ġexclusive": 13761, + "Ġexclusively": 23242, + "Ġexcursion": 94340, + "Ġexcuse": 27291, + "Ġexcuses": 54486, + "Ġexe": 47193, + "Ġexec": 3883, + "Ġexecut": 23494, + "Ġexecutable": 32156, + "Ġexecute": 9026, + "Ġexecuted": 15695, + "Ġexecutes": 51435, + "Ġexecuting": 30220, + "Ġexecution": 11320, + "Ġexecutions": 68146, + "Ġexecutive": 10905, + "Ġexecutives": 28409, + "Ġexecutor": 31558, + "Ġexem": 15377, + "Ġexemp": 93220, + "Ġexempl": 37939, + "Ġexemplary": 76281, + "Ġexemple": 50073, + "Ġexemplo": 79594, + "Ġexempt": 37928, + "Ġexemption": 44698, + "Ġexemptions": 70888, + "Ġexerc": 7432, + "Ġexercise": 10158, + "Ġexercised": 61013, + "Ġexercises": 22932, + "Ġexercising": 50482, + "Ġexercitation": 98449, + "ĠexercÃŃcio": 141448, + "Ġexert": 42744, + "Ġexh": 56659, + "Ġexhaust": 17502, + "Ġexhausted": 37919, + "Ġexhausting": 78853, + "Ġexhaustion": 69563, + "Ġexhaustive": 72503, + "Ġexhib": 12253, + "Ġexhibit": 30224, + "Ġexhibited": 50613, + "Ġexhibiting": 86619, + "Ġexhibition": 27001, + "Ġexhibitions": 69764, + "Ġexhibits": 49729, + "Ġexhilar": 90726, + "Ġexig": 73672, + "Ġexile": 59987, + "Ġexist": 3000, + "Ġexiste": 28153, + "Ġexisted": 24295, + "Ġexistence": 13885, + "Ġexistential": 66639, + "Ġexisting": 6350, + "Ġexists": 6724, + "Ġexit": 4869, + "Ġexited": 51283, + "Ġexiting": 44748, + "Ġexits": 42086, + "Ġexon": 98744, + "Ġexotic": 38318, + "Ġexp": 1343, + "Ġexpand": 9225, + "Ġexpanded": 17183, + "Ġexpanding": 23175, + "Ġexpands": 51856, + "Ġexpans": 38914, + "Ġexpansion": 14461, + "Ġexpansions": 77488, + "Ġexpansive": 60738, + "Ġexpect": 1720, + "Ġexpectancy": 65895, + "Ġexpectation": 30193, + "Ġexpectations": 16665, + "Ġexpected": 3601, + "ĠexpectedResult": 81441, + "Ġexpecting": 22331, + "Ġexpects": 24297, + "Ġexped": 30627, + "Ġexpedition": 50164, + "Ġexpelled": 66231, + "Ġexpend": 27310, + "Ġexpended": 98031, + "Ġexpenditure": 44326, + "Ġexpenditures": 52769, + "Ġexpense": 20284, + "Ġexpenses": 18024, + "Ġexpensive": 11392, + "Ġexper": 2134, + "Ġexperi": 5747, + "Ġexperience": 3139, + "Ġexperienced": 10321, + "Ġexperiences": 11449, + "Ġexperiencia": 53313, + "Ġexperiencing": 24084, + "Ġexperiment": 9342, + "Ġexperimental": 22000, + "Ġexperimentation": 65096, + "Ġexperimented": 97404, + "Ġexperimenting": 59157, + "Ġexperiments": 21204, + "Ġexperiência": 133780, + "Ġexpert": 6203, + "Ġexpertise": 18726, + "Ġexperts": 11647, + "Ġexpiration": 31692, + "Ġexpire": 32563, + "Ġexpired": 26391, + "Ġexpires": 27902, + "ĠexpiresIn": 98400, + "Ġexpiry": 49921, + "Ġexpl": 3247, + "Ġexplain": 10339, + "Ġexplained": 11247, + "Ġexplaining": 25021, + "Ġexplains": 14758, + "Ġexplan": 13357, + "Ġexplanation": 16148, + "Ġexplanations": 40841, + "Ġexplanatory": 92466, + "Ġexplic": 55269, + "Ġexplicit": 11464, + "Ġexplicitly": 20975, + "Ġexplo": 7532, + "Ġexplode": 15758, + "Ġexploded": 43813, + "Ġexplodes": 93913, + "Ġexploding": 72645, + "Ġexploit": 32194, + "Ġexploitation": 39661, + "Ġexploited": 50663, + "Ġexploiting": 70601, + "Ġexploits": 62388, + "Ġexplor": 47439, + "Ġexploration": 26403, + "Ġexplore": 13186, + "Ġexplored": 35031, + "Ġexplorer": 52967, + "Ġexplores": 40324, + "Ġexploring": 23966, + "Ġexplos": 13894, + "Ġexplosion": 24199, + "Ġexplosions": 55406, + "Ġexplosive": 33050, + "Ġexplosives": 56309, + "Ġexpo": 92316, + "Ġexponent": 27690, + "Ġexponential": 58755, + "Ġexponentially": 74151, + "Ġexport": 7485, + "Ġexported": 34890, + "Ġexporter": 57378, + "Ġexporters": 74848, + "Ġexporting": 60221, + "Ġexports": 12794, + "Ġexpos": 45856, + "Ġexpose": 28141, + "Ġexposed": 14900, + "Ġexposes": 58281, + "Ġexposing": 45966, + "Ġexposition": 90418, + "Ġexposure": 14338, + "Ġexposures": 69430, + "Ġexpr": 15169, + "Ġexpres": 66505, + "Ġexpress": 3158, + "Ġexpressed": 13302, + "Ġexpresses": 60020, + "Ġexpressing": 36710, + "Ġexpression": 7493, + "Ġexpressions": 23393, + "Ġexpressive": 77123, + "Ġexpressly": 52511, + "Ġexpulsion": 94889, + "Ġexpérience": 87311, + "Ġexquisite": 58608, + "Ġext": 1303, + "Ġextend": 13036, + "Ġextended": 11577, + "Ġextending": 32359, + "Ġextends": 2239, + "Ġextension": 8894, + "Ġextensions": 19721, + "Ġextensive": 16376, + "Ġextensively": 41717, + "Ġextent": 12818, + "Ġextents": 91489, + "Ġexter": 54267, + "Ġexterior": 27263, + "Ġextern": 15637, + "Ġexternal": 9250, + "Ġexternally": 68342, + "Ġextinct": 68818, + "Ġextinction": 51509, + "Ġextingu": 55707, + "Ġextortion": 95380, + "Ġextr": 11482, + "Ġextra": 4960, + "Ġextract": 8649, + "Ġextracted": 27432, + "Ġextracting": 59408, + "Ġextraction": 32189, + "Ġextractor": 67472, + "Ġextracts": 47962, + "Ġextrad": 66802, + "Ġextradition": 93976, + "Ġextraordin": 63455, + "Ġextraordinarily": 74924, + "Ġextraordinary": 23742, + "Ġextrapol": 70362, + "Ġextras": 36275, + "Ġextrav": 57339, + "Ġextravag": 73064, + "Ġextravagant": 94850, + "Ġextrem": 5876, + "Ġextreme": 14226, + "Ġextremely": 9016, + "Ġextremes": 68871, + "Ġextremism": 69668, + "Ġextremist": 55782, + "Ġextremists": 57907, + "Ġey": 3912, + "Ġeye": 7912, + "Ġeyeb": 35961, + "Ġeyebrow": 83316, + "Ġeyebrows": 60125, + "Ġeyed": 90672, + "Ġeyel": 53857, + "Ġeyes": 6414, + "Ġeyewitness": 88910, + "Ġez": 43320, + "ĠeÄŁ": 126689, + "ĠeÄŁer": 141166, + "ĠeÄŁitim": 129765, + "ĠeÄŁlen": 140298, + "ĠeÅŁ": 129944, + "Ġf": 282, + "ĠfChain": 95183, + "ĠfName": 92409, + "Ġfa": 2218, + "Ġfab": 9570, + "Ġfabric": 13055, + "Ġfabricated": 69454, + "Ġfabrication": 58151, + "Ġfabrics": 51954, + "Ġfabs": 40728, + "Ġfabulous": 34531, + "Ġfac": 3463, + "Ġfacade": 61616, + "Ġface": 3579, + "Ġfacebook": 22943, + "Ġfaced": 16601, + "Ġfaces": 12300, + "Ġfacet": 44507, + "Ġfacets": 61714, + "Ġfacial": 27800, + "Ġfacil": 57324, + "Ġfacile": 50694, + "Ġfacilit": 16617, + "Ġfacilitate": 27596, + "Ġfacilitated": 71749, + "Ġfacilitates": 72533, + "Ġfacilitating": 67265, + "Ġfacilities": 12786, + "Ġfacility": 12481, + "Ġfacing": 12880, + "Ġfact": 2097, + "Ġfaction": 36380, + "Ġfactions": 47652, + "Ġfacto": 60496, + "Ġfactor": 8168, + "Ġfactorial": 52962, + "Ġfactories": 34059, + "Ġfactors": 9363, + "Ġfactory": 8633, + "Ġfacts": 13064, + "Ġfactual": 59901, + "Ġfactura": 99045, + "Ġfacult": 72433, + "Ġfaculties": 96983, + "Ġfaculty": 21564, + "Ġfade": 15016, + "ĠfadeIn": 46510, + "Ġfaded": 53334, + "Ġfades": 86409, + "Ġfading": 58517, + "Ġfail": 3690, + "Ġfailed": 4641, + "Ġfailing": 21394, + "Ġfails": 14525, + "Ġfailure": 7901, + "Ġfailures": 27850, + "Ġfaint": 37578, + "Ġfair": 6624, + "Ġfaire": 19463, + "Ġfairly": 14138, + "Ġfairness": 50741, + "Ġfairy": 44486, + "Ġfais": 65417, + "Ġfait": 19694, + "Ġfaith": 9881, + "Ġfaithful": 36413, + "Ġfaithfully": 93076, + "Ġfake": 12418, + "Ġfaker": 64086, + "Ġfakt": 66792, + "Ġfal": 25484, + "Ġfall": 4399, + "Ġfallback": 32772, + "Ġfallen": 20866, + "Ġfalling": 15679, + "Ġfallout": 64152, + "Ġfalls": 17066, + "Ġfals": 31932, + "Ġfalse": 895, + "Ġfalsehood": 95018, + "Ġfalsely": 60356, + "Ġfalta": 62999, + "Ġfam": 2091, + "Ġfame": 32551, + "Ġfamed": 60303, + "Ġfamil": 43426, + "Ġfamilia": 50947, + "Ġfamilial": 97204, + "Ġfamiliar": 11285, + "Ġfamiliarity": 70440, + "Ġfamilies": 8521, + "Ġfamille": 66528, + "Ġfamily": 2997, + "Ġfamine": 77954, + "Ġfamous": 11245, + "Ġfamously": 50187, + "ĠfamÃŃlia": 132492, + "Ġfan": 8405, + "Ġfanatic": 96799, + "Ġfanc": 80597, + "Ġfancy": 26457, + "Ġfandom": 74325, + "Ġfans": 7211, + "Ġfant": 8587, + "Ġfantas": 39338, + "Ġfantasies": 59177, + "Ġfantast": 63879, + "Ġfantastic": 14622, + "Ġfantasy": 18378, + "Ġfar": 3041, + "Ġfare": 20425, + "Ġfares": 69218, + "Ġfarewell": 71543, + "Ġfark": 129369, + "Ġfarklı": 129370, + "Ġfarm": 8785, + "Ġfarmer": 36400, + "Ġfarmers": 20336, + "Ġfarmhouse": 82308, + "Ġfarming": 32389, + "Ġfarms": 33224, + "Ġfart": 97495, + "Ġfarther": 42626, + "Ġfas": 66518, + "Ġfasc": 15195, + "Ġfascinated": 60814, + "Ġfascinating": 26291, + "Ġfascination": 74344, + "Ġfascism": 80422, + "Ġfascist": 72879, + "Ġfase": 69712, + "Ġfashion": 11153, + "Ġfashionable": 59323, + "Ġfashioned": 67456, + "Ġfast": 4937, + "Ġfasta": 85249, + "Ġfastball": 93962, + "Ġfaster": 10596, + "Ġfastest": 25648, + "Ġfasting": 53460, + "Ġfat": 8664, + "Ġfatal": 18582, + "ĠfatalError": 57863, + "Ġfatalities": 70205, + "Ġfatally": 71361, + "Ġfate": 24382, + "Ġfather": 6981, + "Ġfathers": 39217, + "Ġfatigue": 35609, + "Ġfats": 49027, + "Ġfatt": 86209, + "Ġfatto": 63878, + "Ġfatty": 38985, + "Ġfauc": 49786, + "Ġfaucet": 71453, + "Ġfault": 14527, + "Ġfaults": 56690, + "Ġfaulty": 57571, + "Ġfauna": 98914, + "Ġfaut": 45990, + "Ġfaux": 50427, + "Ġfav": 9244, + "Ġfavicon": 82804, + "Ġfavor": 4694, + "Ġfavorable": 36749, + "Ġfavored": 45503, + "Ġfavorite": 6930, + "Ġfavorites": 26574, + "Ġfavors": 53847, + "Ġfavour": 12337, + "Ġfavourable": 81185, + "Ġfavoured": 91767, + "Ġfavourite": 18696, + "Ġfavourites": 56934, + "Ġfax": 48619, + "Ġfaz": 37069, + "Ġfazer": 39845, + "Ġfaç": 94872, + "Ġfaçon": 75307, + "Ġfb": 25469, + "Ġfc": 25563, + "Ġfclose": 27204, + "Ġfd": 12414, + "Ġfds": 93734, + "Ġfe": 1152, + "Ġfear": 8679, + "Ġfeared": 37469, + "Ġfearful": 62657, + "Ġfearing": 80525, + "Ġfearless": 92011, + "Ġfears": 23209, + "Ġfeas": 31892, + "Ġfeasibility": 68443, + "Ġfeasible": 42203, + "Ġfeast": 52168, + "Ġfeat": 12347, + "Ġfeather": 49776, + "Ġfeathers": 55894, + "Ġfeats": 63301, + "Ġfeature": 4565, + "Ġfeatured": 14766, + "Ġfeatures": 4419, + "Ġfeaturing": 16445, + "Ġfec": 53961, + "Ġfecha": 23028, + "Ġfed": 22313, + "Ġfeder": 41107, + "Ġfederal": 6775, + "Ġfederally": 79537, + "Ġfederation": 79275, + "Ġfee": 11060, + "Ġfeed": 5395, + "Ġfeedback": 11055, + "Ġfeeder": 57007, + "Ġfeeding": 24991, + "Ġfeeds": 34396, + "Ġfeel": 2666, + "Ġfeeling": 8266, + "Ġfeelings": 15650, + "Ġfeels": 11074, + "Ġfees": 12436, + "Ġfeet": 7541, + "Ġfel": 18027, + "Ġfeliz": 94495, + "Ġfell": 11052, + "Ġfellow": 12357, + "Ġfellows": 86719, + "Ġfellowship": 62249, + "Ġfelon": 79914, + "Ġfelony": 36623, + "Ġfelt": 6476, + "Ġfem": 4996, + "Ġfemale": 8778, + "Ġfemales": 27485, + "Ġfemin": 17002, + "Ġfeminine": 45184, + "Ġfeminism": 52010, + "Ġfeminist": 36912, + "Ġfeminists": 64765, + "Ġfemme": 18195, + "Ġfemmes": 30556, + "Ġfen": 43485, + "Ġfence": 24650, + "Ġfenced": 83393, + "Ġfences": 69155, + "Ġfencing": 67427, + "Ġfend": 94027, + "Ġfer": 18227, + "Ġferm": 80582, + "Ġferment": 67636, + "Ġfermentation": 73354, + "Ġfermented": 94980, + "Ġfern": 87024, + "Ġferr": 57039, + "Ġferry": 51550, + "Ġfert": 21207, + "Ġfertil": 35114, + "Ġfertile": 69125, + "Ġfertility": 47011, + "Ġfertilizer": 64291, + "Ġferv": 66375, + "Ġfest": 18859, + "Ġfestival": 18780, + "Ġfestivals": 44417, + "Ġfestive": 58837, + "Ġfestivities": 79359, + "Ġfet": 16467, + "Ġfetal": 62646, + "Ġfetch": 7807, + "ĠfetchData": 77798, + "Ġfetched": 41442, + "Ġfetching": 44234, + "Ġfetisch": 97730, + "Ġfetish": 39152, + "Ġfetus": 75588, + "Ġfeu": 81611, + "Ġfeud": 56710, + "Ġfeudal": 94350, + "Ġfever": 33553, + "Ġfew": 2421, + "Ġfewer": 16745, + "Ġff": 25539, + "Ġffi": 76956, + "Ġfflush": 52686, + "Ġffm": 75892, + "Ġffmpeg": 84912, + "Ġfft": 43700, + "Ġfg": 29799, + "Ġfgets": 65556, + "Ġfh": 36075, + "Ġfi": 9136, + "Ġfian": 75572, + "Ġfiance": 94188, + "Ġfiat": 53830, + "Ġfib": 15801, + "Ġfiber": 23788, + "Ġfiberglass": 93211, + "Ġfibers": 48674, + "Ġfibonacci": 75698, + "Ġfibr": 94135, + "Ġfibre": 56425, + "Ġfic": 41255, + "Ġfica": 95127, + "Ġfich": 28975, + "Ġfichier": 44320, + "Ġfick": 49203, + "Ġficken": 35879, + "Ġfict": 88107, + "Ġfiction": 16989, + "Ġfictional": 43582, + "Ġfid": 32104, + "Ġfidelity": 82504, + "Ġfie": 62305, + "Ġfield": 2070, + "ĠfieldName": 38538, + "ĠfieldType": 98151, + "ĠfieldValue": 95084, + "Ġfields": 5043, + "Ġfier": 30518, + "Ġfierc": 74627, + "Ġfierce": 41342, + "Ġfiercely": 77945, + "Ġfiery": 63842, + "Ġfif": 11128, + "Ġfifo": 63497, + "Ġfifteen": 36655, + "Ġfifth": 17702, + "Ġfifty": 32417, + "Ġfig": 4144, + "Ġfight": 4367, + "Ġfighter": 24860, + "Ġfighters": 23202, + "Ġfighting": 10805, + "Ġfights": 27433, + "Ġfigsize": 78101, + "Ġfigur": 55655, + "Ġfigura": 93715, + "Ġfigure": 7071, + "Ġfigured": 24692, + "Ġfigures": 12396, + "Ġfiguring": 47209, + "Ġfil": 1461, + "Ġfila": 48769, + "Ġfilament": 90780, + "Ġfile": 1034, + "ĠfileId": 91261, + "ĠfileInfo": 73651, + "ĠfileList": 69076, + "ĠfileName": 12665, + "ĠfilePath": 22598, + "ĠfileSize": 71976, + "ĠfileType": 82646, + "Ġfiled": 12729, + "Ġfilename": 3899, + "Ġfilenames": 40540, + "Ġfilepath": 26054, + "Ġfileprivate": 55595, + "Ġfiles": 3542, + "Ġfilesize": 70436, + "Ġfilesystem": 38389, + "Ġfiletype": 90271, + "Ġfilho": 91521, + "Ġfilib": 95686, + "Ġfiling": 25480, + "Ġfilings": 67148, + "Ġfill": 5155, + "ĠfillColor": 73235, + "Ġfille": 36154, + "Ġfilled": 10199, + "Ġfiller": 54710, + "Ġfilles": 55027, + "Ġfilling": 21274, + "Ġfills": 40587, + "Ġfilm": 4531, + "Ġfilme": 41080, + "Ġfilmed": 41408, + "Ġfilmer": 74580, + "Ġfilmes": 97796, + "Ġfilming": 38870, + "Ġfilmm": 33526, + "Ġfilmmaker": 56237, + "Ġfilmmakers": 62103, + "Ġfilmpjes": 94330, + "Ġfilms": 12351, + "Ġfilmy": 89394, + "Ġfils": 88572, + "Ġfilt": 71854, + "Ġfilter": 4051, + "Ġfiltered": 18293, + "Ġfiltering": 29670, + "Ġfilters": 13406, + "Ġfilthy": 77342, + "Ġfiltr": 50691, + "Ġfiltration": 74938, + "Ġfiltro": 74485, + "Ġfim": 52087, + "Ġfin": 1875, + "Ġfinal": 1590, + "Ġfinale": 36298, + "Ġfinalist": 97391, + "Ġfinalists": 82546, + "Ġfinalize": 54090, + "Ġfinalized": 60387, + "Ġfinally": 5499, + "Ġfinals": 40302, + "Ġfinanc": 40545, + "Ġfinance": 17017, + "Ġfinanced": 69664, + "Ġfinances": 39282, + "Ġfinancial": 5896, + "Ġfinancially": 37975, + "Ġfinancier": 96298, + "Ġfinancing": 28542, + "Ġfinanzi": 87878, + "Ġfind": 1477, + "ĠfindAll": 42785, + "ĠfindBy": 44459, + "ĠfindById": 59525, + "ĠfindOne": 78530, + "ĠfindViewById": 10241, + "Ġfinde": 82207, + "Ġfinden": 35103, + "Ġfinder": 43730, + "Ġfindet": 75315, + "Ġfinding": 9271, + "Ġfindings": 14613, + "Ġfinds": 13719, + "Ġfine": 6915, + "Ġfined": 53881, + "Ġfinely": 60702, + "Ġfiner": 72786, + "Ġfines": 33918, + "Ġfinest": 27707, + "Ġfing": 38727, + "Ġfinger": 14317, + "Ġfingerprint": 42597, + "Ġfingerprints": 76677, + "Ġfingers": 19225, + "Ġfingert": 63603, + "Ġfingertips": 71048, + "Ġfinish": 6248, + "Ġfinished": 8060, + "Ġfinishes": 33036, + "Ġfinishing": 24284, + "Ġfinite": 34226, + "Ġfinns": 73065, + "Ġfino": 76130, + "Ġfins": 64979, + "Ġfint": 84285, + "Ġfir": 33992, + "Ġfire": 3940, + "ĠfireEvent": 76755, + "Ġfirearm": 35605, + "Ġfirearms": 31553, + "Ġfirebase": 20424, + "Ġfired": 13895, + "Ġfirefight": 37603, + "Ġfirefighter": 94165, + "Ġfirefighters": 52106, + "Ġfirefox": 81683, + "Ġfireplace": 39411, + "Ġfirepower": 93726, + "Ġfires": 26084, + "Ġfirestore": 73993, + "Ġfirewall": 49877, + "Ġfireworks": 51833, + "Ġfiring": 22830, + "Ġfirm": 7474, + "Ġfirma": 80353, + "Ġfirmly": 31520, + "Ġfirms": 18809, + "Ġfirmware": 29678, + "Ġfirst": 1156, + "ĠfirstName": 21046, + "Ġfirsthand": 82568, + "Ġfirstly": 93952, + "Ġfirstname": 48193, + "Ġfis": 41623, + "Ġfiscal": 19743, + "Ġfish": 7640, + "Ġfisheries": 81496, + "Ġfishermen": 73193, + "Ġfishes": 94361, + "Ġfishing": 19948, + "Ġfiss": 94059, + "Ġfisse": 67662, + "Ġfist": 39802, + "Ġfists": 80684, + "Ġfit": 4946, + "Ġfitness": 17042, + "Ġfits": 18304, + "Ġfitte": 66455, + "Ġfitted": 28341, + "Ġfitting": 26345, + "Ġfittings": 68926, + "Ġfive": 4236, + "Ġfix": 5046, + "Ġfixation": 83762, + "Ġfixed": 8356, + "Ġfixes": 26537, + "Ġfixing": 35251, + "Ġfixture": 12507, + "Ġfixtures": 37664, + "Ġfiyat": 128102, + "Ġfiyatları": 140001, + "Ġfiyatı": 141974, + "Ġfiz": 78211, + "Ġfizz": 86979, + "Ġfj": 75371, + "Ġfk": 62683, + "Ġfkk": 60798, + "Ġfl": 1320, + "Ġfla": 96128, + "Ġflag": 5181, + "Ġflagged": 61209, + "Ġflags": 8042, + "Ġflagship": 42672, + "Ġflair": 68565, + "Ġflakes": 81623, + "Ġflam": 73499, + "Ġflame": 34578, + "Ġflames": 38785, + "Ġflaming": 84623, + "Ġflank": 69492, + "Ġflap": 66433, + "Ġflare": 60263, + "Ġflash": 8217, + "Ġflashback": 97423, + "Ġflashed": 69839, + "Ġflashes": 61285, + "Ġflashing": 49757, + "Ġflashlight": 78415, + "Ġflashy": 95467, + "Ġflask": 19534, + "Ġflat": 10063, + "Ġflats": 60357, + "Ġflatt": 51039, + "Ġflatten": 45343, + "Ġflattened": 67361, + "Ġflattering": 83108, + "Ġflav": 18276, + "Ġflavor": 17172, + "Ġflavored": 87240, + "Ġflavorful": 93531, + "Ġflavors": 31423, + "Ġflavour": 46597, + "Ġflavours": 70082, + "Ġflaw": 25700, + "Ġflawed": 46908, + "Ġflawless": 68516, + "Ġflaws": 40759, + "Ġfld": 59166, + "Ġfle": 12962, + "Ġflea": 95855, + "Ġfled": 28910, + "Ġflee": 27400, + "Ġfleece": 91990, + "Ġfleeing": 49287, + "Ġfleet": 25099, + "Ġfleeting": 95335, + "Ġfleets": 90649, + "Ġflere": 85591, + "Ġflesh": 24955, + "Ġflew": 31022, + "Ġflex": 5763, + "ĠflexDirection": 29397, + "ĠflexGrow": 91770, + "Ġflexibility": 24177, + "Ġflexible": 18776, + "Ġflick": 28347, + "Ġflies": 37104, + "Ġflight": 10971, + "Ġflights": 24908, + "Ġflip": 18287, + "Ġflipped": 46080, + "Ġflipping": 64661, + "Ġflips": 85186, + "Ġflirt": 52630, + "Ġflirting": 87242, + "Ġflo": 9744, + "Ġfloat": 2224, + "ĠfloatValue": 81513, + "Ġfloated": 66035, + "Ġfloating": 19057, + "Ġfloats": 47902, + "Ġflock": 48716, + "Ġflood": 17726, + "Ġflooded": 48786, + "Ġflooding": 38162, + "Ġfloods": 60172, + "Ġfloor": 6422, + "Ġflooring": 36148, + "Ġfloors": 25945, + "Ġflop": 90342, + "Ġfloppy": 91358, + "Ġflor": 69140, + "Ġflora": 80988, + "Ġfloral": 45019, + "Ġflorida": 75654, + "Ġflotation": 67197, + "Ġflour": 19828, + "Ġflourish": 66684, + "Ġflourishing": 98259, + "Ġflow": 6396, + "Ġflowed": 84650, + "Ġflower": 22351, + "Ġflowering": 81239, + "Ġflowers": 19281, + "Ġflowing": 35512, + "Ġflown": 52750, + "Ġflows": 27455, + "Ġflu": 19660, + "Ġfluct": 38288, + "Ġfluctuations": 64549, + "Ġfluent": 57768, + "Ġfluffy": 67561, + "Ġfluid": 15590, + "Ġfluids": 55306, + "Ġfluor": 53636, + "Ġfluores": 50781, + "Ġfluorescence": 96232, + "Ġfluorescent": 73764, + "Ġfluoride": 88130, + "Ġflurry": 87463, + "Ġflush": 18198, + "Ġflushed": 73720, + "Ġflushing": 90532, + "Ġflute": 95712, + "Ġflutter": 73783, + "Ġflux": 30305, + "Ġfly": 11466, + "Ġflyer": 74906, + "Ġflyers": 85008, + "Ġflying": 16307, + "Ġfm": 31121, + "Ġfmap": 69579, + "Ġfmt": 8879, + "Ġfn": 5168, + "Ġfname": 22548, + "Ġfo": 11756, + "Ġfoam": 31083, + "Ġfoc": 8249, + "Ġfocal": 41099, + "Ġfocus": 5244, + "Ġfocused": 10735, + "Ġfocuses": 23497, + "Ġfocusing": 21080, + "Ġfod": 61429, + "Ġfodder": 97795, + "Ġfoe": 51977, + "Ġfoes": 53744, + "Ġfog": 30249, + "Ġfoi": 21679, + "Ġfoil": 46399, + "Ġfois": 36191, + "Ġfol": 9443, + "Ġfold": 11555, + "Ġfolded": 47035, + "Ġfolder": 8527, + "Ġfolders": 29242, + "Ġfolding": 44742, + "Ġfolds": 60507, + "Ġfoliage": 79005, + "Ġfolk": 27936, + "Ġfolklore": 97869, + "Ġfolks": 15327, + "Ġfoll": 51406, + "Ġfollando": 71579, + "Ġfollic": 81614, + "Ġfollow": 1795, + "Ġfollowed": 8110, + "Ġfollower": 48207, + "Ġfollowers": 20119, + "Ġfollowing": 2701, + "Ġfollows": 11017, + "Ġfolly": 80689, + "Ġfon": 27305, + "Ġfonction": 33401, + "Ġfond": 21208, + "Ġfondo": 80529, + "Ġfont": 3301, + "ĠfontFamily": 29135, + "ĠfontSize": 10893, + "ĠfontStyle": 88237, + "ĠfontWeight": 18551, + "ĠfontWithName": 97118, + "Ġfonts": 32980, + "Ġfontsize": 35222, + "Ġfoo": 15229, + "Ġfood": 3607, + "Ġfoods": 15298, + "Ġfool": 23977, + "Ġfooled": 79547, + "Ġfoolish": 45237, + "Ġfools": 83642, + "Ġfoot": 4478, + "Ġfootage": 21852, + "Ġfootball": 8964, + "Ġfooter": 23173, + "Ġfooth": 78363, + "Ġfooting": 73403, + "Ġfootnote": 86810, + "Ġfootprint": 42872, + "Ġfootsteps": 59702, + "Ġfootwear": 67072, + "Ġfopen": 20377, + "Ġfor": 369, + "ĠforCell": 72522, + "ĠforCellReuseIdentifier": 93856, + "ĠforControlEvents": 48801, + "ĠforEach": 82532, + "ĠforIndexPath": 98656, + "ĠforKey": 18091, + "ĠforState": 28228, + "Ġfora": 84984, + "Ġforall": 31647, + "Ġforam": 54534, + "Ġforb": 54744, + "Ġforbid": 73631, + "Ġforbidden": 36813, + "Ġforc": 51742, + "Ġforce": 5344, + "Ġforced": 9575, + "Ġforcefully": 85978, + "Ġforces": 8437, + "Ġforcibly": 73823, + "Ġforcing": 24913, + "Ġford": 44628, + "Ġfore": 2241, + "Ġforeach": 4043, + "Ġforearm": 90565, + "Ġforecast": 17595, + "Ġforecasting": 78382, + "Ġforecasts": 50065, + "Ġforeclosure": 80282, + "Ġforefront": 51201, + "Ġforegoing": 87831, + "Ġforeground": 39305, + "Ġforehead": 51254, + "Ġforeign": 7214, + "Ġforeigners": 47412, + "Ġforemost": 42680, + "Ġforensic": 54221, + "Ġfores": 69756, + "Ġforesee": 56691, + "Ġforeseeable": 77903, + "Ġforest": 13638, + "Ġforestry": 87034, + "Ġforests": 35558, + "Ġforever": 15683, + "Ġforex": 29806, + "Ġforfe": 48932, + "Ġforfeiture": 89621, + "Ġforg": 11548, + "Ġforge": 56628, + "Ġforged": 53199, + "Ġforget": 10667, + "Ġforgetting": 65027, + "Ġforging": 94831, + "Ġforgive": 43391, + "Ġforgiven": 72595, + "Ġforgiveness": 52050, + "Ġforgiving": 92341, + "Ġforgot": 28595, + "Ġforgotten": 24547, + "Ġfork": 22435, + "Ġforks": 73574, + "Ġform": 1352, + "ĠformData": 27977, + "Ġforma": 21791, + "Ġformal": 15908, + "Ġformally": 36302, + "Ġformas": 83089, + "Ġformat": 3561, + "ĠformatDate": 86772, + "Ġformation": 18003, + "Ġformations": 61164, + "Ġformato": 54856, + "Ġformats": 19856, + "Ġformatted": 23126, + "Ġformatter": 24814, + "Ġformatting": 36566, + "Ġforme": 56028, + "Ġformed": 14122, + "Ġformer": 4741, + "Ġformerly": 33733, + "Ġformidable": 53729, + "Ġforming": 29064, + "Ġforms": 7586, + "Ġformul": 80940, + "Ġformula": 14806, + "Ġformulaire": 94895, + "Ġformulario": 62428, + "Ġformulas": 54079, + "Ġformulate": 88859, + "Ġformulated": 59294, + "Ġformulation": 54886, + "Ġformulations": 96977, + "Ġforn": 56009, + "Ġfors": 74651, + "Ġforsk": 83352, + "Ġfort": 11845, + "Ġforte": 63958, + "Ġforth": 13241, + "Ġforthcoming": 44664, + "Ġfortified": 94436, + "Ġfortn": 83211, + "Ġfortress": 69838, + "Ġforts": 74552, + "Ġfortunate": 39410, + "Ġfortunately": 86832, + "Ġfortune": 32315, + "Ġfortunes": 67503, + "Ġforty": 35398, + "Ġforum": 11848, + "Ġforums": 24865, + "Ġforward": 4637, + "Ġforwarded": 52246, + "Ġforwarding": 62104, + "Ġforwards": 41054, + "Ġforça": 135192, + "Ġfos": 48390, + "Ġfoss": 25421, + "Ġfossil": 30276, + "Ġfossils": 80373, + "Ġfost": 36313, + "Ġfoster": 29987, + "Ġfostering": 85544, + "Ġfot": 31840, + "Ġfoto": 25575, + "Ġfotograf": 63246, + "Ġfotos": 32959, + "ĠfotoÄŁ": 134222, + "ĠfotoÄŁraf": 134223, + "Ġfou": 56525, + "Ġfought": 20463, + "Ġfoul": 39784, + "Ġfound": 1730, + "Ġfoundation": 16266, + "Ġfoundational": 88228, + "Ġfoundations": 40482, + "Ġfounded": 18047, + "Ġfounder": 18996, + "Ġfounders": 47627, + "Ġfounding": 35230, + "Ġfountain": 56613, + "Ġfour": 3040, + "Ġfourn": 80894, + "Ġfours": 89620, + "Ġfourteen": 62260, + "Ġfourth": 11737, + "Ġfout": 38273, + "Ġfov": 93473, + "Ġfox": 38835, + "Ġfoyer": 98916, + "Ġfp": 12007, + "Ġfprintf": 9377, + "Ġfps": 33881, + "Ġfputs": 77036, + "Ġfq": 82467, + "Ġfr": 1422, + "Ġfra": 6677, + "Ġfrac": 53676, + "Ġfracking": 70054, + "Ġfract": 26602, + "Ġfraction": 19419, + "Ġfractional": 68209, + "Ġfractions": 64895, + "Ġfracture": 58804, + "Ġfractured": 79108, + "Ġfractures": 77232, + "Ġfrag": 8343, + "Ġfragile": 44250, + "Ġfragment": 12289, + "ĠfragmentManager": 94532, + "Ġfragmentation": 87352, + "Ġfragmented": 86095, + "Ġfragments": 34503, + "Ġfragrance": 55127, + "Ġfrags": 97203, + "Ġfrail": 90123, + "Ġfrais": 90553, + "Ġfram": 59643, + "Ġframe": 4034, + "Ġframeborder": 71462, + "Ġframebuffer": 73019, + "Ġframed": 45993, + "Ġframerate": 97251, + "Ġframes": 14087, + "Ġframework": 12626, + "Ġframeworks": 48025, + "Ġframing": 57949, + "Ġfran": 22156, + "Ġfranc": 43479, + "Ġfranca": 34747, + "Ġfrancais": 50332, + "Ġfrancaise": 66468, + "Ġfrance": 47587, + "Ġfranch": 16545, + "Ġfranchise": 19024, + "Ġfranchises": 63903, + "Ġfrank": 25790, + "Ġfrankfurt": 82175, + "Ġfrankly": 41662, + "Ġfrantic": 88606, + "Ġfranç": 41337, + "Ġfrançais": 54367, + "Ġfrançaise": 92324, + "Ġfrase": 90329, + "Ġfrat": 80972, + "Ġfraternity": 78296, + "Ġfrau": 37031, + "Ġfraud": 16124, + "Ġfraudulent": 53883, + "Ġfrauen": 19835, + "Ġfraught": 95959, + "Ġfray": 97253, + "Ġfre": 3457, + "Ġfread": 46134, + "Ġfreak": 29205, + "Ġfreaking": 73240, + "Ġfrec": 81105, + "Ġfred": 97126, + "Ġfreder": 84076, + "Ġfree": 1910, + "Ġfreed": 27534, + "Ġfreedom": 11290, + "Ġfreedoms": 59977, + "Ġfreeing": 66717, + "Ġfreel": 31131, + "Ġfreelance": 45109, + "Ġfreelancer": 92668, + "Ġfreely": 25580, + "Ġfrees": 71455, + "Ġfreeway": 83574, + "Ġfreeze": 29930, + "Ġfreezer": 49585, + "Ġfreezes": 93377, + "Ġfreezing": 42218, + "Ġfrei": 83423, + "Ġfreight": 45117, + "Ġfrem": 61122, + "Ġfren": 46834, + "Ġfrench": 41193, + "Ġfrente": 56722, + "Ġfrenzy": 73841, + "Ġfreopen": 56443, + "Ġfreq": 20895, + "Ġfrequ": 6166, + "Ġfrequencies": 33773, + "Ġfrequency": 11639, + "Ġfrequent": 20757, + "Ġfrequently": 13814, + "Ġfres": 48873, + "Ġfresh": 7722, + "Ġfreshly": 49899, + "Ġfreshman": 40217, + "Ġfreshmen": 97749, + "Ġfreshness": 98216, + "Ġfreshwater": 79644, + "Ġfret": 53064, + "Ġfreund": 77634, + "Ġfri": 2633, + "Ġfriction": 38576, + "Ġfriday": 79313, + "Ġfridge": 37581, + "Ġfried": 40851, + "Ġfriend": 4238, + "Ġfriendly": 11657, + "Ġfriends": 4780, + "Ġfriendship": 26509, + "Ġfriendships": 61981, + "Ġfries": 51931, + "Ġfright": 29547, + "Ġfrightened": 59078, + "Ġfrightening": 53639, + "Ġfringe": 54807, + "Ġfrivol": 95424, + "Ġfrm": 28354, + "Ġfro": 12799, + "Ġfrog": 59881, + "Ġfrogs": 89446, + "Ġfrom": 504, + "ĠfromDate": 79616, + "Ġfron": 77952, + "Ġfront": 4065, + "Ġfrontal": 65646, + "Ġfrontend": 45645, + "Ġfrontier": 48000, + "Ġfrontline": 96001, + "Ġfronts": 63390, + "Ġfrost": 44734, + "Ġfrosting": 98181, + "Ġfrowned": 76359, + "Ġfroze": 89009, + "Ġfrozen": 19690, + "Ġfruit": 13779, + "Ġfruitful": 88584, + "Ġfruition": 93606, + "Ġfruits": 25322, + "Ġfruity": 91932, + "Ġfrustr": 14457, + "Ġfrustrated": 32530, + "Ġfrustrating": 34611, + "Ġfrustration": 31986, + "Ġfrustrations": 88934, + "Ġfry": 52546, + "Ġfrying": 86096, + "ĠfrÃ¥n": 48106, + "Ġfrække": 98403, + "Ġfrü": 98877, + "Ġfrüh": 134303, + "Ġfs": 8619, + "Ġfscanf": 68486, + "Ġfseek": 79262, + "Ġfsm": 83138, + "Ġfst": 48434, + "Ġft": 10482, + "Ġftp": 45427, + "Ġfu": 18415, + "Ġfuck": 7820, + "Ġfucked": 27153, + "Ġfucking": 17224, + "Ġfucks": 54363, + "Ġfue": 22026, + "Ġfuel": 10416, + "Ġfueled": 58361, + "Ġfuels": 39273, + "Ġfuer": 38651, + "Ġfuera": 69675, + "Ġfueron": 67146, + "Ġfug": 48084, + "Ġful": 5599, + "Ġfulfil": 63101, + "Ġfulfill": 20423, + "Ġfulfilled": 40734, + "Ġfulfilling": 49598, + "Ġfulfillment": 56283, + "Ġfull": 2480, + "ĠfullName": 48008, + "ĠfullPath": 85844, + "ĠfullWidth": 51320, + "Ġfuller": 86918, + "Ġfullest": 76696, + "Ġfullfile": 84616, + "Ġfullname": 61036, + "Ġfullscreen": 56927, + "Ġfully": 7225, + "Ġfun": 2464, + "Ġfunc": 2915, + "Ġfuncion": 27863, + "Ġfunciona": 86896, + "Ġfunciones": 83412, + "Ġfuncionários": 143768, + "Ġfunción": 63077, + "Ġfuncs": 76871, + "Ġfunct": 41051, + "Ġfunction": 729, + "ĠfunctionName": 90519, + "Ġfunctional": 15629, + "Ġfunctionalities": 92080, + "Ġfunctionality": 14944, + "Ġfunctioning": 30201, + "Ġfunctions": 5746, + "Ġfunctools": 56466, + "Ġfunctor": 74712, + "Ġfund": 3802, + "Ġfundament": 12479, + "Ġfundamental": 15811, + "Ġfundamentally": 42493, + "Ġfundamentals": 56840, + "Ġfunded": 23906, + "Ġfunding": 10773, + "Ġfundra": 26318, + "Ġfundraiser": 58659, + "Ġfundraising": 36319, + "Ġfunds": 10514, + "Ġfuneral": 30979, + "Ġfung": 33598, + "Ġfungal": 94347, + "Ġfungi": 80440, + "Ġfungus": 78802, + "Ġfunk": 68292, + "Ġfunkc": 71778, + "Ġfunktion": 68312, + "Ġfunky": 73888, + "Ġfunnel": 60219, + "Ġfunny": 15173, + "Ġfunção": 62753, + "Ġfur": 18241, + "Ġfurious": 52070, + "Ġfurn": 9439, + "Ġfurnace": 54464, + "Ġfurnish": 51159, + "Ġfurnished": 23437, + "Ġfurnishings": 72171, + "Ġfurniture": 14549, + "Ġfurry": 74419, + "Ġfurther": 4623, + "Ġfurthermore": 77537, + "Ġfury": 63059, + "Ġfus": 68478, + "Ġfuse": 31702, + "Ġfused": 74654, + "Ġfusion": 36508, + "Ġfuss": 63764, + "Ġfut": 17645, + "Ġfutile": 84527, + "Ġfuture": 3853, + "Ġfutures": 36823, + "Ġfuturistic": 79220, + "Ġfuturo": 64644, + "Ġfuzz": 76142, + "Ġfuzzy": 52733, + "Ġfv": 61654, + "Ġfw": 33886, + "Ġfwd": 91004, + "Ġfwrite": 45235, + "Ġfx": 33219, + "Ġfy": 51941, + "Ġfz": 92649, + "Ġfácil": 59140, + "ĠfÃ¥": 38605, + "ĠfÃ¥r": 48516, + "Ġfé": 57382, + "Ġfévrier": 135431, + "Ġfête": 140738, + "Ġfö": 40500, + "Ġför": 16641, + "Ġfø": 57370, + "Ġfør": 46483, + "Ġførste": 78436, + "Ġfüh": 50232, + "Ġfün": 96120, + "Ġfünf": 136740, + "Ġfür": 7180, + "ĠfÃŃs": 64846, + "ĠfÃŃsica": 134921, + "ĠfÃŃsico": 139074, + "Ġfı": 136901, + "Ġfırs": 136902, + "Ġfırsat": 136903, + "Ġg": 342, + "Ġga": 13510, + "Ġgaan": 61613, + "Ġgaat": 68045, + "Ġgab": 55879, + "Ġgad": 33216, + "Ġgadget": 58851, + "Ġgadgets": 53375, + "Ġgag": 45592, + "Ġgain": 8722, + "Ġgained": 18163, + "Ġgaining": 29140, + "Ġgains": 19619, + "Ġgal": 15369, + "Ġgala": 96480, + "Ġgalaxies": 64917, + "Ġgalaxy": 33161, + "Ġgaler": 77607, + "Ġgall": 19203, + "Ġgalleries": 42554, + "Ġgallery": 18046, + "Ġgallon": 49580, + "Ġgallons": 50122, + "Ġgam": 9015, + "Ġgamb": 74934, + "Ġgamble": 63801, + "Ġgambling": 24696, + "Ġgame": 1809, + "ĠgameId": 92178, + "ĠgameObject": 36236, + "ĠgameOver": 71837, + "ĠgameState": 54993, + "ĠgameTime": 93254, + "Ġgameplay": 26029, + "Ġgamer": 57918, + "Ġgamers": 34743, + "Ġgames": 3868, + "Ġgaming": 15833, + "Ġgamle": 76552, + "Ġgamm": 85842, + "Ġgamma": 21619, + "Ġgan": 24851, + "Ġgang": 13179, + "Ġgangbang": 40754, + "Ġgangs": 54123, + "Ġganz": 38805, + "Ġganze": 93898, + "Ġgap": 12929, + "Ġgaping": 98936, + "Ġgaps": 32151, + "Ġgar": 7364, + "Ġgarage": 19277, + "Ġgarant": 39092, + "Ġgarbage": 25878, + "Ġgard": 21881, + "Ġgarden": 13551, + "Ġgardening": 59199, + "Ġgardens": 35436, + "Ġgarg": 96083, + "Ġgarlic": 30635, + "Ġgarment": 59618, + "Ġgarments": 66456, + "Ġgarn": 59916, + "Ġgarner": 46101, + "Ġgarnered": 67290, + "Ġgas": 6819, + "Ġgases": 44512, + "Ġgasoline": 45495, + "Ġgast": 87659, + "Ġgastr": 55680, + "Ġgastric": 88285, + "Ġgastro": 58249, + "Ġgastrointestinal": 79211, + "Ġgate": 18126, + "Ġgated": 84711, + "Ġgates": 34534, + "Ġgateway": 28795, + "Ġgather": 9567, + "Ġgathered": 20190, + "Ġgathering": 22888, + "Ġgatherings": 65137, + "Ġgathers": 85681, + "Ġgating": 73399, + "Ġgauche": 95621, + "Ġgauge": 30890, + "Ġgauss": 86869, + "Ġgaussian": 63447, + "Ġgave": 6551, + "Ġgay": 8321, + "Ġgays": 56477, + "Ġgaz": 54986, + "Ġgaze": 35396, + "Ġgb": 47015, + "Ġgbc": 65461, + "Ġgboolean": 48667, + "Ġgc": 22122, + "Ġgcc": 50917, + "Ġgcd": 44858, + "Ġgchar": 38761, + "Ġgd": 32630, + "Ġgdb": 75953, + "Ġgdk": 89939, + "Ġgdy": 75934, + "Ġgdyż": 139578, + "Ġgdzie": 95491, + "Ġge": 3893, + "Ġgear": 14448, + "Ġgearbox": 97873, + "Ġgeared": 58447, + "Ġgearing": 80722, + "Ġgears": 53160, + "Ġgeb": 29167, + "Ġgeben": 56500, + "Ġgebru": 43638, + "Ġgebruik": 83654, + "Ġged": 52193, + "Ġgee": 56505, + "Ġgeek": 55639, + "Ġgeen": 47820, + "Ġgef": 27684, + "Ġgefunden": 81671, + "Ġgeg": 97341, + "Ġgegen": 34699, + "Ġgegenüber": 139115, + "Ġgeh": 29695, + "Ġgehen": 62459, + "Ġgeht": 39264, + "Ġgehören": 142037, + "Ġgehört": 93096, + "Ġgeil": 62022, + "Ġgeile": 24162, + "Ġgek": 48323, + "Ġgel": 17837, + "Ġgeld": 41919, + "ĠgeldiÄŁi": 138315, + "Ġgele": 97502, + "ĠgeleceÄŁi": 143806, + "Ġgeli": 125963, + "ĠgeliÅŁ": 127436, + "ĠgeliÅŁme": 136126, + "ĠgeliÅŁtir": 132187, + "Ġgem": 18747, + "Ġgemacht": 69876, + "Ġgeme": 51637, + "Ġgemeins": 74354, + "Ġgems": 42158, + "Ġgen": 4081, + "Ġgenau": 68876, + "Ġgender": 9825, + "Ġgenders": 81328, + "Ġgene": 14862, + "Ġgener": 1766, + "Ġgenera": 83435, + "Ġgenerado": 79634, + "Ġgeneral": 4586, + "Ġgeneralize": 92540, + "Ġgeneralized": 66117, + "Ġgenerally": 8789, + "Ġgenerals": 70067, + "Ġgenerar": 94979, + "Ġgenerate": 6923, + "Ġgenerated": 7907, + "Ġgenerates": 26885, + "Ġgenerating": 23163, + "Ġgeneration": 9471, + "Ġgenerations": 21787, + "Ġgenerator": 13823, + "Ġgenerators": 43063, + "Ġgeneric": 13954, + "Ġgenerics": 95545, + "Ġgenerosity": 64252, + "Ġgenerous": 27671, + "Ġgenerously": 79120, + "Ġgenes": 20727, + "Ġgenesis": 59366, + "Ġgenetic": 18929, + "Ġgenetically": 50933, + "Ġgenetics": 55004, + "Ġgenie": 82757, + "Ġgenital": 56355, + "Ġgenitals": 90642, + "Ġgenius": 34101, + "ĠgeniÅŁ": 134325, + "Ġgenocide": 52524, + "Ġgenom": 98198, + "Ġgenome": 32769, + "Ġgenomes": 84281, + "Ġgenomic": 79964, + "Ġgenotype": 79185, + "Ġgenre": 17328, + "Ġgenres": 35644, + "Ġgens": 46004, + "Ġgent": 15886, + "Ġgente": 43042, + "Ġgentle": 21700, + "Ġgentleman": 45129, + "Ġgentlemen": 57809, + "Ġgently": 29273, + "Ġgenu": 16497, + "Ġgenuine": 22011, + "Ġgenuinely": 35197, + "Ġgenus": 63577, + "Ġgenç": 132425, + "Ġgeo": 27581, + "Ġgeographic": 45039, + "Ġgeographical": 52901, + "Ġgeography": 53142, + "Ġgeological": 85178, + "Ġgeom": 27256, + "Ġgeomet": 67986, + "Ġgeometric": 52484, + "Ġgeometry": 17047, + "Ġgeopol": 73456, + "Ġgeopolitical": 86898, + "Ġgep": 82942, + "Ġger": 17239, + "Ġgerade": 67058, + "Ġgeral": 97892, + "Ġgere": 66287, + "Ġgerekt": 140248, + "Ġgerekti": 140249, + "ĠgerektiÄŁini": 140250, + "Ġgerm": 81886, + "Ġgerman": 42527, + "Ġgermany": 82824, + "Ġgerne": 54064, + "Ġgerçek": 124962, + "ĠgerçekleÅŁ": 135442, + "ĠgerçekleÅŁtir": 127516, + "ĠgerçekleÅŁtiril": 137483, + "Ġgerçekten": 138837, + "Ġges": 14410, + "Ġgesch": 25620, + "Ġgeschichten": 50737, + "Ġgest": 12743, + "Ġgestion": 58371, + "Ġgestión": 138747, + "Ġgesture": 30157, + "Ġgestures": 52392, + "Ġgestão": 137632, + "Ġget": 633, + "ĠgetActivity": 47934, + "ĠgetAddress": 80623, + "ĠgetAll": 23955, + "ĠgetApp": 76382, + "ĠgetArguments": 97879, + "ĠgetBy": 57169, + "ĠgetById": 81502, + "ĠgetC": 43099, + "ĠgetCategory": 93575, + "ĠgetChild": 76518, + "ĠgetClass": 47237, + "ĠgetClient": 86287, + "ĠgetCode": 82873, + "ĠgetColor": 73468, + "ĠgetColumn": 97240, + "ĠgetConfig": 66763, + "ĠgetConnection": 65313, + "ĠgetContent": 63187, + "ĠgetContentPane": 53576, + "ĠgetContext": 66175, + "ĠgetCount": 54938, + "ĠgetCurrent": 40517, + "ĠgetData": 27616, + "ĠgetDate": 60166, + "ĠgetDefault": 69106, + "ĠgetDescription": 54733, + "ĠgetElement": 71547, + "ĠgetEmail": 67408, + "ĠgetField": 70964, + "ĠgetFile": 62442, + "ĠgetHeight": 62444, + "ĠgetId": 20452, + "ĠgetImage": 54561, + "ĠgetIndex": 89771, + "ĠgetInfo": 91824, + "ĠgetInput": 88977, + "ĠgetInstance": 37886, + "ĠgetInt": 85133, + "ĠgetIntent": 48811, + "ĠgetItem": 26978, + "ĠgetItemCount": 59772, + "ĠgetKey": 74103, + "ĠgetLast": 81479, + "ĠgetList": 59891, + "ĠgetLocation": 83392, + "ĠgetLogger": 96663, + "ĠgetMax": 81785, + "ĠgetMenu": 72827, + "ĠgetMenuInflater": 91000, + "ĠgetMessage": 57753, + "ĠgetModel": 73326, + "ĠgetName": 20299, + "ĠgetNext": 70379, + "ĠgetNode": 81798, + "ĠgetObject": 89173, + "ĠgetOrder": 82627, + "ĠgetP": 57720, + "ĠgetPage": 70853, + "ĠgetParent": 75578, + "ĠgetPassword": 68439, + "ĠgetPath": 82740, + "ĠgetPlayer": 98037, + "ĠgetPosition": 84357, + "ĠgetPrice": 89332, + "ĠgetProduct": 78203, + "ĠgetProperty": 75496, + "ĠgetRandom": 52436, + "ĠgetRequest": 78731, + "ĠgetResource": 87794, + "ĠgetResources": 58145, + "ĠgetResult": 83415, + "ĠgetS": 73683, + "ĠgetService": 85869, + "ĠgetSession": 74694, + "ĠgetSize": 66735, + "ĠgetSource": 71003, + "ĠgetState": 39070, + "ĠgetStatus": 56424, + "ĠgetString": 30907, + "ĠgetSupportActionBar": 72828, + "ĠgetSupportFragmentManager": 87435, + "ĠgetSystemService": 91376, + "ĠgetText": 63038, + "ĠgetTime": 60213, + "ĠgetTitle": 58832, + "ĠgetToken": 54111, + "ĠgetTotal": 74456, + "ĠgetType": 36581, + "ĠgetUrl": 68019, + "ĠgetUser": 23197, + "ĠgetUserId": 87498, + "ĠgetUsername": 85414, + "ĠgetUsers": 77569, + "ĠgetValue": 31720, + "ĠgetVersion": 97245, + "ĠgetView": 49885, + "ĠgetWidth": 69262, + "ĠgetWindow": 82461, + "ĠgetX": 75656, + "ĠgetY": 75280, + "Ġgetattr": 29997, + "Ġgetaway": 69893, + "Ġgetch": 96551, + "Ġgetchar": 60644, + "Ġgetenv": 55474, + "Ġgetline": 55198, + "Ġgetopt": 66902, + "Ġgetpid": 95378, + "Ġgets": 5221, + "Ġgetter": 33429, + "Ġgetters": 52894, + "Ġgettext": 63706, + "Ġgettimeofday": 88579, + "Ġgetting": 3709, + "Ġgev": 94406, + "Ġgew": 22539, + "Ġgez": 41150, + "Ġgezocht": 90274, + "Ġgeç": 76111, + "Ġgeçen": 132762, + "Ġgeçir": 130516, + "ĠgeçmiÅŁ": 137031, + "Ġgf": 41888, + "Ġgfx": 39478, + "Ġgg": 52034, + "Ġgh": 36124, + "Ġghetto": 98321, + "Ġghost": 19866, + "Ġghosts": 53840, + "Ġghé": 136639, + "Ġghế": 136221, + "Ġgi": 13535, + "Ġgia": 94995, + "Ġgiant": 14538, + "Ġgiants": 45209, + "Ġgib": 77327, + "Ġgibi": 67845, + "Ġgibt": 27298, + "Ġgid": 45863, + "Ġgif": 38466, + "Ġgifs": 95913, + "Ġgift": 8189, + "Ġgifted": 46780, + "Ġgifts": 20609, + "Ġgig": 22583, + "Ġgigantic": 57873, + "Ġgigg": 69618, + "Ġgigs": 77820, + "Ġgilt": 69499, + "Ġgim": 83258, + "Ġgimm": 74773, + "Ġgin": 46183, + "Ġging": 66093, + "Ġginger": 43668, + "Ġgint": 48474, + "Ġgio": 53043, + "Ġgioc": 91056, + "Ġgiochi": 72413, + "Ġgiorn": 59010, + "Ġgiorni": 85730, + "Ġgiorno": 80068, + "Ġgiov": 71523, + "Ġgir": 40289, + "ĠgiriÅŁ": 131066, + "Ġgirl": 3743, + "Ġgirlfriend": 22761, + "Ġgirlfriends": 84512, + "Ġgirls": 7571, + "Ġgist": 90500, + "Ġgit": 16345, + "Ġgithub": 31004, + "Ġgive": 2968, + "Ġgiveaway": 59964, + "Ġgiveaways": 86226, + "Ġgiven": 2661, + "Ġgiver": 94515, + "Ġgives": 6696, + "Ġgiving": 7086, + "Ġgiá": 128309, + "Ġgiác": 130780, + "Ġgiám": 133804, + "Ġgiáo": 129023, + "Ġgiãn": 143164, + "Ġgió": 135963, + "Ġgiú": 128429, + "Ġgiúp": 128430, + "ĠgiÃł": 71452, + "ĠgiÃłnh": 139833, + "ĠgiÃłu": 136949, + "Ġgiưá»Ŀng": 140744, + "Ġgiả": 126359, + "Ġgiải": 128486, + "Ġgiảm": 128757, + "Ġgiản": 130631, + "Ġgiảng": 137384, + "Ġgiấ": 141373, + "Ġgiấc": 141374, + "Ġgiấy": 131826, + "Ġgiết": 139950, + "ĠgiáºŃn": 142481, + "Ġgiữ": 129566, + "Ġgiữa": 129181, + "Ġgiá»įng": 140139, + "Ġgiá»ıi": 141734, + "Ġgiá»ijng": 130469, + "ĠgiỼi": 128449, + "Ġgiá»Ŀ": 128626, + "Ġgj": 35001, + "Ġgl": 2770, + "ĠglBegin": 93121, + "ĠglBind": 49609, + "ĠglColor": 69836, + "ĠglEnable": 64271, + "ĠglEnd": 91055, + "ĠglGen": 92849, + "ĠglGet": 50951, + "ĠglGetUniformLocation": 92720, + "ĠglUniform": 76110, + "ĠglVertex": 31128, + "Ġglac": 79516, + "Ġglacier": 93767, + "Ġglaciers": 94690, + "Ġglad": 15713, + "Ġgladly": 70850, + "Ġglam": 41012, + "Ġglamorous": 75914, + "Ġglamour": 78949, + "Ġglance": 33422, + "Ġglanced": 64557, + "Ġgland": 66069, + "Ġglands": 81275, + "Ġglare": 80134, + "Ġglaring": 71121, + "Ġglass": 8991, + "Ġglasses": 28147, + "Ġglazed": 92186, + "Ġgle": 29003, + "Ġglean": 94016, + "Ġgleich": 55248, + "Ġglfw": 50829, + "Ġgli": 27932, + "Ġglide": 85041, + "Ġglimps": 93062, + "Ġglimpse": 39842, + "Ġglitch": 61084, + "Ġglitches": 96380, + "Ġglitter": 54151, + "Ġglm": 17336, + "Ġglo": 13701, + "Ġglob": 13206, + "Ġglobal": 3644, + "Ġglobalization": 79969, + "Ġglobally": 30450, + "Ġglobals": 37785, + "Ġglobe": 23918, + "Ġglor": 57035, + "Ġglorious": 43208, + "Ġglory": 25936, + "Ġgloss": 35351, + "Ġglossy": 73056, + "Ġglove": 55420, + "Ġgloves": 35416, + "Ġglow": 35966, + "Ġglowing": 48492, + "Ġglu": 77276, + "Ġgluc": 29117, + "Ġglucose": 33223, + "Ġglue": 37461, + "Ġglued": 78115, + "Ġglut": 34069, + "Ġgluten": 37355, + "Ġgly": 36707, + "Ġglyc": 71057, + "Ġglyph": 32150, + "Ġglyphicon": 21997, + "Ġglyphs": 84277, + "Ġgm": 37879, + "Ġgmail": 58805, + "Ġgn": 38619, + "Ġgnome": 91061, + "Ġgo": 728, + "ĠgoTo": 77597, + "Ġgoal": 5795, + "Ġgoalie": 73333, + "Ġgoalkeeper": 62200, + "Ġgoals": 8845, + "Ġgoalt": 79622, + "Ġgoat": 53292, + "Ġgoats": 70832, + "Ġgob": 74417, + "Ġgobierno": 80793, + "Ġgod": 9886, + "Ġgoddess": 51090, + "Ġgode": 81257, + "Ġgods": 28813, + "Ġgodt": 80223, + "Ġgoed": 46401, + "Ġgoede": 98560, + "Ġgoes": 5780, + "Ġgoggles": 93990, + "Ġgoing": 2087, + "Ġgol": 44449, + "Ġgold": 6623, + "Ġgolden": 20748, + "Ġgolf": 19120, + "Ġgon": 63492, + "Ġgone": 8048, + "Ġgonna": 16519, + "Ġgoo": 79466, + "Ġgood": 1661, + "Ġgoodbye": 46455, + "Ġgoodies": 69755, + "Ġgoodness": 38426, + "Ġgoods": 11561, + "Ġgoodwill": 81646, + "Ġgoof": 70666, + "Ġgoofy": 99094, + "Ġgoog": 47469, + "Ġgoogle": 11558, + "Ġgoose": 62137, + "Ġgor": 45198, + "Ġgord": 93813, + "Ġgore": 90432, + "Ġgorge": 97377, + "Ġgorgeous": 23511, + "Ġgospel": 41482, + "Ġgossip": 57487, + "Ġgost": 67934, + "Ġgot": 2684, + "Ġgoto": 7986, + "Ġgotta": 36923, + "Ġgotten": 17019, + "Ġgou": 92036, + "Ġgourmet": 72485, + "Ġgouver": 84082, + "Ġgov": 47625, + "Ġgover": 2318, + "Ġgovern": 2569, + "Ġgovernance": 34086, + "Ġgoverned": 26702, + "Ġgoverning": 10012, + "Ġgovernment": 3033, + "Ġgovernmental": 51711, + "Ġgovernments": 16636, + "Ġgoverno": 93845, + "Ġgovernor": 19044, + "Ġgovernors": 67040, + "Ġgown": 64410, + "Ġgoût": 141423, + "ĠgoÅĽci": 137148, + "Ġgp": 28705, + "Ġgpio": 31622, + "Ġgpointer": 58398, + "Ġgps": 45567, + "Ġgpu": 38434, + "Ġgql": 53045, + "Ġgr": 1081, + "Ġgra": 27599, + "Ġgrab": 11633, + "Ġgrabbed": 29318, + "Ġgrabbing": 48930, + "Ġgrabs": 48055, + "Ġgrac": 91466, + "Ġgrace": 20839, + "Ġgraceful": 77362, + "Ġgracefully": 78499, + "Ġgracias": 66548, + "Ġgracious": 81326, + "Ġgrad": 5989, + "Ġgrade": 11972, + "Ġgraded": 79173, + "Ġgrades": 27611, + "Ġgradient": 20169, + "Ġgradients": 52149, + "Ġgrading": 65188, + "Ġgrado": 77645, + "Ġgrads": 69481, + "Ġgradu": 26934, + "Ġgradual": 52622, + "Ġgradually": 26024, + "Ġgraduate": 19023, + "Ġgraduated": 32009, + "Ġgraduates": 37481, + "Ġgraduating": 55917, + "Ġgraduation": 38454, + "Ġgraf": 58300, + "Ġgraffiti": 64843, + "Ġgraft": 68313, + "Ġgrain": 23925, + "Ġgrains": 40836, + "Ġgram": 23018, + "Ġgramm": 68125, + "Ġgrammar": 31428, + "Ġgrams": 33319, + "Ġgran": 15732, + "Ġgrand": 6662, + "Ġgrandchildren": 49751, + "Ġgranddaughter": 91390, + "Ġgrande": 24604, + "Ġgrandes": 37446, + "Ġgrandfather": 37850, + "Ġgrandi": 89774, + "Ġgrandma": 82677, + "Ġgrandmother": 38184, + "Ġgrandparents": 55335, + "Ġgrands": 93427, + "Ġgrandson": 65855, + "Ġgranite": 37290, + "Ġgranny": 62395, + "Ġgrant": 13198, + "ĠgrantResults": 79693, + "Ġgranted": 11676, + "Ġgranting": 48680, + "Ġgrants": 24108, + "Ġgranularity": 95997, + "Ġgrap": 67983, + "Ġgrape": 51152, + "Ġgrapes": 64908, + "Ġgraph": 4771, + "Ġgraphene": 65092, + "Ġgraphic": 20514, + "Ġgraphical": 53143, + "Ġgraphics": 14181, + "Ġgraphite": 94173, + "Ġgraphql": 48865, + "Ġgraphs": 38999, + "Ġgrappling": 89734, + "Ġgras": 48987, + "Ġgrasp": 33377, + "Ġgrass": 16359, + "Ġgrassroots": 56093, + "Ġgrat": 8728, + "Ġgrate": 85354, + "Ġgrated": 96176, + "Ġgrateful": 25195, + "Ġgratis": 7126, + "Ġgratitude": 45035, + "Ġgratuit": 11382, + "Ġgratuita": 95775, + "Ġgratuite": 47309, + "Ġgratuitement": 91934, + "Ġgratuites": 86746, + "Ġgratuiti": 95007, + "Ġgratuito": 77063, + "Ġgratuits": 75523, + "Ġgrav": 27959, + "Ġgrave": 24190, + "Ġgravel": 41523, + "Ġgraves": 61181, + "Ġgraveyard": 71833, + "Ġgravid": 89554, + "Ġgravitational": 69919, + "Ġgravity": 23249, + "Ġgravy": 87764, + "Ġgray": 17545, + "Ġgrayscale": 77357, + "Ġgraz": 50833, + "Ġgrazing": 83450, + "Ġgrd": 87560, + "Ġgre": 2816, + "Ġgrease": 62371, + "Ġgreat": 2244, + "Ġgreater": 7046, + "Ġgreatest": 12196, + "Ġgreatly": 18875, + "Ġgreatness": 62456, + "Ġgreed": 55826, + "Ġgreedy": 55980, + "Ġgreen": 6176, + "Ġgreenhouse": 36547, + "Ġgreens": 50911, + "Ġgreet": 40786, + "Ġgreeted": 43765, + "Ġgreeting": 42113, + "Ġgreetings": 94085, + "Ġgren": 37463, + "Ġgrenade": 66581, + "Ġgrenades": 78959, + "Ġgrep": 20675, + "Ġgrew": 13938, + "Ġgrey": 19780, + "Ġgri": 22993, + "Ġgrid": 5827, + "ĠgridBagConstraints": 54259, + "ĠgridColumn": 93699, + "ĠgridSize": 97511, + "ĠgridView": 69044, + "Ġgrids": 56349, + "Ġgrief": 37284, + "Ġgriev": 49913, + "Ġgrievances": 86936, + "Ġgrieving": 90744, + "Ġgrill": 39058, + "Ġgrille": 86825, + "Ġgrilled": 60633, + "Ġgrim": 43417, + "Ġgrin": 58280, + "Ġgrind": 39236, + "Ġgrinder": 45541, + "Ġgrinding": 16857, + "Ġgrinned": 87279, + "Ġgrip": 24676, + "Ġgripping": 80191, + "Ġgrips": 66239, + "Ġgris": 85830, + "Ġgrit": 40918, + "Ġgritty": 85678, + "Ġgro": 10487, + "Ġgroceries": 65408, + "Ġgrocery": 29587, + "Ġgroin": 97775, + "Ġgroom": 39841, + "Ġgrooming": 74061, + "Ġgroot": 81467, + "Ġgroove": 56406, + "Ġgrop": 91341, + "Ġgros": 24490, + "Ġgross": 19952, + "Ġgrosse": 44271, + "Ġgrote": 43446, + "Ġgrotes": 95179, + "Ġgrou": 93768, + "Ġground": 4910, + "Ġgroundbreaking": 63855, + "Ġgrounded": 44345, + "Ġgrounding": 94741, + "Ġgrounds": 20664, + "Ġgroundwater": 71229, + "Ġgroundwork": 94744, + "Ġgroup": 1874, + "ĠgroupBox": 42734, + "ĠgroupId": 41797, + "ĠgroupName": 65535, + "Ġgroupe": 56851, + "Ġgrouped": 40041, + "Ġgrouping": 49184, + "Ġgroups": 5203, + "Ġgrow": 3063, + "Ġgrowers": 68977, + "Ġgrowing": 7826, + "Ġgrown": 14700, + "Ġgrows": 27715, + "Ġgrowth": 6513, + "ĠgroÃŁ": 66723, + "ĠgroÃŁe": 56381, + "ĠgroÃŁen": 67703, + "Ġgrp": 51859, + "Ġgrpc": 47900, + "Ġgrub": 73706, + "Ġgrues": 75057, + "Ġgruesome": 85928, + "Ġgrund": 93434, + "Ġgrunt": 43043, + "Ġgrup": 33570, + "Ġgrupo": 37510, + "Ġgrupos": 72924, + "Ġgrátis": 74809, + "Ġgrâce": 76333, + "Ġgrö": 59409, + "ĠgrÃ¶ÃŁ": 74539, + "ĠgrÃ¶ÃŁer": 140370, + "Ġgs": 28081, + "Ġgsi": 96970, + "Ġgsl": 44221, + "Ġgson": 43066, + "Ġgst": 43769, + "Ġgt": 25161, + "Ġgtk": 18692, + "Ġgu": 1675, + "Ġguar": 9352, + "Ġguarante": 15389, + "Ġguarantee": 15440, + "Ġguaranteed": 19323, + "Ġguarantees": 35655, + "Ġguard": 7616, + "Ġguarda": 97385, + "Ġguardar": 63100, + "Ġguarded": 63163, + "Ġguardian": 46032, + "Ġguardians": 71781, + "Ġguarding": 79500, + "Ġguards": 26178, + "Ġgubern": 90422, + "Ġguerr": 75170, + "Ġguerra": 84927, + "Ġguess": 7942, + "Ġguessed": 42310, + "Ġguesses": 60537, + "Ġguessing": 46476, + "Ġguest": 8640, + "Ġguests": 14709, + "Ġgui": 19229, + "ĠguiActive": 70827, + "Ġguid": 9643, + "Ġguidance": 18821, + "Ġguide": 8474, + "Ġguided": 32587, + "Ġguideline": 72445, + "Ġguidelines": 17501, + "Ġguides": 27193, + "Ġguiding": 50246, + "Ġguild": 26411, + "Ġguilt": 33851, + "Ġguilty": 16007, + "Ġguint": 51726, + "Ġguise": 83063, + "Ġguit": 54399, + "Ġguitar": 16986, + "Ġguitarist": 61640, + "Ġguitars": 59494, + "Ġgul": 90270, + "Ġgulp": 29779, + "Ġgum": 41265, + "Ġgums": 97654, + "Ġgun": 6038, + "Ġgunfire": 76795, + "Ġgunman": 62097, + "Ġgunmen": 97668, + "Ġguns": 16362, + "Ġgunshot": 73781, + "Ġguru": 59426, + "Ġgust": 29814, + "Ġgusta": 69658, + "Ġgusto": 92239, + "Ġgut": 17859, + "Ġgute": 62702, + "Ġguten": 93615, + "Ġguts": 64873, + "Ġgutter": 48979, + "Ġguy": 7412, + "Ġguys": 7598, + "Ġgv": 54979, + "Ġgw": 56572, + "Ġgx": 61033, + "Ġgy": 28620, + "Ġgym": 18813, + "Ġgymn": 69121, + "Ġgypsum": 87580, + "Ġgyr": 79505, + "Ġgyro": 66868, + "Ġgz": 76963, + "Ġgzip": 57795, + "Ġgái": 129826, + "Ġgây": 128882, + "ĠgÃ¥": 53974, + "ĠgÃ¥r": 70757, + "Ġgé": 73572, + "Ġgén": 77322, + "Ġgénero": 142356, + "Ġgéné": 45580, + "Ġgénér": 135862, + "Ġgénéral": 88075, + "Ġgì": 128382, + "Ġgó": 131071, + "Ġgóc": 138456, + "Ġgói": 135974, + "Ġgóp": 131072, + "Ġgór": 132575, + "Ġgö": 30185, + "Ġgönder": 127520, + "Ġgönderil": 140693, + "Ġgör": 57704, + "Ġgörd": 128056, + "Ġgördü": 139449, + "ĠgördÃ¼ÄŁÃ¼": 142023, + "Ġgöre": 128569, + "Ġgörev": 129864, + "Ġgörm": 139767, + "Ġgörme": 136734, + "Ġgörmek": 139768, + "Ġgörü": 133037, + "Ġgörül": 131217, + "Ġgörün": 131389, + "Ġgörünt": 133038, + "Ġgörüntü": 133039, + "Ġgörüyor": 142301, + "ĠgörÃ¼ÅŁ": 127435, + "ĠgörÃ¼ÅŁme": 137857, + "Ġgöster": 128689, + "Ġgötür": 140438, + "Ġgöz": 126465, + "Ġgözü": 141667, + "Ġgü": 53908, + "Ġgüc": 137171, + "Ġgücü": 137172, + "Ġgün": 87687, + "Ġgünc": 137724, + "Ġgüncel": 137725, + "Ġgünd": 140070, + "Ġgündem": 140071, + "Ġgüne": 142397, + "ĠgüneÅŁ": 142398, + "Ġgünl": 139530, + "Ġgünlük": 139531, + "Ġgünü": 131214, + "Ġgünümüz": 143113, + "Ġgüven": 126802, + "Ġgüvenlik": 138655, + "Ġgüzel": 129032, + "Ġgüç": 127305, + "Ġgüçlü": 136735, + "ĠgÃł": 134530, + "Ġgı": 143677, + "Ġgıda": 143678, + "ĠgÅĤ": 65689, + "ĠgÅĤów": 143811, + "ĠgÅĤówn": 128127, + "ĠgÅĤówna": 140541, + "ĠgÅĤówne": 143812, + "ĠgÅĤównie": 141329, + "Ġgương": 139227, + "Ġgạo": 143692, + "Ġgấp": 139831, + "Ġgần": 128914, + "Ġgắn": 136822, + "Ġgắng": 137827, + "Ġgặp": 129233, + "Ġgợi": 143252, + "Ġgá»įi": 129349, + "Ġgá»įn": 138345, + "Ġgá»ijc": 134766, + "Ġgá»ĵ": 130067, + "Ġgá»ĵm": 130068, + "Ġgá»Ĺ": 133105, + "Ġgá»Ń": 130505, + "Ġgá»Ńi": 130506, + "Ġh": 305, + "ĠhObject": 69770, + "ĠhWnd": 58579, + "Ġha": 6386, + "Ġhaar": 41125, + "Ġhab": 6055, + "Ġhabe": 27726, + "Ġhaben": 17630, + "Ġhaber": 58282, + "Ġhabil": 94215, + "Ġhabit": 14132, + "Ġhabitat": 38546, + "Ġhabitats": 70599, + "Ġhabits": 25785, + "Ġhabitual": 83720, + "Ġhablar": 95743, + "ĠhabÃŃa": 54038, + "Ġhac": 53695, + "Ġhace": 34805, + "Ġhacen": 90965, + "Ġhacer": 25309, + "Ġhaci": 76903, + "Ġhacia": 56909, + "Ġhaciendo": 88053, + "Ġhack": 17086, + "Ġhacked": 46839, + "Ġhacker": 54122, + "Ġhackers": 40367, + "Ġhacking": 38138, + "Ġhacks": 59784, + "Ġhad": 1030, + "Ġhadde": 66334, + "Ġhade": 96884, + "Ġhadn": 18604, + "Ġhaha": 62350, + "Ġhai": 46051, + "Ġhail": 64250, + "Ġhailed": 63795, + "Ġhair": 6869, + "Ġhaircut": 85724, + "Ġhaired": 86102, + "Ġhairs": 64960, + "Ġhairst": 55284, + "Ġhairstyle": 95135, + "Ġhairstyles": 88772, + "Ġhairy": 50033, + "Ġhak": 97659, + "Ġhakk": 126779, + "Ġhakkı": 135932, + "Ġhakkında": 130023, + "Ġhal": 14761, + "Ġhalf": 4279, + "Ġhalftime": 78459, + "Ġhalfway": 41336, + "Ġhalk": 142183, + "Ġhalkın": 142184, + "Ġhall": 13994, + "Ġhallmark": 97699, + "Ġhalls": 51373, + "Ġhalluc": 58023, + "Ġhallway": 50802, + "Ġhalo": 70712, + "Ġhalt": 26269, + "Ġhalted": 60095, + "Ġhalten": 95723, + "Ġhalves": 74112, + "Ġham": 13515, + "Ġhamburg": 56847, + "Ġhamburger": 88747, + "Ġhamm": 82565, + "Ġhammer": 23457, + "Ġhammered": 92024, + "Ġhamstring": 89705, + "Ġhan": 19001, + "Ġhand": 1424, + "Ġhandbook": 75249, + "Ġhandc": 65045, + "Ġhanded": 22593, + "Ġhandful": 22955, + "Ġhandgun": 56805, + "Ġhandguns": 98536, + "Ġhandheld": 64949, + "Ġhandic": 47170, + "Ġhandicap": 74279, + "Ġhanding": 51488, + "Ġhandjob": 90641, + "Ġhandle": 3705, + "ĠhandleChange": 38655, + "ĠhandleClick": 52440, + "ĠhandleClose": 81567, + "ĠhandleError": 68200, + "ĠhandleMessage": 96840, + "ĠhandleSubmit": 42265, + "Ġhandled": 17608, + "Ġhandler": 7013, + "Ġhandlers": 24083, + "Ġhandles": 13469, + "Ġhandling": 11589, + "Ġhandmade": 51686, + "Ġhands": 6078, + "Ġhandset": 82163, + "Ġhandshake": 57020, + "Ġhandsome": 43777, + "Ġhandwriting": 91709, + "Ġhandwritten": 98389, + "Ġhandy": 25163, + "Ġhang": 14678, + "Ġhanging": 20704, + "Ġhangs": 58341, + "Ġhann": 92790, + "Ġhanno": 47784, + "Ġhans": 87544, + "Ġhanya": 69837, + "Ġhap": 45800, + "Ġhapp": 2255, + "Ġhappen": 3537, + "Ġhappened": 6932, + "Ġhappening": 12482, + "Ġhappens": 8573, + "Ġhappier": 43367, + "Ġhappiest": 98852, + "Ġhappily": 36775, + "Ġhappiness": 23009, + "Ġhappy": 6247, + "Ġhar": 4855, + "Ġharass": 18737, + "Ġharassed": 67510, + "Ġharassing": 86170, + "Ġharassment": 25355, + "Ġharb": 68466, + "Ġharbor": 56411, + "Ġharbour": 74642, + "Ġhard": 2588, + "Ġhardcoded": 92959, + "Ġhardcore": 26604, + "Ġhardened": 70736, + "Ġharder": 15750, + "Ġhardest": 36454, + "Ġhardly": 20171, + "Ġhardness": 70903, + "Ġhardship": 66550, + "Ġhardships": 95222, + "Ġhardware": 11773, + "Ġhardwood": 50545, + "Ġhare": 94918, + "Ġharga": 71348, + "Ġhari": 69696, + "Ġharm": 11428, + "Ġharmed": 68425, + "Ġharmful": 27756, + "Ġharming": 89155, + "Ġharmless": 52897, + "Ġharmon": 17774, + "Ġharmonic": 81241, + "Ġharmony": 25281, + "Ġharms": 83396, + "Ġharness": 32408, + "Ġharsh": 24939, + "Ġhart": 73928, + "Ġharus": 58497, + "Ġharvest": 23429, + "Ġharvested": 66472, + "Ġharvesting": 65647, + "Ġhas": 702, + "ĠhasNext": 82624, + "Ġhasattr": 31056, + "Ġhash": 5175, + "ĠhashCode": 29444, + "ĠhashMap": 67959, + "Ġhashed": 50676, + "Ġhasher": 90819, + "Ġhashes": 50257, + "Ġhashing": 72355, + "Ġhashlib": 54495, + "Ġhashmap": 92148, + "Ġhashtable": 82110, + "Ġhashtag": 47077, + "Ġhashtags": 81861, + "Ġhasil": 75043, + "Ġhasn": 12492, + "Ġhass": 36404, + "Ġhassle": 46847, + "Ġhast": 33043, + "Ġhasta": 28526, + "Ġhastalık": 137803, + "Ġhaste": 89439, + "Ġhastily": 93432, + "Ġhat": 8896, + "Ġhatch": 43255, + "Ġhate": 12213, + "Ġhated": 37574, + "Ġhateful": 75715, + "Ġhates": 54306, + "Ġhath": 51577, + "Ġhatred": 34142, + "Ġhats": 44426, + "Ġhatte": 42668, + "Ġhatten": 83572, + "Ġhatırl": 138834, + "Ġhatırla": 138835, + "Ġhaul": 37868, + "Ġhauling": 98446, + "Ġhaunt": 83658, + "Ġhaunted": 56675, + "Ġhaunting": 76654, + "Ġhaus": 70272, + "Ġhaut": 61221, + "Ġhaute": 79998, + "Ġhav": 29981, + "Ġhave": 614, + "Ġhaven": 8990, + "Ġhaving": 3432, + "Ġhavoc": 65492, + "Ġhawk": 75820, + "Ġhay": 17669, + "Ġhaya": 79521, + "Ġhayat": 127672, + "Ġhayatı": 134797, + "Ġhayatını": 140860, + "Ġhaystack": 88447, + "Ġhaz": 20144, + "Ġhazard": 30297, + "Ġhazardous": 49924, + "Ġhazards": 51542, + "Ġhaze": 89309, + "Ġhazır": 130348, + "Ġhazırl": 135754, + "Ġhazırlan": 135755, + "Ġhb": 45135, + "Ġhbox": 96782, + "Ġhc": 50394, + "Ġhd": 17907, + "Ġhdc": 85032, + "Ġhdf": 97045, + "Ġhdr": 36615, + "Ġhe": 566, + "Ġhead": 1968, + "Ġheadache": 46746, + "Ġheadaches": 54303, + "Ġheaded": 19383, + "Ġheader": 4247, + "Ġheaders": 7102, + "Ġheading": 14496, + "Ġheadings": 69452, + "Ġheadlights": 87620, + "Ġheadline": 31163, + "Ġheadlines": 30086, + "Ġheadphone": 76401, + "Ġheadphones": 43001, + "Ġheadquartered": 80196, + "Ġheadquarters": 25042, + "Ġheads": 14629, + "Ġheadset": 49579, + "Ġheal": 26563, + "Ġhealed": 61220, + "Ġhealer": 97111, + "Ġhealing": 21052, + "Ġheals": 86984, + "Ġhealth": 2820, + "Ġhealthcare": 18478, + "Ġhealthier": 38245, + "Ġhealthy": 9314, + "Ġheap": 17364, + "Ġheapq": 88522, + "Ġheaps": 92640, + "Ġhear": 6723, + "Ġheard": 6617, + "Ġhearing": 10778, + "Ġhearings": 44248, + "Ġhears": 52059, + "Ġheart": 4746, + "Ġheartbeat": 52105, + "Ġheartbreaking": 80106, + "Ġheartfelt": 83111, + "Ġhearts": 22662, + "Ġhearty": 81551, + "Ġheat": 8628, + "Ġheated": 31713, + "Ġheater": 42238, + "Ġheaters": 81367, + "Ġheating": 23579, + "Ġheatmap": 96623, + "Ġheats": 76562, + "Ġheav": 13405, + "Ġheaven": 22274, + "Ġheavenly": 66598, + "Ġheavens": 62780, + "Ġheavier": 43822, + "Ġheavily": 16916, + "Ġheavy": 8811, + "Ġheavyweight": 70629, + "Ġheb": 58141, + "Ġhebben": 39758, + "Ġhebt": 88083, + "Ġhecho": 49302, + "Ġheck": 46681, + "Ġhect": 47001, + "Ġhectares": 80643, + "Ġhectic": 90325, + "Ġhed": 60216, + "Ġhedge": 41666, + "Ġheed": 74399, + "Ġheeft": 37398, + "Ġheel": 34328, + "Ġheels": 33360, + "Ġhefty": 64142, + "Ġheg": 89403, + "Ġhei": 64489, + "Ġheight": 2608, + "ĠheightFor": 68349, + "Ġheightened": 59387, + "Ġheights": 35294, + "Ġheir": 50443, + "Ġheirs": 90301, + "ĠheiÃŁ": 79536, + "ĠheiÃŁt": 140385, + "Ġhel": 11338, + "Ġheld": 5644, + "Ġhele": 54095, + "Ġhelf": 74403, + "Ġhelfen": 82197, + "Ġhelicopt": 27367, + "Ġhelicopter": 35025, + "Ġhelicopters": 58332, + "Ġhelium": 96507, + "Ġhell": 14780, + "Ġhello": 23811, + "Ġhelm": 33765, + "Ġhelmet": 31535, + "Ġhelmets": 64620, + "Ġhelp": 1492, + "Ġhelped": 8910, + "Ġhelper": 13137, + "Ġhelpers": 30187, + "Ġhelpful": 10950, + "Ġhelping": 10476, + "Ġhelpless": 62528, + "Ġhelps": 8609, + "Ġhelt": 56442, + "Ġhely": 90115, + "Ġhem": 17280, + "Ġhemat": 95749, + "Ġhemisphere": 68666, + "Ġhemorrh": 70317, + "Ġhemos": 72257, + "Ġhemp": 40963, + "Ġhen": 47816, + "Ġhence": 16085, + "Ġhend": 84379, + "Ġhenne": 91617, + "Ġhentai": 45301, + "Ġhenüz": 141457, + "Ġhep": 79345, + "Ġhepat": 57159, + "Ġhepatitis": 72875, + "Ġher": 1059, + "Ġherald": 64106, + "Ġheraus": 83380, + "Ġherb": 38899, + "Ġherbal": 50387, + "Ġherbs": 41293, + "Ġherd": 58113, + "Ġhere": 1588, + "Ġhereby": 21799, + "Ġherein": 35781, + "Ġheritage": 27848, + "Ġhern": 71809, + "Ġhero": 11821, + "Ġheroes": 22907, + "Ġheroic": 55275, + "Ġheroin": 41634, + "Ġheroine": 80087, + "Ġherpes": 68985, + "Ġherr": 54739, + "Ġherramient": 81990, + "Ġhers": 10837, + "Ġherself": 11675, + "Ġhes": 19779, + "Ġhesabı": 139326, + "Ġhesitant": 70460, + "Ġhesitate": 38566, + "Ġhesitation": 64337, + "Ġhet": 9017, + "Ġhete": 82118, + "Ġheter": 29448, + "Ġheterogeneous": 97782, + "Ġheterosexual": 65308, + "Ġheure": 76718, + "Ġheures": 75045, + "Ġheuristic": 66609, + "Ġheute": 48614, + "Ġhex": 12371, + "Ġhexadecimal": 69722, + "Ġhexatrigesimal": 57966, + "Ġhey": 34209, + "Ġhf": 43242, + "Ġhg": 73294, + "Ġhh": 37914, + "Ġhi": 15588, + "Ġhiatus": 80186, + "Ġhic": 77907, + "Ġhid": 30794, + "Ġhidden": 8177, + "Ġhide": 10265, + "Ġhides": 57604, + "Ġhiding": 25037, + "Ġhier": 12412, + "Ġhierarchical": 69894, + "Ġhierarchy": 28922, + "Ġhieronta": 80203, + "Ġhigh": 1550, + "Ġhigher": 5080, + "Ġhighest": 8426, + "Ġhighlight": 11167, + "Ġhighlighted": 26366, + "Ġhighlighting": 38586, + "Ġhighlights": 21314, + "Ġhighly": 7548, + "Ġhighs": 53779, + "Ġhighway": 26736, + "Ġhighways": 59295, + "Ġhij": 21415, + "Ġhijo": 80045, + "Ġhijos": 88527, + "Ġhike": 34231, + "Ġhikes": 67485, + "Ġhiking": 37364, + "Ġhil": 51327, + "Ġhilar": 35659, + "Ġhilarious": 40267, + "Ġhill": 23946, + "Ġhills": 34131, + "Ġhim": 1435, + "Ġhimself": 5561, + "Ġhin": 51721, + "Ġhind": 47319, + "Ġhinder": 56680, + "Ġhindi": 52017, + "Ġhindsight": 86294, + "Ġhinge": 75593, + "Ġhinges": 79834, + "Ġhint": 13012, + "ĠhintText": 66228, + "Ġhinted": 66369, + "Ġhinter": 71968, + "Ġhints": 30643, + "Ġhip": 18143, + "Ġhipp": 43556, + "Ġhippoc": 70106, + "Ġhips": 42423, + "Ġhire": 17983, + "Ġhired": 21446, + "Ġhires": 71941, + "Ġhiring": 23134, + "Ġhis": 806, + "Ġhist": 12744, + "Ġhisto": 76517, + "Ġhistogram": 30281, + "Ġhistograms": 85645, + "Ġhistoire": 86710, + "Ġhistor": 6990, + "Ġhistoria": 50322, + "Ġhistorian": 42968, + "Ġhistorians": 50701, + "Ġhistoric": 18035, + "Ġhistorical": 13656, + "Ġhistorically": 34801, + "Ġhistories": 52110, + "Ġhistory": 3840, + "Ġhistó": 78486, + "Ġhistória": 93209, + "Ġhistórico": 139363, + "Ġhit": 4201, + "Ġhitch": 58107, + "Ġhits": 12983, + "Ġhitter": 71307, + "Ġhitters": 80033, + "Ġhitting": 19558, + "Ġhive": 65507, + "Ġhizo": 85183, + "Ġhiç": 126922, + "Ġhiçbir": 132303, + "Ġhiếm": 143799, + "Ġhiá»ĥm": 131491, + "Ġhiá»ĥn": 141280, + "Ġhiá»ĥu": 129061, + "Ġhiá»ĩn": 128283, + "Ġhiá»ĩp": 141741, + "Ġhiá»ĩu": 128423, + "Ġhj": 35588, + "Ġhjem": 66985, + "Ġhk": 83381, + "Ġhl": 49586, + "Ġhm": 49362, + "Ġhmac": 83356, + "Ġhmm": 87501, + "Ġhn": 95238, + "Ġho": 11386, + "Ġhoax": 72844, + "Ġhob": 68076, + "Ġhobbies": 63305, + "Ġhobby": 31528, + "Ġhoc": 66390, + "Ġhoch": 60021, + "Ġhockey": 26953, + "Ġhod": 86803, + "Ġhoe": 45876, + "Ġhog": 45925, + "Ġhogy": 32406, + "Ġhoje": 73158, + "Ġhol": 23523, + "Ġhold": 3331, + "Ġholder": 11839, + "Ġholders": 28709, + "Ġholding": 9963, + "Ġholdings": 57248, + "Ġholds": 9982, + "Ġhole": 14178, + "Ġholes": 19766, + "Ġholiday": 13257, + "Ġholidays": 24421, + "Ġholistic": 60776, + "Ġholland": 64664, + "Ġhollow": 41802, + "Ġholog": 71827, + "Ġholster": 97951, + "Ġholy": 26725, + "Ġhom": 4998, + "Ġhomage": 66989, + "Ġhombre": 55704, + "Ġhombres": 61844, + "Ġhome": 2114, + "Ġhomeland": 55236, + "Ġhomeless": 22994, + "Ġhomelessness": 61922, + "Ġhomem": 73415, + "Ġhomemade": 35333, + "Ġhomeowner": 66698, + "Ġhomeowners": 39862, + "Ġhomepage": 34589, + "Ġhomer": 91305, + "Ġhomers": 97598, + "Ġhomes": 10415, + "Ġhomeschool": 89819, + "Ġhometown": 43016, + "Ġhomework": 28459, + "Ġhomic": 36963, + "Ġhomicide": 48977, + "Ġhomicides": 88395, + "Ġhomme": 39251, + "Ġhommes": 52442, + "Ġhomo": 54413, + "Ġhomogeneous": 86182, + "Ġhomophobic": 92418, + "Ġhomosex": 27308, + "Ġhomosexual": 41764, + "Ġhomosexuality": 52351, + "Ġhomosexuals": 96123, + "Ġhon": 5831, + "Ġhonda": 84376, + "Ġhone": 51022, + "Ġhonest": 10745, + "Ġhonestly": 26044, + "Ġhonesty": 47848, + "Ġhoney": 25744, + "Ġhoneymoon": 83793, + "Ġhonor": 15669, + "Ġhonorable": 73490, + "Ġhonorary": 98019, + "Ġhonored": 38277, + "Ġhonoring": 77925, + "Ġhonors": 48493, + "Ġhonour": 33562, + "Ġhonoured": 89733, + "Ġhood": 27215, + "Ġhoodie": 96301, + "Ġhoof": 84730, + "Ġhoog": 98126, + "Ġhook": 9704, + "Ġhooked": 42670, + "Ġhookers": 92627, + "Ġhooks": 29677, + "Ġhookup": 61314, + "Ġhoop": 78671, + "Ġhoops": 96015, + "Ġhop": 7446, + "Ġhope": 3900, + "Ġhoped": 25189, + "Ġhopeful": 37550, + "Ġhopefully": 22326, + "Ġhopeless": 74223, + "Ġhopes": 16005, + "Ġhoping": 15652, + "Ġhopping": 92238, + "Ġhops": 58118, + "Ġhor": 4812, + "Ġhora": 34120, + "Ġhoras": 38892, + "Ġhorizon": 34074, + "Ġhorizontal": 16202, + "Ġhorizontally": 58888, + "Ġhorm": 20878, + "Ġhormonal": 73118, + "Ġhormone": 35808, + "Ġhormones": 43215, + "Ġhorn": 20631, + "Ġhorns": 59984, + "Ġhorny": 37207, + "Ġhorr": 25887, + "Ġhorrend": 92740, + "Ġhorrible": 27102, + "Ġhorribly": 76004, + "Ġhorrific": 46222, + "Ġhorrified": 81223, + "Ġhorrifying": 77550, + "Ġhorror": 21452, + "Ġhorrors": 69518, + "Ġhors": 59375, + "Ġhorse": 15223, + "Ġhorsepower": 69786, + "Ġhorses": 26830, + "Ġhos": 67827, + "Ġhose": 37500, + "Ġhoses": 95836, + "Ġhosp": 20407, + "Ġhospital": 8777, + "Ġhospitality": 46509, + "Ġhospitalized": 69240, + "Ġhospitals": 23551, + "Ġhost": 3468, + "Ġhostage": 58368, + "Ġhostages": 87603, + "Ġhosted": 21009, + "Ġhostel": 80811, + "Ġhostile": 32515, + "Ġhostility": 59929, + "Ġhosting": 19678, + "Ġhostname": 28115, + "Ġhosts": 18432, + "Ġhot": 4017, + "Ġhotel": 9500, + "Ġhotels": 24332, + "Ġhotline": 91875, + "Ġhotspot": 79210, + "Ġhott": 31522, + "Ġhotter": 78905, + "Ġhottest": 37291, + "Ġhotéis": 143286, + "Ġhour": 6460, + "Ġhourly": 46629, + "Ġhours": 4115, + "Ġhous": 7696, + "Ġhouse": 3753, + "Ġhoused": 51158, + "Ġhousehold": 13732, + "Ġhouseholds": 28839, + "Ġhouses": 14967, + "Ġhousing": 11721, + "Ġhover": 19548, + "Ġhovered": 90451, + "Ġhovering": 68607, + "Ġhow": 1246, + "Ġhowever": 4764, + "Ġhoy": 48741, + "Ġhoá": 137692, + "ĠhoÃłn": 128643, + "ĠhoÃłng": 135704, + "ĠhoÅŁ": 138013, + "Ġhoạch": 130361, + "Ġhoạt": 128623, + "Ġhoặc": 128475, + "Ġhp": 20630, + "Ġhr": 18026, + "Ġhref": 1801, + "Ġhrs": 40040, + "Ġhs": 32220, + "Ġhsv": 82333, + "Ġht": 34323, + "Ġhtml": 5272, + "ĠhtmlFor": 32193, + "Ġhtmlentities": 62998, + "Ġhtmlspecialchars": 31437, + "Ġhton": 81743, + "Ġhtonl": 86982, + "Ġhtons": 56367, + "Ġhtt": 54320, + "Ġhttp": 1758, + "ĠhttpClient": 45775, + "ĠhttpRequest": 77547, + "ĠhttpResponse": 74806, + "Ġhttps": 3703, + "Ġhu": 32965, + "Ġhub": 18719, + "Ġhubby": 97324, + "Ġhubs": 68676, + "Ġhud": 64033, + "Ġhue": 39040, + "Ġhues": 81657, + "Ġhug": 29784, + "Ġhuge": 6765, + "Ġhugely": 48637, + "Ġhugged": 93468, + "Ġhugs": 79242, + "Ġhuh": 56743, + "Ġhuis": 67609, + "Ġhull": 40198, + "Ġhum": 2784, + "Ġhuman": 3738, + "Ġhumane": 81026, + "Ġhumanitarian": 37648, + "Ġhumanities": 93035, + "Ġhumanity": 21941, + "Ġhumano": 96457, + "Ġhumanoid": 92109, + "Ġhumans": 12677, + "Ġhumble": 38512, + "Ġhumid": 65938, + "Ġhumidity": 37093, + "Ġhumili": 82169, + "Ġhumiliating": 91738, + "Ġhumiliation": 71718, + "Ġhumility": 74272, + "Ġhumming": 86327, + "Ġhumor": 27385, + "Ġhumorous": 69846, + "Ġhumour": 57369, + "Ġhun": 11533, + "Ġhundred": 7739, + "Ġhundreds": 11499, + "Ġhung": 18295, + "Ġhunger": 33806, + "Ġhungry": 28956, + "Ġhunt": 19073, + "Ġhunted": 79169, + "Ġhunter": 39727, + "Ġhunters": 47973, + "Ġhunting": 22514, + "Ġhunts": 92829, + "Ġhur": 12819, + "Ġhurd": 47481, + "Ġhurdle": 80683, + "Ġhurdles": 72535, + "Ġhurl": 93731, + "Ġhurricane": 48684, + "Ġhurricanes": 89027, + "Ġhurried": 72873, + "Ġhurry": 47235, + "Ġhurt": 12898, + "Ġhurting": 47289, + "Ġhurts": 43064, + "Ġhus": 9045, + "Ġhusband": 9972, + "Ġhusbands": 60854, + "Ġhust": 58650, + "Ġhustle": 87212, + "Ġhut": 75879, + "Ġhuyết": 135596, + "Ġhuyá»ĩn": 131210, + "Ġhv": 22747, + "Ġhva": 82766, + "Ġhvad": 91491, + "Ġhver": 89666, + "Ġhvis": 77575, + "Ġhvor": 50719, + "Ġhvordan": 62159, + "Ġhw": 31256, + "Ġhwnd": 52832, + "Ġhx": 53214, + "Ġhy": 6275, + "Ġhybrid": 24989, + "Ġhybrids": 96105, + "Ġhydr": 44605, + "Ġhydrated": 94731, + "Ġhydration": 86900, + "Ġhydraulic": 43075, + "Ġhydro": 16643, + "Ġhydrogen": 34684, + "Ġhygiene": 51956, + "Ġhym": 66643, + "Ġhyp": 9751, + "Ġhype": 44890, + "Ġhyper": 17071, + "Ġhyperlink": 93580, + "Ġhypers": 86652, + "Ġhypert": 47755, + "Ġhypertension": 62208, + "Ġhypnot": 75451, + "Ġhypo": 21845, + "Ġhypoc": 66486, + "Ġhypocrisy": 71706, + "Ġhypoth": 44355, + "Ġhypotheses": 73413, + "Ġhypothesis": 30078, + "Ġhypothetical": 58059, + "Ġhyster": 53019, + "Ġhysteria": 97666, + "Ġhá": 42122, + "Ġhát": 133951, + "Ġhã": 126704, + "Ġhãng": 132159, + "Ġhãy": 129805, + "Ġhä": 52656, + "Ġhär": 63686, + "Ġhät": 74492, + "Ġhätte": 95035, + "Ġhäufig": 137943, + "ĠhÃ¥": 53613, + "Ġhè": 136252, + "Ġhé": 87923, + "Ġhình": 128338, + "Ġhòa": 130669, + "Ġhóa": 128585, + "Ġhô": 132658, + "Ġhôm": 130535, + "Ġhôn": 133647, + "Ġhö": 42759, + "Ġhöch": 141963, + "Ġhöchst": 141964, + "Ġhöher": 139554, + "Ġhút": 130862, + "Ġhü": 133862, + "Ġhük": 128114, + "Ġhükü": 140149, + "Ġhüküm": 142799, + "Ġhükümet": 140150, + "ĠhÃł": 130734, + "ĠhÃłi": 133844, + "ĠhÃłng": 61758, + "ĠhÃłnh": 128373, + "Ġhız": 127597, + "Ġhızlı": 134255, + "ĠhÆ¡i": 131671, + "ĠhÆ¡n": 128304, + "Ġhư": 136375, + "Ġhương": 133899, + "ĠhÆ°á»Ľng": 128846, + "ĠhÆ°á»Łng": 129589, + "Ġhạ": 131140, + "Ġhại": 131579, + "Ġhạn": 129572, + "Ġhạng": 135303, + "Ġhạnh": 134127, + "Ġhạt": 134761, + "Ġhải": 138650, + "Ġhảo": 137658, + "Ġhấp": 131768, + "Ġhầu": 136402, + "Ġhắn": 129810, + "Ġhẳ": 138122, + "Ġhẳn": 138123, + "Ġhẹ": 140242, + "Ġhẹn": 140243, + "Ġhết": 128906, + "ĠháºŃu": 130990, + "Ġhợ": 128318, + "Ġhợp": 128319, + "Ġhá»§y": 143010, + "Ġhứng": 141163, + "Ġhữu": 129347, + "Ġhá»ģ": 137738, + "Ġhá»ĩ": 128457, + "Ġhá»į": 126247, + "Ġhá»įc": 128292, + "Ġhá»įp": 136856, + "Ġhá»ı": 143225, + "Ġhá»ıi": 129087, + "Ġhá»ıng": 143226, + "Ġhá»ĵ": 129122, + "Ġhá»ĵi": 130344, + "Ġhá»ĵng": 135342, + "Ġhá»Ĺ": 129481, + "Ġhá»Ļ": 126469, + "Ġhá»Ļi": 128410, + "Ġhá»Ļp": 139127, + "Ġi": 600, + "ĠiCloud": 86911, + "ĠiNdEx": 90443, + "ĠiOS": 16048, + "ĠiP": 76486, + "ĠiPad": 22271, + "ĠiParam": 42048, + "ĠiPhone": 12167, + "ĠiPhones": 72065, + "ĠiPod": 44071, + "ĠiT": 12387, + "ĠiTunes": 13025, + "ĠiVar": 56192, + "Ġia": 34824, + "Ġiam": 97148, + "Ġiar": 87721, + "Ġib": 18668, + "Ġic": 17902, + "Ġice": 9853, + "Ġiceberg": 88035, + "Ġich": 10638, + "Ġici": 48633, + "Ġicing": 75006, + "Ġicmp": 91826, + "Ġicon": 4603, + "ĠiconName": 89877, + "Ġiconic": 26277, + "Ġicons": 23765, + "Ġicy": 65904, + "Ġid": 877, + "Ġidade": 65965, + "Ġide": 2614, + "Ġidea": 4522, + "Ġideal": 10507, + "Ġideally": 49563, + "Ġideals": 51705, + "Ġideas": 6708, + "Ġident": 3524, + "Ġidentical": 19516, + "Ġidentifiable": 71350, + "Ġidentification": 21892, + "Ġidentified": 10820, + "Ġidentifier": 12816, + "Ġidentifiers": 47216, + "Ġidentifies": 35511, + "Ġidentify": 10542, + "Ġidentifying": 24588, + "Ġidentities": 39421, + "Ġidentity": 9569, + "Ġideological": 41833, + "Ġideologies": 88871, + "Ġideology": 33549, + "Ġidi": 40660, + "Ġidiot": 42258, + "Ġidiots": 79033, + "Ġidle": 27647, + "Ġidol": 47914, + "Ġidols": 89761, + "Ġids": 14151, + "Ġidx": 7187, + "Ġie": 29858, + "Ġieee": 60055, + "Ġierr": 53398, + "Ġiets": 88333, + "Ġif": 421, + "Ġiface": 49313, + "Ġifdef": 68971, + "Ġiff": 51108, + "Ġifndef": 82290, + "Ġiframe": 48413, + "Ġifs": 84073, + "Ġifstream": 71151, + "Ġig": 19373, + "Ġigen": 91229, + "Ġign": 5222, + "Ġignite": 89436, + "Ġignited": 97888, + "Ġignition": 59592, + "Ġignor": 25357, + "Ġignorance": 40560, + "Ġignorant": 46980, + "Ġignore": 10034, + "Ġignored": 12036, + "Ġignores": 48278, + "Ġignoring": 30903, + "Ġigual": 43475, + "Ġih": 13305, + "Ġihm": 58442, + "Ġihn": 40662, + "Ġihnen": 74021, + "Ġihr": 24440, + "Ġihre": 34749, + "Ġihrem": 71017, + "Ġihren": 52716, + "Ġihrer": 54649, + "Ġiht": 127648, + "Ġihtiy": 136922, + "Ġihtiya": 134616, + "Ġihtiyacı": 136923, + "Ġihtiyaç": 134617, + "Ġii": 14460, + "Ġiid": 91343, + "Ġiii": 62093, + "Ġij": 71861, + "Ġik": 13361, + "Ġikea": 74735, + "Ġikke": 22688, + "Ġil": 3815, + "Ġila": 138987, + "Ġilaç": 138988, + "Ġile": 30805, + "Ġileti": 137081, + "ĠiletiÅŁim": 137082, + "Ġili": 59309, + "ĠiliÅŁ": 127253, + "ĠiliÅŁki": 132156, + "ĠiliÅŁkin": 135047, + "Ġilk": 80338, + "Ġill": 5862, + "Ġilleg": 28967, + "Ġillegal": 11816, + "Ġillegally": 37384, + "Ġillicit": 58129, + "Ġillness": 17125, + "Ġillnesses": 48809, + "Ġillum": 46406, + "Ġillumin": 42987, + "Ġilluminate": 87582, + "Ġilluminated": 74548, + "Ġillumination": 75952, + "Ġillusion": 40819, + "Ġillusions": 87028, + "Ġillust": 96483, + "Ġillustr": 12293, + "Ġillustrate": 40368, + "Ġillustrated": 35662, + "Ġillustrates": 45380, + "Ġillustrating": 92912, + "Ġillustration": 39034, + "Ġillustrations": 44443, + "Ġillustrator": 93365, + "Ġilma": 91221, + "Ġils": 45052, + "Ġilçe": 132807, + "Ġim": 732, + "Ġima": 92693, + "Ġimag": 6532, + "Ġimage": 2168, + "ĠimageData": 65958, + "ĠimageName": 66073, + "ĠimageNamed": 30293, + "ĠimagePath": 69933, + "ĠimageSize": 87559, + "ĠimageURL": 89861, + "ĠimageUrl": 39649, + "ĠimageView": 29965, + "Ġimagem": 62897, + "Ġimagen": 40569, + "Ġimagery": 40445, + "Ġimages": 5335, + "Ġimagin": 42206, + "Ġimaginable": 89167, + "Ġimaginary": 49952, + "Ġimagination": 27799, + "Ġimaginative": 76261, + "Ġimagine": 12793, + "Ġimagined": 34606, + "Ġimaging": 31658, + "Ġimagining": 76624, + "Ġimap": 94717, + "Ġimb": 72254, + "Ġimbalance": 67231, + "Ġimdb": 83245, + "Ġime": 74434, + "Ġimg": 4964, + "ĠimgUrl": 76083, + "Ġimgs": 56165, + "Ġimitation": 86840, + "Ġimm": 4893, + "Ġimmac": 98138, + "Ġimmature": 85425, + "Ġimmedi": 5551, + "Ġimmediate": 13922, + "Ġimmediately": 7069, + "Ġimmense": 37783, + "Ġimmensely": 62589, + "Ġimmer": 25531, + "Ġimmersed": 77208, + "Ġimmersion": 74557, + "Ġimmersive": 64180, + "Ġimmigr": 15285, + "Ġimmigrant": 36550, + "Ġimmigrants": 19955, + "Ġimmigration": 15093, + "Ġimminent": 50451, + "Ġimmobil": 76341, + "Ġimmoral": 78154, + "Ġimmortal": 59114, + "Ġimmun": 32019, + "Ġimmune": 22077, + "Ġimmunity": 39268, + "Ġimmutable": 51572, + "Ġimp": 3163, + "Ġimpact": 5421, + "Ġimpacted": 38928, + "Ġimpactful": 97890, + "Ġimpacting": 72955, + "Ġimpacts": 24905, + "Ġimpair": 37874, + "Ġimpaired": 49060, + "Ġimpairment": 52217, + "Ġimpart": 69672, + "Ġimpartial": 67724, + "Ġimpass": 95589, + "Ġimpatient": 73697, + "Ġimpe": 35026, + "Ġimpeachment": 63654, + "Ġimpecc": 74619, + "Ġimpeccable": 88546, + "Ġimped": 49402, + "Ġimpedance": 89948, + "Ġimpending": 62461, + "Ġimper": 16772, + "Ġimperative": 47596, + "Ġimperfect": 67963, + "Ġimperial": 34279, + "Ġimperialism": 78951, + "Ġimperson": 59749, + "Ġimpl": 11605, + "Ġimplant": 45360, + "Ġimplanted": 94856, + "Ġimplants": 62184, + "Ġimplement": 4211, + "Ġimplementation": 8129, + "Ġimplementations": 38337, + "Ġimplemented": 11537, + "Ġimplementing": 24931, + "Ġimplements": 5169, + "Ġimplic": 91095, + "Ġimplicated": 68602, + "Ġimplication": 60536, + "Ġimplications": 24154, + "Ġimplicit": 17995, + "Ġimplicitly": 51773, + "Ġimplied": 6131, + "Ġimplies": 23945, + "Ġimplode": 26661, + "Ġimply": 33508, + "Ġimplying": 72867, + "Ġimport": 1159, + "Ġimportance": 12650, + "Ġimportant": 2989, + "Ġimportante": 35797, + "Ġimportantes": 74319, + "Ġimportantly": 22813, + "Ġimported": 24928, + "Ġimporter": 62561, + "Ġimporting": 49895, + "Ġimports": 15202, + "Ġimportância": 139669, + "Ġimpose": 32230, + "Ġimposed": 25980, + "Ġimposes": 87281, + "Ġimposing": 48841, + "Ġimposition": 93779, + "Ġimposs": 90082, + "Ġimpossible": 11997, + "Ġimpost": 81409, + "Ġimpover": 76642, + "Ġimpoverished": 87298, + "Ġimpr": 22540, + "Ġimpres": 76898, + "Ġimpress": 9897, + "Ġimpressed": 24404, + "Ġimpression": 20792, + "Ġimpressions": 49991, + "Ġimpressive": 15978, + "Ġimprimir": 92506, + "Ġimprint": 78867, + "Ġimprison": 28219, + "Ġimprisoned": 51842, + "Ġimprisonment": 42274, + "Ġimpro": 3657, + "Ġimprobable": 96256, + "Ġimproper": 39986, + "Ġimproperly": 74198, + "Ġimprov": 86670, + "Ġimprove": 7269, + "Ġimproved": 12944, + "Ġimprovement": 15673, + "Ġimprovements": 18142, + "Ġimproves": 34950, + "Ġimproving": 18392, + "Ġimprovis": 79063, + "Ġimprovised": 92539, + "Ġimpuls": 96779, + "Ġimpulse": 48867, + "Ġimpulses": 86533, + "Ġimpunity": 97670, + "Ġimread": 82419, + "Ġims": 71203, + "Ġimshow": 78599, + "Ġimu": 92861, + "Ġimágenes": 96275, + "Ġin": 304, + "ĠinFile": 94001, + "Ġinability": 37450, + "Ġinaccessible": 81728, + "Ġinaccur": 40925, + "Ġinaccurate": 49404, + "Ġinactive": 31799, + "Ġinade": 39106, + "Ġinadequate": 45479, + "Ġinadvert": 58828, + "Ġinadvertently": 69085, + "Ġinappropriate": 32681, + "Ġinaug": 29520, + "Ġinaugur": 81000, + "Ġinaugural": 53459, + "Ġinauguration": 59710, + "Ġinbound": 64943, + "Ġinbox": 22883, + "Ġinc": 3625, + "Ġincap": 39868, + "Ġincapable": 57171, + "Ġincapac": 93030, + "Ġincarcer": 45528, + "Ġincarcerated": 80124, + "Ġincarceration": 69052, + "Ġincarn": 95972, + "Ġincarnation": 78948, + "Ġincent": 20046, + "Ġincentiv": 82984, + "Ġincentive": 35110, + "Ġincentives": 35480, + "Ġinception": 53429, + "Ġincess": 91847, + "Ġincest": 42880, + "Ġinch": 17122, + "Ġinches": 14924, + "Ġincid": 61860, + "Ġincidence": 38675, + "Ġincident": 10455, + "Ġincidental": 83216, + "Ġincidents": 23546, + "Ġincididunt": 86404, + "Ġincl": 18409, + "Ġinclination": 75904, + "Ġinclined": 42031, + "Ġinclud": 1798, + "Ġinclude": 2924, + "Ġincluded": 5230, + "Ġincludes": 5646, + "Ġincluding": 2670, + "Ġinclus": 44309, + "Ġinclusion": 27187, + "Ġinclusive": 28308, + "Ġincluso": 72211, + "Ġincom": 52890, + "Ġincome": 7911, + "Ġincomes": 45691, + "Ġincoming": 19393, + "Ġincompatible": 52824, + "Ġincompet": 53920, + "Ġincompetence": 98648, + "Ġincompetent": 80812, + "Ġincomplete": 32143, + "Ġincon": 26995, + "Ġincons": 28664, + "Ġinconsist": 79861, + "Ġinconsistencies": 91822, + "Ġinconsistency": 96149, + "Ġinconsistent": 39140, + "Ġincontr": 78619, + "Ġincontri": 15372, + "Ġincontro": 71525, + "Ġinconvenience": 60009, + "Ġinconvenient": 83683, + "Ġincor": 8920, + "Ġincorpor": 14232, + "Ġincorporate": 32335, + "Ġincorporated": 31662, + "Ġincorporates": 51824, + "Ġincorporating": 51813, + "Ġincorporation": 87365, + "Ġincorrect": 15114, + "Ġincorrectly": 43347, + "Ġincr": 66826, + "Ġincre": 2281, + "Ġincrease": 5263, + "Ġincreased": 7172, + "Ġincreases": 12703, + "Ġincreasing": 7703, + "Ġincreasingly": 14756, + "Ġincred": 9653, + "Ġincredible": 15050, + "Ġincredibly": 16815, + "Ġincrement": 16252, + "Ġincremental": 52299, + "Ġincremented": 87819, + "Ġincrements": 61600, + "Ġincub": 48627, + "Ġincumb": 42750, + "Ġincumbent": 50282, + "Ġincur": 61522, + "Ġincurred": 51772, + "Ġind": 1257, + "Ġinde": 5167, + "Ġindeb": 89797, + "Ġindebted": 92898, + "Ġindeed": 12824, + "Ġindef": 36427, + "Ġindefinite": 80856, + "Ġindefinitely": 55234, + "Ġindem": 64289, + "Ġindemn": 89008, + "Ġindent": 17504, + "Ġindentation": 69592, + "Ġindepend": 7011, + "Ġindependence": 23665, + "Ġindependent": 9489, + "Ġindependently": 28135, + "Ġindependents": 95922, + "Ġindex": 1922, + "ĠindexOf": 60623, + "ĠindexPath": 18109, + "Ġindexed": 30581, + "Ġindexer": 87216, + "Ġindexes": 24953, + "Ġindexing": 51980, + "Ġindia": 27711, + "Ġindian": 42225, + "Ġindic": 4447, + "Ġindica": 74680, + "Ġindicate": 13216, + "Ġindicated": 16317, + "Ġindicates": 14807, + "Ġindicating": 18860, + "Ġindication": 27039, + "Ġindications": 55090, + "Ġindicative": 62905, + "Ġindicator": 20438, + "Ġindicators": 33724, + "Ġindice": 56082, + "Ġindices": 14937, + "Ġindict": 37767, + "Ġindicted": 67234, + "Ġindictment": 46860, + "Ġindie": 43478, + "Ġindifference": 84173, + "Ġindifferent": 78908, + "Ġindigenous": 38578, + "Ġindign": 81901, + "Ġindirect": 24614, + "Ġindirectly": 45245, + "Ġindis": 82875, + "Ġindiscrim": 93446, + "Ġindispens": 56460, + "Ġindispensable": 63184, + "Ġindiv": 3603, + "Ġindivid": 22665, + "Ġindividual": 3842, + "Ġindividually": 31299, + "Ġindividuals": 7775, + "Ġindo": 41000, + "Ġindonesia": 84378, + "Ġindoor": 29519, + "Ġindoors": 54054, + "Ġinds": 98513, + "Ġindu": 4408, + "Ġinduce": 48753, + "Ġinduced": 35472, + "Ġinduces": 89874, + "Ġinducing": 95711, + "Ġinduction": 37056, + "Ġindul": 36860, + "Ġindulge": 67090, + "Ġindust": 8871, + "Ġindustri": 66865, + "Ġindustrial": 12785, + "Ġindustries": 19102, + "Ġindustry": 4958, + "Ġindx": 71095, + "Ġine": 19116, + "Ġineff": 89225, + "Ġineffective": 54188, + "Ġinefficient": 63381, + "Ġineligible": 95816, + "Ġinequalities": 92234, + "Ġinequality": 31205, + "Ġinert": 79973, + "Ġinertia": 77452, + "Ġinet": 49476, + "Ġinev": 20659, + "Ġinevitable": 30252, + "Ġinevitably": 39505, + "Ġinex": 54881, + "Ġinexp": 33578, + "Ġinexpensive": 43152, + "Ġinexperienced": 78866, + "Ġinf": 4132, + "Ġinfamous": 38533, + "Ġinfancy": 85517, + "Ġinfant": 30283, + "Ġinfantry": 54427, + "Ġinfants": 41434, + "Ġinfect": 33427, + "Ġinfected": 28274, + "Ġinfection": 18873, + "Ġinfections": 28920, + "Ġinfectious": 49500, + "Ġinfer": 23583, + "Ġinference": 44378, + "Ġinferior": 37179, + "Ġinferred": 67595, + "Ġinfertility": 94543, + "Ġinfield": 99006, + "Ġinfile": 48989, + "Ġinfiltr": 42264, + "Ġinfiltration": 97735, + "Ġinfinit": 81126, + "Ġinfinite": 23809, + "Ġinfinitely": 71500, + "Ġinfinity": 54910, + "Ġinfix": 88157, + "Ġinfl": 4601, + "Ġinflamm": 25222, + "Ġinflammation": 36040, + "Ġinflammatory": 46188, + "Ġinflatable": 96207, + "Ġinflate": 70698, + "Ġinflated": 69854, + "Ġinflater": 21993, + "Ġinflation": 24527, + "Ġinflict": 67468, + "Ġinflicted": 68077, + "Ġinflu": 6653, + "Ġinfluence": 10173, + "Ġinfluenced": 27061, + "Ġinfluencers": 86233, + "Ġinfluences": 33353, + "Ġinfluencing": 65600, + "Ġinfluential": 31449, + "Ġinfluenza": 61837, + "Ġinflux": 52852, + "Ġinfo": 3546, + "Ġinfographic": 97382, + "Ġinform": 6051, + "Ġinformación": 34515, + "Ġinformal": 41787, + "Ġinformant": 92021, + "Ġinformat": 50207, + "Ġinformatie": 97255, + "Ġinformation": 1995, + "Ġinformational": 46635, + "Ġinformations": 43727, + "Ġinformative": 38219, + "Ġinformação": 138188, + "Ġinformações": 64066, + "Ġinformed": 15987, + "Ġinforming": 61925, + "Ġinforms": 63152, + "Ġinfos": 44144, + "Ġinfr": 24631, + "Ġinfra": 48176, + "Ġinfrared": 58041, + "Ġinfrastructure": 13737, + "Ġinfring": 44019, + "Ġinfringement": 53343, + "Ġinfuri": 84125, + "Ġinfused": 71803, + "Ġinfusion": 70208, + "Ġing": 6749, + "Ġingen": 45392, + "Ġingenious": 96699, + "Ġingest": 88272, + "Ġingestion": 87347, + "Ġingin": 87911, + "Ġingl": 97825, + "Ġinglés": 143184, + "Ġingr": 80236, + "Ġingredient": 24763, + "Ġingredients": 13966, + "Ġingres": 42664, + "Ġingresar": 89953, + "Ġingress": 78559, + "Ġinh": 59437, + "Ġinhab": 25252, + "Ġinhabit": 32327, + "Ġinhabitants": 39671, + "Ġinhabited": 69498, + "Ġinhal": 76673, + "Ġinher": 15164, + "Ġinherent": 36988, + "Ġinherently": 48088, + "Ġinherit": 23749, + "Ġinheritance": 41822, + "Ġinherited": 26990, + "Ġinherits": 75482, + "Ġinhib": 20142, + "Ġinhibit": 67933, + "Ġinhibited": 98569, + "Ġinhibition": 60378, + "Ġinhibitor": 69685, + "Ġinhibitors": 67542, + "Ġini": 16806, + "Ġinici": 44023, + "Ġinicial": 48556, + "Ġiniciar": 86027, + "Ġinicio": 50334, + "Ġinit": 2930, + "ĠinitComponents": 52056, + "ĠinitData": 68733, + "ĠinitState": 51138, + "ĠinitView": 59451, + "ĠinitWith": 13864, + "ĠinitWithFrame": 30067, + "ĠinitWithNibName": 96708, + "ĠinitWithStyle": 90188, + "ĠinitWithTitle": 72730, + "Ġiniti": 12672, + "Ġinitial": 2856, + "ĠinitialState": 29427, + "ĠinitialValue": 60062, + "ĠinitialValues": 93483, + "Ġinitialise": 66166, + "Ġinitialised": 87667, + "Ġinitialization": 17465, + "Ġinitialize": 9468, + "Ġinitialized": 17271, + "Ġinitializer": 42312, + "Ġinitializes": 57857, + "Ġinitializing": 57060, + "Ġinitially": 15102, + "Ġinitials": 78437, + "Ġinitiate": 38101, + "Ġinitiated": 32130, + "Ġinitiating": 78416, + "Ġinitiation": 60468, + "Ġinitiative": 20162, + "Ġinitiatives": 27172, + "Ġinitiator": 98040, + "Ġinj": 5811, + "Ġinject": 15551, + "Ġinjected": 40672, + "Ġinjecting": 87285, + "Ġinjection": 25071, + "Ġinjections": 64146, + "Ġinjector": 62643, + "Ġinjunction": 60429, + "Ġinjured": 15532, + "Ġinjuries": 14970, + "Ġinjuring": 85585, + "Ġinjury": 10895, + "Ġinjust": 97653, + "Ġinjustice": 50115, + "Ġink": 26415, + "Ġinkl": 81625, + "Ġinland": 76189, + "Ġinlet": 79107, + "Ġinline": 7381, + "Ġinmate": 66435, + "Ġinmates": 41612, + "Ġinn": 6170, + "Ġinnate": 63970, + "Ġinne": 77225, + "Ġinner": 9179, + "Ġinning": 45286, + "Ġinnings": 35000, + "Ġinnoc": 18619, + "Ġinnocence": 56211, + "Ġinnocent": 24243, + "Ġinnov": 9529, + "Ġinnovate": 90964, + "Ġinnovation": 18770, + "Ġinnovations": 44945, + "Ġinnovative": 18199, + "Ġinoc": 85918, + "Ġinode": 34803, + "Ġinorder": 66041, + "Ġinout": 96277, + "Ġinp": 32344, + "Ġinplace": 41894, + "Ġinput": 1946, + "ĠinputData": 90162, + "ĠinputFile": 76208, + "ĠinputStream": 46456, + "ĠinputValue": 79908, + "Ġinputs": 11127, + "Ġinquire": 67170, + "Ġinquiries": 43883, + "Ġinquiry": 25893, + "Ġins": 1640, + "Ġinsan": 61538, + "Ġinsane": 34018, + "Ġinsanely": 89366, + "Ġinsanity": 72053, + "Ġinsanların": 138680, + "Ġinscription": 50570, + "Ġinse": 53582, + "Ġinsect": 25989, + "Ġinsects": 40811, + "Ġinsecure": 61845, + "Ġinsecurity": 73835, + "Ġinsensitive": 70480, + "Ġinser": 47325, + "Ġinsert": 5656, + "Ġinserted": 21578, + "Ġinserting": 38298, + "Ġinsertion": 35927, + "Ġinserts": 49298, + "Ġinset": 54859, + "Ġinsets": 98069, + "Ġinside": 4766, + "Ġinsider": 54365, + "Ġinsiders": 76714, + "Ġinsight": 20017, + "Ġinsightful": 69062, + "Ġinsights": 25709, + "Ġinsign": 53027, + "Ġinsignificant": 72421, + "Ġinsist": 31162, + "Ġinsisted": 28576, + "Ġinsistence": 78297, + "Ġinsisting": 59627, + "Ġinsists": 42346, + "Ġinsn": 76102, + "Ġinsol": 84843, + "Ġinsomnia": 82411, + "Ġinspect": 24085, + "Ġinspected": 61784, + "Ġinspection": 25683, + "Ġinspections": 56974, + "Ġinspector": 44725, + "Ġinspectors": 68830, + "Ġinspir": 12690, + "Ġinspiration": 19760, + "Ġinspirational": 56214, + "Ġinspire": 30640, + "Ġinspired": 14606, + "Ġinspires": 69290, + "Ġinspiring": 33047, + "Ġinst": 1761, + "Ġinstability": 55299, + "Ġinstagram": 51004, + "Ġinstal": 45598, + "Ġinstall": 4582, + "Ġinstallation": 13713, + "Ġinstallations": 44118, + "Ġinstalled": 10275, + "Ġinstaller": 43052, + "Ġinstalling": 26632, + "Ġinstallment": 46544, + "Ġinstalls": 70320, + "Ġinstance": 2867, + "Ġinstanceof": 8083, + "Ġinstances": 13121, + "Ġinstancia": 97979, + "Ġinstant": 9690, + "Ġinstantaneous": 95119, + "Ġinstantiate": 40902, + "Ġinstantiated": 54586, + "Ġinstantiation": 81128, + "Ġinstantly": 21818, + "Ġinstead": 4518, + "Ġinstinct": 30555, + "Ġinstincts": 61821, + "Ġinstit": 7759, + "Ġinstitute": 43698, + "Ġinstituted": 79999, + "Ġinstitutes": 93712, + "Ġinstitution": 14898, + "Ġinstitutional": 32132, + "Ġinstitutions": 14336, + "Ġinstr": 9491, + "Ġinstruct": 21066, + "Ġinstructed": 40975, + "Ġinstruction": 7600, + "Ġinstructional": 71245, + "Ġinstructions": 11221, + "Ġinstructor": 32215, + "Ġinstructors": 53818, + "Ġinstrument": 14141, + "Ġinstrumental": 40945, + "Ġinstrumentation": 77033, + "Ġinstruments": 23316, + "Ġinsufficient": 38313, + "Ġinsulated": 78279, + "Ġinsulation": 48017, + "Ġinsulin": 31052, + "Ġinsult": 26132, + "Ġinsulting": 67092, + "Ġinsults": 67889, + "Ġinsurance": 8113, + "Ġinsure": 69532, + "Ġinsured": 52042, + "Ġinsurer": 72531, + "Ġinsurers": 55274, + "Ġinsurg": 48114, + "Ġinsurgency": 96508, + "Ġinsurgents": 87338, + "Ġint": 526, + "ĠintValue": 53077, + "Ġintact": 34439, + "Ġintake": 22881, + "Ġinte": 28287, + "Ġinteg": 5388, + "Ġinteger": 7546, + "ĠintegerValue": 70613, + "Ġintegers": 25780, + "Ġintegr": 8761, + "Ġintegral": 25098, + "Ġintegrate": 31072, + "Ġintegrated": 18250, + "Ġintegrates": 74662, + "Ġintegrating": 53852, + "Ġintegration": 17590, + "Ġintegrity": 16614, + "Ġintel": 14157, + "Ġintelig": 89645, + "Ġintellect": 16694, + "Ġintellectual": 19634, + "Ġintellectually": 93291, + "Ġintellectuals": 82836, + "Ġintellig": 59438, + "Ġintelligence": 11229, + "Ġintelligent": 24514, + "Ġintend": 29630, + "Ġintended": 10602, + "Ġintending": 84339, + "Ġintends": 43230, + "Ġintens": 24245, + "Ġintense": 18894, + "Ġintensely": 69633, + "Ġintensified": 78749, + "Ġintensity": 20612, + "Ġintensive": 36195, + "Ġintent": 7385, + "Ġintention": 14602, + "Ġintentional": 46864, + "Ġintentionally": 36204, + "Ġintentions": 33234, + "Ġintents": 93515, + "Ġinter": 946, + "Ġinteract": 16282, + "Ġinteracting": 44730, + "Ġinteraction": 16230, + "Ġinteractions": 21880, + "Ġinteractive": 20753, + "Ġinteracts": 83161, + "Ġintercept": 28639, + "Ġintercepted": 69374, + "Ġinterception": 69468, + "Ġinterceptions": 72457, + "Ġinterceptor": 76945, + "Ġinterchange": 51263, + "Ġinterchangeable": 90129, + "Ġinterconnected": 82316, + "Ġintercourse": 49440, + "Ġinterdisciplinary": 87319, + "Ġinteres": 49416, + "Ġinteress": 39353, + "Ġinteresse": 95421, + "Ġinterest": 2734, + "Ġinterested": 8014, + "Ġinteresting": 7040, + "Ġinterests": 11772, + "Ġinterf": 33184, + "Ġinterface": 3749, + "Ġinterfaces": 24099, + "Ġinterfer": 40205, + "Ġinterfere": 39878, + "Ġinterference": 31217, + "Ġinterfering": 74287, + "Ġinterim": 40553, + "Ġinterior": 14791, + "Ġinteriors": 73464, + "Ġinterle": 94968, + "Ġintermedi": 54175, + "Ġintermediary": 94070, + "Ġintermediate": 28439, + "Ġintermitt": 59669, + "Ġintermittent": 74468, + "Ġintern": 2590, + "Ġinternacional": 85159, + "Ġinternal": 5306, + "Ġinternally": 33067, + "Ġinternals": 91025, + "Ġinternational": 6489, + "Ġinternationally": 36445, + "Ġinternet": 7602, + "Ġinterns": 98834, + "Ġinternship": 59759, + "Ġinteroper": 94130, + "Ġinterp": 47271, + "Ġinterpersonal": 73858, + "Ġinterpol": 22876, + "Ġinterpolate": 59457, + "Ġinterpolated": 85664, + "Ġinterpolation": 36487, + "Ġinterpre": 17929, + "Ġinterpret": 14198, + "Ġinterpretation": 22845, + "Ġinterpretations": 57589, + "Ġinterpreted": 32298, + "Ġinterpreter": 39299, + "Ġinterpreting": 65644, + "Ġinterracial": 52461, + "Ġinterrog": 36439, + "Ġinterrogation": 65265, + "Ġinterrupt": 12667, + "Ġinterrupted": 36783, + "Ġinterruption": 74003, + "Ġinterruptions": 88408, + "Ġinterrupts": 43437, + "Ġinters": 80919, + "Ġintersect": 31796, + "Ġintersection": 19246, + "Ġintersections": 64257, + "Ġintersects": 88184, + "Ġinterstate": 73083, + "Ġintertw": 79950, + "Ġintertwined": 98792, + "Ġinterv": 85853, + "Ġinterval": 9873, + "Ġintervals": 26992, + "Ġinterven": 30857, + "Ġintervene": 54479, + "Ġintervened": 89448, + "Ġintervening": 88401, + "Ġintervention": 20949, + "Ġinterventions": 38355, + "Ġinterview": 7128, + "Ġinterviewed": 29047, + "Ġinterviewer": 86835, + "Ġinterviewing": 64716, + "Ġinterviews": 19344, + "Ġinterés": 138048, + "Ġintest": 38308, + "Ġintestinal": 62800, + "Ġintestine": 91134, + "Ġintf": 93706, + "Ġintim": 21947, + "Ġintimacy": 65164, + "Ġintimate": 31387, + "Ġintimately": 95652, + "Ġintimid": 29834, + "Ġintimidate": 89688, + "Ġintimidated": 84061, + "Ġintimidating": 63030, + "Ġintimidation": 69059, + "Ġintl": 95015, + "Ġinto": 1119, + "Ġintoler": 71629, + "Ġintolerance": 74331, + "Ġintox": 84017, + "Ġintoxic": 71715, + "Ġintoxicated": 86423, + "Ġintptr": 43966, + "Ġintr": 10582, + "Ġintra": 49838, + "Ġintric": 93503, + "Ġintricate": 56116, + "Ġintrig": 30396, + "Ġintrigue": 80513, + "Ġintrigued": 68018, + "Ġintriguing": 40665, + "Ġintrinsic": 46601, + "Ġintro": 19706, + "Ġintrodu": 5983, + "Ġintroduce": 19131, + "Ġintroduced": 11523, + "Ġintroduces": 38919, + "Ġintroducing": 31918, + "Ġintroduction": 16800, + "Ġintroductory": 61369, + "Ġintros": 25528, + "Ġintrusion": 81046, + "Ġintrusive": 84910, + "Ġints": 54724, + "Ġintuit": 96201, + "Ġintuition": 56251, + "Ġintuitive": 41679, + "Ġintval": 26217, + "Ġinté": 56010, + "Ġintéress": 91659, + "Ġintéressant": 141164, + "Ġinund": 91529, + "Ġinv": 1529, + "Ġinvade": 74172, + "Ġinvaded": 63665, + "Ġinvaders": 90221, + "Ġinvading": 82531, + "Ġinval": 48044, + "Ġinvalid": 8318, + "Ġinvalidate": 39943, + "Ġinvalidated": 98306, + "Ġinvaluable": 55093, + "Ġinvariably": 73514, + "Ġinvariant": 57620, + "Ġinvasion": 29115, + "Ġinvasive": 52254, + "Ġinve": 92184, + "Ġinvent": 17023, + "Ġinvented": 35492, + "Ġinvention": 27130, + "Ġinventions": 84049, + "Ġinventive": 90932, + "Ġinventor": 70901, + "Ġinventory": 15444, + "Ġinvers": 64583, + "Ġinverse": 27949, + "Ġinversion": 46488, + "Ġinversión": 143142, + "Ġinvert": 42199, + "Ġinverted": 46701, + "Ġinvest": 2725, + "Ġinvested": 27991, + "Ġinvestig": 4665, + "Ġinvestigación": 97524, + "Ġinvestigate": 19314, + "Ġinvestigated": 26219, + "Ġinvestigates": 83263, + "Ġinvestigating": 23890, + "Ġinvestigation": 8814, + "Ġinvestigations": 25883, + "Ġinvestigative": 46636, + "Ġinvestigator": 48481, + "Ġinvestigators": 25381, + "Ġinvesting": 24965, + "Ġinvestment": 9162, + "Ġinvestments": 21786, + "Ġinvestor": 29593, + "Ġinvestors": 14823, + "Ġinvis": 99037, + "Ġinvisible": 29447, + "Ġinvit": 82067, + "Ġinvitation": 28688, + "Ġinvitations": 58669, + "Ġinvite": 21399, + "Ġinvited": 18218, + "Ġinvites": 44410, + "Ġinviting": 41192, + "Ġinvo": 50644, + "Ġinvocation": 28696, + "Ġinvoice": 24615, + "Ġinvoices": 64052, + "Ġinvoke": 19873, + "Ġinvoked": 26881, + "Ġinvokes": 66051, + "Ġinvokevirtual": 69825, + "Ġinvoking": 46028, + "ĠinvokingState": 87841, + "Ġinvol": 4359, + "Ġinvoluntary": 90505, + "Ġinvolve": 21058, + "Ġinvolved": 6398, + "Ġinvolvement": 21587, + "Ġinvolves": 17601, + "Ġinvolving": 15860, + "Ġinward": 61918, + "ĠinÃŃcio": 135567, + "Ġio": 6399, + "Ġioctl": 64460, + "Ġiod": 78987, + "Ġion": 27672, + "Ġions": 64025, + "Ġios": 27531, + "Ġiota": 81910, + "Ġioutil": 43144, + "Ġiov": 98962, + "Ġip": 5997, + "ĠipAddress": 91715, + "Ġipad": 80713, + "Ġipairs": 56244, + "Ġipc": 60104, + "Ġiphone": 42448, + "Ġips": 59022, + "Ġipsum": 26342, + "Ġipt": 66068, + "Ġipv": 45475, + "Ġiq": 91459, + "Ġir": 6216, + "Ġirc": 79923, + "Ġire": 43750, + "Ġirgend": 87262, + "Ġiris": 63942, + "Ġirm": 97282, + "Ġiron": 11001, + "Ġironic": 58460, + "Ġironically": 79765, + "Ġirony": 50605, + "Ġirq": 38269, + "Ġirr": 36719, + "Ġirradi": 75227, + "Ġirrational": 60654, + "Ġirre": 24870, + "Ġirregular": 41308, + "Ġirrelevant": 39715, + "Ġirres": 33948, + "Ġirresist": 76197, + "Ġirresistible": 85829, + "Ġirrespective": 75553, + "Ġirresponsible": 70279, + "Ġirreversible": 92194, + "Ġirrig": 52005, + "Ġirrigation": 62466, + "Ġirrit": 29018, + "Ġirritated": 83871, + "Ġirritating": 83977, + "Ġirritation": 65410, + "Ġis": 374, + "ĠisActive": 43247, + "ĠisAdmin": 70600, + "ĠisArray": 76969, + "ĠisAuthenticated": 70926, + "ĠisChecked": 69960, + "ĠisConnected": 98796, + "ĠisEmpty": 38948, + "ĠisEnabled": 61302, + "ĠisEqual": 21337, + "ĠisEqualToString": 25522, + "ĠisError": 87612, + "ĠisFirst": 88521, + "ĠisIn": 81247, + "ĠisKindOfClass": 48511, + "ĠisLoading": 33662, + "ĠisLoggedIn": 72018, + "ĠisNaN": 77867, + "ĠisNew": 87478, + "ĠisOpen": 44681, + "ĠisSelected": 61441, + "ĠisSuccess": 76688, + "ĠisValid": 26171, + "ĠisVisible": 71744, + "Ġisa": 51335, + "Ġisbn": 69469, + "Ġisc": 95172, + "Ġischem": 97008, + "Ġise": 76093, + "Ġisempty": 50309, + "Ġisi": 95395, + "Ġisinstance": 11402, + "Ġisize": 91373, + "Ġisl": 81931, + "Ġisland": 12922, + "Ġislands": 29000, + "Ġisn": 4436, + "Ġisnt": 68958, + "Ġiso": 33456, + "Ġisol": 13098, + "Ġisolate": 42123, + "Ġisolated": 24203, + "Ġisolation": 30298, + "Ġisot": 68451, + "Ġisp": 95297, + "Ġisr": 93562, + "Ġiss": 2369, + "Ġisset": 10248, + "Ġisso": 40805, + "Ġissu": 51966, + "Ġissuance": 66499, + "Ġissue": 4265, + "Ġissued": 10897, + "Ġissuer": 54835, + "Ġissues": 4714, + "Ġissuing": 42121, + "Ġist": 5999, + "Ġistedi": 127979, + "ĠistediÄŁ": 143322, + "ĠistediÄŁi": 138396, + "ĠistediÄŁiniz": 143323, + "Ġit": 432, + "Ġitalia": 85776, + "Ġitalian": 27948, + "Ġitaliana": 88359, + "Ġitaliane": 98943, + "Ġitaliani": 79449, + "Ġitaliano": 59804, + "Ġitalic": 60763, + "Ġitch": 94756, + "Ġitching": 84975, + "Ġitem": 1509, + "ĠitemBuilder": 75835, + "ĠitemCount": 55893, + "ĠitemId": 46371, + "ĠitemList": 72274, + "ĠitemName": 66009, + "ĠitemType": 88791, + "ĠitemView": 27888, + "Ġitemprop": 69173, + "Ġitems": 3589, + "Ġitens": 94935, + "Ġiter": 5367, + "Ġiterable": 50834, + "Ġiterate": 29629, + "Ġiterating": 87235, + "Ġiteration": 19569, + "Ġiterations": 25687, + "Ġiterative": 86875, + "Ġiterator": 15091, + "Ġiterators": 90865, + "Ġitertools": 41331, + "Ġith": 77842, + "Ġitinerary": 72904, + "Ġitk": 78635, + "Ġitm": 84237, + "Ġitr": 32833, + "Ġits": 1181, + "Ġitself": 5086, + "Ġitu": 35691, + "Ġiv": 17509, + "Ġive": 98577, + "Ġivory": 69816, + "Ġiw": 72268, + "Ġix": 26864, + "Ġiy": 50414, + "Ġiz": 22137, + "Ġiç": 26517, + "Ġiçer": 127051, + "Ġiçerik": 142240, + "Ġiçeris": 131116, + "Ġiçerisinde": 131117, + "Ġiçin": 33171, + "Ġiçinde": 128972, + "ĠiÅŁ": 54709, + "ĠiÅŁaret": 138296, + "ĠiÅŁle": 140170, + "ĠiÅŁlem": 129850, + "ĠiÅŁleml": 140171, + "ĠiÅŁlemleri": 140172, + "ĠiÅŁlet": 139976, + "ĠiÅŁletme": 139977, + "ĠiÅŁÃ§i": 137660, + "Ġj": 502, + "ĠjButton": 24639, + "ĠjLabel": 13858, + "ĠjMenuItem": 87168, + "ĠjPanel": 19068, + "ĠjQuery": 13215, + "ĠjScrollPane": 55365, + "ĠjTable": 92549, + "ĠjTextField": 51340, + "Ġja": 11937, + "Ġjaar": 50343, + "Ġjab": 60064, + "Ġjac": 59843, + "Ġjack": 25072, + "Ġjacket": 26208, + "Ġjackets": 60142, + "Ġjackpot": 68328, + "Ġjade": 85015, + "Ġjadx": 57543, + "Ġjag": 26742, + "Ġjail": 17540, + "Ġjailed": 54506, + "Ġjails": 98565, + "Ġjak": 19394, + "Ġjaki": 139537, + "Ġjakie": 141227, + "ĠjakieÅĽ": 141228, + "ĠjakiÅĽ": 139538, + "Ġjako": 49726, + "ĠjakoÅĽci": 134241, + "Ġjal": 75974, + "Ġjam": 20072, + "Ġjamais": 55216, + "Ġjames": 84946, + "Ġjams": 89298, + "Ġjan": 32862, + "Ġjane": 98606, + "Ġjanvier": 93711, + "Ġjap": 85185, + "Ġjapan": 54129, + "Ġjapanese": 52948, + "Ġjapon": 79397, + "Ġjar": 29595, + "Ġjardin": 90299, + "Ġjars": 60011, + "Ġjasmine": 65809, + "Ġjav": 1258, + "Ġjava": 1642, + "Ġjavafx": 16718, + "Ġjavascript": 35710, + "Ġjavax": 2871, + "Ġjaw": 16535, + "Ġjaws": 81256, + "Ġjazz": 33897, + "Ġjb": 87118, + "Ġjc": 72818, + "Ġjclass": 76296, + "Ġjd": 73376, + "Ġjdbc": 85902, + "ĠjdbcTemplate": 87317, + "Ġje": 4759, + "Ġjealous": 40590, + "Ġjealousy": 79722, + "Ġjean": 96621, + "Ġjeans": 33289, + "Ġjed": 18302, + "Ġjede": 94572, + "Ġjedem": 94885, + "Ġjeden": 53165, + "Ġjeder": 55928, + "Ġjedis": 70427, + "Ġjednak": 75225, + "Ġjednoc": 140872, + "Ġjednocze": 140873, + "ĠjednoczeÅĽnie": 140874, + "ĠjednÄħ": 140263, + "Ġjedoch": 60637, + "Ġjeep": 85176, + "Ġjeg": 23991, + "Ġjego": 70929, + "Ġjej": 62964, + "Ġjel": 84988, + "Ġjelly": 51341, + "Ġjemand": 77016, + "Ġjenis": 93251, + "Ġjente": 72404, + "Ġjenter": 37159, + "Ġjeopard": 52032, + "Ġjeopardy": 89604, + "Ġjer": 21030, + "Ġjerk": 53424, + "Ġjersey": 36112, + "Ġjerseys": 53051, + "Ġjes": 64746, + "Ġjest": 13296, + "Ġjeste": 128155, + "ĠjesteÅĽ": 141246, + "ĠjesteÅĽmy": 141546, + "Ġjeszcze": 80253, + "Ġjet": 16594, + "Ġjets": 43205, + "Ġjetzt": 37699, + "Ġjeu": 50555, + "Ġjeune": 35863, + "Ġjeunes": 59020, + "Ġjeux": 69075, + "Ġjew": 16868, + "Ġjewel": 65841, + "Ġjewellery": 62391, + "Ġjewelry": 30717, + "Ġjewels": 85031, + "Ġjewish": 88138, + "Ġjeżeli": 138441, + "ĠjeÅĽli": 130541, + "Ġji": 73885, + "Ġjig": 94064, + "Ġjihad": 41834, + "Ġjihadist": 96044, + "Ġjihadists": 93280, + "Ġjika": 70879, + "Ġjim": 95444, + "Ġjint": 52467, + "Ġjit": 84573, + "Ġjitter": 84392, + "Ġjj": 36729, + "Ġjl": 84710, + "Ġjlong": 74254, + "Ġjm": 95667, + "Ġjmp": 90369, + "Ġjo": 7647, + "Ġjob": 2618, + "ĠjobId": 89703, + "Ġjobject": 49045, + "Ġjobs": 6887, + "Ġjoe": 93341, + "Ġjog": 30125, + "Ġjogador": 95458, + "Ġjogging": 98638, + "Ġjogo": 66134, + "Ġjohn": 39642, + "Ġjoin": 5138, + "Ġjoined": 10859, + "Ġjoining": 18169, + "Ġjoins": 28682, + "Ġjoint": 10284, + "Ġjointly": 52158, + "Ġjoints": 34258, + "Ġjoke": 21646, + "Ġjoked": 67229, + "Ġjokes": 31420, + "Ġjoking": 57034, + "Ġjon": 88504, + "Ġjong": 63413, + "Ġjorn": 71349, + "Ġjos": 87886, + "Ġjose": 89644, + "Ġjot": 92380, + "Ġjou": 26970, + "Ġjouer": 99096, + "Ġjoueur": 75949, + "Ġjour": 9682, + "Ġjournal": 8322, + "Ġjournalism": 29004, + "Ġjournalist": 22825, + "Ġjournalistic": 93700, + "Ġjournalists": 22532, + "Ġjournals": 41680, + "Ġjourney": 11618, + "Ġjourneys": 63776, + "Ġjournée": 85223, + "Ġjours": 48201, + "Ġjov": 67261, + "Ġjoven": 76583, + "Ġjoy": 15888, + "Ġjoyful": 82112, + "Ġjoys": 93798, + "Ġjoystick": 61589, + "Ġjp": 43571, + "Ġjpeg": 54566, + "Ġjpg": 71185, + "Ġjq": 44648, + "Ġjquery": 61701, + "Ġjr": 79598, + "Ġjs": 6994, + "Ġjson": 2951, + "ĠjsonArray": 66273, + "ĠjsonData": 53875, + "ĠjsonObj": 79175, + "ĠjsonObject": 30227, + "ĠjsonResponse": 90465, + "ĠjsonString": 79689, + "Ġjsonify": 32058, + "Ġjspb": 63770, + "Ġjsx": 80095, + "Ġjt": 91076, + "Ġju": 10267, + "Ġjub": 96343, + "Ġjud": 5741, + "Ġjudge": 11651, + "Ġjudged": 44387, + "Ġjudgement": 47819, + "Ġjudges": 24001, + "Ġjudging": 49690, + "Ġjudgment": 19407, + "Ġjudgments": 58258, + "Ġjudicial": 30652, + "Ġjudiciary": 71875, + "Ġjue": 46950, + "Ġjuego": 51895, + "Ġjuegos": 66523, + "Ġjug": 22394, + "Ġjuga": 42369, + "Ġjugador": 71676, + "Ġjugar": 93260, + "Ġjugg": 89149, + "Ġjuice": 22815, + "Ġjuices": 64859, + "Ġjuicy": 55053, + "Ġjuin": 84759, + "Ġjul": 40538, + "Ġjumlah": 85140, + "Ġjump": 7784, + "Ġjumped": 26005, + "Ġjumper": 68984, + "Ġjumping": 29002, + "Ġjumps": 34208, + "Ġjun": 17699, + "Ġjunction": 48241, + "Ġjung": 68910, + "Ġjunge": 63657, + "Ġjungle": 44420, + "Ġjuni": 82744, + "Ġjunior": 26052, + "Ġjunit": 70727, + "Ġjunk": 29674, + "Ġjunto": 61988, + "Ġjur": 16037, + "Ġjuris": 45895, + "Ġjurisdiction": 28001, + "Ġjurisdictions": 54943, + "Ġjurors": 72039, + "Ġjury": 21234, + "Ġjus": 44807, + "Ġjusqu": 54298, + "Ġjust": 1101, + "Ġjuste": 69274, + "Ġjustice": 12161, + "Ġjustices": 70612, + "Ġjustification": 41435, + "Ġjustified": 34416, + "Ġjustify": 9357, + "ĠjustifyContent": 22039, + "Ġjusto": 81843, + "Ġjuven": 98445, + "Ġjuvenile": 47670, + "Ġjuxtap": 97853, + "Ġjuż": 53921, + "Ġjwt": 24589, + "Ġjylland": 90978, + "Ġjá": 32092, + "Ġjä": 87528, + "Ġjó": 71648, + "Ġjóvenes": 139011, + "ĠjÄĻ": 132936, + "ĠjÄĻzy": 132937, + "ĠjÄĻzyk": 132938, + "Ġk": 595, + "ĠkB": 84435, + "ĠkHz": 69626, + "ĠkInstruction": 92500, + "ĠkW": 77477, + "ĠkWh": 95887, + "Ġka": 16502, + "Ġkab": 73995, + "Ġkad": 68044, + "Ġkadar": 73909, + "Ġkadın": 129546, + "Ġkafka": 67852, + "Ġkak": 95401, + "Ġkako": 90517, + "Ġkal": 35354, + "Ġkald": 127900, + "Ġkaldı": 140199, + "Ġkaldır": 137071, + "Ġkale": 62477, + "Ġkali": 96368, + "Ġkam": 33661, + "Ġkami": 62669, + "Ġkamu": 85190, + "Ġkan": 13422, + "Ġkang": 98286, + "Ġkann": 15719, + "Ġkannst": 65836, + "Ġkans": 68285, + "Ġkao": 78381, + "Ġkap": 44849, + "Ġkappa": 83995, + "Ġkaps": 135699, + "Ġkapsam": 135700, + "Ġkapsamında": 135701, + "Ġkapı": 136110, + "Ġkar": 23425, + "Ġkarakter": 90755, + "Ġkararı": 137549, + "Ġkarde": 136159, + "ĠkardeÅŁ": 136160, + "Ġkarena": 59469, + "Ġkarma": 70442, + "Ġkart": 57302, + "ĠkarÄ±ÅŁ": 134913, + "ĠkarÅŁ": 141771, + "ĠkarÅŁÄ±": 126285, + "ĠkarÅŁÄ±laÅŁ": 136000, + "ĠkarÅŁÄ±s": 141772, + "ĠkarÅŁÄ±sında": 141773, + "Ġkas": 47656, + "Ġkat": 44256, + "Ġkata": 61688, + "Ġkatkı": 136932, + "Ġkatıl": 131475, + "Ġkatılı": 139787, + "Ġkatılım": 139788, + "Ġkaufen": 51134, + "Ġkaum": 99025, + "Ġkay": 36847, + "Ġkayak": 86978, + "Ġkayna": 143584, + "Ġkaynaģı": 143585, + "Ġkayıt": 136051, + "Ġkaz": 90725, + "Ġkaç": 130268, + "Ġkaż": 63293, + "Ġkażd": 136438, + "Ġkażde": 127792, + "Ġkażdego": 135785, + "Ġkażdej": 139638, + "Ġkażdy": 133246, + "Ġkażdym": 136439, + "Ġkb": 38653, + "Ġkc": 87518, + "Ġkcal": 95003, + "Ġkd": 68150, + "Ġke": 1962, + "Ġked": 79042, + "Ġkeen": 26891, + "Ġkeep": 2506, + "Ġkeeper": 53416, + "Ġkeeping": 10282, + "Ġkeeps": 13598, + "Ġkeer": 91924, + "Ġkein": 48803, + "Ġkeine": 31357, + "Ġkeinen": 79913, + "Ġkel": 48228, + "Ġkell": 78072, + "Ġkem": 67431, + "Ġken": 49232, + "Ġkend": 71748, + "Ġkenn": 74261, + "Ġkennen": 59570, + "Ġkennenlernen": 88862, + "Ġkep": 64184, + "Ġkepada": 76797, + "Ġkept": 8604, + "Ġker": 33701, + "Ġkeras": 25689, + "Ġkern": 82585, + "Ġkernel": 10001, + "Ġkernels": 63515, + "Ġkes": 50977, + "Ġket": 31281, + "Ġketo": 65534, + "Ġketogenic": 93750, + "Ġkettle": 72318, + "Ġkey": 1376, + "ĠkeyCode": 64432, + "ĠkeyPressed": 96104, + "ĠkeyValue": 63937, + "Ġkeyboard": 13625, + "ĠkeyboardType": 75099, + "Ġkeyboards": 70302, + "Ġkeycode": 86991, + "Ġkeyed": 90529, + "Ġkeynote": 70905, + "Ġkeyof": 79863, + "Ġkeypad": 60205, + "Ġkeypoints": 87892, + "Ġkeys": 6894, + "Ġkeyst": 99132, + "Ġkeyword": 16174, + "Ġkeywords": 20844, + "Ġkf": 85580, + "Ġkg": 20972, + "Ġkh": 23165, + "Ġkhi": 94994, + "Ġkhiến": 128893, + "Ġkhiá»ĥn": 139533, + "Ġkho": 72067, + "Ġkhoá": 131821, + "Ġkhoản": 130042, + "Ġkhoảng": 129168, + "Ġkhu": 127980, + "Ġkhuyên": 142076, + "Ġkhuyến": 135984, + "Ġkhuôn": 138411, + "Ġkhuẩn": 139901, + "Ġkhá": 128949, + "Ġkhác": 128354, + "Ġkhách": 128351, + "Ġkhám": 131797, + "Ġkháng": 138075, + "Ġkhó": 128639, + "Ġkhóa": 131774, + "Ġkhô": 132190, + "Ġkhông": 53037, + "Ġkhúc": 138228, + "ĠkhÃŃ": 129019, + "ĠkhÄĥn": 130909, + "Ġkhả": 129363, + "Ġkhảo": 132123, + "Ġkhấu": 143449, + "Ġkhẩu": 97610, + "Ġkhắc": 132611, + "Ġkhắp": 138028, + "Ġkhẳng": 136732, + "Ġkhá»§ng": 139235, + "Ġkhá»ı": 129699, + "Ġkhá»ıe": 129700, + "Ġkhá»ıi": 130553, + "Ġkhá»iji": 135836, + "Ġkhá»ķ": 133787, + "ĠkhỼ": 141709, + "ĠkhỼp": 141710, + "Ġkhợi": 132600, + "Ġki": 19953, + "Ġkick": 10323, + "Ġkicked": 27120, + "Ġkicker": 86176, + "Ġkicking": 39194, + "Ġkickoff": 82148, + "Ġkicks": 36250, + "Ġkid": 10369, + "Ġkidd": 91415, + "Ġkidding": 55725, + "Ġkidn": 32698, + "Ġkidnapped": 54845, + "Ġkidnapping": 61682, + "Ġkidney": 37942, + "Ġkidneys": 80860, + "Ġkids": 6837, + "Ġkijken": 77957, + "Ġkil": 15045, + "Ġkill": 5505, + "Ġkilled": 7425, + "Ġkiller": 24518, + "Ġkillers": 60035, + "Ġkilling": 13118, + "Ġkillings": 47033, + "Ġkills": 28810, + "Ġkilograms": 84302, + "Ġkilomet": 43887, + "Ġkilometers": 40568, + "Ġkilometres": 51857, + "Ġkim": 55784, + "Ġkin": 23938, + "Ġkinase": 81060, + "Ġkind": 3093, + "Ġkinda": 33390, + "Ġkindergarten": 67123, + "Ġkindly": 46526, + "Ġkindness": 44872, + "Ġkinds": 12829, + "Ġkinetic": 70323, + "Ġkinetics": 90368, + "Ġking": 11477, + "Ġkingdom": 25079, + "Ġkingdoms": 95406, + "Ġkings": 44519, + "Ġkinky": 66776, + "Ġkino": 89255, + "Ġkir": 72630, + "Ġkiss": 21057, + "Ġkissed": 58234, + "Ġkisses": 71991, + "Ġkissing": 51046, + "Ġkit": 16138, + "Ġkita": 54211, + "Ġkitabı": 138893, + "Ġkitchen": 9780, + "Ġkitchens": 65338, + "Ġkite": 98119, + "Ġkits": 31496, + "Ġkitt": 66612, + "Ġkitten": 76657, + "Ġkittens": 82090, + "Ġkitty": 79821, + "Ġkiên": 139824, + "ĠkiÅŁ": 127464, + "ĠkiÅŁi": 129043, + "ĠkiÅŁiler": 133354, + "ĠkiÅŁinin": 137522, + "Ġkiếm": 130158, + "Ġkiến": 128935, + "Ġkiá»ĥm": 129522, + "Ġkiá»ĥu": 132310, + "Ġkiá»ĩm": 135159, + "Ġkiá»ĩn": 129065, + "Ġkj": 42041, + "Ġkk": 49865, + "Ġkl": 20226, + "Ġkla": 75886, + "Ġklar": 72734, + "Ġklass": 37955, + "Ġkle": 24285, + "Ġklein": 79409, + "Ġkleine": 53555, + "Ġkleinen": 80378, + "Ġklient": 90108, + "Ġklientów": 138341, + "Ġklik": 89909, + "Ġklub": 82690, + "Ġkm": 13136, + "Ġkms": 96677, + "Ġkn": 1148, + "Ġknack": 93250, + "Ġkne": 13853, + "Ġknee": 21381, + "Ġkneeling": 99015, + "Ġknees": 30524, + "Ġknew": 6876, + "Ġknex": 63355, + "Ġknife": 21430, + "Ġknight": 46609, + "Ġknights": 70530, + "Ġknit": 52103, + "Ġknitting": 71799, + "Ġknives": 48895, + "Ġknob": 58572, + "Ġknobs": 89085, + "Ġknock": 14127, + "Ġknocked": 31985, + "Ġknocking": 49144, + "Ġknockout": 76073, + "Ġknocks": 84669, + "Ġknot": 50800, + "Ġknots": 60217, + "Ġknow": 1414, + "Ġknowing": 14063, + "Ġknowingly": 67013, + "Ġknowledge": 6540, + "Ġknowledgeable": 40966, + "Ġknown": 3881, + "Ġknows": 8788, + "Ġknull": 85949, + "Ġko": 15236, + "Ġkob": 72251, + "Ġkod": 65532, + "Ġkode": 83477, + "Ġkoje": 96499, + "Ġkoji": 59371, + "Ġkok": 89537, + "Ġkol": 47119, + "Ġkole": 58019, + "Ġkolej": 82322, + "Ġkom": 10890, + "Ġkomb": 84464, + "Ġkomen": 81737, + "Ġkomm": 93688, + "Ġkommen": 58065, + "Ġkomment": 83034, + "Ġkommer": 47331, + "Ġkommt": 51528, + "Ġkommun": 87943, + "Ġkomple": 83769, + "Ġkomt": 72633, + "Ġkon": 16540, + "Ġkond": 79164, + "Ġkone": 81690, + "Ġkonk": 96815, + "Ġkonnte": 64015, + "Ġkont": 20153, + "Ġkontakt": 40746, + "Ġkontakte": 70312, + "Ġkontrol": 96034, + "ĠkonuÅŁ": 126065, + "ĠkonuÅŁma": 134180, + "ĠkonuÅŁtu": 134271, + "Ġkop": 81804, + "Ġkor": 31954, + "Ġkort": 66821, + "Ġkos": 40397, + "Ġkosher": 97640, + "Ġkost": 62411, + "Ġkosten": 18926, + "Ġkostenlos": 26073, + "Ġkostenlose": 48522, + "Ġkot": 74925, + "Ġkotlin": 21527, + "Ġkotlinx": 27452, + "ĠkoÅĦ": 127982, + "ĠkoÅĦca": 139368, + "ĠkoÅĦcu": 138434, + "ĠkoÅĽci": 137592, + "ĠkoÅŁ": 136969, + "Ġkp": 60319, + "Ġkr": 23103, + "Ġkra": 96411, + "Ġkrat": 91014, + "Ġkrb": 79132, + "Ġkre": 81070, + "Ġkrij": 69437, + "Ġkrist": 85848, + "Ġkron": 87932, + "Ġkró": 137234, + "Ġkról": 141796, + "Ġkrótk": 137235, + "Ġks": 41282, + "Ġksi": 87527, + "ĠksiÄħż": 127703, + "ĠksiÄħżki": 139679, + "ĠksiÄĻ": 143722, + "ĠksiÄĻg": 143723, + "Ġksz": 141368, + "ĠksztaÅĤ": 141369, + "ĠksztaÅĤt": 141370, + "Ġkt": 18541, + "Ġkter": 99128, + "Ġkto": 135727, + "ĠktoÅĽ": 135728, + "Ġktó": 21342, + "Ġktór": 135685, + "Ġktóra": 82121, + "Ġktóre": 42890, + "Ġktórego": 133031, + "Ġktórej": 131252, + "Ġktóry": 41842, + "Ġktórych": 129925, + "Ġktórym": 130235, + "Ġktórzy": 130404, + "ĠktórÄħ": 135686, + "Ġku": 36600, + "Ġkube": 80958, + "Ġkuk": 93277, + "Ġkul": 74416, + "Ġkull": 142134, + "Ġkullan": 57016, + "Ġkulland": 142135, + "Ġkullandı": 142136, + "Ġkullanıcı": 133558, + "Ġkullanıl": 130002, + "Ġkullanılan": 138287, + "Ġkullanım": 138876, + "Ġkullanımı": 140958, + "Ġkun": 23330, + "Ġkunne": 79190, + "Ġkunnen": 48436, + "Ġkunt": 61777, + "Ġkup": 80708, + "Ġkur": 27265, + "Ġkurs": 92579, + "Ġkurulu": 136634, + "ĠkuruluÅŁ": 136635, + "Ġkurz": 70984, + "Ġkut": 30599, + "Ġkutje": 88343, + "Ġkv": 16178, + "Ġkvin": 40061, + "Ġkvinde": 78492, + "Ġkvinder": 41425, + "Ġkvinn": 49973, + "Ġkvinna": 96331, + "Ġkvinne": 75503, + "Ġkvinner": 47084, + "Ġkvinnor": 66830, + "Ġkvm": 94748, + "Ġkvp": 85555, + "Ġkw": 29525, + "Ġkwargs": 16494, + "Ġky": 54953, + "Ġkz": 80955, + "Ġkä": 59095, + "ĠkÃ¥": 78724, + "Ġkè": 136943, + "Ġkèm": 136944, + "Ġké": 76438, + "Ġkém": 135153, + "Ġkéo": 131778, + "Ġkê": 127444, + "Ġkênh": 142770, + "Ġkì": 136542, + "Ġkö": 13040, + "Ġköln": 92571, + "Ġkön": 41762, + "Ġkönnen": 20048, + "Ġkönnt": 142580, + "Ġkönnte": 67594, + "Ġkötü": 134043, + "Ġköz": 67824, + "Ġkø": 34400, + "Ġkøbenhavn": 44080, + "Ġkü": 65948, + "Ġkültür": 134753, + "Ġküç": 132619, + "Ġküçük": 132620, + "Ġký": 129212, + "ĠkÃŃch": 130767, + "ĠkÃŃnh": 131921, + "Ġkı": 127100, + "Ġkısa": 131370, + "Ġkısm": 135942, + "Ġkısmı": 135943, + "Ġkısı": 139364, + "Ġkız": 130773, + "Ġkẻ": 134248, + "Ġkế": 128495, + "Ġkết": 128403, + "Ġkỳ": 128880, + "Ġká»·": 132551, + "Ġkỹ": 129531, + "Ġká»ĥ": 130291, + "Ġká»ĭ": 137287, + "Ġká»ĭch": 142665, + "Ġká»ĭp": 137288, + "Ġl": 326, + "ĠlParam": 51834, + "Ġla": 1187, + "Ġlaat": 72399, + "Ġlab": 10072, + "Ġlabel": 2383, + "ĠlabelText": 64710, + "Ġlabeled": 29829, + "Ġlabeling": 54302, + "Ġlabelled": 60236, + "Ġlabels": 9201, + "Ġlabor": 9327, + "Ġlaboratories": 69660, + "Ġlaboratory": 26594, + "Ġlabore": 72204, + "Ġlabour": 22939, + "Ġlabs": 49948, + "Ġlabyrinth": 98710, + "Ġlac": 43879, + "Ġlace": 45693, + "Ġlack": 6853, + "Ġlacked": 48001, + "Ġlacking": 31061, + "Ġlacks": 36756, + "Ġlact": 50544, + "Ġlad": 57625, + "Ġladder": 35765, + "Ġladen": 81687, + "Ġladies": 22786, + "Ġlado": 43424, + "Ġlady": 16820, + "Ġlag": 21454, + "Ġlagi": 73944, + "Ġlah": 89045, + "Ġlaid": 17113, + "Ġlain": 50376, + "Ġlaisse": 92469, + "Ġlak": 93686, + "Ġlake": 21800, + "Ġlakes": 43136, + "Ġlakh": 62173, + "Ġlam": 31603, + "Ġlamb": 38306, + "Ġlambda": 12459, + "Ġlame": 60883, + "Ġlament": 50372, + "Ġlamin": 78433, + "Ġlaminate": 84834, + "Ġlamp": 27962, + "Ġlamps": 49452, + "Ġlan": 30691, + "Ġlanc": 75285, + "Ġlance": 85245, + "Ġland": 4268, + "Ġlanded": 26120, + "Ġlandfill": 84534, + "Ġlanding": 20327, + "Ġlandlord": 40147, + "Ġlandlords": 63629, + "Ġlandmark": 37250, + "Ġlandmarks": 59924, + "Ġlands": 11835, + "Ġlandsc": 33515, + "Ġlandscape": 18414, + "Ġlandscapes": 54790, + "Ġlandscaping": 64018, + "Ġlandslide": 95702, + "Ġlane": 21272, + "Ġlanes": 33025, + "Ġlang": 8688, + "Ġlange": 52287, + "Ġlangs": 63952, + "Ġlangu": 82300, + "Ġlanguage": 4128, + "Ġlanguages": 15459, + "Ġlangue": 95182, + "Ġlantern": 73165, + "Ġlanz": 83146, + "Ġlanç": 95969, + "Ġlança": 132763, + "Ġlap": 23410, + "Ġlaps": 49955, + "Ġlapse": 89317, + "Ġlaptop": 20638, + "Ġlaptops": 47885, + "Ġlar": 44455, + "Ġlarg": 4053, + "Ġlarge": 3460, + "Ġlargely": 13771, + "Ġlarger": 8131, + "Ġlargest": 7772, + "Ġlargo": 63913, + "Ġlarvae": 82761, + "Ġlas": 5141, + "Ġlasc": 80880, + "Ġlaser": 20486, + "Ġlasers": 71375, + "Ġlash": 97281, + "Ġlashes": 86611, + "Ġlassen": 47008, + "Ġlast": 1537, + "ĠlastIndex": 86972, + "ĠlastName": 20460, + "Ġlasted": 35413, + "Ġlasting": 28769, + "Ġlastname": 45341, + "Ġlasts": 39658, + "Ġlat": 6844, + "Ġlatch": 56926, + "Ġlate": 3309, + "Ġlateinit": 26130, + "Ġlately": 30345, + "Ġlaten": 92532, + "Ġlatency": 39270, + "Ġlatent": 41667, + "Ġlater": 2937, + "Ġlateral": 44469, + "Ġlatest": 5535, + "Ġlatex": 44536, + "Ġlatin": 62919, + "Ġlatina": 57315, + "Ġlatino": 91882, + "Ġlatitude": 20849, + "Ġlatter": 15271, + "Ġlattice": 54272, + "Ġlaugh": 12550, + "Ġlaughed": 31527, + "Ġlaughing": 31581, + "Ġlaughs": 48236, + "Ġlaughter": 42114, + "Ġlaunch": 7050, + "Ġlaunched": 11626, + "Ġlauncher": 46449, + "Ġlaunches": 37075, + "Ġlaunching": 27718, + "Ġlaunder": 53038, + "Ġlaundering": 63302, + "Ġlaundry": 34089, + "Ġlaure": 66085, + "Ġlaut": 93539, + "Ġlav": 29483, + "Ġlava": 57568, + "Ġlavender": 80360, + "Ġlavish": 79134, + "Ġlavor": 45055, + "Ġlavoro": 65385, + "Ġlaw": 2329, + "Ġlawful": 51221, + "Ġlawmaker": 70376, + "Ġlawmakers": 25081, + "Ġlawn": 36025, + "Ġlaws": 6872, + "Ġlawsuit": 19275, + "Ġlawsuits": 42886, + "Ġlawy": 11010, + "Ġlawyer": 15417, + "Ġlawyers": 21177, + "Ġlax": 73730, + "Ġlay": 10962, + "Ġlayer": 6193, + "Ġlayered": 63141, + "Ġlayers": 13617, + "Ġlaying": 34644, + "Ġlayoffs": 98822, + "Ġlayout": 6789, + "ĠlayoutManager": 84637, + "ĠlayoutParams": 94644, + "Ġlayouts": 49054, + "Ġlays": 56522, + "Ġlayui": 67885, + "Ġlaz": 64436, + "Ġlazy": 15678, + "Ġlazım": 140507, + "Ġlb": 18866, + "Ġlbl": 16421, + "Ġlbs": 28060, + "Ġlc": 36213, + "Ġlcd": 38644, + "Ġlcm": 86079, + "Ġld": 26129, + "Ġlda": 68131, + "Ġldap": 67759, + "Ġldb": 97789, + "Ġldc": 32006, + "Ġle": 512, + "Ġlead": 2990, + "Ġleader": 7653, + "Ġleaderboard": 86453, + "Ġleaders": 6036, + "Ġleadership": 11438, + "Ġleading": 6388, + "Ġleads": 11508, + "Ġleaf": 15933, + "Ġleague": 10734, + "Ġleagues": 38359, + "Ġleak": 23352, + "Ġleakage": 80273, + "Ġleaked": 33847, + "Ġleaking": 51829, + "Ġleaks": 36696, + "Ġlean": 15651, + "Ġleaned": 54839, + "Ġleaning": 48348, + "Ġleans": 92836, + "Ġleap": 31471, + "Ġleaps": 83458, + "Ġlear": 34399, + "Ġlearn": 3960, + "Ġlearned": 9498, + "Ġlearner": 62960, + "Ġlearners": 52143, + "Ġlearning": 6832, + "Ġlearns": 46210, + "Ġlearnt": 49250, + "Ġlease": 25064, + "Ġleased": 81180, + "Ġleases": 72557, + "Ġleash": 78901, + "Ġleasing": 70319, + "Ġleast": 3245, + "Ġleather": 17553, + "Ġleave": 5274, + "Ġleaves": 10901, + "Ġleaving": 9380, + "Ġleben": 97872, + "Ġlebih": 49935, + "Ġlecken": 92561, + "Ġlect": 16513, + "Ġlecture": 30578, + "Ġlecturer": 71135, + "Ġlectures": 45132, + "Ġled": 6069, + "Ġledge": 85833, + "Ġledger": 46933, + "Ġleds": 90972, + "Ġlee": 81944, + "Ġleer": 53138, + "Ġleft": 2115, + "Ġleftist": 71127, + "Ġleftover": 64443, + "Ġleftovers": 92167, + "Ġleg": 2472, + "Ġlegacy": 19588, + "Ġlegal": 5777, + "Ġlegality": 88746, + "Ġlegalization": 56488, + "Ġlegalize": 87913, + "Ġlegalized": 76992, + "Ġlegally": 25202, + "Ġlegend": 13016, + "Ġlegendary": 27712, + "Ġlegends": 48328, + "Ġleggings": 90341, + "Ġlegion": 94168, + "Ġlegis": 8529, + "Ġlegisl": 14963, + "Ġlegislation": 13240, + "Ġlegislative": 26645, + "Ġlegislators": 47862, + "Ġlegislature": 38756, + "Ġlegit": 15563, + "Ġlegitim": 48197, + "Ġlegitimacy": 55908, + "Ġlegitimate": 22745, + "Ġlegitimately": 87251, + "Ġlegs": 14201, + "Ġlehet": 76132, + "Ġlei": 59962, + "Ġleicht": 71935, + "Ġleider": 88028, + "Ġleisure": 40208, + "Ġlekker": 76547, + "Ġlem": 83974, + "Ġlemma": 18810, + "Ġlemon": 29464, + "Ġlen": 2422, + "Ġlend": 38480, + "Ġlender": 47475, + "Ġlenders": 45015, + "Ġlending": 39551, + "Ġlends": 77918, + "Ġleng": 47794, + "Ġlenght": 83254, + "Ġlength": 3084, + "Ġlengths": 28316, + "Ġlengthy": 34206, + "Ġlens": 18342, + "Ġlenses": 34957, + "Ġlent": 47807, + "Ġleo": 88501, + "Ġleopard": 97993, + "Ġlept": 94440, + "Ġler": 75020, + "Ġles": 3541, + "Ġlesb": 22783, + "Ġlesbi": 81227, + "Ġlesbian": 24054, + "Ġlesbians": 58630, + "Ġlesbienne": 60519, + "Ġlesbische": 97854, + "Ġlesbisk": 91556, + "Ġlesen": 75182, + "Ġlesion": 92117, + "Ġlesions": 62224, + "Ġless": 2686, + "Ġlessen": 75870, + "Ġlesser": 31315, + "Ġlesson": 17755, + "Ġlessons": 18366, + "Ġlest": 63506, + "Ġlet": 1077, + "Ġleth": 97422, + "Ġlethal": 43989, + "Ġletra": 66417, + "Ġletras": 89358, + "Ġlets": 15354, + "Ġlett": 50581, + "Ġletter": 6524, + "ĠletterSpacing": 87999, + "Ġletters": 11931, + "Ġletting": 20194, + "Ġletto": 97899, + "Ġlettre": 69456, + "Ġlettuce": 70555, + "Ġletz": 49765, + "Ġletzten": 66968, + "Ġleuk": 56281, + "Ġleukemia": 95206, + "Ġleur": 27032, + "Ġleurs": 42053, + "Ġlev": 22638, + "Ġleve": 55489, + "Ġlevel": 2188, + "Ġleveled": 82249, + "Ġleveling": 72985, + "Ġlevels": 5866, + "Ġleven": 92203, + "Ġlever": 27505, + "Ġleverage": 32064, + "Ġleveraging": 76482, + "Ġlevitra": 67920, + "Ġlevy": 76062, + "Ġlex": 22429, + "Ġlexer": 53259, + "Ġlexical": 77586, + "Ġley": 60415, + "Ġlf": 58080, + "Ġlg": 26088, + "Ġlh": 72273, + "Ġlhs": 22505, + "Ġli": 898, + "Ġlia": 54816, + "Ġliabilities": 57065, + "Ġliability": 23414, + "Ġliable": 34373, + "Ġliaison": 82099, + "Ġliar": 72624, + "Ġlib": 3051, + "Ġlibc": 42142, + "Ġliber": 33829, + "Ġliberal": 17775, + "Ġliberalism": 83481, + "Ġliberals": 44650, + "Ġliberated": 84755, + "Ġliberation": 54533, + "Ġlibero": 79486, + "Ġlibert": 22073, + "Ġlibertarian": 56025, + "Ġliberties": 57355, + "Ġlibertin": 62632, + "Ġlibertine": 69149, + "Ġliberty": 30120, + "Ġliberté": 142341, + "Ġlibido": 93382, + "Ġlibr": 60359, + "Ġlibrarian": 94207, + "Ġlibraries": 20186, + "Ġlibrary": 6733, + "Ġlibre": 47547, + "Ġlibro": 51011, + "Ġlibros": 72346, + "Ġlibs": 63974, + "Ġlic": 12273, + "Ġlicence": 29973, + "Ġlicences": 91984, + "Ġlicens": 66236, + "Ġlicense": 5723, + "Ġlicensed": 16000, + "Ġlicensee": 85634, + "Ġlicenses": 21067, + "Ġlicensing": 27406, + "Ġlicensors": 88104, + "Ġlick": 62199, + "Ġlicked": 92935, + "Ġlicking": 66131, + "Ġlicz": 84521, + "Ġlid": 26334, + "Ġlider": 67319, + "Ġlids": 98978, + "Ġlie": 10246, + "Ġliebe": 91342, + "Ġlied": 46153, + "Ġlief": 82590, + "Ġliegt": 63940, + "Ġlien": 54420, + "Ġlies": 15448, + "Ġlieu": 38281, + "Ġlieutenant": 73536, + "Ġlif": 10136, + "Ġlife": 2272, + "Ġlifecycle": 47508, + "Ġlifelong": 50163, + "Ġlifes": 48685, + "Ġlifespan": 60861, + "Ġlifestyle": 18899, + "Ġlifestyles": 78631, + "Ġlifetime": 19031, + "Ġlift": 11893, + "Ġlifted": 29731, + "Ġlifting": 32410, + "Ġlifts": 53066, + "Ġlig": 28313, + "Ġlight": 3100, + "Ġlighten": 82508, + "Ġlighter": 29573, + "Ġlighting": 17716, + "Ġlightly": 33404, + "Ġlightning": 32438, + "Ġlights": 12711, + "Ġlightweight": 29144, + "Ġlign": 71600, + "Ġligne": 26814, + "Ġlij": 72684, + "Ġlik": 3130, + "Ġlike": 1075, + "Ġliked": 14915, + "Ġlikelihood": 28636, + "Ġlikely": 4363, + "Ġliken": 54285, + "Ġlikeness": 92774, + "Ġlikes": 13151, + "Ġlikewise": 37922, + "Ġliking": 48737, + "Ġlil": 61371, + "Ġlille": 88022, + "Ġlim": 4568, + "Ġlimb": 47594, + "Ġlimbs": 48595, + "Ġlimburg": 60231, + "Ġlime": 41719, + "Ġlimestone": 43916, + "Ġlimit": 3930, + "Ġlimitation": 20278, + "Ġlimitations": 9481, + "Ġlimite": 69980, + "Ġlimited": 7199, + "Ġlimiting": 32894, + "Ġlimitless": 91559, + "Ġlimits": 13388, + "Ġlimp": 82278, + "Ġlin": 9805, + "Ġline": 1555, + "ĠlineHeight": 57592, + "ĠlineNumber": 58941, + "ĠlineWidth": 72840, + "Ġlinea": 53856, + "Ġlineage": 63909, + "Ġlinear": 13482, + "Ġlineback": 40701, + "Ġlinebacker": 47700, + "Ġlined": 31293, + "Ġlineman": 81673, + "Ġlinen": 54015, + "Ġlineno": 85095, + "Ġliner": 52988, + "Ġliners": 95374, + "Ġlines": 5128, + "Ġlinestyle": 68541, + "Ġlineup": 27512, + "Ġlinewidth": 47847, + "Ġling": 37072, + "Ġlinger": 39509, + "Ġlingerie": 56174, + "Ġlingering": 63510, + "Ġlingu": 38503, + "Ġlinguistic": 64667, + "Ġlinha": 47673, + "Ġlining": 35371, + "Ġlink": 2656, + "Ġlinkage": 71441, + "Ġlinked": 10592, + "Ġlinker": 85745, + "Ġlinking": 30699, + "Ġlinks": 7746, + "Ġlinspace": 85796, + "Ġlint": 57920, + "Ġlinux": 36245, + "Ġlion": 39032, + "Ġlions": 68032, + "Ġlip": 19050, + "Ġlipid": 67600, + "Ġlips": 22877, + "Ġlipstick": 73583, + "Ġliqu": 11784, + "Ġlique": 92442, + "Ġliquid": 14473, + "Ġliquidity": 52013, + "Ġliquids": 66749, + "Ġliquor": 44204, + "Ġlire": 77451, + "Ġlis": 40280, + "Ġlist": 1140, + "ĠlistBox": 79328, + "ĠlistItem": 72132, + "ĠlistOf": 35891, + "ĠlistView": 40628, + "Ġlista": 14124, + "Ġlistar": 92868, + "Ġliste": 32924, + "Ġlisted": 10007, + "Ġlisten": 8844, + "Ġlistened": 33693, + "Ġlistener": 11446, + "Ġlisteners": 23562, + "Ġlistening": 14289, + "Ġlistens": 63648, + "Ġlisting": 14838, + "Ġlistings": 25465, + "Ġlists": 11469, + "Ġlit": 13020, + "Ġlite": 49672, + "Ġliter": 6935, + "Ġliteracy": 51982, + "Ġliteral": 23141, + "Ġliterally": 15901, + "Ġliterals": 75275, + "Ġliterary": 31365, + "Ġliterature": 17206, + "Ġliters": 92866, + "Ġlith": 45182, + "Ġlithium": 56807, + "Ġlitigation": 38625, + "Ġlitre": 91369, + "Ġlitres": 87916, + "Ġlitt": 59542, + "Ġlitter": 38582, + "Ġlittle": 2632, + "Ġliv": 15120, + "Ġlive": 3887, + "Ġlived": 12163, + "Ġlivelihood": 63651, + "Ġlively": 48177, + "Ġliver": 25506, + "Ġlives": 6305, + "Ġlivest": 41812, + "Ġlivestock": 50776, + "Ġliving": 5382, + "Ġlivre": 55884, + "Ġlivro": 81423, + "Ġlizard": 94811, + "Ġlié": 141035, + "Ġliên": 128542, + "Ġliá»ģn": 132726, + "Ġliá»ĩt": 137513, + "Ġliá»ĩu": 88057, + "Ġlj": 91502, + "Ġlk": 79016, + "Ġll": 9323, + "Ġllam": 50104, + "Ġllama": 93676, + "Ġlle": 56915, + "Ġlleg": 32502, + "Ġllegar": 74783, + "Ġllegó": 143047, + "Ġllen": 78831, + "Ġllev": 53195, + "Ġllevar": 69580, + "Ġllvm": 34664, + "Ġlm": 40238, + "Ġln": 29390, + "Ġlname": 76139, + "Ġlng": 24820, + "Ġlo": 775, + "Ġload": 2795, + "ĠloadChildren": 63035, + "ĠloadData": 62432, + "ĠloadImage": 69589, + "Ġloaded": 6661, + "Ġloader": 16047, + "Ġloaders": 68574, + "Ġloading": 8277, + "Ġloads": 20907, + "Ġloaf": 76148, + "Ġloan": 11679, + "Ġloans": 16606, + "Ġlob": 40153, + "Ġlobby": 19307, + "Ġlobbying": 46238, + "Ġlobbyist": 90469, + "Ġlobbyists": 72342, + "Ġlobster": 79715, + "Ġloc": 1329, + "Ġlocal": 2205, + "ĠlocalObject": 58272, + "ĠlocalStorage": 15566, + "ĠlocalVar": 42766, + "Ġlocale": 18565, + "Ġlocales": 52297, + "Ġlocalhost": 47422, + "Ġlocality": 68087, + "Ġlocalization": 52304, + "Ġlocalize": 94416, + "Ġlocalized": 43489, + "Ġlocally": 23490, + "Ġlocals": 24914, + "Ġlocaltime": 87494, + "Ġlocate": 24523, + "Ġlocated": 7407, + "Ġlocating": 71891, + "Ġlocation": 3728, + "ĠlocationManager": 68732, + "Ġlocations": 10468, + "Ġlocator": 47117, + "Ġlock": 5296, + "Ġlockdown": 50135, + "Ġlocked": 16061, + "Ġlocker": 34679, + "Ġlocking": 37855, + "Ġlocks": 31676, + "Ġlocksmith": 97960, + "Ġlocom": 62807, + "Ġlocus": 78157, + "Ġlod": 35032, + "Ġlodash": 89464, + "Ġlodge": 59791, + "Ġlodged": 73632, + "Ġlodging": 76647, + "Ġloft": 58427, + "Ġlofty": 86691, + "Ġlog": 1487, + "Ġlogarith": 89936, + "Ġlogfile": 68374, + "Ġlogged": 13726, + "ĠloggedIn": 69546, + "Ġlogger": 5925, + "Ġlogging": 8392, + "Ġlogic": 12218, + "Ġlogical": 19819, + "Ġlogically": 73045, + "Ġlogin": 5858, + "ĠloginUser": 87169, + "Ġlogistic": 71710, + "Ġlogistical": 95868, + "Ġlogistics": 42157, + "Ġlogits": 60788, + "Ġlogo": 12426, + "Ġlogos": 49019, + "Ġlogout": 22359, + "Ġlogs": 18422, + "Ġloi": 81191, + "Ġloin": 76960, + "Ġlok": 40283, + "Ġlokal": 84544, + "Ġlokale": 93887, + "Ġlol": 27409, + "Ġlombok": 16717, + "Ġlon": 17806, + "Ġlondon": 62046, + "Ġlone": 46666, + "Ġloneliness": 73940, + "Ġlonely": 39566, + "Ġlong": 1293, + "Ġlonger": 5021, + "Ġlongest": 22032, + "Ġlongevity": 57119, + "Ġlonging": 78322, + "Ġlongitud": 48604, + "Ġlongitude": 20515, + "Ġlongitudinal": 67002, + "Ġlongstanding": 73129, + "Ġlongtime": 35404, + "Ġlook": 1401, + "Ġlookahead": 86745, + "Ġlooked": 6966, + "Ġlooking": 3330, + "Ġlookout": 70915, + "Ġlooks": 5868, + "Ġlookup": 18615, + "Ġlooming": 64464, + "Ġloop": 6337, + "Ġlooph": 61476, + "Ġloophole": 92995, + "Ġlooping": 62587, + "Ġloops": 29753, + "Ġloos": 44950, + "Ġloose": 20174, + "Ġloosely": 62457, + "Ġloosen": 82742, + "Ġloot": 45303, + "Ġlor": 66211, + "Ġlord": 36931, + "Ġlords": 89113, + "Ġlore": 51222, + "Ġlorem": 92385, + "Ġloro": 47093, + "Ġlors": 36853, + "Ġlorsque": 86058, + "Ġlos": 2478, + "Ġlose": 9052, + "Ġloser": 58891, + "Ġlosers": 67356, + "Ġloses": 32191, + "Ġlosing": 13188, + "Ġloss": 4709, + "Ġlosses": 17683, + "Ġlost": 5558, + "Ġlot": 2696, + "Ġlotion": 86842, + "Ġlots": 10077, + "Ġlottery": 38239, + "Ġlou": 28640, + "Ġloud": 17361, + "Ġlouder": 61796, + "Ġloudly": 53845, + "Ġlouis": 96130, + "Ġloung": 83144, + "Ġlounge": 37932, + "Ġlov": 97792, + "Ġlove": 2948, + "Ġloved": 10245, + "Ġlovely": 16690, + "Ġlover": 30557, + "Ġlovers": 32118, + "Ġloves": 15803, + "Ġloving": 21258, + "Ġlow": 3347, + "Ġlower": 4722, + "Ġlowercase": 42047, + "Ġlowered": 38820, + "Ġlowering": 45201, + "Ġlowers": 72015, + "Ġlowes": 96827, + "Ġlowest": 15457, + "Ġlows": 65146, + "Ġloyal": 28847, + "Ġloyalty": 31783, + "Ġloại": 128433, + "Ġloạn": 138641, + "Ġloạt": 137634, + "Ġlp": 18576, + "Ġlr": 27060, + "Ġls": 19597, + "Ġlsp": 98679, + "Ġlst": 18845, + "Ġlstm": 97805, + "Ġlt": 25175, + "Ġltd": 70893, + "Ġlu": 24743, + "Ġlua": 20357, + "ĠluaL": 72731, + "Ġlub": 27345, + "Ġlubric": 53394, + "Ġluc": 25927, + "Ġluck": 15019, + "Ġluckily": 90045, + "Ġlucky": 17605, + "Ġlucr": 45190, + "Ġlucrative": 50206, + "Ġlud": 43964, + "Ġluder": 94998, + "Ġludicrous": 89171, + "Ġluego": 60394, + "Ġlug": 53410, + "Ġlugar": 33900, + "Ġlugares": 92824, + "Ġluggage": 50497, + "Ġlui": 24397, + "Ġluk": 81280, + "Ġlul": 68063, + "Ġlum": 40163, + "Ġlumber": 61453, + "Ġlumin": 44958, + "Ġlumière": 141096, + "Ġlump": 48529, + "Ġlun": 31973, + "Ġlunar": 56711, + "Ġlunch": 15786, + "Ġlunches": 93630, + "Ġlung": 20622, + "Ġlungs": 44174, + "Ġlup": 98789, + "Ġlur": 46647, + "Ġlure": 56583, + "Ġlurking": 81975, + "Ġlush": 57267, + "Ġlust": 40102, + "Ġlut": 62822, + "Ġlux": 14050, + "Ġluxe": 96100, + "Ġluxurious": 42728, + "Ġluxury": 19352, + "Ġluyá»ĩn": 133247, + "Ġluz": 64466, + "Ġluôn": 128787, + "ĠluáºŃn": 129748, + "ĠluáºŃt": 129963, + "Ġlv": 28378, + "Ġlvl": 46129, + "Ġlw": 41073, + "Ġlx": 63244, + "Ġlxml": 98723, + "Ġly": 14528, + "Ġlying": 20446, + "Ġlymph": 42645, + "Ġlyn": 89203, + "Ġlyon": 88430, + "Ġlyr": 85237, + "Ġlyric": 68335, + "Ġlyrics": 23261, + "Ġlys": 83395, + "Ġlz": 95982, + "Ġlá": 53638, + "Ġlâu": 129986, + "Ġlã": 131399, + "Ġlãi": 140694, + "Ġlãnh": 131400, + "Ġlä": 30005, + "Ġläng": 96976, + "Ġlänger": 139975, + "Ġlässt": 73741, + "ĠlÃ¥": 69369, + "Ġlæ": 77214, + "Ġlé": 62728, + "Ġlég": 71983, + "Ġlégère": 143979, + "Ġlên": 128350, + "Ġlòng": 98170, + "Ġlö": 76133, + "Ġlø": 80020, + "Ġlúc": 128730, + "Ġlý": 128356, + "ĠlÃł": 37915, + "ĠlÃłm": 128266, + "ĠlÃłnh": 136382, + "ĠlÃŃ": 31750, + "ĠlÃŃder": 94696, + "ĠlÃŃnea": 55162, + "ĠlÄ©nh": 132122, + "Ġlưng": 137643, + "Ġlưu": 129480, + "Ġlương": 135803, + "Ġlược": 136312, + "Ġlượng": 128375, + "Ġlượt": 134628, + "ĠlÆ°á»Ľi": 141593, + "Ġlạ": 132143, + "Ġlạc": 133043, + "Ġlại": 85852, + "Ġlạnh": 130823, + "Ġlấy": 129483, + "Ġlần": 128568, + "Ġlẫn": 138618, + "Ġlắm": 133883, + "Ġlắng": 135378, + "Ġlắp": 135765, + "Ġlặng": 141431, + "Ġlẽ": 132358, + "ĠláºŃp": 128980, + "Ġlợi": 128994, + "Ġlá»±a": 129506, + "Ġlá»±c": 128599, + "Ġlá»ħ": 131309, + "Ġlá»ĩ": 130554, + "Ġlá»ĩnh": 139081, + "Ġlá»ĭch": 128717, + "Ġlá»įc": 137154, + "Ġlá»Ĺi": 133015, + "Ġlá»Ļ": 130540, + "ĠlỼ": 126255, + "ĠlỼn": 128465, + "ĠlỼp": 130192, + "Ġlá»Ŀi": 128814, + "Ġlá»Ńa": 136337, + "Ġm": 296, + "ĠmA": 98777, + "ĠmActivity": 98925, + "ĠmAdapter": 62876, + "ĠmAuth": 87912, + "ĠmContext": 38674, + "ĠmCurrent": 87436, + "ĠmData": 63587, + "ĠmHandler": 82863, + "ĠmL": 64070, + "ĠmList": 87680, + "ĠmListener": 82396, + "ĠmM": 83217, + "ĠmMap": 74470, + "ĠmName": 96550, + "ĠmRNA": 77772, + "ĠmRecyclerView": 91523, + "ĠmView": 53585, + "Ġma": 7491, + "Ġmaar": 31591, + "Ġmac": 8978, + "ĠmacOS": 67178, + "Ġmach": 7845, + "Ġmachen": 37896, + "Ġmachine": 5662, + "Ġmachinery": 25868, + "Ġmachines": 12645, + "Ġmachining": 97323, + "Ġmacht": 52549, + "Ġmacro": 18072, + "Ġmacros": 37714, + "Ġmad": 12796, + "Ġmade": 1865, + "Ġmadness": 51719, + "Ġmadre": 70774, + "Ġmadrid": 64787, + "Ġmaduras": 70163, + "Ġmafia": 89885, + "Ġmag": 4878, + "Ġmagazine": 14418, + "Ġmagazines": 31847, + "Ġmage": 51803, + "Ġmagg": 87899, + "Ġmagic": 10963, + "Ġmagical": 23702, + "Ġmagically": 77725, + "Ġmagician": 89341, + "Ġmagistrate": 94091, + "Ġmagma": 63505, + "Ġmagn": 8455, + "Ġmagna": 58917, + "Ġmagnesium": 60833, + "Ġmagnet": 32197, + "Ġmagnetic": 23970, + "Ġmagnets": 72680, + "Ġmagnificent": 40692, + "Ġmagnitude": 25621, + "Ġmah": 43216, + "Ġmai": 16737, + "Ġmaid": 42065, + "Ġmaiden": 73222, + "Ġmail": 8072, + "Ġmailbox": 45742, + "Ġmailed": 75210, + "Ġmailing": 32759, + "Ġmails": 85558, + "Ġmain": 1887, + "ĠmainAxisAlignment": 33692, + "ĠmainBundle": 77258, + "ĠmainScreen": 66406, + "ĠmainWindow": 68137, + "Ġmainland": 50015, + "Ġmainly": 14576, + "Ġmains": 68136, + "Ġmainstream": 20729, + "Ġmaint": 4981, + "Ġmaintain": 10306, + "Ġmaintained": 18401, + "Ġmaintaining": 20337, + "Ġmaintains": 31995, + "Ġmaintenance": 13404, + "Ġmaintenant": 73599, + "Ġmaior": 45491, + "Ġmais": 9870, + "Ġmaison": 46176, + "Ġmaize": 92207, + "Ġmaj": 23953, + "Ġmajestic": 80289, + "Ġmajor": 3598, + "Ġmajority": 8686, + "Ġmajors": 55233, + "ĠmajÄħ": 91823, + "Ġmak": 50916, + "Ġmaka": 82277, + "Ġmake": 1281, + "ĠmakeStyles": 35805, + "Ġmaken": 54056, + "Ġmakeover": 84052, + "Ġmaker": 24232, + "Ġmakers": 28314, + "Ġmakes": 3643, + "Ġmakeshift": 86080, + "Ġmakeup": 26551, + "Ġmaking": 3259, + "Ġmaks": 89144, + "Ġmal": 8641, + "Ġmalaria": 68163, + "Ġmalaysia": 79539, + "Ġmale": 8593, + "Ġmales": 24037, + "Ġmalformed": 79250, + "Ġmalfunction": 71187, + "Ġmalicious": 38170, + "Ġmalign": 59227, + "Ġmalignant": 93229, + "Ġmall": 33253, + "Ġmalloc": 15731, + "Ġmalls": 83368, + "Ġmalt": 54656, + "Ġmalware": 39731, + "Ġmam": 32720, + "Ġmama": 56164, + "Ġmaman": 93431, + "Ġmamm": 34941, + "Ġmamma": 89311, + "Ġmammals": 55569, + "Ġman": 883, + "Ġmana": 30131, + "Ġmanage": 10091, + "Ġmanageable": 70028, + "Ġmanaged": 8975, + "Ġmanagement": 6240, + "Ġmanager": 6645, + "Ġmanagerial": 90150, + "Ġmanagers": 19680, + "Ġmanages": 28872, + "Ġmanaging": 18150, + "Ġmand": 11576, + "Ġmandate": 34281, + "Ġmandated": 64409, + "Ġmandates": 74953, + "Ġmandatory": 23042, + "Ġmane": 52360, + "Ġmanera": 43892, + "Ġmaneu": 37252, + "Ġmaneuver": 44114, + "Ġmaneuvers": 95791, + "Ġmang": 50196, + "Ġmanga": 34250, + "Ġmanganese": 86834, + "Ġmange": 59434, + "Ġmango": 69268, + "Ġmanh": 139272, + "Ġmanhã": 139273, + "Ġmanic": 69122, + "Ġmanifest": 14455, + "Ġmanifestation": 62950, + "Ġmanifestations": 77067, + "Ġmanifested": 78654, + "Ġmanifesto": 74603, + "Ġmanifests": 83232, + "Ġmanifold": 72829, + "Ġmanip": 14609, + "Ġmanipulate": 36635, + "Ġmanipulated": 54215, + "Ġmanipulating": 62514, + "Ġmanipulation": 33686, + "Ġmanière": 84622, + "Ġmankind": 42284, + "Ġmann": 37808, + "Ġmanned": 88325, + "Ġmannen": 76942, + "Ġmanner": 11566, + "Ġmanners": 69470, + "Ġmano": 55945, + "Ġmanoe": 98681, + "Ġmanos": 96249, + "Ġmanpower": 95769, + "Ġmans": 49234, + "Ġmansion": 51428, + "Ġmanslaughter": 88239, + "Ġmant": 25889, + "Ġmanten": 89890, + "Ġmantener": 91870, + "Ġmantle": 70432, + "Ġmantra": 66909, + "Ġmanual": 11376, + "Ġmanually": 20083, + "Ġmanuals": 55194, + "Ġmanufact": 5844, + "Ġmanufacture": 29747, + "Ġmanufactured": 27548, + "Ġmanufacturer": 13963, + "Ġmanufacturers": 16621, + "Ġmanufactures": 69801, + "Ġmanufacturing": 14919, + "Ġmanus": 34337, + "Ġmanuscript": 46813, + "Ġmanuscripts": 78588, + "Ġmany": 1657, + "Ġmap": 2415, + "ĠmapDispatchToProps": 31641, + "ĠmapStateToProps": 30190, + "ĠmapView": 62434, + "Ġmapa": 68110, + "Ġmaple": 54380, + "Ġmapped": 23844, + "ĠmappedBy": 95074, + "Ġmapper": 24501, + "Ġmapping": 12731, + "Ġmappings": 44817, + "Ġmaps": 14043, + "Ġmar": 3594, + "Ġmarathon": 44696, + "Ġmarble": 41290, + "Ġmarc": 59798, + "Ġmarca": 45997, + "Ġmarch": 15217, + "Ġmarched": 58661, + "Ġmarches": 87777, + "Ġmarching": 61098, + "Ġmarché": 83651, + "Ġmare": 54218, + "Ġmarg": 18687, + "Ġmargin": 4745, + "ĠmarginBottom": 32036, + "ĠmarginLeft": 32255, + "ĠmarginRight": 41464, + "ĠmarginTop": 22828, + "Ġmarginal": 31773, + "Ġmarginalized": 75720, + "Ġmargins": 36582, + "Ġmari": 12696, + "Ġmariage": 88260, + "Ġmarijuana": 14593, + "Ġmarin": 89115, + "Ġmarine": 28591, + "Ġmarital": 59339, + "Ġmaritime": 57312, + "Ġmark": 1868, + "Ġmarkdown": 50494, + "Ġmarked": 12864, + "Ġmarkedly": 87001, + "Ġmarker": 11134, + "Ġmarkers": 23962, + "Ġmarket": 3081, + "Ġmarketed": 60049, + "Ġmarketer": 97263, + "Ġmarketers": 54472, + "Ġmarketing": 8493, + "Ġmarketplace": 29533, + "Ġmarkets": 11725, + "Ġmarking": 34924, + "Ġmarkings": 64072, + "Ġmarks": 15423, + "Ġmarkup": 40751, + "Ġmarque": 74367, + "Ġmarriage": 10866, + "Ġmarriages": 46351, + "Ġmarried": 12224, + "Ġmarrow": 82648, + "Ġmarry": 31818, + "Ġmarrying": 77749, + "Ġmars": 47349, + "Ġmarsh": 31996, + "Ġmarshal": 60771, + "Ġmarshaller": 98909, + "Ġmart": 48195, + "Ġmartial": 36859, + "Ġmartin": 94916, + "Ġmarty": 59349, + "Ġmartyr": 89308, + "Ġmarvel": 44782, + "Ġmarvelous": 80192, + "Ġmary": 84630, + "Ġmarzo": 85528, + "Ġmarço": 138323, + "Ġmas": 9243, + "Ġmasa": 70689, + "Ġmasc": 42936, + "Ġmascara": 98378, + "Ġmascot": 84665, + "Ġmascul": 36098, + "Ġmasculine": 60065, + "Ġmasculinity": 79711, + "Ġmash": 62458, + "Ġmashed": 95134, + "Ġmasih": 83471, + "Ġmask": 6911, + "Ġmasked": 42148, + "Ġmasking": 78076, + "Ġmasks": 26135, + "Ġmass": 3072, + "Ġmassa": 70842, + "Ġmassac": 90058, + "Ġmassacre": 48252, + "Ġmassage": 6245, + "Ġmassages": 63179, + "Ġmassaggi": 45709, + "Ġmassasje": 30170, + "Ġmasse": 57517, + "Ġmasses": 31638, + "Ġmassive": 10951, + "Ġmassively": 63707, + "Ġmast": 18700, + "Ġmaster": 7341, + "Ġmastered": 70225, + "Ġmastering": 87761, + "Ġmasterpiece": 58731, + "Ġmasters": 35367, + "Ġmastery": 65663, + "Ġmastur": 78102, + "Ġmasturb": 29343, + "Ġmasturbating": 92007, + "Ġmasturbation": 56662, + "Ġmat": 5517, + "Ġmata": 96193, + "Ġmatch": 2432, + "Ġmatched": 18054, + "Ġmatcher": 36052, + "Ġmatches": 9071, + "Ġmatching": 12579, + "Ġmatchmaking": 61987, + "Ġmatchup": 44837, + "Ġmatchups": 78358, + "Ġmate": 29176, + "Ġmater": 52151, + "Ġmateria": 44458, + "Ġmaterial": 3684, + "Ġmaterially": 89594, + "Ġmaterials": 7236, + "ĠmateriaÅĤ": 132331, + "Ġmaternal": 49050, + "Ġmaternity": 70171, + "Ġmates": 54516, + "Ġmath": 6888, + "Ġmathematic": 20976, + "Ġmathematical": 35972, + "Ġmathematics": 37596, + "Ġmaths": 70708, + "Ġmating": 72119, + "Ġmatière": 135725, + "Ġmatlab": 76001, + "Ġmatplotlib": 16801, + "Ġmatrices": 35195, + "Ġmatrimon": 73749, + "Ġmatrix": 6172, + "Ġmatriz": 52640, + "Ġmats": 61056, + "Ġmatt": 26439, + "Ġmatte": 64157, + "Ġmatter": 4925, + "Ġmattered": 80620, + "Ġmatters": 12850, + "Ġmattress": 32226, + "Ġmattresses": 90953, + "Ġmature": 14851, + "Ġmatures": 88626, + "Ġmaturity": 47161, + "Ġmaté": 85274, + "Ġmatéria": 136169, + "Ġmatériel": 140513, + "Ġmau": 58914, + "Ġmaur": 97978, + "Ġmav": 58370, + "Ġmax": 1932, + "ĠmaxHeight": 58109, + "ĠmaxLength": 35507, + "ĠmaxSize": 61935, + "ĠmaxValue": 52172, + "ĠmaxWidth": 39617, + "ĠmaxX": 67482, + "ĠmaxY": 64618, + "Ġmaxi": 76968, + "Ġmaxim": 30027, + "Ġmaximal": 53129, + "Ġmaximize": 34508, + "Ġmaximizing": 87178, + "Ġmaximum": 7192, + "Ġmaxlen": 75448, + "Ġmaxlength": 29460, + "Ġmaxx": 97211, + "Ġmay": 1231, + "Ġmaybe": 7196, + "Ġmayo": 73331, + "Ġmayor": 16923, + "Ġmayores": 96267, + "ĠmayorÃŃa": 85316, + "Ġmaze": 35096, + "Ġmaç": 133297, + "Ġmaçı": 141262, + "Ġmañana": 95516, + "ĠmaÅĤe": 139680, + "Ġmb": 10016, + "Ġmbedtls": 97289, + "Ġmbox": 91407, + "Ġmc": 19223, + "Ġmcc": 73847, + "Ġmd": 10688, + "Ġmdb": 97776, + "Ġmdi": 68944, + "Ġmdl": 51700, + "Ġme": 752, + "Ġmeal": 15145, + "Ġmeals": 20969, + "Ġmean": 3076, + "Ġmeaning": 7290, + "Ġmeaningful": 22414, + "Ġmeaningless": 55926, + "Ġmeanings": 49700, + "Ġmeans": 3363, + "Ġmeant": 8791, + "Ġmeantime": 32853, + "Ġmeanwhile": 36218, + "Ġmeas": 6893, + "Ġmeasles": 79237, + "Ġmeasurable": 65203, + "Ġmeasure": 6629, + "Ġmeasured": 16878, + "Ġmeasurement": 18662, + "Ġmeasurements": 21595, + "Ġmeasures": 10953, + "Ġmeasuring": 28990, + "Ġmeat": 13041, + "Ġmeats": 62775, + "Ġmec": 74778, + "Ġmech": 83496, + "Ġmechan": 7697, + "Ġmechanic": 44139, + "Ġmechanical": 22148, + "Ġmechanically": 97997, + "Ġmechanics": 29026, + "Ġmechanism": 16953, + "Ġmechanisms": 23783, + "Ġmed": 1774, + "Ġmedal": 36612, + "Ġmedals": 58982, + "Ġmeddling": 80213, + "Ġmedi": 24127, + "Ġmedia": 3687, + "ĠmediaPlayer": 90141, + "Ġmedial": 96248, + "Ġmedian": 22553, + "Ġmediante": 65573, + "Ġmediated": 77826, + "Ġmediation": 76573, + "Ġmediator": 67930, + "Ġmedic": 8089, + "Ġmedical": 6457, + "Ġmedically": 77169, + "Ġmedication": 23221, + "Ġmedications": 29910, + "Ġmedicinal": 56941, + "Ġmedicine": 15712, + "Ġmedicines": 38553, + "Ġmedida": 71528, + "Ġmedidas": 82020, + "Ġmedieval": 41008, + "Ġmedio": 48679, + "Ġmediocre": 67380, + "Ġmedios": 95646, + "Ġmeditation": 32762, + "Ġmedium": 11051, + "Ġmediums": 97812, + "Ġmeds": 83807, + "Ġmee": 73343, + "Ġmeer": 37967, + "Ġmeet": 3367, + "Ġmeeting": 6438, + "Ġmeetings": 16261, + "Ġmeets": 20027, + "Ġmeetup": 95157, + "Ġmeg": 18740, + "Ġmega": 39387, + "Ġmeget": 88625, + "Ġmehr": 18520, + "Ġmeiden": 83577, + "Ġmeille": 30226, + "Ġmeilleur": 48248, + "Ġmeilleure": 93621, + "Ġmeilleurs": 78613, + "Ġmein": 41235, + "Ġmeine": 44390, + "Ġmeinem": 88787, + "Ġmeinen": 81212, + "Ġmeiner": 71126, + "Ġmeio": 65899, + "Ġmeis": 53355, + "Ġmeisje": 92552, + "Ġmeisjes": 77234, + "Ġmeisten": 84701, + "Ġmej": 20326, + "Ġmejor": 25937, + "Ġmejorar": 84116, + "Ġmejores": 51095, + "Ġmel": 10581, + "Ġmelakukan": 81237, + "Ġmelan": 66912, + "Ġmelanch": 86063, + "Ġmeld": 70430, + "Ġmelee": 42815, + "Ġmelhor": 45682, + "Ġmelhores": 94363, + "Ġmell": 53348, + "Ġmelod": 96274, + "Ġmelodies": 89013, + "Ġmelody": 61584, + "Ġmelt": 28999, + "Ġmeltdown": 90040, + "Ġmelted": 49359, + "Ġmelting": 49584, + "Ġmelts": 98091, + "Ġmem": 1833, + "Ġmemb": 38234, + "Ġmember": 4462, + "ĠmemberId": 96339, + "Ġmembers": 3613, + "Ġmembership": 15871, + "Ġmemberships": 87981, + "Ġmembr": 30719, + "Ġmembrane": 38554, + "Ġmembranes": 78248, + "Ġmembres": 77857, + "Ġmembuat": 76234, + "Ġmemcmp": 98718, + "Ġmemcpy": 19797, + "Ġmeme": 41185, + "Ġmemes": 61177, + "Ġmemiliki": 71136, + "Ġmemo": 21438, + "Ġmemoir": 50242, + "Ġmemor": 16035, + "Ġmemorable": 32496, + "Ġmemorandum": 81658, + "Ġmemoria": 71327, + "Ġmemorial": 37917, + "Ġmemories": 18923, + "Ġmemory": 4938, + "Ġmemset": 17800, + "Ġmen": 2953, + "Ġmenace": 80595, + "Ġmenacing": 93972, + "Ġmenc": 51518, + "Ġmend": 44776, + "Ġmeng": 19729, + "Ġmenggunakan": 80327, + "Ġmening": 66879, + "Ġmenj": 41374, + "Ġmenjadi": 54353, + "Ġmenn": 58727, + "Ġmennes": 95118, + "Ġmeno": 88911, + "Ġmenor": 47270, + "Ġmenos": 31795, + "Ġmens": 16049, + "Ġmensagem": 58817, + "Ġmensaje": 35577, + "Ġmensajes": 97305, + "Ġmensen": 64699, + "Ġmenstr": 53530, + "Ġmenstrual": 77627, + "Ġment": 11288, + "Ġmental": 10502, + "Ġmentality": 48295, + "Ġmentally": 33225, + "Ġmente": 87062, + "Ġmention": 6286, + "Ġmentioned": 9733, + "Ġmentioning": 44291, + "Ġmentions": 33845, + "Ġmentor": 35428, + "Ġmentoring": 75479, + "Ġmentors": 75607, + "Ġmentre": 92019, + "Ġmenu": 5022, + "ĠmenuItem": 56754, + "Ġmenus": 34154, + "Ġmeny": 54425, + "Ġmer": 4704, + "Ġmerc": 16481, + "Ġmercado": 48850, + "Ġmercenaries": 98587, + "Ġmerch": 21733, + "Ġmerchandise": 34945, + "Ġmerchant": 29238, + "Ġmerchants": 41770, + "Ġmerciless": 99152, + "Ġmercury": 50752, + "Ġmercy": 38224, + "Ġmere": 17524, + "Ġmereka": 63493, + "Ġmerely": 16234, + "Ġmerg": 18463, + "Ġmerge": 10880, + "Ġmerged": 26001, + "Ġmerger": 46012, + "Ġmerges": 80953, + "Ġmerging": 53377, + "Ġmerit": 37459, + "Ġmerits": 49832, + "Ġmerry": 92452, + "Ġmerupakan": 76983, + "Ġmes": 10846, + "Ġmesa": 70444, + "Ġmeses": 48984, + "Ġmesh": 11294, + "Ġmeshes": 64514, + "Ġmesma": 94310, + "Ġmesmer": 83361, + "Ġmesmo": 40249, + "Ġmess": 9435, + "Ġmessage": 1943, + "ĠmessageId": 74693, + "ĠmessageType": 97650, + "Ġmessagebox": 92420, + "Ġmessages": 6605, + "Ġmessaging": 29522, + "Ġmessed": 64202, + "Ġmessenger": 49496, + "Ġmessing": 75751, + "Ġmessy": 45846, + "Ġmest": 60013, + "Ġmesure": 83821, + "Ġmet": 2270, + "Ġmeta": 8823, + "ĠmetaData": 81928, + "Ġmetab": 19635, + "Ġmetabol": 27069, + "Ġmetabolic": 40761, + "Ġmetabolism": 37997, + "Ġmetadata": 11160, + "Ġmetal": 9317, + "Ġmetall": 69659, + "Ġmetallic": 45158, + "Ġmetals": 36082, + "Ġmetam": 87748, + "Ġmetaph": 30668, + "Ġmetaphor": 45350, + "Ġmetast": 67270, + "Ġmetav": 77520, + "Ġmetavar": 51218, + "Ġmeteor": 41042, + "Ġmeter": 22962, + "Ġmeters": 20044, + "Ġmeth": 21068, + "Ġmethane": 59983, + "Ġmethod": 1714, + "ĠmethodName": 47745, + "Ġmethodologies": 80798, + "Ġmethodology": 37052, + "Ġmethods": 5413, + "Ġmethyl": 78474, + "Ġmetic": 58664, + "Ġmeticulous": 95178, + "Ġmeticulously": 90326, + "Ġmetod": 94384, + "Ġmetodo": 58179, + "Ġmetre": 81573, + "Ġmetres": 36256, + "Ġmetric": 18266, + "Ġmetrics": 16734, + "Ġmetro": 33482, + "Ġmetropolitan": 57406, + "Ġmetros": 75478, + "Ġmettre": 46921, + "Ġmeu": 55209, + "Ġmex": 36949, + "Ġmexico": 76376, + "Ġmez": 76785, + "Ġmf": 43969, + "Ġmg": 13742, + "Ġmga": 76068, + "Ġmgr": 57897, + "Ġmh": 72317, + "Ġmi": 9497, + "Ġmia": 56253, + "ĠmiaÅĤ": 130139, + "ĠmiaÅĤa": 135775, + "Ġmic": 19197, + "Ġmice": 23627, + "Ġmich": 23273, + "Ġmichael": 87906, + "Ġmicro": 8003, + "Ġmicrobes": 79627, + "Ġmicrobi": 52399, + "Ġmicrobial": 74318, + "Ġmicron": 94209, + "Ġmicrophone": 42395, + "Ġmicroscope": 72657, + "Ġmicroscopic": 88990, + "Ġmicroscopy": 91814, + "Ġmicroseconds": 84762, + "Ġmicrosoft": 79896, + "Ġmicrotime": 98471, + "Ġmicrowave": 41274, + "Ġmid": 5099, + "Ġmidd": 64603, + "Ġmiddle": 6149, + "Ġmiddleware": 29679, + "Ġmidfield": 29689, + "Ġmidfielder": 43967, + "Ġmidi": 54040, + "Ġmidnight": 32333, + "Ġmidpoint": 81963, + "Ġmidst": 34346, + "Ġmidterm": 77490, + "Ġmidway": 82151, + "Ġmie": 29065, + "Ġmiejs": 76870, + "Ġmiejsc": 91894, + "ĠmiejscowoÅĽci": 143489, + "Ġmientras": 58594, + "Ġmies": 78581, + "ĠmiesiÄħ": 133703, + "ĠmiesiÄħc": 133704, + "ĠmiesiÄĻ": 141347, + "ĠmiesiÄĻcy": 141348, + "Ġmiesz": 93073, + "Ġmieszka": 128196, + "ĠmieszkaÅĦ": 128197, + "ĠmieszkaÅĦców": 142564, + "Ġmieux": 63872, + "ĠmieÄĩ": 133131, + "Ġmig": 28344, + "Ġmight": 2578, + "Ġmighty": 41627, + "Ġmigli": 44479, + "Ġmiglior": 91744, + "Ġmigliori": 74741, + "Ġmigr": 9772, + "Ġmigraine": 91881, + "Ġmigrant": 51116, + "Ġmigrants": 30987, + "Ġmigrate": 44566, + "Ġmigrated": 72591, + "Ġmigrating": 84526, + "Ġmigration": 11906, + "ĠmigrationBuilder": 16955, + "Ġmigrations": 17063, + "Ġmij": 68397, + "Ġmijn": 52992, + "Ġmik": 78805, + "Ġmike": 97669, + "Ġmil": 7473, + "Ġmilan": 93113, + "Ġmilano": 41774, + "Ġmild": 23034, + "Ġmildly": 75849, + "Ġmile": 14587, + "Ġmileage": 49852, + "Ġmiles": 8756, + "Ġmilestone": 42191, + "Ġmilestones": 68276, + "Ġmilf": 25399, + "Ġmilfs": 91974, + "Ġmilhões": 133893, + "Ġmilieu": 82159, + "Ġmilit": 5327, + "Ġmilitant": 51282, + "Ġmilitants": 38349, + "Ġmilitar": 61433, + "Ġmilitary": 6277, + "Ġmilitia": 57598, + "Ġmilitias": 79492, + "Ġmilk": 14074, + "Ġmill": 2544, + "Ġmillenn": 40599, + "Ġmillennia": 95631, + "Ġmillennials": 69333, + "Ġmillennium": 88685, + "Ġmilling": 45372, + "Ġmillion": 3526, + "Ġmillionaire": 88944, + "Ġmillions": 11728, + "Ġmillis": 57092, + "Ġmilliseconds": 30504, + "Ġmillones": 63024, + "Ġmills": 31908, + "Ġmim": 26905, + "Ġmime": 45270, + "ĠmimeType": 71826, + "Ġmimetype": 80008, + "Ġmimic": 55359, + "Ġmin": 1308, + "ĠminHeight": 64385, + "ĠminLength": 83785, + "ĠminOccurs": 79241, + "ĠminValue": 63292, + "ĠminWidth": 56098, + "ĠminX": 76174, + "ĠminY": 70347, + "Ġminced": 93827, + "Ġmind": 3971, + "Ġminded": 60167, + "Ġminden": 90544, + "Ġmindful": 62342, + "Ġmindfulness": 70605, + "Ġminds": 20062, + "Ġmindset": 39443, + "Ġmine": 10485, + "Ġminecraft": 73873, + "Ġmined": 67668, + "Ġminer": 47163, + "Ġmineral": 24136, + "Ġminerals": 32972, + "Ġminers": 42143, + "Ġmines": 33657, + "Ġming": 72159, + "Ġmingle": 90025, + "Ġminha": 62114, + "Ġmini": 13420, + "Ġminiature": 56763, + "Ġminib": 75917, + "Ġminim": 21187, + "Ġminimal": 17377, + "Ġminimalist": 71670, + "Ġminimise": 95987, + "Ġminimize": 29337, + "Ġminimized": 81078, + "Ġminimizing": 76291, + "Ġminimum": 8028, + "Ġmining": 11673, + "Ġminion": 62535, + "Ġminions": 59805, + "Ġminist": 48804, + "Ġminister": 12725, + "Ġministers": 33950, + "Ġministries": 89024, + "Ġministry": 24846, + "Ġminlength": 77929, + "Ġminor": 8922, + "Ġminorities": 39704, + "Ġminority": 22591, + "Ġminors": 48540, + "Ġmins": 24835, + "Ġmint": 28337, + "Ġminus": 27283, + "Ġminut": 81370, + "Ġminute": 9383, + "Ġminutes": 4420, + "Ġminutos": 51662, + "Ġmio": 69012, + "Ġmip": 72751, + "Ġmir": 8652, + "Ġmirac": 64966, + "Ġmiracle": 44025, + "Ġmiracles": 70812, + "Ġmiraculous": 90713, + "Ġmirror": 17846, + "Ġmirrored": 69037, + "Ġmirrors": 40485, + "Ġmis": 5786, + "Ġmisc": 31125, + "Ġmiscar": 79562, + "Ġmiscellaneous": 92404, + "Ġmischief": 93946, + "Ġmiscon": 29671, + "Ġmisconception": 97331, + "Ġmisconduct": 39994, + "Ġmisd": 50068, + "Ġmisdemean": 50626, + "Ġmisdemeanor": 57992, + "Ġmise": 56359, + "Ġmiser": 97210, + "Ġmiserable": 49639, + "Ġmisery": 57601, + "Ġmisguided": 75506, + "Ġmish": 63396, + "Ġmisinformation": 74059, + "Ġmisleading": 37209, + "Ġmisled": 98425, + "Ġmism": 66414, + "Ġmisma": 60709, + "Ġmismatch": 35301, + "Ġmismo": 35894, + "Ġmisog": 62885, + "Ġmisogyn": 94684, + "Ġmisplaced": 89560, + "Ġmisrepresented": 82421, + "Ġmiss": 3116, + "Ġmissed": 13628, + "Ġmisses": 42294, + "Ġmissile": 25224, + "Ġmissiles": 33341, + "Ġmissing": 7402, + "Ġmission": 8954, + "Ġmissionaries": 85216, + "Ġmissionary": 72898, + "Ġmissions": 24640, + "Ġmist": 8600, + "Ġmistake": 16523, + "Ġmistaken": 36004, + "Ġmistakenly": 76534, + "Ġmistakes": 20643, + "Ġmistr": 91911, + "Ġmistress": 65094, + "Ġmisunder": 37626, + "Ġmisunderstand": 89600, + "Ġmisunderstanding": 69747, + "Ġmisunderstood": 72856, + "Ġmisuse": 61751, + "Ġmit": 5451, + "Ġmitig": 34990, + "Ġmitigate": 49360, + "Ġmitigation": 65760, + "Ġmitochond": 53942, + "Ġmitochondrial": 70961, + "Ġmitt": 47332, + "Ġmix": 6514, + "Ġmixed": 9519, + "Ġmixer": 37778, + "Ġmixes": 62426, + "Ġmixin": 89603, + "Ġmixing": 26792, + "Ġmixins": 93980, + "Ġmixture": 20980, + "ĠmiÄĻd": 127252, + "ĠmiÄĻdz": 143665, + "ĠmiÄĻdzy": 132152, + "ĠmiÄĻdzyn": 143666, + "ĠmiÄĻdzynar": 143667, + "ĠmiÄĻdzynarod": 143668, + "ĠmiÄĻdzynarodow": 143669, + "ĠmiÅĤo": 143970, + "ĠmiÅĤoÅĽci": 143971, + "Ġmiá»ģn": 133318, + "Ġmiá»ħn": 132382, + "Ġmiá»ĩ": 134213, + "Ġmiá»ĩng": 134214, + "Ġmj": 83408, + "Ġmk": 23789, + "Ġmkdir": 30073, + "Ġmktime": 93049, + "Ġml": 15739, + "Ġmlx": 75411, + "Ġmm": 9465, + "Ġmmap": 70866, + "Ġmmc": 84208, + "Ġmn": 27938, + "Ġmnemonic": 86801, + "Ġmnie": 73073, + "Ġmnist": 91264, + "Ġmo": 4544, + "Ġmob": 12595, + "Ġmobil": 28805, + "Ġmobile": 6371, + "Ġmobility": 30039, + "Ġmobs": 72392, + "Ġmoc": 87672, + "Ġmock": 7860, + "ĠmockMvc": 96342, + "Ġmocked": 46149, + "Ġmocker": 93124, + "Ġmockery": 98154, + "Ġmocking": 66483, + "Ġmocks": 68909, + "Ġmod": 1463, + "Ġmoda": 90668, + "Ġmodal": 13228, + "ĠmodalitÃł": 140393, + "Ġmode": 3856, + "Ġmodel": 1614, + "ĠmodelAndView": 80675, + "ĠmodelBuilder": 37504, + "ĠmodelName": 87031, + "Ġmodele": 96725, + "Ġmodeled": 61553, + "Ġmodeling": 33479, + "Ġmodelling": 60866, + "Ġmodelo": 35942, + "Ġmodelos": 82642, + "Ġmodels": 4119, + "Ġmodem": 51504, + "Ġmoden": 99147, + "Ġmoder": 13303, + "Ġmoderate": 23193, + "Ġmoderated": 86216, + "Ġmoderately": 69251, + "Ġmoderation": 63282, + "Ġmoderator": 59427, + "Ġmoderators": 82747, + "Ġmodern": 6481, + "Ġmoderne": 76480, + "Ġmodes": 19777, + "Ġmodest": 26848, + "Ġmodific": 69133, + "Ġmodificar": 56206, + "Ġmodification": 17030, + "Ġmodifications": 28782, + "Ġmodified": 10807, + "Ġmodifier": 22739, + "Ġmodifiers": 35670, + "Ġmodifies": 82949, + "Ġmodify": 5602, + "Ġmodifying": 46041, + "Ġmodne": 80271, + "Ġmodo": 33337, + "Ġmods": 28254, + "Ġmodular": 43893, + "Ġmodulation": 66447, + "Ġmodule": 4688, + "ĠmoduleId": 92609, + "ĠmoduleName": 73788, + "Ġmodules": 13454, + "Ġmodulo": 59864, + "Ġmodulus": 74024, + "Ġmodèle": 82497, + "Ġmoet": 54794, + "Ġmoeten": 92565, + "Ġmog": 26170, + "Ġmogelijk": 74411, + "Ġmogul": 96908, + "ĠmogÄħ": 129928, + "ĠmogÄĻ": 139359, + "ĠmogÅĤa": 142995, + "Ġmoi": 39991, + "Ġmoins": 39870, + "Ġmois": 45315, + "Ġmoist": 20507, + "Ġmoistur": 53802, + "Ġmoisture": 31157, + "Ġmojo": 75416, + "Ġmol": 21609, + "Ġmold": 28461, + "Ġmolded": 84660, + "Ġmolding": 65057, + "Ġmolds": 97734, + "Ġmole": 34651, + "Ġmolec": 16671, + "Ġmolecular": 30106, + "Ġmolecule": 41930, + "Ġmolecules": 34615, + "Ġmolest": 69000, + "Ġmoll": 54409, + "Ġmolt": 90339, + "Ġmolto": 47548, + "Ġmom": 3368, + "Ġmoment": 4445, + "Ġmomentarily": 98397, + "Ġmomento": 30121, + "Ġmomentos": 96153, + "Ġmoments": 13943, + "Ġmomentum": 23270, + "Ġmommy": 81140, + "Ġmoms": 46734, + "Ġmon": 1615, + "Ġmonarch": 62754, + "Ġmonarchy": 86049, + "Ġmonastery": 84761, + "Ġmond": 61840, + "Ġmonday": 93199, + "Ġmonde": 37800, + "Ġmondo": 69709, + "Ġmonet": 57091, + "Ġmonetary": 32284, + "Ġmoney": 3220, + "Ġmong": 74542, + "Ġmongo": 33814, + "Ġmongodb": 77464, + "Ġmongoose": 15607, + "Ġmonitor": 8718, + "Ġmonitored": 40123, + "Ġmonitoring": 16558, + "Ġmonitors": 36477, + "Ġmonk": 62672, + "Ġmonkey": 38703, + "Ġmonkeys": 60720, + "Ġmonks": 65377, + "Ġmono": 39674, + "Ġmonoc": 95057, + "Ġmonopol": 74153, + "Ġmonopoly": 53201, + "Ġmonot": 77216, + "Ġmonster": 17651, + "Ġmonsters": 25787, + "Ġmonstr": 65504, + "Ġmonstrous": 75581, + "Ġmont": 20007, + "Ġmontage": 95944, + "Ġmonth": 2254, + "Ġmonthly": 15088, + "Ġmonths": 3951, + "Ġmonument": 36897, + "Ġmonumental": 77756, + "Ġmonuments": 61903, + "Ġmoo": 45871, + "Ġmood": 19671, + "Ġmooie": 75653, + "Ġmoon": 17788, + "Ġmoons": 80160, + "Ġmoot": 97783, + "Ġmop": 88064, + "Ġmor": 4313, + "Ġmoral": 15659, + "Ġmorale": 62583, + "Ġmorality": 44551, + "Ġmorally": 56223, + "Ġmorals": 80336, + "Ġmorb": 92044, + "Ġmore": 803, + "Ġmoreover": 43543, + "Ġmorning": 6556, + "Ġmornings": 57202, + "Ġmorph": 26351, + "Ġmorphology": 78512, + "Ġmort": 10032, + "Ġmortal": 48872, + "Ġmortality": 28428, + "Ġmortar": 57460, + "Ġmortgage": 20846, + "Ġmortgages": 60898, + "Ġmos": 22686, + "Ġmosaic": 70524, + "Ġmosque": 50012, + "Ġmosques": 86559, + "Ġmosquito": 49546, + "Ġmosquitoes": 82101, + "Ġmoss": 77243, + "Ġmost": 1429, + "Ġmostly": 10008, + "Ġmostr": 55490, + "Ġmostra": 92097, + "Ġmostrar": 43008, + "Ġmot": 3852, + "Ġmote": 87386, + "Ġmotel": 80381, + "Ġmoth": 97078, + "Ġmother": 6554, + "Ġmotherboard": 64243, + "Ġmothers": 26600, + "Ġmotif": 59512, + "Ġmotifs": 83889, + "Ġmotion": 11379, + "Ġmotions": 53145, + "Ġmotiv": 12242, + "Ġmotivate": 61325, + "Ġmotivated": 26664, + "Ġmotivating": 88589, + "Ġmotivation": 24798, + "Ġmotivational": 87359, + "Ġmotivations": 64831, + "Ġmotive": 45994, + "Ġmotives": 51040, + "Ġmotivo": 78464, + "Ġmoto": 72956, + "Ġmotor": 8872, + "Ġmotorcycle": 34304, + "Ġmotorcycles": 70653, + "Ġmotorists": 71898, + "Ġmotors": 37324, + "Ġmots": 77099, + "Ġmotto": 62326, + "Ġmou": 53309, + "Ġmould": 50270, + "Ġmound": 72166, + "Ġmount": 6470, + "Ġmountain": 16301, + "Ġmountains": 23501, + "Ġmounted": 21810, + "Ġmounting": 33639, + "Ġmounts": 56222, + "Ġmour": 66583, + "Ġmourn": 59134, + "Ġmourning": 75250, + "Ġmouse": 8644, + "ĠmouseClicked": 85739, + "ĠmouseX": 80890, + "ĠmouseY": 68047, + "Ġmouth": 10780, + "Ġmouths": 64509, + "Ġmov": 1974, + "Ġmovable": 82579, + "Ġmove": 3271, + "ĠmoveTo": 94781, + "Ġmoved": 7726, + "Ġmovement": 7203, + "Ġmovements": 19029, + "Ġmover": 67999, + "Ġmovers": 95354, + "Ġmoves": 10797, + "Ġmovie": 5700, + "Ġmovies": 9508, + "Ġmovimiento": 88026, + "Ġmoving": 7218, + "Ġmower": 98678, + "Ġmoy": 51038, + "Ġmoyen": 90016, + "Ġmoz": 95339, + "Ġmozilla": 84914, + "Ġmoż": 37947, + "Ġmoże": 53521, + "Ġmożemy": 134403, + "Ġmożesz": 135951, + "Ġmożli": 86057, + "Ġmożliwe": 141708, + "Ġmożliwo": 127600, + "ĠmożliwoÅĽci": 134275, + "ĠmożliwoÅĽÄĩ": 134359, + "Ġmożna": 62573, + "Ġmp": 10490, + "Ġmpfr": 82214, + "Ġmpg": 82454, + "Ġmph": 33961, + "Ġmpi": 58122, + "Ġmpl": 63784, + "Ġmpz": 71228, + "Ġmq": 72298, + "Ġmqtt": 59431, + "Ġmr": 17317, + "Ġmrb": 84261, + "Ġms": 9829, + "Ġmse": 82053, + "Ġmsec": 80359, + "Ġmsg": 3750, + "Ġmsgs": 41270, + "Ġmsm": 58588, + "Ġmt": 11965, + "Ġmtx": 81522, + "Ġmu": 11834, + "Ġmuc": 64004, + "Ġmuch": 1753, + "Ġmuchas": 63459, + "Ġmucho": 41816, + "Ġmuchos": 57206, + "Ġmud": 26183, + "Ġmudança": 136349, + "Ġmuddy": 79473, + "Ġmue": 40439, + "Ġmuestra": 70587, + "Ġmuff": 54304, + "Ġmug": 51489, + "Ġmuit": 53760, + "Ġmuito": 33750, + "Ġmuj": 26839, + "Ġmujer": 52093, + "Ġmujeres": 38808, + "Ġmul": 15602, + "Ġmulher": 38950, + "Ġmulheres": 57666, + "Ġmult": 2745, + "Ġmulter": 80245, + "Ġmulti": 7299, + "Ġmultic": 91420, + "Ġmulticast": 79534, + "Ġmulticultural": 74316, + "Ġmultid": 63386, + "Ġmultif": 61287, + "Ġmultiline": 85218, + "Ġmultim": 79049, + "Ġmultimedia": 56513, + "Ġmultin": 60130, + "Ġmultinational": 67926, + "Ġmultip": 12557, + "Ġmultipart": 68058, + "Ġmultiplayer": 38728, + "Ġmultiple": 5248, + "Ġmultiples": 65060, + "Ġmultiplic": 91707, + "Ġmultiplication": 46444, + "Ġmultiplied": 54916, + "Ġmultiplier": 30559, + "Ġmultiply": 30270, + "Ġmultiplying": 84192, + "Ġmultiprocessing": 57124, + "Ġmultis": 74737, + "Ġmultit": 84260, + "Ġmultitude": 47968, + "Ġmum": 38859, + "Ġmun": 49979, + "Ġmunch": 98245, + "Ġmund": 40566, + "Ġmundane": 68682, + "Ġmundial": 97087, + "Ġmundo": 28352, + "Ġmunic": 136332, + "Ġmunicip": 19615, + "Ġmunicipal": 26414, + "Ġmunicipalities": 58489, + "Ġmunicipality": 56818, + "ĠmunicÃŃpio": 136333, + "Ġmur": 8146, + "Ġmural": 73373, + "Ġmurder": 9901, + "Ġmurdered": 31385, + "Ġmurderer": 67966, + "Ġmurderers": 98165, + "Ġmurdering": 79927, + "Ġmurderous": 84129, + "Ġmurders": 41986, + "Ġmurky": 96792, + "Ġmurm": 86551, + "Ġmus": 3091, + "Ġmuschi": 72020, + "Ġmuscle": 15747, + "Ġmuscles": 23648, + "Ġmuscular": 51092, + "Ġmuse": 38848, + "Ġmuseum": 23971, + "Ġmuseums": 50577, + "Ġmush": 29723, + "Ġmushroom": 57366, + "Ġmushrooms": 47151, + "ĠmusiaÅĤ": 140610, + "Ġmusic": 4627, + "Ġmusica": 76587, + "Ġmusical": 17795, + "Ġmusician": 38744, + "Ġmusicians": 31529, + "Ġmusique": 88569, + "Ġmuslim": 60956, + "Ġmuss": 35141, + "Ġmust": 1969, + "Ġmustard": 64307, + "Ġmuster": 83650, + "Ġmut": 5206, + "Ġmutable": 25804, + "ĠmutableListOf": 76253, + "Ġmutant": 60518, + "Ġmutants": 87654, + "Ġmutate": 67182, + "Ġmutated": 84822, + "Ġmutating": 96518, + "Ġmutation": 26374, + "Ġmutations": 33584, + "Ġmute": 53221, + "Ġmuted": 59021, + "Ġmutex": 30863, + "Ġmutil": 75995, + "Ġmuttered": 86594, + "Ġmutual": 26750, + "Ġmutually": 52479, + "Ġmux": 59807, + "Ġmuy": 22506, + "Ġmuzzle": 74229, + "Ġmuá»ijn": 128491, + "Ġmv": 23164, + "Ġmw": 52810, + "Ġmx": 14619, + "ĠmxArray": 65838, + "Ġmy": 847, + "ĠmyList": 91727, + "Ġmycket": 79761, + "Ġmyfile": 77886, + "Ġmyocard": 94636, + "Ġmyriad": 51809, + "Ġmys": 9717, + "Ġmyself": 7037, + "Ġmysql": 10564, + "Ġmysqli": 9938, + "Ġmyst": 11425, + "Ġmyster": 23329, + "Ġmysteries": 56600, + "Ġmysterious": 25382, + "Ġmystery": 22531, + "Ġmystical": 68432, + "Ġmyth": 21163, + "Ġmythical": 83119, + "Ġmythology": 58392, + "Ġmyths": 50186, + "ĠmyÅĽli": 140603, + "Ġmz": 95392, + "Ġmá": 28730, + "Ġmái": 132273, + "Ġmáquina": 135702, + "Ġmár": 79979, + "Ġmás": 10918, + "Ġmát": 135614, + "Ġmáu": 131641, + "Ġmáxima": 141883, + "Ġmáximo": 75617, + "Ġmáy": 128392, + "Ġmâ": 142162, + "Ġmã": 58865, + "Ġmãe": 93147, + "Ġmãi": 134752, + "Ġmão": 135069, + "Ġmädchen": 58295, + "Ġmänn": 88765, + "Ġmänner": 64669, + "ĠmÃ¥": 30233, + "Ġmænd": 70560, + "Ġmé": 31018, + "Ġméd": 33930, + "Ġmédec": 139593, + "Ġmédecin": 139594, + "Ġmédica": 133272, + "Ġmédico": 132744, + "Ġmég": 82691, + "Ġmét": 35152, + "Ġméth": 74501, + "Ġméthode": 97552, + "Ġmétier": 141715, + "Ġmétodo": 57509, + "Ġmê": 130435, + "Ġmême": 26486, + "Ġmình": 128312, + "Ġmó": 65742, + "Ġmóc": 140703, + "Ġmóg": 138968, + "ĠmógÅĤ": 138969, + "Ġmón": 129994, + "Ġmóvil": 139672, + "Ġmówi": 129912, + "Ġmô": 130179, + "Ġmôi": 129991, + "Ġmôn": 130929, + "Ġmö": 22163, + "Ġmöchte": 66316, + "Ġmöchten": 85970, + "Ġmöglich": 50837, + "Ġmø": 55845, + "Ġmù": 126918, + "Ġmùa": 130567, + "Ġmùi": 138106, + "Ġmú": 91724, + "Ġmús": 59857, + "Ġmúsica": 70345, + "Ġmü": 58285, + "Ġmücade": 133118, + "Ġmücadele": 133119, + "Ġmüd": 143554, + "Ġmüdah": 143555, + "Ġmüdahale": 143556, + "Ġmük": 143547, + "Ġmükem": 143548, + "Ġmükemmel": 143549, + "Ġmümk": 132815, + "Ġmümkün": 132816, + "Ġmünchen": 78534, + "Ġmüssen": 47827, + "Ġmüzik": 141009, + "ĠmÃ¼ÅŁ": 133573, + "ĠmÃ¼ÅŁter": 133574, + "ĠmÃ¼ÅŁteri": 133575, + "ĠmÃł": 126199, + "ĠmÃłn": 131718, + "ĠmÃłu": 128865, + "ĠmÃŃ": 76326, + "ĠmÃŃn": 70224, + "ĠmÃŃnimo": 97299, + "Ġmı": 130446, + "ĠmÄĻ": 143930, + "ĠmÄĻsk": 143931, + "ĠmÄĻż": 138603, + "ĠmÄĻżczy": 138604, + "ĠmÄĻżczyzn": 138605, + "ĠmÅ©": 134968, + "ĠmÅ©i": 134969, + "ĠmÅĤ": 127433, + "ĠmÅĤod": 127434, + "ĠmÅĤodzie": 141319, + "ĠmÅĤodzież": 141320, + "ĠmÆ¡": 134796, + "Ġmưa": 136733, + "Ġmưá»Ŀi": 142837, + "Ġmạch": 135498, + "Ġmại": 131537, + "Ġmạng": 129461, + "Ġmạnh": 128818, + "Ġmất": 129276, + "Ġmấy": 132222, + "Ġmẫu": 129643, + "Ġmắc": 131880, + "Ġmắn": 138643, + "Ġmắt": 128668, + "Ġmặc": 131945, + "Ġmặt": 128387, + "Ġmẹ": 128850, + "Ġmẽ": 137125, + "ĠmáºŃt": 130821, + "Ġmỡ": 140268, + "Ġmụ": 136350, + "Ġmục": 129203, + "Ġmụn": 136351, + "Ġmức": 129240, + "Ġmừng": 137098, + "Ġmỹ": 131594, + "Ġmá»ģm": 131328, + "Ġmá»ĩnh": 140432, + "Ġmá»įi": 128728, + "Ġmá»ıi": 142306, + "Ġmá»iji": 134666, + "Ġmá»Ĺi": 129058, + "Ġmá»Ļ": 135682, + "Ġmá»Ļt": 128249, + "ĠmỼi": 92105, + "Ġmá»Ŀi": 137558, + "Ġmợ": 129367, + "Ġn": 308, + "ĠnIndex": 91024, + "Ġna": 4317, + "Ġnaam": 99105, + "Ġnaar": 29848, + "Ġnab": 72691, + "Ġnach": 14047, + "Ġnacional": 57967, + "Ġnack": 34343, + "Ġnackt": 56668, + "Ġnackte": 62002, + "Ġnad": 33534, + "Ġnada": 38211, + "Ġnadzie": 142241, + "ĠnadziejÄĻ": 142242, + "Ġnag": 36528, + "Ġnah": 68917, + "Ġnail": 30995, + "Ġnailed": 74874, + "Ġnails": 39843, + "Ġnaive": 49665, + "Ġnaj": 29174, + "Ġnajb": 143244, + "Ġnajbli": 143245, + "Ġnajbliż": 143246, + "Ġnajbliższ": 143247, + "ĠnajczÄĻ": 143594, + "ĠnajczÄĻÅĽcie": 143595, + "ĠnajczÄĻÅĽciej": 143596, + "Ġnajle": 93007, + "ĠnajwiÄĻks": 142848, + "ĠnajwiÄĻkszy": 142849, + "Ġnajwyż": 140760, + "Ġnajwyższ": 140761, + "Ġnak": 40886, + "Ġnaked": 19020, + "Ġnaken": 43845, + "Ġnakne": 67853, + "Ġnal": 84533, + "Ġnale": 87390, + "Ġnależy": 130220, + "Ġnam": 16449, + "Ġnama": 45067, + "Ġname": 829, + "ĠnameLabel": 96108, + "Ġnamed": 6941, + "Ġnamedtuple": 80536, + "Ġnamely": 31025, + "Ġnameof": 38894, + "Ġnames": 5036, + "Ġnamespace": 4473, + "Ġnamespaces": 58091, + "Ġnaming": 34948, + "Ġnamoro": 73489, + "Ġnan": 20021, + "Ġnanny": 98257, + "Ġnano": 50493, + "Ġnanop": 75207, + "Ġnanoparticles": 93653, + "Ġnao": 64273, + "Ġnap": 25859, + "Ġnapisa": 141276, + "ĠnapisaÅĤ": 141277, + "Ġnapraw": 133731, + "ĠnaprawdÄĻ": 133732, + "Ġnar": 43569, + "Ġnarc": 38688, + "Ġnarciss": 57198, + "Ġnarcotics": 91052, + "Ġnargin": 58985, + "Ġnargs": 51616, + "Ġnarr": 13408, + "Ġnarration": 85080, + "Ġnarrative": 19221, + "Ġnarratives": 57278, + "Ġnarrator": 64171, + "Ġnarrow": 15026, + "Ġnarrowed": 72935, + "Ġnarrower": 90429, + "Ġnarrowing": 97767, + "Ġnarrowly": 63947, + "Ġnarz": 140251, + "ĠnarzÄĻdzi": 140252, + "Ġnas": 17141, + "Ġnasal": 71088, + "Ġnast": 52715, + "Ġnasty": 34407, + "ĠnastÄĻ": 143287, + "ĠnastÄĻp": 141208, + "ĠnastÄĻpn": 143288, + "ĠnastÄĻpnie": 141209, + "Ġnasıl": 129384, + "Ġnat": 17588, + "Ġnation": 6995, + "Ġnational": 5313, + "Ġnationalism": 59433, + "Ġnationalist": 51894, + "Ġnationalists": 87380, + "Ġnationality": 58243, + "Ġnationally": 39243, + "Ġnationals": 55911, + "Ġnations": 16675, + "Ġnationwide": 27954, + "Ġnative": 9867, + "Ġnatives": 67373, + "Ġnatur": 14478, + "Ġnatural": 5810, + "Ġnaturally": 17712, + "Ġnature": 6993, + "Ġnatuur": 88610, + "Ġnatürlich": 67658, + "Ġnau": 92462, + "Ġnaughty": 52943, + "Ġnause": 48294, + "Ġnausea": 60780, + "Ġnav": 8075, + "ĠnavCtrl": 95168, + "Ġnaval": 45298, + "Ġnavbar": 15548, + "Ġnave": 43491, + "Ġnaveg": 55443, + "Ġnavegador": 95410, + "Ġnavig": 33412, + "Ġnavigate": 20876, + "Ġnavigating": 59399, + "Ġnavigation": 10646, + "ĠnavigationController": 65462, + "ĠnavigationOptions": 99121, + "Ġnavigator": 35409, + "Ġnavy": 44774, + "Ġnaw": 75916, + "Ġnawet": 90400, + "Ġnaz": 45802, + "Ġnaï": 94880, + "Ġnb": 16649, + "Ġnbr": 68090, + "Ġnbytes": 74647, + "Ġnc": 25126, + "Ġncols": 80235, + "Ġnd": 15581, + "Ġndarray": 66883, + "Ġne": 834, + "Ġnear": 3143, + "Ġnearby": 14046, + "Ġnearer": 86820, + "Ġnearest": 23480, + "Ġnearing": 82299, + "Ġnearly": 7009, + "Ġneat": 28485, + "Ġneatly": 62166, + "Ġneb": 80867, + "Ġneben": 84732, + "Ġnec": 19053, + "Ġneces": 24223, + "Ġnecesario": 68274, + "Ġnecesita": 93966, + "Ġnecess": 4441, + "Ġnecessarily": 14312, + "Ġnecessary": 5871, + "Ġnecessities": 75400, + "Ġnecessity": 31861, + "ĠnecessitÃł": 142091, + "Ġnecessário": 95544, + "Ġneck": 12975, + "Ġnecklace": 54447, + "Ġneckline": 91561, + "Ġned": 53705, + "Ġnederland": 62269, + "Ġneed": 1184, + "Ġneeded": 4362, + "Ġneeding": 32821, + "Ġneedle": 30309, + "Ġneedles": 56167, + "Ġneedless": 96830, + "Ġneeds": 3880, + "Ġneedy": 82463, + "Ġnef": 86124, + "Ġneg": 4184, + "Ġnegate": 71219, + "Ġnegative": 8225, + "Ġnegatively": 47191, + "Ġnegatives": 74085, + "Ġnegativity": 92695, + "Ġneger": 73733, + "Ġneglect": 38232, + "Ġneglected": 50433, + "Ġneglig": 35056, + "Ġnegligence": 58257, + "Ġnegligent": 94761, + "Ġnegligible": 81702, + "Ġnegoci": 81159, + "Ġnegocio": 95474, + "Ġnegot": 11642, + "Ġnegotiate": 36567, + "Ġnegotiated": 50478, + "Ġnegotiating": 43625, + "Ġnegotiation": 48048, + "Ġnegotiations": 23449, + "Ġnegro": 61733, + "Ġnegó": 135144, + "Ġnegócio": 135145, + "Ġneh": 68453, + "Ġnehmen": 80617, + "Ġnei": 55997, + "Ġneigh": 96166, + "Ġneighb": 6024, + "Ġneighbor": 9565, + "Ġneighborhood": 12534, + "Ġneighborhoods": 32000, + "Ġneighboring": 41517, + "Ġneighbors": 18709, + "Ġneighbour": 21922, + "Ġneighbourhood": 39342, + "Ġneighbouring": 60927, + "Ġneighbours": 35856, + "Ġneither": 13866, + "Ġnel": 24340, + "Ġnell": 65802, + "Ġnella": 35922, + "Ġnelle": 71764, + "Ġnem": 23645, + "Ġnen": 70298, + "Ġneo": 35082, + "Ġneoliberal": 80468, + "Ġneon": 46652, + "Ġnep": 91534, + "Ġneph": 43864, + "Ġnephew": 62804, + "Ġner": 40183, + "Ġnerd": 72187, + "Ġnerv": 19658, + "Ġnerve": 30915, + "Ġnerves": 49481, + "Ġnervous": 22596, + "Ġness": 85396, + "Ġnest": 22791, + "Ġnesta": 92958, + "Ġneste": 77459, + "Ġnested": 24034, + "Ġnesting": 66710, + "Ġnestled": 88677, + "Ġnests": 95121, + "Ġnet": 4179, + "Ġnets": 52457, + "Ġnett": 27770, + "Ġnettsteder": 43732, + "Ġnetwork": 3922, + "Ġnetworking": 28030, + "Ġnetworks": 14155, + "Ġneu": 20504, + "Ġneue": 38383, + "Ġneuen": 47669, + "Ġneuken": 31283, + "Ġneur": 21164, + "Ġneural": 29728, + "Ġneuro": 17773, + "Ġneurological": 63808, + "Ġneuron": 48284, + "Ġneuronal": 78302, + "Ġneurons": 33213, + "Ġneurop": 97566, + "Ġneuroscience": 91948, + "Ġneurotrans": 89251, + "Ġneut": 25793, + "Ġneutr": 41599, + "Ġneutral": 20628, + "Ġneutrality": 51764, + "Ġneutron": 72304, + "Ġnever": 2581, + "Ġnevertheless": 37230, + "Ġnew": 501, + "ĠnewArr": 91983, + "ĠnewArray": 62917, + "ĠnewData": 51523, + "ĠnewIndex": 94530, + "ĠnewInstance": 66662, + "ĠnewItem": 47988, + "ĠnewList": 67418, + "ĠnewName": 48919, + "ĠnewNode": 33560, + "ĠnewObj": 74259, + "ĠnewPassword": 75838, + "ĠnewPath": 96037, + "ĠnewPos": 82375, + "ĠnewPosition": 90899, + "ĠnewRow": 83838, + "ĠnewSize": 80843, + "ĠnewState": 36848, + "ĠnewText": 94653, + "ĠnewUser": 46246, + "ĠnewVal": 72328, + "ĠnewValue": 24174, + "ĠnewX": 88803, + "ĠnewY": 80384, + "Ġnewbie": 94578, + "Ġnewborn": 45297, + "Ġnewcom": 48476, + "Ġnewcomer": 87305, + "Ġnewcomers": 67458, + "Ġnewer": 25546, + "Ġnewest": 23601, + "Ġnewfound": 93521, + "Ġnewline": 39027, + "Ġnewly": 13631, + "Ġnews": 3669, + "Ġnewsletter": 20233, + "Ġnewsletters": 34388, + "Ġnewsp": 13377, + "Ġnewspaper": 16803, + "Ġnewspapers": 31494, + "Ġnex": 91808, + "Ġnext": 1790, + "ĠnextPage": 83595, + "ĠnextProps": 67322, + "ĠnextState": 69815, + "Ġnexus": 81401, + "Ġnf": 42128, + "Ġnfl": 85804, + "Ġnfs": 77472, + "Ġng": 7777, + "ĠngOn": 16305, + "ĠngOnDestroy": 76738, + "ĠngOnInit": 17807, + "Ġnga": 93146, + "Ġngh": 124414, + "Ġnghi": 125021, + "Ġnghiêm": 134168, + "Ġnghiên": 130693, + "Ġnghiá»ĩm": 128990, + "Ġnghiá»ĩp": 128386, + "Ġnghè": 138815, + "Ġnghèo": 138816, + "Ġnghìn": 142876, + "ĠnghÄ©": 126568, + "ĠnghÄ©a": 129598, + "Ġnghá»ģ": 132197, + "Ġnghá»ĩ": 128781, + "Ġnghá»ī": 131840, + "Ġnghá»ĭ": 130097, + "Ġnginx": 70482, + "Ġngo": 126339, + "ĠngoÃłi": 128703, + "Ġngoại": 130906, + "Ġngu": 129737, + "Ġnguyên": 129031, + "Ġnguyá»ĩn": 138030, + "Ġnguá»ĵn": 129738, + "Ġngx": 43392, + "Ġngân": 131538, + "Ġngôi": 132406, + "Ġngôn": 137425, + "ĠngÃłn": 139063, + "ĠngÃłnh": 129857, + "ĠngÃły": 128300, + "ĠngÄĥn": 133776, + "ĠngÅ©": 135518, + "Ġngươi": 133185, + "Ġngược": 138239, + "Ġngưá»Ŀi": 128251, + "Ġngạc": 143152, + "Ġngại": 138039, + "Ġngắn": 134780, + "Ġngá»§": 130370, + "Ġngừa": 142154, + "Ġngừng": 139474, + "Ġngữ": 133687, + "Ġngá»±c": 139997, + "Ġngá»į": 139137, + "Ġngá»įt": 139138, + "Ġngá»ĵi": 132055, + "Ġngá»Ļ": 142587, + "Ġngá»Ŀ": 132132, + "Ġnh": 20303, + "Ġnhi": 124599, + "Ġnhiên": 128452, + "Ġnhiêu": 135594, + "Ġnhiá»ģu": 128271, + "Ġnhiá»ħ": 133636, + "Ġnhiá»ħm": 133637, + "Ġnhiá»ĩm": 129414, + "Ġnhiá»ĩt": 130884, + "Ġnhu": 141580, + "ĠnhuáºŃn": 141581, + "Ġnhân": 128310, + "Ġnhãn": 143702, + "Ġnhé": 130236, + "Ġnhìn": 128780, + "Ġnhóm": 131270, + "ĠnhÃł": 128275, + "Ġnhư": 128260, + "Ġnhưng": 128344, + "Ġnhạc": 131245, + "Ġnhấn": 137429, + "Ġnhất": 128281, + "Ġnhắc": 136674, + "Ġnhắn": 142045, + "Ġnhằm": 131763, + "Ġnhẹ": 131314, + "ĠnháºŃn": 128370, + "ĠnháºŃp": 62572, + "ĠnháºŃt": 131330, + "Ġnhững": 128253, + "Ġnhá»±a": 140531, + "Ġnhá»ĭ": 137627, + "Ġnhá»ı": 128821, + "ĠnhỼ": 130537, + "Ġnhá»Ŀ": 132666, + "Ġni": 12788, + "Ġnib": 30285, + "Ġnic": 17327, + "Ġnice": 6419, + "Ġnicely": 28821, + "Ġnicer": 65184, + "Ġnich": 57871, + "Ġniche": 34049, + "Ġnicht": 8793, + "Ġnichts": 59765, + "Ġnick": 15988, + "Ġnickel": 51249, + "Ġnickname": 29399, + "Ġnicknamed": 89134, + "Ġnicotine": 48911, + "Ġnid": 70939, + "Ġnie": 11316, + "Ġniece": 79015, + "Ġnied": 75260, + "Ġniekt": 143841, + "Ġniektóry": 143842, + "Ġniektórych": 143843, + "Ġnier": 140327, + "Ġnieruch": 140328, + "Ġnieruchomo": 140329, + "ĠnieruchomoÅĽci": 140330, + "Ġniet": 20791, + "Ġnieu": 53801, + "Ġnieuwe": 60836, + "Ġniezb": 140433, + "ĠniezbÄĻ": 140434, + "ĠniezbÄĻdn": 140435, + "Ġnig": 82770, + "Ġnigeria": 90977, + "Ġnight": 3729, + "Ġnightclub": 69732, + "Ġnightlife": 92911, + "Ġnightly": 74760, + "Ġnightmare": 37811, + "Ġnightmares": 74262, + "Ġnights": 21461, + "Ġnighttime": 92644, + "Ġnihil": 98072, + "Ġnije": 86493, + "Ġnik": 75102, + "Ġnike": 90614, + "Ġnil": 2092, + "Ġnilai": 83367, + "Ġnim": 45692, + "Ġnin": 19550, + "Ġnine": 11627, + "Ġninete": 48742, + "Ġnineteen": 93835, + "Ġnineteenth": 64989, + "Ġninety": 77876, + "Ġning": 45507, + "Ġningu": 143259, + "Ġninguna": 98083, + "Ġninguém": 143260, + "Ġningún": 89599, + "Ġninja": 63851, + "Ġninth": 42541, + "Ġnip": 98470, + "Ġnipple": 87880, + "Ġnipples": 67112, + "Ġnir": 96824, + "Ġnisi": 68152, + "Ġnit": 24691, + "Ġnitrogen": 46403, + "Ġnive": 81332, + "Ġniveau": 51952, + "Ġnivel": 42345, + "Ġniên": 132883, + "Ġniño": 141339, + "Ġniños": 67256, + "Ġniż": 93759, + "Ġniá»ģm": 136141, + "Ġniá»ĩm": 132686, + "Ġnj": 36085, + "Ġnk": 79491, + "Ġnl": 20216, + "Ġnltk": 51064, + "Ġnm": 25723, + "Ġnn": 10883, + "Ġno": 902, + "Ġnob": 13049, + "Ġnoble": 34382, + "Ġnobody": 18581, + "Ġnoc": 75497, + "Ġnoch": 17790, + "Ġnoche": 74422, + "Ġnoct": 92001, + "Ġnod": 16004, + "Ġnodded": 34965, + "Ġnode": 2436, + "ĠnodeId": 66023, + "ĠnodeList": 89461, + "ĠnodeName": 76103, + "Ġnodes": 7798, + "Ġnodo": 52900, + "Ġnods": 82659, + "Ġnoe": 66439, + "Ġnoen": 71762, + "Ġnoexcept": 20583, + "Ġnog": 33310, + "Ġnoi": 57110, + "Ġnoir": 55562, + "Ġnoise": 11980, + "Ġnoises": 57348, + "Ġnoisy": 49280, + "Ġnok": 79733, + "Ġnoktas": 143934, + "Ġnoktası": 143935, + "Ġnom": 9662, + "Ġnombre": 12736, + "Ġnombres": 80598, + "Ġnombreux": 80686, + "Ġnome": 17129, + "Ġnomin": 24215, + "Ġnominal": 46755, + "Ġnominate": 87300, + "Ġnominated": 37948, + "Ġnomination": 28704, + "Ġnominations": 59598, + "Ġnomine": 23295, + "Ġnominee": 28211, + "Ġnominees": 59042, + "Ġnon": 2477, + "Ġnonatomic": 17861, + "Ġnonce": 39676, + "Ġnond": 73898, + "Ġnone": 6857, + "Ġnonetheless": 37813, + "Ġnonexistent": 86934, + "Ġnonlinear": 73998, + "Ġnonprofit": 32084, + "Ġnonprofits": 92260, + "Ġnons": 31695, + "Ġnonsense": 40802, + "Ġnonzero": 72232, + "Ġnood": 45795, + "Ġnoodles": 60160, + "Ġnoon": 37145, + "Ġnoop": 60829, + "Ġnop": 65026, + "Ġnoqa": 45621, + "Ġnor": 6329, + "Ġnord": 47634, + "Ġnoreferrer": 71604, + "Ġnorge": 65922, + "Ġnorm": 7465, + "Ġnormal": 4622, + "Ġnormalization": 48723, + "Ġnormalize": 21694, + "Ġnormalized": 29410, + "Ġnormally": 14279, + "Ġnormals": 53452, + "Ġnorms": 39751, + "Ġnors": 39202, + "Ġnorsk": 64267, + "Ġnorske": 60046, + "Ġnorth": 10200, + "Ġnortheast": 40829, + "Ġnortheastern": 86144, + "Ġnorthern": 18172, + "Ġnorthwest": 52242, + "Ġnos": 11891, + "Ġnose": 19142, + "Ġnoses": 96714, + "Ġnosotros": 67581, + "Ġnoss": 78630, + "Ġnossa": 76739, + "Ġnosso": 69692, + "Ġnost": 30612, + "Ġnostalg": 45089, + "Ġnostalgia": 64775, + "Ġnostalgic": 82274, + "Ġnostr": 56326, + "Ġnostra": 90027, + "Ġnostro": 83569, + "Ġnot": 537, + "Ġnota": 44038, + "Ġnotable": 27190, + "Ġnotably": 34046, + "Ġnotamment": 80086, + "Ġnotas": 85418, + "Ġnotation": 44197, + "Ġnotch": 50454, + "Ġnote": 5185, + "Ġnotebook": 37166, + "Ġnotebooks": 78531, + "Ġnoted": 10342, + "Ġnotes": 8388, + "Ġnoteworthy": 66379, + "Ġnothing": 4302, + "Ġnotice": 5293, + "Ġnoticeable": 42326, + "Ġnoticeably": 77888, + "Ġnoticed": 13686, + "Ġnotices": 34771, + "Ġnoticias": 95601, + "Ġnoticing": 61364, + "Ġnotification": 11540, + "Ġnotifications": 21969, + "Ġnotified": 29216, + "Ġnotifier": 73169, + "Ġnotifies": 96589, + "Ġnotify": 15456, + "ĠnotifyDataSetChanged": 91202, + "Ġnotifying": 97501, + "Ġnoting": 26305, + "Ġnotion": 22240, + "Ġnotions": 58582, + "Ġnotorious": 42981, + "Ġnotoriously": 72735, + "Ġnotre": 28349, + "Ġnotwithstanding": 71565, + "ĠnotÃŃcia": 141862, + "Ġnou": 19547, + "Ġnoun": 36921, + "Ġnouns": 89838, + "Ġnour": 45698, + "Ġnous": 16890, + "Ġnouve": 76601, + "Ġnouveau": 42353, + "Ġnouveaux": 86225, + "Ġnouvel": 34362, + "Ġnouvelle": 47559, + "Ġnouvelles": 78980, + "Ġnov": 6609, + "Ġnova": 40534, + "Ġnovamente": 86919, + "Ġnovel": 11514, + "Ġnovelist": 80647, + "Ġnoveller": 85614, + "Ġnovels": 31863, + "Ġnovelty": 66309, + "Ġnovember": 84223, + "Ġnovembre": 82515, + "Ġnovice": 71545, + "ĠnovitÃł": 143145, + "Ġnovo": 38323, + "Ġnow": 1431, + "Ġnowadays": 45777, + "Ġnowhere": 27112, + "Ġnowrap": 48744, + "Ġnozzle": 81881, + "Ġnp": 2595, + "Ġnpc": 36383, + "Ġnpm": 46422, + "Ġnr": 20262, + "Ġnrows": 85189, + "Ġnrw": 76506, + "Ġns": 12268, + "Ġnt": 31412, + "Ġnth": 55129, + "Ġnto": 61818, + "Ġntohs": 93464, + "Ġnu": 10908, + "Ġnuanced": 81791, + "Ġnuances": 83789, + "Ġnucle": 30384, + "Ġnuclear": 11249, + "Ġnuclei": 96092, + "Ġnucleus": 61507, + "Ġnud": 41441, + "Ġnude": 20036, + "Ġnudity": 91372, + "Ġnue": 11854, + "Ġnues": 84760, + "Ġnuest": 30887, + "Ġnuestra": 41691, + "Ġnuestras": 88510, + "Ġnuestro": 42462, + "Ġnuestros": 56913, + "Ġnueva": 37009, + "Ġnuevas": 79196, + "Ġnuevo": 28296, + "Ġnuevos": 70563, + "Ġnug": 81780, + "Ġnuis": 69438, + "Ġnuisance": 76641, + "Ġnuit": 84334, + "Ġnull": 845, + "Ġnulla": 60450, + "Ġnullable": 13258, + "Ġnullptr": 6827, + "Ġnum": 1629, + "ĠnumOf": 61575, + "ĠnumRows": 79050, + "Ġnuma": 92094, + "Ġnumb": 56271, + "Ġnumber": 1372, + "ĠnumberOf": 17512, + "ĠnumberOfRows": 44910, + "ĠnumberOfRowsInSection": 50150, + "ĠnumberWith": 49798, + "ĠnumberWithInt": 98277, + "Ġnumbered": 48826, + "Ġnumbering": 78262, + "Ġnumbers": 5109, + "Ġnumel": 76211, + "Ġnumer": 7857, + "Ġnumeral": 83292, + "Ġnumerator": 63533, + "Ġnumeric": 24064, + "Ġnumerical": 34776, + "Ġnumero": 20050, + "Ġnumeros": 55626, + "Ġnumerous": 12114, + "Ġnumpy": 8591, + "Ġnums": 10307, + "Ġnumérique": 141418, + "Ġnuméro": 97977, + "Ġnun": 28095, + "Ġnunca": 52743, + "Ġnuova": 85547, + "Ġnuovo": 72729, + "Ġnur": 12222, + "Ġnurs": 16996, + "Ġnurse": 28098, + "Ġnursery": 55126, + "Ġnurses": 36365, + "Ġnursing": 27356, + "Ġnurt": 59096, + "Ġnurture": 78430, + "Ġnurturing": 83811, + "Ġnuru": 75111, + "Ġnut": 9979, + "Ġnutrient": 49023, + "Ġnutrients": 36393, + "Ġnutrit": 68614, + "Ġnutrition": 25595, + "Ġnutritional": 42126, + "Ġnutritious": 77116, + "Ġnuts": 29949, + "Ġnutshell": 92913, + "Ġnutzen": 77919, + "Ġnuôi": 132645, + "Ġnv": 31440, + "Ġnw": 62532, + "Ġnx": 24496, + "Ġnxt": 70452, + "Ġny": 19004, + "Ġnya": 75189, + "Ġnye": 78667, + "Ġnylon": 51055, + "Ġnymph": 98301, + "Ġnz": 42942, + "Ġnº": 136352, + "Ġnâ": 130313, + "Ġnâng": 130314, + "Ġnão": 12393, + "Ġnä": 43117, + "Ġnäch": 62668, + "Ġnächste": 141257, + "Ġnächsten": 90935, + "Ġnäm": 143524, + "Ġnämlich": 143525, + "Ġnär": 64375, + "ĠnÃ¥": 42135, + "ĠnÃ¥r": 58690, + "Ġnæ": 81498, + "Ġnær": 56697, + "Ġné": 37505, + "Ġnécess": 62746, + "Ġnécessaire": 99143, + "Ġnét": 135437, + "Ġnên": 128317, + "Ġnë": 66143, + "Ġnó": 128525, + "Ġnói": 128339, + "Ġnóng": 131484, + "Ġnós": 91775, + "Ġnông": 132234, + "Ġnø": 54905, + "Ġnú": 136191, + "Ġnúi": 136192, + "Ġnúmero": 30211, + "Ġnúmeros": 69426, + "ĠnÃłng": 130453, + "ĠnÃło": 128352, + "ĠnÃły": 96535, + "ĠnÃŃvel": 136234, + "ĠnÄĥm": 128290, + "ĠnÄĥng": 128365, + "ĠnÆ¡i": 128837, + "ĠnÆ°á»Ľc": 128301, + "Ġnạ": 138420, + "Ġnạn": 132864, + "Ġnấu": 138102, + "Ġnắm": 135905, + "Ġnắng": 136583, + "Ġnằm": 130668, + "Ġnặng": 132932, + "Ġnếu": 128725, + "Ġnợ": 138178, + "Ġnữ": 128872, + "Ġnữa": 128870, + "Ġná»ģn": 131167, + "Ġná»iji": 132364, + "Ġná»ķi": 129144, + "Ġná»Ļ": 143280, + "Ġná»Ļi": 128811, + "Ġná»Ļp": 143281, + "Ġná»Ńa": 138405, + "Ġo": 297, + "Ġoa": 90812, + "Ġoak": 37871, + "Ġoasis": 85710, + "Ġoat": 71654, + "Ġoath": 47942, + "Ġoats": 84170, + "Ġoauth": 46415, + "Ġob": 1508, + "Ġobe": 97402, + "Ġobed": 48190, + "Ġobedience": 66094, + "Ġobedient": 95656, + "Ġoben": 89005, + "Ġober": 91245, + "Ġobese": 49220, + "Ġobesity": 31948, + "Ġobey": 40601, + "Ġobj": 2839, + "Ġobjc": 63326, + "Ġobject": 1633, + "ĠobjectAtIndex": 29320, + "ĠobjectForKey": 22684, + "ĠobjectId": 82887, + "ĠobjectMapper": 83912, + "ĠobjectType": 92802, + "Ġobjected": 75779, + "Ġobjection": 45555, + "Ġobjections": 53011, + "Ġobjective": 16538, + "Ġobjectively": 84021, + "Ġobjectives": 25398, + "Ġobjects": 6171, + "Ġobjet": 30160, + "Ġobjetivo": 54226, + "Ġobjeto": 35543, + "Ġobjetos": 74048, + "Ġobjs": 52937, + "Ġobl": 93344, + "Ġoblig": 12331, + "Ġobligated": 66876, + "Ġobligation": 28572, + "Ġobligations": 29155, + "Ġobligatory": 96924, + "Ġobliged": 53017, + "Ġobliv": 66005, + "Ġoblivious": 93270, + "Ġobr": 38194, + "Ġobra": 61734, + "Ġobras": 84749, + "Ġobrig": 69337, + "Ġobs": 7448, + "Ġobsc": 25647, + "Ġobscene": 83009, + "Ġobscure": 39534, + "Ġobscured": 95701, + "Ġobserv": 9282, + "Ġobservable": 39535, + "Ġobservation": 21930, + "Ġobservational": 89280, + "Ġobservations": 23722, + "Ġobserve": 22986, + "Ġobserved": 13166, + "Ġobserver": 22067, + "Ġobservers": 36543, + "Ġobserves": 80199, + "Ġobserving": 44971, + "Ġobsess": 55259, + "Ġobsessed": 43798, + "Ġobsession": 48535, + "Ġobsessive": 83032, + "Ġobsolete": 46066, + "Ġobst": 17864, + "Ġobstacle": 32187, + "Ġobstacles": 31016, + "Ġobstruct": 53192, + "Ġobstruction": 59951, + "ĠobsÅĤ": 142870, + "ĠobsÅĤugi": 142871, + "Ġobt": 4514, + "Ġobtain": 6851, + "Ġobtained": 12180, + "Ġobtaining": 19009, + "Ġobtains": 93563, + "Ġobten": 63190, + "Ġobtener": 49504, + "Ġobvious": 8036, + "Ġobviously": 13900, + "Ġoc": 17796, + "Ġocas": 93252, + "Ġocc": 2983, + "Ġoccas": 8986, + "Ġoccasion": 13101, + "Ġoccasional": 27861, + "Ġoccasionally": 22930, + "Ġoccasions": 24930, + "Ġoccult": 73826, + "Ġoccup": 10156, + "Ġoccupancy": 65319, + "Ġoccupants": 74672, + "Ġoccupation": 29647, + "Ġoccupational": 64095, + "Ġoccupations": 59866, + "Ġoccupied": 24367, + "Ġoccupies": 75754, + "Ġoccupy": 47578, + "Ġoccupying": 71180, + "Ġoccur": 12170, + "Ġoccured": 34942, + "Ġoccurred": 10017, + "Ġoccurrence": 31559, + "Ġoccurrences": 56015, + "Ġoccurring": 30865, + "Ġoccurs": 13666, + "Ġocean": 17951, + "Ġoceans": 53180, + "Ġoch": 11952, + "Ġoci": 93975, + "ĠocksÃ¥": 87316, + "Ġocor": 83972, + "Ġoct": 18491, + "Ġoctave": 73932, + "Ġocup": 74666, + "Ġocur": 91220, + "Ġoczy": 138959, + "ĠoczywiÅĽcie": 138960, + "Ġod": 10785, + "Ġodby": 140993, + "ĠodbyÅĤ": 140994, + "Ġodd": 10322, + "Ġoddly": 73444, + "Ġodds": 20785, + "Ġode": 72894, + "Ġodense": 74887, + "Ġoder": 12385, + "Ġodio": 86734, + "Ġodom": 87106, + "Ġodor": 52997, + "Ġodp": 50906, + "Ġodpowied": 70986, + "Ġoe": 68383, + "Ġof": 315, + "ĠofApp": 97742, + "ĠofType": 71754, + "Ġofere": 80952, + "Ġoferta": 84220, + "Ġoff": 1007, + "Ġoffen": 87533, + "Ġoffence": 34255, + "Ġoffences": 50212, + "Ġoffend": 83986, + "Ġoffended": 53121, + "Ġoffender": 46653, + "Ġoffenders": 40602, + "Ġoffending": 69569, + "Ġoffense": 16775, + "Ġoffenses": 46143, + "Ġoffensive": 15183, + "Ġoffer": 3010, + "Ġoffered": 8900, + "Ġoffering": 10004, + "Ġofferings": 32835, + "Ġoffers": 6081, + "Ġoffic": 2786, + "Ġoffice": 5163, + "Ġofficer": 9452, + "Ġofficers": 9611, + "Ġoffices": 19126, + "Ġofficial": 3946, + "Ġofficially": 18562, + "Ġofficials": 7359, + "Ġoffline": 26166, + "Ġoffre": 67805, + "Ġoffs": 73217, + "Ġoffseason": 39872, + "Ġoffset": 4347, + "ĠoffsetX": 73272, + "ĠoffsetY": 68038, + "Ġoffsetof": 61647, + "Ġoffsets": 35046, + "Ġoffshore": 38400, + "Ġoffspring": 45371, + "Ġoficial": 74440, + "Ġofrec": 63095, + "Ġofrece": 88439, + "Ġofs": 58061, + "Ġofstream": 75393, + "Ġoft": 42046, + "Ġoften": 3545, + "Ġog": 7350, + "Ġoggi": 88644, + "Ġogl": 93956, + "ĠoglÄħda": 141409, + "Ġogni": 54374, + "Ġogr": 95261, + "Ġogs": 43734, + "ĠogsÃ¥": 48387, + "Ġogó": 143809, + "Ġogóle": 143810, + "Ġogóln": 137415, + "ĠogÅĤ": 143303, + "ĠogÅĤos": 143304, + "ĠogÅĤoszen": 143305, + "ĠogÅĤoszeni": 143306, + "Ġoh": 14019, + "Ġohio": 95885, + "Ġohne": 28883, + "Ġoi": 74633, + "Ġoid": 48766, + "Ġoil": 5590, + "Ġoils": 31362, + "Ġoily": 77471, + "Ġok": 5394, + "Ġokay": 16910, + "Ġokhttp": 48032, + "Ġoko": 135249, + "ĠokoÅĤo": 135250, + "Ġokre": 138166, + "ĠokreÅĽ": 138167, + "ĠokreÅĽlon": 138168, + "Ġol": 8328, + "Ġolacaktır": 135651, + "Ġolacaģı": 139005, + "Ġolan": 56298, + "Ġolarak": 49894, + "Ġold": 2310, + "ĠoldValue": 49102, + "Ġolder": 9014, + "Ġoldest": 23513, + "Ġolds": 78473, + "Ġoldu": 57000, + "Ġolduk": 127365, + "Ġoldukları": 139796, + "Ġoldukça": 132821, + "ĠolduÄŁ": 66465, + "ĠolduÄŁu": 128508, + "ĠolduÄŁunu": 128776, + "Ġole": 31778, + "Ġoleh": 51316, + "Ġoli": 59136, + "Ġolig": 54884, + "Ġolive": 32113, + "Ġolmad": 127956, + "Ġolmadıģ": 139952, + "Ġolmadıģı": 137890, + "Ġolmadıģını": 139953, + "Ġolması": 129280, + "Ġolmasını": 142721, + "ĠolmuÅŁ": 136506, + "ĠolmuÅŁtur": 142843, + "Ġoltre": 91102, + "Ġolu": 127079, + "ĠoluÅŁ": 128065, + "ĠoluÅŁan": 139527, + "ĠoluÅŁtur": 127080, + "ĠoluÅŁturul": 142288, + "ĠoluÅŁu": 139656, + "Ġom": 7861, + "Ġoma": 47153, + "Ġomap": 88236, + "Ġomas": 94676, + "Ġomdat": 94020, + "Ġomega": 33898, + "Ġomin": 71175, + "Ġominous": 85496, + "Ġomission": 83118, + "Ġomit": 51044, + "Ġomitted": 39442, + "Ġomn": 49097, + "Ġomnip": 94429, + "Ġomp": 42925, + "Ġon": 389, + "ĠonActivityResult": 78840, + "ĠonAnimation": 88456, + "ĠonBackPressed": 70609, + "ĠonBind": 51234, + "ĠonBindViewHolder": 57494, + "ĠonBlur": 63973, + "ĠonCancel": 65125, + "ĠonCancelled": 88500, + "ĠonChange": 9095, + "ĠonChangeText": 60868, + "ĠonChanged": 55360, + "ĠonClick": 5697, + "ĠonClose": 39876, + "ĠonComplete": 55246, + "ĠonCreate": 12057, + "ĠonCreateOptionsMenu": 67094, + "ĠonCreateView": 41438, + "ĠonCreateViewHolder": 58164, + "ĠonData": 77662, + "ĠonDataChange": 98908, + "ĠonDelete": 39553, + "ĠonDestroy": 54189, + "ĠonError": 32430, + "ĠonFailure": 48334, + "ĠonFinish": 96905, + "ĠonFocus": 85974, + "ĠonHide": 95604, + "ĠonItemClick": 66774, + "ĠonKeyDown": 92608, + "ĠonLoad": 53431, + "ĠonMouse": 42280, + "ĠonNext": 94612, + "ĠonOptionsItemSelected": 63967, + "ĠonPage": 74871, + "ĠonPause": 81149, + "ĠonPostExecute": 83163, + "ĠonPress": 22976, + "ĠonPressed": 27002, + "ĠonRequest": 79008, + "ĠonResponse": 56236, + "ĠonResume": 61170, + "ĠonSave": 77444, + "ĠonSelect": 63518, + "ĠonStart": 57113, + "ĠonStop": 85230, + "ĠonSubmit": 22487, + "ĠonSuccess": 38702, + "ĠonTap": 35384, + "ĠonTouch": 70779, + "ĠonUpdate": 80402, + "ĠonView": 50734, + "ĠonViewCreated": 86634, + "Ġonboard": 55058, + "Ġonc": 77870, + "Ġonce": 3055, + "Ġonchange": 49252, + "Ġonclick": 12172, + "Ġonde": 49003, + "Ġonder": 37761, + "Ġone": 825, + "Ġones": 6174, + "Ġoneself": 56569, + "Ġongoing": 14195, + "Ġonion": 37327, + "Ġonions": 44597, + "Ġonline": 2860, + "Ġonload": 96964, + "Ġonly": 1172, + "Ġons": 48449, + "Ġonset": 40980, + "Ġonsite": 83869, + "Ġonslaught": 87692, + "Ġonstage": 92535, + "Ġont": 14508, + "Ġonto": 8630, + "Ġontology": 83269, + "Ġontvang": 95052, + "Ġontvangst": 50290, + "Ġonward": 78187, + "Ġonwards": 59425, + "Ġonze": 55805, + "Ġoo": 55909, + "Ġook": 27479, + "Ġop": 1179, + "Ġopacity": 18655, + "Ġopaque": 46484, + "Ġopc": 37027, + "Ġopcion": 72808, + "Ġopciones": 75436, + "Ġopción": 83689, + "Ġopcode": 30028, + "Ġopen": 1787, + "ĠopenFileDialog": 97371, + "Ġopendir": 95142, + "Ġopened": 8930, + "Ġopener": 35153, + "Ġopenid": 98726, + "Ġopening": 8568, + "Ġopenings": 48549, + "Ġopenly": 29347, + "Ġopenness": 70060, + "Ġopens": 15885, + "Ġopenssl": 80733, + "Ġoper": 1997, + "Ġopera": 42416, + "Ġoperand": 27213, + "Ġoperands": 54510, + "Ġoperate": 14476, + "Ġoperated": 23151, + "Ġoperates": 26057, + "Ġoperating": 10350, + "Ġoperation": 5666, + "Ġoperational": 24586, + "Ġoperations": 7525, + "Ġoperative": 63785, + "Ġoperatives": 72190, + "Ġoperator": 5675, + "Ġoperators": 19624, + "Ġopi": 35458, + "Ġopin": 7334, + "Ġopinion": 9459, + "Ġopinions": 17979, + "Ġopinión": 139421, + "Ġopioid": 47725, + "Ġopioids": 83360, + "Ġoportun": 61998, + "Ġopp": 3995, + "Ġoppon": 10359, + "Ġopponent": 14704, + "Ġopponents": 19386, + "Ġopport": 4800, + "Ġopportun": 85854, + "Ġopportunities": 10488, + "Ġopportunity": 6638, + "Ġoppos": 7922, + "Ġoppose": 33034, + "Ġopposed": 16087, + "Ġopposes": 75212, + "Ġopposing": 30222, + "Ġopposite": 14002, + "Ġopposition": 13758, + "Ġoppress": 89738, + "Ġoppressed": 66416, + "Ġoppression": 46648, + "Ġoppressive": 77512, + "Ġopr": 71063, + "Ġops": 27132, + "Ġopt": 3387, + "Ġoptarg": 82191, + "Ġopted": 40767, + "Ġoptic": 80109, + "Ġoptical": 28293, + "Ġoptics": 69885, + "Ġoptim": 7553, + "Ġoptimal": 22823, + "Ġoptimism": 53408, + "Ġoptimistic": 35936, + "Ġoptimization": 25262, + "Ġoptimizations": 81178, + "Ġoptimize": 29436, + "Ġoptimized": 33340, + "Ġoptimizer": 25632, + "Ġoptimizing": 73042, + "Ġoptimum": 53667, + "Ġopting": 74601, + "Ġoption": 2999, + "Ġoptional": 10101, + "Ġoptionally": 45524, + "Ġoptions": 2606, + "Ġopts": 12185, + "Ġor": 476, + "Ġora": 59325, + "Ġoracle": 52741, + "Ġoral": 20655, + "Ġorally": 97022, + "Ġoran": 135683, + "Ġorang": 49607, + "Ġorange": 18575, + "Ġoranges": 84038, + "Ġoranı": 135684, + "Ġoraz": 35470, + "Ġorb": 36366, + "Ġorbit": 26507, + "Ġorbital": 64591, + "Ġorbits": 85128, + "Ġorbs": 98405, + "Ġorc": 75691, + "Ġorch": 40145, + "Ġorchest": 69884, + "Ġorchestr": 65128, + "Ġorchestra": 68055, + "Ġorchestrated": 92067, + "Ġord": 6013, + "Ġordained": 85624, + "Ġordeal": 73681, + "Ġorden": 38388, + "Ġorder": 1973, + "ĠorderBy": 39694, + "ĠorderId": 48973, + "Ġorderby": 85836, + "Ġordered": 11457, + "Ġordering": 21391, + "Ġorderly": 84769, + "Ġorders": 10163, + "Ġordin": 48642, + "Ġordinal": 67948, + "Ġordinance": 47621, + "Ġordinances": 88764, + "Ġordinarily": 96981, + "Ġordinary": 19119, + "Ġore": 16144, + "Ġores": 75058, + "Ġorg": 1240, + "Ġorgan": 2872, + "Ġorganic": 17356, + "Ġorganis": 21185, + "Ġorganisation": 21424, + "Ġorganisations": 28433, + "Ġorganise": 69940, + "Ġorganised": 38333, + "Ġorganisers": 90460, + "Ġorganising": 87071, + "Ġorganism": 46020, + "Ġorganisms": 43204, + "Ġorganisé": 142892, + "Ġorganiz": 24299, + "Ġorganización": 140051, + "Ġorganization": 7321, + "Ġorganizational": 40195, + "Ġorganizations": 11104, + "Ġorganize": 30235, + "Ġorganized": 16645, + "Ġorganizer": 46202, + "Ġorganizers": 43218, + "Ġorganizing": 34721, + "Ġorgans": 35753, + "Ġorgas": 51473, + "Ġorgasm": 37665, + "Ġorgy": 84431, + "Ġori": 58635, + "Ġorient": 11485, + "Ġorientation": 16725, + "Ġorientations": 88835, + "Ġoriented": 41108, + "Ġorig": 2713, + "Ġorigen": 87389, + "Ġorigin": 6238, + "Ġoriginal": 4024, + "Ġoriginally": 13214, + "Ġoriginals": 85590, + "Ġoriginate": 81216, + "Ġoriginated": 43753, + "Ġoriginates": 98870, + "Ġoriginating": 70273, + "Ġorigins": 32372, + "Ġorm": 67602, + "Ġorn": 39445, + "Ġornament": 78660, + "Ġornaments": 84268, + "Ġoro": 94893, + "Ġorphan": 12746, + "Ġort": 63509, + "Ġorth": 29199, + "Ġorthodox": 91242, + "Ġorthogonal": 94580, + "Ġos": 2643, + "Ġosc": 30947, + "Ġoscill": 42424, + "Ġoscillator": 83641, + "Ġosg": 50848, + "Ġoslo": 36262, + "Ġosob": 71657, + "Ġosp": 63490, + "Ġoss": 30059, + "Ġost": 36896, + "Ġostat": 95369, + "Ġoste": 51268, + "Ġostensibly": 80494, + "Ġostr": 65994, + "Ġostream": 63511, + "Ġosób": 93443, + "Ġot": 14147, + "Ġother": 1008, + "ĠotherButtonTitles": 81187, + "Ġothers": 3800, + "Ġotherwise": 5937, + "Ġotp": 83431, + "Ġotra": 45395, + "Ġotras": 50866, + "Ġotro": 38092, + "Ġotros": 38405, + "Ġott": 52922, + "Ġou": 5908, + "Ġoud": 76973, + "Ġought": 21776, + "Ġoui": 96407, + "Ġounce": 53708, + "Ġounces": 48038, + "Ġour": 1039, + "Ġours": 11350, + "Ġourselves": 13217, + "Ġousted": 97455, + "Ġout": 700, + "ĠoutFile": 72809, + "Ġoutage": 88129, + "Ġoutbound": 73795, + "Ġoutbreak": 30088, + "Ġoutbreaks": 76771, + "Ġoutcome": 15274, + "Ġoutcomes": 19554, + "Ġoutcry": 87863, + "Ġoutdated": 40526, + "Ġoutdoor": 15789, + "Ġoutdoors": 33532, + "Ġouter": 15955, + "Ġoutf": 75881, + "Ġoutfield": 66337, + "Ġoutfile": 33178, + "Ġoutfit": 27303, + "Ġoutfits": 47274, + "Ġoutgoing": 36698, + "Ġouting": 64181, + "Ġoutings": 95627, + "Ġoutlaw": 64006, + "Ġoutlet": 26389, + "Ġoutlets": 27084, + "Ġoutlier": 88160, + "Ġoutliers": 86663, + "Ġoutline": 21101, + "Ġoutlined": 32640, + "Ġoutlines": 49629, + "Ġoutlining": 78674, + "Ġoutlook": 35621, + "Ġoutnumber": 84592, + "Ġoutpatient": 86130, + "Ġoutpost": 95578, + "Ġoutput": 2550, + "ĠoutputFile": 62843, + "ĠoutputPath": 95017, + "ĠoutputStream": 57961, + "Ġoutputs": 16275, + "Ġoutr": 17852, + "Ġoutra": 80707, + "Ġoutrage": 29129, + "Ġoutraged": 71264, + "Ġoutrageous": 54488, + "Ġoutras": 73702, + "Ġoutreach": 46110, + "Ġoutright": 39128, + "Ġoutro": 60658, + "Ġoutros": 56873, + "Ġouts": 22806, + "Ġoutset": 67764, + "Ġoutside": 4889, + "Ġoutsider": 74583, + "Ġoutsiders": 72189, + "Ġoutskirts": 70233, + "Ġoutsourcing": 74692, + "Ġoutspoken": 69279, + "Ġoutstanding": 18781, + "Ġoutward": 50946, + "Ġoutweigh": 61974, + "Ġov": 24550, + "Ġoval": 60037, + "Ġovar": 65068, + "Ġovarian": 84442, + "Ġoven": 23387, + "Ġover": 916, + "Ġoverall": 8084, + "Ġoverarching": 97436, + "Ġoverclock": 80194, + "Ġovercome": 22279, + "Ġovercoming": 72917, + "Ġovercrow": 87104, + "Ġoverd": 39452, + "Ġoverdose": 59453, + "Ġoverdue": 72526, + "Ġoverflow": 16484, + "Ġoverflowing": 93215, + "Ġoverhaul": 51674, + "Ġoverhe": 71289, + "Ġoverhead": 31015, + "Ġoverl": 28104, + "Ġoverlap": 27248, + "Ġoverlapping": 49817, + "Ġoverlaps": 76007, + "Ġoverlay": 24629, + "Ġoverlays": 96872, + "Ġoverload": 42184, + "Ġoverloaded": 74591, + "Ġoverlook": 30705, + "Ġoverlooked": 44436, + "Ġoverlooking": 52224, + "Ġoverly": 38432, + "Ġovernight": 24398, + "Ġoverposting": 94053, + "Ġoverpower": 97787, + "Ġoverridden": 51456, + "Ġoverride": 2812, + "Ġoverrides": 35455, + "Ġoverriding": 67871, + "Ġoverrun": 84278, + "Ġovers": 17460, + "Ġoverse": 19692, + "Ġoverseas": 24357, + "Ġoversee": 40903, + "Ġoverseeing": 67604, + "Ġoversees": 70845, + "Ġoversh": 65471, + "Ġovershadow": 84205, + "Ġoversight": 37089, + "Ġoversized": 61205, + "Ġoverst": 82409, + "Ġovert": 42561, + "Ġoverthrow": 62231, + "Ġovertime": 38499, + "Ġoverturn": 66587, + "Ġoverturned": 67998, + "Ġoverview": 23251, + "Ġoverweight": 49154, + "Ġoverwhel": 16783, + "Ġoverwhelm": 95748, + "Ġoverwhelmed": 42106, + "Ġoverwhelming": 22024, + "Ġoverwhelmingly": 54634, + "Ġoverwrite": 32216, + "Ġoverwritten": 59173, + "Ġow": 15570, + "Ġowe": 41110, + "Ġowed": 47909, + "Ġowes": 68844, + "Ġowing": 55512, + "Ġowl": 52269, + "Ġown": 1828, + "Ġowned": 12938, + "Ġowner": 6372, + "ĠownerId": 96923, + "Ġowners": 7824, + "Ġownership": 15278, + "Ġowning": 40277, + "Ġowns": 24256, + "Ġox": 18952, + "Ġoxid": 35072, + "Ġoxidation": 70062, + "Ġoxidative": 78301, + "Ġoxide": 50080, + "Ġoxy": 98590, + "Ġoxygen": 23552, + "Ġoy": 69348, + "Ġoz": 24596, + "Ġozone": 76437, + "Ġoù": 38623, + "Ġp": 281, + "ĠpBuffer": 92400, + "ĠpData": 48639, + "ĠpH": 36043, + "ĠpInfo": 93583, + "ĠpItem": 93640, + "ĠpNode": 64566, + "ĠpObj": 59952, + "ĠpParent": 94220, + "Ġpa": 7106, + "Ġpaar": 48100, + "Ġpac": 22894, + "Ġpace": 17857, + "Ġpaced": 77680, + "Ġpaciente": 72412, + "Ġpacientes": 83924, + "Ġpacing": 82140, + "Ġpack": 3769, + "Ġpackage": 6328, + "ĠpackageName": 66604, + "Ġpackaged": 46137, + "Ġpackages": 14185, + "Ġpackaging": 23189, + "Ġpacked": 19375, + "Ġpacket": 10151, + "Ġpackets": 27035, + "Ġpacking": 35713, + "Ġpacks": 30306, + "Ġpacman": 96864, + "Ġpact": 59721, + "Ġpad": 11016, + "Ġpada": 31811, + "Ġpadd": 53169, + "Ġpadded": 43868, + "Ġpadding": 5300, + "ĠpaddingBottom": 60665, + "ĠpaddingHorizontal": 84098, + "ĠpaddingLeft": 54191, + "ĠpaddingRight": 70015, + "ĠpaddingTop": 48534, + "Ġpaddle": 39303, + "Ġpadr": 93060, + "Ġpadre": 61020, + "Ġpads": 35995, + "Ġpadx": 54822, + "Ġpady": 52632, + "Ġpaed": 94147, + "Ġpag": 14774, + "Ġpagamento": 78807, + "Ġpagan": 83802, + "Ġpagar": 83799, + "Ġpage": 2150, + "ĠpageCount": 85751, + "ĠpageIndex": 59246, + "ĠpageInfo": 97267, + "ĠpageNo": 84882, + "ĠpageNum": 67571, + "ĠpageNumber": 60141, + "ĠpageSize": 22635, + "ĠpageTitle": 87221, + "Ġpageable": 70147, + "Ġpager": 52588, + "Ġpages": 6816, + "Ġpagina": 45965, + "Ġpaginate": 81974, + "Ġpagination": 28495, + "Ġpaginator": 69395, + "Ġpaging": 56287, + "Ġpago": 64377, + "Ġpai": 60600, + "Ġpaid": 7171, + "Ġpain": 6646, + "Ġpainful": 25118, + "Ġpainfully": 75599, + "Ġpains": 50166, + "Ġpainstaking": 97613, + "Ġpaint": 6177, + "Ġpainted": 23983, + "Ġpainter": 29481, + "Ġpainters": 96853, + "Ġpainting": 18824, + "Ġpaintings": 35592, + "Ġpaints": 56848, + "Ġpair": 6716, + "Ġpaired": 34426, + "Ġpairing": 47713, + "Ġpairs": 13530, + "Ġpairwise": 92759, + "Ġpais": 61283, + "Ġpaj": 90610, + "Ġpak": 38014, + "Ġpakistan": 85998, + "Ġpal": 10854, + "Ġpalabra": 72937, + "Ġpalabras": 73204, + "Ġpalace": 43339, + "Ġpalate": 76238, + "Ġpalavra": 94647, + "Ġpale": 27539, + "Ġpalette": 26308, + "Ġpalindrome": 73097, + "Ġpall": 66497, + "Ġpallet": 58803, + "Ġpalm": 32452, + "Ġpalms": 74718, + "Ġpalp": 74549, + "Ġpals": 83087, + "Ġpam": 41190, + "Ġpami": 137695, + "ĠpamiÄĻ": 137696, + "ĠpamiÄĻta": 137697, + "Ġpamph": 74831, + "Ġpan": 7215, + "Ġpanc": 53474, + "Ġpancakes": 79860, + "Ġpancre": 61168, + "Ġpancreatic": 97044, + "Ġpand": 12217, + "Ġpanda": 88222, + "Ġpandas": 18617, + "Ġpandemic": 27422, + "Ġpane": 37322, + "Ġpanel": 6945, + "Ġpanels": 21285, + "Ġpang": 96874, + "Ġpanic": 21975, + "Ġpanicked": 93010, + "Ġpanor": 67804, + "Ġpanorama": 91678, + "Ġpanoramic": 83257, + "Ġpans": 63783, + "Ġpant": 25279, + "Ġpantalla": 70024, + "Ġpanties": 55357, + "Ġpantry": 68257, + "Ġpants": 24549, + "Ġpap": 25298, + "Ġpapel": 69157, + "Ġpaper": 5567, + "Ġpaperback": 85731, + "Ġpapers": 15689, + "Ġpaperwork": 44631, + "Ġpar": 1346, + "Ġpara": 3348, + "Ġparach": 73945, + "Ġparachute": 98120, + "Ġparad": 27317, + "Ġparade": 36504, + "Ġparadigm": 48240, + "Ġparadise": 49752, + "Ġparadox": 51213, + "Ġparagraph": 14311, + "Ġparagraphs": 42643, + "Ġparalle": 57030, + "Ġparallel": 15279, + "Ġparallels": 71374, + "Ġparalysis": 85039, + "Ġparalyzed": 87868, + "Ġparam": 1685, + "ĠparamInt": 40750, + "ĠparamMap": 87056, + "ĠparamName": 90696, + "ĠparamString": 49641, + "Ġparameter": 5733, + "Ġparameters": 5029, + "Ġparametro": 75137, + "Ġparametros": 60405, + "Ġparamount": 73410, + "Ġparams": 3628, + "Ġparan": 47582, + "Ġparanoia": 82444, + "Ġparanoid": 68663, + "Ġparanormal": 91597, + "Ġparaph": 62230, + "Ġparas": 32303, + "Ġparasite": 84473, + "Ġparasites": 78283, + "Ġparc": 44640, + "Ġparce": 64921, + "Ġparcel": 29309, + "Ġparcels": 83927, + "Ġparch": 65202, + "Ġparchment": 86104, + "Ġpard": 89947, + "Ġpardon": 64168, + "Ġpare": 23191, + "Ġparece": 64017, + "Ġpared": 97011, + "Ġpareja": 74786, + "Ġparen": 37940, + "Ġparent": 2681, + "ĠparentId": 55885, + "ĠparentNode": 74955, + "Ġparental": 45579, + "Ġparentheses": 73975, + "Ġparenthesis": 95356, + "Ġparenting": 47958, + "Ġparents": 6562, + "Ġparfait": 85866, + "Ġpari": 66982, + "Ġparis": 40858, + "Ġparish": 51197, + "Ġparity": 49615, + "Ġpark": 6118, + "Ġparked": 42235, + "Ġparking": 12921, + "Ġparks": 26845, + "Ġparl": 43214, + "Ġparler": 88404, + "Ġparliament": 21130, + "Ġparliamentary": 43150, + "Ġparm": 59818, + "Ġparms": 58477, + "Ġparody": 66165, + "Ġparole": 48261, + "Ġpars": 10154, + "Ġparse": 4715, + "ĠparseFloat": 25763, + "ĠparseInt": 12891, + "Ġparsed": 15676, + "Ġparser": 6729, + "Ġparsers": 87073, + "Ġparses": 70835, + "Ġparsing": 22314, + "Ġparsley": 87521, + "Ġpart": 949, + "Ġpartager": 96348, + "Ġparte": 19538, + "Ġparted": 84056, + "Ġparten": 97294, + "Ġpartes": 80182, + "Ġparti": 78888, + "Ġpartial": 7130, + "Ġpartially": 25244, + "Ġpartic": 2536, + "Ġparticip": 5182, + "Ġparticipación": 140594, + "Ġparticipant": 24880, + "Ġparticipants": 13026, + "Ġparticipate": 15759, + "Ġparticipated": 30308, + "Ġparticipates": 90187, + "Ġparticipating": 23528, + "Ġparticipation": 20239, + "Ġparticipação": 140294, + "Ġparticle": 18790, + "Ġparticles": 18730, + "Ġparticul": 52908, + "Ġparticular": 3953, + "Ġparticularly": 7945, + "Ġparticulars": 85874, + "Ġparticulière": 136547, + "Ġpartida": 96547, + "Ġpartido": 73041, + "Ġpartie": 47801, + "Ġparties": 9677, + "Ġpartir": 30532, + "Ġpartisan": 48538, + "Ġpartition": 16658, + "Ġpartitions": 46688, + "Ġpartly": 27037, + "Ġpartner": 8263, + "Ġpartnered": 52219, + "Ġpartnering": 69120, + "Ġpartners": 8549, + "Ġpartnership": 15305, + "Ġpartnerships": 39605, + "Ġparts": 5479, + "Ġparty": 4614, + "Ġparça": 132831, + "Ġpas": 6368, + "Ġpasa": 86676, + "Ġpasado": 57416, + "Ġpasar": 66614, + "Ġpaso": 61677, + "Ġpass": 1494, + "Ġpassage": 21085, + "Ġpassages": 46769, + "Ġpasse": 38804, + "Ġpassed": 5823, + "Ġpassenger": 23148, + "Ġpassengers": 22172, + "Ġpasser": 44247, + "Ġpasses": 16211, + "Ġpassing": 12299, + "Ġpassion": 11677, + "Ġpassionate": 24425, + "Ġpassionately": 91825, + "Ġpassions": 57624, + "Ġpassive": 27879, + "Ġpassphrase": 75493, + "Ġpassport": 25458, + "Ġpassports": 74426, + "Ġpasswd": 52037, + "Ġpassword": 3552, + "Ġpasswords": 33716, + "Ġpassé": 80521, + "Ġpast": 3267, + "Ġpasta": 36447, + "Ġpaste": 24937, + "Ġpastor": 42944, + "Ġpastoral": 89271, + "Ġpastors": 91320, + "Ġpastry": 73055, + "Ġpasture": 88255, + "Ġpat": 3272, + "Ġpatch": 10900, + "Ġpatched": 71732, + "Ġpatches": 28660, + "Ġpatent": 24571, + "Ġpatented": 62612, + "Ġpatents": 45897, + "Ġpaternal": 92267, + "Ġpath": 1815, + "ĠpathMatch": 86018, + "Ġpathetic": 66063, + "Ġpathlib": 53159, + "Ġpathname": 39683, + "Ġpathogens": 77184, + "Ġpathological": 88861, + "Ġpathology": 75941, + "Ġpaths": 12716, + "Ġpathway": 37870, + "Ġpathways": 42914, + "Ġpatience": 29301, + "Ġpatient": 8720, + "Ġpatiently": 78161, + "Ġpatients": 6835, + "Ġpatio": 31178, + "Ġpatië": 142207, + "Ġpatiënt": 142208, + "Ġpatri": 28225, + "Ġpatriarch": 70221, + "Ġpatriot": 66517, + "Ġpatriotic": 76684, + "Ġpatriotism": 89342, + "Ġpatrol": 32522, + "Ġpatrols": 87042, + "Ġpatron": 35280, + "Ġpatrons": 51760, + "Ġpatt": 83067, + "Ġpatter": 91993, + "Ġpattern": 5383, + "Ġpatterns": 12624, + "Ġpau": 83195, + "Ġpaul": 67011, + "Ġpause": 18087, + "Ġpaused": 34495, + "Ġpauses": 85832, + "Ġpav": 39116, + "Ġpave": 93023, + "Ġpaved": 62575, + "Ġpavement": 64342, + "Ġpaving": 81468, + "Ġpaw": 76838, + "Ġpawn": 53455, + "Ġpay": 2291, + "Ġpayable": 44591, + "Ġpaycheck": 76979, + "Ġpayday": 52606, + "Ġpayer": 69275, + "Ġpaying": 12515, + "ĠpaylaÅŁ": 131056, + "Ġpayload": 7729, + "Ġpayloads": 89523, + "Ġpayment": 8160, + "Ġpayments": 14173, + "Ġpayoff": 84265, + "Ġpayout": 45888, + "Ġpayouts": 89529, + "Ġpaypal": 67793, + "Ġpayroll": 45108, + "Ġpays": 21241, + "Ġpaz": 79485, + "ĠpaÃŃs": 41665, + "ĠpaÃŃses": 70877, + "Ġpaź": 138111, + "Ġpaździ": 138112, + "Ġpaździern": 138113, + "Ġpaździernika": 138114, + "ĠpaÅĦst": 134434, + "ĠpaÅĦstw": 134435, + "Ġpb": 17310, + "Ġpc": 13312, + "Ġpca": 80896, + "Ġpcap": 94554, + "Ġpcb": 86240, + "Ġpci": 45471, + "Ġpcl": 51881, + "Ġpcm": 68647, + "Ġpcs": 70299, + "Ġpct": 81847, + "Ġpd": 7744, + "Ġpdata": 70311, + "Ġpdb": 47984, + "Ġpdf": 12782, + "Ġpdo": 90133, + "Ġpdu": 96969, + "Ġpe": 1051, + "Ġpea": 97663, + "Ġpeace": 8919, + "Ġpeaceful": 25650, + "Ġpeacefully": 66779, + "Ġpeach": 72088, + "Ġpeak": 16162, + "Ġpeaked": 77192, + "Ġpeaks": 38935, + "Ġpeanut": 49833, + "Ġpeanuts": 89077, + "Ġpear": 37690, + "Ġpearl": 69623, + "Ġpearls": 91632, + "Ġpeas": 46036, + "Ġpeasant": 88938, + "Ġpeasants": 75747, + "Ġpec": 30000, + "Ġpeculiar": 43697, + "Ġped": 10477, + "Ġpedal": 48601, + "Ġpedals": 86406, + "Ġpedest": 29226, + "Ġpedestal": 91655, + "Ġpedestrian": 46754, + "Ġpedestrians": 61458, + "Ġpediatric": 52066, + "Ġpedido": 51794, + "Ġpedig": 65682, + "Ġpedigree": 92954, + "Ġpee": 64138, + "Ġpeek": 32592, + "Ġpeel": 57112, + "Ġpeeled": 82512, + "Ġpeer": 14397, + "Ġpeers": 25029, + "Ġpeg": 30687, + "Ġpel": 11814, + "Ġpela": 32723, + "Ġpelic": 75333, + "Ġpellet": 67410, + "Ġpellets": 77320, + "Ġpelo": 27525, + "Ġpelos": 82839, + "Ġpelvic": 83068, + "ĠpelÃŃcul": 65740, + "ĠpelÃŃcula": 96216, + "Ġpem": 54184, + "Ġpemb": 63567, + "Ġpen": 5750, + "Ġpena": 94457, + "Ġpenal": 46326, + "Ġpenalties": 29986, + "Ġpenalty": 16346, + "Ġpenc": 65274, + "Ġpenchant": 98680, + "Ġpencil": 46118, + "Ġpencils": 96338, + "Ġpend": 41530, + "Ġpendant": 41744, + "Ġpending": 15280, + "Ġpenet": 31656, + "Ġpenetr": 43705, + "Ġpenetrate": 63755, + "Ġpenetrated": 96822, + "Ġpenetrating": 89200, + "Ġpenetration": 43496, + "Ġpeng": 36686, + "Ġpeninsula": 65395, + "Ġpenis": 22951, + "Ġpenn": 42949, + "Ġpenned": 77697, + "Ġpenny": 49069, + "Ġpens": 22601, + "Ġpensar": 93051, + "Ġpense": 72853, + "Ġpension": 27681, + "Ġpensions": 63763, + "Ġpensé": 140016, + "Ġpent": 20189, + "Ġpentru": 26220, + "Ġpeny": 92933, + "Ġpeople": 1251, + "Ġpeoples": 31438, + "Ġpep": 86595, + "Ġpepp": 82704, + "Ġpepper": 24353, + "Ġpeppers": 57473, + "Ġpeptide": 71149, + "Ġpeptides": 95334, + "Ġpequ": 80672, + "Ġpeque": 47367, + "Ġpequeña": 137005, + "Ġpequeño": 133894, + "Ġper": 817, + "Ġperc": 43463, + "Ġperce": 61722, + "Ġperceive": 44393, + "Ġperceived": 25536, + "Ġpercent": 3266, + "Ġpercentage": 11414, + "Ġpercentages": 53108, + "Ġpercentile": 65157, + "Ġpercept": 77532, + "Ġperception": 20431, + "Ġperceptions": 50766, + "Ġperch": 41883, + "Ġperchè": 138275, + "Ġperché": 60955, + "Ġpercussion": 83749, + "Ġperd": 40418, + "Ġperder": 77896, + "Ġperennial": 73618, + "Ġperf": 21784, + "Ġperfect": 4727, + "Ġperfected": 94146, + "Ġperfection": 38043, + "Ġperfectly": 13942, + "Ġperfil": 63112, + "Ġperfor": 83768, + "Ġperform": 2736, + "Ġperformance": 5068, + "Ġperformances": 23675, + "Ġperformed": 10660, + "Ġperformer": 48154, + "Ġperformers": 44418, + "Ġperforming": 16380, + "Ġperforms": 26674, + "Ġperfume": 66343, + "Ġperg": 89785, + "Ġperhaps": 8365, + "Ġperi": 61157, + "Ġperil": 59292, + "Ġperimeter": 46342, + "Ġperiod": 4168, + "Ġperiodic": 38345, + "Ġperiodically": 47482, + "Ġperiodo": 70455, + "Ġperiods": 18346, + "Ġperipheral": 34588, + "Ġperipherals": 90884, + "Ġperish": 82117, + "Ġperk": 65093, + "Ġperks": 57186, + "Ġperl": 56056, + "Ġperm": 13854, + "Ġpermalink": 79228, + "Ġperman": 11624, + "Ġpermanent": 15330, + "Ġpermanently": 30759, + "Ġperme": 54324, + "Ġpermet": 28457, + "Ġpermissible": 75245, + "Ġpermission": 7882, + "Ġpermissions": 8541, + "Ġpermit": 11549, + "Ġpermite": 51503, + "Ġpermits": 28690, + "Ġpermitted": 15129, + "Ġpermitting": 61074, + "Ġperms": 82282, + "Ġpermutation": 58441, + "Ġpermutations": 71949, + "Ġpero": 19438, + "Ġperpendicular": 76833, + "Ġperpet": 21585, + "Ġperpetr": 38354, + "Ġperpetrated": 92330, + "Ġperpetrator": 84536, + "Ġperpetrators": 65346, + "Ġperpetual": 65593, + "Ġperplex": 73152, + "Ġperror": 40731, + "Ġpers": 7413, + "Ġperse": 54150, + "Ġpersec": 37468, + "Ġpersecuted": 92598, + "Ġpersecution": 55429, + "Ġpersever": 72978, + "Ġperseverance": 98741, + "Ġpersist": 22334, + "Ġpersisted": 52995, + "Ġpersistence": 40956, + "Ġpersistent": 24999, + "Ġpersists": 66045, + "Ġperso": 76972, + "Ġperson": 1697, + "Ġpersona": 27955, + "Ġpersonal": 4345, + "Ġpersonalised": 84248, + "Ġpersonalities": 43808, + "Ġpersonality": 17294, + "Ġpersonalize": 80838, + "Ġpersonalized": 34549, + "Ġpersonally": 15726, + "Ġpersonals": 47730, + "Ġpersonas": 31425, + "Ġpersone": 60688, + "Ġpersonn": 96152, + "Ġpersonne": 53513, + "Ġpersonnel": 16849, + "Ġpersonnes": 47477, + "Ġpersons": 11186, + "Ġperspective": 13057, + "Ġperspectives": 38455, + "Ġpersu": 23685, + "Ġpersuade": 49941, + "Ġpersuaded": 64001, + "Ġpersuasion": 97124, + "Ġpersuasive": 65243, + "Ġpersön": 77420, + "Ġpersönlich": 134776, + "Ġpert": 18212, + "Ġpertaining": 45931, + "Ġpertinent": 61523, + "Ġperu": 91859, + "Ġperv": 46174, + "Ġpervasive": 70767, + "Ġperverse": 94804, + "Ġperò": 132085, + "ĠperÃŃ": 76082, + "ĠperÃŃodo": 86624, + "Ġpes": 18050, + "Ġpeso": 36935, + "Ġpesos": 84147, + "Ġpesquisa": 94630, + "Ġpessim": 72523, + "Ġpesso": 39575, + "Ġpessoa": 56630, + "Ġpessoas": 45962, + "Ġpest": 43785, + "Ġpestic": 44095, + "Ġpesticide": 83573, + "Ġpesticides": 60157, + "Ġpests": 74956, + "Ġpet": 6753, + "Ġpetals": 95640, + "Ġpeter": 93987, + "Ġpetit": 45010, + "Ġpetite": 37820, + "Ġpetites": 97285, + "Ġpetition": 20361, + "Ġpetitioner": 83403, + "Ġpetitions": 74454, + "Ġpetits": 85596, + "Ġpetrol": 57083, + "Ġpetroleum": 58963, + "Ġpets": 25103, + "Ġpetty": 60485, + "Ġpeu": 27700, + "Ġpeut": 24666, + "Ġpeuvent": 55211, + "Ġpeux": 93197, + "Ġpew": 61416, + "ĠpewnoÅĽci": 142372, + "ĠpewnoÅĽciÄħ": 142373, + "Ġpeça": 137395, + "ĠpeÅĤ": 143904, + "ĠpeÅĤne": 143905, + "Ġpf": 27418, + "Ġpg": 17495, + "Ġph": 1319, + "Ġphantom": 69708, + "Ġpharm": 15080, + "Ġpharmac": 35349, + "Ġpharmaceutical": 34310, + "Ġpharmacies": 76505, + "Ġpharmacist": 89294, + "Ġpharmacy": 32300, + "Ġphase": 10262, + "Ġphased": 85229, + "Ġphases": 34430, + "Ġphen": 14018, + "Ġphenomen": 16735, + "Ġphenomena": 43147, + "Ġphenomenal": 56392, + "Ġphenomenon": 24844, + "Ġphenotype": 81323, + "Ġphi": 22693, + "Ġphil": 10664, + "Ġphilanth": 57638, + "Ġphilippines": 79904, + "Ġphilosoph": 13556, + "Ġphilosopher": 54375, + "Ġphilosophers": 60687, + "Ġphilosophical": 40803, + "Ġphilosophy": 19128, + "Ġphishing": 98097, + "Ġphiên": 133234, + "Ġphiếu": 137535, + "Ġphoenix": 89715, + "Ġphon": 50823, + "Ġphone": 4540, + "ĠphoneNumber": 42088, + "Ġphones": 17619, + "Ġphosph": 31988, + "Ġphosphate": 78006, + "Ġphosphory": 93989, + "Ġphot": 4503, + "Ġphoto": 6548, + "Ġphotoc": 98236, + "Ġphotograph": 10300, + "Ġphotographed": 56203, + "Ġphotographer": 28767, + "Ġphotographers": 47177, + "Ġphotographic": 69064, + "Ġphotographs": 24248, + "Ġphotography": 23751, + "Ġphoton": 67910, + "Ġphotons": 88135, + "Ġphotos": 7249, + "Ġphotoshop": 98761, + "Ġphp": 24362, + "Ġphrase": 17133, + "Ġphrases": 31747, + "Ġphy": 36455, + "Ġphys": 4471, + "Ġphysic": 54024, + "Ġphysical": 6961, + "Ġphysically": 21893, + "Ġphysician": 27279, + "Ġphysicians": 34844, + "Ġphysicist": 82223, + "Ġphysicists": 97317, + "Ġphysics": 21321, + "Ġphysiological": 52094, + "Ġphysiology": 77052, + "Ġphysique": 63341, + "Ġphá": 129108, + "Ġpháp": 128513, + "Ġphát": 128362, + "Ġphân": 129036, + "Ġphé": 130223, + "Ġphép": 130224, + "Ġphê": 134634, + "Ġphòng": 128441, + "Ġphó": 142267, + "Ġphóng": 136607, + "Ġphù": 129831, + "Ġphú": 140116, + "Ġphúc": 133765, + "Ġphút": 130747, + "ĠphÃŃ": 128812, + "ĠphÃŃa": 129854, + "Ġphương": 128675, + "Ġphưá»Ŀng": 140515, + "Ġphạm": 130375, + "Ġphạt": 138675, + "Ġphải": 128282, + "Ġphản": 131084, + "Ġphấn": 142094, + "Ġphần": 128463, + "Ġphẩm": 79479, + "Ġphẫu": 140552, + "ĠpháºŃn": 134851, + "Ġphụ": 128877, + "Ġphục": 128904, + "Ġphá»§": 130388, + "Ġphức": 142244, + "Ġphá»ij": 128974, + "Ġphá»iji": 131628, + "Ġphá»ķ": 132175, + "Ġpi": 8938, + "Ġpian": 59066, + "Ġpiano": 26278, + "Ġpic": 10319, + "Ġpick": 3735, + "Ġpicked": 12771, + "Ġpicker": 31456, + "ĠpickerView": 88412, + "Ġpicking": 21132, + "Ġpickle": 22186, + "Ġpicks": 21895, + "Ġpickup": 29586, + "Ġpickups": 92038, + "Ġpicnic": 54462, + "Ġpics": 21356, + "Ġpict": 71729, + "Ġpicture": 6802, + "ĠpictureBox": 38009, + "Ġpictured": 41566, + "Ġpictures": 9185, + "Ġpicturesque": 73901, + "Ġpid": 14814, + "Ġpie": 4349, + "Ġpiece": 6573, + "Ġpieces": 9666, + "Ġpied": 43530, + "Ġpiel": 80846, + "Ġpien": 76226, + "ĠpieniÄħ": 142637, + "ĠpieniÄħd": 142638, + "ĠpieniÄħdze": 142639, + "Ġpier": 21944, + "Ġpierced": 88413, + "Ġpiercing": 59120, + "Ġpierws": 67460, + "Ġpies": 60677, + "Ġpig": 23694, + "Ġpige": 49355, + "Ġpigeon": 92544, + "Ġpiger": 38771, + "Ġpigment": 76578, + "Ġpigs": 48810, + "Ġpii": 74030, + "Ġpij": 65234, + "Ġpik": 47150, + "Ġpil": 11459, + "Ġpile": 26306, + "Ġpiled": 76331, + "Ġpiles": 58772, + "Ġpilgr": 54598, + "Ġpilgrimage": 89313, + "Ġpill": 15177, + "Ġpillar": 61207, + "Ġpillars": 63882, + "Ġpillow": 42325, + "Ġpillows": 65926, + "Ġpills": 25097, + "Ġpilot": 17708, + "Ġpilots": 35277, + "Ġpimp": 84348, + "Ġpin": 8983, + "ĠpinMode": 90552, + "Ġpinch": 49246, + "Ġpine": 33597, + "Ġpineapple": 77082, + "Ġping": 29998, + "Ġpink": 18217, + "Ġpinnacle": 98489, + "Ġpinned": 47709, + "Ġpinpoint": 66538, + "Ġpins": 26944, + "Ġpint": 45646, + "Ġpinterest": 61016, + "Ġpione": 28843, + "Ġpioneer": 52947, + "Ġpioneered": 95634, + "Ġpioneering": 70574, + "Ġpioneers": 82307, + "Ġpip": 23626, + "Ġpipe": 13647, + "Ġpipeline": 15301, + "Ġpipelines": 57673, + "Ġpipes": 31175, + "Ġpiping": 76479, + "Ġpir": 29611, + "Ġpiracy": 77212, + "Ġpirate": 53966, + "Ġpirates": 61475, + "Ġpis": 62314, + "Ġpisc": 67368, + "Ġpiss": 39940, + "Ġpissed": 63436, + "Ġpist": 23875, + "Ġpistol": 39436, + "Ġpistols": 83647, + "Ġpiston": 75325, + "Ġpit": 19838, + "Ġpitch": 9649, + "Ġpitched": 48998, + "Ġpitcher": 40970, + "Ġpitchers": 60550, + "Ġpitches": 48541, + "Ġpitching": 46441, + "Ġpitfalls": 80975, + "Ġpits": 61908, + "Ġpity": 56943, + "Ġpiv": 41330, + "Ġpivot": 26045, + "Ġpivotal": 59750, + "Ġpix": 31574, + "Ġpixel": 12955, + "Ġpixels": 15751, + "Ġpixmap": 84351, + "Ġpizza": 22502, + "Ġpizzas": 87770, + "Ġpiè": 69377, + "Ġpièce": 133472, + "Ġpiù": 22765, + "Ġpj": 72613, + "Ġpk": 22458, + "Ġpkg": 24793, + "Ġpkt": 34654, + "Ġpl": 625, + "Ġpla": 39752, + "Ġplaats": 73857, + "Ġplac": 28860, + "Ġplace": 1992, + "Ġplacebo": 42615, + "Ġplaced": 9099, + "Ġplaceholder": 5878, + "Ġplaceholders": 78428, + "Ġplacement": 21448, + "Ġplacements": 88393, + "Ġplacer": 71696, + "Ġplaces": 7482, + "Ġplacing": 24047, + "Ġplag": 38037, + "Ġplage": 85351, + "Ġplagiar": 67573, + "Ġplagiarism": 83898, + "Ġplague": 54437, + "Ġplagued": 65302, + "Ġplain": 14396, + "Ġplainly": 64295, + "Ġplains": 77366, + "Ġplaint": 23372, + "Ġplaintext": 60847, + "Ġplaintiff": 42002, + "Ġplaintiffs": 50564, + "Ġplais": 70268, + "Ġplaisir": 79310, + "Ġplan": 3119, + "Ġplane": 11031, + "Ġplanes": 24730, + "Ġplanet": 11580, + "Ġplanetary": 61838, + "Ġplanets": 32875, + "Ġplank": 72087, + "Ġplanned": 12909, + "Ġplanner": 49711, + "Ġplanners": 67452, + "Ġplanning": 9115, + "Ġplano": 84328, + "Ġplans": 6649, + "Ġplant": 6008, + "Ġplanta": 81665, + "Ġplantation": 81952, + "Ġplanted": 38341, + "Ġplanting": 47014, + "Ġplants": 10779, + "Ġplaque": 60364, + "Ġplasma": 31326, + "Ġplast": 87147, + "Ġplaster": 61927, + "Ġplastic": 12188, + "Ġplastics": 67286, + "Ġplat": 44989, + "Ġplata": 62786, + "Ġplataforma": 91208, + "Ġplate": 11968, + "Ġplateau": 64288, + "Ġplated": 87133, + "Ġplates": 24477, + "Ġplatform": 5339, + "Ġplatforms": 15409, + "Ġplatinum": 62227, + "Ġplausible": 49334, + "Ġplay": 1486, + "Ġplaya": 97440, + "Ġplayable": 51035, + "Ġplayback": 31959, + "Ġplaybook": 86053, + "Ġplayed": 6342, + "Ġplayer": 2781, + "ĠplayerId": 75961, + "ĠplayerName": 83547, + "Ġplayers": 4217, + "Ġplayful": 56069, + "Ġplayground": 41615, + "Ġplaying": 5619, + "Ġplaylist": 26791, + "Ġplaylists": 68808, + "Ġplayoff": 19380, + "Ġplayoffs": 30755, + "Ġplays": 11088, + "Ġplaywright": 98368, + "Ġplaza": 71998, + "Ġplc": 96462, + "Ġple": 7100, + "Ġplea": 38967, + "Ġplead": 72193, + "Ġpleaded": 38901, + "Ġpleading": 72973, + "Ġpleas": 87700, + "Ġpleasant": 23795, + "Ġpleasantly": 81274, + "Ġplease": 4486, + "Ġpleased": 18442, + "Ġpleasing": 53699, + "Ġpleasure": 16656, + "Ġpleasures": 71874, + "Ġpled": 29786, + "Ġpledge": 35079, + "Ġpledged": 42247, + "Ġpledges": 78018, + "Ġplein": 83771, + "Ġplentiful": 80726, + "Ġplenty": 11260, + "Ġplethora": 67606, + "Ġplight": 69975, + "Ġplist": 68635, + "Ġpll": 83080, + "Ġplot": 7089, + "Ġplots": 30694, + "Ġplotted": 67583, + "Ġplotting": 43902, + "Ġpls": 86605, + "Ġplt": 6516, + "Ġplug": 19633, + "Ġplugged": 58229, + "Ġplugin": 9006, + "Ġplugins": 17215, + "Ġplugs": 62534, + "Ġplum": 41172, + "Ġplumber": 88586, + "Ġplumbing": 43188, + "Ġplummet": 70178, + "Ġplunder": 87702, + "Ġplung": 52772, + "Ġplunge": 66578, + "Ġplunged": 74703, + "Ġplur": 59117, + "Ġplural": 38498, + "Ġplurality": 72305, + "Ġplus": 5519, + "Ġplush": 71679, + "Ġplusieurs": 50754, + "Ġplut": 59487, + "Ġplutôt": 95208, + "Ġply": 38573, + "Ġplywood": 87898, + "Ġpm": 8836, + "Ġpn": 43050, + "Ġpname": 64946, + "Ġpne": 43526, + "Ġpneum": 56123, + "Ġpneumonia": 68229, + "Ġpng": 36151, + "Ġpnl": 72693, + "Ġpo": 3193, + "Ġpoate": 78555, + "Ġpob": 76722, + "Ġpobl": 73575, + "Ġpoblación": 139202, + "Ġpobli": 142092, + "Ġpobliżu": 142093, + "Ġpoc": 56303, + "Ġpocket": 17822, + "Ġpockets": 29178, + "Ġpoco": 38794, + "ĠpoczÄħ": 135407, + "ĠpoczÄħtk": 135408, + "ĠpoczÄħtku": 135409, + "Ġpod": 7509, + "Ġpodcast": 17711, + "Ġpodcasts": 54246, + "Ġpode": 28194, + "Ġpodem": 62676, + "Ġpodemos": 62504, + "Ġpoder": 28538, + "Ġpoderá": 136866, + "Ġpodium": 58036, + "Ġpodr": 41915, + "Ġpodrá": 132608, + "Ġpodróż": 135038, + "ĠpodrÃŃa": 82442, + "Ġpods": 54587, + "ĠpodÃŃa": 140319, + "Ġpoem": 32794, + "Ġpoems": 44219, + "Ġpoet": 39260, + "Ġpoetic": 75434, + "Ġpoetry": 31249, + "Ġpoets": 72972, + "Ġpog": 80269, + "Ġpoi": 46927, + "Ġpoids": 96318, + "Ġpoignant": 92347, + "Ġpoil": 73511, + "Ġpoint": 1459, + "Ġpointed": 14283, + "Ġpointer": 7445, + "Ġpointers": 27454, + "Ġpointing": 21633, + "Ġpointless": 59691, + "Ġpoints": 3501, + "Ġpois": 57783, + "Ġpoised": 53846, + "Ġpoison": 20476, + "Ġpoisoned": 70498, + "Ġpoisoning": 51136, + "Ġpoisonous": 73408, + "Ġpoj": 87030, + "Ġpok": 44195, + "Ġpoke": 51551, + "Ġpokemon": 37651, + "Ġpoker": 20814, + "Ġpoking": 91363, + "Ġpol": 1471, + "Ġpolar": 24660, + "Ġpolarity": 75690, + "Ġpolarization": 82145, + "Ġpole": 25026, + "Ġpoles": 50779, + "Ġpolic": 8540, + "Ġpolice": 4282, + "Ġpoliceman": 78580, + "Ġpolicemen": 85119, + "Ġpolicies": 10186, + "Ġpolicing": 54571, + "Ġpolicy": 4842, + "Ġpolicym": 61418, + "Ġpolicymakers": 69878, + "Ġpolish": 44029, + "Ġpolished": 43361, + "Ġpolishing": 84463, + "Ġpolit": 3354, + "Ġpolite": 47787, + "Ġpolitely": 80767, + "Ġpolitic": 75937, + "Ġpolitical": 4948, + "Ġpolitically": 30105, + "Ġpolitician": 35938, + "Ġpoliticians": 18761, + "Ġpolitics": 11500, + "Ġpolitique": 77936, + "Ġpoll": 7085, + "Ġpolled": 83621, + "Ġpollen": 78684, + "Ġpolling": 30644, + "Ġpolls": 23056, + "Ġpollut": 70034, + "Ġpollutants": 82561, + "Ġpolluted": 94769, + "Ġpollution": 24761, + "Ġpolo": 77785, + "Ġpoly": 9861, + "Ġpolyester": 55070, + "Ġpolygon": 29372, + "Ġpolygons": 68159, + "Ġpolyline": 92241, + "Ġpolym": 44933, + "Ġpolymer": 46293, + "Ġpolynomial": 47311, + "Ġpolys": 98819, + "ĠpolÃŃt": 38844, + "ĠpolÃŃtica": 61489, + "ĠpolÃŃtico": 132168, + "Ġpom": 29484, + "Ġpomi": 141263, + "ĠpomiÄĻd": 141264, + "ĠpomiÄĻdzy": 141265, + "Ġpomoc": 94711, + "ĠpomocÄħ": 137292, + "Ġpomp": 72770, + "Ġpomys": 135669, + "ĠpomysÅĤ": 135670, + "Ġpon": 33923, + "Ġpond": 35570, + "Ġponder": 47783, + "Ġponds": 88355, + "Ġponer": 72723, + "Ġpong": 77622, + "Ġponiew": 134308, + "Ġponieważ": 134309, + "Ġpont": 36094, + "Ġponto": 74188, + "Ġpontos": 76115, + "Ġpony": 52636, + "Ġpoo": 82710, + "Ġpool": 7314, + "Ġpooled": 75376, + "Ġpooling": 74410, + "Ġpools": 33139, + "Ġpoop": 75816, + "Ġpoor": 7852, + "Ġpoorer": 65181, + "Ġpoorest": 67651, + "Ġpoorly": 30455, + "Ġpop": 2420, + "Ġpopcorn": 69705, + "Ġpope": 63272, + "Ġpopover": 89379, + "Ġpopped": 42233, + "Ġpopping": 49849, + "Ġpops": 47600, + "Ġpopul": 31749, + "Ġpopulace": 93820, + "Ġpopular": 5411, + "Ġpopularity": 22538, + "Ġpopulate": 30446, + "Ġpopulated": 34359, + "Ġpopulation": 7042, + "Ġpopulations": 21910, + "Ġpopulação": 138489, + "Ġpopulist": 68968, + "Ġpopulous": 94451, + "Ġpopup": 21960, + "Ġpor": 4154, + "Ġporcelain": 77642, + "Ġporch": 44647, + "Ġpore": 96451, + "Ġpores": 70928, + "Ġpork": 35067, + "Ġporn": 3480, + "Ġporno": 5850, + "Ġpornofil": 70270, + "Ġpornofilm": 89234, + "Ġpornografia": 95784, + "Ġpornography": 43787, + "Ġpornos": 39383, + "Ġpornost": 62423, + "Ġpornstar": 72183, + "Ġpornô": 47357, + "Ġpornôs": 90745, + "Ġporous": 93661, + "Ġporque": 27795, + "Ġporr": 48352, + "Ġporrf": 88817, + "Ġport": 2635, + "Ġporta": 60990, + "Ġportable": 22819, + "Ġportal": 23132, + "Ġportals": 75463, + "Ġporte": 59072, + "Ġporter": 87186, + "Ġportfolio": 19565, + "Ġportfolios": 75708, + "Ġportion": 13348, + "Ġportions": 19325, + "Ġportrait": 33033, + "Ġportraits": 58134, + "Ġportray": 24878, + "Ġportrayal": 73933, + "Ġportrayed": 44777, + "Ġportraying": 98497, + "Ġportrays": 96506, + "Ġports": 20325, + "Ġporém": 139567, + "Ġporówn": 142467, + "Ġporówna": 142468, + "Ġpos": 1133, + "ĠposX": 66690, + "ĠposY": 57797, + "Ġpose": 17040, + "Ġposed": 36160, + "Ġposes": 33003, + "Ġposi": 80893, + "Ġposible": 55032, + "Ġposicion": 63653, + "Ġposición": 90231, + "Ġposing": 51904, + "Ġposit": 20321, + "Ġposition": 2309, + "Ġpositional": 67547, + "Ġpositioned": 34228, + "Ġpositioning": 38725, + "Ġpositions": 9892, + "Ġpositive": 6785, + "Ġpositively": 39546, + "Ġpositives": 63656, + "Ġpositivity": 96942, + "Ġposix": 98343, + "Ġposição": 98194, + "Ġposs": 2229, + "Ġpossess": 15218, + "Ġpossessed": 42790, + "Ġpossesses": 49226, + "Ġpossessing": 59562, + "Ġpossession": 18721, + "Ġpossessions": 52535, + "Ġpossibile": 79946, + "Ġpossibilit": 66139, + "Ġpossibilities": 23607, + "Ġpossibility": 13038, + "Ġpossibilité": 137759, + "ĠpossibilitÃł": 133231, + "Ġpossible": 3204, + "Ġpossibly": 10767, + "Ġpossono": 81018, + "Ġpossui": 87274, + "ĠpossÃŃvel": 63181, + "Ġpost": 1736, + "ĠpostData": 54183, + "ĠpostId": 66030, + "Ġpostage": 77041, + "Ġpostal": 39754, + "Ġpostcode": 72836, + "Ġposte": 91046, + "Ġposted": 8454, + "Ġposter": 22361, + "Ġposterior": 44900, + "Ġposters": 38468, + "Ġpostfix": 68122, + "Ġpostgres": 59826, + "Ġposting": 17004, + "Ġpostings": 74481, + "Ġposto": 87599, + "Ġpostpon": 50001, + "Ġpostpone": 94658, + "Ġpostponed": 61620, + "Ġposts": 7999, + "Ġpostseason": 55371, + "Ġposture": 47278, + "Ġpot": 3338, + "Ġpotassium": 61175, + "Ġpotato": 38734, + "Ġpotatoes": 34167, + "Ġpotency": 76129, + "Ġpotent": 35775, + "Ġpotential": 4650, + "Ġpotentially": 13581, + "Ġpotentials": 94258, + "Ġpotion": 60108, + "Ġpotions": 78583, + "Ġpotrze": 88350, + "ĠpotrÃł": 142008, + "Ġpots": 50031, + "Ġpottery": 82663, + "Ġpou": 16836, + "Ġpouch": 55852, + "Ġpouco": 82837, + "Ġpoultry": 65528, + "Ġpound": 30023, + "Ġpounded": 94479, + "Ġpounding": 69527, + "Ġpounds": 16302, + "Ġpour": 4914, + "Ġpoured": 42933, + "Ġpouring": 50413, + "Ġpourquoi": 86220, + "Ġpourrait": 91115, + "Ġpours": 66707, + "Ġpouvez": 44266, + "Ġpouvoir": 67126, + "Ġpov": 55214, + "Ġpoverty": 19005, + "Ġpow": 6875, + "Ġpowder": 16723, + "Ġpowdered": 82483, + "Ġpower": 2355, + "Ġpowered": 22333, + "Ġpowerful": 7988, + "Ġpowerhouse": 74114, + "Ġpowering": 89900, + "Ġpowerless": 84352, + "Ġpowers": 13430, + "Ġpowsta": 140098, + "ĠpowstaÅĤ": 140099, + "Ġpoz": 40721, + "Ġpozosta": 139666, + "ĠpozostaÅĤ": 139667, + "Ġpożycz": 137108, + "Ġpożyczk": 137109, + "Ġpp": 11823, + "Ġppl": 88428, + "Ġppm": 63597, + "Ġpprint": 46446, + "Ġppt": 77484, + "Ġpq": 39639, + "Ġpr": 548, + "Ġpra": 44003, + "Ġprac": 60886, + "Ġpracownik": 139872, + "Ġpracowników": 139873, + "Ġpract": 3771, + "Ġpractical": 14976, + "Ġpractically": 31267, + "Ġpractice": 6588, + "Ġpracticed": 43564, + "Ġpractices": 12378, + "Ġpracticing": 35566, + "Ġpractise": 87588, + "Ġpractition": 30171, + "Ġpractitioner": 54372, + "Ġpractitioners": 42095, + "Ġpracy": 81209, + "ĠpracÄĻ": 138263, + "Ġprag": 51351, + "Ġpragma": 62120, + "Ġpragmatic": 74236, + "Ġpraise": 28388, + "Ġpraised": 36375, + "Ġpraises": 86078, + "Ġpraising": 80578, + "Ġprakt": 73455, + "Ġprank": 75273, + "Ġprat": 54644, + "Ġpratic": 97039, + "Ġpratique": 94134, + "Ġpraw": 53472, + "Ġpray": 23803, + "Ġprayed": 66697, + "Ġprayer": 22936, + "Ġprayers": 33196, + "Ġpraying": 51132, + "Ġpre": 855, + "Ġpreach": 70621, + "Ġpreached": 88094, + "Ġpreacher": 88753, + "Ġpreaching": 68323, + "Ġpreamble": 89454, + "Ġprec": 5833, + "Ġpreca": 40507, + "Ġprecarious": 92122, + "Ġprecaution": 71047, + "Ġprecautions": 59903, + "Ġpreced": 16201, + "Ġpreceded": 52480, + "Ġprecedence": 53056, + "Ġprecedent": 46791, + "Ġpreceding": 37746, + "Ġprecinct": 67899, + "Ġprecio": 36520, + "Ġprecios": 83781, + "Ġprecious": 26400, + "Ġprecip": 35741, + "Ġprecipitation": 59950, + "Ġprecis": 20234, + "Ġprecisa": 72766, + "Ġprecise": 23560, + "Ġprecisely": 23638, + "Ġprecision": 16052, + "Ġpreco": 83742, + "Ġprecondition": 79438, + "Ġprecursor": 70161, + "Ġpred": 4162, + "Ġpredator": 66906, + "Ġpredators": 55117, + "Ġpredatory": 87070, + "Ġpredecess": 30082, + "Ġpredecessor": 39174, + "Ġpredecessors": 61440, + "Ġpredefined": 63062, + "Ġpredetermined": 85910, + "Ġpredic": 78334, + "Ġpredicate": 24283, + "Ġpredicates": 80027, + "Ġpredict": 7023, + "Ġpredictable": 44107, + "Ġpredicted": 19149, + "Ġpredicting": 51897, + "Ġprediction": 19639, + "Ġpredictions": 19898, + "Ġpredictive": 59236, + "Ġpredictor": 61154, + "Ġpredictors": 94122, + "Ġpredicts": 55878, + "Ġpredis": 79532, + "Ġpredomin": 33964, + "Ġpredominant": 95431, + "Ġpredominantly": 46804, + "Ġpreds": 60721, + "Ġpreempt": 81747, + "Ġpref": 18735, + "Ġprefab": 75314, + "Ġprefect": 89445, + "Ġprefer": 10702, + "Ġpreferable": 69568, + "Ġpreferably": 51654, + "Ġpreference": 21933, + "Ġpreferences": 19322, + "Ġpreferred": 14890, + "ĠpreferredStyle": 63437, + "Ġpreferring": 97617, + "Ġprefers": 53964, + "Ġprefetch": 90477, + "Ġprefix": 9252, + "Ġprefixed": 81495, + "Ġprefixes": 62576, + "Ġprefs": 36013, + "Ġpreg": 19978, + "Ġpregn": 11925, + "Ġpregnancies": 81208, + "Ġpregnancy": 19636, + "Ġpregnant": 20280, + "Ġpregunta": 79340, + "Ġprejud": 33820, + "Ġprejudice": 47686, + "Ġprejudices": 98776, + "Ġprelim": 29613, + "Ġpreliminary": 32169, + "Ġpreload": 60457, + "Ġprem": 6811, + "Ġpremature": 41127, + "Ġprematurely": 87919, + "Ġpremier": 20496, + "Ġpremiere": 34852, + "Ġpremiered": 84070, + "Ġpremise": 40202, + "Ġpremises": 33922, + "Ġpremium": 14848, + "Ġpremiums": 47507, + "Ġpremière": 54033, + "Ġprenatal": 96028, + "Ġprend": 34356, + "Ġprendre": 58294, + "Ġpreocup": 87085, + "Ġpreorder": 71850, + "Ġprep": 21327, + "Ġprepaid": 82676, + "Ġprepar": 9872, + "Ġpreparation": 17975, + "Ġpreparations": 46879, + "Ġprepare": 10549, + "Ġprepared": 10030, + "ĠpreparedStatement": 55510, + "Ġprepares": 47442, + "Ġpreparing": 20045, + "Ġprepend": 73647, + "Ġpreprocess": 53465, + "Ġpreprocessing": 63631, + "Ġprer": 98861, + "Ġprere": 43901, + "Ġprerequisite": 79784, + "Ġprerequisites": 85258, + "Ġpres": 1652, + "Ġpreschool": 60805, + "Ġprescribe": 71233, + "Ġprescribed": 30931, + "Ġprescribing": 88344, + "Ġprescription": 22091, + "Ġprescriptions": 64735, + "Ġpreseason": 47574, + "Ġpresence": 9362, + "Ġpresent": 3042, + "ĠpresentViewController": 82448, + "Ġpresenta": 96080, + "Ġpresentation": 15496, + "Ġpresentations": 37380, + "Ġpresente": 53172, + "Ġpresented": 10449, + "Ġpresenter": 33656, + "Ġpresenting": 31544, + "Ġpresently": 49701, + "Ġpresents": 18404, + "Ġpresença": 140571, + "Ġpreservation": 45543, + "Ġpreserve": 21129, + "Ġpreserved": 33583, + "Ġpreserves": 74898, + "Ġpreserving": 46895, + "Ġpreset": 42921, + "Ġpresets": 62196, + "Ġpresidency": 31758, + "Ġpresident": 4767, + "Ġpresidente": 51385, + "Ġpresidential": 13318, + "Ġpresidents": 46508, + "Ġpress": 3493, + "Ġpressed": 17320, + "Ġpresses": 52963, + "Ġpressing": 25352, + "Ġpressure": 7262, + "Ġpressured": 74351, + "Ġpressures": 39750, + "Ġprest": 22298, + "Ġprestige": 65121, + "Ġprestigious": 40285, + "Ġpresum": 22277, + "Ġpresumably": 35448, + "Ġpresume": 77395, + "Ġpresumed": 61257, + "Ġpresumption": 86845, + "Ġpresup": 70292, + "Ġpret": 4410, + "Ġpretend": 34740, + "Ġpretended": 91740, + "Ġpretending": 54806, + "Ġpretext": 93814, + "Ġpretrained": 80669, + "Ġpretty": 5020, + "Ġprev": 7872, + "ĠprevState": 74946, + "Ġprevail": 65728, + "Ġprevailed": 84714, + "Ġprevailing": 60029, + "Ġpreval": 25344, + "Ġprevalence": 36909, + "Ġprevalent": 45841, + "Ġprevent": 5358, + "Ġpreventative": 96320, + "Ġprevented": 30998, + "Ġpreventing": 26160, + "Ġprevention": 26248, + "Ġpreventive": 70023, + "Ġprevents": 27934, + "Ġpreview": 17124, + "Ġpreviews": 56149, + "Ġprevious": 3681, + "Ġpreviously": 8597, + "Ġprey": 36593, + "Ġprez": 79942, + "Ġpreço": 97230, + "Ġpri": 12493, + "Ġprice": 3349, + "Ġpriced": 32605, + "Ġpriceless": 90944, + "Ġprices": 7576, + "Ġpricey": 78362, + "Ġpricing": 21219, + "Ġprick": 88700, + "Ġpride": 21770, + "Ġpriest": 27086, + "Ġpriesthood": 85616, + "Ġpriests": 41863, + "Ġprim": 8860, + "Ġprima": 35313, + "Ġprimal": 86181, + "Ġprimaries": 65412, + "Ġprimarily": 15503, + "Ġprimary": 6028, + "ĠprimaryKey": 65858, + "ĠprimaryStage": 89750, + "Ġprime": 10250, + "Ġprimeira": 67731, + "Ġprimeiro": 56081, + "Ġprimer": 26811, + "Ġprimera": 43669, + "Ġprimero": 70136, + "Ġprimes": 49433, + "Ġprimitive": 27594, + "Ġprimitives": 71194, + "Ġprimo": 61232, + "Ġprin": 63027, + "Ġprinc": 81560, + "Ġprince": 41726, + "Ġprinces": 83193, + "Ġprincess": 39588, + "Ġprincip": 7394, + "Ġprincipal": 12435, + "ĠprincipalColumn": 58801, + "ĠprincipalTable": 58651, + "Ġprincipales": 81412, + "Ġprincipally": 91281, + "Ġprincipalmente": 94337, + "Ġprincipals": 81039, + "Ġprincipio": 92053, + "Ġprinciple": 17508, + "Ġprinciples": 16170, + "ĠprincÃŃpio": 143160, + "Ġprint": 1173, + "Ġprintable": 41995, + "Ġprinted": 16709, + "Ġprinter": 22380, + "Ġprinters": 55953, + "Ġprintf": 4100, + "Ġprinting": 18484, + "Ġprintk": 58323, + "Ġprintln": 13751, + "Ġprints": 23473, + "Ġprio": 92908, + "Ġprior": 4867, + "Ġpriorit": 44677, + "Ġpriorities": 29501, + "Ġprioritize": 62552, + "Ġpriority": 10619, + "Ġpris": 51479, + "Ġprise": 84324, + "Ġprism": 93610, + "Ġprison": 9343, + "Ġprisoner": 41850, + "Ġprisoners": 27141, + "Ġprisons": 48329, + "Ġpristine": 64985, + "Ġpriv": 6095, + "Ġprivacy": 12345, + "Ġprivat": 37521, + "Ġprivate": 869, + "ĠprivateKey": 70565, + "Ġprivately": 37071, + "Ġprive": 33373, + "Ġprivile": 15531, + "Ġprivilege": 25551, + "Ġprivileged": 46261, + "Ġprivileges": 35732, + "Ġprivé": 134283, + "Ġprix": 37747, + "Ġprize": 21882, + "Ġprized": 86530, + "Ġprizes": 37408, + "Ġprm": 82719, + "Ġpro": 462, + "Ġproactive": 58214, + "Ġprob": 3566, + "Ġprobabil": 84093, + "Ġprobabilities": 48216, + "Ġprobability": 18927, + "Ġprobable": 34877, + "Ġprobably": 4658, + "Ġprobation": 45927, + "Ġprobe": 21730, + "Ġprobes": 62510, + "Ġprobing": 82972, + "Ġprobl": 50988, + "Ġproble": 2499, + "Ġproblem": 3491, + "Ġproblema": 41076, + "Ġproblemas": 53769, + "Ġproblematic": 34933, + "Ġproblems": 5322, + "Ġproblème": 79159, + "Ġprobs": 69370, + "Ġproc": 13674, + "Ġproced": 7103, + "Ġprocedural": 61681, + "Ġprocedure": 10324, + "Ġprocedures": 15966, + "Ġproceed": 10354, + "Ġproceeded": 44274, + "Ġproceeding": 38447, + "Ġproceedings": 28307, + "Ġproceeds": 33455, + "Ġproces": 28294, + "Ġproceso": 49009, + "Ġprocess": 1882, + "ĠprocessData": 94372, + "Ġprocessed": 15233, + "Ġprocesses": 11364, + "Ġprocessing": 8692, + "Ġprocession": 83714, + "Ġprocesso": 58896, + "Ġprocessor": 17654, + "Ġprocessors": 36586, + "Ġproclaim": 64984, + "Ġproclaimed": 61911, + "Ġproclamation": 95430, + "Ġprocrast": 96444, + "Ġprocur": 70502, + "Ġprocure": 79679, + "Ġprocurement": 52578, + "Ġprod": 14475, + "Ġprodu": 2497, + "Ġproducción": 91036, + "Ġproduce": 8193, + "Ġproduced": 8947, + "Ġproducer": 16851, + "Ġproducers": 23308, + "Ġproduces": 18644, + "Ġproducing": 17387, + "Ġproduct": 1985, + "ĠproductId": 42848, + "ĠproductList": 97592, + "ĠproductName": 65587, + "ĠproductService": 98995, + "Ġproduction": 5670, + "Ġproductions": 53148, + "Ġproductive": 26236, + "Ġproductivity": 25148, + "Ġproducto": 34774, + "Ġproductos": 41325, + "Ġproducts": 3871, + "Ġproduit": 53106, + "Ġproduits": 52921, + "Ġproduk": 60874, + "Ġprodukt": 72295, + "Ġproduktów": 137368, + "Ġproduto": 43670, + "Ġprodutos": 61862, + "Ġprodução": 134791, + "Ġprof": 2778, + "Ġprofes": 34708, + "Ġprofesional": 75248, + "Ġprofess": 34478, + "Ġprofession": 4808, + "Ġprofessional": 6584, + "Ġprofessionalism": 65329, + "Ġprofessionally": 44929, + "Ġprofessionals": 15387, + "Ġprofessionnel": 84405, + "Ġprofessions": 68692, + "Ġprofessor": 14227, + "Ġprofessors": 44624, + "Ġproficiency": 62139, + "Ġproficient": 68265, + "Ġprofil": 52354, + "Ġprofile": 5526, + "Ġprofiler": 87497, + "Ġprofiles": 20872, + "Ġprofiling": 55086, + "Ġprofit": 11372, + "Ġprofitability": 62236, + "Ġprofitable": 33135, + "Ġprofits": 21856, + "Ġprofound": 27155, + "Ġprofoundly": 65178, + "Ġprofund": 97570, + "Ġprog": 29271, + "Ġprogen": 83260, + "Ġprogn": 62803, + "Ġprognosis": 94250, + "Ġprogram": 2025, + "Ġprograma": 35760, + "Ġprogramas": 93839, + "Ġprogramm": 55486, + "Ġprogramma": 97139, + "Ġprogramme": 18136, + "Ġprogrammed": 55068, + "Ġprogrammer": 47788, + "Ġprogrammers": 54846, + "Ġprogrammes": 37637, + "Ġprogramming": 15473, + "Ġprograms": 7468, + "Ġprogress": 5098, + "ĠprogressBar": 47076, + "ĠprogressDialog": 57438, + "Ġprogressed": 61816, + "Ġprogresses": 67611, + "Ġprogressing": 75622, + "Ġprogression": 32724, + "Ġprogressive": 22257, + "Ġprogressively": 71759, + "Ġprogressives": 81759, + "Ġprohib": 14207, + "Ġprohibit": 47386, + "Ġprohibited": 25921, + "Ġprohibiting": 78366, + "Ġprohibition": 49867, + "Ġprohibits": 58438, + "Ġproj": 21333, + "Ġproject": 2390, + "ĠprojectId": 41696, + "ĠprojectName": 72516, + "Ġprojected": 27348, + "Ġprojectile": 37957, + "Ġprojectiles": 79037, + "Ġprojecting": 81282, + "Ġprojection": 21615, + "Ġprojections": 40479, + "Ġprojector": 67985, + "Ġprojects": 7079, + "Ġprojekt": 73894, + "Ġprojet": 45394, + "Ġprojeto": 68859, + "Ġprol": 28420, + "Ġprolet": 63117, + "Ġproletariat": 87247, + "Ġprolifer": 41936, + "Ġproliferation": 52740, + "Ġprolific": 62000, + "Ġprolong": 32382, + "Ġprolonged": 43287, + "Ġprom": 2706, + "Ġpromin": 44802, + "Ġprominence": 70099, + "Ġprominent": 20469, + "Ġprominently": 72988, + "Ġpromise": 11222, + "Ġpromised": 18951, + "Ġpromises": 20647, + "Ġpromising": 25383, + "Ġpromo": 26871, + "Ġpromot": 18408, + "Ġpromote": 11926, + "Ġpromoted": 28926, + "Ġpromoter": 65542, + "Ġpromoters": 94160, + "Ġpromotes": 38890, + "Ġpromoting": 22136, + "Ġpromotion": 20249, + "Ġpromotional": 34282, + "Ġpromotions": 35971, + "Ġprompt": 9934, + "Ġprompted": 28646, + "Ġprompting": 49645, + "Ġpromptly": 39422, + "Ġprompts": 50932, + "Ġpron": 18613, + "Ġprone": 36997, + "Ġpronounce": 87982, + "Ġpronounced": 37517, + "Ġpronto": 89698, + "Ġpronunciation": 70622, + "Ġproof": 11064, + "Ġproofs": 77159, + "Ġprop": 2004, + "ĠpropName": 90114, + "ĠpropTypes": 55516, + "Ġpropag": 16992, + "Ġpropaganda": 29517, + "Ġpropagate": 57414, + "Ġpropagated": 85050, + "Ġpropagation": 53643, + "Ġpropane": 90938, + "Ġpropel": 88383, + "Ġpropelled": 91685, + "Ġpropensity": 94223, + "Ġproper": 6169, + "Ġproperly": 10277, + "Ġproperties": 5888, + "Ġproperty": 3343, + "ĠpropertyName": 44276, + "Ġproph": 39234, + "Ġprophecy": 65784, + "Ġprophet": 52504, + "Ġprophets": 71668, + "Ġpropia": 97329, + "Ġpropiedad": 87623, + "Ġpropio": 91092, + "Ġproponents": 70696, + "Ġpropor": 70908, + "Ġproporcion": 79843, + "Ġproport": 32703, + "Ġproportion": 21117, + "Ġproportional": 54172, + "Ġproportions": 48792, + "Ġpropos": 9844, + "Ġproposal": 13734, + "Ġproposals": 24258, + "Ġpropose": 29614, + "Ġproposed": 10981, + "Ġproposes": 54625, + "Ġproposing": 56415, + "Ġproposition": 40080, + "Ġpropositions": 83273, + "Ġproposé": 139280, + "Ġpropre": 82087, + "Ġpropri": 21065, + "Ġpropriet": 53173, + "Ġproprietary": 33233, + "Ġproprietor": 91923, + "ĠproprietÃł": 141375, + "Ġproprio": 65910, + "Ġproprié": 143447, + "Ġpropriétaire": 143448, + "Ġpropriété": 142779, + "Ġprops": 6914, + "Ġpropulsion": 90128, + "Ġpropósito": 142043, + "Ġpros": 8709, + "Ġprose": 60701, + "Ġprosec": 12297, + "Ġprosecute": 65843, + "Ġprosecuted": 57459, + "Ġprosecuting": 97174, + "Ġprosecution": 31599, + "Ġprosecutions": 97411, + "Ġprosecutor": 35051, + "Ġprosecutors": 30820, + "Ġprospect": 21479, + "Ġprospective": 32447, + "Ġprospects": 26851, + "Ġprosper": 28661, + "Ġprosperity": 43102, + "Ġprosperous": 70307, + "Ġprost": 35221, + "Ġprostate": 46347, + "Ġprostit": 11197, + "Ġprostitu": 38622, + "Ġprostituer": 39924, + "Ġprostituerade": 42723, + "Ġprostituerte": 50617, + "Ġprostitut": 56180, + "Ġprostitutas": 20580, + "Ġprostitute": 65541, + "Ġprostitutes": 62582, + "Ġprostitution": 49750, + "Ġprot": 1724, + "Ġprotagon": 32986, + "Ġprotagonist": 45584, + "Ġprotagonists": 94635, + "Ġprote": 5425, + "Ġprotección": 142502, + "Ġprotect": 6016, + "Ġprotected": 2617, + "Ġprotecting": 22184, + "Ġprotection": 9135, + "Ġprotections": 36324, + "Ġprotective": 28119, + "Ġprotector": 62161, + "Ġprotects": 35136, + "Ġprotein": 12833, + "Ġproteins": 27796, + "Ġprotest": 8665, + "Ġprotested": 69405, + "Ġprotester": 93334, + "Ġprotesters": 25743, + "Ġprotesting": 58210, + "Ġprotestors": 88356, + "Ġprotests": 21907, + "Ġproto": 18433, + "Ġprotobuf": 69634, + "Ġprotocol": 11507, + "Ġprotocols": 31785, + "Ġproton": 81486, + "Ġprototype": 24052, + "Ġprototypes": 46628, + "Ġprotr": 80358, + "Ġproud": 12409, + "Ġproudly": 44809, + "Ġprov": 2543, + "Ġprova": 96893, + "Ġprove": 12118, + "Ġproved": 18651, + "Ġproven": 16622, + "Ġproverb": 78438, + "Ġproves": 35762, + "Ġprovid": 2059, + "Ġprovide": 3410, + "Ġprovided": 3897, + "ĠprovidedIn": 44010, + "Ġprovider": 9109, + "Ġproviders": 12565, + "Ġprovides": 5707, + "Ġproviding": 8241, + "Ġprovinc": 27807, + "Ġprovince": 16847, + "Ġprovinces": 39921, + "Ġprovincia": 78826, + "Ġprovincial": 34931, + "Ġproving": 38758, + "Ġprovision": 17136, + "Ġprovisional": 76990, + "Ġprovisioning": 97056, + "Ġprovisions": 19156, + "Ġprovoc": 41402, + "Ġprovocative": 65898, + "Ġprovoke": 71993, + "Ġprovoked": 75466, + "Ġprow": 47558, + "Ġprowess": 73060, + "Ġprox": 21542, + "Ġproxies": 59365, + "Ġproximity": 36743, + "Ġproximité": 141910, + "Ġproxy": 13291, + "Ġproyecto": 49851, + "Ġproyectos": 88743, + "Ġprt": 95684, + "Ġprudent": 73878, + "Ġprueba": 74990, + "Ġprune": 79526, + "Ġpruning": 85192, + "Ġprv": 84888, + "Ġpry": 81279, + "Ġprz": 21254, + "Ġprze": 30329, + "Ġprzec": 141118, + "Ġprzecież": 141119, + "Ġprzed": 49172, + "Ġprzegl": 141719, + "ĠprzeglÄħd": 141720, + "Ġprzez": 37624, + "Ġprzy": 30848, + "ĠprzykÅĤad": 133644, + "Ġprzypad": 97217, + "Ġprá": 63167, + "Ġpráctica": 134408, + "Ġprática": 134801, + "Ġprès": 83264, + "Ġpré": 26293, + "Ġpréc": 50525, + "Ġprécis": 132971, + "Ġprécéd": 140057, + "Ġprécédent": 140058, + "Ġprés": 29079, + "Ġprésence": 140201, + "Ġprésent": 56385, + "Ġprésente": 132047, + "Ġprésenté": 142133, + "Ġprésident": 88702, + "Ġprést": 143038, + "Ġpréstamo": 143039, + "Ġprévu": 140234, + "Ġprêt": 134827, + "Ġpró": 29291, + "Ġprób": 134276, + "Ġpróp": 71321, + "Ġprópria": 135321, + "Ġpróprio": 134453, + "Ġpróxima": 90207, + "Ġpróximo": 70798, + "Ġps": 4726, + "Ġpseud": 50643, + "Ġpseudo": 34750, + "Ġpsi": 38026, + "Ġpst": 43802, + "Ġpstmt": 49575, + "Ġpsy": 25142, + "Ġpsych": 8671, + "Ġpsyche": 87809, + "Ġpsyched": 60733, + "Ġpsychedelic": 80982, + "Ġpsychiat": 45776, + "Ġpsychiatric": 46557, + "Ġpsychiatrist": 72869, + "Ġpsychic": 49117, + "Ġpsycho": 41011, + "Ġpsychological": 23187, + "Ġpsychologically": 95555, + "Ġpsychologist": 53495, + "Ġpsychologists": 73091, + "Ġpsychology": 29918, + "Ġpsychosis": 97047, + "Ġpsychotic": 93141, + "Ġpsycopg": 86061, + "Ġpsz": 45083, + "Ġpt": 10817, + "Ġpthread": 19937, + "Ġptr": 10087, + "Ġpts": 29993, + "Ġpu": 18256, + "Ġpub": 6675, + "Ġpubb": 60149, + "Ġpuberty": 86298, + "Ġpubkey": 95116, + "Ġpubli": 77411, + "Ġpublic": 584, + "ĠpublicKey": 69347, + "Ġpublication": 16599, + "Ġpublications": 27985, + "Ġpublicity": 42663, + "Ġpublicly": 17424, + "Ġpublish": 3415, + "Ġpublished": 4652, + "Ġpublisher": 21262, + "Ġpublishers": 35617, + "Ġpublishes": 64485, + "Ġpublishing": 22913, + "Ġpublié": 139341, + "Ġpubs": 73762, + "Ġpuck": 48635, + "Ġpud": 84295, + "Ġpudd": 54582, + "Ġpudding": 81427, + "Ġpudo": 74725, + "Ġpueblo": 92441, + "Ġpued": 13419, + "Ġpueda": 89304, + "Ġpuede": 20673, + "Ġpueden": 40504, + "Ġpuedes": 58945, + "Ġpuedo": 80815, + "Ġpuerto": 89971, + "Ġpues": 68510, + "Ġpuesto": 84709, + "Ġpuff": 62072, + "Ġpuis": 43729, + "Ġpul": 7737, + "Ġpull": 6815, + "Ġpulled": 13238, + "Ġpulling": 22266, + "Ġpulls": 33045, + "Ġpulmonary": 69424, + "Ġpulp": 63088, + "Ġpuls": 54937, + "Ġpulse": 27235, + "Ġpulses": 65457, + "Ġpulumi": 88556, + "Ġpulver": 70155, + "Ġpump": 13834, + "Ġpumped": 61354, + "Ġpumping": 52126, + "Ġpumpkin": 49525, + "Ġpumps": 42775, + "Ġpun": 8310, + "Ġpunch": 20380, + "Ġpunched": 60918, + "Ġpunches": 60751, + "Ġpunching": 67881, + "Ġpunct": 31355, + "Ġpunctuation": 61503, + "Ġpund": 61112, + "Ġpundits": 78206, + "Ġpunish": 31051, + "Ġpunishable": 80570, + "Ġpunished": 40898, + "Ġpunishing": 74886, + "Ġpunishment": 24093, + "Ġpunishments": 89825, + "Ġpunitive": 81432, + "Ġpunk": 35758, + "Ġpunt": 31849, + "Ġpunto": 41514, + "Ġpuntos": 60889, + "Ġpup": 14895, + "Ġpupil": 59972, + "Ġpupils": 44072, + "Ġpupper": 72629, + "Ġpuppet": 49448, + "Ġpuppies": 56625, + "Ġpuppy": 41189, + "Ġpups": 89721, + "Ġpur": 3999, + "Ġpurch": 4902, + "Ġpurchase": 7627, + "Ġpurchased": 14733, + "Ġpurchaser": 76250, + "Ġpurchasers": 93655, + "Ġpurchases": 23491, + "Ġpurchasing": 22577, + "Ġpure": 10526, + "Ġpurely": 31127, + "Ġpurge": 54100, + "Ġpurification": 93436, + "Ġpurified": 91500, + "Ġpurity": 52400, + "Ġpurple": 24932, + "Ġpurported": 58760, + "Ġpurpos": 83796, + "Ġpurpose": 7428, + "Ġpurposely": 89087, + "Ġpurposes": 9895, + "Ġpurs": 9836, + "Ġpurse": 52001, + "Ġpursuant": 32449, + "Ġpursue": 22729, + "Ġpursued": 45431, + "Ġpursuing": 33018, + "Ġpursuit": 32336, + "Ġpursuits": 89787, + "Ġpus": 74576, + "Ġpush": 4484, + "ĠpushViewController": 62059, + "Ġpushed": 15391, + "Ġpushes": 40702, + "Ġpushing": 17461, + "Ġpuss": 83700, + "Ġpussy": 16189, + "Ġput": 2182, + "ĠputStrLn": 78598, + "Ġputa": 75390, + "Ġputas": 41824, + "Ġputchar": 89246, + "Ġpute": 24290, + "Ġputs": 9521, + "Ġputt": 90334, + "Ġputting": 10687, + "Ġpuzz": 31754, + "Ġpuzzle": 24626, + "Ġpuzzled": 86320, + "Ġpuzzles": 46523, + "Ġpuò": 50879, + "Ġpv": 33491, + "Ġpw": 26581, + "Ġpwd": 32706, + "Ġpwm": 64551, + "Ġpx": 17146, + "Ġpy": 4510, + "Ġpygame": 14699, + "Ġpyl": 34285, + "Ġpylab": 78414, + "Ġpylint": 50371, + "Ġpym": 45760, + "Ġpymongo": 84776, + "Ġpymysql": 95373, + "Ġpyplot": 70254, + "Ġpyramid": 46797, + "Ġpys": 66089, + "Ġpyt": 94671, + "Ġpytest": 27764, + "Ġpython": 10135, + "Ġpz": 97102, + "Ġpá": 64745, + "Ġpágina": 46301, + "Ġpä": 92364, + "ĠpÃ¥": 9114, + "Ġpé": 62932, + "Ġpérd": 143906, + "Ġpérdida": 143907, + "Ġpéri": 83311, + "Ġpériode": 136868, + "Ġpër": 93160, + "Ġpó": 127397, + "Ġpóź": 135696, + "Ġpóźniej": 135697, + "Ġpúb": 35905, + "Ġpública": 87949, + "Ġpúblico": 58950, + "ĠpÅĤ": 94232, + "ĠpÅĤyn": 140916, + "ĠpÅĤyt": 136434, + "ĠpÅĻ": 58306, + "Ġq": 2804, + "ĠqDebug": 34853, + "Ġqa": 88496, + "Ġqb": 75807, + "Ġqc": 77066, + "Ġqed": 49678, + "Ġqemu": 85108, + "Ġqi": 74105, + "Ġqint": 84717, + "Ġql": 66384, + "Ġqos": 92973, + "Ġqp": 62369, + "Ġqq": 55539, + "Ġqr": 49290, + "Ġqreal": 75574, + "Ġqry": 68960, + "Ġqs": 32421, + "Ġqt": 38949, + "Ġqty": 43185, + "Ġqu": 922, + "Ġqua": 73525, + "Ġquad": 27082, + "Ġquadr": 29136, + "Ġquadrant": 93286, + "Ġquadratic": 79151, + "Ġquaint": 87508, + "Ġquake": 87370, + "Ġqual": 5841, + "Ġqualche": 89016, + "Ġquale": 83484, + "Ġquali": 94525, + "Ġqualidade": 85883, + "Ġqualification": 45060, + "Ġqualifications": 42684, + "Ġqualified": 14988, + "Ġqualifier": 60647, + "Ġqualifiers": 82318, + "Ġqualifies": 79907, + "Ġqualify": 25384, + "Ġqualifying": 36114, + "Ġqualitative": 61029, + "Ġqualities": 28500, + "Ġquality": 4271, + "Ġqualité": 59028, + "ĠqualitÃł": 131848, + "Ġqualquer": 57799, + "Ġquam": 89839, + "Ġquand": 41576, + "Ġquando": 29705, + "Ġquant": 10272, + "Ġquantidade": 63015, + "Ġquantify": 75398, + "Ġquantitative": 46516, + "Ġquantities": 32676, + "Ġquantity": 12194, + "Ġquanto": 50112, + "Ġquantum": 30128, + "Ġquar": 36865, + "Ġquarantine": 70389, + "Ġquarry": 36070, + "Ġquart": 40276, + "Ġquarter": 8338, + "Ġquarterback": 19481, + "Ġquarterbacks": 60491, + "Ġquarterly": 50434, + "Ġquarters": 31473, + "Ġquartz": 51155, + "Ġquasi": 47744, + "Ġquat": 73221, + "Ġquaternion": 66775, + "Ġquatre": 98209, + "Ġque": 1709, + "Ġqued": 59186, + "Ġqueda": 90876, + "Ġqueen": 27906, + "Ġqueens": 68997, + "Ġqueer": 54541, + "Ġquel": 25025, + "Ġquella": 78282, + "Ġquelle": 73840, + "Ġquello": 63954, + "Ġquelque": 77971, + "Ġquelques": 44789, + "Ġquem": 57608, + "Ġquer": 29134, + "Ġqueried": 77902, + "Ġqueries": 19556, + "Ġquery": 3239, + "ĠqueryInterface": 94915, + "ĠqueryParams": 53469, + "ĠqueryString": 50679, + "Ġquerying": 81098, + "Ġqueryset": 39527, + "Ġques": 91384, + "Ġquest": 2222, + "Ġquesta": 51601, + "Ġquesti": 84681, + "Ġquestion": 3405, + "Ġquestionable": 43278, + "Ġquestioned": 28340, + "Ġquestioning": 33585, + "Ġquestionnaire": 47864, + "Ġquestions": 4755, + "Ġquesto": 33460, + "Ġquests": 49177, + "Ġquestão": 138521, + "Ġquestões": 143486, + "Ġqueue": 7177, + "Ġqueued": 57163, + "Ġqueues": 48094, + "Ġqui": 7774, + "Ġquick": 3974, + "Ġquicker": 38292, + "Ġquickest": 80510, + "Ġquickly": 6157, + "Ġquien": 43889, + "Ġquienes": 92390, + "Ġquiere": 73068, + "Ġquieres": 94295, + "Ġquiero": 91720, + "Ġquiet": 11340, + "Ġquieter": 89745, + "Ġquietly": 29566, + "Ġquil": 74864, + "Ġquilt": 60736, + "Ġquindi": 76608, + "Ġquint": 39657, + "Ġquir": 49626, + "Ġquirky": 67132, + "Ġquis": 48480, + "Ġquit": 16835, + "Ġquite": 5008, + "Ġquits": 69477, + "Ġquitting": 68084, + "Ġquiz": 27124, + "Ġquizzes": 88223, + "Ġquizá": 143876, + "Ġquo": 40508, + "Ġquoi": 74149, + "Ġquot": 26370, + "Ġquota": 42042, + "Ġquotas": 84818, + "Ġquotation": 54231, + "Ġquotations": 85987, + "Ġquote": 12641, + "Ġquoted": 23237, + "Ġquotes": 17194, + "Ġquotid": 98468, + "Ġquotient": 74762, + "Ġquoting": 56518, + "Ġquy": 126370, + "Ġquyết": 128784, + "Ġquyá»ģn": 128918, + "Ġquá": 128459, + "Ġquán": 133751, + "Ġquân": 129866, + "Ġqué": 42288, + "Ġquê": 135348, + "Ġquên": 135967, + "Ġquý": 129681, + "ĠquÃł": 135282, + "Ġquả": 128432, + "Ġquản": 129270, + "Ġquảng": 133503, + "Ġquần": 132543, + "ĠquáºŃn": 132859, + "Ġquá»ijc": 128664, + "Ġqw": 41178, + "Ġqx": 96030, + "Ġr": 435, + "Ġra": 15122, + "Ġrab": 30574, + "Ġrabbit": 38724, + "Ġrabbits": 69144, + "Ġrac": 8971, + "Ġrace": 6957, + "Ġraced": 71054, + "Ġracer": 84832, + "Ġraces": 20588, + "Ġracial": 19189, + "Ġracially": 74804, + "Ġracing": 21313, + "Ġracism": 25962, + "Ġracist": 24207, + "Ġrack": 29659, + "Ġracked": 99041, + "Ġracket": 81512, + "Ġracks": 65868, + "Ġrad": 8862, + "Ġradar": 27508, + "Ġradi": 11900, + "Ġradial": 56836, + "Ġradians": 50784, + "Ġradiant": 76343, + "Ġradiation": 24403, + "Ġradiator": 77090, + "Ġradical": 17855, + "Ġradically": 60027, + "Ġradicals": 73256, + "Ġradio": 8887, + "ĠradioButton": 75420, + "Ġradioactive": 58762, + "Ġradios": 70069, + "Ġradius": 10578, + "Ġradix": 75658, + "Ġraft": 52455, + "Ġrag": 20475, + "Ġragaz": 32433, + "Ġragazza": 76195, + "Ġragazze": 47599, + "Ġragazzi": 97136, + "Ġragazzo": 92998, + "Ġrage": 32949, + "Ġraging": 69417, + "Ġrah": 141846, + "Ġrahats": 141847, + "Ġrahatsız": 141848, + "Ġraid": 24797, + "Ġraided": 78396, + "Ġraids": 51481, + "Ġrail": 13569, + "Ġrailing": 91830, + "Ġrailroad": 50476, + "Ġrails": 42600, + "Ġrailway": 39006, + "Ġrailways": 95147, + "Ġrain": 11174, + "Ġrainbow": 47613, + "Ġrainfall": 52858, + "Ġraining": 83253, + "Ġrains": 61455, + "Ġrainy": 62757, + "Ġrais": 6369, + "Ġraise": 4828, + "Ġraised": 9226, + "Ġraises": 24886, + "Ġraising": 18545, + "Ġraison": 72073, + "Ġraj": 91428, + "Ġrak": 82944, + "Ġrake": 75418, + "Ġrall": 38819, + "Ġrallied": 75337, + "Ġrallies": 56331, + "Ġrally": 19217, + "Ġrallying": 90056, + "Ġram": 17823, + "Ġramifications": 84350, + "Ġramp": 22294, + "Ġrampage": 93063, + "Ġrampant": 62577, + "Ġramps": 86116, + "Ġran": 10613, + "Ġranch": 38061, + "Ġrand": 10382, + "Ġrandint": 53182, + "Ġrandom": 4194, + "ĠrandomNumber": 87720, + "Ġrandomized": 46241, + "Ġrandomly": 26618, + "Ġrandomness": 86690, + "Ġrang": 42100, + "Ġrange": 2088, + "Ġranged": 40729, + "Ġranger": 78059, + "Ġranges": 21283, + "Ġranging": 23994, + "Ġrank": 7077, + "Ġranked": 21006, + "Ġranking": 23001, + "Ġrankings": 32307, + "Ġranks": 20803, + "Ġransom": 57586, + "Ġrant": 56843, + "Ġrap": 7327, + "Ġrape": 16781, + "Ġraped": 39377, + "Ġrapes": 95230, + "Ġrapid": 11048, + "Ġrapide": 75508, + "Ġrapidement": 99098, + "Ġrapidly": 18512, + "Ġraping": 86446, + "Ġrapp": 60058, + "Ġrapper": 49337, + "Ġrapport": 38374, + "Ġrar": 95566, + "Ġrare": 8848, + "Ġrarely": 18521, + "Ġrarity": 58771, + "Ġras": 49910, + "Ġrash": 56242, + "Ġrasp": 84046, + "Ġraspberry": 93702, + "Ġraster": 47208, + "Ġrat": 11244, + "Ġrate": 4379, + "Ġrated": 21628, + "Ġrates": 7813, + "Ġrather": 4751, + "Ġratified": 96565, + "Ġrating": 10728, + "Ġratings": 18102, + "Ġratio": 11341, + "Ġration": 65801, + "Ġrational": 24438, + "Ġrationale": 56816, + "Ġratios": 41238, + "Ġrats": 31410, + "Ġratt": 56545, + "Ġrav": 42543, + "Ġrave": 90403, + "Ġraw": 7112, + "ĠrawData": 83160, + "Ġray": 18299, + "Ġrays": 44120, + "Ġraz": 23848, + "Ġrazor": 59130, + "Ġrazón": 140100, + "ĠraÄŁ": 134492, + "ĠraÄŁmen": 134493, + "Ġrb": 18717, + "Ġrc": 10192, + "Ġrd": 22694, + "Ġrdf": 66065, + "Ġrdr": 66984, + "Ġre": 312, + "Ġreach": 5545, + "Ġreachable": 68284, + "Ġreached": 8643, + "Ġreaches": 24491, + "Ġreaching": 18739, + "Ġreact": 13767, + "Ġreacted": 54741, + "Ġreacting": 73050, + "Ġreaction": 12720, + "Ġreactionary": 95138, + "Ġreactions": 24473, + "Ġreactive": 55463, + "Ġreactor": 37629, + "Ġreactors": 70473, + "Ġreacts": 68845, + "Ġread": 1349, + "ĠreadFile": 61732, + "ĠreadOnly": 61451, + "Ġreadability": 91494, + "Ġreadable": 33798, + "Ġreaddir": 80132, + "Ġreader": 6604, + "Ġreaders": 12726, + "Ġreadily": 30268, + "Ġreadiness": 61692, + "Ġreading": 5290, + "Ġreadings": 39074, + "Ġreadline": 74766, + "Ġreadme": 83684, + "Ġreadonly": 5762, + "Ġreads": 15804, + "Ġready": 5527, + "Ġreaff": 72000, + "Ġreal": 1931, + "Ġrealidad": 80504, + "Ġrealise": 38156, + "Ġrealised": 40093, + "Ġrealism": 68771, + "Ġrealistic": 25489, + "Ġrealistically": 88616, + "Ġrealities": 49346, + "Ġreality": 8729, + "Ġrealiz": 33760, + "Ġrealiza": 91735, + "Ġrealizado": 83045, + "Ġrealizar": 41558, + "Ġrealization": 48703, + "Ġrealize": 13083, + "Ġrealized": 15043, + "Ġrealizes": 51594, + "Ġrealizing": 43014, + "Ġrealloc": 66015, + "Ġreally": 2167, + "Ġrealm": 21889, + "Ġrealmente": 68847, + "Ġrealms": 76463, + "Ġrealpath": 91794, + "Ġrealtime": 84978, + "ĠrealtÃł": 134523, + "Ġreap": 56903, + "Ġrear": 14639, + "Ġrearr": 55327, + "Ġreason": 2874, + "Ġreasonable": 13276, + "Ġreasonably": 28446, + "Ġreasoned": 92369, + "Ġreasoning": 32711, + "Ġreasons": 7966, + "Ġreass": 31734, + "Ġreassure": 90047, + "Ġreassuring": 81864, + "Ġreb": 12366, + "Ġrebate": 89946, + "Ġrebel": 35317, + "Ġrebell": 41201, + "Ġrebellion": 52748, + "Ġrebels": 35623, + "Ġreboot": 34592, + "Ġrebound": 41532, + "Ġrebounds": 42662, + "Ġrebuild": 31493, + "Ġrebuilding": 55128, + "Ġrebuilt": 59585, + "Ġrebut": 83029, + "Ġrec": 1395, + "Ġrecal": 98683, + "Ġrecall": 19091, + "Ġrecalled": 31130, + "Ġrecalling": 88646, + "Ġrecalls": 40131, + "Ġrecap": 53999, + "Ġrece": 2166, + "Ġrecebe": 96730, + "Ġreceber": 90145, + "Ġreceipt": 22567, + "Ġreceipts": 56480, + "Ġreceive": 5258, + "Ġreceived": 3949, + "Ġreceiver": 13964, + "Ġreceivers": 39906, + "Ġreceives": 21189, + "Ġreceiving": 12308, + "Ġrecent": 3213, + "Ġrecently": 5926, + "Ġrecept": 28818, + "Ġreception": 23117, + "Ġreceptions": 80193, + "Ġreceptive": 87551, + "Ġreceptor": 34168, + "Ġreceptors": 43440, + "Ġrecess": 46762, + "Ġrecession": 38521, + "Ġrecharge": 46823, + "Ġrecher": 87717, + "Ġrecherche": 37217, + "Ġrecht": 67519, + "Ġrecib": 45269, + "Ġrecibir": 93180, + "Ġrecieved": 80892, + "Ġrecip": 24450, + "Ġrecipe": 11116, + "Ġrecipes": 18627, + "Ġrecipient": 21713, + "Ġrecipients": 33776, + "Ġrecipro": 66542, + "Ġreciprocal": 86198, + "Ġreck": 28199, + "Ġreckless": 53217, + "Ġreckon": 87465, + "Ġrecl": 47420, + "Ġreclaim": 57483, + "Ġreclaimed": 88659, + "Ġreco": 47655, + "Ġrecogn": 5829, + "Ġrecognise": 43817, + "Ġrecognised": 38664, + "Ġrecognition": 17843, + "Ġrecognizable": 65252, + "Ġrecognize": 15282, + "Ġrecognized": 14975, + "Ġrecognizer": 93159, + "Ġrecognizes": 44699, + "Ġrecognizing": 48083, + "Ġrecoil": 73322, + "Ġrecom": 37201, + "Ġrecomend": 70394, + "Ġrecomm": 4861, + "Ġrecommand": 83093, + "Ġrecommend": 6934, + "Ġrecommendation": 27682, + "Ġrecommendations": 18564, + "Ġrecommended": 11102, + "Ġrecommending": 64674, + "Ġrecommends": 39812, + "Ġrecon": 16070, + "Ġreconc": 31445, + "Ġreconcile": 63408, + "Ġreconciliation": 57655, + "Ġreconnaissance": 87505, + "Ġreconnect": 50036, + "Ġreconoc": 82911, + "Ġreconsider": 39075, + "Ġreconstruct": 43828, + "Ġreconstructed": 82004, + "Ġreconstruction": 42638, + "Ġrecord": 3255, + "Ġrecorded": 12433, + "Ġrecorder": 48835, + "Ġrecording": 14633, + "Ġrecordings": 37040, + "Ġrecords": 7424, + "Ġrecount": 40186, + "Ġrecounted": 88491, + "Ġrecounts": 94789, + "Ġrecourse": 92270, + "Ġrecover": 11731, + "Ġrecovered": 25335, + "Ġrecovering": 41286, + "Ġrecovery": 13351, + "Ġrecre": 22824, + "Ġrecreate": 52775, + "Ġrecreated": 36580, + "Ġrecreation": 45944, + "Ġrecreational": 36000, + "Ġrecru": 12117, + "Ġrecruit": 20309, + "Ġrecruited": 44325, + "Ġrecruiter": 96601, + "Ġrecruiters": 92840, + "Ġrecruiting": 31859, + "Ġrecruitment": 33002, + "Ġrecruits": 55097, + "Ġrect": 7608, + "Ġrectangle": 22756, + "Ġrectangles": 76192, + "Ġrectangular": 51424, + "Ġrects": 82656, + "Ġrecuper": 51521, + "Ġrecur": 63548, + "Ġrecurrence": 75193, + "Ġrecurrent": 64074, + "Ġrecurring": 45250, + "Ġrecurs": 24724, + "Ġrecurse": 73299, + "Ġrecursion": 50262, + "Ġrecursive": 30819, + "Ġrecursively": 52847, + "Ġrecursos": 59080, + "Ġrecv": 27006, + "Ġrecycl": 91365, + "Ġrecycle": 60743, + "Ġrecycled": 46558, + "ĠrecyclerView": 37149, + "Ġrecycling": 33878, + "Ġred": 2518, + "Ġredd": 62144, + "Ġreddit": 42174, + "Ġrede": 30165, + "Ġredeem": 48778, + "Ġredeemed": 83243, + "Ġredefine": 83275, + "Ġredemption": 55652, + "Ġredes": 46078, + "Ġredesign": 55619, + "Ġredesigned": 75472, + "Ġredevelopment": 87639, + "Ġredhead": 91705, + "Ġredirect": 6423, + "ĠredirectTo": 64698, + "Ġredirected": 48928, + "Ġredirection": 84059, + "Ġredirects": 71171, + "Ġredis": 20870, + "Ġredistrib": 8702, + "Ġredistribute": 9385, + "Ġredistributed": 73394, + "Ġredistribution": 71060, + "Ġredo": 57923, + "Ġredraw": 56259, + "Ġredu": 3595, + "Ġreduce": 7949, + "Ġreduced": 11046, + "Ġreducer": 37004, + "Ġreducers": 75209, + "Ġreduces": 25271, + "Ġreducing": 17719, + "Ġreduction": 13951, + "Ġreductions": 46211, + "Ġredund": 36573, + "Ġredundancy": 89373, + "Ġredundant": 47732, + "Ġredux": 67616, + "Ġreef": 70045, + "Ġreefs": 91722, + "Ġreel": 56282, + "Ġreelection": 92711, + "Ġreels": 78500, + "Ġref": 2053, + "Ġrefactor": 92295, + "Ġrefer": 8300, + "Ġrefere": 38908, + "Ġreferee": 54202, + "Ġreferees": 95826, + "Ġreference": 5785, + "Ġreferenced": 24784, + "ĠreferencedColumnName": 82347, + "Ġreferences": 15057, + "Ġreferencia": 81909, + "Ġreferencing": 56516, + "Ġreferendum": 29698, + "Ġreferral": 44780, + "Ġreferrals": 69301, + "Ġreferred": 13862, + "Ġreferring": 22023, + "Ġrefers": 19257, + "Ġrefill": 73715, + "Ġrefin": 59696, + "Ġrefine": 45364, + "Ġrefined": 37191, + "Ġrefinement": 72913, + "Ġrefinery": 93634, + "Ġrefining": 73185, + "Ġrefl": 82435, + "Ġreflect": 8708, + "Ġreflected": 25911, + "Ġreflecting": 41752, + "Ġreflection": 21844, + "Ġreflections": 62751, + "Ġreflective": 51728, + "Ġreflects": 25963, + "Ġreflex": 32666, + "Ġreflexivity": 68333, + "Ġreflux": 92162, + "Ġreform": 14836, + "Ġreforms": 30243, + "Ġrefr": 19353, + "Ġrefrain": 57508, + "Ġrefresh": 10408, + "ĠrefreshToken": 74541, + "Ġrefreshed": 65129, + "Ġrefreshing": 35918, + "Ġrefriger": 29377, + "Ġrefrigerator": 44944, + "Ġrefs": 43143, + "Ġrefuge": 14510, + "Ġrefugee": 33167, + "Ġrefugees": 21728, + "Ġrefund": 20965, + "Ġrefunded": 93452, + "Ġrefunds": 72518, + "Ġrefurb": 55656, + "Ġrefurbished": 82315, + "Ġrefusal": 43151, + "Ġrefuse": 25066, + "Ġrefused": 16051, + "Ġrefuses": 40616, + "Ġrefusing": 37989, + "Ġrefute": 83177, + "Ġreg": 1217, + "Ġregain": 48555, + "Ġregained": 84925, + "Ġregard": 5250, + "Ġregarded": 26361, + "Ġregarding": 8826, + "Ġregardless": 15484, + "Ġregards": 23935, + "Ġregelm": 142832, + "ĠregelmÃ¤ÃŁig": 142833, + "Ġregenerate": 79451, + "Ġregenerated": 70353, + "Ġregeneration": 59417, + "Ġregex": 20180, + "Ġregexp": 41877, + "Ġregime": 17484, + "Ġregimen": 67028, + "Ġregiment": 74441, + "Ġregimes": 60811, + "Ġregion": 5537, + "Ġregional": 15130, + "Ġregions": 13604, + "Ġregist": 31247, + "Ġregister": 4161, + "Ġregistered": 9681, + "Ġregistering": 38729, + "Ġregisters": 24740, + "Ġregistr": 22855, + "Ġregistrado": 96311, + "Ġregistrar": 61467, + "Ġregistration": 12227, + "Ġregistrations": 69490, + "Ġregistrazione": 95835, + "Ġregistro": 36785, + "Ġregistros": 63527, + "Ġregistry": 19424, + "Ġregião": 134306, + "Ġregión": 141661, + "Ġregress": 69780, + "Ġregression": 30549, + "Ġregret": 22231, + "Ġregrets": 67925, + "Ġregs": 44498, + "Ġregul": 14089, + "Ġregular": 5792, + "Ġregularization": 87294, + "Ġregularly": 15502, + "Ġregulate": 36277, + "Ġregulated": 34219, + "Ġregulates": 79312, + "Ġregulating": 57399, + "Ġregulation": 19256, + "Ġregulations": 14305, + "Ġregulator": 39604, + "Ġregulators": 39142, + "Ġregulatory": 22515, + "Ġrehab": 26778, + "Ġrehabilit": 88166, + "Ġrehabilitation": 43338, + "Ġrehe": 47206, + "Ġrehears": 68572, + "Ġrehearsal": 88855, + "Ġreife": 46161, + "Ġreign": 30302, + "Ġreigning": 89956, + "Ġreim": 43987, + "Ġreimb": 97578, + "Ġreimburse": 95160, + "Ġreimbursement": 70011, + "Ġrein": 15244, + "Ġreinc": 94707, + "Ġreinforce": 54314, + "Ġreinforced": 48580, + "Ġreinforcement": 71278, + "Ġreinforcements": 98500, + "Ġreinforces": 94493, + "Ġreinforcing": 90015, + "Ġreins": 89418, + "Ġreinst": 68667, + "Ġreinstall": 91861, + "Ġreint": 75167, + "Ġreinterpret": 25186, + "Ġreiterated": 65747, + "Ġrej": 77367, + "Ġreject": 7850, + "Ġrejected": 17551, + "Ġrejecting": 62586, + "Ġrejection": 36901, + "Ġrejects": 59415, + "Ġrejo": 46941, + "Ġrejoice": 91759, + "Ġrejuven": 72384, + "Ġrek": 77281, + "Ġrekl": 99080, + "Ġrel": 1351, + "Ġrelacion": 46005, + "Ġrelaciones": 96573, + "Ġrelación": 68197, + "Ġrelat": 90803, + "Ġrelate": 28143, + "Ġrelated": 5435, + "Ġrelates": 35616, + "Ġrelating": 22527, + "Ġrelation": 12687, + "Ġrelational": 71183, + "Ġrelations": 4300, + "Ġrelationship": 5025, + "Ġrelationships": 11871, + "Ġrelativ": 58325, + "Ġrelative": 8674, + "Ġrelatively": 12040, + "Ġrelatives": 28558, + "Ġrelax": 11967, + "Ġrelaxation": 42585, + "Ġrelaxed": 30367, + "Ġrelaxing": 33848, + "Ġrelay": 31851, + "Ġrelação": 96103, + "Ġrele": 13733, + "Ġrelease": 4879, + "Ġreleased": 5880, + "Ġreleases": 19232, + "Ġreleasing": 27865, + "Ġreleg": 58682, + "Ġrelegated": 86452, + "Ġrelent": 42846, + "Ġrelentless": 59035, + "Ġrelentlessly": 90170, + "Ġrelev": 78315, + "Ġrelevance": 40861, + "Ġrelevant": 9760, + "Ġreli": 9375, + "Ġreliability": 30538, + "Ġreliable": 14720, + "Ġreliably": 56382, + "Ġreliance": 53080, + "Ġreliant": 88127, + "Ġrelic": 60937, + "Ġrelics": 98416, + "Ġrelie": 58544, + "Ġrelied": 39913, + "Ġrelief": 15957, + "Ġrelies": 33644, + "Ġrelieve": 47739, + "Ġrelieved": 50412, + "Ġrelig": 6694, + "Ġreligion": 13587, + "Ġreligions": 43130, + "Ġreligious": 10381, + "Ġrelinqu": 84398, + "Ġreload": 18646, + "ĠreloadData": 57847, + "Ġreloading": 85630, + "Ġreloc": 74295, + "Ġrelocate": 69572, + "Ġrelocated": 69109, + "Ġrelocation": 59895, + "Ġrelu": 92874, + "Ġreluct": 30244, + "Ġreluctance": 82950, + "Ġreluctant": 42940, + "Ġreluctantly": 83085, + "Ġrely": 17188, + "Ġrelying": 38561, + "Ġrem": 1299, + "Ġremain": 7146, + "Ġremainder": 26313, + "Ġremained": 14616, + "Ġremaining": 9664, + "Ġremains": 8458, + "Ġremake": 58055, + "Ġremar": 83128, + "Ġremark": 13841, + "Ġremarkable": 22804, + "Ġremarkably": 48623, + "Ġremarked": 60314, + "Ġremarks": 23325, + "Ġrematch": 92226, + "Ġremed": 33530, + "Ġremedies": 50136, + "Ġremedy": 39139, + "Ġremember": 6099, + "Ġremembered": 26471, + "Ġremembering": 47284, + "Ġremembers": 42357, + "Ġremin": 42550, + "Ġremind": 23974, + "Ġreminded": 30156, + "Ġreminder": 26528, + "Ġreminders": 65597, + "Ġreminding": 62942, + "Ġreminds": 34610, + "Ġreminis": 48579, + "Ġreminiscent": 54985, + "Ġremix": 56366, + "Ġremnants": 72340, + "Ġremodel": 45986, + "Ġremodeling": 69330, + "Ġremorse": 78057, + "Ġremot": 35793, + "Ġremote": 8699, + "Ġremotely": 38429, + "Ġremovable": 53791, + "Ġremoval": 16652, + "Ġremove": 4057, + "ĠremoveAll": 65099, + "ĠremoveFrom": 57785, + "ĠremoveFromSuperview": 77180, + "ĠremoveObject": 81622, + "Ġremoved": 6963, + "Ġremover": 76248, + "Ġremoves": 28160, + "Ġremoving": 17592, + "Ġrempl": 68141, + "Ġren": 5672, + "Ġrenal": 62815, + "Ġrename": 29074, + "Ġrenamed": 27280, + "Ġrenaming": 92890, + "Ġrencont": 8875, + "Ġrencontr": 65183, + "Ġrencontre": 10391, + "Ġrencontrer": 89926, + "Ġrencontres": 26820, + "Ġrend": 17191, + "Ġrender": 3141, + "ĠrenderItem": 86667, + "Ġrendered": 22383, + "Ġrenderer": 19715, + "Ġrendering": 20898, + "Ġrenders": 44282, + "Ġrendez": 65551, + "Ġrendition": 82865, + "Ġrendre": 96954, + "Ġrenegot": 98677, + "Ġrenew": 13873, + "Ġrenewable": 32168, + "Ġrenewables": 87985, + "Ġrenewal": 42801, + "Ġrenewed": 35546, + "Ġrenov": 31552, + "Ġrenovated": 56834, + "Ġrenovation": 49455, + "Ġrenovations": 77564, + "Ġrenown": 33717, + "Ġrenowned": 35948, + "Ġrent": 8016, + "Ġrental": 18645, + "Ġrentals": 47004, + "Ġrented": 48859, + "Ġrenters": 91851, + "Ġrenting": 52227, + "Ġrents": 61990, + "Ġreopen": 36759, + "Ġreopened": 77781, + "Ġreopening": 93400, + "Ġreorder": 83184, + "Ġreordered": 49032, + "Ġrep": 2064, + "Ġrepaint": 85689, + "Ġrepair": 12733, + "Ġrepaired": 51734, + "Ġrepairing": 67237, + "Ġrepairs": 30186, + "Ġrepar": 75926, + "Ġrepay": 53338, + "Ġrepayment": 70018, + "Ġrepe": 6219, + "Ġrepeal": 39579, + "Ġrepealed": 85914, + "Ġrepeat": 13153, + "Ġrepeated": 11504, + "Ġrepeatedly": 18822, + "Ġrepeating": 39816, + "Ġrepeats": 43594, + "Ġrepell": 87774, + "Ġrepent": 62178, + "Ġreperc": 73388, + "Ġrepercussions": 80809, + "Ġrepertoire": 76668, + "Ġrepet": 26330, + "Ġrepetition": 53415, + "Ġrepetitions": 84966, + "Ġrepetitive": 58077, + "Ġrepl": 5791, + "Ġreplace": 8290, + "Ġreplaced": 12575, + "Ġreplacement": 13723, + "Ġreplacements": 53055, + "Ġreplaces": 40700, + "Ġreplacing": 24891, + "Ġreplay": 34485, + "Ġreplen": 69443, + "Ġreplic": 28541, + "Ġreplica": 36954, + "Ġreplicas": 80801, + "Ġreplicate": 45013, + "Ġreplicated": 71380, + "Ġreplication": 47791, + "Ġreplied": 19995, + "Ġreplies": 30637, + "Ġreply": 9851, + "Ġrepmat": 79339, + "Ġrepo": 15867, + "Ġreport": 1895, + "Ġreported": 4961, + "Ġreportedly": 17828, + "Ġreporter": 18960, + "Ġreporters": 19040, + "Ġreporting": 12827, + "Ġreports": 6682, + "Ġrepos": 45774, + "Ġrepositories": 49657, + "Ġrepository": 12542, + "Ġrepost": 94187, + "Ġrepr": 30636, + "Ġrepreh": 77837, + "Ġrepresent": 4009, + "Ġrepresenta": 95115, + "Ġrepresentation": 13042, + "Ġrepresentations": 43613, + "Ġrepresentative": 18239, + "Ġrepresentatives": 23130, + "Ġrepresented": 15251, + "Ġrepresenting": 14064, + "Ġrepresents": 10868, + "Ġrepression": 71398, + "Ġreprint": 80471, + "Ġrepro": 52723, + "Ġreprodu": 14503, + "Ġreproduce": 22800, + "Ġreproduced": 54617, + "Ġreproduction": 38556, + "Ġreproductive": 41789, + "Ġreprés": 70825, + "Ġreprésent": 90578, + "Ġreprésente": 140786, + "Ġreps": 40411, + "Ġrept": 63773, + "Ġrepublic": 34444, + "Ġrepublican": 90436, + "Ġreput": 42739, + "Ġreputable": 55840, + "Ġreputation": 17011, + "Ġreputed": 91972, + "Ġreq": 4232, + "Ġrequ": 5892, + "Ġrequer": 66687, + "Ġrequest": 1681, + "ĠrequestBody": 85626, + "ĠrequestCode": 55374, + "ĠrequestData": 97264, + "ĠrequestId": 69861, + "ĠrequestOptions": 91094, + "Ġrequested": 11223, + "Ġrequester": 83584, + "Ġrequesting": 34692, + "Ġrequests": 7388, + "Ġrequire": 1373, + "Ġrequired": 2567, + "Ġrequirement": 16287, + "Ġrequirements": 8502, + "Ġrequires": 7460, + "Ġrequiring": 22703, + "Ġrequis": 44582, + "Ġrequisite": 81680, + "Ġrer": 48547, + "Ġres": 592, + "Ġresale": 98224, + "Ġresc": 12912, + "Ġrescue": 17186, + "Ġrescued": 44333, + "Ġresearch": 3412, + "Ġresearched": 55998, + "Ġresearcher": 31085, + "Ġresearchers": 11811, + "Ġresearching": 44143, + "Ġresembl": 66320, + "Ġresemblance": 68900, + "Ġresemble": 51180, + "Ġresembled": 95758, + "Ġresembles": 52191, + "Ġresembling": 70607, + "Ġresend": 93888, + "Ġresent": 46440, + "Ġresentment": 70554, + "Ġreserv": 21196, + "Ġreserva": 98644, + "Ġreservation": 27667, + "Ġreservations": 40062, + "Ġreserve": 20499, + "Ġreserved": 4591, + "Ġreserves": 29500, + "Ġreservoir": 44412, + "Ġreset": 7585, + "Ġresets": 66197, + "Ġresett": 90388, + "Ġresetting": 78028, + "Ġresh": 63693, + "Ġreshape": 43809, + "Ġresid": 21151, + "Ġreside": 47283, + "Ġresidence": 21682, + "Ġresidences": 84771, + "Ġresidency": 52866, + "Ġresident": 18967, + "Ġresidential": 20057, + "Ġresidents": 10826, + "Ġresides": 52968, + "Ġresiding": 66412, + "Ġresidual": 32147, + "Ġresiduals": 97540, + "Ġresidue": 48132, + "Ġresidues": 70683, + "Ġresign": 23592, + "Ġresignation": 43634, + "Ġresigned": 39107, + "Ġresil": 36030, + "Ġresilience": 54962, + "Ġresilient": 58680, + "Ġresin": 53514, + "Ġresist": 22106, + "Ġresistance": 13643, + "Ġresistant": 30685, + "Ġresisted": 66314, + "Ġresisting": 69341, + "Ġresistor": 77636, + "Ġresizable": 97083, + "Ġresize": 20925, + "ĠresizeMode": 98579, + "Ġresized": 58549, + "Ġresizing": 86243, + "Ġresolution": 10935, + "Ġresolutions": 42495, + "Ġresolve": 8830, + "Ġresolved": 19673, + "Ġresolver": 36220, + "Ġresolves": 63297, + "Ġresolving": 52483, + "Ġreson": 28180, + "Ġresonance": 56981, + "Ġresonate": 88886, + "Ġresort": 21788, + "Ġresorts": 60445, + "Ġresource": 5101, + "ĠresourceId": 97246, + "ĠresourceName": 98485, + "Ġresources": 4963, + "Ġresp": 9039, + "Ġrespawn": 96032, + "Ġrespect": 5091, + "Ġrespectable": 61386, + "Ġrespected": 30287, + "Ġrespectful": 48050, + "Ġrespectfully": 92345, + "Ġrespecting": 68415, + "Ġrespective": 19511, + "Ġrespectively": 15576, + "Ġrespecto": 88060, + "Ġrespects": 42566, + "Ġrespir": 32415, + "Ġrespiratory": 41531, + "Ġrespond": 5889, + "Ġresponded": 16441, + "Ġrespondent": 66888, + "Ġrespondents": 33181, + "Ġresponder": 64034, + "Ġresponders": 69539, + "Ġresponding": 29338, + "Ġresponds": 30580, + "ĠrespondsToSelector": 62293, + "Ġrespons": 4200, + "Ġresponsable": 76832, + "Ġresponse": 2033, + "ĠresponseBody": 98065, + "ĠresponseData": 67445, + "ĠresponseObject": 69040, + "ĠresponseType": 90444, + "Ġresponses": 14507, + "Ġresponsibilities": 27323, + "Ġresponsibility": 11752, + "Ġresponsible": 8480, + "Ġresponsibly": 86288, + "Ġresponsive": 25988, + "Ġresponsiveness": 98939, + "Ġresponsável": 141402, + "Ġresposta": 74506, + "Ġrespuesta": 43360, + "Ġrest": 2732, + "ĠrestTemplate": 79829, + "Ġresta": 96952, + "Ġrestart": 17024, + "Ġrestarted": 80654, + "Ġrestarting": 92524, + "Ġrestaur": 7544, + "Ġrestaurant": 10729, + "Ġrestaurants": 15556, + "Ġreste": 65067, + "Ġrested": 65429, + "Ġresting": 40119, + "Ġrestitution": 89209, + "Ġrestless": 88258, + "Ġresto": 62716, + "Ġrestoration": 33993, + "Ġrestore": 14952, + "Ġrestored": 27003, + "Ġrestores": 87755, + "Ġrestoring": 49103, + "Ġrestr": 15583, + "Ġrestrain": 96776, + "Ġrestrained": 76200, + "Ġrestraining": 83293, + "Ġrestraint": 58356, + "Ġrestrict": 8891, + "Ġrestricted": 21739, + "Ġrestricting": 64851, + "Ġrestriction": 19455, + "Ġrestrictions": 16869, + "Ġrestrictive": 56996, + "Ġrestroom": 88809, + "Ġrestructuring": 67927, + "Ġrests": 53231, + "Ġresult": 1102, + "ĠresultCode": 42631, + "ĠresultList": 71596, + "ĠresultMap": 90461, + "ĠresultSet": 41973, + "Ġresultado": 26192, + "Ġresultados": 46722, + "Ġresultant": 83813, + "Ġresultat": 90272, + "Ġresulted": 19006, + "Ġresulting": 12942, + "Ġresults": 3059, + "Ġresume": 15688, + "Ġresumed": 57405, + "Ġresumes": 65213, + "Ġresurgence": 90490, + "Ġresurrect": 65719, + "Ġresurrection": 61432, + "Ġret": 2112, + "ĠretVal": 34730, + "Ġretail": 10806, + "Ġretailer": 36791, + "Ġretailers": 29182, + "Ġretain": 14060, + "Ġretained": 34263, + "Ġretaining": 50010, + "Ġretains": 51844, + "Ġretal": 50146, + "Ġretali": 90055, + "Ġretaliation": 60349, + "Ġretard": 75972, + "Ġretarded": 97359, + "Ġretention": 37131, + "Ġrethink": 75655, + "Ġretina": 83727, + "Ġretir": 81679, + "Ġretire": 15800, + "Ġretired": 21583, + "Ġretirees": 96723, + "Ġretirement": 20950, + "Ġretiring": 59773, + "Ġretorn": 87594, + "Ġretorna": 71675, + "Ġretorno": 50132, + "Ġretour": 53506, + "Ġretr": 37731, + "Ġretract": 72200, + "Ġretreat": 30014, + "Ġretreated": 90937, + "Ġretrie": 10759, + "Ġretries": 60601, + "Ġretrieval": 56370, + "Ġretrieve": 17179, + "Ġretrieved": 30403, + "Ġretrieves": 70671, + "Ġretrieving": 48224, + "Ġretro": 16771, + "Ġretrofit": 24653, + "Ġretros": 48358, + "Ġretrospect": 76553, + "Ġretrospective": 78829, + "Ġretry": 22683, + "Ġreturn": 470, + "ĠreturnType": 84509, + "ĠreturnUrl": 95012, + "ĠreturnValue": 41597, + "Ġreturned": 5927, + "Ġreturning": 13451, + "Ġreturns": 4675, + "Ġretval": 22625, + "Ġreun": 34640, + "Ġreunion": 57284, + "Ġreunited": 76879, + "Ġreusable": 61571, + "Ġreuse": 25978, + "ĠreuseIdentifier": 52660, + "Ġreused": 68743, + "Ġrev": 5772, + "Ġrevamped": 95496, + "Ġreve": 6195, + "Ġreveal": 16400, + "Ġrevealed": 10457, + "Ġrevealing": 30620, + "Ġreveals": 20992, + "Ġrevel": 22118, + "Ġrevelation": 38910, + "Ġrevelations": 49808, + "Ġreven": 18909, + "Ġrevenge": 36069, + "Ġrevenue": 12957, + "Ġrevenues": 29366, + "Ġrever": 35137, + "Ġrevered": 84661, + "Ġreverence": 97364, + "Ġrevers": 17431, + "Ġreversal": 58114, + "Ġreverse": 9931, + "Ġreversed": 27437, + "Ġreversible": 80093, + "Ġreversing": 75183, + "Ġrevert": 41128, + "Ġreverted": 93593, + "Ġreview": 3395, + "Ġreviewed": 21926, + "Ġreviewer": 55514, + "Ġreviewers": 59038, + "Ġreviewing": 33888, + "Ġreviews": 8379, + "Ġrevis": 17493, + "Ġrevise": 64736, + "Ġrevised": 31728, + "Ġrevision": 23578, + "Ġrevisions": 53762, + "Ġrevisit": 64778, + "Ġrevital": 78855, + "Ġrevival": 56005, + "Ġrevive": 58716, + "Ġrevived": 68275, + "Ġrevoke": 72038, + "Ġrevoked": 64346, + "Ġrevolt": 64835, + "Ġrevolution": 13791, + "Ġrevolutionary": 29091, + "Ġrevolutions": 92474, + "Ġrevolver": 97798, + "Ġrevolves": 78376, + "Ġrevolving": 86026, + "Ġrew": 78622, + "Ġreward": 11313, + "Ġrewarded": 43837, + "Ġrewarding": 40993, + "Ġrewards": 21160, + "Ġrewind": 77665, + "Ġrewrite": 18130, + "Ġrewriting": 94445, + "Ġrewritten": 58524, + "Ġrex": 86510, + "Ġrez": 48058, + "Ġrf": 36026, + "Ġrfl": 69335, + "Ġrg": 18361, + "Ġrgb": 17993, + "Ġrgba": 23524, + "Ġrh": 21669, + "Ġrhe": 67032, + "Ġrhet": 28386, + "Ġrhetoric": 33631, + "Ġrhetorical": 85968, + "Ġrho": 31381, + "Ġrhs": 12916, + "Ġrhyme": 96934, + "Ġrhyth": 28071, + "Ġrhythm": 36290, + "Ġrhythms": 80721, + "Ġri": 24185, + "Ġrib": 20131, + "Ġribbon": 44567, + "Ġribs": 55149, + "Ġric": 29950, + "Ġrice": 19653, + "Ġricerca": 82035, + "Ġrich": 9080, + "ĠrichTextBox": 97429, + "Ġricher": 58674, + "Ġriches": 69834, + "Ġrichest": 56371, + "Ġrichness": 88930, + "Ġricht": 61430, + "Ġrichtig": 65681, + "Ġrico": 93598, + "Ġrid": 9279, + "Ġridden": 92227, + "Ġride": 11877, + "Ġrider": 37126, + "Ġriders": 29703, + "Ġrides": 31227, + "Ġridge": 57833, + "Ġridic": 19965, + "Ġridicule": 93747, + "Ġridiculous": 26775, + "Ġridiculously": 71911, + "Ġriding": 19837, + "Ġrien": 54355, + "Ġries": 53598, + "Ġrif": 36924, + "Ġriff": 80250, + "Ġrifle": 28132, + "Ġrifles": 45370, + "Ġrift": 77157, + "Ġrig": 13249, + "Ġrigged": 80073, + "Ġright": 1290, + "Ġrighteous": 41705, + "Ġrighteousness": 70229, + "Ġrightful": 97944, + "Ġrightfully": 98024, + "Ġrightly": 53366, + "Ġrights": 3188, + "Ġrigid": 32856, + "Ġrigor": 77377, + "Ġrigorous": 46899, + "Ġrigs": 95051, + "Ġrij": 69345, + "Ġrim": 30130, + "Ġrims": 97162, + "Ġrin": 53721, + "Ġring": 10058, + "Ġringing": 67043, + "Ġrings": 24544, + "Ġrins": 96039, + "Ġrinse": 73241, + "Ġrio": 94704, + "Ġriot": 41497, + "Ġriots": 59996, + "Ġrip": 23288, + "Ġripe": 56696, + "Ġripped": 42992, + "Ġripping": 77114, + "Ġripple": 69625, + "Ġris": 9824, + "Ġrise": 10000, + "Ġrisen": 40382, + "Ġrises": 37168, + "Ġrising": 16062, + "Ġrisk": 5214, + "Ġrisking": 84922, + "Ġrisks": 15276, + "Ġrisky": 40173, + "Ġrisult": 88112, + "Ġrit": 21198, + "Ġrites": 98452, + "Ġritual": 27179, + "Ġrituals": 56579, + "Ġriv": 43129, + "Ġrival": 20762, + "Ġrivalry": 59922, + "Ġrivals": 34838, + "Ġriver": 14796, + "Ġrivers": 35517, + "Ġriê": 130007, + "Ġriêng": 130008, + "Ġrk": 58006, + "Ġrl": 38877, + "Ġrm": 18998, + "Ġrms": 79994, + "Ġrn": 55839, + "Ġrnd": 37193, + "Ġrng": 28422, + "Ġrnn": 86400, + "Ġro": 926, + "Ġroad": 5636, + "Ġroadmap": 68817, + "Ġroads": 19241, + "Ġroadside": 79643, + "Ġroadway": 81111, + "Ġroam": 74967, + "Ġroaming": 68253, + "Ġroar": 77624, + "Ġroaring": 92756, + "Ġroast": 57685, + "Ġroasted": 53562, + "Ġrob": 10550, + "Ġrobbed": 62254, + "Ġrobber": 70550, + "Ġrobbery": 43114, + "Ġrobe": 62619, + "Ġrobert": 88893, + "Ġrobes": 70168, + "Ġrobin": 98585, + "Ġrobot": 12305, + "Ġrobotic": 58001, + "Ġrobotics": 73606, + "Ġrobots": 28707, + "Ġrobust": 21765, + "Ġroc": 63330, + "Ġrock": 6946, + "Ġrocked": 77260, + "Ġrocker": 95126, + "Ġrocket": 24306, + "Ġrockets": 51998, + "Ġrocking": 66624, + "Ġrocks": 23035, + "Ġrocky": 55517, + "Ġrod": 20589, + "Ġrode": 40661, + "Ġrodents": 93109, + "Ġrods": 56900, + "Ġrodz": 79541, + "Ġrog": 94156, + "Ġrogue": 53891, + "Ġroi": 44413, + "Ġroku": 55302, + "Ġrol": 17679, + "Ġrole": 3476, + "ĠroleId": 69645, + "ĠroleName": 90947, + "Ġroles": 12783, + "Ġroll": 6502, + "Ġrollback": 60414, + "Ġrolled": 22447, + "Ġroller": 28451, + "Ġrollers": 88613, + "Ġrolling": 20097, + "Ġrollout": 71730, + "Ġrolls": 27373, + "Ġrom": 9882, + "Ġroma": 57467, + "Ġroman": 47776, + "Ġromance": 29263, + "Ġromant": 67467, + "Ġromantic": 23467, + "Ġrond": 88681, + "Ġroof": 15134, + "Ġroofing": 65425, + "Ġroofs": 76295, + "Ġrooft": 55605, + "Ġrooftop": 66192, + "Ġrookie": 30525, + "Ġrookies": 99009, + "Ġroom": 3054, + "ĠroomId": 77506, + "Ġroommate": 69314, + "Ġrooms": 12026, + "Ġroot": 3704, + "ĠrootNode": 77975, + "ĠrootReducer": 84608, + "ĠrootView": 51886, + "Ġrooted": 40876, + "Ġrooting": 86338, + "Ġroots": 19703, + "Ġrop": 48395, + "Ġrope": 33420, + "Ġropes": 63946, + "Ġros": 26190, + "Ġrosa": 95244, + "Ġrose": 16009, + "Ġroses": 60641, + "Ġrospy": 35578, + "Ġroster": 21445, + "Ġrot": 5749, + "Ġrotary": 61817, + "Ġrotate": 16919, + "Ġrotated": 45620, + "Ġrotates": 89059, + "Ġrotating": 41396, + "Ġrotation": 12695, + "Ġrotational": 91271, + "Ġrotations": 69201, + "Ġrotor": 62025, + "Ġrotten": 77692, + "Ġrotterdam": 70429, + "Ġrou": 28365, + "Ġrouge": 83959, + "Ġrough": 11165, + "Ġroughly": 17267, + "Ġroulette": 48111, + "Ġround": 4778, + "Ġrounded": 17976, + "Ġrounding": 51562, + "Ġrounds": 19531, + "Ġroundup": 86669, + "Ġrout": 8242, + "Ġroute": 6021, + "Ġrouted": 72217, + "Ġrouter": 9273, + "Ġrouters": 59175, + "Ġroutes": 11291, + "Ġroutine": 14021, + "Ġroutinely": 38976, + "Ġroutines": 29497, + "Ġrouting": 29058, + "Ġrover": 61252, + "Ġrow": 2802, + "ĠrowCount": 65145, + "ĠrowData": 73658, + "ĠrowIndex": 51438, + "ĠrowNum": 99035, + "Ġrows": 6978, + "Ġrowspan": 51288, + "Ġroy": 54242, + "Ġroyal": 29236, + "Ġroyalties": 84874, + "Ġroyalty": 52717, + "Ġroz": 38752, + "Ġrozp": 140775, + "Ġrozpoc": 140776, + "ĠrozpoczÄĻ": 140777, + "ĠrozwiÄħz": 128126, + "ĠrozwiÄħzania": 142027, + "ĠrozwiÄħzanie": 140491, + "ĠroÅĽ": 139443, + "ĠroÅĽlin": 139444, + "Ġrp": 33109, + "Ġrpc": 35596, + "Ġrpm": 49925, + "Ġrpt": 94952, + "Ġrq": 60602, + "Ġrr": 34393, + "Ġrs": 10036, + "Ġrsa": 68570, + "Ġrsp": 42160, + "Ġrss": 71082, + "Ġrst": 55515, + "Ġrt": 16677, + "Ġrtc": 71842, + "Ġrte": 77391, + "Ġrtl": 61060, + "Ġrtn": 56335, + "Ġrtrim": 84465, + "Ġru": 11122, + "Ġrua": 89910, + "Ġrub": 10273, + "Ġrubbed": 66754, + "Ġrubber": 22674, + "Ġrubbing": 55935, + "Ġrubbish": 60924, + "Ġrubble": 73435, + "Ġruby": 45207, + "Ġrud": 67230, + "Ġrude": 46001, + "Ġrue": 55849, + "Ġrug": 28642, + "Ġrugby": 46450, + "Ġrugged": 54783, + "Ġrugs": 72571, + "Ġruin": 37239, + "Ġruined": 46068, + "Ġruining": 91171, + "Ġruins": 45662, + "Ġrule": 5912, + "Ġruled": 21286, + "Ġruler": 47980, + "Ġrulers": 59896, + "Ġrules": 5601, + "Ġruling": 17408, + "Ġrulings": 77588, + "Ġrum": 13995, + "Ġrumor": 57901, + "Ġrumored": 69368, + "Ġrumors": 34392, + "Ġrumours": 64925, + "Ġrumpe": 98580, + "Ġrun": 1598, + "ĠrunApp": 84680, + "Ġrunaway": 90640, + "Ġrund": 40897, + "Ġrundown": 94648, + "Ġrune": 63499, + "Ġrunes": 79918, + "Ġrunnable": 78679, + "Ġrunner": 22259, + "Ġrunners": 38280, + "Ġrunning": 4303, + "Ġrunoff": 78052, + "Ġruns": 8473, + "Ġruntime": 15592, + "Ġrunway": 51761, + "Ġrupt": 59953, + "Ġrupture": 88769, + "Ġrural": 19082, + "Ġrus": 52454, + "Ġrush": 12973, + "Ġrushed": 32241, + "Ġrushes": 94222, + "Ġrushing": 37017, + "Ġruss": 63563, + "Ġrusse": 91539, + "Ġrussian": 63145, + "Ġrust": 23071, + "Ġrustic": 57272, + "Ġrusty": 97453, + "Ġrut": 54619, + "Ġruta": 67463, + "Ġruth": 53146, + "Ġruthless": 68410, + "Ġrv": 17570, + "Ġrval": 62846, + "Ġrw": 25991, + "Ġrx": 19111, + "Ġry": 33919, + "Ġrz": 53001, + "Ġrá": 48669, + "Ġráp": 69738, + "Ġrápida": 135290, + "Ġrápido": 97320, + "Ġrãi": 142423, + "Ġrè": 82129, + "Ġrègle": 136072, + "Ġré": 9333, + "Ġréal": 58057, + "Ġréalis": 85495, + "Ġréalisé": 137883, + "Ġréalité": 141307, + "Ġrécup": 78096, + "Ġréfé": 90458, + "Ġréférence": 140500, + "Ġrég": 37582, + "Ġrégime": 137269, + "Ġrégion": 87000, + "Ġrégl": 139240, + "Ġrépond": 74771, + "Ġréponse": 89409, + "Ġrés": 30707, + "Ġréseau": 97847, + "Ġréserve": 141595, + "Ġrésult": 64185, + "Ġrésultat": 133078, + "Ġrésultats": 99146, + "Ġréuss": 83688, + "Ġréussi": 135750, + "Ġrê": 141138, + "Ġrêve": 141139, + "Ġrówn": 60374, + "Ġrównież": 67741, + "Ġróż": 79944, + "Ġróżne": 136528, + "Ġróżnic": 143614, + "Ġróżnych": 134713, + "Ġrôle": 137614, + "Ġrõ": 129398, + "Ġrút": 137335, + "ĠrÃłng": 135111, + "ĠrÄĥng": 131771, + "ĠrÄĻk": 135182, + "ĠrÆ¡i": 136978, + "Ġrượ": 136287, + "Ġrượu": 136288, + "Ġrất": 128323, + "Ġrằng": 128640, + "Ġrẻ": 130908, + "Ġrừng": 137730, + "Ġrá»iji": 138615, + "Ġrá»ĵi": 128558, + "Ġrá»Ļng": 129922, + "Ġrá»Ŀi": 138559, + "Ġrá»Ńa": 135525, + "Ġs": 274, + "Ġsa": 822, + "Ġsaat": 51365, + "Ġsab": 19408, + "Ġsabe": 75294, + "Ġsaber": 41004, + "Ġsabot": 60958, + "Ġsabotage": 79653, + "Ġsac": 11100, + "Ġsach": 97576, + "Ġsack": 52333, + "Ġsacked": 76461, + "Ġsacks": 56771, + "Ġsacr": 17038, + "Ġsacram": 85802, + "Ġsacred": 31342, + "Ġsacrific": 30861, + "Ġsacrifice": 27136, + "Ġsacrificed": 66804, + "Ġsacrifices": 56923, + "Ġsacrificing": 72028, + "Ġsad": 12421, + "Ġsadd": 57172, + "Ġsaddened": 93917, + "Ġsaddle": 60758, + "Ġsadly": 43045, + "Ġsadness": 50878, + "Ġsaf": 5536, + "Ġsafari": 82306, + "Ġsafe": 6092, + "Ġsafeg": 37286, + "Ġsafeguard": 47971, + "Ġsafeguards": 76920, + "Ġsafely": 21000, + "Ġsafer": 29449, + "Ġsafest": 61744, + "Ġsafety": 7149, + "Ġsafezone": 86157, + "Ġsag": 29711, + "Ġsaga": 46303, + "Ġsage": 38261, + "Ġsagen": 73689, + "Ġsagt": 75018, + "Ġsagte": 85389, + "Ġsai": 71632, + "Ġsaid": 1053, + "Ġsail": 29403, + "Ġsailed": 75744, + "Ġsailing": 50029, + "Ġsailor": 92537, + "Ġsailors": 68838, + "Ġsails": 85005, + "Ġsaint": 35859, + "Ġsaints": 67120, + "Ġsais": 62662, + "Ġsaison": 78991, + "Ġsaja": 78786, + "Ġsak": 77310, + "Ġsake": 24483, + "Ġsal": 4274, + "Ġsala": 57933, + "Ġsalad": 32466, + "Ġsalads": 76752, + "Ġsalah": 77324, + "Ġsalari": 143576, + "Ġsalaries": 36432, + "Ġsalario": 93091, + "Ġsalarié": 143577, + "Ġsalary": 16107, + "Ġsaldo": 73851, + "Ġsaldır": 135915, + "Ġsaldırı": 135916, + "Ġsale": 6278, + "Ġsales": 6625, + "Ġsalesman": 78447, + "Ġsalida": 67104, + "Ġsaline": 98007, + "Ġsalir": 78122, + "Ġsaliva": 84557, + "Ġsalle": 70532, + "Ġsalmon": 40320, + "Ġsalon": 39728, + "Ġsalope": 18393, + "Ġsalopes": 81322, + "Ġsalsa": 73214, + "Ġsalt": 12021, + "Ġsalts": 77135, + "Ġsalty": 73875, + "Ġsalud": 58012, + "Ġsalute": 70111, + "Ġsalv": 27059, + "Ġsalvage": 71736, + "Ġsalvar": 92956, + "Ġsalvation": 44935, + "Ġsam": 9962, + "Ġsama": 66623, + "Ġsame": 1852, + "Ġsamen": 81457, + "Ġsammen": 86279, + "Ġsamo": 82378, + "Ġsamp": 61481, + "Ġsample": 6077, + "Ġsampled": 48876, + "Ġsampler": 41799, + "Ġsamples": 10469, + "Ġsampling": 24892, + "Ġsamsung": 72984, + "Ġsamt": 75243, + "Ġsan": 8363, + "Ġsanct": 34489, + "Ġsanction": 44261, + "Ġsanctioned": 73244, + "Ġsanctions": 23746, + "Ġsanctuary": 50539, + "Ġsand": 9278, + "Ġsandals": 80257, + "Ġsandbox": 42754, + "Ġsands": 69063, + "Ġsandwich": 27874, + "Ġsandwiches": 56658, + "Ġsandy": 67439, + "Ġsane": 61100, + "Ġsang": 28240, + "Ġsangat": 70464, + "Ġsanit": 19812, + "Ġsanitary": 94529, + "Ġsanitation": 76523, + "Ġsanitize": 45183, + "Ġsanitized": 39232, + "Ġsanitizer": 98096, + "Ġsanity": 46842, + "Ġsank": 64230, + "Ġsans": 15510, + "Ġsant": 51508, + "Ġsanta": 71751, + "Ġsanté": 76223, + "Ġsap": 34635, + "Ġsapi": 81195, + "Ġsar": 28047, + "Ġsarc": 78766, + "Ġsarcast": 82267, + "Ġsare": 77345, + "ĠsarÃł": 79044, + "Ġsass": 74853, + "Ġsat": 7578, + "Ġsatellite": 23210, + "Ġsatellites": 46610, + "Ġsatin": 77084, + "Ġsatire": 81395, + "Ġsatisf": 8625, + "Ġsatisfaction": 23689, + "Ġsatisfactory": 57731, + "Ġsatisfied": 19527, + "Ġsatisfies": 67901, + "Ġsatisfy": 26553, + "Ġsatisfying": 36054, + "Ġsatu": 58566, + "Ġsatur": 93477, + "Ġsaturated": 49485, + "Ġsaturation": 49743, + "Ġsaturn": 84689, + "Ġsatın": 137747, + "ĠsatÄ±ÅŁ": 133980, + "Ġsau": 32154, + "Ġsauce": 19187, + "Ġsauces": 89970, + "Ġsauna": 46858, + "Ġsausage": 58886, + "Ġsav": 11179, + "Ġsavage": 72035, + "ĠsavaÅŁ": 133629, + "Ġsave": 3581, + "Ġsaved": 6781, + "ĠsavedInstanceState": 14724, + "Ġsaver": 60162, + "Ġsaves": 25935, + "Ġsaving": 13997, + "Ġsavings": 18986, + "Ġsavoir": 46829, + "Ġsavory": 93660, + "Ġsavvy": 63728, + "Ġsaw": 5485, + "Ġsax": 63008, + "Ġsay": 1977, + "Ġsaya": 53937, + "Ġsaygı": 140758, + "Ġsaying": 5488, + "Ġsays": 2727, + "Ġsayı": 131404, + "Ġsayılı": 137714, + "Ġsayısı": 134052, + "Ġsaç": 135271, + "Ġsaúde": 96895, + "ĠsaÄŁ": 84839, + "ĠsaÄŁlam": 128048, + "ĠsaÄŁlamak": 140065, + "ĠsaÄŁlan": 136445, + "ĠsaÄŁlayan": 142669, + "ĠsaÄŁlık": 133874, + "ĠsaÄŁlıkl": 140695, + "ĠsaÄŁlıklı": 140696, + "Ġsb": 7898, + "Ġsc": 1136, + "Ġsca": 15350, + "Ġscaff": 56150, + "Ġscaffold": 84452, + "Ġscal": 24004, + "Ġscala": 15771, + "Ġscalability": 93740, + "Ġscalable": 68211, + "Ġscalar": 17274, + "Ġscale": 5452, + "ĠscaleFactor": 84819, + "ĠscaleX": 92618, + "ĠscaleY": 90333, + "Ġscaled": 30690, + "Ġscaler": 68724, + "Ġscales": 28405, + "Ġscaling": 26943, + "Ġscalp": 64326, + "Ġscam": 34626, + "Ġscams": 63055, + "Ġscan": 8569, + "Ġscand": 54292, + "Ġscandal": 25599, + "Ġscandals": 68029, + "Ġscanf": 16865, + "Ġscanned": 47448, + "Ġscanner": 20775, + "Ġscanners": 94046, + "Ġscanning": 35101, + "Ġscans": 42639, + "Ġscant": 83855, + "Ġscape": 98013, + "Ġscar": 22290, + "Ġscarc": 51315, + "Ġscarce": 58190, + "Ġscarcely": 79887, + "Ġscarcity": 81384, + "Ġscare": 42930, + "Ġscared": 26115, + "Ġscares": 96371, + "Ġscarf": 67271, + "Ġscars": 60599, + "Ġscary": 28465, + "Ġscatter": 44477, + "Ġscattered": 36967, + "Ġscattering": 71816, + "Ġscav": 82650, + "Ġsce": 93810, + "Ġscen": 7529, + "Ġscenario": 15048, + "Ġscenarios": 25283, + "Ġscene": 6109, + "Ġscenery": 50231, + "Ġscenes": 16065, + "Ġscenic": 60981, + "Ġscent": 40366, + "Ġscept": 66351, + "Ġsch": 5699, + "Ġsche": 70251, + "Ġsched": 5575, + "Ġschedule": 9700, + "Ġscheduled": 13537, + "Ġscheduler": 28809, + "Ġschedules": 36769, + "Ġscheduling": 37852, + "Ġschem": 19650, + "Ġschema": 10802, + "Ġschemas": 61800, + "Ġschematic": 67044, + "Ġscheme": 12859, + "Ġschemes": 30856, + "Ġschizophren": 50181, + "Ġschizophrenia": 57433, + "Ġschl": 71417, + "Ġschle": 84187, + "Ġschn": 41130, + "Ġschnell": 58052, + "Ġscho": 77040, + "Ġscholar": 18145, + "Ġscholarly": 62581, + "Ġscholars": 30739, + "Ġscholarship": 33125, + "Ġscholarships": 54755, + "Ġschon": 28086, + "Ġschool": 2906, + "Ġschooling": 67621, + "Ġschools": 8682, + "Ġschw": 35515, + "Ġschwar": 81828, + "Ġschwer": 83758, + "Ġschö": 58459, + "Ġschön": 74764, + "Ġschöne": 91896, + "Ġsci": 37974, + "Ġscience": 8038, + "Ġsciences": 35688, + "Ġscient": 7182, + "Ġscientific": 12344, + "Ġscientifically": 73547, + "Ġscientist": 27468, + "Ġscientists": 13923, + "Ġscipy": 28090, + "Ġscissors": 71624, + "Ġscl": 56119, + "Ġsclerosis": 90257, + "Ġscm": 85520, + "Ġscn": 98988, + "Ġsco": 24637, + "Ġscoff": 89556, + "Ġscoop": 56407, + "Ġscooter": 75040, + "Ġscop": 47521, + "Ġscope": 6891, + "Ġscoped": 46960, + "Ġscopes": 50598, + "Ġscor": 61095, + "Ġscore": 5456, + "Ġscoreboard": 82329, + "Ġscored": 16548, + "Ġscorer": 63669, + "Ġscores": 12205, + "Ġscoring": 19826, + "Ġscorn": 87006, + "Ġscour": 69255, + "Ġscout": 53494, + "Ġscouting": 66380, + "Ġscouts": 69871, + "Ġscp": 69657, + "Ġscr": 9224, + "Ġscram": 34961, + "Ġscramble": 76287, + "Ġscrambled": 70478, + "Ġscrambling": 84227, + "Ġscrap": 20843, + "Ġscrape": 57128, + "Ġscraped": 87837, + "Ġscraper": 88034, + "Ġscraping": 81438, + "Ġscrapped": 83247, + "Ġscraps": 91785, + "Ġscrapy": 47734, + "Ġscratch": 18778, + "Ġscratched": 82833, + "Ġscratches": 65739, + "Ġscratching": 71919, + "Ġscre": 7862, + "Ġscream": 45689, + "Ġscreamed": 60366, + "Ġscreaming": 34001, + "Ġscreams": 60008, + "Ġscreen": 4171, + "ĠscreenHeight": 77555, + "ĠscreenSize": 98695, + "ĠscreenWidth": 70370, + "Ġscreened": 57577, + "Ġscreening": 22265, + "Ġscreenings": 83506, + "Ġscreenplay": 84775, + "Ġscreens": 15310, + "Ġscreenshot": 36190, + "Ġscreenshots": 48720, + "Ġscrew": 21966, + "Ġscrewed": 56345, + "Ġscrews": 44922, + "Ġscri": 67222, + "Ġscrim": 80557, + "Ġscrimmage": 98176, + "Ġscript": 5316, + "Ġscripted": 85025, + "Ġscripting": 67422, + "Ġscripts": 19502, + "Ġscripture": 72788, + "Ġscriptures": 86633, + "Ġscroll": 9059, + "ĠscrollTo": 88295, + "ĠscrollTop": 64454, + "ĠscrollView": 44753, + "Ġscrollbar": 79655, + "Ġscrolled": 95619, + "Ġscrolling": 37976, + "Ġscrolls": 78564, + "Ġscrub": 44901, + "Ġscrut": 28917, + "Ġscrutin": 69242, + "Ġscrutiny": 35652, + "Ġsculpt": 26765, + "Ġsculpture": 49967, + "Ġsculptures": 74873, + "Ġscène": 138938, + "Ġsd": 20585, + "Ġsdf": 50157, + "Ġsdk": 45402, + "Ġsdl": 86330, + "Ġse": 511, + "Ġsea": 9396, + "Ġseab": 65491, + "Ġseaborn": 94760, + "Ġseafood": 52305, + "Ġseal": 25349, + "Ġsealed": 19046, + "Ġsealing": 65834, + "Ġseals": 56777, + "Ġseam": 26396, + "Ġseamless": 46870, + "Ġseamlessly": 60340, + "Ġseams": 73688, + "Ġsean": 84422, + "Ġsearch": 2711, + "ĠsearchBar": 71801, + "ĠsearchData": 59131, + "ĠsearchString": 94040, + "ĠsearchTerm": 89026, + "ĠsearchText": 72508, + "Ġsearchable": 73643, + "Ġsearched": 26502, + "Ġsearcher": 94674, + "Ġsearches": 26475, + "Ġsearching": 15039, + "Ġseas": 51740, + "Ġseaside": 92776, + "Ġseason": 3200, + "Ġseasonal": 35799, + "Ġseasoned": 51514, + "Ġseasoning": 86219, + "Ġseasons": 15584, + "Ġseat": 10723, + "Ġseated": 46313, + "Ġseating": 37299, + "Ġseats": 16312, + "Ġseaw": 66229, + "Ġsebagai": 58622, + "Ġsebuah": 95466, + "Ġsec": 5701, + "Ġsecara": 72327, + "Ġsecluded": 88015, + "Ġsecond": 2086, + "Ġsecondary": 14246, + "Ġsecondo": 85936, + "Ġseconds": 6486, + "Ġsecre": 19183, + "Ġsecrecy": 59959, + "Ġsecret": 6234, + "Ġsecretary": 19067, + "Ġsecretion": 85967, + "Ġsecretive": 82999, + "Ġsecretly": 41739, + "Ġsecrets": 23594, + "Ġsecs": 49749, + "Ġsect": 30137, + "Ġsectarian": 78780, + "Ġsection": 3772, + "Ġsectional": 92360, + "Ġsections": 14158, + "Ġsector": 10486, + "Ġsectors": 25512, + "Ġsecular": 35919, + "Ġsecure": 9767, + "Ġsecured": 26430, + "Ġsecurely": 51023, + "Ġsecuring": 40477, + "Ġsecurities": 33819, + "Ġsecurity": 4763, + "Ġsed": 10923, + "Ġsedan": 54740, + "Ġsede": 95587, + "Ġsediment": 58032, + "Ġsee": 1490, + "Ġseed": 10320, + "Ġseeded": 89816, + "Ġseeding": 94415, + "Ġseeds": 19056, + "Ġseedu": 95284, + "Ġseeing": 9120, + "Ġseek": 5931, + "Ġseeker": 89133, + "Ġseekers": 49333, + "Ġseeking": 10887, + "Ġseeks": 25654, + "Ġseem": 2803, + "Ġseemed": 9324, + "Ġseeming": 75695, + "Ġseemingly": 22660, + "Ġseems": 4977, + "Ġseen": 3884, + "Ġsees": 15634, + "Ġseg": 4810, + "Ġsegment": 10238, + "Ġsegmentation": 59752, + "Ġsegmented": 84945, + "Ġsegments": 20632, + "Ġsegreg": 43067, + "Ġsegregated": 91298, + "Ġsegregation": 63144, + "Ġsegu": 81302, + "Ġsegue": 36770, + "Ġseguint": 93072, + "Ġseguir": 57448, + "Ġsegunda": 54972, + "Ġsegundo": 44287, + "Ġsegundos": 89616, + "Ġsegurança": 133665, + "Ġseguridad": 63304, + "Ġseguro": 61027, + "Ġsegún": 62015, + "Ġseh": 91665, + "Ġsehen": 53835, + "Ġsehr": 25494, + "Ġsei": 42137, + "Ġsein": 19349, + "Ġseine": 38897, + "Ġseinem": 72440, + "Ġseinen": 59228, + "Ġseiner": 53936, + "Ġseins": 54328, + "Ġseis": 86412, + "Ġseismic": 71479, + "Ġseit": 41488, + "Ġseiz": 33113, + "Ġseize": 49985, + "Ġseized": 30489, + "Ġseizing": 97206, + "Ġseizure": 52150, + "Ġseizures": 54898, + "Ġseja": 52596, + "Ġsek": 43834, + "Ġseks": 17510, + "Ġseksi": 61782, + "Ġsektör": 132655, + "Ġsel": 11806, + "Ġselber": 84193, + "Ġselbst": 34542, + "Ġseldom": 55352, + "Ġsele": 27319, + "Ġseleccion": 47043, + "Ġselecion": 85286, + "Ġselect": 3293, + "Ġselectable": 81051, + "Ġselected": 4091, + "ĠselectedIndex": 83617, + "ĠselectedItem": 70992, + "Ġselecting": 26301, + "Ġselection": 6589, + "Ġselections": 37399, + "Ġselective": 42910, + "Ġselectively": 81675, + "Ġselector": 9367, + "Ġselectors": 56037, + "Ġselects": 49143, + "Ġselenium": 35945, + "Ġself": 656, + "Ġselfie": 64294, + "Ġselfies": 85297, + "Ġselfish": 44643, + "Ġsell": 4559, + "Ġseller": 22856, + "Ġsellers": 36149, + "Ġselling": 11236, + "Ġsells": 30778, + "Ġselon": 64635, + "Ġselv": 60251, + "Ġselves": 83037, + "Ġsem": 5234, + "Ġsemaine": 71609, + "Ġsemana": 44975, + "Ġsemanas": 96456, + "Ġsemantic": 41733, + "Ġsemantics": 52694, + "Ġsemaphore": 55918, + "Ġsembl": 79564, + "Ġsemble": 96387, + "Ġsemen": 74657, + "Ġsemester": 33153, + "Ġsemi": 18267, + "Ġsemiclass": 94277, + "Ġsemiclassical": 94278, + "Ġsemiconductor": 86736, + "Ġsemif": 65500, + "Ġseminal": 89673, + "Ġseminar": 53575, + "Ġseminars": 67469, + "Ġsemp": 74597, + "Ġsempre": 34562, + "Ġsemua": 77497, + "Ġsen": 6124, + "Ġsenate": 76370, + "Ġsenator": 31697, + "Ġsenators": 38920, + "Ġsend": 3624, + "ĠsendData": 92323, + "ĠsendMessage": 52358, + "Ġsender": 4646, + "Ġsending": 11628, + "Ġsendo": 61346, + "Ġsends": 21308, + "Ġsenha": 48802, + "Ġsenior": 9990, + "Ġseniors": 37249, + "Ġsens": 6097, + "Ġsensation": 36292, + "Ġsensational": 67403, + "Ġsensations": 71497, + "Ġsense": 5530, + "Ġsensed": 88442, + "Ġsenses": 37107, + "Ġsensible": 37661, + "Ġsensing": 59099, + "Ġsensit": 73533, + "Ġsensitive": 16216, + "Ġsensitivity": 26443, + "Ġsensor": 12002, + "Ġsensors": 25092, + "Ġsensory": 47969, + "Ġsensual": 59617, + "Ġsent": 3208, + "Ġsentence": 11652, + "Ġsentenced": 29131, + "Ġsentences": 22870, + "Ġsentencing": 50517, + "Ġsentido": 64384, + "Ġsentient": 97770, + "Ġsentiment": 25975, + "Ġsentimental": 91938, + "Ġsentiments": 57114, + "Ġsentinel": 80778, + "Ġsentir": 74702, + "Ġsenza": 41926, + "Ġseo": 90608, + "Ġsep": 21017, + "Ġsepar": 4836, + "Ġseparat": 61382, + "Ġseparate": 8651, + "Ġseparated": 18663, + "Ġseparately": 25156, + "Ġseparates": 61749, + "Ġseparating": 49445, + "Ġseparation": 24737, + "Ġseparator": 24792, + "Ġseparators": 82502, + "Ġseper": 48354, + "Ġseperate": 83883, + "Ġseperti": 67720, + "Ġsept": 34329, + "Ġseptember": 83649, + "Ġseptembre": 94643, + "Ġseq": 12981, + "Ġsequ": 11256, + "Ġsequel": 34761, + "Ġsequelize": 58785, + "Ġsequence": 8500, + "Ġsequences": 23700, + "Ġsequencing": 61019, + "Ġsequential": 51000, + "Ġsequentially": 94559, + "Ġser": 1420, + "Ġsera": 34637, + "Ġserait": 86164, + "Ġserde": 61570, + "Ġserene": 94763, + "Ġserge": 71201, + "Ġsergeant": 83121, + "Ġseri": 93346, + "Ġseria": 83365, + "Ġserial": 6146, + "ĠserialVersionUID": 23673, + "Ġserialization": 47443, + "Ġserialize": 24235, + "Ġserialized": 32916, + "Ġserializer": 21759, + "Ġserializers": 38221, + "Ġserie": 32878, + "Ġseries": 4013, + "Ġserif": 89748, + "Ġserious": 6001, + "Ġseriously": 13919, + "Ġseriousness": 74381, + "Ġsermon": 75382, + "Ġseront": 77771, + "Ġserotonin": 76030, + "Ġserpent": 91842, + "Ġserr": 61934, + "Ġserta": 92619, + "Ġserum": 40429, + "Ġserv": 4853, + "Ġservant": 40464, + "Ġservants": 41732, + "Ġserve": 8683, + "Ġserved": 10223, + "Ġserver": 3538, + "Ġservers": 16293, + "Ġserves": 17045, + "Ġservi": 54405, + "Ġservic": 81359, + "Ġservice": 2473, + "ĠserviceName": 69191, + "ĠserviceProvider": 98009, + "Ġserviced": 80123, + "Ġservices": 3516, + "Ġservicing": 64818, + "Ġservicio": 46157, + "Ġservicios": 52123, + "Ġservidor": 57226, + "Ġserving": 13480, + "Ġservings": 77949, + "Ġserviço": 90575, + "Ġserviços": 93743, + "Ġservlet": 36496, + "Ġservo": 58312, + "Ġserá": 32898, + "Ġserão": 98424, + "ĠserÃŃa": 136302, + "Ġses": 15537, + "Ġsesame": 91123, + "Ġsesión": 68745, + "Ġsess": 21875, + "Ġsession": 3797, + "ĠsessionFactory": 70476, + "ĠsessionId": 54154, + "ĠsessionStorage": 50143, + "Ġsessions": 15704, + "Ġsesso": 18249, + "Ġsessuali": 94869, + "Ġset": 738, + "ĠsetActive": 74859, + "ĠsetAddress": 98420, + "ĠsetBackground": 70655, + "ĠsetBackgroundColor": 79874, + "ĠsetBackgroundImage": 95266, + "ĠsetC": 84901, + "ĠsetColor": 75968, + "ĠsetContent": 68260, + "ĠsetContentView": 21109, + "ĠsetCurrent": 53914, + "ĠsetData": 42149, + "ĠsetDate": 74412, + "ĠsetDefaultCloseOperation": 89017, + "ĠsetDescription": 68325, + "ĠsetEmail": 53320, + "ĠsetError": 52215, + "ĠsetFrame": 85289, + "ĠsetHidden": 98415, + "ĠsetId": 26635, + "ĠsetImage": 39962, + "ĠsetInput": 92591, + "ĠsetInterval": 35479, + "ĠsetIs": 38398, + "ĠsetLoading": 48589, + "ĠsetLocation": 88332, + "ĠsetMessage": 66891, + "ĠsetName": 27780, + "ĠsetObject": 38983, + "ĠsetOpen": 66619, + "ĠsetPage": 82828, + "ĠsetPassword": 53115, + "ĠsetPosition": 91197, + "ĠsetResult": 82906, + "ĠsetSearch": 85682, + "ĠsetSelected": 45263, + "ĠsetShow": 66557, + "ĠsetSize": 83859, + "ĠsetState": 21450, + "ĠsetStatus": 56304, + "ĠsetSupportActionBar": 83048, + "ĠsetText": 44562, + "ĠsetTime": 89521, + "ĠsetTimeout": 14258, + "ĠsetTitle": 26226, + "ĠsetTitleColor": 77138, + "ĠsetType": 62709, + "ĠsetUp": 18620, + "ĠsetUser": 42413, + "ĠsetUsername": 86055, + "ĠsetValue": 28052, + "ĠsetVisible": 83252, + "Ġsetattr": 61114, + "Ġsetback": 78813, + "Ġsetbacks": 96637, + "Ġsets": 7289, + "Ġsett": 8227, + "Ġsetter": 40388, + "Ġsetters": 80842, + "Ġsetting": 6243, + "Ġsettings": 5003, + "Ġsettle": 24729, + "Ġsettled": 22378, + "Ġsettlement": 17079, + "Ġsettlements": 44604, + "Ġsettlers": 60007, + "Ġsettles": 91918, + "Ġsettling": 51845, + "Ġsetup": 6505, + "Ġsetups": 83723, + "Ġsetuptools": 63797, + "Ġsetw": 50063, + "Ġsetzen": 97696, + "Ġseu": 20009, + "Ġseul": 73106, + "Ġseule": 85706, + "Ġseulement": 70586, + "Ġseus": 41398, + "Ġseven": 8094, + "Ġsevent": 21373, + "Ġseventeen": 81397, + "Ġseventh": 30387, + "Ġseventy": 69949, + "Ġsever": 3205, + "Ġseveral": 3807, + "Ġsevere": 15386, + "Ġsevered": 83350, + "Ġseverely": 34806, + "Ġseverity": 29920, + "Ġsew": 22352, + "Ġsewage": 71117, + "Ġsewer": 61920, + "Ġsewing": 51219, + "Ġsex": 1839, + "Ġsexdate": 61902, + "Ġsexe": 15974, + "Ġsexes": 74932, + "Ġsexism": 67995, + "Ġsexist": 63782, + "Ġsexkontakte": 86278, + "Ġsexle": 71195, + "Ġsexo": 19616, + "Ġsext": 37003, + "Ġsextreffen": 96835, + "Ġsexual": 7244, + "Ġsexuales": 98113, + "Ġsexuality": 38175, + "Ġsexually": 26583, + "Ġsexy": 10886, + "Ġsez": 78472, + "Ġseznam": 94868, + "Ġseç": 125312, + "Ġseçen": 138155, + "Ġseçenek": 138156, + "Ġseçil": 141162, + "Ġseçim": 131544, + "Ġseñ": 54230, + "Ġseñal": 132456, + "Ġseñor": 140315, + "Ġsf": 13124, + "Ġsg": 30673, + "Ġsh": 557, + "Ġsha": 15870, + "Ġshack": 76290, + "Ġshade": 27501, + "Ġshaded": 91766, + "Ġshader": 21013, + "Ġshaders": 67550, + "Ġshades": 36099, + "Ġshading": 71734, + "Ġshadow": 12455, + "Ġshadows": 34512, + "Ġshady": 77081, + "Ġshaft": 30368, + "Ġshake": 26025, + "Ġshaken": 65800, + "Ġshakes": 62030, + "Ġshaking": 37739, + "Ġshaky": 83413, + "Ġshale": 74005, + "Ġshall": 4880, + "Ġshallow": 25600, + "Ġshalt": 88535, + "Ġsham": 71876, + "Ġshame": 20973, + "Ġshameful": 83194, + "Ġshampoo": 61932, + "Ġshape": 6083, + "Ġshaped": 26271, + "Ġshapes": 20816, + "Ġshaping": 45520, + "Ġshar": 25626, + "Ġshard": 52069, + "Ġshards": 74110, + "Ġshare": 4332, + "Ġshared": 6094, + "ĠsharedApplication": 60318, + "ĠsharedInstance": 51958, + "ĠsharedPreferences": 61057, + "Ġshareholder": 62242, + "Ġshareholders": 40677, + "Ġshares": 13248, + "Ġsharing": 11560, + "Ġshark": 43792, + "Ġsharks": 60435, + "Ġsharp": 17232, + "Ġsharpen": 56363, + "Ġsharper": 95469, + "Ġsharply": 45373, + "Ġshattered": 55367, + "Ġshave": 72607, + "Ġshaved": 65199, + "Ġshaving": 80009, + "Ġshe": 1340, + "Ġshear": 64244, + "Ġshed": 24354, + "Ġshedding": 79317, + "Ġsheds": 75939, + "Ġsheep": 31912, + "Ġsheer": 32538, + "Ġsheet": 10834, + "Ġsheets": 24140, + "Ġshel": 73035, + "Ġshelf": 27645, + "Ġshell": 12528, + "Ġshells": 42954, + "Ġshelter": 22906, + "Ġshelters": 51788, + "Ġshelves": 35210, + "Ġshemale": 33627, + "Ġshepherd": 90378, + "Ġsher": 34213, + "Ġsheriff": 39739, + "Ġshield": 20245, + "Ġshielding": 96676, + "Ġshields": 52665, + "Ġshift": 6407, + "Ġshifted": 28973, + "Ġshifting": 31831, + "Ġshifts": 28635, + "Ġshim": 62132, + "Ġshimmer": 75361, + "Ġshin": 76011, + "Ġshine": 32405, + "Ġshines": 64780, + "Ġshining": 47925, + "Ġshiny": 41199, + "Ġship": 8284, + "Ġshipment": 43754, + "Ġshipments": 58621, + "Ġshipped": 27259, + "Ġshipping": 11601, + "Ġships": 17727, + "Ġshirt": 15478, + "Ġshirts": 36724, + "Ġshit": 17176, + "Ġshitty": 77458, + "Ġshl": 95910, + "Ġshm": 76500, + "Ġsho": 71519, + "Ġshock": 10755, + "Ġshocked": 26620, + "Ġshocking": 33634, + "Ġshocks": 67160, + "Ġshoe": 28977, + "Ġshoes": 15294, + "Ġshook": 34914, + "Ġshoot": 6102, + "Ġshooter": 31571, + "Ġshooters": 62070, + "Ġshooting": 10441, + "Ġshootings": 43761, + "Ġshootout": 87181, + "Ġshoots": 43627, + "Ġshop": 8061, + "Ġshopper": 93203, + "Ġshoppers": 48735, + "Ġshopping": 11919, + "Ġshops": 19456, + "Ġshore": 30184, + "Ġshoreline": 98064, + "Ġshores": 62163, + "Ġshort": 2805, + "Ġshortage": 38159, + "Ġshortages": 66176, + "Ġshortcode": 75813, + "Ġshortcomings": 71921, + "Ġshortcut": 37115, + "Ġshortcuts": 54920, + "Ġshorten": 73381, + "Ġshortened": 65563, + "Ġshorter": 23327, + "Ġshortest": 39143, + "Ġshortfall": 99049, + "Ġshorthand": 78339, + "Ġshortly": 19620, + "Ġshorts": 35776, + "Ġshot": 6552, + "Ġshotgun": 51248, + "Ġshots": 14951, + "Ġshould": 1265, + "ĠshouldBe": 45483, + "Ġshoulder": 16882, + "Ġshoulders": 26906, + "Ġshouldn": 13133, + "Ġshout": 41123, + "Ġshouted": 43443, + "Ġshouting": 47555, + "Ġshouts": 83836, + "Ġshove": 84615, + "Ġshoved": 82098, + "Ġshovel": 91991, + "Ġshow": 1473, + "ĠshowAlert": 86189, + "ĠshowDialog": 75709, + "ĠshowError": 63003, + "ĠshowMessage": 81649, + "ĠshowModal": 94734, + "ĠshowToast": 81231, + "Ġshowc": 47210, + "Ġshowcase": 34783, + "Ġshowcased": 82760, + "Ġshowcases": 70749, + "Ġshowcasing": 66808, + "Ġshowdown": 71608, + "Ġshowed": 8542, + "Ġshower": 17196, + "Ġshowers": 51802, + "Ġshowing": 9027, + "Ġshown": 6839, + "Ġshowroom": 86135, + "Ġshows": 4933, + "Ġshr": 14035, + "Ġshred": 97762, + "Ġshredd": 52362, + "Ġshredded": 74120, + "Ġshri": 96846, + "Ġshrimp": 55526, + "Ġshrine": 73548, + "Ġshrink": 28900, + "Ġshrinking": 62085, + "Ġshrugged": 69656, + "Ġshuffle": 25947, + "Ġshuffled": 74271, + "Ġshut": 9311, + "Ġshutdown": 23766, + "Ġshutil": 33960, + "Ġshuts": 88578, + "Ġshutter": 47956, + "Ġshutting": 50026, + "Ġshuttle": 44541, + "Ġshy": 32294, + "Ġsi": 4403, + "Ġsia": 49889, + "Ġsibling": 44223, + "Ġsiblings": 36683, + "Ġsic": 52647, + "Ġsich": 9089, + "Ġsicher": 54548, + "Ġsick": 14036, + "Ġsickness": 53850, + "Ġsid": 14084, + "Ġside": 3108, + "Ġsidebar": 27226, + "Ġsided": 67936, + "Ġsidel": 85077, + "Ġsideline": 78953, + "Ġsidelined": 97948, + "Ġsidelines": 69413, + "Ġsiden": 86816, + "Ġsider": 86574, + "Ġsides": 11067, + "Ġsidew": 31517, + "Ġsidewalk": 41090, + "Ġsidewalks": 81735, + "Ġsideways": 73729, + "Ġsiding": 75685, + "Ġsidl": 85791, + "Ġsido": 33971, + "Ġsie": 9911, + "Ġsiege": 51691, + "Ġsieht": 86899, + "Ġsiempre": 41598, + "Ġsiendo": 81266, + "Ġsieve": 74336, + "Ġsift": 80744, + "Ġsig": 8366, + "Ġsigh": 30138, + "Ġsighed": 52814, + "Ġsight": 13929, + "Ġsighting": 92679, + "Ġsightings": 82654, + "Ġsights": 41166, + "Ġsigma": 20254, + "Ġsigmoid": 64890, + "Ġsign": 1841, + "ĠsignIn": 50995, + "ĠsignUp": 71030, + "Ġsignage": 79080, + "Ġsignal": 8286, + "Ġsignaled": 74748, + "Ġsignaling": 41980, + "Ġsignalling": 90877, + "Ġsignals": 17289, + "Ġsignature": 11957, + "Ġsignatures": 32628, + "Ġsigned": 8499, + "Ġsigner": 70039, + "Ġsignific": 4595, + "Ġsignifica": 67732, + "Ġsignificance": 25361, + "Ġsignificant": 5089, + "Ġsignificantly": 11941, + "Ġsignifies": 77764, + "Ġsignify": 88422, + "Ġsignin": 87992, + "Ġsigning": 15971, + "Ġsignings": 98556, + "Ġsigns": 11929, + "Ġsignup": 40048, + "Ġsigu": 41645, + "Ġsigue": 96229, + "Ġsiguiente": 54901, + "Ġsiguientes": 85861, + "Ġsik": 75461, + "Ġsil": 5437, + "Ġsildenafil": 91801, + "Ġsilence": 21162, + "Ġsilenced": 91681, + "Ġsilent": 21059, + "Ġsilently": 49394, + "Ġsilhouette": 56727, + "Ġsilica": 76789, + "Ġsilicon": 50592, + "Ġsilicone": 52925, + "Ġsilk": 39944, + "Ġsilky": 86318, + "Ġsill": 84267, + "Ġsilly": 29471, + "Ġsilver": 14961, + "Ġsim": 1643, + "Ġsimd": 92257, + "Ġsimil": 91792, + "Ġsimilar": 4428, + "Ġsimilarities": 42774, + "Ġsimilarity": 37623, + "Ġsimilarly": 29193, + "Ġsimmer": 61815, + "Ġsimp": 12432, + "Ġsimpl": 15491, + "Ġsimple": 4285, + "Ġsimplement": 70257, + "Ġsimpler": 34288, + "Ġsimples": 68306, + "Ġsimplest": 44548, + "Ġsimplex": 91651, + "Ġsimplicity": 38975, + "Ġsimplified": 43799, + "Ġsimplify": 39721, + "Ġsimplistic": 77530, + "Ġsimply": 4936, + "Ġsims": 90913, + "Ġsimul": 75673, + "Ġsimulate": 37453, + "Ġsimulated": 45736, + "Ġsimulation": 19038, + "Ġsimulations": 46490, + "Ġsimulator": 41891, + "Ġsimult": 20127, + "Ġsimultaneous": 57532, + "Ġsimultaneously": 24303, + "Ġsin": 7437, + "Ġsina": 96263, + "Ġsinc": 26241, + "Ġsince": 2474, + "Ġsincer": 43049, + "Ġsincere": 48324, + "Ġsincerely": 58283, + "Ġsincerity": 99126, + "Ġsind": 12583, + "Ġsine": 57668, + "Ġsinful": 96627, + "Ġsing": 7780, + "Ġsingapore": 95762, + "Ġsinger": 22757, + "Ġsingers": 67041, + "Ġsinging": 25083, + "Ġsingle": 3175, + "Ġsingled": 78873, + "Ġsingles": 17389, + "Ġsingleton": 36903, + "Ġsingly": 85499, + "Ġsings": 67241, + "Ġsingular": 33944, + "Ġsinh": 97832, + "Ġsinister": 62567, + "Ġsink": 19309, + "Ġsinking": 61093, + "Ġsinks": 56952, + "Ġsino": 51804, + "Ġsinon": 36472, + "Ġsins": 39344, + "Ġsint": 42229, + "Ġsinus": 75814, + "Ġsip": 46615, + "Ġsir": 27048, + "Ġsire": 85911, + "Ġsis": 52315, + "Ġsist": 20896, + "Ġsistem": 68886, + "Ġsistema": 28828, + "Ġsistemas": 88149, + "Ġsister": 12923, + "Ġsisters": 29293, + "Ġsit": 2444, + "Ġsitcom": 80366, + "Ġsite": 2747, + "Ġsites": 6594, + "Ġsiti": 42692, + "Ġsitio": 46956, + "Ġsitios": 70561, + "Ġsito": 50619, + "Ġsits": 23011, + "Ġsitting": 11699, + "Ġsitu": 9908, + "Ġsituación": 78295, + "Ġsituated": 30083, + "Ġsituation": 6534, + "Ġsituations": 14740, + "Ġsituação": 136578, + "Ġsitué": 136435, + "Ġsix": 4743, + "Ġsixteen": 58039, + "Ġsixth": 25031, + "Ġsixty": 52374, + "Ġsiz": 64896, + "Ġsizable": 78098, + "Ġsize": 1379, + "Ġsized": 29287, + "Ġsizeof": 3935, + "Ġsizes": 12282, + "Ġsizing": 63001, + "Ġsiè": 96639, + "Ġsiècle": 136971, + "Ġsiêu": 132539, + "ĠsiÄĻ": 12662, + "Ġsj": 37436, + "Ġsk": 1901, + "Ġska": 53727, + "Ġskal": 41813, + "Ġskate": 42778, + "Ġskateboard": 97982, + "Ġskating": 64404, + "Ġskb": 67823, + "Ġske": 15722, + "Ġskeletal": 68297, + "Ġskeleton": 29435, + "Ġskeletons": 74307, + "Ġskept": 36579, + "Ġskeptic": 54850, + "Ġskeptical": 43829, + "Ġskepticism": 66133, + "Ġsketch": 25529, + "Ġsketches": 67682, + "Ġskew": 40543, + "Ġskewed": 86472, + "Ġski": 28679, + "Ġskies": 49293, + "Ġskiing": 62017, + "Ġskill": 9947, + "Ġskilled": 25530, + "Ġskillet": 77985, + "Ġskills": 7361, + "Ġskim": 78569, + "Ġskimage": 92276, + "Ġskin": 6787, + "Ġskincare": 79605, + "Ġskinny": 47677, + "Ġskins": 49950, + "Ġskip": 10706, + "Ġskipped": 35157, + "Ġskipping": 42659, + "Ġskips": 91520, + "Ġskirm": 95280, + "Ġskirt": 37280, + "Ġskirts": 76205, + "Ġskl": 82542, + "Ġsklearn": 17987, + "Ġsku": 52322, + "Ġskull": 34013, + "Ġskulle": 69258, + "Ġskulls": 97612, + "Ġsky": 12884, + "Ġskyline": 87739, + "Ġskype": 86672, + "Ġskyrocket": 74622, + "Ġskys": 84321, + "Ġskóry": 137772, + "ĠskÅĤ": 137941, + "ĠskÅĤad": 131185, + "ĠskÅĤada": 137942, + "Ġsl": 1739, + "Ġsla": 31893, + "Ġslab": 54691, + "Ġslack": 45619, + "Ġslag": 66591, + "Ġslain": 69186, + "Ġslam": 44892, + "Ġslammed": 49265, + "Ġslamming": 91860, + "Ġslander": 98458, + "Ġslang": 79912, + "Ġslap": 50052, + "Ġslapped": 71788, + "Ġslash": 36626, + "Ġslashed": 83524, + "Ġslashes": 87396, + "Ġslashing": 91943, + "Ġslate": 50580, + "Ġslated": 52270, + "Ġslaughter": 39756, + "Ġslaughtered": 85747, + "Ġslave": 20362, + "Ġslavery": 32692, + "Ġslaves": 31660, + "Ġsle": 19747, + "Ġsled": 94412, + "Ġslee": 21661, + "Ġsleek": 47394, + "Ġsleep": 6084, + "Ġsleeper": 82547, + "Ġsleeping": 21127, + "Ġsleeps": 71390, + "Ġsleepy": 81793, + "Ġsleeve": 36153, + "Ġsleeves": 42375, + "Ġslender": 72242, + "Ġslept": 45398, + "Ġslew": 59032, + "Ġslic": 21652, + "Ġslice": 15983, + "Ġsliced": 47615, + "Ġslices": 34254, + "Ġslicing": 84915, + "Ġslick": 49638, + "Ġslid": 59914, + "Ġslide": 14983, + "Ġslider": 21412, + "Ġsliders": 81050, + "Ġslides": 21521, + "Ġslideshow": 64447, + "Ġsliding": 33832, + "Ġslight": 8112, + "Ġslightest": 59047, + "Ġslightly": 10078, + "Ġslik": 75224, + "Ġslim": 29353, + "Ġslime": 77844, + "Ġsling": 87633, + "Ġslip": 21134, + "Ġslipped": 42478, + "Ġslippery": 70572, + "Ġslipping": 63702, + "Ġslips": 67654, + "Ġslit": 78030, + "Ġslo": 17603, + "Ġslog": 34949, + "Ġslogan": 51537, + "Ġslogans": 77001, + "Ġslope": 30232, + "Ġslopes": 59008, + "Ġsloppy": 73982, + "Ġslot": 9446, + "Ġslots": 15711, + "Ġslow": 6301, + "Ġslowdown": 73917, + "Ġslowed": 47846, + "Ġslower": 28393, + "Ġslowing": 47308, + "Ġslowly": 13970, + "Ġslows": 69170, + "Ġslu": 98455, + "Ġslug": 24470, + "Ġslugg": 79022, + "Ġsluggish": 86946, + "Ġslump": 85728, + "Ġslur": 83571, + "Ġslut": 38397, + "Ġsluts": 57595, + "Ġsm": 1525, + "Ġsmack": 90494, + "Ġsmall": 2613, + "Ġsmaller": 9155, + "Ġsmallest": 24632, + "Ġsmart": 7785, + "Ġsmarter": 46478, + "Ġsmartphone": 21511, + "Ġsmartphones": 35022, + "Ġsmarty": 76048, + "Ġsmash": 51906, + "Ġsmashed": 56420, + "Ġsmashing": 82488, + "Ġsmb": 77792, + "Ġsme": 90467, + "Ġsmear": 80503, + "Ġsmell": 22068, + "Ġsmelled": 85698, + "Ġsmelling": 96003, + "Ġsmells": 49637, + "Ġsmile": 15289, + "Ġsmiled": 30545, + "Ġsmiles": 42532, + "Ġsmiling": 36063, + "Ġsmirk": 85024, + "Ġsmith": 76721, + "Ġsmo": 43573, + "Ġsmoke": 16205, + "Ġsmoked": 47537, + "Ġsmoker": 77220, + "Ġsmokers": 56657, + "Ġsmoking": 19578, + "Ġsmooth": 10876, + "Ġsmoothed": 92839, + "Ġsmoother": 60305, + "Ġsmoothing": 61961, + "Ġsmoothly": 38411, + "Ġsms": 37929, + "Ġsmtp": 54397, + "Ġsmugg": 40737, + "Ġsmuggling": 72934, + "ĠsmÃ¥": 82844, + "Ġsn": 4131, + "Ġsnack": 39359, + "Ġsnacks": 42302, + "Ġsnag": 71802, + "Ġsnake": 25265, + "Ġsnakes": 56096, + "Ġsnap": 10658, + "Ġsnapchat": 98050, + "Ġsnapped": 47010, + "Ġsnapping": 79789, + "Ġsnaps": 51304, + "Ġsnapshot": 16295, + "Ġsnapshots": 61823, + "Ġsnatch": 84560, + "Ġsnd": 28612, + "Ġsne": 20760, + "Ġsneak": 44164, + "Ġsneakers": 67191, + "Ġsnel": 89711, + "Ġsniff": 57287, + "Ġsniper": 62282, + "Ġsnippet": 43065, + "Ġsnippets": 68642, + "Ġsno": 65084, + "Ġsnork": 96706, + "Ġsnow": 11794, + "Ġsnowy": 89773, + "Ġsnprintf": 36446, + "Ġsns": 50101, + "Ġsnug": 85601, + "Ġso": 773, + "Ġsoak": 60692, + "Ġsoaked": 68875, + "Ġsoaking": 77346, + "Ġsoap": 26785, + "Ġsoar": 98465, + "Ġsoared": 85955, + "Ġsoaring": 68897, + "Ġsob": 11493, + "Ġsober": 49405, + "Ġsobie": 76258, + "Ġsobre": 15131, + "ĠsobÄħ": 138495, + "Ġsoc": 12019, + "Ġsoccer": 22174, + "Ġsoci": 15610, + "Ġsocial": 3590, + "Ġsociale": 74007, + "Ġsociales": 64107, + "Ġsocialism": 50518, + "Ġsocialist": 40189, + "Ġsocially": 39318, + "Ġsociedad": 89687, + "Ġsocietal": 58429, + "Ġsocieties": 33675, + "Ġsociety": 8232, + "ĠsocietÃł": 132926, + "Ġsocio": 40489, + "Ġsocioeconomic": 79331, + "Ġsociology": 77108, + "Ġsociété": 76724, + "Ġsock": 11087, + "Ġsockaddr": 24169, + "Ġsocket": 7575, + "Ġsockets": 47557, + "Ġsockfd": 80773, + "Ġsocks": 38986, + "Ġsod": 23952, + "Ġsoda": 38862, + "Ġsodium": 38595, + "Ġsodom": 82637, + "Ġsof": 85806, + "Ġsofa": 31069, + "Ġsofas": 96406, + "Ġsofort": 76720, + "Ġsoft": 8413, + "Ġsoftball": 95268, + "Ġsoften": 79945, + "Ġsoftened": 89058, + "Ġsofter": 63130, + "Ġsoftly": 56050, + "Ġsoftmax": 76369, + "Ġsoftware": 3162, + "Ġsog": 44842, + "Ġsogar": 71765, + "Ġsoil": 17171, + "Ġsoils": 68461, + "Ġsoir": 56892, + "Ġsoirée": 142434, + "Ġsoit": 44341, + "Ġsok": 83726, + "Ġsol": 2048, + "Ġsola": 95089, + "Ġsolar": 12941, + "Ġsold": 6088, + "Ġsolder": 59777, + "Ġsoldier": 26110, + "Ġsoldiers": 14856, + "Ġsole": 13309, + "Ġsolely": 21063, + "Ġsolemn": 63602, + "Ġsolic": 29744, + "Ġsolicit": 55146, + "Ġsolicitud": 95857, + "Ġsolid": 6437, + "Ġsolidarity": 43154, + "Ġsolidity": 72163, + "Ġsolids": 81386, + "Ġsolitary": 57129, + "Ġsolitude": 98444, + "Ġsoll": 22589, + "Ġsollen": 81146, + "Ġsollte": 50926, + "Ġsollten": 74453, + "Ġsolo": 13529, + "Ġsoluble": 86945, + "Ġsolución": 142706, + "Ġsolution": 6291, + "Ġsolutions": 9904, + "Ġsolução": 143396, + "Ġsolve": 11625, + "Ġsolved": 27956, + "Ġsolvent": 68896, + "Ġsolver": 28961, + "Ġsolves": 67477, + "Ġsolving": 21828, + "Ġsom": 1757, + "Ġsoma": 57759, + "Ġsome": 1045, + "Ġsomebody": 18079, + "Ġsomeday": 53008, + "Ġsomehow": 16925, + "Ġsomeone": 4325, + "Ġsomething": 2494, + "Ġsometime": 35013, + "Ġsometimes": 7025, + "Ġsomew": 9002, + "Ġsomewhat": 14400, + "Ġsomewhere": 14696, + "Ġsommes": 73433, + "Ġson": 4438, + "Ġsond": 46837, + "Ġsondern": 51290, + "Ġsong": 5492, + "Ġsongs": 11674, + "Ġsongwriter": 84657, + "Ġsonic": 71336, + "Ġsono": 20821, + "Ġsonra": 84126, + "Ġsonras": 138857, + "Ġsonrası": 138858, + "Ġsonrasında": 140820, + "Ġsons": 25350, + "Ġsonst": 95243, + "Ġsont": 14789, + "Ġsonuç": 131900, + "Ġsoo": 43176, + "Ġsoon": 5135, + "Ġsooner": 30273, + "Ġsoothing": 64364, + "Ġsop": 70204, + "Ġsoph": 18701, + "Ġsophistic": 23984, + "Ġsophisticated": 26779, + "Ġsophistication": 84436, + "Ġsophomore": 50398, + "Ġsopr": 79129, + "Ġsor": 24534, + "Ġsore": 35266, + "Ġsorrow": 57496, + "Ġsorry": 14589, + "Ġsort": 3378, + "ĠsortBy": 66913, + "ĠsortOrder": 78651, + "Ġsortable": 56775, + "Ġsorte": 79344, + "Ġsorted": 10615, + "Ġsorter": 90452, + "Ġsortie": 83043, + "Ġsorting": 28273, + "Ġsorts": 20853, + "Ġsos": 60877, + "Ġsost": 77948, + "Ġsotto": 95241, + "Ġsou": 5822, + "Ġsought": 16105, + "Ġsouha": 76338, + "Ġsoul": 13527, + "Ġsouls": 34726, + "Ġsound": 5112, + "Ġsounded": 34308, + "Ġsounding": 48127, + "Ġsounds": 10362, + "Ġsoundtrack": 44899, + "Ġsoup": 19174, + "Ġsour": 20282, + "Ġsource": 2530, + "ĠsourceMapping": 44584, + "ĠsourceMappingURL": 44695, + "Ġsourced": 41111, + "Ġsources": 8173, + "Ġsourcing": 73181, + "Ġsous": 31492, + "Ġsout": 64727, + "Ġsouth": 9806, + "Ġsoutheast": 41452, + "Ġsoutheastern": 82109, + "Ġsouthern": 18070, + "Ġsouthwest": 52985, + "Ġsouthwestern": 98811, + "Ġsouvenir": 97295, + "Ġsouvent": 70051, + "Ġsovere": 28997, + "Ġsovereign": 45284, + "Ġsovereignty": 45959, + "Ġsow": 57465, + "Ġsowie": 39307, + "Ġsoy": 28739, + "Ġsoát": 134814, + "Ġsp": 978, + "Ġspa": 30393, + "Ġspac": 99008, + "Ġspace": 3550, + "Ġspacecraft": 41540, + "Ġspaced": 63828, + "Ġspacer": 78849, + "Ġspaces": 12621, + "Ġspaceship": 84842, + "Ġspacing": 25943, + "Ġspacious": 32136, + "Ġspaghetti": 86910, + "Ġspam": 25328, + "Ġspan": 9390, + "Ġspanish": 77032, + "Ġspanking": 75735, + "Ġspanning": 55786, + "Ġspans": 44295, + "Ġspar": 79869, + "Ġspare": 23863, + "Ġspared": 73588, + "Ġsparing": 94955, + "Ġspark": 15186, + "Ġsparked": 40444, + "Ġsparking": 98982, + "Ġsparkle": 95070, + "Ġsparkling": 63512, + "Ġsparks": 75783, + "Ġsparse": 33444, + "Ġspas": 81960, + "Ġspat": 62883, + "Ġspatial": 27979, + "Ġspawn": 18042, + "Ġspawned": 52534, + "Ġspawning": 78429, + "Ġspawns": 92676, + "Ġspd": 81462, + "Ġspe": 2341, + "Ġspeak": 6468, + "Ġspeaker": 18601, + "Ġspeakers": 21326, + "Ġspeaking": 12094, + "Ġspeaks": 21191, + "Ġspear": 40863, + "Ġspec": 1398, + "Ġspecial": 3281, + "Ġspecialised": 87904, + "Ġspecialist": 23753, + "Ġspecialists": 34316, + "Ġspecialization": 65879, + "Ġspecialize": 47344, + "Ġspecialized": 27076, + "Ġspecializes": 45572, + "Ġspecializing": 57294, + "Ġspecially": 34326, + "Ġspecials": 59774, + "Ġspecialties": 93336, + "Ġspecialty": 35874, + "Ġspecies": 9419, + "Ġspecific": 3151, + "Ġspecifically": 11689, + "Ġspecification": 25128, + "Ġspecifications": 28703, + "Ġspecificity": 74941, + "Ġspecifics": 48349, + "Ġspecified": 5189, + "Ġspecifier": 97616, + "Ġspecifies": 29102, + "Ġspecify": 13837, + "Ġspecifying": 37838, + "Ġspecimen": 57084, + "Ġspecimens": 56649, + "Ġspecs": 32247, + "Ġspect": 9431, + "Ġspectacle": 56791, + "Ġspectacular": 27709, + "Ġspectator": 88623, + "Ġspectators": 69302, + "Ġspectra": 62597, + "Ġspectral": 55977, + "Ġspectro": 65325, + "Ġspectrum": 19745, + "Ġspecular": 85417, + "Ġspeculate": 63501, + "Ġspeculated": 73777, + "Ġspeculation": 32322, + "Ġspeculative": 65736, + "Ġsped": 85610, + "Ġspeech": 8806, + "Ġspeeches": 43340, + "Ġspeed": 4628, + "Ġspeeding": 57575, + "Ġspeeds": 24722, + "Ġspeedy": 65483, + "Ġspel": 70935, + "Ġspell": 12845, + "Ġspelled": 67818, + "Ġspelling": 42429, + "Ġspells": 25619, + "Ġspend": 8329, + "Ġspender": 83888, + "Ġspending": 10164, + "Ġspends": 37102, + "Ġspent": 7391, + "Ġspep": 69236, + "Ġsper": 46016, + "Ġsperm": 39599, + "Ġsperma": 73368, + "Ġsph": 41722, + "Ġsphere": 25366, + "Ġspheres": 65718, + "Ġspherical": 64151, + "Ġspi": 26949, + "Ġspice": 41686, + "Ġspices": 48559, + "Ġspicy": 47251, + "Ġspider": 34354, + "Ġspiders": 62136, + "Ġspiel": 78152, + "Ġspiele": 91855, + "Ġspielen": 58237, + "Ġspies": 68898, + "Ġspike": 36293, + "Ġspiked": 92242, + "Ġspikes": 50660, + "Ġspill": 38797, + "Ġspilled": 73025, + "Ġspills": 82951, + "Ġspin": 12616, + "Ġspinach": 67698, + "Ġspinal": 49012, + "Ġspindle": 98191, + "Ġspine": 34676, + "Ġspinner": 35491, + "Ġspinning": 37860, + "Ġspins": 44758, + "Ġspir": 18318, + "Ġspiral": 41097, + "Ġspirit": 8913, + "Ġspirited": 97043, + "Ġspirits": 30639, + "Ġspiritual": 17849, + "Ġspirituality": 74356, + "Ġspiritually": 87753, + "Ġspit": 55987, + "Ġspite": 33681, + "Ġspl": 12503, + "Ġsplash": 34632, + "Ġsple": 86350, + "Ġsplendid": 69860, + "Ġsplice": 88290, + "Ġspline": 70086, + "Ġsplit": 6718, + "Ġsplits": 40467, + "Ġsplitted": 84889, + "Ġsplitter": 72986, + "Ġsplitting": 44373, + "Ġspo": 19451, + "Ġspoil": 64793, + "Ġspoiled": 71424, + "Ġspoiler": 72206, + "Ġspoilers": 65327, + "Ġspoke": 12290, + "Ġspoken": 21355, + "Ġspokes": 11200, + "Ġspokesman": 17885, + "Ġspokesperson": 27903, + "Ġspokeswoman": 33141, + "Ġsponge": 68348, + "Ġspons": 13461, + "Ġsponsor": 24434, + "Ġsponsored": 29538, + "Ġsponsoring": 84341, + "Ġsponsors": 38601, + "Ġsponsorship": 54431, + "Ġspont": 36291, + "Ġspontaneous": 53457, + "Ġspontaneously": 87458, + "Ġspoof": 87890, + "Ġspooky": 97076, + "Ġspoon": 45505, + "Ġspor": 60916, + "Ġsport": 10552, + "Ġsporting": 34927, + "Ġsports": 9833, + "Ġspos": 53407, + "Ġsposób": 98621, + "Ġspot": 7702, + "Ġspotify": 87790, + "Ġspotlight": 36873, + "Ġspots": 18773, + "Ġspotted": 29552, + "Ġspotting": 89219, + "Ġspouse": 30980, + "Ġspouses": 65212, + "Ġspp": 90699, + "Ġspr": 8151, + "Ġspraw": 34808, + "Ġsprawling": 70892, + "Ġspray": 22899, + "Ġsprayed": 77621, + "Ġspraying": 82536, + "Ġspre": 65562, + "Ġspread": 8865, + "Ġspreading": 30035, + "Ġspreads": 42553, + "Ġspreadsheet": 45886, + "Ġspree": 78628, + "Ġspring": 10464, + "Ġspringfox": 96377, + "Ġsprings": 41142, + "Ġsprink": 50698, + "Ġsprinkle": 84847, + "Ġsprint": 37849, + "Ġsprintf": 11812, + "Ġsprite": 17051, + "ĠspriteBatch": 77188, + "Ġsprites": 46757, + "Ġsprung": 91210, + "Ġsprz": 127977, + "Ġsprzeda": 127978, + "Ġsprzedaż": 139648, + "Ġsprzedaży": 138326, + "ĠsprzÄĻ": 136640, + "ĠsprzÄĻt": 136641, + "Ġspun": 56485, + "Ġspur": 59031, + "Ġspurred": 84647, + "Ġspy": 21236, + "ĠspyOn": 66534, + "Ġspying": 62214, + "Ġspäter": 87791, + "Ġspéc": 70169, + "Ġspécial": 75205, + "Ġspécialisé": 143530, + "Ġspécifique": 142943, + "ĠspÄĻd": 140781, + "ĠspÄĻdz": 140782, + "Ġsq": 18031, + "Ġsql": 5704, + "ĠsqlCommand": 85304, + "ĠsqlSession": 90065, + "Ġsqlalchemy": 42175, + "Ġsqlite": 22003, + "Ġsqr": 85029, + "Ġsqrt": 17946, + "Ġsqu": 8167, + "Ġsquad": 18529, + "Ġsquadron": 97280, + "Ġsquads": 77196, + "Ġsquare": 9334, + "Ġsquared": 52263, + "Ġsquarely": 94463, + "Ġsquares": 31340, + "Ġsquash": 58476, + "Ġsquat": 57470, + "Ġsque": 31132, + "Ġsqueez": 43746, + "Ġsqueeze": 36563, + "Ġsqueezed": 64162, + "Ġsqueezing": 87707, + "Ġsquid": 89175, + "Ġsquir": 45950, + "Ġsquirrel": 90524, + "Ġsquirt": 72481, + "Ġsr": 18962, + "Ġsrand": 68268, + "Ġsrc": 2286, + "Ġsrv": 43578, + "Ġss": 10870, + "Ġsscanf": 61913, + "Ġssh": 29230, + "Ġssid": 96361, + "Ġssize": 42951, + "Ġssl": 33537, + "Ġst": 357, + "Ġsta": 28511, + "Ġstaat": 73824, + "Ġstab": 26964, + "Ġstabbed": 50180, + "Ġstabbing": 70922, + "Ġstabil": 26184, + "Ġstability": 19753, + "Ġstabilization": 82838, + "Ġstabilize": 69136, + "Ġstabilized": 92063, + "Ġstable": 15175, + "Ġstack": 5611, + "Ġstacked": 41315, + "Ġstacking": 74072, + "Ġstacks": 39950, + "Ġstad": 86123, + "Ġstadium": 23889, + "Ġstadiums": 89084, + "Ġstaff": 5570, + "Ġstaffer": 77134, + "Ġstaffers": 60250, + "Ġstaffing": 64051, + "Ġstag": 76436, + "Ġstage": 6430, + "Ġstaged": 50057, + "Ġstages": 17628, + "Ġstagger": 39156, + "Ġstaggering": 54844, + "Ġstaging": 47762, + "Ġstagn": 53263, + "Ġstagnant": 95943, + "Ġstain": 51964, + "Ġstained": 60052, + "Ġstaining": 87796, + "Ġstainless": 24461, + "Ġstains": 62463, + "Ġstair": 36549, + "Ġstaircase": 61762, + "Ġstairs": 31149, + "Ġstake": 18279, + "Ġstakeholders": 38110, + "Ġstakes": 44425, + "Ġstal": 79476, + "Ġstale": 50351, + "Ġstalk": 54872, + "Ġstalking": 81795, + "Ġstall": 42693, + "Ġstalled": 67841, + "Ġstalls": 73573, + "Ġstamina": 60684, + "Ġstamp": 21206, + "Ġstamped": 64039, + "Ġstamps": 49212, + "Ġstan": 45194, + "Ġstance": 28931, + "Ġstand": 2498, + "Ġstandalone": 43388, + "Ġstandard": 5297, + "ĠstandardUserDefaults": 55989, + "Ġstandardized": 50014, + "Ġstandards": 10659, + "Ġstandby": 71738, + "Ġstanding": 11259, + "Ġstandings": 67563, + "Ġstandoff": 84327, + "Ġstandout": 71460, + "Ġstandpoint": 50782, + "Ġstands": 13352, + "Ġstanza": 95436, + "Ġstap": 35014, + "Ġstaple": 48956, + "Ġstaples": 86829, + "Ġstar": 6774, + "Ġstarch": 69081, + "Ġstare": 45843, + "Ġstared": 44035, + "Ġstares": 94402, + "Ġstaring": 36774, + "Ġstark": 37146, + "Ġstarred": 58235, + "Ġstarring": 39400, + "Ġstars": 9759, + "Ġstart": 1191, + "ĠstartActivity": 18273, + "ĠstartActivityForResult": 73627, + "ĠstartDate": 29686, + "ĠstartIndex": 40693, + "ĠstartPoint": 82429, + "ĠstartPos": 68791, + "ĠstartPosition": 98342, + "ĠstartTime": 22858, + "ĠstartX": 68599, + "ĠstartY": 61702, + "Ġstarted": 3855, + "Ġstarter": 26697, + "Ġstarters": 42706, + "Ġstarting": 5916, + "Ġstartled": 88715, + "Ġstartling": 67734, + "Ġstarts": 8471, + "Ġstarttime": 92395, + "Ġstartup": 20567, + "Ġstartups": 43383, + "Ġstarvation": 77145, + "Ġstarving": 77401, + "Ġstash": 64037, + "Ġstat": 2793, + "Ġstata": 79221, + "Ġstate": 1584, + "Ġstated": 10982, + "Ġstatement": 5114, + "Ġstatements": 12239, + "Ġstates": 5302, + "Ġstatewide": 50941, + "Ġstati": 95877, + "Ġstatic": 1099, + "Ġstatically": 91174, + "Ġstating": 27798, + "Ġstation": 8056, + "Ġstationary": 52635, + "Ġstationed": 62520, + "Ġstations": 17337, + "Ġstatist": 20115, + "Ġstatistic": 42489, + "Ġstatistical": 28464, + "Ġstatistically": 46852, + "Ġstatistics": 13142, + "Ġstato": 48153, + "Ġstats": 10472, + "Ġstatt": 67317, + "Ġstatue": 34272, + "Ġstatues": 57902, + "Ġstature": 76623, + "Ġstatus": 2639, + "ĠstatusBar": 90120, + "ĠstatusCode": 35532, + "Ġstatuses": 59358, + "Ġstatute": 34686, + "Ġstatutes": 61182, + "Ġstatutory": 44430, + "Ġstaunch": 81621, + "Ġstav": 74235, + "Ġstay": 4717, + "Ġstayed": 19613, + "Ġstaying": 19429, + "Ġstays": 26558, + "ĠstaÅĤ": 134118, + "Ġstd": 1460, + "ĠstdClass": 55287, + "Ġstddev": 87867, + "Ġstderr": 26436, + "Ġstdin": 31369, + "Ġstdout": 20075, + "Ġste": 4087, + "Ġstead": 16845, + "Ġsteadfast": 84287, + "Ġsteadily": 41971, + "Ġsteady": 24020, + "Ġsteak": 48959, + "Ġsteal": 26571, + "Ġstealing": 37998, + "Ġsteals": 61466, + "Ġstealth": 46965, + "Ġsteam": 20312, + "Ġsteel": 9509, + "Ġsteep": 31266, + "Ġsteer": 48615, + "Ġsteering": 32177, + "Ġstehen": 64857, + "Ġsteht": 54383, + "Ġstellar": 47217, + "Ġstellen": 68102, + "Ġstellt": 97096, + "Ġstem": 19101, + "Ġstemmed": 98518, + "Ġstemming": 75944, + "Ġstems": 43714, + "Ġstencil": 72370, + "Ġstep": 3019, + "Ġstepped": 24326, + "Ġstepper": 95932, + "Ġstepping": 35467, + "Ġsteps": 7354, + "Ġster": 21809, + "Ġstere": 22646, + "Ġstereo": 37952, + "Ġstereotype": 73646, + "Ġstereotypes": 53068, + "Ġsteril": 65640, + "Ġsterile": 81643, + "Ġsterling": 71937, + "Ġstern": 57544, + "Ġsteroid": 76748, + "Ġsteroids": 57061, + "Ġstesso": 82588, + "Ġstew": 60343, + "Ġsteward": 68506, + "Ġstick": 9214, + "Ġsticker": 46177, + "Ġstickers": 49296, + "Ġsticking": 36972, + "Ġsticks": 33140, + "Ġsticky": 27402, + "Ġstif": 91226, + "Ġstiff": 30061, + "Ġstiffness": 69234, + "Ġstigma": 59281, + "Ġstil": 90766, + "Ġstill": 2058, + "Ġstim": 12647, + "Ġstimulate": 49977, + "Ġstimulated": 80371, + "Ġstimulates": 94355, + "Ġstimulating": 64692, + "Ġstimulation": 40859, + "Ġstimuli": 55588, + "Ġstimulus": 38435, + "Ġsting": 54313, + "Ġstint": 60708, + "Ġstip": 42673, + "Ġstir": 23065, + "Ġstirred": 74840, + "Ġstirring": 53954, + "Ġstitch": 32861, + "Ġstitched": 94608, + "Ġstitches": 61936, + "Ġstitching": 79131, + "Ġstk": 55003, + "Ġstm": 55258, + "Ġstmt": 20020, + "Ġsto": 42032, + "Ġstochastic": 95240, + "Ġstock": 5591, + "Ġstocked": 70016, + "Ġstockholm": 92090, + "Ġstocking": 79149, + "Ġstockings": 75605, + "Ġstocks": 22488, + "Ġstoi": 87344, + "Ġstole": 39506, + "Ġstolen": 22329, + "Ġstom": 20100, + "Ġstomach": 22350, + "Ġstone": 9798, + "Ġstones": 26210, + "Ġstood": 14638, + "Ġstool": 63072, + "Ġstools": 89772, + "Ġstop": 2936, + "Ġstopped": 10497, + "Ġstopping": 22535, + "Ġstops": 17933, + "Ġstopwatch": 85572, + "Ġstopwords": 91044, + "Ġstor": 33689, + "Ġstorage": 5819, + "Ġstore": 3553, + "Ġstored": 9768, + "Ġstorefront": 90728, + "Ġstores": 10533, + "Ġstoria": 98889, + "Ġstories": 7343, + "Ġstoring": 27572, + "Ġstorm": 13458, + "Ġstormed": 84987, + "Ġstorms": 43483, + "Ġstory": 3364, + "Ġstoryboard": 42033, + "Ġstoryline": 50628, + "Ġstoryt": 40006, + "Ġstorytelling": 47829, + "Ġstos": 95983, + "Ġstout": 89672, + "Ġstove": 44015, + "Ġstr": 607, + "ĠstrSQL": 80153, + "ĠstrSql": 52131, + "Ġstraight": 7678, + "Ġstraightforward": 30339, + "Ġstrain": 25716, + "Ġstrained": 72679, + "Ġstrains": 41300, + "Ġstrand": 41489, + "Ġstranded": 62105, + "Ġstrands": 68764, + "Ġstrang": 46205, + "Ġstrange": 14888, + "Ġstrangely": 69606, + "Ġstranger": 33958, + "Ġstrangers": 39621, + "Ġstrap": 33547, + "Ġstrapon": 92246, + "Ġstrapped": 90024, + "Ġstraps": 48249, + "Ġstrat": 43297, + "Ġstrateg": 5275, + "Ġstrategic": 18580, + "Ġstrategically": 67674, + "Ġstrategies": 14830, + "Ġstrategist": 64135, + "Ġstrategy": 8282, + "Ġstratég": 142681, + "Ġstratégie": 142682, + "Ġstraw": 30007, + "Ġstrawberries": 75103, + "Ġstrawberry": 72600, + "Ġstray": 61390, + "Ġstrcat": 33512, + "Ġstrchr": 64035, + "Ġstrcmp": 24272, + "Ġstrconv": 33317, + "Ġstrcpy": 23072, + "Ġstrdup": 48819, + "Ġstre": 5412, + "Ġstreak": 29214, + "Ġstream": 4269, + "Ġstreamed": 73245, + "Ġstreaming": 16842, + "Ġstreamline": 81603, + "Ġstreamlined": 74060, + "Ġstreams": 23045, + "Ġstreet": 8592, + "Ġstreets": 14371, + "Ġstren": 92219, + "Ġstrength": 8170, + "Ġstrengthen": 19681, + "Ġstrengthened": 59762, + "Ġstrengthening": 47413, + "Ġstrengthens": 95831, + "Ġstrengths": 35386, + "Ġstrerror": 43673, + "Ġstress": 8464, + "Ġstressed": 31547, + "Ġstresses": 58523, + "Ġstressful": 45783, + "Ġstressing": 93691, + "Ġstret": 26185, + "Ġstretch": 14501, + "Ġstretched": 40298, + "Ġstretches": 49599, + "Ġstretching": 41849, + "Ġstrftime": 81400, + "Ġstri": 5950, + "Ġstrict": 7304, + "Ġstricter": 83718, + "Ġstrictly": 25470, + "Ġstride": 19228, + "Ġstrides": 36287, + "Ġstrife": 96612, + "Ġstrike": 13169, + "Ġstrikeouts": 88926, + "Ġstriker": 43941, + "Ġstrikes": 22368, + "Ġstriking": 21239, + "Ġstring": 914, + "ĠstringBuffer": 60707, + "ĠstringBuilder": 39050, + "ĠstringBy": 42098, + "ĠstringByAppending": 59261, + "ĠstringByAppendingString": 96510, + "ĠstringValue": 78468, + "ĠstringWith": 19027, + "ĠstringWithFormat": 20067, + "Ġstringent": 68999, + "Ġstringify": 83166, + "Ġstrings": 9069, + "Ġstringstream": 80927, + "Ġstrip": 13316, + "Ġstripe": 45542, + "Ġstriped": 67590, + "Ġstripes": 54688, + "Ġstripped": 36679, + "Ġstripper": 96687, + "Ġstripping": 66670, + "Ġstrips": 29922, + "Ġstripslashes": 63223, + "Ġstrive": 36006, + "Ġstrives": 74551, + "Ġstriving": 67627, + "Ġstrlen": 12420, + "Ġstrm": 88776, + "Ġstrncmp": 93202, + "Ġstrncpy": 57329, + "Ġstro": 30411, + "Ġstroke": 12654, + "ĠstrokeLine": 77829, + "ĠstrokeWidth": 62825, + "Ġstrokes": 52472, + "Ġstroll": 68883, + "Ġstron": 77061, + "Ġstrong": 3746, + "Ġstronger": 16245, + "Ġstrongest": 29905, + "Ġstronghold": 85468, + "Ġstrongly": 16510, + "ĠstronÄĻ": 139926, + "Ġstrpos": 43468, + "Ġstrr": 79452, + "Ġstrs": 57584, + "Ġstrstr": 66295, + "Ġstrt": 13613, + "Ġstrtok": 48345, + "Ġstrtol": 33624, + "Ġstrtolower": 38814, + "Ġstrtotime": 24646, + "Ġstrtoupper": 60440, + "Ġstruck": 17490, + "Ġstruct": 2036, + "Ġstructs": 62845, + "Ġstructural": 23759, + "Ġstructure": 5944, + "Ġstructured": 32930, + "Ġstructures": 14389, + "Ġstrugg": 7930, + "Ġstruggle": 14651, + "Ġstruggled": 27115, + "Ġstruggles": 27870, + "Ġstruggling": 19962, + "Ġstrut": 82343, + "Ġsts": 51756, + "Ġstu": 54754, + "Ġstub": 13633, + "Ġstubborn": 51396, + "Ġstuck": 15700, + "Ġstud": 1673, + "Ġstudent": 5458, + "Ġstudents": 4143, + "Ġstudied": 19476, + "Ġstudies": 7822, + "Ġstudio": 14029, + "Ġstudios": 41695, + "Ġstuds": 85621, + "Ġstudy": 3920, + "Ġstudying": 20956, + "Ġstuff": 6259, + "Ġstuffed": 44313, + "Ġstuffing": 82516, + "Ġstumble": 74137, + "Ġstumbled": 49057, + "Ġstumbling": 87355, + "Ġstump": 88612, + "Ġstun": 65428, + "Ġstunned": 46926, + "Ġstunning": 19850, + "Ġstunt": 59025, + "Ġstup": 90032, + "Ġstupid": 18253, + "Ġstupidity": 87537, + "Ġsturdy": 44552, + "Ġstutter": 90537, + "Ġsty": 17613, + "Ġstyl": 48204, + "Ġstyle": 1707, + "ĠstyleUrls": 17650, + "Ġstyled": 11096, + "Ġstyles": 9222, + "Ġstylesheet": 73802, + "Ġstyling": 41328, + "Ġstylish": 31361, + "Ġstylist": 66373, + "Ġstør": 74350, + "Ġsu": 913, + "Ġsua": 19345, + "Ġsuas": 52097, + "Ġsub": 1186, + "ĠsubTitle": 91910, + "Ġsubclass": 37190, + "Ġsubclasses": 68183, + "Ġsubconscious": 83851, + "Ġsubcontract": 94379, + "Ġsubdir": 98021, + "Ġsubdiv": 66509, + "Ġsubdivision": 59822, + "Ġsubdivisions": 91445, + "Ġsubdued": 92419, + "Ġsubgroup": 80115, + "Ġsubj": 80317, + "Ġsubject": 3832, + "Ġsubjected": 37026, + "Ġsubjective": 43022, + "Ġsubjects": 14878, + "Ġsublic": 28491, + "Ġsublicense": 28787, + "Ġsublime": 70146, + "Ġsublist": 93893, + "Ġsubmar": 42171, + "Ġsubmarine": 57529, + "Ġsubmarines": 85542, + "Ġsubmenu": 77397, + "Ġsubmerged": 84077, + "Ġsubmission": 20503, + "Ġsubmissions": 36835, + "Ġsubmissive": 96240, + "Ġsubmit": 9318, + "Ġsubmits": 94035, + "Ġsubmitted": 14634, + "Ġsubmitting": 33094, + "Ġsubmodule": 95570, + "Ġsubnet": 51457, + "Ġsubordinate": 78163, + "Ġsubparagraph": 74815, + "Ġsubplot": 46626, + "Ġsubpo": 60509, + "Ġsubpoena": 72769, + "Ġsubprocess": 23514, + "Ġsubrange": 93791, + "Ġsubreddit": 62092, + "Ġsubroutine": 88334, + "Ġsubs": 5147, + "Ġsubscri": 12398, + "Ġsubscribe": 17963, + "Ġsubscribed": 52100, + "Ġsubscriber": 32115, + "Ġsubscribers": 30020, + "Ġsubscribing": 39785, + "Ġsubscript": 98250, + "Ġsubscription": 15142, + "Ġsubscriptions": 40355, + "Ġsubsection": 22848, + "Ġsubsequ": 13085, + "Ġsubsequent": 17420, + "Ġsubsequently": 27420, + "Ġsubset": 25993, + "Ġsubsets": 74505, + "Ġsubsid": 24863, + "Ġsubsidi": 31519, + "Ġsubsidiaries": 71102, + "Ġsubsidiary": 40064, + "Ġsubsidies": 42021, + "Ġsubsidized": 94014, + "Ġsubsidy": 67647, + "Ġsubst": 37315, + "Ġsubstance": 19699, + "Ġsubstances": 32055, + "Ġsubstant": 10913, + "Ġsubstantial": 11924, + "Ġsubstantially": 31202, + "Ġsubstantive": 66649, + "Ġsubstit": 31334, + "Ġsubstitute": 27679, + "Ġsubstituted": 62096, + "Ġsubstitutes": 90262, + "Ġsubstitution": 48968, + "Ġsubstitutions": 93650, + "Ġsubstr": 15769, + "Ġsubstrate": 52957, + "Ġsubstring": 38449, + "Ġsubsystem": 52785, + "Ġsubt": 41029, + "Ġsubtitle": 31735, + "Ġsubtitles": 66666, + "Ġsubtle": 26447, + "Ġsubtly": 86317, + "Ġsubtotal": 90381, + "Ġsubtract": 32256, + "Ġsubtraction": 75240, + "Ġsubtree": 53464, + "Ġsubtype": 52482, + "Ġsubur": 29322, + "Ġsuburb": 51165, + "Ġsuburban": 45218, + "Ġsuburbs": 46913, + "Ġsubway": 43246, + "Ġsuc": 11157, + "Ġsucc": 20038, + "Ġsucceed": 11996, + "Ġsucceeded": 25331, + "Ġsucceeding": 72720, + "Ġsucceeds": 50081, + "Ġsucces": 49451, + "Ġsuccesfully": 92636, + "Ġsuccess": 2393, + "Ġsuccesses": 47088, + "Ġsuccessful": 6849, + "Ġsuccessfully": 7790, + "Ġsuccession": 49687, + "Ġsuccessive": 48924, + "Ġsuccessor": 33565, + "Ġsuccessors": 74896, + "Ġsuccinct": 98632, + "Ġsuccès": 93399, + "Ġsuce": 74200, + "Ġsucess": 77247, + "Ġsucesso": 43462, + "Ġsuch": 1741, + "Ġsuche": 77654, + "Ġsuchen": 66124, + "Ġsucht": 40416, + "Ġsuck": 18886, + "Ġsucked": 52212, + "Ġsucker": 93788, + "Ġsucking": 43525, + "Ġsucks": 39099, + "Ġsuction": 83778, + "Ġsud": 30737, + "Ġsudah": 49796, + "Ġsudden": 10968, + "Ġsuddenly": 14843, + "Ġsudo": 25408, + "Ġsudoku": 90809, + "Ġsue": 33772, + "Ġsued": 41084, + "Ġsuede": 96635, + "Ġsuf": 55768, + "Ġsuff": 8489, + "Ġsuffer": 7676, + "Ġsuffered": 16256, + "Ġsufferers": 95360, + "Ġsuffering": 15691, + "Ġsuffers": 46421, + "Ġsuffice": 76156, + "Ġsufficient": 14016, + "Ġsufficiently": 38944, + "Ġsuffix": 20525, + "Ġsug": 30605, + "Ġsugar": 13163, + "Ġsugars": 69813, + "Ġsuger": 97584, + "Ġsuggest": 4190, + "Ġsuggested": 11827, + "Ġsuggesting": 22561, + "Ġsuggestion": 23776, + "Ġsuggestions": 18225, + "Ġsuggestive": 98478, + "Ġsuggests": 13230, + "Ġsui": 83653, + "Ġsuic": 42819, + "Ġsuicidal": 65053, + "Ġsuicide": 18144, + "Ġsuicides": 83949, + "Ġsuing": 69407, + "Ġsuis": 35631, + "Ġsuit": 7781, + "Ġsuitability": 80628, + "Ġsuitable": 14452, + "Ġsuitcase": 87849, + "Ġsuite": 16182, + "Ġsuited": 31499, + "Ġsuites": 55164, + "Ġsuits": 27976, + "Ġsuiv": 45832, + "Ġsujet": 69827, + "Ġsuk": 95790, + "Ġsul": 25774, + "Ġsulf": 39669, + "Ġsulfate": 88318, + "Ġsulfur": 71491, + "Ġsulla": 56970, + "Ġsulph": 82678, + "Ġsum": 2629, + "Ġsuma": 51313, + "Ġsumm": 34259, + "Ġsummar": 28285, + "Ġsummaries": 68922, + "Ġsummarize": 62079, + "Ġsummarized": 68629, + "Ġsummarizes": 93494, + "Ġsummary": 12126, + "Ġsummed": 65666, + "Ġsummer": 7324, + "Ġsummers": 79669, + "Ġsummit": 28948, + "Ġsummon": 27545, + "Ġsummoned": 51710, + "Ġsummons": 74300, + "Ġsums": 36398, + "Ġsun": 7015, + "Ġsund": 50582, + "Ġsunday": 92363, + "Ġsung": 39339, + "Ġsunglasses": 59369, + "Ġsunk": 67165, + "Ġsunlight": 39020, + "Ġsunny": 39698, + "Ġsunrise": 63819, + "Ġsunscreen": 82084, + "Ġsunset": 42984, + "Ġsunshine": 51395, + "Ġsunt": 39695, + "Ġsuo": 51512, + "Ġsuoi": 96020, + "Ġsup": 1026, + "Ġsuper": 2256, + "ĠsuperClass": 59657, + "Ġsuperb": 32589, + "Ġsuperclass": 77614, + "Ġsuperf": 42017, + "Ġsuperficial": 58815, + "Ġsuperhero": 45144, + "Ġsuperheroes": 90718, + "Ġsuperintendent": 73033, + "Ġsuperior": 16353, + "Ġsuperiority": 65632, + "Ġsupermarket": 48213, + "Ġsupermarkets": 76499, + "Ġsupern": 38787, + "Ġsupernatural": 48958, + "Ġsupers": 51690, + "Ġsuperst": 93605, + "Ġsuperstar": 57247, + "Ġsuperv": 15574, + "Ġsupervise": 99101, + "Ġsupervised": 58989, + "Ġsupervision": 37117, + "Ġsupervisor": 37319, + "Ġsupervisors": 72951, + "Ġsupp": 90035, + "Ġsupper": 86017, + "Ġsuppl": 81013, + "Ġsupplement": 22047, + "Ġsupplemental": 67951, + "Ġsupplementary": 79406, + "Ġsupplementation": 72696, + "Ġsupplemented": 88704, + "Ġsupplements": 32590, + "Ġsupplied": 17221, + "Ġsupplier": 18823, + "Ġsuppliers": 20350, + "Ġsupplies": 16720, + "Ġsupply": 8149, + "Ġsupplying": 43313, + "Ġsupport": 1824, + "Ġsupported": 7248, + "Ġsupporter": 36463, + "Ġsupporters": 15511, + "Ġsupporting": 12613, + "Ġsupportive": 32345, + "Ġsupports": 11554, + "Ġsuppose": 22477, + "Ġsupposed": 9966, + "Ġsupposedly": 32728, + "Ġsuppress": 27222, + "Ġsuppressed": 54989, + "Ġsuppressing": 97695, + "Ġsuppression": 45635, + "Ġsupra": 66555, + "Ġsuprem": 34125, + "Ġsupremacist": 82011, + "Ġsupremacy": 64403, + "Ġsupreme": 43122, + "Ġsupérieur": 140134, + "Ġsur": 1729, + "Ġsure": 2704, + "Ġsurely": 21502, + "Ġsurf": 20269, + "Ġsurface": 7329, + "Ġsurfaced": 57546, + "Ġsurfaces": 26431, + "Ġsurfing": 57740, + "Ġsurg": 12005, + "Ġsurge": 21781, + "Ġsurged": 83907, + "Ġsurgeon": 44998, + "Ġsurgeons": 73172, + "Ġsurgeries": 67723, + "Ġsurgery": 14829, + "Ġsurgical": 33833, + "Ġsurname": 39753, + "Ġsurpass": 52020, + "Ġsurpassed": 67228, + "Ġsurplus": 40448, + "Ġsurpr": 6218, + "Ġsurprise": 12761, + "Ġsurprised": 14453, + "Ġsurprises": 45440, + "Ġsurprising": 14861, + "Ġsurprisingly": 28292, + "Ġsurre": 24165, + "Ġsurreal": 64523, + "Ġsurrender": 31759, + "Ġsurrendered": 69944, + "Ġsurrogate": 72850, + "Ġsurround": 8995, + "Ġsurrounded": 22865, + "Ġsurrounding": 14590, + "Ġsurroundings": 39090, + "Ġsurrounds": 70274, + "Ġsurtout": 75676, + "Ġsurv": 7398, + "Ġsurve": 15403, + "Ġsurveillance": 21440, + "Ġsurvey": 10572, + "Ġsurveyed": 47998, + "Ġsurveys": 31213, + "Ġsurviv": 22750, + "Ġsurvival": 19661, + "Ġsurvive": 17697, + "Ġsurvived": 25882, + "Ġsurvives": 82317, + "Ġsurviving": 39646, + "Ġsurvivor": 48648, + "Ġsurvivors": 31596, + "Ġsus": 4481, + "Ġsuscept": 36229, + "Ġsusceptibility": 87536, + "Ġsusceptible": 46181, + "Ġsushi": 66222, + "Ġsusp": 9298, + "Ġsuspect": 15207, + "Ġsuspected": 23804, + "Ġsuspects": 29761, + "Ġsuspend": 26247, + "Ġsuspended": 21612, + "Ġsuspense": 71830, + "Ġsuspension": 24301, + "Ġsuspicion": 37041, + "Ġsuspicions": 78507, + "Ġsuspicious": 31327, + "Ġsust": 10235, + "Ġsustain": 13879, + "Ġsustainability": 40229, + "Ġsustainable": 21803, + "Ġsustained": 28659, + "Ġsustaining": 79084, + "Ġsut": 34212, + "Ġsuç": 135023, + "Ġsuất": 131531, + "Ġsuá»ijt": 134601, + "Ġsv": 13559, + "Ġsvc": 46154, + "Ġsve": 80029, + "Ġsvens": 48697, + "Ġsvensk": 79161, + "Ġsvenska": 86953, + "Ġsvg": 26852, + "Ġsvm": 66311, + "Ġsvn": 60446, + "Ġsvo": 66007, + "Ġsw": 2021, + "Ġswagger": 60823, + "Ġswal": 49354, + "Ġswallow": 41176, + "Ġswallowed": 63337, + "Ġswallowing": 90647, + "Ġswamp": 66072, + "Ġswap": 14291, + "Ġswapped": 56950, + "Ġswapping": 63719, + "Ġswaps": 73151, + "Ġswarm": 60841, + "Ġswath": 92412, + "Ġsway": 60129, + "Ġswe": 10546, + "Ġswear": 46077, + "Ġswearing": 90681, + "Ġsweat": 27466, + "Ġsweater": 60121, + "Ġsweating": 80266, + "Ġsweaty": 97421, + "Ġsweep": 23146, + "Ġsweeping": 40971, + "Ġsweeps": 96467, + "Ġsweet": 10226, + "Ġsweetheart": 91254, + "Ġsweetness": 63450, + "Ġsweets": 75557, + "Ġswell": 77253, + "Ġswelling": 54207, + "Ġswept": 40223, + "Ġswift": 29362, + "Ġswiftly": 55551, + "Ġswim": 16191, + "Ġswimming": 23380, + "Ġswims": 80636, + "Ġswing": 18806, + "Ġswinger": 23624, + "Ġswingerclub": 45854, + "Ġswingers": 54026, + "Ġswinging": 53960, + "Ġswings": 56790, + "Ġswipe": 38520, + "Ġswiper": 78319, + "Ġswirl": 86360, + "Ġswirling": 92278, + "Ġswitch": 3398, + "Ġswitched": 29875, + "Ġswitches": 31167, + "Ġswitching": 27765, + "Ġswo": 33127, + "ĠswojÄħ": 134106, + "Ġswollen": 75093, + "Ġsword": 20214, + "Ġswords": 52057, + "Ġswore": 91207, + "Ġsworn": 50648, + "Ġswung": 69855, + "Ġswój": 136865, + "Ġsx": 31808, + "Ġsy": 6568, + "Ġsyll": 47604, + "Ġsym": 7886, + "Ġsymb": 66654, + "Ġsymbol": 7735, + "Ġsymbolic": 35296, + "Ġsymbolism": 82376, + "Ġsymbols": 17738, + "Ġsymlink": 83221, + "Ġsymmetric": 54343, + "Ġsymmetry": 45120, + "Ġsymp": 21459, + "Ġsympath": 32718, + "Ġsympathetic": 52364, + "Ġsympathy": 45984, + "Ġsympt": 11841, + "Ġsymptom": 48548, + "Ġsymptoms": 13495, + "Ġsyn": 6782, + "Ġsynagogue": 94865, + "Ġsynaptic": 98727, + "Ġsync": 12811, + "Ġsynced": 85028, + "Ġsynchron": 14121, + "Ġsynchronization": 57912, + "Ġsynchronize": 63799, + "Ġsynchronized": 21466, + "Ġsynchronous": 65949, + "Ġsyncing": 96145, + "Ġsynd": 21307, + "Ġsyndrome": 27339, + "Ġsynerg": 79426, + "Ġsynergy": 92040, + "Ġsynonym": 73350, + "Ġsynonymous": 68493, + "Ġsynonyms": 85406, + "Ġsynopsis": 80663, + "Ġsynt": 83187, + "Ġsyntax": 19482, + "Ġsynth": 42898, + "Ġsynthes": 51289, + "Ġsynthesis": 38875, + "Ġsynthesized": 91006, + "Ġsynthetic": 27268, + "Ġsyrup": 44222, + "Ġsys": 5708, + "Ġsyscall": 49345, + "Ġsyslog": 74487, + "Ġsyst": 41245, + "Ġsystem": 1849, + "ĠsystemFontOfSize": 51816, + "Ġsystematic": 36438, + "Ġsystematically": 59726, + "Ġsystemctl": 89421, + "Ġsystemd": 74966, + "Ġsystemic": 45317, + "Ġsystems": 5942, + "Ġsystème": 71501, + "Ġsz": 10038, + "Ġszcz": 75573, + "Ġszczeg": 126161, + "Ġszczególn": 139071, + "Ġszczególnie": 139072, + "ĠszczegóÅĤ": 128067, + "ĠszczegóÅĤowo": 140930, + "Ġszer": 57117, + "Ġszko": 140013, + "ĠszkoÅĤy": 140014, + "Ġszy": 91765, + "Ġsábado": 97459, + "Ġsách": 95082, + "Ġsáng": 128688, + "Ġsát": 129261, + "Ġsân": 130514, + "Ġsâu": 131307, + "Ġsão": 29610, + "Ġsä": 64774, + "ĠsÃ¥": 21072, + "Ġsæ": 92325, + "Ġsé": 18744, + "Ġséculo": 143042, + "Ġsécur": 64361, + "Ġsécurité": 76363, + "Ġséjour": 137612, + "Ġsélection": 70032, + "Ġsérie": 46428, + "Ġsó": 31090, + "Ġsóc": 131998, + "Ġsólo": 52188, + "Ġsóng": 136847, + "Ġsông": 134782, + "Ġsö": 50287, + "Ġsöyle": 129284, + "Ġsöyled": 131982, + "Ġsöyledi": 131983, + "Ġsöz": 126519, + "ĠsözleÅŁme": 139406, + "Ġsø": 30330, + "Ġsøger": 61271, + "Ġsøker": 92744, + "Ġsû": 77605, + "Ġsûr": 96617, + "Ġsü": 59739, + "Ġsür": 127146, + "Ġsürdür": 135239, + "Ġsüre": 126369, + "Ġsürec": 139889, + "Ġsürecin": 139890, + "Ġsürek": 136454, + "Ġsürekli": 136455, + "Ġsüreç": 137302, + "Ġsürü": 134446, + "ĠsÃłn": 136071, + "ĠsÃłng": 137968, + "ĠsÃŃ": 44715, + "ĠsÄ©": 128916, + "Ġsı": 125715, + "Ġsıc": 135956, + "Ġsıcak": 135957, + "Ġsık": 126998, + "Ġsıkıntı": 138636, + "Ġsın": 133011, + "Ġsınav": 138575, + "Ġsını": 135964, + "Ġsınıf": 135965, + "Ġsınır": 133012, + "Ġsıra": 129723, + "Ġsırasında": 137424, + "ĠsÄĥ": 28994, + "ĠsÄħ": 37127, + "ĠsÅĤ": 74258, + "ĠsÅĤab": 142363, + "ĠsÅĤow": 133452, + "ĠsÅĤucha": 142039, + "ĠsÅĤuż": 143054, + "ĠsÅĤużb": 143055, + "ĠsÆ¡": 129582, + "Ġsư": 129773, + "Ġsạch": 131560, + "Ġsạn": 133724, + "Ġsản": 85636, + "Ġsắc": 129222, + "Ġsắp": 134406, + "Ġsắt": 139961, + "Ġsẵ": 133690, + "Ġsẵn": 133691, + "Ġsẻ": 129958, + "Ġsẽ": 128267, + "Ġsợ": 131146, + "Ġsứ": 138624, + "Ġsức": 128738, + "Ġsữa": 133044, + "Ġsá»±": 128277, + "Ġsá»ij": 65526, + "Ġsá»ijng": 128480, + "Ġsá»ķ": 138504, + "ĠsỼ": 131968, + "ĠsỼm": 131969, + "Ġsợ": 128760, + "Ġsá»Ń": 128353, + "Ġsá»Ńa": 132038, + "Ġt": 259, + "Ġta": 9450, + "Ġtab": 5651, + "ĠtabBar": 76781, + "ĠtabIndex": 84941, + "ĠtabPage": 87810, + "Ġtabel": 89782, + "Ġtabela": 67215, + "Ġtabindex": 30173, + "Ġtabla": 40546, + "Ġtable": 1965, + "ĠtableLayoutPanel": 97755, + "ĠtableName": 29544, + "ĠtableView": 13756, + "Ġtableau": 56134, + "Ġtablename": 86365, + "Ġtables": 12632, + "Ġtablesp": 37090, + "Ġtablespoon": 61511, + "Ġtablespoons": 55488, + "Ġtablet": 20697, + "Ġtabletop": 88471, + "Ġtablets": 28579, + "Ġtaboo": 75945, + "Ġtabs": 22398, + "Ġtac": 59288, + "Ġtack": 13870, + "Ġtackle": 21403, + "Ġtackled": 81357, + "Ġtackles": 42481, + "Ġtackling": 56811, + "Ġtaco": 90841, + "Ġtacos": 84401, + "Ġtact": 14083, + "Ġtactic": 37901, + "Ġtactical": 38647, + "Ġtactics": 25342, + "Ġtactile": 98583, + "Ġtad": 63806, + "Ġtag": 4772, + "ĠtagName": 62367, + "Ġtagged": 37113, + "Ġtagging": 78153, + "Ġtags": 9492, + "Ġtah": 90093, + "Ġtahun": 57523, + "Ġtai": 52537, + "Ġtail": 9787, + "Ġtaille": 61727, + "Ġtailor": 50956, + "Ġtailored": 40791, + "Ġtails": 63514, + "Ġtainted": 83973, + "Ġtak": 18116, + "Ġtake": 1896, + "Ġtakeaway": 88374, + "Ġtaken": 4429, + "Ġtakeover": 63210, + "Ġtakes": 4990, + "Ġtaking": 4633, + "Ġtako": 97033, + "Ġtakım": 131967, + "Ġtakże": 63854, + "Ġtal": 8210, + "Ġtale": 23281, + "Ġtalent": 10772, + "Ġtalented": 23074, + "Ġtalents": 33932, + "Ġtales": 36465, + "Ġtalk": 3061, + "Ġtalked": 14897, + "Ġtalking": 7404, + "Ġtalks": 13433, + "Ġtall": 16217, + "Ġtaller": 49909, + "Ġtallest": 81617, + "Ġtally": 52295, + "Ġtam": 25476, + "Ġtamanho": 69238, + "Ġtamaño": 80252, + "Ġtamb": 17379, + "Ġtambién": 28471, + "Ġtambém": 34601, + "Ġtame": 81823, + "Ġtamil": 84535, + "Ġtamp": 75029, + "Ġtan": 14197, + "Ġtand": 53567, + "Ġtandem": 65942, + "Ġtang": 21878, + "Ġtangent": 68660, + "Ġtanggal": 90902, + "Ġtangible": 49301, + "Ġtangled": 92841, + "Ġtank": 12896, + "Ġtanker": 92394, + "Ġtanks": 25866, + "Ġtant": 36522, + "Ġtantal": 94585, + "Ġtanto": 32086, + "Ġtantr": 63007, + "Ġtantra": 32102, + "Ġtanı": 129409, + "Ġtap": 15239, + "Ġtape": 16971, + "Ġtaped": 82514, + "Ġtaper": 53935, + "Ġtapered": 94055, + "Ġtapes": 51696, + "Ġtapi": 98076, + "Ġtapped": 48423, + "Ġtapping": 52628, + "Ġtaps": 61970, + "Ġtar": 12183, + "Ġtaraf": 95798, + "Ġtarafından": 128691, + "Ġtard": 61097, + "Ġtarde": 60645, + "Ġtarea": 89907, + "Ġtarg": 88980, + "Ġtarget": 2169, + "ĠtargetType": 71330, + "Ġtargeted": 17112, + "Ġtargeting": 24132, + "Ġtargets": 11550, + "Ġtariff": 60261, + "Ġtariffs": 43895, + "Ġtarn": 86480, + "Ġtart": 44815, + "ĠtartÄ±ÅŁ": 142124, + "ĠtartÄ±ÅŁma": 142125, + "Ġtas": 51214, + "Ġtasar": 135291, + "Ġtasarım": 135292, + "Ġtask": 3383, + "ĠtaskId": 60527, + "Ġtasked": 50820, + "Ġtasks": 9079, + "Ġtast": 57072, + "Ġtaste": 12656, + "Ġtasted": 60480, + "Ġtastes": 35163, + "Ġtasting": 46400, + "Ġtasty": 38853, + "Ġtat": 71713, + "Ġtats": 140525, + "Ġtatsäch": 140526, + "Ġtatsächlich": 140527, + "Ġtatto": 24781, + "Ġtattoo": 31794, + "Ġtattoos": 53642, + "Ġtau": 31823, + "Ġtaught": 15599, + "Ġtav": 92209, + "Ġtavern": 90135, + "Ġtax": 3742, + "Ġtaxa": 76214, + "Ġtaxable": 58353, + "Ġtaxation": 50627, + "Ġtaxed": 71415, + "Ġtaxes": 13125, + "Ġtaxi": 32505, + "Ġtaxing": 86904, + "Ġtaxis": 89503, + "Ġtaxonomy": 71806, + "Ġtaxp": 22979, + "Ġtaxpayer": 38173, + "Ġtaxpayers": 34648, + "ĠtaÅŁÄ±": 129818, + "Ġtb": 16363, + "Ġtbl": 21173, + "Ġtbody": 45420, + "Ġtbsp": 63875, + "Ġtc": 17130, + "Ġtcb": 93179, + "Ġtcp": 28051, + "Ġtd": 17941, + "Ġte": 1013, + "Ġtea": 15243, + "Ġteach": 4538, + "Ġteacher": 11079, + "Ġteachers": 13336, + "Ġteaches": 32892, + "Ġteaching": 12629, + "Ġteachings": 44472, + "Ġteal": 94607, + "Ġteam": 2083, + "Ġteamed": 60210, + "Ġteammate": 47566, + "Ġteammates": 31646, + "Ġteams": 7263, + "Ġteamwork": 83533, + "Ġtear": 17576, + "ĠtearDown": 32825, + "Ġteardown": 49304, + "Ġtearing": 55736, + "Ġtears": 23139, + "Ġteas": 29568, + "Ġtease": 66026, + "Ġteased": 78775, + "Ġteaser": 69553, + "Ġteasing": 70363, + "Ġteaspoon": 41284, + "Ġteaspoons": 92100, + "Ġtec": 85571, + "Ġtech": 13014, + "Ġtechn": 2799, + "Ġtechnical": 10916, + "Ġtechnically": 31554, + "Ġtechnician": 53511, + "Ġtechnicians": 52359, + "Ġtechnique": 14762, + "Ġtechniques": 12538, + "Ġtechno": 88899, + "Ġtechnological": 29016, + "Ġtechnologies": 14310, + "Ġtechnology": 5440, + "Ġtecn": 40834, + "ĠtecnologÃŃa": 136037, + "Ġtecrü": 143451, + "Ġtecrübe": 143452, + "Ġted": 41223, + "Ġtedious": 65738, + "Ġtee": 35076, + "Ġteen": 9378, + "Ġteenage": 33168, + "Ġteenager": 31539, + "Ġteenagers": 36619, + "Ġteens": 26202, + "Ġteeth": 17832, + "Ġteg": 55955, + "Ġtegen": 71862, + "Ġtego": 49547, + "Ġteh": 79906, + "Ġteil": 77414, + "Ġtej": 61980, + "Ġtek": 42085, + "Ġtekn": 80026, + "Ġtekst": 72578, + "Ġtel": 18708, + "Ġtela": 69766, + "Ġtelah": 57635, + "Ġtele": 7963, + "Ġtelecom": 59405, + "Ġtelecommunications": 61766, + "Ġtelefon": 42150, + "Ġtelefone": 73164, + "Ġtelefono": 62430, + "Ġtelegram": 67544, + "Ġtelemetry": 61037, + "Ġtelephone": 20544, + "Ġteleport": 49489, + "Ġtelesc": 77413, + "Ġtelescope": 55825, + "Ġtelevis": 89965, + "Ġtelevised": 75610, + "Ġtelevision": 12425, + "Ġtell": 3291, + "Ġtelling": 11629, + "Ġtells": 10742, + "Ġteléfono": 140948, + "Ġtem": 1562, + "Ġtema": 45915, + "Ġtemas": 83906, + "Ġtemp": 2730, + "ĠtempList": 94864, + "Ġtemper": 6797, + "Ġtemperament": 78864, + "Ġtemperatura": 82953, + "Ġtemperature": 9315, + "Ġtemperatures": 19879, + "Ġtempered": 89661, + "Ġtempfile": 54819, + "Ġtempl": 67530, + "Ġtemplate": 3811, + "ĠtemplateUrl": 13031, + "Ġtemplates": 19911, + "Ġtemple": 26752, + "Ġtemples": 56897, + "Ġtempo": 23230, + "Ġtempor": 18965, + "Ġtemporada": 81461, + "Ġtemporal": 35915, + "Ġtemporarily": 27092, + "Ġtemporary": 13340, + "Ġtemps": 26046, + "Ġtempt": 24983, + "Ġtemptation": 61361, + "Ġtempted": 53880, + "Ġtempting": 60199, + "Ġten": 5779, + "Ġtenant": 25239, + "Ġtenants": 39916, + "Ġtend": 8376, + "Ġtended": 48790, + "Ġtendencies": 60455, + "Ġtendency": 30854, + "Ġtender": 27582, + "Ġtendon": 87832, + "Ġtendr": 90246, + "Ġtendrá": 135246, + "Ġtends": 27236, + "Ġtendência": 143559, + "Ġtenemos": 70448, + "Ġtener": 33430, + "Ġteng": 41149, + "Ġtenga": 92332, + "Ġtengo": 66890, + "Ġtennis": 31415, + "Ġtens": 22008, + "Ġtense": 42687, + "Ġtension": 23504, + "Ġtensions": 38410, + "Ġtensor": 15626, + "Ġtensorflow": 28087, + "Ġtensors": 77087, + "Ġtent": 15632, + "Ġtentang": 92064, + "Ġtentative": 75120, + "Ġtenth": 55666, + "Ġtents": 63370, + "Ġtenure": 38961, + "ĠtenÃŃa": 91565, + "Ġtep": 55052, + "Ġter": 1982, + "Ġtercer": 88881, + "Ġterm": 4647, + "Ġterme": 96920, + "Ġtermed": 60837, + "Ġtermin": 10204, + "Ġterminal": 15022, + "Ġterminals": 52979, + "Ġterminate": 29654, + "Ġterminated": 31272, + "Ġterminates": 87611, + "Ġterminating": 70581, + "Ġtermination": 34408, + "Ġterminator": 80025, + "Ġterminology": 56626, + "Ġterms": 3793, + "Ġtern": 71617, + "Ġterr": 7170, + "Ġterra": 59561, + "Ġterrace": 51478, + "Ġterrain": 24869, + "Ġterre": 86926, + "Ġterrestrial": 79392, + "Ġterrible": 17478, + "Ġterribly": 49036, + "Ġterrific": 42560, + "Ġterrified": 52631, + "Ġterrifying": 41251, + "Ġterritor": 39356, + "Ġterritorial": 51382, + "Ġterritories": 38443, + "Ġterritory": 17971, + "Ġterror": 8648, + "Ġterrorism": 23145, + "Ġterrorist": 19740, + "Ġterrorists": 28231, + "Ġters": 87630, + "Ġterse": 50537, + "Ġtersebut": 55766, + "Ġtert": 97114, + "Ġtertiary": 79323, + "Ġterug": 89501, + "Ġterör": 137047, + "Ġtes": 50209, + "Ġtess": 79830, + "Ġtest": 1273, + "ĠtestCase": 54452, + "ĠtestData": 67348, + "ĠtestName": 94396, + "Ġtestament": 60200, + "Ġtestcase": 70080, + "Ġteste": 74665, + "Ġtested": 12510, + "Ġtester": 37111, + "Ġtesters": 89719, + "Ġtestified": 38773, + "Ġtestify": 48162, + "Ġtestim": 18992, + "Ġtestimon": 46662, + "Ġtestimonials": 83510, + "Ġtestimony": 23468, + "Ġtesting": 7497, + "Ġtestosterone": 38932, + "Ġtests": 7032, + "Ġtet": 27853, + "Ġtetas": 87515, + "Ġtether": 79565, + "Ġtex": 23112, + "Ġtexas": 59248, + "Ġtext": 1467, + "ĠtextAlign": 23784, + "ĠtextBox": 19418, + "ĠtextColor": 47059, + "ĠtextDecoration": 91945, + "ĠtextField": 31388, + "ĠtextSize": 77400, + "ĠtextStatus": 60783, + "ĠtextStyle": 73096, + "ĠtextView": 31189, + "Ġtextarea": 52624, + "Ġtextbook": 55085, + "Ġtextbooks": 64203, + "Ġtextbox": 73999, + "Ġtexte": 67967, + "Ġtextile": 65538, + "Ġtextiles": 92982, + "Ġtexting": 67994, + "Ġtexto": 32025, + "Ġtexts": 21984, + "Ġtextual": 62533, + "Ġtexture": 10434, + "Ġtextured": 73544, + "Ġtextures": 29853, + "Ġteż": 65820, + "ĠteÅŁek": 133939, + "ĠteÅŁekkür": 133940, + "Ġtf": 6409, + "Ġtg": 53188, + "Ġtgt": 50475, + "Ġth": 270, + "Ġtha": 48488, + "Ġthai": 17936, + "Ġthaimassage": 65878, + "Ġthan": 1091, + "Ġthank": 9702, + "Ġthanked": 56495, + "Ġthankful": 42688, + "Ġthankfully": 75985, + "Ġthanking": 93021, + "Ġthanks": 9339, + "Ġthat": 429, + "Ġthats": 40036, + "Ġthaw": 85378, + "Ġthe": 279, + "Ġtheano": 95899, + "Ġtheat": 46113, + "Ġtheater": 26705, + "Ġtheaters": 43766, + "Ġtheatre": 33496, + "Ġtheatrical": 65233, + "Ġthee": 39244, + "Ġtheft": 27383, + "Ġtheir": 862, + "Ġtheirs": 47201, + "Ġthem": 1105, + "Ġthematic": 94768, + "Ġtheme": 6912, + "Ġthemed": 48544, + "Ġthemes": 21386, + "Ġthems": 5568, + "Ġthemselves": 5577, + "Ġthen": 1221, + "Ġtheolog": 89502, + "Ġtheological": 69534, + "Ġtheology": 53541, + "Ġtheor": 45720, + "Ġtheorem": 57817, + "Ġtheoret": 25131, + "Ġtheoretical": 31787, + "Ġtheoretically": 62134, + "Ġtheories": 24970, + "Ġtheorists": 82962, + "Ġtheory": 10126, + "Ġther": 8962, + "Ġtherap": 24702, + "Ġtherapeutic": 36371, + "Ġtherapies": 51212, + "Ġtherapist": 41763, + "Ġtherapists": 67516, + "Ġtherapy": 15069, + "Ġthere": 1052, + "Ġthereafter": 44995, + "Ġthereby": 27492, + "Ġtherefore": 8916, + "Ġtherein": 57079, + "Ġthereof": 33266, + "Ġtheres": 91679, + "Ġtherm": 29845, + "Ġthermal": 28387, + "Ġthermo": 96147, + "Ġthermometer": 92197, + "Ġthermostat": 74070, + "Ġthese": 1493, + "Ġthesis": 17394, + "Ġtheta": 18526, + "Ġthey": 807, + "Ġthi": 61328, + "Ġthick": 12045, + "Ġthicker": 58784, + "Ġthickness": 25755, + "Ġthief": 57735, + "Ġthieves": 60567, + "Ġthigh": 38671, + "Ġthighs": 59511, + "Ġthin": 15430, + "Ġthing": 3166, + "Ġthings": 2513, + "Ġthink": 1744, + "Ġthinker": 97536, + "Ġthinkers": 68022, + "Ġthinking": 7274, + "Ġthinks": 15482, + "Ġthinly": 98431, + "Ġthinner": 64255, + "Ġthird": 4843, + "Ġthirds": 93137, + "Ġthirst": 49590, + "Ġthirsty": 97108, + "Ġthirteen": 60659, + "Ġthirty": 26127, + "Ġthis": 419, + "Ġthiên": 131265, + "Ġthiết": 128422, + "Ġthiếu": 130502, + "Ġthiá»ĥu": 141700, + "Ġthiá»ĩn": 130509, + "Ġthiá»ĩt": 141753, + "Ġthiá»ĩu": 132786, + "Ġtho": 39325, + "Ġthor": 72733, + "Ġthorough": 17423, + "Ġthoroughly": 26364, + "Ġthose": 1846, + "Ġthou": 33123, + "Ġthough": 3498, + "Ġthought": 3381, + "Ġthoughtful": 42666, + "Ġthoughts": 11303, + "Ġthous": 7048, + "Ġthousand": 16183, + "Ġthousands": 9037, + "Ġthoáng": 141192, + "Ġthoát": 136122, + "Ġthoại": 129752, + "Ġthoải": 136938, + "Ġthr": 8807, + "Ġthread": 4516, + "ĠthreadIdx": 72971, + "Ġthreaded": 66787, + "Ġthreading": 30159, + "Ġthreads": 14564, + "Ġthreat": 5899, + "Ġthreaten": 39150, + "Ġthreatened": 21023, + "Ġthreatening": 26805, + "Ġthreatens": 47826, + "Ġthreats": 17736, + "Ġthree": 2326, + "Ġthreesome": 68964, + "Ġthresh": 51744, + "Ġthreshold": 12171, + "Ġthresholds": 56140, + "Ġthrew": 22192, + "Ġthrift": 81542, + "Ġthrill": 53804, + "Ġthrilled": 37464, + "Ġthriller": 53361, + "Ġthrilling": 56299, + "Ġthrive": 40872, + "Ġthriving": 52314, + "Ġthro": 59087, + "Ġthroat": 27591, + "Ġthrom": 83564, + "Ġthrone": 43621, + "Ġthrott": 79972, + "Ġthrottle": 42166, + "Ġthrough": 1526, + "Ġthroughout": 6814, + "Ġthroughput": 63762, + "Ġthrow": 2510, + "ĠthrowError": 80780, + "Ġthrowable": 64531, + "Ġthrowing": 21244, + "Ġthrown": 14989, + "Ġthrows": 3787, + "Ġthru": 40078, + "Ġthrust": 29885, + "Ġthu": 77374, + "Ġthugs": 85319, + "Ġthuis": 54284, + "Ġthuisontvangst": 78323, + "Ġthumb": 24050, + "Ġthumbnail": 28574, + "Ġthumbnails": 80065, + "Ġthumbs": 56849, + "Ġthunder": 36843, + "Ġthunk": 78156, + "Ġthus": 8450, + "Ġthuyết": 136957, + "Ġthuyá»ģn": 143897, + "Ġthuê": 131500, + "Ġthuế": 134375, + "ĠthuáºŃn": 131678, + "ĠthuáºŃt": 128859, + "Ġthuá»ijc": 129374, + "Ġthuá»Ļc": 129281, + "Ġthwart": 63688, + "Ġthy": 25174, + "Ġthyroid": 52960, + "Ġthác": 133457, + "Ġthái": 131860, + "Ġtháng": 128526, + "Ġthánh": 143821, + "Ġthân": 128672, + "Ġthé": 75580, + "Ġthép": 142842, + "Ġthêm": 128622, + "Ġthì": 128291, + "Ġthói": 143720, + "Ġthôi": 131021, + "Ġthôn": 138757, + "Ġthông": 92430, + "Ġthù": 136211, + "Ġthú": 131999, + "Ġthúc": 133600, + "ĠthÃłnh": 64061, + "ĠthÃŃ": 135863, + "ĠthÃŃch": 128745, + "ĠthÄĥm": 133171, + "ĠthÆ¡": 130917, + "Ġthư": 129101, + "Ġthương": 128560, + "Ġthượng": 140156, + "ĠthÆ°á»Ľc": 136543, + "Ġthưá»Ŀng": 128454, + "ĠthÆ°á»Łng": 132720, + "Ġthả": 130978, + "Ġthảo": 132972, + "Ġthấp": 132058, + "Ġthất": 130034, + "Ġthấy": 128391, + "Ġthần": 129719, + "Ġthầy": 135560, + "Ġthẩm": 133024, + "Ġthắ": 136139, + "Ġthắng": 131457, + "Ġthẳng": 133863, + "Ġthẻ": 137118, + "Ġthế": 128314, + "ĠtháºŃm": 136023, + "ĠtháºŃn": 136408, + "ĠtháºŃp": 142122, + "ĠtháºŃt": 128869, + "Ġthụ": 135364, + "Ġthá»§": 128799, + "Ġthá»§y": 132601, + "Ġthứ": 128720, + "Ġthức": 128482, + "Ġthừa": 135991, + "Ġthá»±": 139514, + "Ġthá»±c": 128336, + "Ġthá»ĥ": 128254, + "Ġthá»ĭ": 126260, + "Ġthá»ĭt": 134101, + "Ġthá»ı": 139987, + "Ġthá»ıa": 139988, + "Ġthá»ijng": 128573, + "Ġthá»ķ": 140926, + "Ġthá»Ŀ": 139900, + "Ġthá»Ŀi": 128327, + "Ġthợ": 138762, + "Ġthá»Ń": 131885, + "Ġti": 8988, + "Ġtic": 61447, + "Ġtick": 9341, + "Ġticker": 46987, + "Ġticket": 11727, + "Ġtickets": 14403, + "Ġticking": 82337, + "Ġticks": 29957, + "Ġtid": 13112, + "Ġtidak": 27281, + "Ġtidal": 85459, + "Ġtide": 41938, + "Ġtidy": 57805, + "Ġtie": 18131, + "Ġtied": 17339, + "Ġtiempo": 28742, + "Ġtiene": 23332, + "Ġtienen": 46439, + "Ġtiener": 86646, + "Ġtienes": 71097, + "Ġtier": 25996, + "Ġtiers": 63171, + "Ġties": 19818, + "Ġtieten": 71005, + "Ġtiger": 51735, + "Ġtight": 10296, + "Ġtighten": 66745, + "Ġtightened": 84720, + "Ġtightening": 76780, + "Ġtighter": 62962, + "Ġtightly": 38969, + "Ġtijd": 52690, + "Ġtik": 86172, + "Ġtil": 10266, + "Ġtile": 10023, + "ĠtileSize": 93775, + "Ġtiled": 81870, + "Ġtiles": 20493, + "Ġtill": 11956, + "Ġtilt": 36200, + "Ġtilted": 85917, + "Ġtim": 6792, + "Ġtimber": 44788, + "Ġtime": 882, + "ĠtimeStamp": 64276, + "ĠtimeZone": 98764, + "Ġtimed": 25182, + "Ġtimedelta": 42255, + "Ġtimeframe": 69953, + "Ġtimeless": 57005, + "Ġtimeline": 24808, + "Ġtimelines": 87144, + "Ġtimely": 31000, + "Ġtimeout": 9632, + "Ġtimeouts": 91444, + "Ġtimer": 9021, + "Ġtimers": 44522, + "Ġtimes": 3039, + "Ġtimespec": 71156, + "Ġtimestamp": 11441, + "Ġtimestamps": 48781, + "Ġtimestep": 74578, + "Ġtimetable": 66540, + "Ġtimeval": 51334, + "Ġtimezone": 33957, + "Ġtimid": 98049, + "Ġtiming": 18405, + "Ġtimings": 84798, + "Ġtimp": 76645, + "Ġtin": 24790, + "Ġtinder": 89165, + "Ġting": 42227, + "Ġtinha": 94948, + "Ġtink": 90584, + "Ġtint": 48004, + "Ġtiny": 13673, + "Ġtion": 72252, + "Ġtip": 11552, + "Ġtipo": 16298, + "Ġtipos": 63862, + "Ġtipped": 65372, + "Ġtipping": 71193, + "Ġtips": 10414, + "Ġtir": 44612, + "Ġtire": 27287, + "Ġtired": 19227, + "Ġtirelessly": 98129, + "Ġtires": 30700, + "Ġtissue": 19847, + "Ġtissues": 38781, + "Ġtit": 8142, + "Ġtitan": 85512, + "Ġtitanium": 70304, + "Ġtitle": 2265, + "ĠtitleLabel": 60820, + "Ġtitled": 24849, + "Ġtitles": 15311, + "Ġtitre": 63060, + "Ġtits": 29582, + "Ġtitten": 92375, + "Ġtitular": 80105, + "Ġtitulo": 51063, + "Ġtiên": 128924, + "Ġtiêu": 128575, + "Ġtiế": 128606, + "Ġtiến": 129263, + "Ġtiếng": 128607, + "Ġtiếp": 128424, + "Ġtiết": 128861, + "Ġtiá»ģm": 141673, + "Ġtiá»ģn": 128515, + "Ġtiá»ĥu": 132037, + "Ġtiá»ĩn": 129698, + "Ġtj": 65254, + "Ġtjejer": 71907, + "Ġtk": 17162, + "Ġtkinter": 40490, + "Ġtl": 29796, + "Ġtls": 55026, + "Ġtm": 17333, + "Ġtmp": 4174, + "Ġtmpl": 79839, + "Ġtn": 43308, + "Ġto": 311, + "ĠtoArray": 78190, + "ĠtoDate": 81103, + "ĠtoItem": 98360, + "ĠtoJSON": 79522, + "ĠtoJson": 65257, + "ĠtoReturn": 81998, + "ĠtoString": 14330, + "Ġtoast": 22405, + "Ġtoasted": 92872, + "Ġtoaster": 89818, + "Ġtoastr": 58352, + "Ġtob": 24406, + "Ġtobacco": 26433, + "Ġtoc": 54861, + "Ġtoch": 97698, + "Ġtod": 11530, + "Ġtoda": 40178, + "Ġtodas": 30101, + "Ġtodav": 140022, + "ĠtodavÃŃa": 140023, + "Ġtoday": 3351, + "Ġtodd": 37934, + "Ġtoddler": 51235, + "Ġtoddlers": 78835, + "Ġtodo": 11804, + "Ġtodos": 15519, + "Ġtoe": 30216, + "Ġtoen": 99034, + "Ġtoes": 44613, + "Ġtofu": 79147, + "Ġtog": 13144, + "Ġtogether": 3786, + "Ġtogg": 53806, + "Ġtoggle": 14999, + "Ġtoi": 75320, + "Ġtoile": 48675, + "Ġtoilet": 26214, + "Ġtoilets": 62187, + "Ġtok": 9628, + "Ġtoken": 3950, + "ĠtokenId": 87038, + "Ġtokenize": 77651, + "Ġtokenizer": 45958, + "Ġtokens": 11211, + "Ġtol": 38574, + "Ġtold": 3229, + "Ġtoler": 14885, + "Ġtolerance": 24098, + "Ġtolerant": 73482, + "Ġtolerate": 49034, + "Ġtolerated": 65341, + "Ġtoll": 25851, + "Ġtolua": 65524, + "Ġtom": 10180, + "Ġtomar": 63711, + "Ġtomato": 41020, + "Ġtomatoes": 40513, + "Ġtomb": 43080, + "Ġtome": 84742, + "Ġtomorrow": 16577, + "Ġton": 8766, + "Ġtone": 16232, + "Ġtoned": 97731, + "Ġtones": 41976, + "Ġtong": 21365, + "Ġtongue": 24459, + "Ġtongues": 88335, + "Ġtonic": 98205, + "Ġtonight": 17913, + "Ġtonnes": 50921, + "Ġtons": 19608, + "Ġtonumber": 66296, + "Ġtoo": 2238, + "Ġtook": 3867, + "Ġtool": 5392, + "ĠtoolStrip": 46902, + "Ġtoolbar": 25942, + "Ġtoolbox": 67870, + "Ġtoolkit": 65894, + "Ġtools": 7375, + "Ġtooltip": 25451, + "Ġtooltips": 91234, + "Ġtooth": 25507, + "Ġtop": 1909, + "ĠtopLeft": 79103, + "ĠtopLevel": 90420, + "Ġtopic": 8544, + "Ġtopical": 65276, + "Ġtopics": 13347, + "Ġtoplant": 135368, + "Ġtoplantı": 135369, + "Ġtopo": 72519, + "Ġtopology": 44882, + "Ġtopp": 44041, + "Ġtopped": 39801, + "Ġtopping": 60276, + "Ġtoppings": 89671, + "Ġtops": 32422, + "Ġtor": 5683, + "Ġtorch": 7834, + "Ġtorchvision": 67825, + "Ġtore": 62008, + "Ġtorino": 70443, + "Ġtorment": 72670, + "Ġtorn": 21145, + "Ġtornado": 44172, + "Ġtorpedo": 91883, + "Ġtorque": 41031, + "Ġtorrent": 30575, + "Ġtorrents": 97831, + "Ġtors": 94460, + "Ġtorso": 63200, + "Ġtort": 16426, + "Ġtorture": 29567, + "Ġtortured": 55934, + "Ġtoss": 25187, + "Ġtossed": 50149, + "Ġtossing": 84051, + "Ġtostring": 70890, + "Ġtot": 2401, + "Ġtotal": 2790, + "ĠtotalCount": 59713, + "ĠtotalPages": 92294, + "ĠtotalPrice": 72205, + "ĠtotalTime": 91746, + "Ġtotaled": 88418, + "Ġtotalement": 93673, + "Ġtotaling": 81123, + "Ġtotalitarian": 89264, + "Ġtotally": 12473, + "Ġtotalmente": 80141, + "Ġtotals": 43645, + "Ġtote": 87519, + "Ġtou": 14776, + "Ġtouch": 5796, + "Ġtouchdown": 25027, + "Ġtouchdowns": 42779, + "Ġtouched": 23939, + "Ġtouches": 28627, + "Ġtouching": 30587, + "Ġtouchscreen": 76757, + "Ġtough": 11045, + "Ġtougher": 48231, + "Ġtoughest": 59306, + "Ġtoughness": 76218, + "Ġtoujours": 42993, + "Ġtour": 7216, + "Ġtoured": 90613, + "Ġtouring": 47451, + "Ġtourism": 30983, + "Ġtourist": 29970, + "Ġtourists": 31653, + "Ġtournament": 16129, + "Ġtournaments": 39061, + "Ġtours": 30161, + "Ġtous": 25642, + "Ġtout": 16559, + "Ġtoute": 42805, + "Ġtouted": 66428, + "Ġtoutes": 39644, + "Ġtow": 15813, + "Ġtoward": 8841, + "Ġtowards": 6974, + "Ġtowel": 42613, + "Ġtowels": 52884, + "Ġtower": 21271, + "Ġtowering": 86694, + "Ġtowers": 39725, + "Ġtowing": 88692, + "Ġtown": 6290, + "Ġtowns": 24824, + "Ġtownship": 87175, + "Ġtox": 40469, + "Ġtoxic": 20836, + "Ġtoxicity": 56911, + "Ġtoxin": 97434, + "Ġtoxins": 61086, + "Ġtoy": 21357, + "Ġtoys": 23069, + "Ġtoán": 130051, + "ĠtoÃłn": 128401, + "Ġtp": 18101, + "Ġtph": 73779, + "Ġtpl": 60979, + "Ġtq": 35885, + "Ġtqdm": 39404, + "Ġtr": 489, + "Ġtra": 9586, + "Ġtrab": 17598, + "Ġtrabaj": 38874, + "Ġtrabajar": 94565, + "Ġtrabajo": 38218, + "Ġtrabal": 36268, + "Ġtrabalho": 54639, + "Ġtrace": 11655, + "Ġtraceback": 46058, + "Ġtraced": 50300, + "Ġtracer": 64306, + "Ġtraces": 34383, + "Ġtracing": 45415, + "Ġtrack": 3754, + "Ġtracked": 33056, + "Ġtracker": 28331, + "Ġtrackers": 90982, + "Ġtracking": 14849, + "Ġtracks": 13918, + "Ġtract": 41829, + "Ġtraction": 46855, + "Ġtractor": 58685, + "Ġtrad": 4685, + "Ġtrade": 6559, + "Ġtraded": 30107, + "Ġtrademark": 31364, + "Ġtrademarks": 41034, + "Ġtrader": 44365, + "Ġtraders": 34587, + "Ġtrades": 30342, + "Ġtrading": 11133, + "Ġtradition": 13815, + "Ġtraditional": 8606, + "Ġtraditionally": 35242, + "Ġtraditions": 30906, + "Ġtraf": 98888, + "Ġtraff": 8258, + "Ġtraffic": 9442, + "Ġtrafficking": 33463, + "Ġtrag": 71602, + "Ġtraged": 27276, + "Ġtragedies": 98770, + "Ġtragedy": 30826, + "Ġtragic": 34179, + "Ġtrail": 8849, + "Ġtrailed": 96071, + "Ġtrailer": 19253, + "Ġtrailers": 49705, + "Ġtrailing": 27748, + "Ġtrails": 32865, + "Ġtrain": 5426, + "Ġtrainable": 87661, + "Ġtrained": 16176, + "Ġtrainer": 28894, + "Ġtrainers": 50091, + "Ġtraining": 4862, + "Ġtrains": 27688, + "Ġtrait": 17567, + "Ġtraitement": 88614, + "Ġtraits": 24055, + "Ġtraj": 62028, + "Ġtrajectories": 85548, + "Ġtrajectory": 34682, + "Ġtram": 52245, + "Ġtran": 23686, + "Ġtrance": 97431, + "Ġtranny": 69598, + "Ġtranqu": 40564, + "Ġtranquil": 66940, + "Ġtrans": 1356, + "Ġtransaction": 7745, + "Ġtransactions": 14131, + "Ġtransc": 41490, + "Ġtranscend": 73709, + "Ġtranscript": 35715, + "Ġtranscription": 45840, + "Ġtranscripts": 60312, + "Ġtransf": 19839, + "Ġtransfer": 8317, + "Ġtransferred": 22409, + "Ġtransferring": 49951, + "Ġtransfers": 30611, + "Ġtransform": 5165, + "Ġtransformation": 17991, + "Ġtransformations": 52970, + "Ġtransformative": 86335, + "Ġtransformed": 23507, + "Ġtransformer": 42578, + "Ġtransformers": 86870, + "Ġtransforming": 45790, + "Ġtransforms": 28475, + "Ġtransgender": 27208, + "Ġtransient": 40558, + "Ġtransistor": 97941, + "Ġtransit": 24065, + "Ġtransition": 9142, + "Ġtransitional": 65643, + "Ġtransitioning": 72094, + "Ġtransitions": 33592, + "Ġtransl": 11949, + "Ġtranslate": 14683, + "ĠtranslateY": 87749, + "Ġtranslated": 24531, + "Ġtranslates": 46918, + "Ġtranslating": 66271, + "Ġtranslation": 14468, + "Ġtranslations": 36693, + "Ġtranslator": 45488, + "Ġtranslators": 72704, + "Ġtranslucent": 84264, + "Ġtransmission": 18368, + "Ġtransmissions": 90996, + "Ġtransmit": 29282, + "Ġtransmitted": 33599, + "Ġtransmitter": 61110, + "Ġtransmitting": 77668, + "Ġtransparency": 27231, + "Ġtransparent": 17821, + "Ġtransplant": 42829, + "Ġtransplantation": 97602, + "Ġtransport": 7557, + "Ġtransportation": 17903, + "Ġtransporte": 95388, + "Ġtransported": 39360, + "Ġtransporter": 72465, + "Ġtransporting": 66657, + "Ġtransports": 68069, + "Ġtranspose": 50923, + "Ġtrap": 22862, + "Ġtrapped": 30591, + "Ġtrapping": 92569, + "Ġtraps": 44560, + "Ġtras": 33164, + "Ġtrash": 22854, + "Ġtrat": 40821, + "Ġtrata": 63267, + "Ġtratamiento": 88645, + "Ġtratt": 92934, + "Ġtrauma": 29888, + "Ġtraumat": 98036, + "Ġtraumatic": 53186, + "Ġtrav": 10137, + "Ġtrava": 57573, + "Ġtravail": 41675, + "Ġtravel": 5821, + "Ġtraveled": 30696, + "Ġtraveler": 62765, + "Ġtravelers": 39286, + "Ġtraveling": 20971, + "Ġtravelled": 45268, + "Ġtraveller": 85194, + "Ġtravellers": 58968, + "Ġtravelling": 30729, + "Ġtravels": 34192, + "Ġtravers": 29417, + "Ġtraversal": 56302, + "Ġtraverse": 37546, + "Ġtravés": 49352, + "Ġtray": 34688, + "Ġtrays": 91725, + "Ġtre": 4258, + "Ġtread": 47714, + "Ġtreadmill": 84646, + "Ġtreason": 78259, + "Ġtreasure": 31626, + "Ġtreasurer": 95391, + "Ġtreasures": 58849, + "Ġtreasury": 78253, + "Ġtreat": 4228, + "Ġtreated": 11758, + "Ġtreaties": 75377, + "Ġtreating": 26625, + "Ġtreatment": 6380, + "Ġtreatments": 22183, + "Ġtreats": 31739, + "Ġtreaty": 37897, + "Ġtrebuie": 94558, + "Ġtree": 4916, + "ĠtreeNode": 65275, + "Ġtrees": 12408, + "Ġtreff": 75687, + "Ġtreffen": 51873, + "Ġtrek": 44588, + "Ġtrem": 18162, + "Ġtrembling": 91988, + "Ġtremend": 23451, + "Ġtremendous": 26942, + "Ġtremendously": 71323, + "Ġtren": 73600, + "Ġtrench": 53252, + "Ġtrenches": 94978, + "Ġtrend": 9149, + "Ġtrending": 50850, + "Ġtrends": 18339, + "Ġtrendy": 57830, + "Ġtres": 30623, + "Ġtrespass": 74230, + "ĠtreÅĽci": 140889, + "Ġtrg": 97260, + "Ġtri": 2406, + "Ġtrial": 9091, + "Ġtrials": 19080, + "Ġtriang": 76405, + "Ġtriangle": 21495, + "Ġtriangles": 42446, + "Ġtriangular": 65494, + "Ġtrib": 13802, + "Ġtribal": 39389, + "Ġtribe": 33571, + "Ġtribes": 38403, + "Ġtribunal": 62892, + "Ġtribute": 34391, + "Ġtrick": 14068, + "Ġtrickle": 97344, + "Ġtricks": 28762, + "Ġtricky": 33453, + "Ġtrie": 59067, + "Ġtried": 6679, + "Ġtries": 16297, + "Ġtrif": 89103, + "Ġtrig": 52933, + "Ġtrigger": 8183, + "Ġtriggered": 22119, + "Ġtriggering": 51505, + "Ġtriggers": 30754, + "Ġtrillion": 31510, + "Ġtrilogy": 56786, + "Ġtrim": 11013, + "Ġtrimest": 74010, + "Ġtrimmed": 50981, + "Ġtrimming": 81887, + "Ġtrio": 33962, + "Ġtrip": 8411, + "Ġtriple": 23725, + "Ġtriples": 88561, + "Ġtriplet": 98709, + "Ġtripod": 91054, + "Ġtrips": 22466, + "Ġtritur": 66141, + "Ġtrium": 91806, + "Ġtriumph": 37486, + "Ġtrivia": 72932, + "Ġtrivial": 35647, + "Ġtriá»ĥn": 128531, + "Ġtriá»ĩu": 129077, + "Ġtro": 8185, + "Ġtroch": 135218, + "ĠtrochÄĻ": 135219, + "Ġtrois": 55221, + "Ġtroll": 57434, + "Ġtrolling": 89510, + "Ġtrolls": 64585, + "Ġtrom": 64536, + "Ġtrong": 69086, + "Ġtroop": 70716, + "Ġtroops": 16886, + "Ġtrop": 21267, + "Ġtrope": 84148, + "Ġtrophies": 53444, + "Ġtrophy": 37682, + "Ġtropical": 34048, + "Ġtrot": 56577, + "Ġtrotz": 90039, + "Ġtrou": 25410, + "Ġtroub": 19904, + "Ġtrouble": 12264, + "Ġtroubled": 41032, + "Ġtroubles": 34565, + "Ġtroubleshooting": 68671, + "Ġtroublesome": 77418, + "Ġtroubling": 51229, + "Ġtrough": 77965, + "Ġtrous": 60028, + "Ġtrousers": 67676, + "Ġtrout": 79727, + "Ġtrouve": 70351, + "Ġtrouver": 45424, + "Ġtrouvé": 99062, + "Ġtrov": 53539, + "Ġtrovare": 82583, + "Ġtruck": 10855, + "Ġtrucks": 26763, + "Ġtrue": 830, + "Ġtruly": 9428, + "Ġtrump": 38055, + "Ġtrumpet": 91259, + "Ġtrunc": 62850, + "Ġtruncate": 56772, + "Ġtruncated": 59756, + "Ġtrunk": 37311, + "Ġtrust": 6950, + "Ġtrusted": 21789, + "Ġtrustee": 75766, + "Ġtrustees": 78711, + "Ġtrusting": 68244, + "Ġtrusts": 65514, + "Ġtrustworthy": 55942, + "Ġtruth": 8046, + "Ġtruthful": 89867, + "Ġtruths": 52119, + "Ġtruyá»ģn": 129045, + "Ġtruyá»ĩn": 135867, + "ĠtrwaÅĤ": 141124, + "Ġtrx": 73021, + "Ġtry": 1430, + "Ġtrying": 4460, + "Ġtrách": 131368, + "Ġtrái": 130641, + "Ġtránh": 131282, + "Ġträ": 85623, + "Ġtrès": 24901, + "Ġtrên": 128273, + "Ġtrês": 81529, + "Ġtrì": 132630, + "Ġtrình": 128361, + "Ġtrò": 129659, + "Ġtrông": 140494, + "Ġtrùng": 136658, + "Ġtrú": 136003, + "Ġtrúc": 134383, + "ĠtrÃł": 130491, + "ĠtrÃŃ": 128597, + "ĠtrÄĥm": 137451, + "Ġtrưng": 138397, + "Ġtrương": 140996, + "ĠtrÆ°á»Ľc": 128371, + "Ġtrưá»Ŀng": 128331, + "ĠtrÆ°á»Łng": 129528, + "Ġtrạng": 129916, + "Ġtrả": 129477, + "Ġtrải": 132465, + "Ġtrầm": 141631, + "Ġtrần": 142035, + "Ġtrắng": 130595, + "Ġtrẻ": 128634, + "ĠtráºŃn": 130121, + "Ġtrợ": 128886, + "Ġtrụ": 134902, + "Ġtrứng": 138432, + "Ġtrừ": 137329, + "Ġtrữ": 141207, + "Ġtrá»±c": 129372, + "Ġtrá»ĭ": 128466, + "Ġtrá»įng": 128550, + "Ġtrá»ĵng": 135149, + "Ġtrá»Ŀi": 130868, + "Ġtrợ": 128570, + "Ġts": 10591, + "Ġtsl": 46386, + "Ġtslib": 94261, + "Ġtslint": 48681, + "Ġtsp": 37920, + "Ġtst": 71707, + "Ġtsunami": 71055, + "Ġtt": 17853, + "Ġttk": 66464, + "Ġttl": 53932, + "Ġtts": 98540, + "Ġtty": 60067, + "Ġtu": 9765, + "Ġtua": 63884, + "Ġtub": 14938, + "Ġtube": 13705, + "Ġtuberculosis": 87134, + "Ġtubes": 32983, + "Ġtubing": 73902, + "Ġtucked": 60118, + "Ġtud": 70059, + "Ġtudo": 55696, + "Ġtug": 81615, + "Ġtuition": 37938, + "Ġtul": 60936, + "Ġtum": 15394, + "Ġtumble": 86950, + "Ġtumblr": 45300, + "Ġtumor": 35154, + "Ġtumors": 54971, + "Ġtumult": 84577, + "Ġtun": 11460, + "Ġtuna": 73957, + "Ġtune": 25240, + "Ġtuned": 32419, + "Ġtuner": 90695, + "Ġtunes": 53990, + "Ġtung": 84207, + "Ġtuning": 41338, + "Ġtunnel": 25629, + "Ġtunnels": 52727, + "Ġtuo": 62158, + "Ġtup": 57385, + "Ġtuple": 14405, + "Ġtuples": 45225, + "Ġtur": 13232, + "Ġturb": 28012, + "Ġturbine": 65170, + "Ġturbines": 75675, + "Ġturbo": 47027, + "Ġturbulence": 94067, + "Ġturbulent": 82221, + "Ġturf": 58990, + "Ġturkey": 44305, + "Ġturmoil": 62255, + "Ġturn": 2484, + "Ġturnaround": 66510, + "Ġturned": 6519, + "Ġturning": 13054, + "Ġturno": 97867, + "Ġturnout": 51919, + "Ġturnover": 47539, + "Ġturnovers": 80627, + "Ġturns": 10577, + "Ġturquoise": 95947, + "Ġturret": 64386, + "Ġturtle": 36089, + "Ġturtles": 71403, + "Ġtus": 40684, + "Ġtussen": 88722, + "Ġtut": 18060, + "Ġtutor": 25302, + "Ġtutorial": 21514, + "Ġtutorials": 43812, + "Ġtutoring": 77043, + "Ġtutors": 82932, + "Ġtutte": 74157, + "Ġtutti": 43280, + "Ġtutto": 51182, + "Ġtuy": 130544, + "Ġtuyên": 135593, + "Ġtuyến": 130980, + "Ġtuyá»ĥn": 131535, + "Ġtuyá»ĩt": 130545, + "Ġtuần": 130681, + "Ġtuá»ķi": 128790, + "Ġtv": 11086, + "Ġtvb": 85907, + "Ġtw": 4384, + "Ġtwe": 14545, + "Ġtweak": 51043, + "Ġtweaked": 88460, + "Ġtweaking": 88247, + "Ġtweaks": 62337, + "Ġtwee": 72608, + "Ġtween": 65517, + "Ġtweet": 11809, + "Ġtweeted": 25955, + "Ġtweeting": 68715, + "Ġtweets": 23150, + "Ġtwelve": 29235, + "Ġtwenties": 96596, + "Ġtwentieth": 61420, + "Ġtwenty": 17073, + "Ġtwice": 10917, + "Ġtwig": 51246, + "Ġtwilight": 92622, + "Ġtwin": 27397, + "Ġtwink": 96400, + "Ġtwins": 43636, + "Ġtwist": 26646, + "Ġtwisted": 37717, + "Ġtwisting": 79554, + "Ġtwists": 61890, + "Ġtwitch": 59413, + "Ġtwitter": 22272, + "Ġtwo": 1378, + "Ġtwó": 141779, + "Ġtwórc": 141780, + "Ġtx": 9854, + "Ġtxn": 49721, + "Ġtxt": 7932, + "Ġty": 13580, + "Ġtying": 62043, + "Ġtylko": 52044, + "Ġtym": 38377, + "Ġtyp": 3582, + "Ġtype": 943, + "ĠtypeId": 89747, + "ĠtypeName": 52600, + "Ġtypealias": 87230, + "Ġtyped": 31969, + "Ġtypedef": 13501, + "Ġtypeid": 86806, + "Ġtypename": 11246, + "Ġtypeof": 7804, + "Ġtypes": 4494, + "Ġtypical": 14260, + "Ġtypically": 11136, + "Ġtyping": 19496, + "Ġtypings": 24921, + "ĠtypingsJapgolly": 71640, + "ĠtypingsSlinky": 59834, + "Ġtypo": 85105, + "Ġtypography": 96282, + "Ġtyr": 41949, + "Ġtyranny": 76901, + "Ġtyre": 72928, + "Ġtyres": 68693, + "Ġtys": 77439, + "Ġtytu": 136602, + "ĠtytuÅĤ": 136603, + "Ġtz": 42528, + "Ġtá": 79584, + "Ġtác": 128400, + "Ġtái": 135578, + "Ġtán": 141873, + "Ġtâm": 128396, + "Ġtây": 141567, + "Ġtão": 95672, + "Ġtä": 50790, + "Ġtäglich": 140729, + "Ġté": 46189, + "Ġtéc": 86163, + "Ġtécn": 70523, + "Ġtécnica": 131322, + "Ġtécnico": 133251, + "Ġtélé": 42016, + "Ġtélécharg": 141504, + "Ġtéléphone": 98907, + "Ġtér": 89635, + "Ġtérmino": 139413, + "Ġtêm": 98860, + "Ġtên": 129064, + "Ġtête": 81967, + "Ġtë": 41095, + "Ġtìm": 128611, + "Ġtình": 128419, + "Ġtòa": 136977, + "Ġtóc": 132591, + "Ġtô": 130003, + "Ġtôi": 128296, + "Ġtôn": 134342, + "Ġtö": 63573, + "Ġtù": 135735, + "Ġtùy": 135736, + "Ġtú": 89218, + "Ġtúi": 136610, + "Ġtü": 127669, + "Ġtüket": 134783, + "Ġtüm": 128817, + "Ġtürl": 136166, + "Ġtürlü": 136167, + "ĠtÃł": 133464, + "ĠtÃłi": 128556, + "ĠtÃłn": 138878, + "ĠtÃłu": 133465, + "ĠtÃŃ": 76407, + "ĠtÃŃch": 128761, + "ĠtÃŃn": 129913, + "ĠtÃŃnh": 128439, + "ĠtÃŃtulo": 84951, + "ĠtÄ©nh": 139843, + "ĠtÄĥng": 128529, + "ĠtÄĻ": 134692, + "ĠtÅĤ": 141082, + "ĠtÅĤum": 141083, + "ĠtÅĤumacz": 141084, + "Ġtư": 128379, + "Ġtươi": 133683, + "Ġtương": 129509, + "Ġtượng": 129014, + "ĠtÆ°á»Ľng": 131943, + "Ġtưá»Ŀng": 137173, + "ĠtÆ°á»Łng": 129608, + "Ġtại": 128272, + "Ġtạm": 139511, + "Ġtạo": 128425, + "Ġtạp": 138628, + "Ġtả": 134358, + "Ġtải": 130897, + "Ġtấm": 136562, + "Ġtấn": 133270, + "Ġtất": 129202, + "Ġtầ": 131519, + "Ġtầm": 134059, + "Ġtầng": 131520, + "Ġtắc": 135352, + "Ġtắm": 137565, + "Ġtắt": 141070, + "Ġtặng": 133269, + "Ġtế": 128384, + "ĠtáºŃn": 132247, + "ĠtáºŃp": 128553, + "Ġtá»": 46500, + "Ġtụ": 126411, + "Ġtục": 128884, + "Ġtá»§": 135783, + "Ġtứ": 129594, + "Ġtức": 129595, + "Ġtừ": 94576, + "Ġtừng": 128981, + "Ġtá»±": 128377, + "Ġtá»·": 129427, + "Ġtá»ĩ": 139091, + "Ġtá»ī": 126450, + "Ġtá»īnh": 129018, + "Ġtá»ĭch": 131443, + "Ġtá»ı": 132367, + "Ġtá»ij": 129556, + "Ġtá»ijc": 133061, + "Ġtá»iji": 129597, + "Ġtá»ijt": 128418, + "Ġtá»ĵ": 137105, + "Ġtá»ĵn": 137106, + "Ġtá»ķ": 126341, + "Ġtá»ķng": 129822, + "Ġtá»Ļc": 133952, + "Ġtá»Ļi": 134057, + "ĠtỼi": 128447, + "Ġtá»Ŀ": 137971, + "Ġtá»Ń": 128804, + "Ġu": 575, + "ĠuLocal": 81368, + "Ġua": 45559, + "Ġuart": 46868, + "Ġub": 17910, + "Ġuber": 92039, + "Ġubic": 62774, + "Ġubiqu": 52236, + "Ġubiquitous": 63464, + "Ġubuntu": 84214, + "Ġuc": 25721, + "Ġucfirst": 66921, + "Ġuchar": 61646, + "Ġucwords": 95404, + "Ġucz": 87426, + "Ġuczni": 143697, + "Ġuczniów": 143698, + "Ġud": 26994, + "Ġuda": 142590, + "ĠudaÅĤo": 142591, + "Ġuden": 92755, + "Ġudp": 49219, + "ĠudziaÅĤ": 131905, + "Ġue": 60098, + "Ġuf": 54051, + "Ġug": 67656, + "Ġugl": 91118, + "Ġugly": 27261, + "Ġuh": 43744, + "Ġui": 7505, + "Ġuid": 14617, + "Ġuint": 2622, + "Ġuintptr": 38190, + "Ġuit": 24302, + "Ġuk": 14670, + "ĠukÅĤad": 135033, + "Ġul": 8557, + "Ġula": 129679, + "ĠulaÅŁ": 129680, + "Ġulcer": 95871, + "Ġull": 52296, + "Ġullam": 77008, + "Ġulong": 41991, + "Ġult": 8507, + "Ġultimate": 16724, + "Ġultimately": 13653, + "Ġultimo": 92006, + "Ġultr": 36132, + "Ġultra": 23998, + "Ġultrasound": 62759, + "Ġulus": 141884, + "Ġuluslararası": 141885, + "Ġum": 4443, + "Ġuma": 10608, + "Ġumb": 36681, + "Ġumbrella": 47898, + "Ġumie": 142903, + "ĠumiejÄĻ": 142904, + "ĠumiejÄĻtn": 142905, + "ĠumiejÄĻtnoÅĽci": 142906, + "Ġumo": 141252, + "Ġumoż": 141253, + "Ġumożli": 141254, + "Ġumożliw": 141255, + "Ġumożliwia": 141256, + "Ġump": 85042, + "Ġun": 650, + "Ġuna": 5093, + "Ġunab": 95618, + "Ġunable": 11889, + "Ġunacceptable": 42985, + "Ġunaffected": 77522, + "Ġuname": 74474, + "Ġunanim": 39975, + "Ġunanimous": 67409, + "Ġunanimously": 64078, + "Ġunanswered": 75447, + "Ġunarmed": 63899, + "Ġunary": 75678, + "Ġunas": 76826, + "Ġunauthorized": 44471, + "Ġunavailable": 34987, + "Ġunavoid": 77202, + "Ġunavoidable": 83016, + "Ġunaware": 40647, + "Ġunb": 41544, + "Ġunbe": 65571, + "Ġunbearable": 98830, + "Ġunbelie": 38937, + "Ġunbelievable": 51129, + "Ġunbiased": 73215, + "Ġunborn": 85233, + "Ġunc": 20815, + "Ġuncert": 23082, + "Ġuncertain": 35118, + "Ġuncertainties": 68767, + "Ġuncertainty": 26826, + "Ġunchanged": 34857, + "Ġunchecked": 53191, + "Ġuncle": 37494, + "Ġunclear": 24416, + "Ġuncomfort": 26726, + "Ġuncomfortable": 28113, + "Ġuncomment": 62073, + "Ġuncommon": 40196, + "Ġuncomp": 65537, + "Ġuncompressed": 92382, + "Ġuncon": 25062, + "Ġunconditional": 83563, + "Ġunconscious": 39611, + "Ġunconstitutional": 52825, + "Ġuncont": 69993, + "Ġunconventional": 72878, + "Ġuncover": 43963, + "Ġuncovered": 42422, + "Ġund": 2030, + "Ġunde": 58895, + "Ġundead": 69104, + "Ġundecided": 93455, + "Ġundef": 31144, + "Ġundefeated": 92476, + "Ġundefined": 5614, + "Ġundeniable": 80732, + "Ġunder": 1212, + "Ġunderage": 87012, + "Ġundercover": 57481, + "Ġundercut": 95009, + "Ġunderestimate": 78483, + "Ġunderestimated": 86036, + "Ġundergo": 36671, + "Ġundergoing": 46297, + "Ġundergone": 63138, + "Ġundergrad": 85072, + "Ġundergraduate": 40434, + "Ġunderground": 25260, + "Ġunderline": 51874, + "Ġunderlying": 16533, + "Ġundermin": 41086, + "Ġundermine": 41947, + "Ġundermined": 79517, + "Ġundermines": 95136, + "Ġundermining": 77369, + "Ġunderneath": 29356, + "Ġunderrated": 93679, + "Ġunders": 25373, + "Ġunderscore": 52109, + "Ġunderscores": 71790, + "Ġunderside": 95221, + "Ġunderst": 14316, + "Ġunderstand": 3535, + "Ġunderstandable": 48739, + "Ġunderstandably": 86998, + "Ġunderstanding": 8660, + "Ġunderstands": 30769, + "Ġunderstood": 15985, + "Ġundert": 18649, + "Ġundertake": 48543, + "Ġundertaken": 44079, + "Ġundertaking": 48226, + "Ġunderwater": 45374, + "Ġunderway": 37099, + "Ġunderwear": 54144, + "Ġunderwent": 53993, + "Ġunderworld": 99125, + "Ġundes": 55738, + "Ġundesirable": 76244, + "Ġundis": 80937, + "Ġundisclosed": 82857, + "Ġundo": 28721, + "Ġundocumented": 46657, + "Ġundone": 78741, + "Ġundoubtedly": 36948, + "Ġundue": 90542, + "Ġune": 6185, + "Ġunearth": 78575, + "Ġuneasy": 82272, + "Ġunemployed": 48889, + "Ġunemployment": 25608, + "Ġunequal": 77195, + "Ġunequiv": 82935, + "Ġunethical": 88635, + "Ġuneven": 60337, + "Ġunexpected": 16500, + "Ġunexpectedly": 50609, + "Ġunf": 9474, + "Ġunfair": 27643, + "Ġunfairly": 75625, + "Ġunfamiliar": 49283, + "Ġunfavor": 80283, + "Ġunfavorable": 91971, + "Ġunfinished": 63616, + "Ġunfit": 93467, + "Ġunfold": 27472, + "Ġunfolded": 75682, + "Ġunfolding": 32731, + "Ġunfolds": 89665, + "Ġunfore": 95591, + "Ġunforgettable": 59998, + "Ġunfortunate": 33607, + "Ġunfortunately": 25822, + "Ġung": 50490, + "Ġunge": 66796, + "Ġunh": 30075, + "Ġunhappy": 42151, + "Ġunhealthy": 52708, + "Ġunheard": 85460, + "Ġuni": 47886, + "Ġunic": 87549, + "Ġunicode": 26077, + "Ġunicorn": 81830, + "Ġunidad": 96722, + "Ġunidades": 94431, + "Ġunidentified": 64478, + "Ġunified": 42690, + "Ġuniform": 13794, + "Ġuniformly": 77809, + "Ġuniforms": 44133, + "Ġunify": 85340, + "Ġunilateral": 85878, + "Ġunimagin": 86112, + "Ġunin": 70013, + "Ġuninitialized": 97287, + "Ġunins": 65103, + "Ġuninstall": 53635, + "Ġuninsured": 81607, + "Ġunint": 34777, + "Ġunintended": 72343, + "Ġunintention": 69476, + "Ġuninterrupted": 98150, + "Ġunion": 11300, + "Ġunions": 28490, + "Ġuniq": 54152, + "Ġuniqu": 28021, + "Ġunique": 4911, + "Ġuniquely": 41229, + "Ġuniqueness": 59057, + "Ġunit": 4982, + "ĠunitOfWork": 84053, + "Ġunite": 51596, + "Ġunited": 28192, + "Ġunits": 8153, + "Ġunittest": 19905, + "Ġunity": 30326, + "Ġunivers": 10919, + "Ġuniversal": 20178, + "Ġuniversally": 60428, + "Ġuniverse": 15494, + "Ġuniversities": 23106, + "Ġuniversity": 12103, + "Ġunix": 51866, + "Ġunjust": 51483, + "Ġunk": 64431, + "Ġunknow": 83544, + "Ġunknown": 9788, + "Ġunl": 64885, + "Ġunlaw": 38857, + "Ġunlawful": 44422, + "Ġunle": 37726, + "Ġunleash": 77435, + "Ġunleashed": 68264, + "Ġunless": 7241, + "Ġunlike": 19836, + "Ġunlikely": 17367, + "Ġunlimited": 26764, + "Ġunlink": 47479, + "Ġunload": 62714, + "Ġunloaded": 87706, + "Ġunlock": 15055, + "Ġunlocked": 35618, + "Ġunlocking": 79378, + "Ġunlocks": 73191, + "Ġunlucky": 88444, + "Ġunm": 37431, + "Ġunmanned": 87840, + "Ġunmarried": 94487, + "Ġunmatched": 68377, + "Ġunmist": 91258, + "Ġunmistak": 95107, + "Ġunn": 16950, + "Ġunnamed": 52773, + "Ġunnatural": 80746, + "Ġunnecessarily": 82374, + "Ġunnecessary": 25165, + "Ġunnoticed": 76923, + "Ġuno": 23895, + "Ġunofficial": 56651, + "Ġunordered": 55733, + "Ġunos": 52898, + "Ġunp": 21624, + "Ġunpack": 31065, + "Ġunpaid": 54084, + "Ġunparalleled": 70157, + "Ġunpl": 83845, + "Ġunpleasant": 46889, + "Ġunpopular": 65758, + "Ġunprecedented": 29969, + "Ġunpredict": 43596, + "Ġunpredictable": 49135, + "Ġunprotected": 93351, + "Ġunpublished": 84217, + "Ġunquestion": 74533, + "Ġunr": 40380, + "Ġunravel": 73973, + "Ġunre": 37123, + "Ġunreachable": 69322, + "Ġunread": 60291, + "Ġunreal": 49104, + "Ġunrealistic": 70885, + "Ġunreasonable": 58425, + "Ġunrecognized": 96139, + "Ġunregister": 62113, + "Ġunrelated": 45205, + "Ġunreliable": 66029, + "Ġunresolved": 80161, + "Ġunrest": 58222, + "Ġunrestricted": 77839, + "Ġuns": 6975, + "Ġunsafe": 19860, + "Ġunseen": 63133, + "Ġunser": 23898, + "Ġunsere": 44367, + "Ġunserem": 77012, + "Ġunseren": 66482, + "Ġunserer": 44160, + "Ġunserialize": 75494, + "Ġunset": 18000, + "Ġunsett": 67670, + "Ġunsettling": 90980, + "Ġunsigned": 3774, + "Ġunspecified": 54311, + "Ġunst": 64790, + "Ġunstable": 44211, + "Ġunstoppable": 92664, + "Ġunsub": 79204, + "Ġunsubscribe": 44374, + "Ġunsuccessful": 44925, + "Ġunsuccessfully": 96625, + "Ġunsupported": 40409, + "Ġunsur": 68117, + "Ġunsure": 42903, + "Ġunsus": 91551, + "Ġunsustainable": 97509, + "Ġunt": 13065, + "Ġunten": 82816, + "Ġunter": 21506, + "Ġunters": 57686, + "Ġunterschied": 80260, + "Ġunterstüt": 81429, + "Ġunterstützen": 143562, + "Ġunterstützt": 143037, + "Ġunthinkable": 95503, + "Ġuntil": 3080, + "Ġunto": 29349, + "Ġuntouched": 67522, + "Ġuntranslated": 90348, + "Ġuntreated": 82820, + "Ġuntrue": 87877, + "Ġuntuk": 19593, + "Ġunus": 15857, + "Ġunused": 20006, + "Ġunusual": 18511, + "Ġunusually": 56799, + "Ġunve": 27747, + "Ġunveil": 91031, + "Ġunveiled": 38197, + "Ġunveiling": 91488, + "Ġunw": 15025, + "Ġunwanted": 34921, + "Ġunwilling": 45143, + "Ġunwind": 81510, + "Ġunwitting": 94019, + "Ġunwrap": 79813, + "Ġunzip": 75913, + "Ġuom": 50053, + "Ġuomini": 55413, + "Ġuomo": 35126, + "Ġup": 705, + "Ġupbeat": 84673, + "Ġupbringing": 91285, + "Ġupcoming": 14487, + "Ġupd": 79003, + "Ġupdate": 2647, + "ĠupdateTime": 75335, + "ĠupdateUser": 60391, + "Ġupdated": 6049, + "ĠupdatedAt": 48831, + "Ġupdater": 68249, + "Ġupdates": 8837, + "Ġupdating": 21010, + "Ġupfront": 64351, + "Ġupgrade": 13910, + "Ġupgraded": 32524, + "Ġupgrades": 31614, + "Ġupgrading": 45935, + "Ġuphe": 95610, + "Ġupheld": 61311, + "Ġuphill": 76658, + "Ġuphol": 69204, + "Ġuphold": 63423, + "Ġupholstery": 95688, + "Ġupkeep": 96096, + "Ġuplift": 96447, + "Ġuplifting": 94509, + "Ġupload": 8135, + "Ġuploaded": 22853, + "Ġuploader": 82257, + "Ġuploading": 47329, + "Ġuploads": 66563, + "Ġupon": 5193, + "Ġupp": 63433, + "Ġupper": 8416, + "Ġuppercase": 39482, + "Ġupright": 48585, + "Ġuprising": 69406, + "Ġupro": 96832, + "Ġups": 32734, + "Ġupscale": 73923, + "Ġupset": 22459, + "Ġupsetting": 89969, + "Ġupside": 35841, + "Ġupstairs": 48907, + "Ġupstream": 41730, + "Ġupt": 34335, + "Ġuptake": 68475, + "Ġuptime": 86689, + "Ġupto": 80126, + "Ġupward": 39772, + "Ġupwards": 43954, + "Ġur": 4335, + "Ġuranium": 53844, + "Ġuranus": 95228, + "Ġurb": 77710, + "Ġurban": 15662, + "Ġure": 52460, + "Ġurg": 60839, + "Ġurge": 32047, + "Ġurged": 27832, + "Ġurgency": 53817, + "Ġurgent": 33671, + "Ġurgently": 76620, + "Ġurges": 62837, + "Ġurging": 42576, + "Ġuri": 13071, + "Ġurinary": 71635, + "Ġurine": 39235, + "Ġurl": 2515, + "ĠurlString": 82489, + "Ġurlencode": 58691, + "Ġurllib": 24090, + "Ġurlparse": 80067, + "Ġurlpatterns": 63434, + "Ġurls": 29984, + "Ġurn": 65867, + "Ġurz": 143475, + "ĠurzÄħd": 140821, + "ĠurzÄħdzenia": 140822, + "ĠurzÄĻd": 143476, + "Ġus": 601, + "Ġusa": 33715, + "Ġusability": 75060, + "Ġusable": 39930, + "Ġusado": 97343, + "Ġusage": 10431, + "Ġusando": 57141, + "Ġusar": 46515, + "Ġusb": 26983, + "Ġuse": 990, + "ĠuseCallback": 43505, + "ĠuseClass": 97675, + "ĠuseContext": 34652, + "ĠuseDispatch": 39894, + "ĠuseEffect": 14233, + "ĠuseForm": 90926, + "ĠuseHistory": 44033, + "ĠuseMemo": 72479, + "ĠuseNewUrlParser": 92630, + "ĠuseParams": 84189, + "ĠuseRef": 40133, + "ĠuseRouter": 93915, + "ĠuseSelector": 32392, + "ĠuseState": 8102, + "ĠuseStyles": 34129, + "Ġused": 1483, + "Ġuseful": 5390, + "Ġusefulness": 75864, + "Ġuseless": 28848, + "Ġuser": 1196, + "ĠuserAgent": 94196, + "ĠuserDao": 77961, + "ĠuserData": 34385, + "ĠuserDetails": 88103, + "ĠuserEmail": 87570, + "ĠuserID": 35204, + "ĠuserId": 10329, + "ĠuserInfo": 24469, + "ĠuserInput": 64173, + "ĠuserList": 64374, + "ĠuserManager": 65662, + "ĠuserModel": 82785, + "ĠuserName": 19855, + "ĠuserProfile": 86656, + "ĠuserRepository": 49508, + "ĠuserService": 30662, + "ĠuserType": 87547, + "Ġuserdata": 66874, + "Ġuserid": 32270, + "Ġuserinfo": 91178, + "Ġusern": 75012, + "Ġusername": 5934, + "Ġusernames": 82051, + "Ġusers": 3847, + "Ġuses": 5711, + "Ġusher": 63160, + "Ġushort": 43024, + "Ġusing": 1667, + "Ġusize": 22301, + "Ġusleep": 82601, + "Ġuso": 33910, + "Ġusr": 43071, + "Ġust": 87589, + "Ġusted": 60295, + "Ġusu": 40288, + "Ġusual": 13475, + "Ġusually": 5990, + "Ġusuario": 16382, + "Ġusuarios": 41756, + "Ġusur": 95088, + "Ġusuário": 59018, + "ĠusÅĤug": 130680, + "Ġut": 8621, + "Ġutan": 81764, + "Ġutc": 69596, + "Ġuten": 85473, + "Ġutens": 80936, + "Ġuterus": 84556, + "Ġutf": 10644, + "Ġutil": 4094, + "Ġutilis": 41487, + "Ġutilisateur": 86526, + "Ġutilise": 68052, + "Ġutiliser": 74044, + "Ġutilisé": 134449, + "Ġutilities": 29419, + "Ġutility": 15549, + "Ġutiliz": 32172, + "Ġutiliza": 76061, + "Ġutilizado": 98533, + "Ġutilizando": 75990, + "Ġutilizar": 63980, + "Ġutilization": 49449, + "Ġutilize": 28066, + "Ġutilized": 33616, + "Ġutilizes": 59780, + "Ġutilizing": 34888, + "Ġutils": 12439, + "Ġutmost": 53611, + "Ġutrecht": 95073, + "Ġutter": 21532, + "Ġuttered": 89739, + "Ġutterly": 37531, + "Ġuu": 41179, + "Ġuuid": 16040, + "Ġuur": 88760, + "Ġuv": 29663, + "Ġuw": 37343, + "Ġuwag": 134907, + "ĠuwagÄĻ": 134908, + "Ġuważ": 137800, + "Ġuważa": 137801, + "Ġux": 89803, + "Ġuy": 81900, + "Ġuyar": 140741, + "Ġuyarı": 140742, + "Ġuz": 44476, + "Ġuç": 141477, + "Ġuçak": 141478, + "Ġuçu": 143783, + "ĠuçuÅŁ": 143784, + "ĠuÄŁ": 136119, + "ĠuÄŁra": 136120, + "Ġuży": 61408, + "Ġużytk": 88324, + "Ġużytkow": 131897, + "Ġużywa": 133290, + "Ġuá»ijng": 130069, + "Ġv": 348, + "ĠvX": 92845, + "Ġva": 11164, + "Ġvac": 9283, + "Ġvacancies": 66651, + "Ġvacancy": 61462, + "Ġvacant": 48691, + "Ġvacation": 20161, + "Ġvacations": 74567, + "Ġvacc": 13712, + "Ġvaccinated": 69458, + "Ġvaccination": 46065, + "Ġvaccinations": 90871, + "Ġvaccine": 24467, + "Ġvaccines": 39200, + "Ġvacuum": 28202, + "Ġvad": 70852, + "Ġvag": 17314, + "Ġvagina": 38788, + "Ġvaginal": 57059, + "Ġvague": 39046, + "Ġvaguely": 71959, + "Ġvagy": 61532, + "Ġvai": 39486, + "Ġvain": 45504, + "Ġvais": 90407, + "Ġvak": 74825, + "Ġval": 1044, + "Ġvale": 33722, + "Ġvaleur": 50204, + "Ġvaleurs": 95935, + "Ġvalid": 2697, + "Ġvalida": 91687, + "Ġvalidar": 83115, + "Ġvalidate": 9593, + "Ġvalidated": 32332, + "Ġvalidates": 26257, + "Ġvalidating": 68672, + "Ġvalidation": 10519, + "ĠvalidationResult": 96533, + "Ġvalidations": 83388, + "Ġvalidator": 22935, + "Ġvalidators": 38588, + "Ġvalide": 98756, + "Ġvalidity": 31839, + "Ġvalido": 89315, + "Ġvalign": 23858, + "Ġvalley": 33581, + "Ġvalleys": 85397, + "Ġvalor": 14116, + "Ġvalore": 93942, + "Ġvalores": 36152, + "Ġvals": 28356, + "Ġvalu": 92931, + "Ġvaluable": 15172, + "Ġvaluation": 50162, + "Ġvalue": 897, + "ĠvalueForKey": 48588, + "ĠvalueType": 96438, + "Ġvalued": 32547, + "Ġvalues": 2750, + "Ġvalve": 31430, + "Ġvalves": 63000, + "Ġvamos": 69670, + "Ġvamp": 59055, + "Ġvampire": 50487, + "Ġvampires": 77190, + "Ġvan": 5242, + "Ġvandal": 61407, + "Ġvandalism": 88852, + "Ġvanilla": 32065, + "Ġvanish": 91127, + "Ġvanished": 58481, + "Ġvanity": 51571, + "Ġvans": 76024, + "Ġvant": 96671, + "Ġvap": 67757, + "Ġvape": 68170, + "Ġvaping": 80288, + "Ġvapor": 37652, + "Ġvar": 762, + "Ġvara": 62801, + "Ġvarargin": 52439, + "Ġvarchar": 32958, + "Ġvard": 127127, + "Ġvardı": 136959, + "Ġvardır": 131477, + "Ġvari": 4151, + "Ġvariability": 53609, + "Ġvariable": 3890, + "Ġvariables": 7332, + "Ġvariance": 32273, + "Ġvariant": 11424, + "Ġvariants": 26012, + "Ġvarias": 84950, + "Ġvariation": 22990, + "Ġvariations": 26244, + "Ġvaried": 27730, + "Ġvaries": 34227, + "Ġvarieties": 35580, + "Ġvariety": 8045, + "Ġvarios": 73818, + "Ġvarious": 5257, + "Ġvars": 19942, + "Ġvary": 13289, + "Ġvarying": 28765, + "Ġvas": 43396, + "Ġvascular": 63503, + "Ġvase": 92384, + "Ġvast": 12767, + "Ġvastly": 52008, + "Ġvat": 78526, + "Ġvatanda": 133277, + "ĠvatandaÅŁ": 133278, + "Ġvault": 34584, + "Ġvaz": 143854, + "Ġvazge": 143855, + "Ġvazgeç": 143856, + "ĠvaÌĢ": 139607, + "Ġvb": 34719, + "ĠvbCrLf": 89407, + "Ġvbox": 85977, + "Ġvc": 24553, + "Ġvd": 61527, + "Ġve": 5208, + "Ġvec": 7486, + "Ġveces": 56329, + "Ġvect": 55635, + "Ġvector": 4621, + "Ġvectors": 22879, + "Ġved": 27933, + "Ġvedere": 87745, + "Ġveel": 50137, + "Ġveg": 47338, + "Ġvegan": 29149, + "Ġvegas": 72439, + "Ġveget": 12997, + "Ġvegetable": 35481, + "Ġvegetables": 23880, + "Ġvegetarian": 45382, + "Ġvegetation": 53732, + "Ġveggies": 59299, + "Ġveh": 5164, + "Ġvehement": 91368, + "Ġvehicle": 7310, + "Ġvehicles": 11474, + "ĠvehÃŃculo": 139759, + "Ġveil": 58255, + "Ġvein": 53244, + "Ġveins": 59664, + "Ġvej": 92460, + "Ġvel": 9054, + "Ġvelit": 71548, + "Ġveloc": 40509, + "Ġvelocidad": 98094, + "Ġvelocities": 74057, + "Ġvelocity": 15435, + "Ġvelvet": 71326, + "Ġvem": 97062, + "Ġven": 11208, + "Ġvend": 18672, + "Ġvenda": 93545, + "Ġvending": 85044, + "Ġvendor": 20728, + "Ġvendors": 28529, + "Ġvene": 76751, + "Ġvener": 90131, + "Ġvenez": 73990, + "Ġvengeance": 71688, + "Ġveniam": 87970, + "Ġvenir": 95486, + "Ġvenom": 73708, + "Ġvent": 10378, + "Ġventa": 55402, + "Ġventana": 63146, + "Ġventas": 99019, + "Ġvente": 73090, + "Ġventil": 70602, + "Ġventilation": 55134, + "Ġvents": 80207, + "Ġventure": 25191, + "Ġventured": 98407, + "Ġventures": 65624, + "Ġvenue": 21434, + "Ġvenues": 36178, + "Ġver": 2739, + "Ġvera": 91244, + "Ġverb": 18607, + "Ġverbal": 35770, + "Ġverbally": 84107, + "Ġverbess": 98038, + "Ġverbose": 13694, + "Ġverbosity": 71387, + "Ġverbs": 61846, + "Ġverd": 25658, + "Ġverdad": 81171, + "Ġverdade": 81585, + "Ġverde": 73161, + "Ġverdict": 35443, + "Ġverdienen": 91440, + "ĠverdiÄŁi": 136704, + "Ġvere": 72943, + "Ġverfüg": 141436, + "Ġverfügbar": 141437, + "Ġverg": 37801, + "Ġverge": 58718, + "Ġverifica": 91071, + "Ġverificar": 91819, + "Ġverification": 22901, + "Ġverified": 23933, + "Ġverifier": 88737, + "Ġverifies": 87856, + "Ġverify": 10146, + "Ġverifying": 68863, + "Ġverk": 68336, + "Ġverm": 95898, + "Ġvern": 71593, + "Ġvero": 78287, + "Ġverr": 71467, + "Ġvers": 5436, + "Ġversa": 45291, + "Ġversatile": 31945, + "Ġversatility": 59012, + "Ġversch": 47708, + "Ġverschied": 51892, + "Ġverschiedene": 95710, + "Ġverschiedenen": 88687, + "Ġverschill": 91864, + "Ġverse": 32387, + "Ġverses": 49299, + "Ġversion": 2319, + "Ġversions": 10795, + "Ġversión": 96491, + "Ġverso": 91865, + "Ġverst": 85104, + "Ġversus": 19041, + "Ġvert": 5198, + "Ġverte": 66761, + "Ġvertex": 11936, + "Ġvertical": 12140, + "Ġvertically": 50935, + "Ġvertices": 17228, + "Ġverts": 59414, + "Ġverv": 96441, + "Ġverw": 31100, + "Ġverwenden": 79551, + "Ġverwendet": 74191, + "Ġvery": 1602, + "Ġverz": 73197, + "Ġveröffent": 141870, + "Ġveröffentlicht": 141871, + "Ġves": 64535, + "Ġvess": 17441, + "Ġvessel": 26182, + "Ġvessels": 29980, + "Ġvest": 27605, + "Ġvested": 77951, + "Ġvestib": 90934, + "Ġvests": 97273, + "Ġvet": 23313, + "Ġveter": 12569, + "Ġveteran": 20820, + "Ġveterans": 25820, + "Ġveterin": 60571, + "Ġveterinarian": 81240, + "Ġveterinary": 68240, + "Ġveto": 47860, + "Ġvetor": 75457, + "Ġvets": 97018, + "Ġveut": 75505, + "Ġvex": 83165, + "Ġveya": 86959, + "Ġvez": 20563, + "Ġvezes": 76693, + "ĠveÃŃculo": 138736, + "Ġvf": 58454, + "Ġvfs": 92941, + "Ġvg": 84530, + "Ġvh": 67888, + "Ġvi": 3275, + "Ġvia": 4566, + "Ġviability": 67407, + "Ġviable": 30428, + "Ġviagra": 19695, + "Ġvib": 55300, + "Ġvibe": 46711, + "Ġvibes": 89849, + "Ġvibr": 16929, + "Ġvibrant": 32976, + "Ġvibrating": 49937, + "Ġvibration": 47813, + "Ġvibrations": 77252, + "Ġvibrator": 60724, + "Ġvic": 31431, + "Ġvice": 16774, + "Ġvicinity": 52751, + "Ġvicious": 42410, + "Ġvict": 5199, + "Ġvictim": 11734, + "Ġvictims": 12415, + "Ġvictories": 45046, + "Ġvictorious": 73710, + "Ġvictory": 12560, + "Ġvid": 18127, + "Ġvida": 24949, + "Ġvide": 23029, + "Ġvideo": 2766, + "Ġvideoer": 69057, + "Ġvideog": 88017, + "Ġvideos": 6803, + "Ġvideot": 71274, + "Ġvids": 70512, + "Ġvidé": 39638, + "Ġvidéo": 56046, + "Ġvidéos": 71786, + "Ġvie": 17098, + "Ġvieille": 69211, + "Ġviel": 36077, + "Ġviele": 41983, + "Ġvielen": 68042, + "Ġvielleicht": 83835, + "Ġviene": 55404, + "Ġvient": 85000, + "Ġvier": 57093, + "Ġviet": 84359, + "Ġvieux": 91522, + "Ġview": 1651, + "ĠviewBox": 37996, + "ĠviewController": 52722, + "ĠviewDidLoad": 21625, + "ĠviewHolder": 43473, + "ĠviewModel": 20710, + "ĠviewPager": 85972, + "ĠviewType": 53081, + "ĠviewWillAppear": 65665, + "Ġviewed": 19334, + "Ġviewer": 25708, + "Ġviewers": 21762, + "Ġviewing": 20102, + "Ġviewpoint": 58385, + "Ġviewpoints": 89809, + "Ġviewport": 32253, + "Ġviews": 6194, + "Ġvig": 20802, + "Ġvigil": 37394, + "Ġvigilant": 80234, + "Ġvign": 92792, + "Ġvigor": 54922, + "Ġvigorous": 70820, + "Ġvigorously": 76749, + "Ġvil": 25801, + "Ġvile": 80214, + "Ġvill": 8901, + "Ġvilla": 46525, + "Ġvillage": 14126, + "Ġvillagers": 60821, + "Ġvillages": 32789, + "Ġvillain": 39048, + "Ġvillains": 60710, + "Ġville": 38873, + "Ġvim": 36157, + "Ġvin": 23902, + "Ġvinc": 81710, + "Ġvind": 37805, + "Ġvinden": 92651, + "Ġvine": 28950, + "Ġvinegar": 46105, + "Ġvines": 85429, + "Ġvintage": 23790, + "Ġvinyl": 33449, + "Ġviol": 4634, + "Ġviolate": 40487, + "Ġviolated": 33421, + "Ġviolates": 50313, + "Ġviolating": 44362, + "Ġviolation": 19940, + "Ġviolations": 26557, + "Ġviolence": 9170, + "Ġviolent": 16401, + "Ġviolently": 64200, + "Ġviolet": 79736, + "Ġviolin": 62037, + "Ġvip": 62868, + "Ġviper": 95132, + "Ġvir": 8867, + "Ġviral": 28862, + "Ġvirgin": 35899, + "Ġvirt": 12763, + "Ġvirtual": 4108, + "Ġvirtually": 21214, + "Ġvirtue": 34360, + "Ġvirtues": 65527, + "Ġvirus": 16770, + "Ġviruses": 40968, + "Ġvis": 2098, + "Ġvisa": 26655, + "Ġvisas": 55677, + "Ġvisc": 58565, + "Ġvisceral": 88016, + "Ġviscosity": 98430, + "Ġvisibility": 23160, + "Ġvisible": 9434, + "Ġvisibly": 73006, + "Ġvision": 11129, + "Ġvisionary": 86390, + "Ġvisions": 57295, + "Ġvisit": 3947, + "Ġvisita": 80833, + "Ġvisite": 74736, + "Ġvisited": 11994, + "Ġvisiting": 16721, + "Ġvisitor": 20181, + "Ġvisitors": 15255, + "Ġvisits": 21051, + "Ġvista": 39036, + "Ġvistas": 95141, + "Ġvisto": 65418, + "Ġvisual": 9124, + "Ġvisualization": 41048, + "Ġvisualize": 50087, + "Ġvisually": 42295, + "Ġvisuals": 52977, + "Ġvit": 13157, + "Ġvita": 54476, + "Ġvitae": 63120, + "Ġvital": 16198, + "Ġvitality": 74616, + "Ġvitamin": 27071, + "Ġvitamins": 45092, + "Ġvite": 82357, + "Ġvitro": 53904, + "Ġviv": 17950, + "Ġvivastreet": 64792, + "Ġvive": 91902, + "Ġvivid": 42020, + "Ġvivo": 40194, + "Ġviz": 47750, + "Ġviêm": 132785, + "Ġviên": 97483, + "Ġviết": 128839, + "Ġviá»ĩc": 128289, + "Ġviá»ĩn": 129972, + "Ġvk": 34283, + "Ġvl": 41136, + "Ġvlan": 65420, + "Ġvm": 10995, + "Ġvmax": 77015, + "Ġvmin": 88263, + "Ġvn": 51017, + "Ġvnode": 68721, + "Ġvo": 4069, + "Ġvoc": 11984, + "Ġvocab": 23820, + "Ġvocabulary": 34918, + "Ġvocal": 25407, + "Ġvocalist": 95565, + "Ġvocals": 46096, + "Ġvocational": 88386, + "Ġvocê": 24709, + "Ġvodka": 75424, + "Ġvog": 62939, + "Ġvoi": 67652, + "Ġvoice": 7743, + "Ġvoiced": 51368, + "Ġvoices": 23314, + "Ġvoid": 737, + "Ġvoir": 45031, + "Ġvois": 99064, + "Ġvoiture": 80545, + "Ġvoks": 92413, + "Ġvoksen": 70920, + "Ġvoksne": 98030, + "Ġvol": 4400, + "Ġvolatile": 17072, + "Ġvolatility": 52738, + "Ġvolcan": 35273, + "Ġvolcanic": 66064, + "Ġvolcano": 65206, + "Ġvoll": 56354, + "Ġvolley": 55573, + "Ġvolleyball": 76067, + "Ġvolont": 97633, + "Ġvolt": 16393, + "Ġvolta": 53986, + "Ġvoltage": 21720, + "Ġvolte": 96459, + "Ġvolts": 86067, + "Ġvolum": 62820, + "Ġvolume": 8123, + "Ġvolumes": 26282, + "Ġvolunt": 24858, + "Ġvoluntarily": 53423, + "Ġvoluntary": 35979, + "Ġvolunte": 13376, + "Ġvolunteer": 25144, + "Ġvolunteered": 73671, + "Ġvolunteering": 69338, + "Ġvolunteers": 23010, + "Ġvolupt": 50710, + "Ġvolver": 86362, + "Ġvom": 21982, + "Ġvomiting": 62471, + "Ġvon": 6538, + "Ġvont": 81033, + "Ġvoor": 12931, + "Ġvor": 13906, + "Ġvorhand": 92670, + "Ġvortex": 86527, + "Ġvos": 25251, + "Ġvot": 4170, + "Ġvote": 6910, + "Ġvoted": 16228, + "Ġvoter": 25323, + "Ġvoters": 11707, + "Ġvotes": 12684, + "Ġvoting": 15668, + "Ġvotre": 14918, + "Ġvou": 54062, + "Ġvoucher": 48658, + "Ġvouchers": 71980, + "Ġvous": 9012, + "Ġvow": 39343, + "Ġvowed": 50030, + "Ġvowel": 76181, + "Ġvowels": 78456, + "Ġvows": 76441, + "Ġvox": 43739, + "Ġvoxel": 60184, + "Ġvoy": 23063, + "Ġvoyage": 44540, + "Ġvoyeur": 47472, + "Ġvoz": 84090, + "Ġvp": 34823, + "Ġvpn": 70806, + "Ġvr": 36100, + "Ġvra": 33823, + "Ġvrai": 82095, + "Ġvraiment": 58919, + "Ġvriend": 79782, + "Ġvrij": 69605, + "Ġvrou": 22451, + "Ġvrouw": 36379, + "Ġvrouwen": 35196, + "Ġvs": 6165, + "Ġvscode": 55380, + "Ġvt": 39105, + "Ġvtk": 29534, + "Ġvtx": 96879, + "Ġvu": 32514, + "Ġvue": 47134, + "Ġvuel": 63613, + "Ġvul": 11719, + "Ġvulgar": 73160, + "Ġvulner": 13521, + "Ġvulnerabilities": 51127, + "Ġvulnerability": 33004, + "Ġvulnerable": 19563, + "Ġvv": 56333, + "Ġvw": 74086, + "Ġvx": 43460, + "Ġvy": 45259, + "Ġvz": 87744, + "Ġvá": 57297, + "Ġvál": 85980, + "Ġválido": 97074, + "Ġvárias": 137417, + "Ġvários": 135626, + "Ġvão": 136622, + "Ġvä": 50618, + "Ġväl": 82924, + "ĠvÃ¥": 58309, + "ĠvÃ¥r": 96497, + "Ġvæ": 31370, + "Ġvære": 46073, + "Ġvé": 44065, + "Ġvéhic": 140422, + "Ġvéhicule": 140423, + "Ġvér": 76580, + "Ġvéritable": 139819, + "Ġvì": 128378, + "Ġvòng": 130604, + "Ġvô": 128867, + "Ġvõ": 79195, + "Ġvö": 143435, + "Ġvöll": 143436, + "Ġvöllig": 143437, + "Ġvùng": 129838, + "Ġvüc": 143837, + "Ġvücud": 143838, + "ĠvÃł": 47742, + "ĠvÃłi": 131302, + "ĠvÃłng": 130180, + "ĠvÃło": 86139, + "ĠvÃŃ": 30544, + "ĠvÃŃct": 142853, + "ĠvÃŃctima": 142854, + "ĠvÃŃde": 37050, + "ĠvÃŃdeo": 51230, + "ĠvÃŃdeos": 67428, + "ĠvÃŃt": 143874, + "ĠvÃŃtima": 143875, + "ĠvÄĥn": 128683, + "ĠvÅ©": 135343, + "Ġvượt": 132629, + "Ġvưá»Ŀ": 138790, + "Ġvưá»Ŀn": 138791, + "Ġvấn": 128580, + "Ġvẫn": 128487, + "Ġvẻ": 130525, + "Ġvẽ": 142077, + "Ġvết": 137866, + "ĠváºŃn": 129984, + "ĠváºŃt": 128853, + "ĠváºŃy": 128490, + "Ġvợ": 130888, + "Ġvụ": 128367, + "Ġvừa": 128727, + "Ġvững": 135584, + "Ġvá»±c": 128978, + "Ġvá»ģ": 128265, + "Ġvá»ĩ": 129067, + "Ġvá»ĭ": 128444, + "Ġvá»įng": 130949, + "Ġvá»ı": 141419, + "Ġvá»ijn": 130498, + "ĠvỼi": 128250, + "Ġvá»Ŀi": 135421, + "Ġw": 289, + "ĠwParam": 56624, + "Ġwa": 10450, + "Ġwaar": 37295, + "Ġwag": 64325, + "Ġwage": 19062, + "Ġwaged": 91400, + "Ġwager": 55343, + "Ġwages": 23123, + "Ġwagon": 60419, + "Ġwah": 85227, + "Ġwaist": 28042, + "Ġwait": 3783, + "ĠwaitFor": 52223, + "Ġwaited": 29215, + "Ġwaiter": 67169, + "Ġwaiting": 8580, + "Ġwaitress": 95207, + "Ġwaits": 46936, + "Ġwaive": 90283, + "Ġwaived": 81482, + "Ġwaiver": 53105, + "Ġwaivers": 90900, + "Ġwake": 15156, + "Ġwakes": 61567, + "Ġwakeup": 91331, + "Ġwaking": 47628, + "Ġwaktu": 85882, + "Ġwal": 40826, + "Ġwalk": 4227, + "Ġwalked": 14858, + "Ġwalker": 62524, + "Ġwalkers": 94637, + "Ġwalking": 11435, + "Ġwalks": 22479, + "Ġwalkthrough": 98944, + "Ġwall": 7002, + "Ġwallet": 15085, + "Ġwallets": 53308, + "Ġwallpaper": 43586, + "Ġwallpapers": 81895, + "Ġwalls": 14285, + "Ġwalmart": 70125, + "Ġwalnut": 92629, + "Ġwan": 77753, + "Ġwand": 28710, + "Ġwander": 39220, + "Ġwandered": 81194, + "Ġwandering": 53963, + "Ġwang": 96944, + "Ġwann": 54505, + "Ġwanna": 32733, + "Ġwannonce": 83576, + "Ġwant": 1366, + "Ġwanted": 4829, + "Ġwanting": 19211, + "Ġwants": 6801, + "Ġwar": 4116, + "Ġward": 25657, + "Ġwardrobe": 45746, + "Ġwards": 77151, + "Ġware": 50706, + "Ġwarehouse": 30112, + "Ġwarehouses": 82173, + "Ġwaren": 43303, + "Ġwarfare": 38207, + "Ġwarm": 8205, + "Ġwarmed": 81730, + "Ġwarmer": 44939, + "Ġwarming": 23867, + "Ġwarmly": 96370, + "Ġwarmth": 40363, + "Ġwarn": 8809, + "Ġwarned": 18683, + "Ġwarning": 9958, + "Ġwarnings": 18993, + "Ġwarns": 48040, + "Ġwarp": 46320, + "Ġwarped": 86212, + "Ġwarrant": 7091, + "Ġwarranted": 73180, + "Ġwarranties": 60813, + "Ġwarrants": 54588, + "Ġwarranty": 8475, + "Ġwarrior": 37390, + "Ġwarriors": 42500, + "Ġwars": 24936, + "Ġwart": 59945, + "Ġwartime": 86852, + "Ġwarto": 78834, + "ĠwartoÅĽci": 135646, + "Ġwary": 54174, + "Ġwas": 572, + "Ġwash": 11369, + "Ġwashed": 37493, + "Ġwasher": 52892, + "Ġwashing": 27686, + "Ġwashington": 93671, + "Ġwasm": 98263, + "Ġwasn": 5710, + "Ġwast": 40321, + "Ġwaste": 12291, + "Ġwasted": 38828, + "Ġwastes": 81220, + "Ġwastewater": 76581, + "Ġwasting": 47797, + "Ġwat": 29025, + "Ġwatch": 3736, + "Ġwatchdog": 46911, + "Ġwatched": 15384, + "Ġwatcher": 55727, + "Ġwatchers": 86572, + "Ġwatches": 31760, + "Ġwatching": 10099, + "Ġwater": 3015, + "Ġwaterfall": 69051, + "Ġwaterfront": 68927, + "Ġwatering": 72741, + "Ġwatermark": 88006, + "Ġwaterproof": 46419, + "Ġwaters": 20520, + "Ġwatershed": 91356, + "Ġwatt": 66172, + "Ġwatts": 71022, + "Ġwav": 53807, + "Ġwave": 12060, + "Ġwaved": 64582, + "Ġwaveform": 72364, + "Ġwavelength": 45306, + "Ġwavelengths": 92859, + "Ġwaves": 16876, + "Ġwaving": 63111, + "Ġwax": 36023, + "Ġway": 1616, + "Ġwaypoint": 65176, + "Ġwaypoints": 81286, + "Ġways": 5510, + "Ġważ": 139215, + "Ġważne": 139216, + "Ġwb": 37858, + "Ġwc": 26548, + "Ġwchar": 41049, + "Ġwcs": 96668, + "ĠwczeÅĽ": 134098, + "ĠwczeÅĽniej": 134099, + "Ġwd": 45404, + "Ġwe": 582, + "Ġweak": 7469, + "ĠweakSelf": 73759, + "Ġweaken": 32456, + "Ġweakened": 57664, + "Ġweakening": 82426, + "Ġweaker": 42283, + "Ġweakest": 80763, + "Ġweakness": 23078, + "Ġweaknesses": 43567, + "Ġwealth": 11939, + "Ġwealthiest": 92018, + "Ġwealthy": 27894, + "Ġweap": 6158, + "Ġweapon": 10288, + "Ġweaponry": 77022, + "Ġweapons": 9893, + "Ġwear": 9850, + "Ġwearable": 77248, + "Ġwearer": 84602, + "Ġwearing": 12233, + "Ġwears": 37300, + "Ġweary": 75556, + "Ġweather": 9104, + "Ġweave": 67508, + "Ġweaving": 83609, + "Ġweb": 3482, + "ĠwebView": 56358, + "Ġwebcam": 25951, + "Ġwebdriver": 45681, + "Ġwebhook": 75268, + "Ġwebinar": 62560, + "Ġweblog": 93298, + "Ġwebpack": 31179, + "Ġwebpage": 44610, + "Ġwebs": 80920, + "Ġwebsite": 3910, + "Ġwebsites": 13037, + "Ġwebsocket": 58943, + "Ġwed": 10840, + "Ġwedding": 13008, + "Ġweddings": 54244, + "Ġwedge": 63685, + "ĠwedÅĤug": 140919, + "Ġwee": 74065, + "Ġweed": 39375, + "Ġweeds": 69435, + "Ġweek": 2003, + "Ġweekday": 46578, + "Ġweekdays": 71995, + "Ġweekend": 9001, + "Ġweekends": 37002, + "Ġweekly": 17059, + "Ġweeks": 5555, + "Ġweer": 65150, + "Ġweg": 70511, + "Ġwegen": 83281, + "Ġwei": 83904, + "Ġweiber": 70349, + "Ġweigh": 17529, + "Ġweighed": 46612, + "Ġweighing": 46726, + "Ġweighs": 49442, + "Ġweight": 4680, + "Ġweighted": 36824, + "Ġweighting": 88939, + "Ġweights": 14324, + "Ġweil": 52507, + "Ġweir": 96986, + "Ġweird": 16283, + "Ġweit": 85117, + "Ġweiter": 30632, + "Ġweitere": 61550, + "ĠweiÃŁ": 67278, + "Ġwel": 12442, + "Ġwelche": 58319, + "Ġwelcome": 10565, + "Ġwelcomed": 30969, + "Ġwelcomes": 56258, + "Ġwelcoming": 35287, + "Ġweld": 32766, + "Ġwelded": 81061, + "Ġwelding": 58922, + "Ġwelfare": 22634, + "Ġwell": 1632, + "Ġwellbeing": 56830, + "Ġwellness": 38790, + "Ġwells": 47296, + "Ġwelt": 77748, + "Ġwen": 37357, + "Ġwenig": 69987, + "Ġweniger": 83801, + "Ġwenn": 22075, + "Ġwent": 3937, + "Ġwer": 7060, + "Ġwerd": 72597, + "Ġwerde": 86913, + "Ġwerden": 12643, + "Ġwere": 1033, + "Ġwereld": 87885, + "Ġweren": 14716, + "Ġwerk": 58618, + "Ġwes": 97168, + "Ġwest": 9710, + "Ġwestern": 18494, + "Ġwet": 14401, + "ĠwewnÄĻtrzn": 141456, + "Ġwf": 41248, + "Ġwg": 63581, + "Ġwget": 64508, + "Ġwh": 420, + "Ġwhale": 50019, + "Ġwhales": 56774, + "Ġwhat": 1128, + "Ġwhatever": 8820, + "Ġwhats": 40109, + "Ġwhatsapp": 60208, + "Ġwhatsoever": 35634, + "Ġwhe": 14894, + "Ġwheat": 33053, + "Ġwheel": 13284, + "Ġwheelchair": 53518, + "Ġwheels": 22696, + "Ġwhen": 979, + "Ġwhence": 87449, + "Ġwhenever": 15356, + "Ġwhere": 1380, + "Ġwhereabouts": 79502, + "Ġwhereas": 19853, + "Ġwhereby": 47901, + "Ġwherein": 41318, + "Ġwherever": 27478, + "Ġwhether": 3425, + "Ġwhich": 892, + "Ġwhichever": 53684, + "Ġwhile": 1393, + "Ġwhilst": 23856, + "Ġwhim": 73471, + "Ġwhims": 69105, + "Ġwhip": 50076, + "Ġwhipped": 63257, + "Ġwhipping": 97522, + "Ġwhirl": 63683, + "Ġwhisk": 40659, + "Ġwhiskey": 64254, + "Ġwhisky": 80483, + "Ġwhisper": 34855, + "Ġwhispered": 57266, + "Ġwhispers": 88148, + "Ġwhistle": 39549, + "Ġwhistlebl": 64666, + "Ġwhistleblower": 90968, + "Ġwhit": 33699, + "Ġwhite": 4158, + "ĠwhiteColor": 45281, + "Ġwhitelist": 67727, + "Ġwhites": 21874, + "Ġwhitespace": 36372, + "Ġwho": 879, + "Ġwhoever": 38271, + "Ġwhole": 4361, + "Ġwholes": 25211, + "Ġwholesale": 34457, + "Ġwholesalers": 89897, + "Ġwholesome": 87218, + "Ġwholly": 41141, + "Ġwhom": 8711, + "Ġwhopping": 62869, + "Ġwhore": 71925, + "Ġwhose": 6693, + "Ġwhy": 3170, + "Ġwi": 23250, + "Ġwich": 70367, + "Ġwicht": 90796, + "Ġwichtig": 65733, + "Ġwicked": 43977, + "Ġwid": 9724, + "Ġwida": 142268, + "ĠwidaÄĩ": 142269, + "Ġwide": 6884, + "Ġwidely": 13570, + "Ġwiden": 93315, + "Ġwidened": 85888, + "Ġwidening": 84400, + "Ġwider": 21864, + "Ġwides": 23003, + "Ġwidespread": 23782, + "Ġwidest": 80239, + "Ġwidget": 9086, + "Ġwidgets": 31855, + "Ġwidow": 56796, + "Ġwidth": 2374, + "Ġwidths": 64411, + "Ġwie": 13368, + "Ġwieder": 26252, + "Ġwiel": 63979, + "Ġwield": 41845, + "Ġwielding": 86823, + "Ġwife": 7403, + "Ġwifi": 33417, + "Ġwig": 58741, + "Ġwij": 49551, + "Ġwik": 97175, + "Ġwiki": 28609, + "Ġwikipedia": 58218, + "Ġwil": 30231, + "Ġwild": 8380, + "Ġwildcard": 59104, + "Ġwilderness": 48262, + "Ġwildfire": 92125, + "Ġwildfires": 98382, + "Ġwildlife": 29305, + "Ġwildly": 47195, + "Ġwill": 686, + "Ġwillen": 84337, + "Ġwilling": 9831, + "Ġwillingly": 72552, + "Ġwillingness": 38275, + "Ġwilt": 74496, + "Ġwin": 3164, + "Ġwind": 9956, + "Ġwinding": 53726, + "Ġwindow": 3241, + "ĠwindowHeight": 88547, + "Ġwindows": 11030, + "Ġwinds": 28492, + "Ġwindshield": 73939, + "Ġwindy": 92921, + "Ġwine": 13078, + "Ġwines": 42755, + "Ġwing": 20013, + "Ġwinger": 73439, + "Ġwings": 26204, + "Ġwink": 68339, + "Ġwinner": 13632, + "Ġwinners": 25448, + "Ġwinning": 10987, + "Ġwinnings": 86349, + "Ġwins": 14816, + "Ġwinter": 12406, + "Ġwinters": 84982, + "Ġwipe": 39094, + "Ġwiped": 48166, + "Ġwipes": 83538, + "Ġwiping": 83139, + "Ġwir": 16111, + "Ġwird": 14821, + "Ġwire": 9067, + "ĠwireType": 89243, + "Ġwired": 52121, + "Ġwireless": 20739, + "Ġwires": 35008, + "Ġwiring": 18827, + "Ġwirk": 50368, + "Ġwirklich": 55813, + "Ġwis": 18599, + "Ġwisdom": 23389, + "Ġwise": 23335, + "Ġwisely": 65613, + "Ġwish": 6426, + "Ġwished": 36187, + "Ġwishes": 24240, + "Ġwishing": 45127, + "Ġwishlist": 71662, + "Ġwissen": 75130, + "Ġwit": 37367, + "Ġwitch": 36382, + "Ġwitches": 80162, + "Ġwith": 448, + "ĠwithObject": 75298, + "ĠwithRouter": 54532, + "ĠwithString": 94974, + "ĠwithStyles": 63310, + "Ġwithd": 44974, + "Ġwithdraw": 14798, + "Ġwithdrawal": 29736, + "Ġwithdrawals": 85204, + "Ġwithdrawing": 82960, + "Ġwithdrawn": 49582, + "Ġwithdrew": 61665, + "Ġwithheld": 76003, + "Ġwithhold": 86800, + "Ġwithholding": 80512, + "Ġwithin": 2878, + "Ġwithout": 2041, + "Ġwithstand": 50471, + "Ġwitness": 11298, + "Ġwitnessed": 31026, + "Ġwitnesses": 27723, + "Ġwitnessing": 68942, + "Ġwitty": 82633, + "Ġwives": 38620, + "Ġwizard": 33968, + "Ġwizards": 88163, + "ĠwiÄĻ": 44916, + "ĠwiÄĻc": 98701, + "ĠwiÄĻcej": 80444, + "ĠwiÄĻks": 128169, + "ĠwiÄĻksze": 141841, + "ĠwiÄĻkszo": 143022, + "ĠwiÄĻkszoÅĽÄĩ": 143023, + "Ġwk": 73760, + "Ġwl": 44709, + "Ġwlan": 78353, + "Ġwm": 51634, + "Ġwn": 65480, + "Ġwnd": 80432, + "ĠwnÄĻtr": 138976, + "ĠwnÄĻtrz": 138977, + "Ġwo": 23738, + "Ġwoes": 76725, + "Ġwohl": 72170, + "Ġwoke": 38726, + "Ġwol": 79614, + "Ġwolf": 36542, + "Ġwoll": 73184, + "Ġwollen": 56945, + "Ġwollte": 96434, + "Ġwolves": 55171, + "Ġwom": 2358, + "Ġwoman": 5220, + "Ġwomb": 72837, + "Ġwomen": 3198, + "Ġwomens": 87361, + "Ġwon": 2765, + "Ġwonder": 5775, + "Ġwondered": 30056, + "Ġwonderful": 11117, + "Ġwonderfully": 59985, + "Ġwondering": 20293, + "Ġwonders": 39064, + "Ġwont": 39364, + "Ġwoo": 68632, + "Ġwood": 7579, + "Ġwooded": 93371, + "Ġwooden": 22360, + "Ġwoodland": 97631, + "Ġwoods": 32533, + "Ġwoodworking": 95953, + "Ġwool": 38540, + "Ġwor": 4099, + "Ġword": 3409, + "Ġworden": 30179, + "Ġwording": 60227, + "Ġwordpress": 75113, + "Ġwords": 4244, + "Ġwordt": 40260, + "Ġwore": 27570, + "Ġwork": 975, + "Ġworkaround": 59030, + "Ġworkbook": 45155, + "Ġworked": 6439, + "Ġworker": 11864, + "Ġworkers": 7337, + "Ġworkflow": 28288, + "Ġworkflows": 87468, + "Ġworkforce": 30927, + "Ġworking": 3238, + "Ġworkings": 78136, + "Ġworkload": 53596, + "Ġworkout": 25242, + "Ġworkouts": 45844, + "Ġworkplace": 26368, + "Ġworkplaces": 92062, + "Ġworks": 4278, + "Ġworksheet": 36636, + "Ġworksheets": 67525, + "Ġworkshop": 25073, + "Ġworkshops": 34836, + "Ġworkspace": 27514, + "Ġworkstation": 95891, + "Ġworld": 1879, + "Ġworldly": 96606, + "Ġworlds": 23859, + "Ġworldview": 77062, + "Ġworldwide": 15245, + "Ġworm": 34211, + "Ġworms": 61900, + "Ġworn": 23704, + "Ġworried": 17811, + "Ġworries": 37045, + "Ġworry": 10955, + "Ġworrying": 39776, + "Ġwors": 46193, + "Ġworse": 10960, + "Ġworsening": 92305, + "Ġworsh": 81978, + "Ġworship": 23916, + "Ġworst": 11785, + "Ġworth": 5802, + "Ġworthless": 65021, + "Ġworthwhile": 47676, + "Ġworthy": 27290, + "Ġwould": 1035, + "Ġwouldn": 8270, + "Ġwound": 26555, + "Ġwounded": 27493, + "Ġwounds": 33154, + "Ġwoven": 55774, + "Ġwow": 35665, + "Ġwp": 12609, + "ĠwpÅĤyw": 132631, + "Ġwr": 3112, + "Ġwrap": 15061, + "Ġwrapped": 19472, + "Ġwrapper": 13261, + "Ġwrappers": 78094, + "Ġwrapping": 41195, + "Ġwraps": 39709, + "Ġwrath": 63291, + "Ġwre": 66508, + "Ġwreak": 82575, + "Ġwreck": 35750, + "Ġwreckage": 87095, + "Ġwrench": 59488, + "Ġwrest": 23055, + "Ġwrestler": 82062, + "Ġwrestlers": 97830, + "Ġwrestling": 35440, + "Ġwrink": 57270, + "Ġwrinkles": 73107, + "Ġwrist": 32171, + "Ġwrists": 85778, + "Ġwrit": 2107, + "Ġwritable": 45905, + "Ġwrite": 3270, + "ĠwriteFile": 92820, + "ĠwriteTo": 83122, + "Ġwriteln": 80947, + "Ġwriter": 6916, + "Ġwriters": 16093, + "Ġwrites": 13914, + "Ġwriting": 4378, + "Ġwritings": 42322, + "Ġwritten": 5326, + "Ġwrong": 4969, + "Ġwrongdoing": 64228, + "Ġwrongful": 92632, + "Ġwrongly": 71209, + "Ġwrote": 6139, + "Ġwrought": 78603, + "ĠwrzeÅĽ": 137718, + "ĠwrzeÅĽnia": 137719, + "Ġws": 17624, + "Ġwsp": 62507, + "Ġwspóln": 133374, + "ĠwspóÅĤ": 127549, + "ĠwspóÅĤpr": 133908, + "ĠwspóÅĤprac": 133909, + "Ġwsz": 38391, + "Ġwszyst": 44827, + "Ġwt": 40473, + "Ġwur": 20600, + "Ġwurde": 26109, + "Ġwurden": 47650, + "Ġww": 65732, + "Ġwww": 8438, + "Ġwx": 9076, + "ĠwxDefault": 83586, + "ĠwxString": 56616, + "ĠwxT": 74197, + "Ġwy": 16590, + "Ġwyb": 143557, + "Ġwybra": 141362, + "ĠwybraÄĩ": 141363, + "Ġwybór": 143558, + "Ġwygl": 133563, + "ĠwyglÄħda": 133564, + "Ġwyja": 142442, + "ĠwyjaÅĽni": 142443, + "ĠwyjÄħ": 135759, + "ĠwyjÄħtk": 135760, + "Ġwykon": 98161, + "Ġwym": 71241, + "Ġwyn": 60934, + "Ġwypos": 136043, + "Ġwyposaż": 136044, + "Ġwys": 54547, + "Ġwysoko": 140256, + "ĠwysokoÅĽci": 140257, + "Ġwyst": 87564, + "ĠwystÄĻp": 135052, + "ĠwyÅĤÄħ": 141464, + "ĠwyÅĤÄħcznie": 141465, + "ĠwyÅĽ": 143084, + "ĠwyÅĽw": 143085, + "ĠwyÅĽwiet": 143086, + "ĠwyÅĽwietl": 143087, + "Ġwz": 62920, + "Ġwzgl": 127904, + "ĠwzglÄĻ": 137102, + "ĠwzglÄĻd": 140140, + "ĠwzglÄĻdu": 137103, + "Ġwäh": 52367, + "Ġwährend": 66983, + "Ġwäre": 68573, + "Ġwür": 43243, + "Ġwürde": 58881, + "ĠwÅĤ": 79224, + "ĠwÅĤa": 81906, + "ĠwÅĤad": 139693, + "ĠwÅĤadz": 139694, + "ĠwÅĤas": 136443, + "ĠwÅĤasne": 136444, + "ĠwÅĤaÅĽ": 140617, + "ĠwÅĤaÅĽc": 140618, + "ĠwÅĤaÅĽciciel": 140619, + "ĠwÅĤaÅĽnie": 131489, + "ĠwÅĤos": 141866, + "ĠwÅĤosów": 141867, + "ĠwÅĽ": 136781, + "ĠwÅĽród": 136782, + "Ġx": 856, + "ĠxAxis": 61185, + "ĠxOffset": 90638, + "ĠxPos": 75041, + "Ġxa": 71800, + "Ġxamarin": 95479, + "Ġxb": 55932, + "Ġxbmc": 82265, + "Ġxbox": 81472, + "Ġxc": 56901, + "Ġxcb": 95569, + "Ġxd": 79052, + "Ġxe": 82790, + "Ġxen": 52165, + "Ġxf": 59364, + "Ġxhr": 28282, + "Ġxhttp": 96197, + "Ġxi": 37590, + "Ġxl": 33008, + "Ġxlabel": 54778, + "Ġxlim": 93669, + "Ġxlink": 96239, + "Ġxls": 96578, + "Ġxm": 55249, + "Ġxmax": 52286, + "Ġxmin": 56459, + "Ġxml": 8396, + "ĠxmlDoc": 91467, + "Ġxmlhttp": 69300, + "Ġxmlns": 24967, + "Ġxmm": 57607, + "Ġxn": 54895, + "Ġxnxx": 94051, + "Ġxo": 97021, + "Ġxor": 53941, + "Ġxp": 37245, + "Ġxpath": 64525, + "Ġxpos": 74395, + "Ġxr": 38670, + "Ġxrange": 33073, + "Ġxs": 11943, + "Ġxsi": 37757, + "Ġxt": 62390, + "Ġxtype": 60614, + "Ġxu": 92654, + "Ġxuyên": 132278, + "Ġxuân": 142893, + "Ġxuất": 128415, + "Ġxuá»ijng": 129630, + "Ġxv": 83411, + "Ġxx": 20908, + "Ġxxx": 12319, + "Ġxxxx": 83976, + "Ġxy": 30784, + "Ġxyz": 40511, + "Ġxác": 129207, + "Ġxây": 128943, + "Ġxã": 128847, + "Ġxét": 130985, + "Ġxúc": 132242, + "Ġxưa": 139518, + "Ġxương": 138036, + "Ġxảy": 131871, + "Ġxấu": 135429, + "Ġxếp": 134143, + "Ġxứ": 134640, + "Ġxá»Ń": 129888, + "Ġy": 379, + "ĠyAxis": 62578, + "ĠyOffset": 81938, + "ĠyPos": 64679, + "Ġya": 13526, + "Ġyab": 139321, + "Ġyabanc": 139322, + "Ġyabancı": 139323, + "Ġyacc": 47364, + "Ġyacht": 71932, + "Ġyahoo": 64528, + "Ġyak": 73678, + "ĠyaklaÅŁ": 127814, + "ĠyaklaÅŁÄ±k": 136073, + "Ġyakın": 130009, + "Ġyal": 137824, + "Ġyalnız": 137825, + "Ġyaml": 32246, + "Ġyan": 82792, + "Ġyang": 10371, + "Ġyanlı": 137667, + "ĠyanlÄ±ÅŁ": 137668, + "Ġyans": 140230, + "Ġyansı": 140231, + "Ġyanı": 130057, + "Ġyanında": 134800, + "Ġyap": 39428, + "Ġyapt": 125801, + "Ġyaptı": 130583, + "Ġyaptır": 137466, + "Ġyaptıģı": 130333, + "Ġyapı": 84700, + "Ġyapıl": 125620, + "Ġyapılacak": 136966, + "Ġyapılan": 129573, + "Ġyapılması": 138541, + "Ġyapılır": 140646, + "Ġyapıyor": 135919, + "Ġyard": 19515, + "Ġyards": 13692, + "Ġyardı": 133344, + "Ġyardım": 132874, + "Ġyardımc": 133345, + "Ġyardımcı": 133346, + "Ġyarg": 141473, + "Ġyargı": 141474, + "Ġyarn": 38247, + "Ġyarı": 130397, + "Ġyat": 131454, + "Ġyatırım": 131455, + "ĠyavaÅŁ": 142050, + "Ġyaw": 45672, + "Ġyay": 97559, + "Ġyayg": 143839, + "Ġyaygın": 143840, + "Ġyayı": 143640, + "Ġyayıml": 143641, + "Ġyayımla": 143642, + "Ġyayın": 127525, + "Ġyayınlan": 140436, + "Ġyaz": 78114, + "Ġyazı": 129654, + "Ġyazılı": 135050, + "ĠyaÄŁ": 132724, + "ĠyaÅŁ": 83568, + "ĠyaÅŁad": 141664, + "ĠyaÅŁadıģı": 141665, + "ĠyaÅŁam": 131750, + "ĠyaÅŁama": 137788, + "ĠyaÅŁan": 131697, + "ĠyaÅŁayan": 139862, + "ĠyaÅŁÄ±nda": 135521, + "Ġyc": 81858, + "Ġydk": 62328, + "Ġye": 19478, + "Ġyeah": 21639, + "Ġyear": 1042, + "Ġyearly": 44270, + "Ġyears": 1635, + "Ġyeast": 40480, + "Ġyell": 64313, + "Ġyelled": 55077, + "Ġyelling": 52202, + "Ġyellow": 13753, + "Ġyen": 57340, + "Ġyeni": 98713, + "Ġyer": 54195, + "ĠyerleÅŁ": 141421, + "Ġyes": 9834, + "Ġyesterday": 13671, + "Ġyet": 3602, + "Ġyeti": 143417, + "ĠyetiÅŁtir": 143418, + "Ġyg": 92786, + "Ġyi": 61697, + "Ġyield": 7540, + "Ġyielded": 57387, + "Ġyielding": 77404, + "Ġyields": 35408, + "Ġyii": 20788, + "Ġylabel": 49812, + "Ġylim": 83914, + "Ġymax": 56078, + "Ġymin": 59115, + "Ġyn": 80761, + "Ġyo": 29496, + "Ġyog": 41353, + "Ġyoga": 26011, + "Ġyogurt": 54475, + "Ġyok": 89677, + "Ġyol": 93482, + "Ġyork": 49572, + "Ġyou": 498, + "Ġyoung": 3908, + "Ġyounger": 14650, + "Ġyoungest": 38537, + "Ġyoungster": 82271, + "Ġyoungsters": 59995, + "Ġyour": 697, + "Ġyoure": 70075, + "Ġyours": 18316, + "Ġyourself": 6133, + "Ġyourselves": 57896, + "Ġyouth": 12537, + "Ġyouthful": 64555, + "Ġyouths": 63962, + "Ġyoutube": 27178, + "ĠyoÄŁun": 135373, + "Ġyp": 75975, + "Ġypos": 70593, + "Ġyr": 42338, + "Ġyrs": 54023, + "Ġys": 31810, + "Ġyt": 68753, + "Ġyuan": 72283, + "Ġyuk": 142725, + "Ġyukarı": 142726, + "Ġyum": 68870, + "Ġyummy": 75354, + "Ġyup": 75446, + "Ġyy": 13197, + "Ġyyn": 85607, + "Ġyytype": 86396, + "Ġyyyy": 57989, + "Ġyên": 134226, + "Ġyêu": 128498, + "Ġyö": 97413, + "Ġyön": 130490, + "Ġyönel": 134158, + "Ġyönelik": 134159, + "Ġyönet": 125981, + "Ġyönetici": 141399, + "Ġyönetim": 132452, + "Ġyönt": 133028, + "Ġyöntem": 133029, + "Ġyü": 134135, + "Ġyük": 126736, + "Ġyüks": 134136, + "Ġyüksek": 129876, + "Ġyüksel": 134137, + "Ġyür": 137042, + "Ġyürüt": 137043, + "Ġyüz": 125723, + "Ġyüzde": 131008, + "Ġyüzden": 137599, + "Ġyı": 128034, + "Ġyık": 136998, + "Ġyıl": 128503, + "Ġyıll": 128035, + "Ġyıllarda": 142054, + "Ġyıllık": 139038, + "Ġyılı": 126851, + "Ġyılında": 130300, + "Ġyếu": 129553, + "Ġz": 1147, + "ĠzIndex": 59717, + "Ġza": 14694, + "Ġzab": 73071, + "Ġzach": 96812, + "Ġzad": 70648, + "Ġzag": 89019, + "Ġzahl": 88628, + "Ġzaj": 81347, + "Ġzak": 50315, + "ĠzakÅĤad": 135230, + "Ġzal": 65558, + "Ġzale": 142463, + "Ġzależy": 142464, + "Ġzam": 60476, + "Ġzaman": 83140, + "Ġzamów": 142343, + "Ġzamówienia": 142344, + "Ġzap": 32978, + "Ġzar": 58824, + "Ġzarówn": 134237, + "Ġzarówno": 134238, + "Ġzaw": 64007, + "Ġzb": 74355, + "Ġzd": 36249, + "Ġzdarma": 96052, + "ĠzdjÄĻ": 83737, + "ĠzdjÄĻcia": 136813, + "ĠzdjÄĻÄĩ": 143297, + "Ġze": 13703, + "Ġzeal": 68679, + "Ġzeigen": 85239, + "Ġzeigt": 77509, + "Ġzeit": 77158, + "Ġzelf": 59011, + "Ġzen": 53462, + "Ġzend": 78568, + "Ġzenith": 98429, + "Ġzer": 76178, + "Ġzero": 7168, + "Ġzeroes": 97443, + "Ġzeros": 17516, + "Ġzes": 137584, + "Ġzespo": 137585, + "ĠzespoÅĤ": 137586, + "Ġzest": 80531, + "ĠzewnÄĻtrzn": 141881, + "Ġzf": 99048, + "Ġzg": 92058, + "Ġzh": 59748, + "Ġzi": 63473, + "Ġzich": 52419, + "Ġzie": 74247, + "Ġzien": 79407, + "Ġzig": 85812, + "Ġzij": 91914, + "Ġzijn": 18488, + "Ġzinc": 48501, + "Ġzip": 10308, + "Ġzipcode": 80349, + "Ġzipfile": 86952, + "Ġzipper": 68761, + "Ġzk": 94528, + "Ġzlib": 81218, + "Ġzm": 47129, + "Ġzmq": 90740, + "Ġzn": 29874, + "Ġznaj": 67715, + "Ġznajdu": 139698, + "ĠznajdujÄħ": 139699, + "Ġznal": 141024, + "Ġznalaz": 141025, + "ĠznalazÅĤ": 141026, + "Ġznale": 137836, + "Ġznaleź": 137837, + "ĠznaleźÄĩ": 137838, + "Ġzo": 18741, + "Ġzoals": 83441, + "Ġzob": 128165, + "Ġzobac": 141611, + "ĠzobaczyÄĩ": 141612, + "Ġzobow": 142818, + "ĠzobowiÄħ": 142819, + "ĠzobowiÄħz": 142820, + "Ġzoek": 73800, + "Ġzoekt": 43702, + "Ġzombie": 37736, + "Ġzombies": 47073, + "Ġzona": 44012, + "Ġzonder": 74160, + "Ġzone": 10143, + "Ġzones": 25395, + "Ġzoning": 64978, + "Ġzoo": 40914, + "Ġzoom": 15562, + "Ġzost": 95926, + "Ġzosta": 61875, + "ĠzostaÄĩ": 142671, + "ĠzostaÅĤ": 130473, + "ĠzostaÅĤa": 134882, + "ĠzostaÅĤy": 137218, + "Ġzou": 72121, + "Ġzro": 135396, + "Ġzrobi": 135397, + "ĠzrobiÄĩ": 135398, + "Ġzs": 68512, + "Ġzu": 6395, + "Ġzug": 84911, + "Ġzum": 16034, + "Ġzun": 98175, + "Ġzunächst": 142596, + "ĠzupeÅĤ": 141998, + "ĠzupeÅĤnie": 141999, + "Ġzur": 17312, + "Ġzurück": 49503, + "Ġzus": 38158, + "Ġzusammen": 53383, + "Ġzusätzlich": 136249, + "Ġzw": 24396, + "Ġzwar": 82664, + "Ġzwarte": 92086, + "Ġzwe": 54936, + "Ġzwei": 40462, + "Ġzwischen": 47226, + "ĠzwiÄħz": 127970, + "ĠzwiÄħzane": 139067, + "ĠzwiÄħzku": 138136, + "Ġzwy": 143597, + "Ġzwyk": 143598, + "ĠzwykÅĤ": 143599, + "ĠzwÅĤ": 142384, + "ĠzwÅĤas": 142385, + "ĠzwÅĤaszc": 142386, + "ĠzwÅĤaszcza": 142387, + "Ġzx": 71322, + "Ġzz": 32633, + "Ġzza": 97649, + "ĠzÅĤ": 58068, + "Ġ{": 314, + "Ġ{!": 30344, + "Ġ{!!": 21353, + "Ġ{\"": 5212, + "Ġ{$": 13924, + "Ġ{%": 52261, + "Ġ{'": 5360, + "Ġ{(": 32798, + "Ġ{*": 29808, + "Ġ{*}": 47437, + "Ġ{-": 29669, + "Ġ{.": 50336, + "Ġ{...": 14916, + "Ġ{/*": 17234, + "Ġ{//": 17154, + "Ġ{:": 22069, + "Ġ{:.": 47518, + "Ġ{:?}\",": 71964, + "Ġ{?": 49376, + "Ġ{?>Ċ": 67615, + "Ġ{?}": 94469, + "Ġ{@": 6418, + "Ġ{[": 45591, + "Ġ{\\": 28152, + "Ġ{\\Ċ": 95519, + "Ġ{_": 48617, + "Ġ{{": 5867, + "Ġ{{$": 30108, + "Ġ{{--": 23896, + "Ġ{{--<": 57931, + "Ġ{{{": 47908, + "Ġ{{Ċ": 80505, + "Ġ{|": 40960, + "Ġ{}": 4687, + "Ġ{}\",": 24689, + "Ġ{}\".": 20863, + "Ġ{}'.": 25241, + "Ġ{})": 37713, + "Ġ{}));Ċ": 99007, + "Ġ{}),Ċ": 77872, + "Ġ{}).": 72727, + "Ġ{});Ċ": 35311, + "Ġ{});ĊĊ": 87894, + "Ġ{})Ċ": 35248, + "Ġ{},": 16452, + "Ġ{},Ċ": 14573, + "Ġ{}.": 84412, + "Ġ{}:": 90912, + "Ġ{};": 52166, + "Ġ{};Ċ": 9321, + "Ġ{};ĊĊ": 20375, + "Ġ{};čĊ": 55122, + "Ġ{}\\": 56057, + "Ġ{}Ċ": 5613, + "Ġ{}ĊĊ": 10086, + "Ġ{}ĊĊĊ": 91935, + "Ġ{}čĊ": 34141, + "Ġ{}čĊčĊ": 70846, + "Ġ{Ċ": 341, + "Ġ{ĊĊ": 1476, + "Ġ{ĊĊĊ": 18507, + "Ġ{ĊĊĊĊ": 70180, + "Ġ{čĊ": 972, + "Ġ{čĊčĊ": 8022, + "Ġ{čĊčĊčĊ": 94679, + "Ġ{ččĊ": 50474, + "Ġ|": 760, + "Ġ|-": 44561, + "Ġ|--": 30588, + "Ġ|--------------------------------------------------------------------------Ċ": 61250, + "Ġ|/": 97845, + "Ġ|=": 8662, + "Ġ|>": 58721, + "Ġ|\\": 63596, + "Ġ|_": 70886, + "Ġ|_|": 66091, + "Ġ||": 1369, + "Ġ||=": 35002, + "Ġ||Ċ": 8244, + "Ġ||čĊ": 45748, + "Ġ|Ċ": 9248, + "Ġ|ĊĊ": 35721, + "Ġ|čĊ": 49777, + "Ġ}": 335, + "Ġ}(": 95916, + "Ġ}()Ċ": 50746, + "Ġ})": 6413, + "Ġ})(": 48885, + "Ġ})();Ċ": 92228, + "Ġ}))": 87771, + "Ġ})),Ċ": 78384, + "Ġ})).": 70457, + "Ġ}));Ċ": 17830, + "Ġ}));ĊĊ": 27126, + "Ġ}))Ċ": 28712, + "Ġ}),": 31706, + "Ġ}),Ċ": 11973, + "Ġ}),ĊĊ": 50940, + "Ġ}).": 8533, + "Ġ}):": 86709, + "Ġ});": 18113, + "Ġ});Ċ": 1625, + "Ġ});ĊĊ": 3011, + "Ġ});ĊĊĊ": 27311, + "Ġ});ĊĊĊĊ": 81767, + "Ġ});čĊ": 12397, + "Ġ});čĊčĊ": 28375, + "Ġ})}Ċ": 29324, + "Ġ})Ċ": 2751, + "Ġ})ĊĊ": 9568, + "Ġ})ĊĊĊ": 61277, + "Ġ})čĊ": 25821, + "Ġ})čĊčĊ": 91276, + "Ġ}*/Ċ": 42833, + "Ġ}*/ĊĊ": 62021, + "Ġ},": 2470, + "Ġ},{": 89072, + "Ġ},{Ċ": 51387, + "Ġ},Ċ": 1153, + "Ġ},ĊĊ": 7331, + "Ġ},ĊĊĊ": 73952, + "Ġ},čĊ": 9879, + "Ġ},čĊčĊ": 54622, + "Ġ}.": 16908, + "Ġ}//": 17629, + "Ġ}:": 35374, + "Ġ};": 20066, + "Ġ};Ċ": 2605, + "Ġ};ĊĊ": 3634, + "Ġ};ĊĊĊ": 38497, + "Ġ};čĊ": 16511, + "Ġ};čĊčĊ": 26828, + "Ġ}": 83287, + "Ġ}>Ċ": 63156, + "Ġ}?>Ċ": 55556, + "Ġ}\\": 51300, + "Ġ}]": 58592, + "Ġ}]);Ċ": 88853, + "Ġ}])Ċ": 94239, + "Ġ}],Ċ": 29043, + "Ġ}];Ċ": 19376, + "Ġ}];ĊĊ": 85604, + "Ġ}]Ċ": 28394, + "Ġ}}": 3869, + "Ġ}}\"": 14483, + "Ġ}}\">": 41800, + "Ġ}}\"><": 75015, + "Ġ}}\">{{": 68177, + "Ġ}}\">Ċ": 22084, + "Ġ}}\"Ċ": 60559, + "Ġ}},Ċ": 64395, + "Ġ}}/": 92502, + "Ġ}};Ċ": 87037, + "Ġ}}": 45521, + "Ġ}}>{": 87715, + "Ġ}}>Ċ": 17844, + "Ġ}}}": 97901, + "Ġ}}Ċ": 8096, + "Ġ}}ĊĊ": 67436, + "Ġ}}čĊ": 82672, + "Ġ}Ċ": 456, + "Ġ}ĊĊ": 555, + "Ġ}ĊĊĊ": 4455, + "Ġ}ĊĊĊĊ": 16968, + "Ġ}ĊĊĊĊĊ": 43153, + "Ġ}ĊĊĊĊĊĊ": 79739, + "Ġ}čĊ": 1686, + "Ġ}čĊčĊ": 2553, + "Ġ}čĊčĊčĊ": 24097, + "Ġ}čĊčĊčĊčĊ": 75947, + "Ġ}ččĊ": 86729, + "Ġ~": 3968, + "Ġ~(": 34546, + "Ġ~/": 41495, + "Ġ~/.": 39958, + "Ġ~=": 21684, + "Ġ~>": 84632, + "Ġ~~": 77777, + "ĠÂ": 2139, + "Ġ¡": 48813, + "Ġ£": 6938, + "ĠÂ¥": 71488, + "Ġ§": 16625, + "Ġ§§": 64331, + "Ġ©": 7240, + "Ġ«": 12486, + "Ġ¬": 96165, + "Ġ®": 60479, + "Ġ°": 36286, + "Ġ±": 20287, + "Ġµ": 62912, + "Ġ¶": 78846, + "Ġ·": 9592, + "Ġ»": 8182, + "Ġ»,": 64701, + "Ġ».": 87891, + "Ġ»ĊĊ": 18292, + "Ġ¿": 28286, + "ĠÂł": 17091, + "ĠÂłĠÂł": 32045, + "ĠÂłĠÂłĠÂłĠÂł": 75620, + "ĠÂŃ": 49855, + "ĠÃ": 1683, + "Ġá": 18698, + "Ġágua": 132461, + "Ġál": 142865, + "Ġálbum": 142866, + "Ġáll": 94785, + "Ġámb": 143477, + "Ġámbito": 143478, + "Ġán": 128595, + "Ġáo": 129970, + "Ġáp": 129470, + "Ġárea": 54171, + "Ġáreas": 94062, + "Ġâ": 27905, + "Ġâm": 129610, + "Ġä": 12709, + "Ġähn": 141152, + "Ġähnlich": 141153, + "Ġäl": 88768, + "Ġän": 97985, + "Ġär": 18593, + "Ġäven": 93603, + "ĠÃ¥": 13076, + "ĠÃ¥r": 33980, + "ĠÃ¥rhus": 69124, + "Ġæ": 65013, + "Ġç": 18086, + "Ġça": 37943, + "ĠçalÄ±ÅŁ": 77487, + "ĠçalÄ±ÅŁan": 132384, + "ĠçalÄ±ÅŁm": 127721, + "ĠçalÄ±ÅŁma": 129711, + "ĠçalÄ±ÅŁmalar": 127722, + "ĠçalÄ±ÅŁmaları": 135122, + "ĠçalÄ±ÅŁtı": 139010, + "ĠçalÄ±ÅŁÄ±yor": 140363, + "Ġçarp": 140935, + "ĠçaÄŁ": 142850, + "ĠçaÄŁrı": 142851, + "Ġçe": 134147, + "Ġçek": 126428, + "Ġçekil": 140033, + "Ġçer": 141513, + "Ġçerç": 141514, + "Ġçerçeve": 141515, + "Ġçev": 127486, + "Ġçevir": 140383, + "Ġçevre": 133515, + "ĠçeÅŁ": 134148, + "ĠçeÅŁit": 134149, + "ĠçeÅŁitli": 134150, + "Ġçift": 136747, + "Ġçiz": 134810, + "Ġço": 135135, + "Ġçoc": 125321, + "Ġçocuk": 125939, + "Ġçocuklar": 136049, + "Ġçocukların": 143034, + "ĠçocuÄŁu": 138711, + "Ġçok": 57012, + "ĠçoÄŁu": 135136, + "Ġçöz": 127489, + "Ġçözüm": 133541, + "Ġçünkü": 137698, + "Ġçı": 125614, + "Ġçık": 125652, + "Ġçıkan": 134669, + "Ġçıkar": 129155, + "Ġçıkma": 137372, + "Ġçıkt": 134693, + "Ġçıktı": 134694, + "ĠçıkÄ±ÅŁ": 143310, + "Ġè": 11422, + "Ġé": 3958, + "Ġéc": 90229, + "Ġéconom": 81553, + "Ġéconomique": 135779, + "Ġécrit": 88432, + "Ġégalement": 43667, + "Ġél": 31913, + "Ġélect": 97051, + "Ġélectrique": 139390, + "Ġélev": 141449, + "Ġélevé": 141450, + "Ġélè": 142918, + "Ġélèves": 142919, + "Ġéléments": 141067, + "Ġép": 49410, + "Ġépoca": 133285, + "Ġéquip": 93516, + "Ġéquipé": 143886, + "Ġés": 21533, + "Ġét": 13916, + "Ġétaient": 96858, + "Ġétait": 42889, + "Ġétant": 138157, + "Ġétape": 141333, + "Ġété": 23639, + "Ġév": 50813, + "Ġéxito": 80886, + "Ġéén": 130269, + "Ġê": 19610, + "Ġêtes": 61199, + "Ġêtre": 22911, + "Ġî": 14364, + "Ġîn": 18531, + "Ġînt": 65574, + "Ġó": 40867, + "Ġórg": 143013, + "Ġórgão": 143014, + "Ġô": 128435, + "Ġông": 128436, + "Ġö": 16942, + "Ġöd": 127554, + "Ġöde": 135659, + "Ġödeme": 133941, + "Ġödül": 140579, + "Ġöff": 63318, + "Ġöl": 127909, + "Ġöld": 138925, + "Ġöldür": 138926, + "Ġölç": 137202, + "Ġölçü": 137203, + "Ġölüm": 137792, + "Ġön": 78605, + "Ġönce": 128489, + "Ġönem": 126392, + "Ġönemli": 128848, + "Ġöner": 136818, + "Ġöneri": 136819, + "Ġönlem": 140962, + "Ġönü": 143017, + "Ġönünde": 139455, + "Ġönüne": 143018, + "Ġör": 136235, + "Ġörg": 136493, + "Ġörgüt": 136494, + "Ġörnek": 136236, + "Ġöver": 76050, + "Ġöz": 125952, + "Ġözel": 129485, + "Ġözelli": 139782, + "Ġözellik": 127250, + "Ġözellikle": 132139, + "Ġözellikleri": 138118, + "ĠözelliÄŁi": 139783, + "Ġözgü": 135705, + "ĠÃ¶ÄŁ": 125281, + "ĠÃ¶ÄŁren": 130947, + "ĠÃ¶ÄŁrenc": 127589, + "ĠÃ¶ÄŁrenci": 135992, + "ĠÃ¶ÄŁrenciler": 134176, + "ĠÃ¶ÄŁret": 137092, + "ĠÃ¶ÄŁretmen": 137093, + "Ġø": 38118, + "Ġøns": 87292, + "Ġú": 17964, + "Ġúlt": 31429, + "Ġúltima": 71028, + "Ġúltimo": 61478, + "Ġúltimos": 78421, + "Ġún": 51163, + "Ġúnica": 88172, + "Ġúnico": 74301, + "Ġútil": 137741, + "Ġü": 10489, + "Ġüber": 13785, + "Ġüberh": 139905, + "Ġüberhaupt": 139906, + "Ġücret": 134799, + "Ġücrets": 137568, + "Ġücretsiz": 137569, + "Ġülk": 127868, + "Ġülke": 129871, + "Ġülkeler": 136675, + "Ġülkem": 139210, + "Ġülkemiz": 139211, + "Ġünivers": 137761, + "Ġüniversite": 137762, + "Ġünl": 142137, + "Ġünlü": 142138, + "Ġüret": 127016, + "Ġüretim": 133352, + "Ġürün": 125970, + "Ġürünler": 132309, + "Ġürünü": 142545, + "Ġüst": 127242, + "Ġüstün": 136195, + "Ġüy": 141747, + "Ġüye": 131153, + "Ġüyeler": 141748, + "Ġüyeleri": 141749, + "Ġüz": 130429, + "Ġüzer": 91557, + "Ġüzere": 130430, + "Ġüzerinde": 129947, + "Ġüzerinden": 132643, + "Ġüzerine": 130563, + "Ġüç": 132773, + "Ġý": 128431, + "Ġþ": 79607, + "ĠÃĢ": 64281, + "ĠÃģ": 42812, + "ĠÃģrea": 143733, + "ĠÃĤ": 45333, + "ĠÃĤu": 135285, + "ĠÃĥ": 73780, + "ĠÃĦ": 48686, + "ĠÃħ": 79252, + "ĠÃĩ": 63005, + "ĠÃĩa": 131098, + "ĠÃĩalÄ±ÅŁ": 137874, + "ĠÃĩok": 133987, + "ĠÃĩünkü": 133829, + "ĠÃĪ": 131390, + "ĠÃī": 28024, + "ĠÃİ": 59108, + "ĠÃĶ": 130797, + "ĠÃĶng": 130798, + "ĠÃĸ": 34037, + "ĠÃĸl": 136990, + "ĠÃĸn": 133470, + "ĠÃĸnce": 143567, + "ĠÃĸnceki": 143568, + "ĠÃĸsterreich": 140108, + "ĠÃĸz": 130307, + "ĠÃĸzel": 137215, + "ĠÃĸzellik": 136636, + "ĠÃĸzellikle": 136637, + "ĠÃĹ": 24768, + "ĠÃĺ": 88943, + "ĠÃľ": 30512, + "ĠÃľber": 49371, + "ĠÃľlke": 140320, + "ĠÃľnivers": 133696, + "ĠÃľniversites": 133697, + "ĠÃľniversitesi": 133698, + "ĠÃľrün": 140046, + "ĠÃľye": 137160, + "ĠÃł": 3784, + "ĠÃłs": 52529, + "ĠÃŃ": 40136, + "ĠÃŃch": 130123, + "ĠÃŃnd": 143282, + "ĠÃŃndice": 143283, + "ĠÃŃt": 129530, + "ĠÄ": 9843, + "Ġİ": 38421, + "Ġİki": 134361, + "Ġİl": 126803, + "Ġİlk": 133447, + "Ġİlçe": 140194, + "Ġİn": 131693, + "Ġİns": 135129, + "Ġİnsan": 135130, + "Ġİnt": 142600, + "Ġİnternet": 142601, + "Ġİs": 130173, + "Ġİsl": 137655, + "Ġİslam": 137656, + "Ġİstanbul": 129514, + "Ġİyi": 136681, + "Ġİz": 136284, + "Ġİzmir": 136285, + "Ġİç": 127607, + "Ġİçin": 137822, + "ĠÄ°ÅŁ": 126664, + "ĠÄ°ÅŁte": 137361, + "Ġısı": 139515, + "ĠÄĥn": 128442, + "ĠÄĩe": 81431, + "ĠÄį": 32802, + "ĠÄIJ": 50327, + "ĠÄIJi": 125807, + "ĠÄIJiá»ģu": 130406, + "ĠÄIJiá»ĥm": 141915, + "ĠÄIJiá»ĩn": 135601, + "ĠÄIJo": 138119, + "ĠÄIJoÃłn": 138120, + "ĠÄIJây": 130244, + "ĠÄIJình": 136929, + "ĠÄIJó": 134845, + "ĠÄIJông": 130215, + "ĠÄIJÃł": 129329, + "ĠÄIJÄĥng": 132719, + "ĠÄIJược": 135632, + "ĠÄIJưá»Ŀng": 136002, + "ĠÄIJại": 129393, + "ĠÄIJảng": 134355, + "ĠÄIJầu": 137720, + "ĠÄIJặc": 135359, + "ĠÄIJến": 136146, + "ĠÄIJức": 129938, + "ĠÄIJá»ģ": 142303, + "ĠÄIJá»ĥ": 130526, + "ĠÄIJá»ĭa": 134467, + "ĠÄIJá»ĭnh": 140401, + "ĠÄIJá»iji": 134914, + "ĠÄIJá»ĵng": 130717, + "ĠÄIJá»Ļ": 131283, + "ĠÄij": 14854, + "ĠÄija": 128907, + "ĠÄijang": 128348, + "ĠÄijau": 129952, + "ĠÄijem": 131881, + "ĠÄijen": 133549, + "ĠÄiji": 73586, + "ĠÄijiá»ģu": 128334, + "ĠÄijiá»ĥm": 128406, + "ĠÄijiá»ĥn": 138713, + "ĠÄijiá»ĩn": 128484, + "ĠÄijo": 125190, + "ĠÄijoán": 136149, + "ĠÄijoÃłn": 129675, + "ĠÄijoạn": 130393, + "ĠÄiju": 142920, + "ĠÄijua": 141305, + "ĠÄijuá»ķi": 142921, + "ĠÄijá": 129411, + "ĠÄijám": 136864, + "ĠÄijáng": 129861, + "ĠÄijánh": 128826, + "ĠÄijáo": 136335, + "ĠÄijáp": 130925, + "ĠÄijâu": 129625, + "ĠÄijây": 128358, + "ĠÄijã": 79126, + "ĠÄijè": 134399, + "ĠÄijèn": 134400, + "ĠÄijêm": 131571, + "ĠÄijình": 128890, + "ĠÄijò": 138584, + "ĠÄijòi": 138585, + "ĠÄijó": 128270, + "ĠÄijóng": 130889, + "ĠÄijô": 130287, + "ĠÄijôi": 130351, + "ĠÄijông": 131026, + "ĠÄijúng": 129315, + "ĠÄijÃłn": 129560, + "ĠÄijÃło": 131912, + "ĠÄijÃŃch": 134569, + "ĠÄijÄĥng": 129418, + "ĠÄijÆ¡n": 128666, + "ĠÄijưa": 128930, + "ĠÄijương": 138051, + "ĠÄijược": 63478, + "ĠÄijưá»Ŀng": 128448, + "ĠÄijại": 128479, + "ĠÄijạo": 128854, + "ĠÄijạt": 129664, + "ĠÄijả": 129550, + "ĠÄijảm": 129551, + "ĠÄijảng": 140431, + "ĠÄijảo": 131638, + "ĠÄijất": 128750, + "ĠÄijấu": 129507, + "ĠÄijấy": 135077, + "ĠÄijầu": 128278, + "ĠÄijầy": 129433, + "ĠÄijẩ": 132667, + "ĠÄijẩy": 132668, + "ĠÄijẳng": 136338, + "ĠÄijặc": 128788, + "ĠÄijặt": 128845, + "ĠÄijẹ": 128500, + "ĠÄijẹp": 128501, + "ĠÄijến": 128263, + "ĠÄijáºŃm": 143219, + "ĠÄijáºŃu": 140747, + "ĠÄijỡ": 137205, + "ĠÄijợi": 137183, + "ĠÄijá»§": 129078, + "ĠÄijứ": 137067, + "ĠÄijứa": 137068, + "ĠÄijức": 138063, + "ĠÄijứng": 130719, + "ĠÄijừng": 135094, + "ĠÄijá»±": 143200, + "ĠÄijá»±ng": 143201, + "ĠÄijá»ģ": 128456, + "ĠÄijá»ģu": 128582, + "ĠÄijá»ĥ": 79668, + "ĠÄijá»ĩ": 139112, + "ĠÄijá»īnh": 140868, + "ĠÄijá»ĭa": 128704, + "ĠÄijá»ĭch": 140787, + "ĠÄijá»ĭnh": 128333, + "ĠÄijá»įc": 130606, + "ĠÄijá»ı": 131928, + "ĠÄijá»ijc": 133796, + "ĠÄijá»iji": 128421, + "ĠÄijá»ĵ": 128913, + "ĠÄijá»ĵng": 128329, + "ĠÄijá»ķ": 136008, + "ĠÄijá»ķi": 128726, + "ĠÄijá»Ļ": 128368, + "ĠÄijá»Ļc": 129703, + "ĠÄijá»Ļi": 129188, + "ĠÄijá»Ļng": 128285, + "ĠÄijá»Ŀi": 129169, + "ĠÅ": 25917, + "ĠÅ¡": 36424, + "ĠÅ¡to": 79700, + "ĠÅ»": 138128, + "ĠÅ»y": 138129, + "Ġż": 46053, + "Ġże": 23345, + "Ġżeby": 131119, + "Ġży": 134625, + "Ġżyc": 73492, + "Ġżycia": 131472, + "Ġżycie": 133096, + "Ġżyciu": 139319, + "Ġżyw": 134626, + "Ġž": 30963, + "Ġže": 84774, + "ĠÅģ": 141297, + "ĠÅģa": 141298, + "ĠÅĤ": 74427, + "ĠÅĤad": 143481, + "ĠÅĤadn": 143482, + "ĠÅĤat": 142663, + "ĠÅĤatwo": 142664, + "ĠÅĤaz": 143161, + "ĠÅĤazien": 143162, + "ĠÅĤazienk": 143163, + "ĠÅĵ": 98125, + "ĠÅļ": 92027, + "ĠÅļw": 128193, + "ĠÅļwiat": 142466, + "ĠÅļwiÄĻt": 139340, + "ĠÅĽ": 29186, + "ĠÅĽc": 139587, + "ĠÅĽcian": 139588, + "ĠÅĽl": 141577, + "ĠÅĽlub": 141578, + "ĠÅĽm": 143254, + "ĠÅĽmier": 143255, + "ĠÅĽmierci": 143256, + "ĠÅĽredni": 135416, + "ĠÅĽro": 140339, + "ĠÅĽrod": 137537, + "ĠÅĽrodk": 140340, + "ĠÅĽrodków": 140341, + "ĠÅĽrodow": 137538, + "ĠÅĽrodowisk": 137539, + "ĠÅĽw": 81908, + "ĠÅĽwi": 127852, + "ĠÅĽwiad": 142289, + "ĠÅĽwiadom": 142290, + "ĠÅĽwiat": 130863, + "ĠÅĽwiata": 136431, + "ĠÅĽwie": 127784, + "ĠÅĽwiecie": 135703, + "ĠÅĽwież": 143607, + "ĠÅĽwiÄĻt": 138092, + "ĠÅŀ": 78819, + "ĠÅŀa": 130998, + "ĠÅŀe": 129964, + "ĠÅŀi": 132484, + "ĠÅŀimdi": 137748, + "ĠÅŀu": 133805, + "ĠÅŀub": 136040, + "ĠÅŀubat": 136041, + "ĠÅŁ": 22066, + "ĠÅŁa": 130338, + "ĠÅŁark": 139887, + "ĠÅŁarkı": 139888, + "ĠÅŁart": 133968, + "ĠÅŁeh": 133124, + "ĠÅŁek": 93295, + "ĠÅŁeker": 143041, + "ĠÅŁekil": 142693, + "ĠÅŁekilde": 128813, + "ĠÅŁeklin": 136031, + "ĠÅŁeklinde": 136032, + "ĠÅŁey": 126417, + "ĠÅŁeyler": 135434, + "ĠÅŁi": 35350, + "ĠÅŁidd": 140364, + "ĠÅŁiddet": 140365, + "ĠÅŁiir": 141697, + "ĠÅŁik": 144000, + "ĠÅŁikay": 144001, + "ĠÅŁikayet": 144002, + "ĠÅŁimdi": 133875, + "ĠÅŁirk": 131023, + "ĠÅŁirket": 131024, + "ĠÅŁu": 129607, + "ĠÅŁÃ¶yle": 134881, + "ĠÅŁÃ¼": 142507, + "ĠÅŁÃ¼phe": 142508, + "ĠÆ¡n": 133170, + "Ġưu": 130483, + "ĠÈĻi": 29334, + "ĠÎ": 7851, + "Ġα": 19043, + "Ġβ": 33218, + "Ġγ": 62027, + "Ġδ": 69334, + "Ġε": 59147, + "Ġκ": 71638, + "Ġλ": 48338, + "Ġμ": 32883, + "Ġν": 98686, + "ĠÎĵ": 84216, + "ĠÎĶ": 81163, + "ĠÏ": 17383, + "ĠÏĢ": 51745, + "ĠÏĥ": 47723, + "ĠÏĦ": 38470, + "ĠÏĨ": 97875, + "ĠÐ": 1278, + "ĠС": 27499, + "ĠСШ": 133917, + "ĠСШÐIJ": 133918, + "ĠСам": 131492, + "ĠСан": 140024, + "ĠСанкÑĤ": 140025, + "ĠСегоднÑı": 138679, + "ĠСейÑĩаÑģ": 140644, + "ĠСеÑĢ": 127418, + "ĠСеÑĢг": 142419, + "ĠСеÑĢгей": 142420, + "ĠСмоÑĤÑĢ": 142243, + "ĠСо": 126509, + "ĠСов": 137551, + "ĠСовеÑĤ": 137552, + "ĠСозд": 95679, + "ĠСообÑī": 142095, + "ĠСÑĤа": 131486, + "ĠТ": 50318, + "ĠТа": 131877, + "ĠТак": 125658, + "ĠТакже": 134075, + "ĠТаким": 141852, + "ĠТе": 130833, + "ĠТем": 136623, + "ĠТеп": 140649, + "ĠТепеÑĢÑĮ": 140650, + "ĠТо": 131345, + "ĠТолÑĮко": 141584, + "ĠТÑĥÑĢ": 138825, + "ĠУ": 58742, + "ĠУкÑĢа": 127872, + "ĠУкÑĢаин": 137829, + "ĠУкÑĢаинÑĭ": 136727, + "ĠФ": 66791, + "ĠФедеÑĢ": 127828, + "ĠФедеÑĢаÑĨии": 136238, + "ĠÐ¥": 125560, + "ĠХо": 137247, + "ĠХоÑĤ": 143955, + "ĠХоÑĤÑı": 143956, + "ĠЦ": 126699, + "ĠЦена": 142492, + "ĠЧ": 124713, + "ĠЧа": 134456, + "ĠЧе": 131234, + "ĠЧеÑĢ": 137594, + "ĠЧи": 134985, + "ĠЧÑĤ": 135579, + "ĠЧÑĤо": 130802, + "ĠЧÑĤобÑĭ": 135580, + "ĠШ": 128936, + "ĠЮ": 142055, + "ĠЮÑĢ": 142056, + "ĠЯ": 128557, + "Ġа": 20396, + "Ġаб": 142749, + "ĠабÑģол": 142750, + "ĠабÑģолÑİÑĤ": 142751, + "ĠабÑģолÑİÑĤно": 142752, + "Ġав": 67728, + "Ġавг": 141638, + "ĠавгÑĥÑģÑĤ": 141639, + "ĠавгÑĥÑģÑĤа": 141640, + "Ġави": 140273, + "Ġавиа": 140274, + "ĠавÑĤо": 129722, + "ĠавÑĤом": 125347, + "ĠавÑĤомаÑĤ": 127363, + "ĠавÑĤомаÑĤÑĭ": 137018, + "ĠавÑĤомоб": 141235, + "ĠавÑĤомоби": 141236, + "ĠавÑĤомобилÑĮ": 140663, + "ĠавÑĤомобилÑı": 141237, + "ĠавÑĤоÑĢ": 132718, + "Ġад": 142742, + "ĠадминиÑģÑĤ": 142743, + "ĠадминиÑģÑĤÑĢа": 142744, + "ĠадминиÑģÑĤÑĢаÑĨи": 142745, + "ĠадÑĢеÑģ": 90895, + "Ġак": 85330, + "ĠакÑĤив": 130304, + "ĠакÑĤÑĥ": 142991, + "ĠакÑĤÑĥалÑĮн": 142992, + "Ġал": 141812, + "Ġалк": 141813, + "Ġалког": 141814, + "Ġалкогол": 141815, + "ĠамеÑĢ": 142078, + "ĠамеÑĢик": 142079, + "ĠамеÑĢикан": 142080, + "ĠамеÑĢиканÑģк": 142081, + "Ġан": 127704, + "Ġана": 127723, + "Ġанализ": 135126, + "Ġаналог": 140892, + "Ġанг": 141020, + "Ġангли": 141021, + "Ġанглий": 141022, + "ĠанглийÑģк": 141023, + "ĠанÑĤи": 134986, + "Ġапп": 137693, + "ĠаппаÑĢаÑĤ": 137694, + "ĠапÑĢ": 139630, + "ĠапÑĢелÑı": 139631, + "ĠаÑĢенд": 139786, + "Ġб": 14062, + "Ġба": 128986, + "Ġбаз": 78273, + "Ġбан": 141931, + "Ġбанков": 141932, + "Ġбез": 91357, + "Ġбезоп": 127874, + "ĠбезопаÑģ": 127875, + "ĠбезопаÑģн": 140599, + "ĠбезопаÑģноÑģÑĤи": 136762, + "Ġбел": 131395, + "ĠбеÑĢ": 141052, + "ĠбеÑĢем": 141053, + "ĠбеÑĢеменн": 141054, + "ĠбеÑģ": 130014, + "ĠбеÑģп": 130060, + "ĠбеÑģплаÑĤ": 130061, + "ĠбеÑģплаÑĤно": 130062, + "Ġбиз": 132198, + "ĠбизнеÑģ": 132199, + "Ġбил": 142249, + "ĠбилеÑĤ": 142250, + "Ġбл": 90294, + "Ġблаг": 138422, + "Ġблагод": 138423, + "ĠблагодаÑĢ": 138424, + "ĠблагодаÑĢÑı": 138425, + "Ġбо": 129396, + "Ġбог": 137887, + "Ġбол": 128713, + "Ġболее": 128714, + "ĠболÑĮ": 83599, + "ĠболÑĮÑĪ": 126129, + "ĠболÑĮÑĪе": 129258, + "ĠболÑĮÑĪим": 143936, + "ĠболÑĮÑĪое": 143193, + "ĠболÑĮÑĪой": 136365, + "Ġбон": 135114, + "ĠбонÑĥÑģ": 135115, + "ĠбÑĢенд": 143140, + "ĠбÑĥд": 50470, + "ĠбÑĥдеÑĤ": 68823, + "ĠбÑĥдÑĥ": 134437, + "ĠбÑĥдÑĥÑĤ": 130414, + "ĠбÑĥк": 141750, + "ĠбÑĥкв": 141751, + "ĠбÑĥм": 142354, + "ĠбÑĥмаг": 142355, + "ĠбÑĭ": 37660, + "ĠбÑĭл": 128897, + "ĠбÑĭла": 129760, + "ĠбÑĭли": 129482, + "ĠбÑĭло": 128656, + "ĠбÑĭÑģÑĤÑĢо": 131993, + "ĠбÑĭÑĤÑĮ": 61631, + "ĠбÑİ": 136763, + "ĠбÑİдж": 136764, + "ĠбÑİджеÑĤ": 136765, + "Ġв": 5805, + "Ġва": 140745, + "Ġваж": 135496, + "Ġважно": 135497, + "Ġвал": 143490, + "ĠвалÑİÑĤ": 143491, + "Ġвам": 129379, + "ĠваÑĢи": 131939, + "ĠваÑĢианÑĤ": 131940, + "ĠваÑģ": 129998, + "ĠваÑĪ": 97017, + "ĠваÑĪего": 142044, + "ĠваÑĪи": 140746, + "ĠведÑĮ": 133820, + "Ġвек": 141626, + "Ġвека": 141627, + "Ġвели": 138486, + "Ġвелик": 138487, + "ĠвеÑĢ": 63262, + "ĠвеÑģ": 131480, + "ĠвеÑģÑĮ": 128113, + "ĠвеÑģÑĮма": 143021, + "ĠвеÑĩ": 137244, + "ĠвеÑĩеÑĢ": 137245, + "ĠвеÑī": 136788, + "ĠвеÑīеÑģÑĤв": 136789, + "Ġвз": 125811, + "ĠвзглÑıд": 140262, + "ĠвзÑıÑĤÑĮ": 141618, + "Ġвид": 72920, + "Ġвиде": 126994, + "Ġвидео": 130866, + "ĠвкÑĥÑģ": 135042, + "Ġвла": 138920, + "Ġвлад": 143319, + "ĠвладелÑĮ": 143320, + "ĠвладелÑĮÑĨ": 143321, + "ĠвлаÑģÑĤи": 138921, + "Ġвли": 134904, + "ĠвлиÑı": 134905, + "ĠвмеÑģÑĤ": 135013, + "ĠвмеÑģÑĤе": 135014, + "Ġвн": 143472, + "Ġвне": 132739, + "ĠвнеÑĪ": 143473, + "ĠвнеÑĪне": 143474, + "Ġвним": 133186, + "Ġвнимание": 133187, + "ĠвнÑĥÑĤ": 140467, + "ĠвнÑĥÑĤÑĢи": 140468, + "Ġво": 126231, + "Ġвод": 127680, + "Ġводо": 134841, + "ĠводÑĭ": 135091, + "Ġвоз": 85191, + "Ġвозд": 127924, + "ĠвоздейÑģÑĤв": 140736, + "ĠвоздейÑģÑĤви": 140737, + "ĠвоздÑĥÑħ": 137460, + "Ġвозмож": 125898, + "Ġвозможно": 132656, + "ĠвозможноÑģÑĤи": 137168, + "ĠвозможноÑģÑĤÑĮ": 131169, + "ĠвозÑĢаÑģÑĤ": 132887, + "Ġвой": 141967, + "ĠвойнÑĭ": 141968, + "Ġвок": 142983, + "ĠвокÑĢÑĥг": 142984, + "Ġволн": 141724, + "ĠволоÑģ": 135228, + "ĠвообÑīе": 135167, + "Ġвоп": 126608, + "ĠвопÑĢоÑģ": 126609, + "ĠвопÑĢоÑģÑĭ": 135693, + "ĠвоÑģ": 129430, + "ĠвоÑĤ": 130607, + "Ġвполн": 140794, + "Ġвполне": 140795, + "ĠвÑĢа": 133759, + "ĠвÑĢаÑĩ": 133760, + "ĠвÑĢем": 64711, + "ĠвÑĢемен": 134654, + "ĠвÑĢемени": 130775, + "ĠвÑĢемÑı": 91778, + "ĠвÑģ": 34614, + "ĠвÑģе": 43993, + "ĠвÑģег": 130355, + "ĠвÑģегда": 130356, + "ĠвÑģего": 129291, + "ĠвÑģей": 137189, + "ĠвÑģем": 130642, + "ĠвÑģеÑħ": 90993, + "ĠвÑģÑĤÑĢ": 127958, + "ĠвÑģÑĤÑĢеÑĩ": 137920, + "ĠвÑģÑĤÑĢеÑĩа": 140557, + "ĠвÑģÑİ": 140002, + "ĠвÑģÑı": 142478, + "ĠвÑģÑıк": 142479, + "ĠвÑģÑij": 131896, + "ĠвÑĤоÑĢ": 138616, + "ĠвÑĤоÑĢой": 138617, + "ĠвÑħод": 137514, + "ĠвÑĭ": 20811, + "ĠвÑĭб": 62129, + "ĠвÑĭбиÑĢа": 140209, + "ĠвÑĭбоÑĢ": 130699, + "ĠвÑĭбÑĢ": 137944, + "ĠвÑĭбÑĢаÑĤÑĮ": 137945, + "ĠвÑĭвод": 84294, + "ĠвÑĭгод": 139059, + "ĠвÑĭз": 96666, + "ĠвÑĭп": 70960, + "ĠвÑĭпол": 136721, + "ĠвÑĭполн": 98296, + "ĠвÑĭполнен": 133880, + "ĠвÑĭполнÑı": 136722, + "ĠвÑĭпÑĥÑģ": 136116, + "ĠвÑĭпÑĥÑģк": 136117, + "ĠвÑĭÑĢаж": 143587, + "ĠвÑĭÑĢажен": 143588, + "ĠвÑĭÑģ": 131799, + "ĠвÑĭÑģоко": 131800, + "ĠвÑĭÑħод": 132399, + "ĠвÑĭÑĪе": 134058, + "Ġг": 24725, + "ĠгаÑĢ": 137112, + "ĠгаÑĢан": 137113, + "ĠгаÑĢанÑĤи": 137114, + "Ġгде": 129162, + "ĠгеÑĢ": 137797, + "ĠгеÑĢо": 137798, + "Ġгл": 132853, + "Ġглав": 143003, + "Ġглавное": 143004, + "Ġглаз": 132854, + "ĠговоÑĢ": 132173, + "ĠговоÑĢиÑĤ": 132174, + "Ġгод": 125632, + "Ġгода": 128536, + "ĠгодÑĥ": 129257, + "ĠгодÑĭ": 139146, + "Ġгол": 138325, + "Ġголов": 132169, + "ĠголоÑģ": 138430, + "ĠгоÑĢ": 125191, + "ĠгоÑĢод": 126887, + "ĠгоÑĢода": 131644, + "ĠгоÑĢоде": 138462, + "ĠгоÑģ": 126081, + "ĠгоÑģÑĥд": 127637, + "ĠгоÑģÑĥдаÑĢ": 127638, + "ĠгоÑģÑĥдаÑĢÑģÑĤв": 127639, + "ĠгоÑģÑĥдаÑĢÑģÑĤвенн": 139653, + "ĠгоÑĤов": 130645, + "ĠгÑĢ": 140911, + "ĠгÑĢаж": 131363, + "ĠгÑĢаждан": 131364, + "ĠгÑĢаниÑĨ": 141283, + "ĠгÑĢаÑĦ": 143414, + "ĠгÑĢаÑĦик": 143415, + "ĠгÑĢн": 140912, + "ĠгÑĢÑĥпп": 127942, + "ĠгÑĢÑĥппа": 143298, + "ĠгÑĢÑĥппÑĭ": 137715, + "Ġд": 7796, + "Ġда": 91810, + "Ġдав": 139834, + "Ġдавно": 139835, + "Ġдаж": 129127, + "Ġдаже": 129128, + "Ġдал": 143624, + "Ġдалек": 143625, + "Ġдалеко": 143626, + "ĠдалÑĮ": 142377, + "ĠдалÑĮн": 141766, + "ĠдалÑĮÑĪе": 142378, + "Ġдан": 97499, + "Ġданного": 143390, + "Ġданной": 140785, + "Ġданном": 141536, + "ĠданнÑĭ": 34364, + "ĠданнÑĭе": 66081, + "ĠданнÑĭй": 140397, + "ĠданнÑĭм": 140856, + "ĠданнÑĭÑħ": 52276, + "Ġдв": 73154, + "Ġдва": 130876, + "Ġдве": 132676, + "Ġдвиг": 141744, + "ĠдвигаÑĤ": 141745, + "ĠдвигаÑĤел": 141746, + "Ġдвиж": 141046, + "ĠдвижениÑı": 141047, + "ĠдвоÑĢ": 140732, + "ĠдвÑĥÑħ": 132006, + "Ġде": 125543, + "Ġдев": 127913, + "ĠдевÑĥÑĪ": 137980, + "ĠдевÑĥÑĪк": 137981, + "ĠдейÑģÑĤв": 126114, + "ĠдейÑģÑĤвие": 142668, + "ĠдейÑģÑĤвиÑĤелÑĮно": 136896, + "ĠдейÑģÑĤвиÑı": 135470, + "Ġдек": 139384, + "Ġдекаб": 139385, + "ĠдекабÑĢÑı": 139386, + "Ġдел": 94269, + "Ġдела": 129691, + "ĠделаÑĤÑĮ": 133106, + "Ġделе": 137915, + "Ġдело": 133279, + "Ġдем": 141834, + "Ġдемо": 141835, + "Ġден": 127981, + "Ġденег": 138419, + "Ġденеж": 141916, + "Ġденежн": 141917, + "ĠденÑĮ": 126585, + "ĠденÑĮги": 131659, + "Ġдеп": 143381, + "ĠдепÑĥÑĤ": 143382, + "ĠдепÑĥÑĤаÑĤ": 143383, + "ĠдеÑĢ": 134817, + "ĠдеÑĢев": 134818, + "ĠдеÑĤ": 125870, + "ĠдеÑĤей": 130963, + "ĠдеÑĤи": 140497, + "ĠдеÑĤÑģк": 132951, + "ĠдеÑıÑĤ": 127358, + "ĠдеÑıÑĤелÑĮ": 127359, + "ĠдеÑıÑĤелÑĮноÑģÑĤи": 132796, + "ĠдеÑıÑĤелÑĮноÑģÑĤÑĮ": 140855, + "Ġди": 143755, + "Ġдиаг": 143756, + "ĠдиагноÑģÑĤ": 143757, + "Ġдиз": 135385, + "Ġдизайн": 135386, + "ĠдиÑĢ": 138364, + "ĠдиÑĢекÑĤоÑĢ": 138365, + "ĠдиÑģк": 136151, + "Ġдл": 142191, + "Ġдлин": 136319, + "ĠдлиÑĤелÑĮн": 142192, + "ĠдлÑı": 19849, + "Ġдней": 133141, + "ĠднÑı": 133072, + "Ġдо": 56197, + "Ġдоб": 140444, + "Ġдобав": 58269, + "ĠдобÑĢо": 140445, + "Ġдов": 138162, + "ĠдоволÑĮ": 138163, + "ĠдоволÑĮно": 138164, + "Ġдог": 131647, + "ĠдоговоÑĢ": 131648, + "ĠдокÑĥменÑĤ": 127067, + "ĠдокÑĥменÑĤов": 139475, + "Ġдол": 126092, + "Ġдолг": 139760, + "Ġдолго": 134806, + "Ġдолж": 55349, + "Ġдолжен": 132563, + "Ġдолжна": 136230, + "Ġдолжно": 136648, + "ĠдолжноÑģÑĤ": 142889, + "ĠдолжнÑĭ": 132128, + "Ġдолл": 139714, + "ĠдоллаÑĢ": 139715, + "Ġдом": 126711, + "Ġдома": 130174, + "Ġдоп": 142861, + "Ġдополн": 142862, + "ĠдополниÑĤелÑĮно": 142863, + "ĠдоÑĢ": 131487, + "ĠдоÑĢог": 131488, + "ĠдоÑģÑĤ": 66325, + "ĠдоÑģÑĤаÑĤ": 132459, + "ĠдоÑģÑĤаÑĤоÑĩно": 132460, + "ĠдоÑģÑĤиг": 139062, + "ĠдоÑģÑĤиж": 143365, + "ĠдоÑģÑĤижени": 143366, + "ĠдоÑģÑĤÑĥп": 91041, + "ĠдоÑħод": 136937, + "ĠдÑĢев": 139671, + "ĠдÑĢÑĥг": 80742, + "ĠдÑĢÑĥгие": 132083, + "ĠдÑĢÑĥгими": 142746, + "ĠдÑĢÑĥгиÑħ": 130599, + "ĠдÑĢÑĥгой": 135361, + "ĠдÑĢÑĥз": 143438, + "ĠдÑĢÑĥзÑĮÑı": 143439, + "ĠдÑĥÑħ": 141787, + "ĠдÑĥÑħов": 141788, + "Ġе": 22132, + "Ġев": 143943, + "ĠевÑĢ": 143944, + "ĠевÑĢо": 143125, + "ĠевÑĢоп": 143945, + "ĠевÑĢопей": 143946, + "Ġего": 72285, + "Ġедин": 141426, + "ĠединÑģÑĤв": 141427, + "Ġее": 128992, + "ĠемÑĥ": 131596, + "ĠеÑģли": 46253, + "ĠеÑģÑĤÑĮ": 79165, + "ĠеÑī": 134412, + "ĠеÑīе": 128863, + "ĠеÑīÑij": 134413, + "ĠеÑij": 134709, + "Ġж": 57217, + "Ġже": 128346, + "Ġжел": 139573, + "Ġжелез": 139574, + "Ġжен": 131161, + "ĠженÑīин": 131162, + "Ġжив": 134030, + "ĠживоÑĤ": 134031, + "Ġжид": 141996, + "Ġжидк": 141997, + "Ġжиз": 126768, + "Ġжизни": 129982, + "ĠжизнÑĮ": 132902, + "ĠжÑĥÑĢ": 137012, + "ĠжÑĥÑĢнал": 137013, + "Ġз": 10885, + "Ġза": 43846, + "Ġзаб": 140335, + "Ġзабол": 140336, + "Ġзаболев": 140337, + "ĠзаболеваниÑı": 140338, + "Ġзав": 92223, + "ĠзавиÑģ": 128039, + "ĠзавиÑģим": 139132, + "ĠзавиÑģимоÑģÑĤи": 139133, + "ĠзавиÑģиÑĤ": 139182, + "Ġзавод": 138651, + "Ġзаг": 71490, + "Ġзад": 52180, + "ĠзадаÑĩ": 133771, + "Ġзай": 140516, + "Ġзайм": 140517, + "Ġзак": 58213, + "Ġзаказ": 86905, + "Ġзакон": 129707, + "Ġзам": 127912, + "ĠзамеÑĤ": 137275, + "ĠзамеÑĩа": 141559, + "Ġзан": 135674, + "Ġзани": 131668, + "Ġзанима": 131669, + "ĠзанÑıÑĤ": 135675, + "Ġзап": 29789, + "ĠзапиÑģ": 55511, + "ĠзапиÑģи": 142573, + "ĠзапÑĢоÑģ": 69083, + "ĠзаÑĢ": 143917, + "ĠзаÑĢегиÑģÑĤ": 143918, + "ĠзаÑĢегиÑģÑĤÑĢи": 143919, + "ĠзаÑĢегиÑģÑĤÑĢиÑĢова": 143920, + "ĠзаÑĤ": 138264, + "ĠзаÑĤем": 138265, + "ĠзаÑīиÑĤ": 141946, + "ĠзаÑīиÑĤÑĭ": 141947, + "ĠзаÑı": 128103, + "ĠзаÑıв": 128104, + "ĠзаÑıвил": 140009, + "ĠзаÑıвк": 143734, + "ĠздеÑģÑĮ": 131337, + "ĠздоÑĢов": 142421, + "ĠздоÑĢовÑĮÑı": 142422, + "Ġзим": 138812, + "Ġзло": 143112, + "Ġзн": 143198, + "Ġзна": 127903, + "ĠзнаеÑĤ": 140131, + "Ġзнак": 143119, + "ĠзнаÑĤÑĮ": 143199, + "ĠзнаÑĩ": 39550, + "ĠзнаÑĩение": 65359, + "ĠзнаÑĩениÑı": 87512, + "ĠзнаÑĩиÑĤ": 138721, + "ĠзнаÑĩиÑĤелÑĮно": 139951, + "ĠзнаÑİ": 137096, + "ĠзÑĢ": 143220, + "ĠзÑĢениÑı": 143221, + "ĠзÑĥб": 140212, + "Ġи": 7587, + "Ġиг": 70522, + "ĠигÑĢ": 125635, + "ĠигÑĢаÑĤÑĮ": 135242, + "ĠигÑĢов": 138601, + "ĠигÑĢовÑĭе": 138602, + "ĠигÑĢок": 134633, + "ĠигÑĢÑĭ": 131420, + "Ġиде": 142567, + "ĠидеалÑĮн": 142568, + "Ġиз": 23064, + "ĠизвеÑģÑĤ": 131199, + "ĠизгоÑĤов": 139968, + "ĠизгоÑĤовлен": 139969, + "Ġизд": 143879, + "Ġиздели": 143880, + "ĠизделиÑı": 143881, + "Ġизмен": 70021, + "ĠизменениÑı": 138506, + "Ġили": 45077, + "Ġим": 47735, + "Ġиме": 125798, + "ĠимееÑĤ": 130318, + "Ġимени": 142112, + "Ġименно": 131122, + "ĠимеÑĤÑĮ": 140292, + "ĠимеÑİÑĤ": 134850, + "ĠимÑı": 84656, + "Ġин": 40978, + "ĠинвеÑģÑĤи": 143347, + "ĠинвеÑģÑĤиÑĨи": 143348, + "Ġиногда": 142013, + "ĠиноÑģÑĤ": 143058, + "ĠиноÑģÑĤÑĢан": 143059, + "ĠиноÑģÑĤÑĢанн": 143060, + "ĠинÑĤеÑĢ": 96951, + "ĠинÑĤеÑĢеÑģ": 129611, + "ĠинÑĤеÑĢнеÑĤ": 130231, + "ĠинÑĤеÑĢÑĮеÑĢ": 143275, + "ĠинÑĦоÑĢм": 81983, + "ĠинÑĦоÑĢмаÑĨии": 135074, + "ĠинÑĦоÑĢмаÑĨиÑİ": 137673, + "ĠинÑĦоÑĢмаÑĨиÑı": 139493, + "ĠиÑģ": 125109, + "ĠиÑģкÑĥÑģ": 136456, + "ĠиÑģкÑĥÑģÑģÑĤв": 136457, + "ĠиÑģп": 52587, + "ĠиÑģпол": 142546, + "ĠиÑģполни": 142547, + "ĠиÑģполÑĮз": 89695, + "ĠиÑģполÑĮзов": 88227, + "ĠиÑģполÑĮзова": 135817, + "ĠиÑģполÑĮзование": 138710, + "ĠиÑģполÑĮзованиÑı": 138719, + "ĠиÑģполÑĮзоваÑĤÑĮ": 132805, + "ĠиÑģполÑĮзÑĥеÑĤÑģÑı": 143248, + "ĠиÑģполÑĮзÑĥÑİÑĤ": 139193, + "ĠиÑģпÑĭÑĤа": 143302, + "ĠиÑģÑģлед": 127959, + "ĠиÑģÑģледова": 137938, + "ĠиÑģÑģледованиÑı": 141441, + "ĠиÑģÑĤ": 126144, + "ĠиÑģÑĤоÑĢ": 127927, + "ĠиÑģÑĤоÑĢии": 137524, + "ĠиÑģÑĤоÑĢиÑı": 143982, + "ĠиÑģÑĤоÑĩ": 137403, + "ĠиÑģÑĤоÑĩник": 137404, + "ĠиÑħ": 128488, + "ĠиÑİ": 128082, + "ĠиÑİлÑı": 140460, + "ĠиÑİнÑı": 139775, + "Ġк": 7665, + "Ġка": 143986, + "Ġкаж": 143845, + "Ġкажд": 88566, + "Ġкаждого": 138393, + "Ġкаждой": 143285, + "Ġкаждом": 140567, + "ĠкаждÑĭй": 132946, + "ĠкажеÑĤÑģÑı": 143846, + "Ġказ": 131046, + "Ġказино": 131047, + "Ġкак": 51670, + "Ġкакие": 136048, + "Ġкакой": 134207, + "Ġкам": 140945, + "ĠкамеÑĢ": 140946, + "Ġкан": 138205, + "Ġканал": 138206, + "Ġкап": 127894, + "ĠкапиÑĤ": 141159, + "ĠкапиÑĤал": 141160, + "ĠкаÑĢ": 129297, + "ĠкаÑĢÑĤ": 90634, + "ĠкаÑĢÑĤин": 134341, + "ĠкаÑĤ": 143890, + "ĠкаÑĤал": 143891, + "ĠкаÑĤалог": 143892, + "ĠкаÑĤегоÑĢ": 143117, + "ĠкаÑĤегоÑĢии": 143118, + "ĠкаÑĦе": 143987, + "ĠкаÑĩ": 125296, + "ĠкаÑĩеÑģÑĤв": 125902, + "ĠкаÑĩеÑģÑĤва": 137094, + "ĠкаÑĩеÑģÑĤве": 131191, + "ĠкаÑĩеÑģÑĤво": 139403, + "ĠкваÑĢ": 130828, + "ĠкваÑĢÑĤиÑĢ": 130829, + "Ġкино": 138288, + "Ġкл": 47346, + "ĠклаÑģÑģ": 59648, + "Ġкли": 89089, + "ĠклиенÑĤ": 131562, + "ĠклÑĥб": 135549, + "ĠклÑİÑĩ": 96355, + "Ġкни": 127398, + "Ġкниг": 132981, + "Ġкниги": 139984, + "Ġкноп": 84949, + "Ġкнопк": 143197, + "Ġко": 87415, + "Ġкогда": 128895, + "Ġкод": 68026, + "Ġкож": 139044, + "Ġкожи": 139045, + "Ġкол": 50448, + "ĠколиÑĩеÑģÑĤво": 88601, + "Ġколл": 143747, + "ĠколлекÑĤив": 143748, + "Ġком": 43116, + "Ġкоманд": 131418, + "ĠкомиÑģ": 142047, + "ĠкомиÑģÑģи": 142048, + "ĠкомнаÑĤ": 135040, + "Ġкомп": 76473, + "Ġкомпани": 126821, + "Ġкомпании": 130188, + "Ġкомпаний": 142490, + "ĠкомпаниÑı": 134979, + "ĠкомплекÑģ": 133516, + "ĠкомпÑĮÑİ": 134231, + "ĠкомпÑĮÑİÑĤеÑĢ": 134232, + "ĠкомÑĦ": 139095, + "ĠкомÑĦоÑĢÑĤ": 139096, + "Ġкон": 36998, + "ĠконеÑĩно": 133859, + "ĠконкÑĢ": 138897, + "ĠконкÑĢеÑĤ": 138898, + "ĠконкÑĥÑĢ": 137507, + "ĠконкÑĥÑĢÑģ": 137508, + "ĠконÑģÑĤÑĢÑĥк": 139686, + "ĠконÑģÑĤÑĢÑĥкÑĨи": 139687, + "ĠконÑĤ": 76985, + "ĠконÑĤак": 139404, + "ĠконÑĤакÑĤ": 139405, + "ĠконÑĤÑĢ": 132160, + "ĠконÑĤÑĢол": 132161, + "ĠконÑĨе": 136229, + "ĠкоÑĢ": 66869, + "ĠкоÑĢпÑĥÑģ": 142574, + "ĠкоÑĤоÑĢ": 37053, + "ĠкоÑĤоÑĢаÑı": 130407, + "ĠкоÑĤоÑĢого": 133785, + "ĠкоÑĤоÑĢое": 134546, + "ĠкоÑĤоÑĢой": 132226, + "ĠкоÑĤоÑĢом": 133669, + "ĠкоÑĤоÑĢÑĥÑİ": 137778, + "ĠкоÑĤоÑĢÑĭе": 128548, + "ĠкоÑĤоÑĢÑĭй": 97751, + "ĠкоÑĤоÑĢÑĭм": 135810, + "ĠкоÑĤоÑĢÑĭÑħ": 130455, + "ĠкÑĢай": 143110, + "ĠкÑĢайне": 143111, + "ĠкÑĢаÑģ": 127736, + "ĠкÑĢаÑģив": 135329, + "ĠкÑĢаÑģоÑĤ": 142171, + "ĠкÑĢед": 130926, + "ĠкÑĢедиÑĤ": 130927, + "ĠкÑĢов": 142238, + "ĠкÑĢови": 142239, + "ĠкÑĢоме": 141727, + "ĠкÑĢÑĥ": 134419, + "ĠкÑĢÑĥг": 137023, + "ĠкÑĢÑĥп": 134420, + "ĠкÑĢÑĭ": 143615, + "ĠкÑĢÑĭÑĪ": 143616, + "ĠкÑĤо": 129796, + "ĠкÑĥд": 141012, + "ĠкÑĥда": 141013, + "ĠкÑĥлÑĮÑĤ": 141959, + "ĠкÑĥлÑĮÑĤÑĥ": 141960, + "ĠкÑĥлÑĮÑĤÑĥÑĢÑĭ": 141961, + "ĠкÑĥпиÑĤÑĮ": 130074, + "ĠкÑĥÑĢÑģ": 132961, + "Ġл": 25460, + "Ġлег": 133362, + "Ġлегко": 133363, + "ĠлеÑĤ": 128696, + "ĠлеÑĩ": 138560, + "ĠлеÑĩение": 140381, + "ĠлеÑĩениÑı": 138561, + "Ġли": 58095, + "Ġлид": 142950, + "ĠлидеÑĢ": 142951, + "ĠлиÑĤ": 140564, + "ĠлиÑĤеÑĢ": 140565, + "ĠлиÑĤеÑĢаÑĤÑĥÑĢ": 140566, + "ĠлиÑĨ": 131111, + "ĠлиÑĨа": 135279, + "ĠлиÑĩно": 134609, + "ĠлиÑĪ": 127225, + "ĠлиÑĪÑĮ": 131991, + "ĠлÑĥÑĩ": 126891, + "ĠлÑĥÑĩÑĪе": 130458, + "ĠлÑĥÑĩÑĪиÑħ": 143937, + "ĠлÑİ": 93036, + "ĠлÑİб": 125389, + "ĠлÑİбим": 138158, + "ĠлÑİбого": 143736, + "ĠлÑİбой": 134706, + "ĠлÑİбом": 141793, + "ĠлÑİдей": 131007, + "ĠлÑİди": 132782, + "Ġм": 11310, + "Ġма": 128692, + "ĠмакÑģ": 140684, + "ĠмакÑģималÑĮ": 140685, + "ĠмакÑģималÑĮно": 140686, + "Ġмал": 127575, + "Ġмало": 134083, + "ĠмалÑĭÑĪ": 142283, + "Ġмам": 141617, + "ĠмаÑĢ": 126064, + "ĠмаÑĢÑĤа": 138856, + "ĠмаÑĢÑĪ": 143872, + "ĠмаÑĢÑĪÑĢÑĥÑĤ": 143873, + "ĠмаÑģ": 142762, + "ĠмаÑģло": 142763, + "ĠмаÑģÑģ": 131434, + "ĠмаÑģÑģив": 50237, + "ĠмаÑģÑĤ": 135903, + "ĠмаÑģÑĤеÑĢ": 135904, + "ĠмаÑĤ": 95473, + "ĠмаÑĤеÑĢи": 136970, + "ĠмаÑĤеÑĢиал": 128076, + "ĠмаÑĤеÑĢиалов": 139696, + "ĠмаÑĤеÑĢиалÑĭ": 141158, + "ĠмаÑĤÑĩ": 138357, + "ĠмаÑĪ": 130838, + "ĠмаÑĪин": 130839, + "ĠмаÑı": 137659, + "Ġме": 127336, + "Ġмед": 127620, + "Ġмеди": 137972, + "ĠмеждÑĥ": 130724, + "Ġмен": 68744, + "Ġменее": 133959, + "ĠменÑĮ": 137200, + "ĠменÑĮÑĪе": 137201, + "ĠменÑı": 129052, + "ĠмеÑĢ": 143779, + "ĠмеÑĢоп": 143780, + "ĠмеÑĢопÑĢи": 143781, + "ĠмеÑĢопÑĢиÑıÑĤиÑı": 143782, + "ĠмеÑģÑĤ": 125270, + "ĠмеÑģÑĤа": 134548, + "ĠмеÑģÑĤе": 136052, + "ĠмеÑģÑĤо": 130746, + "ĠмеÑģÑı": 127337, + "ĠмеÑģÑıÑĨ": 132698, + "ĠмеÑģÑıÑĨев": 138693, + "ĠмеÑĤ": 127765, + "ĠмеÑĤал": 135565, + "ĠмеÑĤалл": 135566, + "ĠмеÑĤод": 55629, + "ĠмеÑĤÑĢо": 140184, + "ĠмеÑħ": 143430, + "ĠмеÑħаниз": 143431, + "ĠмеÑħанизм": 143432, + "Ġмик": 135189, + "ĠмикÑĢо": 135190, + "Ġмил": 128073, + "Ġмилли": 139641, + "Ġмиллион": 139642, + "Ġмин": 97868, + "ĠминималÑĮ": 143857, + "ĠминималÑĮн": 143858, + "ĠминÑĥÑĤ": 131687, + "ĠмиÑĢ": 137144, + "ĠмиÑĢа": 135093, + "ĠмиÑĢе": 137145, + "ĠмиÑĢов": 138061, + "Ġмл": 137958, + "Ġмлн": 137959, + "Ġмм": 134774, + "Ġмн": 126133, + "Ġмне": 129075, + "Ġмног": 127869, + "Ġмногие": 136676, + "ĠмногиÑħ": 137908, + "Ġмного": 128991, + "Ġмнож": 141311, + "ĠмножеÑģÑĤво": 141312, + "Ġмо": 128468, + "ĠмобилÑĮ": 138625, + "ĠмобилÑĮн": 138626, + "Ġмог": 125649, + "ĠмогÑĥ": 139572, + "ĠмогÑĥÑĤ": 129348, + "Ġмод": 70139, + "Ġмодели": 138082, + "ĠмоделÑĮ": 141335, + "Ġмож": 44483, + "ĠможеÑĤ": 66988, + "ĠможеÑĤе": 130619, + "Ġможно": 90298, + "Ġмой": 139732, + "Ġмолод": 132210, + "Ġмом": 130485, + "ĠмоменÑĤ": 130486, + "ĠмоÑī": 137034, + "ĠмоÑīн": 137035, + "ĠмÑĥ": 133294, + "ĠмÑĥж": 127584, + "ĠмÑĥжÑĩин": 134144, + "ĠмÑĥзÑĭк": 133295, + "ĠмÑĭ": 87200, + "ĠмÑı": 141224, + "ĠмÑıг": 141225, + "ĠмÑıгк": 141226, + "ĠмÑıÑģ": 143499, + "Ġн": 6709, + "Ġна": 13073, + "Ġнаб": 140202, + "ĠнаблÑİ": 140203, + "ĠнаблÑİда": 140204, + "Ġнаг": 143960, + "ĠнагÑĢÑĥз": 143961, + "ĠнагÑĢÑĥзк": 143962, + "Ġнад": 126939, + "Ġнадо": 131507, + "Ġназ": 127550, + "Ġназад": 133915, + "Ġназвание": 138943, + "ĠназнаÑĩен": 139779, + "Ġнаи": 135804, + "Ġнаибол": 135805, + "Ġнаиболее": 135806, + "Ġнай": 134173, + "Ġнайд": 86758, + "ĠнайÑĤи": 134174, + "ĠналиÑĩ": 128139, + "ĠналиÑĩие": 142132, + "ĠналиÑĩии": 140885, + "Ġналог": 139674, + "Ġналогов": 139675, + "Ġнам": 130512, + "Ġнап": 90005, + "ĠнапÑĢав": 131584, + "ĠнапÑĢавлен": 131585, + "ĠнаÑĢ": 141824, + "ĠнаÑĢко": 141825, + "ĠнаÑĢÑĥ": 135338, + "ĠнаÑĢÑĥÑĪен": 135339, + "ĠнаÑģ": 126460, + "ĠнаÑģел": 140877, + "ĠнаÑģелениÑı": 140878, + "ĠнаÑģÑĤоÑı": 127754, + "ĠнаÑģÑĤоÑıÑī": 127755, + "ĠнаÑģÑĤоÑıÑīее": 142526, + "ĠнаÑĤÑĥÑĢ": 143114, + "ĠнаÑĤÑĥÑĢалÑĮн": 143115, + "ĠнаÑĥк": 138912, + "ĠнаÑĥÑĩ": 133695, + "ĠнаÑħод": 127254, + "ĠнаÑħодиÑĤÑģÑı": 135480, + "ĠнаÑĩ": 72542, + "ĠнаÑĩал": 129971, + "ĠнаÑĩала": 135883, + "ĠнаÑĩина": 131107, + "ĠнаÑĪ": 126096, + "ĠнаÑĪего": 140072, + "ĠнаÑĪей": 134978, + "ĠнаÑĪем": 135999, + "ĠнаÑĪи": 137122, + "ĠнаÑĪиÑħ": 139290, + "Ġне": 18658, + "Ġнев": 142323, + "Ġневоз": 142324, + "Ġневозможно": 142325, + "Ġнего": 131326, + "Ġнед": 127782, + "Ġнедел": 143223, + "ĠнеделÑİ": 143224, + "Ġнедо": 135681, + "Ġнек": 126164, + "ĠнекоÑĤоÑĢ": 126165, + "ĠнекоÑĤоÑĢÑĭе": 139437, + "ĠнекоÑĤоÑĢÑĭÑħ": 141059, + "ĠнелÑĮ": 138914, + "ĠнелÑĮз": 138915, + "ĠнелÑĮзÑı": 138916, + "Ġнем": 137966, + "Ġнемного": 137967, + "ĠнеобÑħодим": 125742, + "ĠнеобÑħодимо": 129945, + "ĠнеобÑħодимоÑģÑĤи": 142280, + "ĠнеÑģк": 126814, + "ĠнеÑģколÑĮ": 141497, + "ĠнеÑģколÑĮки": 141498, + "ĠнеÑģколÑĮкиÑħ": 141499, + "ĠнеÑģколÑĮко": 130157, + "ĠнеÑĤ": 76638, + "Ġни": 126491, + "Ġниже": 139784, + "Ġник": 126073, + "Ġникак": 134212, + "Ġникогда": 139544, + "ĠникÑĤо": 142084, + "ĠниÑħ": 129690, + "ĠниÑĩ": 134009, + "ĠниÑĩего": 134010, + "Ġно": 128330, + "Ġнов": 48951, + "Ġнового": 137008, + "ĠновоÑģÑĤи": 141268, + "ĠновÑĭе": 136035, + "ĠновÑĭй": 136415, + "ĠновÑĭÑħ": 136820, + "Ġног": 135960, + "ĠномеÑĢ": 73934, + "ĠноÑĢм": 127844, + "ĠноÑĢмалÑĮн": 141177, + "ĠноÑĩ": 142430, + "ĠноÑĩÑĮ": 142431, + "ĠноÑı": 140522, + "ĠноÑıбÑĢÑı": 140523, + "ĠнÑĥж": 75983, + "ĠнÑĥжно": 129100, + "ĠнÑĥжнÑĭ": 141217, + "Ġо": 8215, + "Ġоб": 21229, + "ĠобеÑģп": 137457, + "ĠобеÑģпеÑĩ": 137458, + "ĠобеÑģпеÑĩива": 137459, + "Ġобла": 126684, + "ĠоблаÑģÑĤи": 129729, + "ĠоблаÑģÑĤÑĮ": 137291, + "Ġобо": 130584, + "ĠобоÑĢ": 128148, + "ĠобоÑĢÑĥд": 128149, + "ĠобоÑĢÑĥдование": 141681, + "ĠобоÑĢÑĥдованиÑı": 141123, + "ĠобÑĢаз": 127273, + "ĠобÑĢазованиÑı": 136731, + "ĠобÑĢазом": 132256, + "ĠобÑĢаÑĤ": 131705, + "ĠобÑĢаÑī": 139617, + "ĠобÑĢаÑīа": 139618, + "ĠобÑģлÑĥж": 138446, + "ĠобÑģлÑĥжива": 138447, + "ĠобÑĥÑĩ": 143180, + "ĠобÑĥÑĩениÑı": 143181, + "ĠобÑī": 142217, + "ĠобÑīеÑģÑĤв": 142218, + "ĠобÑīеÑģÑĤвенно": 142219, + "ĠобÑĬ": 47340, + "ĠобÑĬекÑĤ": 53953, + "ĠобÑĬем": 134718, + "ĠобÑĬÑı": 136914, + "ĠобÑĬÑıвл": 136915, + "ĠобÑĭ": 135925, + "ĠобÑĭÑĩ": 135926, + "ĠобÑĭÑĩно": 135927, + "ĠобÑıз": 127838, + "ĠобÑıзан": 138823, + "ĠобÑıзаÑĤелÑĮ": 136346, + "ĠобÑıзаÑĤелÑĮно": 136347, + "ĠогÑĢ": 127776, + "ĠогÑĢани": 137872, + "ĠогÑĢаниÑĩен": 137873, + "ĠогÑĢомн": 135663, + "Ġод": 52319, + "Ġодеж": 137073, + "Ġодежд": 137074, + "Ġодин": 129385, + "Ġодна": 136109, + "Ġоднако": 138284, + "Ġодним": 138502, + "Ġодно": 131550, + "Ġодного": 134682, + "Ġодной": 133304, + "Ġодном": 138433, + "ĠоднÑĥ": 142060, + "Ġожи": 137928, + "Ġожида": 137929, + "Ġок": 71564, + "Ġоказ": 127595, + "Ġоказа": 134236, + "ĠоказÑĭва": 142131, + "Ġокол": 134304, + "Ġоколо": 134305, + "ĠокÑĢÑĥ": 142482, + "ĠокÑĢÑĥг": 143652, + "ĠокÑĢÑĥжа": 142483, + "ĠокÑĢÑĥжаÑİÑī": 142484, + "ĠокÑĤÑı": 139791, + "ĠокÑĤÑıбÑĢÑı": 139792, + "Ġон": 98745, + "Ġона": 129171, + "Ġони": 128729, + "Ġонл": 129210, + "Ġонлайн": 129211, + "Ġоп": 45239, + "ĠопаÑģ": 136129, + "ĠопеÑĢ": 96958, + "ĠопиÑģание": 143696, + "ĠопÑĨион": 141131, + "ĠопÑĭ": 132389, + "ĠопÑĭÑĤ": 132390, + "ĠоÑĢг": 125186, + "ĠоÑĢган": 129651, + "ĠоÑĢганиз": 126030, + "ĠоÑĢганизаÑĨии": 133248, + "ĠоÑĢганизаÑĨиÑı": 142390, + "ĠоÑĢганизм": 134280, + "ĠоÑĢи": 140980, + "ĠоÑĢиг": 140981, + "ĠоÑĢигина": 140982, + "ĠоÑĢигинал": 140983, + "ĠоÑģ": 80276, + "ĠоÑģнов": 126068, + "ĠоÑģнован": 134067, + "ĠоÑģнове": 137854, + "ĠоÑģновном": 143498, + "ĠоÑģоб": 127071, + "ĠоÑģобенно": 131222, + "ĠоÑģÑĤ": 139856, + "ĠоÑģÑĤа": 129993, + "ĠоÑģÑĤалÑĮн": 139857, + "ĠоÑģÑĥÑīеÑģÑĤв": 134673, + "ĠоÑģÑĥÑīеÑģÑĤвлÑı": 134674, + "ĠоÑĤ": 20264, + "ĠоÑĤвеÑĤ": 92647, + "ĠоÑĤд": 135982, + "ĠоÑĤдел": 132650, + "ĠоÑĤдÑĭÑħ": 135983, + "ĠоÑĤз": 127789, + "ĠоÑĤзÑĭв": 127790, + "ĠоÑĤзÑĭвÑĭ": 135758, + "ĠоÑĤказ": 135738, + "ĠоÑĤкÑĢ": 142474, + "ĠоÑĤкÑĢÑĭва": 142475, + "ĠоÑĤкÑĢÑĭÑĤ": 131187, + "ĠоÑĤлиÑĩ": 127888, + "ĠоÑĤлиÑĩа": 136867, + "ĠоÑĤлиÑĩно": 141233, + "ĠоÑĤмеÑĤ": 128176, + "ĠоÑĤмеÑĤил": 142041, + "ĠоÑĤмеÑĤиÑĤÑĮ": 143758, + "ĠоÑĤно": 128001, + "ĠоÑĤноÑĪ": 128002, + "ĠоÑĤноÑĪении": 142723, + "ĠоÑĤноÑĪениÑı": 138623, + "ĠоÑĤпÑĢав": 86088, + "ĠоÑĦ": 139923, + "ĠоÑĦи": 133835, + "ĠоÑĦиÑģ": 139924, + "ĠоÑĦиÑĨиалÑĮн": 133836, + "ĠоÑĦоÑĢм": 137412, + "ĠоÑĦоÑĢмлен": 137413, + "ĠоÑĩ": 94812, + "ĠоÑĩенÑĮ": 128632, + "ĠоÑĩеÑĢ": 128015, + "ĠоÑĩеÑĢед": 140995, + "ĠоÑĩеÑĢедÑĮ": 138738, + "ĠоÑĪиб": 88663, + "Ġп": 5063, + "Ġпа": 129642, + "ĠпаÑĢ": 62680, + "ĠпаÑĢа": 133499, + "ĠпаÑĢамеÑĤ": 66345, + "ĠпаÑĢÑĤ": 140509, + "ĠпаÑĢÑĤи": 140216, + "ĠпаÑĢÑĤнеÑĢ": 140510, + "ĠпаÑĨи": 138299, + "ĠпаÑĨиенÑĤ": 138300, + "Ġпед": 142553, + "Ġпедаг": 142554, + "Ġпедагог": 142555, + "Ġпен": 139507, + "ĠпенÑģи": 139508, + "ĠпеÑĢ": 27016, + "ĠпеÑĢв": 76801, + "ĠпеÑĢвого": 142188, + "ĠпеÑĢвой": 141695, + "ĠпеÑĢвÑĥÑİ": 141517, + "ĠпеÑĢвÑĭй": 135756, + "ĠпеÑĢе": 126219, + "ĠпеÑĢев": 139054, + "ĠпеÑĢевод": 139055, + "ĠпеÑĢед": 77657, + "ĠпеÑĢем": 70801, + "ĠпеÑĢемен": 80099, + "ĠпеÑĢеÑħод": 140312, + "ĠпеÑĢи": 131243, + "ĠпеÑĢиод": 131244, + "ĠпеÑģ": 132415, + "ĠпеÑĩ": 137158, + "Ġпи": 129856, + "ĠпиÑĤ": 141684, + "ĠпиÑĤаниÑı": 141685, + "Ġпл": 125717, + "Ġплан": 129740, + "ĠплаÑģÑĤ": 136356, + "ĠплаÑĤ": 127026, + "ĠплаÑĤеж": 141706, + "ĠплиÑĤ": 143057, + "Ġпло": 127265, + "ĠплоÑĤ": 142284, + "ĠплоÑħо": 141466, + "ĠплоÑī": 132227, + "ĠплоÑīад": 132228, + "Ġпо": 17686, + "Ġпоб": 138021, + "Ġпобед": 138022, + "Ġпов": 94934, + "ĠповеÑĢ": 136225, + "ĠповеÑĢÑħ": 136226, + "ĠповеÑĢÑħноÑģÑĤ": 136227, + "ĠповÑĤоÑĢ": 139716, + "ĠповÑĭ": 133825, + "ĠповÑĭÑĪен": 133826, + "Ġпод": 39742, + "ĠподаÑĢ": 135049, + "ĠподгоÑĤов": 137846, + "ĠподгоÑĤовк": 137847, + "ĠподдеÑĢжива": 141403, + "ĠподпиÑģ": 139585, + "ĠподÑĢоб": 137641, + "ĠподÑĢобн": 137642, + "ĠподÑħод": 133903, + "Ġпозв": 126060, + "Ġпозвол": 142224, + "ĠпозволиÑĤ": 142225, + "ĠпозволÑı": 127556, + "ĠпозволÑıеÑĤ": 133973, + "ĠпоиÑģк": 133570, + "Ġпок": 134227, + "Ġпока": 131636, + "Ġпоказ": 131121, + "ĠпокÑĥп": 134228, + "ĠпокÑĥпа": 134229, + "Ġпол": 28519, + "Ġполи": 133261, + "ĠполиÑĤ": 137745, + "ĠполиÑĤик": 137746, + "Ġполно": 135870, + "ĠполноÑģÑĤÑĮÑİ": 134825, + "Ġполов": 141557, + "Ġполовин": 141558, + "Ġполож": 142189, + "ĠположиÑĤелÑĮн": 142190, + "ĠполÑĥ": 136739, + "ĠполÑĥÑĩ": 51332, + "ĠполÑĥÑĩа": 130783, + "ĠполÑĥÑĩен": 135435, + "ĠполÑĥÑĩениÑı": 140368, + "ĠполÑĥÑĩи": 133652, + "ĠполÑĥÑĩил": 135164, + "ĠполÑĥÑĩиÑĤÑĮ": 132693, + "ĠполÑĮз": 127417, + "ĠполÑĮзоваÑĤ": 45295, + "ĠполÑĮзоваÑĤелÑı": 66149, + "ĠполÑĮзÑĥ": 139032, + "ĠполÑı": 88202, + "Ġпом": 75301, + "ĠпомеÑī": 143299, + "ĠпомеÑīениÑı": 143300, + "Ġпомог": 132050, + "ĠпоможеÑĤ": 141075, + "ĠпомоÑī": 125978, + "ĠпомоÑīи": 135993, + "ĠпомоÑīÑĮ": 140130, + "ĠпомоÑīÑĮÑİ": 132409, + "Ġпон": 141688, + "ĠпонÑĢав": 141689, + "ĠпонÑĢави": 141690, + "Ġпоп": 133049, + "ĠпопÑĥ": 133050, + "ĠпопÑĥлÑıÑĢ": 133051, + "ĠпоÑĢ": 90046, + "ĠпоÑĢÑĤал": 142606, + "ĠпоÑĢÑıд": 142265, + "ĠпоÑĢÑıдке": 142266, + "ĠпоÑģк": 138727, + "ĠпоÑģколÑĮ": 138728, + "ĠпоÑģколÑĮкÑĥ": 138729, + "ĠпоÑģл": 60713, + "ĠпоÑģле": 128693, + "ĠпоÑģлед": 90567, + "ĠпоÑģледние": 142396, + "ĠпоÑģÑĤоÑı": 127768, + "ĠпоÑģÑĤоÑıн": 127769, + "ĠпоÑģÑĤоÑıнн": 143545, + "ĠпоÑģÑĤоÑıнно": 135605, + "ĠпоÑĤ": 80703, + "ĠпоÑĤеÑĢ": 142523, + "ĠпоÑĤеÑĢÑı": 142524, + "ĠпоÑĤом": 133931, + "ĠпоÑĤомÑĥ": 132903, + "ĠпоÑĤÑĢеб": 142516, + "ĠпоÑĤÑĢебиÑĤел": 142517, + "ĠпоÑĩ": 135003, + "ĠпоÑĩем": 137323, + "ĠпоÑĩемÑĥ": 137324, + "ĠпоÑĩÑĤи": 135004, + "ĠпоÑįÑĤомÑĥ": 131603, + "ĠпоÑı": 134484, + "ĠпоÑıви": 134485, + "ĠпÑĢ": 12281, + "ĠпÑĢав": 72126, + "ĠпÑĢава": 136468, + "ĠпÑĢавила": 138700, + "ĠпÑĢавило": 139577, + "ĠпÑĢавилÑĮно": 135544, + "ĠпÑĢавиÑĤелÑĮ": 140727, + "ĠпÑĢавиÑĤелÑĮÑģÑĤв": 140728, + "ĠпÑĢаво": 132092, + "ĠпÑĢаз": 137963, + "ĠпÑĢазд": 137964, + "ĠпÑĢаздник": 137965, + "ĠпÑĢакÑĤи": 135083, + "ĠпÑĢакÑĤиÑĩ": 135084, + "ĠпÑĢакÑĤиÑĩеÑģ": 135085, + "ĠпÑĢакÑĤиÑĩеÑģки": 135086, + "ĠпÑĢе": 126384, + "ĠпÑĢед": 56825, + "ĠпÑĢедлаг": 138900, + "ĠпÑĢедлагаеÑĤ": 138901, + "ĠпÑĢедложениÑı": 142927, + "ĠпÑĢедмеÑĤ": 136705, + "ĠпÑĢедн": 141259, + "ĠпÑĢедназ": 141260, + "ĠпÑĢедназнаÑĩен": 141261, + "ĠпÑĢедоÑģÑĤ": 127987, + "ĠпÑĢедоÑģÑĤав": 127988, + "ĠпÑĢедоÑģÑĤавлен": 140796, + "ĠпÑĢедоÑģÑĤавлÑı": 138499, + "ĠпÑĢедпоÑĩ": 143620, + "ĠпÑĢедпÑĢи": 139827, + "ĠпÑĢедпÑĢиÑıÑĤиÑı": 139828, + "ĠпÑĢедÑģÑĤав": 125975, + "ĠпÑĢедÑģÑĤавлен": 132363, + "ĠпÑĢедÑģÑĤавлÑı": 142364, + "ĠпÑĢедÑģÑĤавлÑıеÑĤ": 141107, + "ĠпÑĢедÑĥÑģ": 142424, + "ĠпÑĢедÑĥÑģмоÑĤÑĢ": 142425, + "ĠпÑĢедÑĥÑģмоÑĤÑĢен": 142426, + "ĠпÑĢеж": 142388, + "ĠпÑĢежде": 142389, + "ĠпÑĢез": 137703, + "ĠпÑĢезиденÑĤ": 137704, + "ĠпÑĢеим": 139338, + "ĠпÑĢеимÑĥÑīеÑģÑĤв": 139339, + "ĠпÑĢек": 137061, + "ĠпÑĢекÑĢаÑģ": 137062, + "ĠпÑĢекÑĢаÑģн": 137063, + "ĠпÑĢеп": 133693, + "ĠпÑĢепаÑĢаÑĤ": 133694, + "ĠпÑĢи": 36305, + "ĠпÑĢид": 142895, + "ĠпÑĢидеÑĤÑģÑı": 142896, + "ĠпÑĢием": 134011, + "ĠпÑĢизна": 132545, + "ĠпÑĢим": 127670, + "ĠпÑĢимен": 127671, + "ĠпÑĢименениÑı": 143175, + "ĠпÑĢименÑı": 134792, + "ĠпÑĢимеÑĢ": 131616, + "ĠпÑĢин": 126017, + "ĠпÑĢиним": 133026, + "ĠпÑĢинима": 133027, + "ĠпÑĢинÑĨип": 136396, + "ĠпÑĢинÑı": 141365, + "ĠпÑĢинÑıл": 141366, + "ĠпÑĢиÑĩ": 131207, + "ĠпÑĢиÑĩин": 131208, + "ĠпÑĢиÑıÑĤ": 138513, + "ĠпÑĢиÑıÑĤн": 138514, + "ĠпÑĢо": 40191, + "ĠпÑĢоб": 126963, + "ĠпÑĢоблем": 126964, + "ĠпÑĢоблемÑĭ": 136696, + "ĠпÑĢов": 61215, + "ĠпÑĢовед": 127950, + "ĠпÑĢоведен": 137831, + "ĠпÑĢоведениÑı": 139867, + "ĠпÑĢовеÑĢ": 68456, + "ĠпÑĢог": 141079, + "ĠпÑĢогн": 141080, + "ĠпÑĢогноз": 141081, + "ĠпÑĢогÑĢам": 80362, + "ĠпÑĢогÑĢамм": 93408, + "ĠпÑĢогÑĢаммÑĭ": 136059, + "ĠпÑĢод": 88615, + "ĠпÑĢодол": 139935, + "ĠпÑĢодолж": 136702, + "ĠпÑĢодолжа": 136703, + "ĠпÑĢодÑĥк": 128086, + "ĠпÑĢодÑĥкÑĤов": 143578, + "ĠпÑĢодÑĥкÑĨии": 139814, + "ĠпÑĢоекÑĤ": 129708, + "ĠпÑĢож": 141933, + "ĠпÑĢожива": 141934, + "ĠпÑĢоиз": 126034, + "ĠпÑĢоизведен": 142292, + "ĠпÑĢоизвод": 137852, + "ĠпÑĢоизводÑģÑĤва": 137853, + "ĠпÑĢоиÑģ": 136829, + "ĠпÑĢоиÑģÑħодиÑĤ": 136830, + "ĠпÑĢоÑģ": 140668, + "ĠпÑĢоÑģмоÑĤÑĢ": 140669, + "ĠпÑĢоÑģÑĤ": 129020, + "ĠпÑĢоÑģÑĤо": 129021, + "ĠпÑĢоÑĤ": 127351, + "ĠпÑĢоÑĤив": 127352, + "ĠпÑĢоÑĤиво": 135095, + "ĠпÑĢоÑĦеÑģÑģиона": 134391, + "ĠпÑĢоÑĦеÑģÑģионалÑĮн": 134392, + "ĠпÑĢоÑĨ": 125697, + "ĠпÑĢоÑĨед": 136920, + "ĠпÑĢоÑĨедÑĥÑĢ": 136921, + "ĠпÑĢоÑĨенÑĤ": 138083, + "ĠпÑĢоÑĨеÑģÑģ": 129644, + "ĠпÑĢоÑĩ": 133140, + "ĠпÑĢÑı": 135150, + "ĠпÑĢÑıмо": 135151, + "ĠпÑģ": 140581, + "ĠпÑģиÑħ": 140582, + "ĠпÑģиÑħолог": 140583, + "ĠпÑĥ": 141385, + "ĠпÑĥнк": 136095, + "ĠпÑĥнкÑĤ": 136096, + "ĠпÑĥÑĤ": 141859, + "ĠпÑĥÑĤеÑĪ": 141860, + "ĠпÑĥÑĤеÑĪеÑģÑĤв": 141861, + "ĠпÑĥÑĤи": 141386, + "ĠпÑĭÑĤа": 138725, + "ĠпÑıÑĤ": 132000, + "ĠÐIJ": 56755, + "ĠÐIJв": 138307, + "ĠÐIJвÑĤоÑĢ": 138308, + "ĠÐIJл": 134653, + "ĠÐIJлекÑģ": 133865, + "ĠÐIJлекÑģан": 133866, + "ĠÐIJлекÑģандÑĢ": 133867, + "ĠÐIJн": 127128, + "ĠÐIJна": 138937, + "ĠÐIJндÑĢ": 137882, + "ĠÐIJÑĢ": 131083, + "ĠÐij": 63471, + "ĠÐijа": 132223, + "ĠÐijез": 139335, + "ĠÐijеÑģ": 141180, + "ĠÐijо": 131419, + "ĠÐijог": 140506, + "ĠÐijолÑĮ": 136500, + "ĠÐijолÑĮÑĪ": 136501, + "ĠÐijÑĢ": 138930, + "ĠÐijÑĥд": 141920, + "ĠÐĴ": 22933, + "ĠÐĴам": 133535, + "ĠÐĴаÑģ": 133371, + "ĠÐĴаÑĪ": 132165, + "ĠÐĴедÑĮ": 141922, + "ĠÐĴеÑĢ": 137604, + "ĠÐĴлад": 136182, + "ĠÐĴладим": 136183, + "ĠÐĴладимиÑĢ": 136184, + "ĠÐĴо": 129307, + "ĠÐĴоз": 128115, + "ĠÐĴозможно": 143711, + "ĠÐĴоÑĤ": 134948, + "ĠÐĴÑģе": 129117, + "ĠÐĴÑĭ": 71631, + "ĠÐĵ": 60465, + "ĠÐĵлав": 136201, + "ĠÐĵоÑĢ": 139776, + "ĠÐĵÑĢ": 138908, + "ĠÐĶ": 38803, + "ĠÐĶа": 129279, + "ĠÐĶан": 138999, + "ĠÐĶе": 129762, + "ĠÐĶж": 132657, + "ĠÐĶи": 133441, + "ĠÐĶлÑı": 129002, + "ĠÐĶм": 142577, + "ĠÐĶмиÑĤ": 142578, + "ĠÐĶмиÑĤÑĢ": 142579, + "ĠÐĶо": 128734, + "ĠÐķ": 59276, + "ĠÐķго": 138328, + "ĠÐķÑģли": 71144, + "ĠÐķÑģÑĤÑĮ": 139065, + "ĠÐķÑīе": 140725, + "ĠÐĸ": 129219, + "ĠÐĹ": 52577, + "ĠÐĹа": 128460, + "ĠÐĹдеÑģÑĮ": 138505, + "ĠÐĹна": 140226, + "ĠÐĺ": 42796, + "ĠÐĺгÑĢ": 133978, + "ĠÐĺз": 129977, + "ĠÐĺменно": 141552, + "ĠÐĺн": 126895, + "ĠÐĺнÑĤеÑĢ": 138586, + "ĠÐĺнÑĤеÑĢнеÑĤ": 138587, + "ĠÐĺÑģ": 133158, + "ĠÐļ": 35379, + "ĠÐļа": 128889, + "ĠÐļак": 128654, + "ĠÐļаÑĢ": 133387, + "ĠÐļи": 127465, + "ĠÐļиев": 138868, + "ĠÐļо": 129911, + "ĠÐļогда": 134937, + "ĠÐļол": 137686, + "ĠÐļом": 133412, + "ĠÐļон": 127443, + "ĠÐļонеÑĩно": 143194, + "ĠÐļоÑĢ": 134610, + "ĠÐļÑĢоме": 134429, + "ĠÐļÑĥ": 132100, + "ĠÐĽ": 124805, + "ĠÐĽÐ°": 135427, + "ĠÐĽÐµ": 131884, + "ĠÐĽÐ¸": 131895, + "ĠÐĽÑİ": 138817, + "ĠÐĽÑİб": 138818, + "ĠÐľ": 44358, + "ĠÐľÐ°": 129047, + "ĠÐľÐ°Ðº": 138047, + "ĠÐľÐ°ÑĢи": 141058, + "ĠÐľÐµ": 129633, + "ĠÐľÐµÐ¶Ð´Ñĥ": 142160, + "ĠÐľÐ¸": 132511, + "ĠÐľÐ¸Ð½": 134279, + "ĠÐľÐ½Ðµ": 140704, + "ĠÐľÐ¾": 129494, + "ĠÐľÐ¾Ð¶ÐµÑĤ": 143685, + "ĠÐľÐ¾Ð¶Ð½Ð¾": 139490, + "ĠÐľÐ¾Ñģ": 125386, + "ĠÐľÐ¾Ñģк": 126082, + "ĠÐľÐ¾Ñģква": 137639, + "ĠÐľÐ¾Ñģкве": 134464, + "ĠÐľÐ¾ÑģквÑĭ": 141594, + "ĠÐľÐ¾Ñģков": 141897, + "ĠÐľÐ¾ÑģковÑģк": 141898, + "ĠÐľÑĭ": 130275, + "ĠÐĿ": 34789, + "ĠÐĿа": 128313, + "ĠÐĿапÑĢимеÑĢ": 140085, + "ĠÐĿе": 128544, + "ĠÐĿи": 130952, + "ĠÐĿо": 128543, + "ĠÐĿово": 130922, + "ĠÐĿÑĥ": 132965, + "ĠÐŀ": 34348, + "ĠÐŀб": 79698, + "ĠÐŀд": 132593, + "ĠÐŀднако": 132594, + "ĠÐŀн": 129917, + "ĠÐŀна": 135432, + "ĠÐŀни": 133482, + "ĠÐŀÐŀ": 139283, + "ĠÐŀÐŀÐŀ": 139284, + "ĠÐŀÑģоб": 143345, + "ĠÐŀÑģобенно": 143346, + "ĠÐŀÑĤ": 128876, + "ĠÐŀÑĩ": 140152, + "ĠÐŀÑĩенÑĮ": 140153, + "ĠÐŁ": 22419, + "ĠÐŁÐ°": 133059, + "ĠÐŁÐ°ÑĢ": 137353, + "ĠÐŁÐµÑĢ": 135067, + "ĠÐŁÐµÑĢв": 135068, + "ĠÐŁÐµÑĢе": 132578, + "ĠÐŁÐµÑĤ": 140710, + "ĠÐŁÐµÑĤÑĢ": 140711, + "ĠÐŁÐ¾": 126206, + "ĠÐŁÐ¾Ð´": 130041, + "ĠÐŁÐ¾Ð»": 134922, + "ĠÐŁÐ¾Ð»ÑĥÑĩ": 88933, + "ĠÐŁÐ¾Ñģ": 131374, + "ĠÐŁÐ¾Ñģле": 131375, + "ĠÐŁÐ¾Ñĩ": 143171, + "ĠÐŁÐ¾Ñĩем": 143172, + "ĠÐŁÐ¾ÑĩемÑĥ": 143173, + "ĠÐŁÐ¾ÑįÑĤомÑĥ": 133841, + "ĠÐŁÑĢ": 79484, + "ĠÐŁÑĢе": 134004, + "ĠÐŁÑĢед": 132419, + "ĠÐŁÑĢи": 128462, + "ĠÐŁÑĢо": 128598, + "ĠÐł": 48420, + "ĠÐłÐ¤": 131177, + "ĠÐłÐ°": 130114, + "ĠÐłÐ°Ð·": 127088, + "ĠÐłÐ°Ð·Ð¼ÐµÑĢ": 143040, + "ĠÐłÐ°Ñģ": 133199, + "ĠÐłÐµ": 129170, + "ĠÐłÐµÑģп": 142155, + "ĠÐłÐµÑģпÑĥб": 142156, + "ĠÐłÐµÑģпÑĥблик": 142157, + "ĠÐłÐ¾": 130026, + "ĠÐłÐ¾Ñģ": 125598, + "ĠÐłÐ¾ÑģÑģ": 125599, + "ĠÐłÐ¾ÑģÑģии": 129069, + "ĠÐłÐ¾ÑģÑģий": 136069, + "ĠÐłÐ¾ÑģÑģийÑģкой": 136070, + "ĠÐłÐ¾ÑģÑģиÑı": 137421, + "ĠÐłÑĥÑģ": 140869, + "ĠÐŃ": 80429, + "ĠÐŃÑĤ": 139470, + "ĠÐŃÑĤи": 141726, + "ĠÐŃÑĤо": 128868, + "ĠÐŃÑĤоÑĤ": 139471, + "ĠÑ": 8839, + "ĠÑĢ": 18108, + "ĠÑĢа": 142014, + "ĠÑĢаÐ": 24212, + "ĠÑĢабоÑĤ": 53104, + "ĠÑĢабоÑĤа": 129897, + "ĠÑĢабоÑĤаеÑĤ": 136533, + "ĠÑĢабоÑĤе": 135088, + "ĠÑĢабоÑĤник": 137271, + "ĠÑĢабоÑĤÑĥ": 133400, + "ĠÑĢабоÑĤÑĭ": 94867, + "ĠÑĢав": 135933, + "ĠÑĢавно": 135934, + "ĠÑĢад": 134296, + "ĠÑĢади": 134297, + "ĠÑĢаз": 38379, + "ĠÑĢазв": 139098, + "ĠÑĢазвива": 139099, + "ĠÑĢазвиÑĤ": 127416, + "ĠÑĢазвиÑĤие": 139853, + "ĠÑĢазвиÑĤиÑı": 133079, + "ĠÑĢаздел": 132528, + "ĠÑĢазлиÑĩ": 127810, + "ĠÑĢазлиÑĩнÑĭе": 142491, + "ĠÑĢазлиÑĩнÑĭÑħ": 136004, + "ĠÑĢазмеÑĢ": 99047, + "ĠÑĢазмеÑīен": 141011, + "ĠÑĢазнÑĭÑħ": 137396, + "ĠÑĢай": 127233, + "ĠÑĢайон": 132025, + "ĠÑĢайона": 138452, + "ĠÑĢам": 138722, + "ĠÑĢамкаÑħ": 138723, + "ĠÑĢан": 128177, + "ĠÑĢанее": 142113, + "ĠÑĢанÑĮ": 142816, + "ĠÑĢанÑĮÑĪе": 142817, + "ĠÑĢаÑģ": 77776, + "ĠÑĢаÑģпол": 134425, + "ĠÑĢаÑģполож": 134426, + "ĠÑĢаÑģположен": 134427, + "ĠÑĢаÑģпÑĢ": 140972, + "ĠÑĢаÑģпÑĢоÑģÑĤ": 140973, + "ĠÑĢаÑģпÑĢоÑģÑĤÑĢан": 140974, + "ĠÑĢаÑģпÑĢоÑģÑĤÑĢанен": 140975, + "ĠÑĢаÑģÑģ": 125326, + "ĠÑĢаÑģÑģказ": 143631, + "ĠÑĢаÑģÑģказал": 141195, + "ĠÑĢаÑģÑģказÑĭва": 143632, + "ĠÑĢаÑģÑģмаÑĤ": 142145, + "ĠÑĢаÑģÑģмаÑĤÑĢ": 142146, + "ĠÑĢаÑģÑģмаÑĤÑĢива": 142147, + "ĠÑĢаÑģÑĤв": 142015, + "ĠÑĢаÑģÑĤвоÑĢ": 142016, + "ĠÑĢаÑģÑħод": 137540, + "ĠÑĢаÑģÑĩеÑĤ": 139973, + "ĠÑĢе": 124619, + "ĠÑĢеак": 143605, + "ĠÑĢеакÑĨи": 143606, + "ĠÑĢеализ": 139754, + "ĠÑĢеализаÑĨи": 139755, + "ĠÑĢеалÑĮно": 139395, + "ĠÑĢеб": 135821, + "ĠÑĢебен": 135822, + "ĠÑĢебенка": 135823, + "ĠÑĢег": 142405, + "ĠÑĢегиона": 142406, + "ĠÑĢегиÑģÑĤ": 135146, + "ĠÑĢегиÑģÑĤÑĢа": 135147, + "ĠÑĢегиÑģÑĤÑĢаÑĨии": 135148, + "ĠÑĢеж": 133887, + "ĠÑĢежим": 133888, + "ĠÑĢезÑĥлÑĮÑĤаÑĤ": 98200, + "ĠÑĢезÑĥлÑĮÑĤаÑĤе": 137580, + "ĠÑĢей": 143894, + "ĠÑĢейÑĤ": 143895, + "ĠÑĢейÑĤинг": 143896, + "ĠÑĢек": 133962, + "ĠÑĢекл": 133963, + "ĠÑĢеклам": 133964, + "ĠÑĢем": 132912, + "ĠÑĢемонÑĤ": 132913, + "ĠÑĢеÑģÑĤ": 142710, + "ĠÑĢеÑģÑĤоÑĢ": 142711, + "ĠÑĢеÑģÑĤоÑĢан": 142712, + "ĠÑĢеÑĨеп": 138852, + "ĠÑĢеÑĨепÑĤ": 138853, + "ĠÑĢеÑĪ": 126091, + "ĠÑĢеÑĪа": 141682, + "ĠÑĢеÑĪение": 134746, + "ĠÑĢеÑĪениÑı": 136080, + "ĠÑĢеÑĪи": 136991, + "ĠÑĢеÑĪил": 136992, + "ĠÑĢиÑģк": 136802, + "ĠÑĢод": 129757, + "ĠÑĢож": 142486, + "ĠÑĢожд": 142487, + "ĠÑĢождениÑı": 142488, + "ĠÑĢол": 143528, + "ĠÑĢоли": 143529, + "ĠÑĢолÑĮ": 141432, + "ĠÑĢом": 143724, + "ĠÑĢоман": 143725, + "ĠÑĢÑĥб": 126582, + "ĠÑĢÑĥблей": 131473, + "ĠÑĢÑĥк": 126722, + "ĠÑĢÑĥков": 131850, + "ĠÑĢÑĥковод": 131851, + "ĠÑĢÑĭ": 127633, + "ĠÑĢÑĭб": 134542, + "ĠÑĢÑĭн": 139929, + "ĠÑĢÑĭнке": 139930, + "ĠÑĢÑıд": 140283, + "ĠÑĢÑıдом": 140284, + "ĠÑģ": 5409, + "ĠÑģай": 82362, + "ĠÑģайÑĤ": 130644, + "ĠÑģайÑĤа": 134249, + "ĠÑģайÑĤе": 131595, + "ĠÑģам": 96085, + "ĠÑģами": 137830, + "ĠÑģамо": 129355, + "ĠÑģамого": 140205, + "ĠÑģамом": 136147, + "ĠÑģамоÑģÑĤоÑı": 139025, + "ĠÑģамоÑģÑĤоÑıÑĤелÑĮно": 139026, + "ĠÑģамÑĭе": 139149, + "ĠÑģамÑĭй": 139524, + "ĠÑģамÑĭм": 138766, + "ĠÑģамÑĭÑħ": 135041, + "ĠÑģаÑħ": 143708, + "ĠÑģаÑħаÑĢ": 143709, + "ĠÑģбоÑĢ": 134005, + "ĠÑģв": 44816, + "ĠÑģвеÑĤ": 130865, + "ĠÑģво": 124460, + "ĠÑģвоб": 134387, + "ĠÑģвобод": 134388, + "ĠÑģвое": 133375, + "ĠÑģвоего": 133520, + "ĠÑģвоей": 131189, + "ĠÑģвоем": 136548, + "ĠÑģвои": 131382, + "ĠÑģвоим": 136173, + "ĠÑģвоими": 137581, + "ĠÑģвоиÑħ": 132393, + "ĠÑģвой": 130210, + "ĠÑģвоÑİ": 131398, + "ĠÑģвÑıз": 126032, + "ĠÑģвÑıзан": 133287, + "ĠÑģвÑıзи": 134023, + "ĠÑģдел": 125871, + "ĠÑģдела": 134943, + "ĠÑģделал": 140896, + "ĠÑģделаÑĤÑĮ": 130965, + "ĠÑģеб": 126564, + "ĠÑģебе": 130688, + "ĠÑģебÑı": 129364, + "ĠÑģегоднÑı": 132110, + "ĠÑģез": 135070, + "ĠÑģезон": 135071, + "ĠÑģейÑĩаÑģ": 131405, + "ĠÑģек": 143542, + "ĠÑģекÑĢ": 143543, + "ĠÑģекÑĢеÑĤ": 143544, + "ĠÑģем": 127258, + "ĠÑģемей": 140054, + "ĠÑģемÑĮ": 132176, + "ĠÑģенÑĤ": 140185, + "ĠÑģенÑĤÑı": 140186, + "ĠÑģенÑĤÑıбÑĢÑı": 140187, + "ĠÑģеÑĢ": 65286, + "ĠÑģеÑĢвеÑĢ": 92764, + "ĠÑģеÑĢвиÑģ": 137317, + "ĠÑģеÑĢи": 143182, + "ĠÑģеÑĢиал": 143183, + "ĠÑģеÑĢии": 141827, + "ĠÑģеÑĢÑĮ": 139998, + "ĠÑģеÑĢÑĮезн": 139999, + "ĠÑģеÑĤ": 137528, + "ĠÑģеÑĤи": 137529, + "ĠÑģиг": 141204, + "ĠÑģигнал": 141205, + "ĠÑģил": 130386, + "ĠÑģилÑĮно": 137947, + "ĠÑģимв": 77912, + "ĠÑģимвол": 141802, + "ĠÑģиÑģÑĤем": 91999, + "ĠÑģиÑģÑĤема": 134450, + "ĠÑģиÑģÑĤемÑĭ": 131773, + "ĠÑģиÑĤ": 128050, + "ĠÑģиÑĤÑĥ": 128051, + "ĠÑģиÑĤÑĥаÑĨии": 139367, + "ĠÑģиÑĤÑĥаÑĨиÑı": 142985, + "ĠÑģк": 68371, + "ĠÑģка": 131682, + "ĠÑģказал": 133260, + "ĠÑģказаÑĤÑĮ": 136670, + "ĠÑģкаÑĩ": 131683, + "ĠÑģкаÑĩаÑĤÑĮ": 131684, + "ĠÑģкоÑĢ": 143688, + "ĠÑģкоÑĢо": 138366, + "ĠÑģкоÑĢоÑģÑĤÑĮ": 143689, + "ĠÑģл": 34775, + "ĠÑģлед": 92029, + "ĠÑģледÑĥеÑĤ": 132996, + "ĠÑģли": 140827, + "ĠÑģлиÑĪ": 140828, + "ĠÑģлиÑĪком": 140829, + "ĠÑģлов": 91107, + "ĠÑģлова": 135381, + "ĠÑģловам": 138469, + "ĠÑģложно": 136927, + "ĠÑģлÑĥж": 141975, + "ĠÑģлÑĥжбÑĭ": 141976, + "ĠÑģлÑĥÑĩ": 89126, + "ĠÑģлÑĥÑĩа": 129401, + "ĠÑģлÑĥÑĩае": 129402, + "ĠÑģлÑĥÑĩай": 140664, + "ĠÑģлÑĥÑĩаÑı": 141113, + "ĠÑģлÑĥÑĩаÑıÑħ": 141114, + "ĠÑģм": 126762, + "ĠÑģмог": 136406, + "ĠÑģмож": 138931, + "ĠÑģможеÑĤе": 138932, + "ĠÑģмÑĭÑģл": 138869, + "ĠÑģни": 141736, + "ĠÑģнижен": 141737, + "ĠÑģнов": 141029, + "ĠÑģнова": 141030, + "ĠÑģо": 33843, + "ĠÑģоб": 124972, + "ĠÑģобак": 141534, + "ĠÑģобиÑĢа": 138483, + "ĠÑģобой": 133104, + "ĠÑģобÑĢан": 142631, + "ĠÑģобÑģÑĤвенно": 133620, + "ĠÑģобÑĭÑĤи": 143921, + "ĠÑģобÑĭÑĤиÑı": 143922, + "ĠÑģов": 124955, + "ĠÑģовеÑĢ": 128079, + "ĠÑģовеÑĢÑĪ": 139730, + "ĠÑģовеÑĢÑĪен": 142855, + "ĠÑģовеÑĢÑĪенно": 139731, + "ĠÑģовеÑĤ": 131847, + "ĠÑģовÑĢем": 131662, + "ĠÑģовÑĢеменн": 131663, + "ĠÑģовÑģ": 136844, + "ĠÑģовÑģем": 136845, + "ĠÑģог": 132801, + "ĠÑģоглаÑģ": 132802, + "ĠÑģодеÑĢж": 79495, + "ĠÑģодеÑĢжа": 132636, + "ĠÑģодеÑĢжиÑĤ": 143132, + "ĠÑģозд": 51224, + "ĠÑģозда": 130222, + "ĠÑģоздан": 136295, + "ĠÑģозданиÑı": 141132, + "ĠÑģообÑī": 66021, + "ĠÑģообÑīаеÑĤ": 142894, + "ĠÑģообÑīил": 140102, + "ĠÑģооÑĤвеÑĤ": 136783, + "ĠÑģооÑĤвеÑĤÑģÑĤв": 136784, + "ĠÑģооÑĤвеÑĤÑģÑĤвии": 136785, + "ĠÑģоÑģ": 143410, + "ĠÑģоÑģед": 143411, + "ĠÑģоÑģÑĤ": 91075, + "ĠÑģоÑģÑĤав": 126737, + "ĠÑģоÑģÑĤавлÑıеÑĤ": 137851, + "ĠÑģоÑģÑĤо": 140271, + "ĠÑģоÑģÑĤоиÑĤ": 140272, + "ĠÑģоÑģÑĤоÑı": 128134, + "ĠÑģоÑģÑĤоÑıние": 140716, + "ĠÑģоÑģÑĤоÑıнии": 141894, + "ĠÑģоÑĤ": 134715, + "ĠÑģоÑĤÑĢÑĥд": 134716, + "ĠÑģоÑĤÑĢÑĥдник": 134717, + "ĠÑģоÑħ": 143362, + "ĠÑģоÑħÑĢ": 143363, + "ĠÑģоÑħÑĢан": 92681, + "ĠÑģоÑħÑĢани": 143364, + "ĠÑģоÑĨиалÑĮно": 137936, + "ĠÑģоÑĩеÑĤ": 142151, + "ĠÑģоÑĩеÑĤа": 142152, + "ĠÑģп": 83614, + "ĠÑģпе": 136429, + "ĠÑģпеÑĨ": 131548, + "ĠÑģпеÑĨиалÑĮно": 136430, + "ĠÑģпин": 142186, + "ĠÑģпиÑģ": 45630, + "ĠÑģпиÑģка": 97533, + "ĠÑģпиÑģок": 73332, + "ĠÑģпоÑĢÑĤ": 132454, + "ĠÑģпоÑģоб": 129187, + "ĠÑģпÑĢав": 134876, + "ĠÑģÑĢав": 134855, + "ĠÑģÑĢавн": 134856, + "ĠÑģÑĢаз": 133306, + "ĠÑģÑĢазÑĥ": 133307, + "ĠÑģÑĢед": 125265, + "ĠÑģÑĢеди": 134987, + "ĠÑģÑĢедне": 135141, + "ĠÑģÑĢедÑģÑĤв": 130589, + "ĠÑģÑĢедÑģÑĤва": 133198, + "ĠÑģÑĢок": 131561, + "ĠÑģÑģÑĭл": 94394, + "ĠÑģÑĤ": 18362, + "ĠÑģÑĤал": 139426, + "ĠÑģÑĤала": 140325, + "ĠÑģÑĤали": 135906, + "ĠÑģÑĤало": 139427, + "ĠÑģÑĤан": 128130, + "ĠÑģÑĤанд": 137254, + "ĠÑģÑĤандаÑĢÑĤ": 137255, + "ĠÑģÑĤанеÑĤ": 140735, + "ĠÑģÑĤанов": 140661, + "ĠÑģÑĤановиÑĤÑģÑı": 140662, + "ĠÑģÑĤаÑĢ": 129511, + "ĠÑģÑĤаÑĤÑĮ": 129934, + "ĠÑģÑĤен": 134253, + "ĠÑģÑĤеп": 143589, + "ĠÑģÑĤепени": 143590, + "ĠÑģÑĤилÑĮ": 143813, + "ĠÑģÑĤо": 125868, + "ĠÑģÑĤоим": 127854, + "ĠÑģÑĤоимоÑģÑĤи": 143608, + "ĠÑģÑĤоимоÑģÑĤÑĮ": 136467, + "ĠÑģÑĤоиÑĤ": 130960, + "ĠÑģÑĤол": 130297, + "ĠÑģÑĤоÑĢ": 132519, + "ĠÑģÑĤоÑĢо": 133418, + "ĠÑģÑĤоÑĢон": 132520, + "ĠÑģÑĤоÑĢонÑĭ": 133419, + "ĠÑģÑĤÑĢ": 29292, + "ĠÑģÑĤÑĢа": 138032, + "ĠÑģÑĤÑĢан": 126540, + "ĠÑģÑĤÑĢаниÑĨ": 61480, + "ĠÑģÑĤÑĢанÑĭ": 135096, + "ĠÑģÑĤÑĢаÑħ": 138033, + "ĠÑģÑĤÑĢаÑħов": 138034, + "ĠÑģÑĤÑĢем": 140867, + "ĠÑģÑĤÑĢок": 43922, + "ĠÑģÑĤÑĢоки": 92714, + "ĠÑģÑĤÑĥд": 140830, + "ĠÑģÑĤÑĥденÑĤ": 140831, + "ĠÑģÑĥ": 128975, + "ĠÑģÑĥд": 126869, + "ĠÑģÑĥдеб": 142594, + "ĠÑģÑĥдебн": 142595, + "ĠÑģÑĥм": 143433, + "ĠÑģÑĥммÑĥ": 143434, + "ĠÑģÑĥÑīеÑģÑĤв": 127339, + "ĠÑģÑĥÑīеÑģÑĤвÑĥеÑĤ": 142738, + "ĠÑģÑĦ": 142728, + "ĠÑģÑĦеÑĢе": 142729, + "ĠÑģÑĩеÑĤ": 131206, + "ĠÑģÑĩиÑĤ": 130480, + "ĠÑģÑĩиÑĤа": 130481, + "ĠÑģÑĭ": 128167, + "ĠÑģÑĭн": 141723, + "ĠÑģÑĭÑĢ": 142195, + "ĠÑĤ": 10813, + "ĠÑĤаблиÑĨ": 78227, + "ĠÑĤак": 49845, + "ĠÑĤакже": 128547, + "ĠÑĤакие": 132842, + "ĠÑĤаким": 135536, + "ĠÑĤакиÑħ": 133360, + "ĠÑĤакого": 137566, + "ĠÑĤакое": 134322, + "ĠÑĤакой": 131392, + "ĠÑĤам": 130769, + "ĠÑĤе": 126515, + "ĠÑĤеаÑĤ": 142633, + "ĠÑĤеаÑĤÑĢ": 142634, + "ĠÑĤеб": 128071, + "ĠÑĤебе": 141763, + "ĠÑĤебÑı": 139596, + "ĠÑĤек": 49090, + "ĠÑĤекÑģÑĤ": 70895, + "ĠÑĤекÑĥÑī": 85146, + "ĠÑĤеле": 136148, + "ĠÑĤелеÑĦ": 131019, + "ĠÑĤелеÑĦон": 131020, + "ĠÑĤем": 128680, + "ĠÑĤеп": 127623, + "ĠÑĤепеÑĢÑĮ": 134475, + "ĠÑĤепло": 136282, + "ĠÑĤеÑĢ": 135010, + "ĠÑĤеÑĢÑĢиÑĤоÑĢ": 135011, + "ĠÑĤеÑĢÑĢиÑĤоÑĢии": 135012, + "ĠÑĤеÑģÑĤ": 137901, + "ĠÑĤеÑħ": 126468, + "ĠÑĤеÑħнолог": 142650, + "ĠÑĤеÑħнологии": 142651, + "ĠÑĤеÑĩение": 133945, + "ĠÑĤип": 63833, + "ĠÑĤипа": 137796, + "ĠÑĤо": 58246, + "ĠÑĤоваÑĢ": 73712, + "ĠÑĤоваÑĢов": 140659, + "ĠÑĤогда": 134564, + "ĠÑĤого": 128798, + "ĠÑĤож": 131558, + "ĠÑĤоже": 131559, + "ĠÑĤолÑĮко": 73626, + "ĠÑĤом": 128856, + "ĠÑĤомÑĥ": 136392, + "ĠÑĤоÑĢ": 127367, + "ĠÑĤоÑĢг": 132834, + "ĠÑĤоÑĢгов": 132835, + "ĠÑĤоÑĢÑĢ": 137877, + "ĠÑĤоÑĢÑĢенÑĤ": 137878, + "ĠÑĤоÑĤ": 134316, + "ĠÑĤоÑĩ": 81602, + "ĠÑĤоÑĩки": 142996, + "ĠÑĤоÑĩно": 136268, + "ĠÑĤÑĢ": 85895, + "ĠÑĤÑĢав": 137486, + "ĠÑĤÑĢе": 131125, + "ĠÑĤÑĢеб": 126155, + "ĠÑĤÑĢебова": 139907, + "ĠÑĤÑĢебованиÑı": 138412, + "ĠÑĤÑĢебÑĥеÑĤÑģÑı": 143157, + "ĠÑĤÑĢеÑĤÑĮ": 140528, + "ĠÑĤÑĢеÑħ": 138733, + "ĠÑĤÑĢи": 131703, + "ĠÑĤÑĢÑĥб": 138379, + "ĠÑĤÑĢÑĥд": 129771, + "ĠÑĤÑĥ": 141634, + "ĠÑĤÑĥÑĢиÑģÑĤ": 141635, + "ĠÑĤÑĭ": 127429, + "ĠÑĤÑĭÑģ": 138994, + "ĠÑĤÑĭÑģÑı": 133155, + "ĠÑĤÑĭÑģÑıÑĩ": 133156, + "ĠÑĥ": 13932, + "ĠÑĥвели": 136325, + "ĠÑĥвелиÑĩи": 136326, + "ĠÑĥвеÑĢ": 137892, + "ĠÑĥвеÑĢен": 137893, + "ĠÑĥг": 143263, + "ĠÑĥгол": 143264, + "ĠÑĥд": 127189, + "ĠÑĥдал": 57391, + "ĠÑĥдален": 142499, + "ĠÑĥдалоÑģÑĮ": 143954, + "ĠÑĥдаÑĢ": 141344, + "ĠÑĥдоб": 131810, + "ĠÑĥже": 90777, + "ĠÑĥзн": 140802, + "ĠÑĥзна": 140380, + "ĠÑĥзнаÑĤÑĮ": 140803, + "ĠÑĥказ": 67114, + "ĠÑĥказан": 134321, + "ĠÑĥкÑĢа": 138131, + "ĠÑĥкÑĢаин": 138132, + "ĠÑĥме": 137504, + "ĠÑĥни": 142644, + "ĠÑĥнивеÑĢ": 142645, + "ĠÑĥнивеÑĢÑģ": 142646, + "ĠÑĥнивеÑĢÑģиÑĤеÑĤ": 142647, + "ĠÑĥника": 140511, + "ĠÑĥникалÑĮн": 140512, + "ĠÑĥпÑĢав": 135860, + "ĠÑĥпÑĢавлениÑı": 135861, + "ĠÑĥÑĢов": 126146, + "ĠÑĥÑĢовенÑĮ": 137779, + "ĠÑĥÑĢовне": 137485, + "ĠÑĥÑĢовнÑı": 138646, + "ĠÑĥÑģи": 139864, + "ĠÑĥÑģили": 139865, + "ĠÑĥÑģл": 96891, + "ĠÑĥÑģлов": 127757, + "ĠÑĥÑģловиÑı": 127758, + "ĠÑĥÑģловиÑıÑħ": 135447, + "ĠÑĥÑģлÑĥг": 127387, + "ĠÑĥÑģлÑĥги": 136911, + "ĠÑĥÑģп": 81533, + "ĠÑĥÑģпеÑħ": 141656, + "ĠÑĥÑģпеÑĪ": 90513, + "ĠÑĥÑģпеÑĪно": 142982, + "ĠÑĥÑģÑĤанов": 133837, + "ĠÑĥÑģÑĤановлен": 133838, + "ĠÑĥÑģÑĤÑĢой": 142285, + "ĠÑĥÑģÑĤÑĢойÑģÑĤва": 142286, + "ĠÑĥÑĩ": 134889, + "ĠÑĥÑĩаÑģÑĤ": 127609, + "ĠÑĥÑĩаÑģÑĤие": 135119, + "ĠÑĥÑĩаÑģÑĤник": 134389, + "ĠÑĥÑĩеб": 134890, + "ĠÑĥÑĩеÑĤ": 134931, + "ĠÑĥÑĩиÑĤÑĭва": 141816, + "ĠÑĦ": 17574, + "ĠÑĦа": 131494, + "ĠÑĦай": 35256, + "ĠÑĦайл": 46318, + "ĠÑĦайла": 70773, + "ĠÑĦак": 127481, + "ĠÑĦакÑĤ": 133480, + "ĠÑĦакÑĤоÑĢ": 138786, + "ĠÑĦев": 140718, + "ĠÑĦевÑĢа": 140719, + "ĠÑĦевÑĢалÑı": 140720, + "ĠÑĦедеÑĢ": 142868, + "ĠÑĦедеÑĢалÑĮн": 142869, + "ĠÑĦиг": 143372, + "ĠÑĦигÑĥÑĢ": 143373, + "ĠÑĦилÑĮ": 131158, + "ĠÑĦилÑĮм": 131159, + "ĠÑĦин": 135156, + "ĠÑĦинанÑģ": 135157, + "ĠÑĦинанÑģов": 135158, + "ĠÑĦиÑĢ": 138634, + "ĠÑĦиÑĢм": 138635, + "ĠÑĦл": 141316, + "ĠÑĦон": 140406, + "ĠÑĦонд": 140407, + "ĠÑĦоÑĢ": 140167, + "ĠÑĦоÑĢм": 53161, + "ĠÑĦоÑĢма": 135005, + "ĠÑĦоÑĢме": 141342, + "ĠÑĦоÑĢмиÑĢов": 140955, + "ĠÑĦоÑĢмиÑĢова": 140956, + "ĠÑĦоÑĢмÑĥ": 139677, + "ĠÑĦоÑĢмÑĭ": 141092, + "ĠÑĦоÑĢÑĥм": 140168, + "ĠÑĦоÑĤ": 127020, + "ĠÑĦоÑĤо": 130942, + "ĠÑĦоÑĤогÑĢаÑĦ": 137243, + "ĠÑĦÑĥнкÑĨи": 54713, + "ĠÑĦÑĥÑĤ": 138370, + "ĠÑĦÑĥÑĤбол": 138371, + "ĠÑħ": 44558, + "ĠÑħа": 127504, + "ĠÑħаÑĢак": 127505, + "ĠÑħаÑĢакÑĤеÑĢ": 127506, + "ĠÑħаÑĢакÑĤеÑĢиÑģÑĤ": 137677, + "ĠÑħаÑĢакÑĤеÑĢиÑģÑĤик": 137678, + "ĠÑħоз": 140219, + "ĠÑħозÑı": 140220, + "ĠÑħозÑıй": 140221, + "ĠÑħозÑıйÑģÑĤв": 140222, + "ĠÑħолод": 140625, + "ĠÑħоÑĢо": 126824, + "ĠÑħоÑĢоÑĪ": 126825, + "ĠÑħоÑĢоÑĪо": 133097, + "ĠÑħоÑĤ": 125351, + "ĠÑħоÑĤел": 136730, + "ĠÑħоÑĤиÑĤе": 142942, + "ĠÑħоÑĤÑĮ": 142929, + "ĠÑħоÑĤÑı": 132957, + "ĠÑħоÑĩ": 127893, + "ĠÑħоÑĩеÑĤ": 136953, + "ĠÑħоÑĩÑĥ": 143612, + "ĠÑħÑĢ": 133708, + "ĠÑĨ": 38133, + "ĠÑĨвеÑĤ": 129596, + "ĠÑĨел": 129151, + "ĠÑĨелÑĮ": 142963, + "ĠÑĨелÑĮÑİ": 142964, + "ĠÑĨен": 129827, + "ĠÑĨена": 133599, + "ĠÑĨенÑĭ": 134972, + "ĠÑĨиÑĦ": 141615, + "ĠÑĨиÑĦÑĢ": 141616, + "ĠÑĩ": 17307, + "ĠÑĩа": 125948, + "ĠÑĩаÑģ": 127770, + "ĠÑĩаÑģа": 141031, + "ĠÑĩаÑģов": 135612, + "ĠÑĩаÑģÑĤ": 127174, + "ĠÑĩаÑģÑĤи": 132153, + "ĠÑĩаÑģÑĤноÑģÑĤи": 143311, + "ĠÑĩаÑģÑĤо": 131704, + "ĠÑĩаÑģÑĤÑĮ": 131964, + "ĠÑĩаÑīе": 141648, + "ĠÑĩего": 133197, + "ĠÑĩел": 126633, + "ĠÑĩелов": 126634, + "ĠÑĩеловек": 126635, + "ĠÑĩеловека": 130914, + "ĠÑĩем": 128743, + "ĠÑĩеÑĢ": 89111, + "ĠÑĩеÑĢез": 129334, + "ĠÑĩеÑĤ": 139845, + "ĠÑĩеÑĤÑĭ": 139846, + "ĠÑĩеÑĤÑĭÑĢе": 139847, + "ĠÑĩиÑģ": 40076, + "ĠÑĩиÑģл": 135752, + "ĠÑĩиÑģле": 134307, + "ĠÑĩиÑģло": 64519, + "ĠÑĩÑĤ": 69360, + "ĠÑĩÑĤо": 47389, + "ĠÑĩÑĤобÑĭ": 81150, + "ĠÑĩÑĥв": 143774, + "ĠÑĩÑĥвÑģÑĤво": 143775, + "ĠÑĩÑĥÑĤÑĮ": 143425, + "ĠÑĪ": 54517, + "ĠÑĪаг": 140697, + "ĠÑĪиÑĢ": 142704, + "ĠÑĪиÑĢоко": 142705, + "ĠÑĪкол": 131862, + "ĠÑĪколÑĮ": 141712, + "ĠÑĪколÑĮн": 141713, + "ĠÑį": 20928, + "ĠÑįк": 80657, + "ĠÑįкон": 138360, + "ĠÑįконом": 138361, + "ĠÑįкономиÑĩеÑģк": 138362, + "ĠÑįкÑĢан": 141244, + "ĠÑįкÑģ": 132664, + "ĠÑįкÑģп": 137978, + "ĠÑįкÑģпеÑĢÑĤ": 137979, + "ĠÑįлекÑĤ": 131228, + "ĠÑįлекÑĤÑĢо": 131229, + "ĠÑįлеменÑĤ": 44988, + "ĠÑįнеÑĢ": 140228, + "ĠÑįнеÑĢги": 140229, + "ĠÑįÑĤ": 48003, + "ĠÑįÑĤа": 134891, + "ĠÑįÑĤап": 137232, + "ĠÑįÑĤи": 126979, + "ĠÑįÑĤим": 134100, + "ĠÑįÑĤиÑħ": 133142, + "ĠÑįÑĤо": 67879, + "ĠÑįÑĤого": 128866, + "ĠÑįÑĤой": 129892, + "ĠÑįÑĤом": 128649, + "ĠÑįÑĤомÑĥ": 138627, + "ĠÑįÑĤоÑĤ": 129863, + "ĠÑįÑĤÑĥ": 135401, + "ĠÑįÑĦÑĦ": 132716, + "ĠÑįÑĦÑĦекÑĤив": 132717, + "ĠÑı": 45310, + "ĠÑıв": 125609, + "ĠÑıвлÑı": 125610, + "ĠÑıвлÑıеÑĤÑģÑı": 129143, + "ĠÑıвлÑıÑİÑĤÑģÑı": 134958, + "ĠÑıзÑĭк": 130844, + "ĠÑıн": 139550, + "ĠÑıнваÑĢ": 139551, + "ĠÑıнваÑĢÑı": 139552, + "ĠÑĸ": 83854, + "Ġ×": 17284, + "Ġס": 124313, + "Ġספר": 126528, + "Ġספר×Ļ×Ŀ": 135592, + "Ġסר×ĺ": 131079, + "Ġס×ij×Ļ×ij": 136036, + "Ġס×ķ": 129206, + "Ġס×ķ×Ĵ": 137308, + "Ġס×ķ×Ĵ×Ļ": 137309, + "Ġס×Ļ": 128497, + "Ġס×Ļפ×ķר": 131737, + "Ġ×¢": 123855, + "Ġעסק": 136549, + "Ġעסק×Ļ": 137910, + "Ġעסק×Ļ×Ŀ": 136550, + "Ġעצ": 125272, + "Ġעצ×ŀ": 139608, + "Ġעצ×ŀ×ķ": 130758, + "Ġעצ×ŀ×Ļ": 132755, + "Ġעצ×ŀ×Ŀ": 139609, + "Ġ×¢×§×ij": 135654, + "Ġער": 136297, + "Ġער×Ļ׼×Ķ": 136298, + "Ġעש": 134727, + "Ġעשר×ķת": 141077, + "Ġעש×ķ×Ļ": 134728, + "Ġ×¢×ijר": 131917, + "Ġ×¢×ij×ķר": 128952, + "Ġ×¢×ij×ķ×ĵ": 125751, + "Ġ×¢×ij×ķ×ĵת": 139619, + "Ġ×¢×ij×ķ×ĵ×Ķ": 130006, + "Ġ×¢×ij×ķ×ĵ×ķת": 142881, + "Ġ×¢×ĵ": 128372, + "Ġ×¢×ĵ×Ļ": 130713, + "Ġ×¢×ĵ×Ļ×Ļף": 130714, + "Ġ×¢×ķ": 128686, + "Ġ×¢×ķס": 141657, + "Ġ×¢×ķסק": 141658, + "Ġ×¢×ķר": 140857, + "Ġ×¢×ķר׼×Ļ": 140858, + "Ġ×¢×ķש": 127381, + "Ġ×¢×ķש×Ķ": 132888, + "Ġ×¢×ķש×Ļ×Ŀ": 133605, + "Ġ×¢×ķ×ij": 136618, + "Ġ×¢×ķ×ijר": 136554, + "Ġ×¢×ķ×ij×ĵ×Ļ×Ŀ": 136619, + "Ġ×¢×ķ×ĵ": 128524, + "Ġ×¢×ķ׾×Ķ": 134677, + "Ġ×¢×ķ×ŀ": 133273, + "Ġ×¢×ķ×ŀ×ĵ": 133274, + "Ġ×¢×Ļצ×ķ×ij": 131782, + "Ġ×¢×Ľ×©×Ļ×ķ": 130581, + "Ġ×¢×ľ": 126194, + "Ġ×¢×ľ×Ļ×Ķ×Ŀ": 137547, + "Ġ×¢×ľ×Ļ×ķ": 130368, + "Ġ×¢×Ŀ": 128256, + "Ġ×¢×ŀ": 126753, + "Ġ×¢×ŀ×ķ×§": 140788, + "Ġפ": 123991, + "Ġפע": 126639, + "Ġפע×ķ׾×Ķ": 133413, + "Ġפע×Ŀ": 129601, + "Ġפר": 124550, + "Ġפרס": 137467, + "Ġפרס×ķ×Ŀ": 137468, + "Ġפרשת": 141045, + "Ġפר×ķפ": 139946, + "Ġפר×ĺ×Ļ": 127094, + "Ġפר×ĺ×Ļ×ķת": 137588, + "Ġפר×ĺ×Ļ×Ŀ": 137147, + "Ġפר×Ļ": 131760, + "Ġפש×ķ×ĺ": 129389, + "Ġפת": 139262, + "Ġפת×ķ×Ĺ": 139263, + "Ġפת×Ĺ": 136659, + "Ġפ×ķ": 129908, + "Ġפ×ķס": 142321, + "Ġפ×ķס×ĺ×Ļ×Ŀ": 142322, + "Ġפ×Ĺ×ķת": 131115, + "Ġפ×Ļ": 128483, + "Ġפ׳×Ļ": 130265, + "Ġצ": 124393, + "Ġצע×Ļר": 139478, + "Ġצפ×ķ": 138286, + "Ġצר": 125606, + "Ġצר×Ļ×ļ": 129107, + "Ġצר×Ļ׼": 134416, + "Ġצר×Ļ׼×Ķ": 137033, + "Ġצר×Ļ׼×Ļ×Ŀ": 134417, + "Ġצ×ķר×ļ": 134046, + "Ġצ×Ļ×ij×ķר": 134847, + "Ġ×§": 123974, + "Ġקצר": 134751, + "Ġקצת": 130667, + "Ġקר": 129350, + "Ġקשר": 129478, + "Ġקש×Ķ": 131230, + "Ġ×§×ij×ķ×¢": 140103, + "Ġ×§×ķ": 129088, + "Ġ×§×ķפ": 142730, + "Ġ×§×ķר×Ķ": 138633, + "Ġ×§×ķ×ĵ": 127142, + "Ġ×§×ķ×ĵ×Ŀ": 133992, + "Ġ×§×ķ׳": 138981, + "Ġ×§×ĺ": 126035, + "Ġ×§×ĺף": 133389, + "Ġ×§×ĺ׳×Ķ": 139504, + "Ġ×§×ĺ׳×Ļ×Ŀ": 138970, + "Ġ×§×Ļש": 137036, + "Ġ×§×Ļש×ķר×Ļ×Ŀ": 137037, + "Ġ×§×Ļ×ij׾": 132360, + "Ġק׾": 129465, + "Ġק׳": 136488, + "Ġר": 123937, + "Ġרפ×ķ×IJ×Ļ": 134567, + "Ġרק": 128390, + "Ġרש": 126162, + "Ġרש×IJ×Ļ": 143101, + "Ġרש×ķ×ŀ": 139108, + "Ġרש×Ļ×ŀת": 141856, + "Ġר×IJ": 124869, + "Ġר×IJש": 126525, + "Ġר×IJש×ķף": 132455, + "Ġר×IJש×Ļ": 133513, + "Ġר×IJ×ķ×Ļ": 139927, + "Ġר×IJ×Ļת×Ļ": 143537, + "Ġר×ij": 129079, + "Ġר×ij×Ķ": 133524, + "Ġר×ij×ķת": 131674, + "Ġר×ij×Ļ×Ŀ": 129747, + "Ġר×Ĵ׾": 141642, + "Ġר×ķ": 129226, + "Ġר×ķצ×Ķ": 129469, + "Ġר×ķצ×Ļ×Ŀ": 131812, + "Ġר×ķ×IJ×Ķ": 136409, + "Ġר×ķ×ij": 131913, + "Ġר×Ĺ×ij": 134085, + "Ġר×Ĺ×ķ×§": 142118, + "Ġר×Ļ": 129716, + "Ġש": 87798, + "Ġשע": 130202, + "Ġשע×ijר": 135587, + "Ġשע×ķת": 131717, + "Ġ×©×¢×ľ": 134635, + "Ġשצר": 138202, + "Ġשצר×Ļ×ļ": 138203, + "Ġשק": 143487, + "Ġשק×Ļ×ij׾": 143488, + "Ġשק׾": 127798, + "Ġשק׾×Ļ×Ŀ": 139195, + "Ġשר": 129684, + "Ġשש": 132979, + "Ġשת": 129158, + "Ġשת×Ļ": 131513, + "Ġש×IJ": 124508, + "Ġש×IJת": 127598, + "Ġש×IJת×Ķ": 134273, + "Ġש×IJת×Ŀ": 134906, + "Ġש×IJ×Ļ": 134073, + "Ġש×IJ×Ļף": 131794, + "Ġש×IJ×Ļ׳×ķ": 140643, + "Ġש×IJ×Ļ׳×Ŀ": 143693, + "Ġש×IJ׾×Ķ": 136576, + "Ġש×IJ׳": 136370, + "Ġש×IJ׳×Ĺ׳×ķ": 136371, + "Ġש×IJ׳×Ļ": 129557, + "Ġש×ij": 126234, + "Ġש×ij×Ķ": 131861, + "Ġש×ij×ķ": 129670, + "Ġש×Ĵ×Ŀ": 135721, + "Ġש×Ķ": 124601, + "Ġש×Ķת": 131960, + "Ġש×Ķ×ķ×IJ": 128770, + "Ġש×Ķ×Ļ×IJ": 130381, + "Ġש×Ķ×Ļ×Ķ": 132369, + "Ġש×Ķ×Ļ×ķ": 138555, + "Ġש×Ķ×Ŀ": 130325, + "Ġש×ķ": 128945, + "Ġש×ķ×ij": 130953, + "Ġש×ķ׳×Ķ": 136894, + "Ġש×ķ׳×ķת": 133497, + "Ġש×ķ׳×Ļ×Ŀ": 130384, + "Ġש×ĸ×Ķ": 130424, + "Ġש×Ĺ": 129929, + "Ġש×Ĺ×ķר": 139387, + "Ġש×Ļ": 128326, + "Ġש×Ļר×ķת": 127009, + "Ġש×Ļר×ķת×Ļ": 136014, + "Ġש×Ļש": 128756, + "Ġש×Ļ×Ķ": 138409, + "Ġש×Ļ×Ķ×Ļ×Ķ": 138410, + "Ġש×Ļ׼×ķ׾": 139800, + "Ġש×Ļ×ŀ×ķש": 130251, + "Ġש×Ļ׳×ķ×Ļ": 139200, + "Ġש×Ļ׳×ķ×Ļ×Ļ×Ŀ": 139201, + "Ġש׼": 126494, + "Ġש׼×ijר": 140213, + "Ġש׼׾": 132974, + "Ġש׼ף": 135672, + "Ġש׾": 124756, + "Ġש׾×IJ": 128517, + "Ġש׾×Ķ": 126617, + "Ġש׾×Ķ×Ŀ": 129812, + "Ġש׾×ķ": 128581, + "Ġש׾×Ļ": 128554, + "Ġש׾×Ļש×Ļ": 140135, + "Ġש׾×ļ": 128723, + "Ġש׾׼×Ŀ": 129406, + "Ġש׾׳×ķ": 128733, + "Ġש×Ŀ": 128768, + "Ġש×ŀ": 126222, + "Ġש×ŀ×ķר×ķת": 131739, + "Ġש׳": 128383, + "Ġ×©×ł×ª": 133581, + "Ġש׳×Ķ": 129391, + "Ġש׳×Ļ": 128735, + "Ġש׳×Ļת": 139259, + "Ġש׳×Ļ×ª×Ł": 139260, + "Ġש׳×Ļ×Ŀ": 129051, + "Ġת": 123921, + "Ġתע": 133739, + "Ġתפ": 135655, + "Ġתפק×Ļ×ĵ": 136066, + "Ġתש": 131784, + "Ġ×ª×©×ľ×ķ×Ŀ": 140452, + "Ġתת": 134831, + "Ġת×IJ": 132517, + "Ġת×Ĵ×ķ×ij×Ķ": 131690, + "Ġת×Ĵ×ķ×ij×ķת": 132081, + "Ġת×Ķ": 134020, + "Ġת×Ķ×Ļ×Ķ": 134021, + "Ġת×ķ": 129440, + "Ġת×ķר×Ķ": 140534, + "Ġת×ķ×ĵ×Ķ": 131027, + "Ġת×ķ×ļ": 130792, + "Ġת×ķ׼": 127557, + "Ġת×ķ׼׾": 140418, + "Ġת×ķ׼׾×ķ": 133974, + "Ġת×ķ׼׳": 139689, + "Ġת×Ļ×§": 136638, + "Ġ×ª×ľ": 126758, + "Ġ×ª×ľ×ķ×Ļ": 139590, + "Ġת×ŀ": 133596, + "Ġת×ŀ×ķ׳×Ķ": 139823, + "Ġת×ŀ×Ļ×ĵ": 130308, + "Ġ×ª×ł": 131934, + "Ġ×ª×ł×IJ×Ļ": 131935, + "Ġ×IJ": 61960, + "Ġ×IJס×ķר": 141705, + "Ġ×IJ×£": 129540, + "Ġ×IJפשר": 126378, + "Ġ×IJפשר×ķת": 137710, + "Ġ×IJפ×Ļ׾×ķ": 131061, + "Ġ×IJצ׾": 126820, + "Ġ×IJצ׾׳×ķ": 138877, + "Ġ×IJר": 128789, + "Ġ×IJש": 130782, + "Ġ×IJשר": 128645, + "Ġ×IJת": 125005, + "Ġ×IJתר": 126788, + "Ġ×IJתר×Ļ×Ŀ": 130605, + "Ġ×IJת×Ķ": 128810, + "Ġ×IJ×ª×Ľ×Ŀ": 136104, + "Ġ×IJת×Ŀ": 129146, + "Ġ×IJ×ijר×Ķ": 143354, + "Ġ×IJ×ijר×Ķ×Ŀ": 143355, + "Ġ×IJ×ij×Ļ": 131785, + "Ġ×IJ×ij×Ļ×ij": 129811, + "Ġ×IJ×ij׾": 128343, + "Ġ×IJ×ĵ×Ŀ": 129412, + "Ġ×IJ×Ķ": 139452, + "Ġ×IJ×Ķ×ijת×Ļ": 139453, + "Ġ×IJ×ķ": 128255, + "Ġ×IJ×ķר": 129997, + "Ġ×IJ×ķת": 124507, + "Ġ×IJ×ķת×Ķ": 129272, + "Ġ×IJ×ķת×ķ": 128532, + "Ġ×IJ×ķת×Ļ": 129990, + "Ġ×IJ×ķת×ļ": 133025, + "Ġ×IJ×ķת×Ŀ": 129394, + "Ġ×IJ×ķ×ª×Ł": 138305, + "Ġ×IJ×ķ×ª×ł×ķ": 132817, + "Ġ×IJ×ķ×ĵ×ķת": 134767, + "Ġ×IJ×ķ×Ķ": 128163, + "Ġ×IJ×ķ×Ķ×ijת": 141505, + "Ġ×IJ×ķ×Ķ×ij×Ļ×Ŀ": 143361, + "Ġ×IJ×ķ×ķ×Ļר": 142005, + "Ġ×IJ×ķ׾×Ļ": 130432, + "Ġ×IJ×ķ×ŀר": 129552, + "Ġ×IJ×ĸ": 128619, + "Ġ×IJ×ĸר": 142399, + "Ġ×IJ×ĸר×Ĺ×Ļ": 142400, + "Ġ×IJ×Ĺ": 127576, + "Ġ×IJ×Ĺר": 124521, + "Ġ×IJ×Ĺרת": 131406, + "Ġ×IJ×Ĺר×ķת": 135113, + "Ġ×IJ×Ĺר×ķ׳×ķת": 141924, + "Ġ×IJ×Ĺר×ķ׳×Ļ×Ŀ": 142433, + "Ġ×IJ×Ĺר×Ļ": 128709, + "Ġ×IJ×Ĺר×Ļ×Ŀ": 130387, + "Ġ×IJ×Ĺת": 128966, + "Ġ×IJ×Ĺ×ĵ": 128395, + "Ġ×IJ×Ĺ×ķ×ĸ": 138613, + "Ġ×IJ×Ĺ×Ļ": 139934, + "Ġ×IJ×Ļ": 128303, + "Ġ×IJ×Ļפ×Ķ": 139414, + "Ġ×IJ×Ļר×ķ×¢": 136419, + "Ġ×IJ×Ļר×ķ×¢×Ļ×Ŀ": 136420, + "Ġ×IJ×Ļש": 135793, + "Ġ×IJ×Ļש×Ļ": 130982, + "Ġ×IJ×Ļש×Ļת": 135794, + "Ġ×IJ×Ļת": 139430, + "Ġ×IJ×Ļת×ķ": 139431, + "Ġ×IJ×Ļ×ļ": 128829, + "Ġ×IJ×Ļ׼×ķת": 138438, + "Ġ×IJ×Ļ׼×ķת×Ļ": 138439, + "Ġ×IJ×Ļ׾": 138372, + "Ġ×IJ×Ļ×ŀ×Ļ": 142457, + "Ġ×IJ×Ļ×ŀ×Ļ×Ļ׾": 142458, + "Ġ×IJ×Ļף": 128455, + "Ġ×IJ×Ļ׳×Ķ": 133183, + "Ġ×IJ×Ļ׳×ķ": 130256, + "Ġ×IJ×Ļ׳×Ŀ": 135479, + "Ġ×IJ×ļ": 128594, + "Ġ×IJ׼ף": 137152, + "Ġ×IJ׾": 124612, + "Ġ×IJ×ľ×£": 138642, + "Ġ×IJ׾פ×Ļ": 139879, + "Ġ×IJ׾ק": 135299, + "Ġ×IJ׾ק×ĺר": 135300, + "Ġ×IJ׾ק×ĺר×ķ׳×Ļ": 135301, + "Ġ×IJ׾×IJ": 128919, + "Ġ×IJ׾×Ķ": 129672, + "Ġ×IJ׾×ķ": 129474, + "Ġ×IJ׾×Ļ×Ķ": 134498, + "Ġ×IJ׾×Ļ×ķ": 134332, + "Ġ×IJ׾×Ļ׳×ķ": 141006, + "Ġ×IJ×Ŀ": 128316, + "Ġ×IJ×ŀר": 129042, + "Ġ×IJ×ŀ×Ļת×Ļ": 135924, + "Ġ×IJ×ŀ׳": 139129, + "Ġ×IJ×ŀ׳×Ŀ": 139130, + "Ġ×IJ׳": 125594, + "Ġ×IJ׳ש×Ļ×Ŀ": 129403, + "Ġ×IJ׳×ķ": 129172, + "Ġ×IJ׳×Ĺ׳×ķ": 129055, + "Ġ×IJ׳×Ļ": 128297, + "Ġ×ij": 88531, + "Ġ×ijס": 125684, + "Ġ×ijספר": 136321, + "Ġ×ijס×ĵר": 136099, + "Ġ×ij×¢": 124182, + "Ġ×ijעצ": 138257, + "Ġ×ijעצ×Ŀ": 138258, + "Ġ×ijעת": 134555, + "Ġ×ij×¢×ijר": 136477, + "Ġ×ij×¢×ķ×ĵ": 135857, + "Ġ×ij×¢×ķ׾×Ŀ": 129973, + "Ġ×ij×¢×ĸר": 138297, + "Ġ×ij×¢×ĸרת": 138298, + "Ġ×ij×¢×Ļקר": 132480, + "Ġ×ij×¢×Ļר": 134474, + "Ġ×ij×¢×Ļ׳×Ļ": 136918, + "Ġ×ij×¢×ľ": 129746, + "Ġ×ij×¢×ľ×Ļ": 130703, + "Ġ×ijפ×Ļ×Ļס×ij×ķ×§": 138042, + "Ġ×ijפ׳×Ļ": 131995, + "Ġ×ijצ": 127047, + "Ġ×ijצ×ķר×Ķ": 131086, + "Ġ×ijקר×ij": 136377, + "Ġ×ij×§×ķ": 135044, + "Ġ×ijק׾": 142472, + "Ġ×ijק׾×ķת": 142473, + "Ġ×ijר": 124803, + "Ġ×ijרשת": 136394, + "Ġ×ijר×IJש": 133833, + "Ġ×ijר×ķר": 131733, + "Ġ×ijר×Ĺ": 139937, + "Ġ×ijר×Ĺ×ij×Ļ": 139938, + "Ġ×ijר×Ļ×IJ": 131588, + "Ġ×ijש": 124823, + "Ġ×ijש×ij": 127212, + "Ġ×ijש×ij×ķ×¢": 139662, + "Ġ×ijש×ij×Ļ׾": 131914, + "Ġ×ijש׾": 127185, + "Ġ×ijש׾×ij": 137393, + "Ġ×ijש×Ŀ": 136318, + "Ġ×ij×©×ł×ª": 131676, + "Ġ×ijש׳×Ķ": 139198, + "Ġ×ijש׳×Ļ×Ŀ": 138572, + "Ġ×ijת": 125025, + "Ġ×ijת×ķר": 138814, + "Ġ×ijת×ķ×ļ": 131515, + "Ġ×ijת×Ĺ": 130971, + "Ġ×ijת×Ĺ×ķ×Ŀ": 130972, + "Ġ×ij×ª×ľ": 134973, + "Ġ×ij×IJ": 124820, + "Ġ×ij×IJר": 127062, + "Ġ×ij×IJרץ": 131197, + "Ġ×ij×IJר×Ķ": 140698, + "Ġ×ij×IJתר": 129011, + "Ġ×ij×IJ×ķ": 132355, + "Ġ×ij×IJ×ķפף": 129110, + "Ġ×ij×IJ×ķת": 127935, + "Ġ×ij×IJ×ķת×Ķ": 140702, + "Ġ×ij×IJ×ķת×ķ": 137605, + "Ġ×ij×IJ×ĸ×ķר": 134941, + "Ġ×ij×IJ×Ļ": 131992, + "Ġ×ij×IJ×Ļ׳×ĺר׳×ĺ": 136795, + "Ġ×ij×IJ×ŀ": 126536, + "Ġ×ij×IJ×ŀצע": 129267, + "Ġ×ij×IJ×ŀצע×ķת": 129268, + "Ġ×ij×IJ×ŀת": 130979, + "Ġ×ij×ij": 125644, + "Ġ×ij×ij×§": 143393, + "Ġ×ij×ijקש×Ķ": 143394, + "Ġ×ij×ij×ķקר": 142880, + "Ġ×ij×ij×Ļת": 129326, + "Ġ×ij×Ĵ": 125773, + "Ġ×ij×Ĵ×Ļ׾": 137363, + "Ġ×ij×Ĵ×Ļף": 138958, + "Ġ×ij×Ĵ׾": 131948, + "Ġ×ij×Ĵ׾׾": 131949, + "Ġ×ij×ĵר×ļ": 130687, + "Ġ×ij×ĵ×ijר": 139237, + "Ġ×ij×ĵ×Ļ": 131624, + "Ġ×ij×ĵ×Ļ×ķ×§": 131625, + "Ġ×ij×Ķ": 125443, + "Ġ×ij×Ķת": 134130, + "Ġ×ij×Ķת×IJ×Ŀ": 134131, + "Ġ×ij×Ķ×Ĺ׾×ĺ": 136536, + "Ġ×ij×Ķ×Ŀ": 130586, + "Ġ×ij×ķ": 128408, + "Ġ×ij×ĸ×Ķ": 137598, + "Ġ×ij×ĸ׼": 141170, + "Ġ×ij×ĸ׼×ķת": 141171, + "Ġ×ij×ĸ×ŀף": 131814, + "Ġ×ij×Ĺ": 124842, + "Ġ×ij×Ĺש×ij×ķף": 141573, + "Ġ×ij×Ĺ×ĵר": 140471, + "Ġ×ij×Ĺ×ķ": 131259, + "Ġ×ij×Ĺ×Ļ": 127983, + "Ġ×ij×Ĺ×Ļ×Ļ": 138454, + "Ġ×ij×Ĺ×Ļ×Ļ×Ŀ": 140390, + "Ġ×ij×Ĺ×Ļ׳×Ŀ": 143356, + "Ġ×ij×ĺ": 130120, + "Ġ×ij×Ļ": 128851, + "Ġ×ij×Ļ×§": 141561, + "Ġ×ij×Ļקש": 141562, + "Ġ×ij×Ļר×ķש׾×Ļ×Ŀ": 137341, + "Ġ×ij×Ļשר×IJ׾": 129564, + "Ġ×ij×Ļת": 128520, + "Ġ×ij×Ļ×ĵ×Ļ": 135422, + "Ġ×ij×Ļ×ķתר": 128567, + "Ġ×ij×Ļ×ķ×Ŀ": 130433, + "Ġ×ij×Ļ×Ĺ": 140991, + "Ġ×ij×Ļ×Ĺ×ĵ": 140992, + "Ġ×ij×Ļף": 128389, + "Ġ×ij×Ļ׳": 127660, + "Ġ×ij×Ļ׳׾×IJ×ķ×ŀ×Ļ": 143963, + "Ġ×ij׼": 129199, + "Ġ×ij×Ľ×ª×ij": 138781, + "Ġ×ij׼×ļ": 135000, + "Ġ×ij׼׾": 126262, + "Ġ×ij׼׾׾": 131650, + "Ġ×ij׾": 130661, + "Ġ×ij׾×ij×ĵ": 130662, + "Ġ×ij׾×ķ×Ĵ": 137567, + "Ġ×ij׾×Ļ": 129563, + "Ġ×ij×ŀ": 125014, + "Ġ×ij×ŀס": 132928, + "Ġ×ij×ŀס×Ĵר": 132929, + "Ġ×ij×ŀס×Ĵרת": 132930, + "Ġ×ij×ŀצ×ij": 136895, + "Ġ×ij×ŀקר×Ķ": 133664, + "Ġ×ij×ŀקר×Ļ×Ŀ": 142619, + "Ġ×ij×ŀ×§×ķ×Ŀ": 129995, + "Ġ×ij×ŀר׼×ĸ": 137807, + "Ġ×ij×ŀש×ļ": 133517, + "Ġ×ij×ŀ×Ķ": 130806, + "Ġ×ij×ŀ×Ķ׾": 130807, + "Ġ×ij×ŀ×Ķ׾×ļ": 130808, + "Ġ×ij×ŀ×Ĺ×Ļר": 139006, + "Ġ×ij×ŀ×Ļ": 131447, + "Ġ×ij×ŀ×Ļ×ĵ": 135936, + "Ġ×ij×ŀ×Ļ×ĵ×Ķ": 135937, + "Ġ×ij×ŀ×Ļ×ķ×Ĺ×ĵ": 131448, + "Ġ×ijף": 129218, + "Ġ×ij׳": 125278, + "Ġ×ij׳×ķסף": 132993, + "Ġ×ij׳×ķש×IJ": 130886, + "Ġ×ij׳×ķ×Ĵ": 143694, + "Ġ×ij׳×ķ×Ĵ×¢": 143695, + "Ġ×ij׳×Ļ": 126466, + "Ġ×ij׳×Ļ×Ļת": 135548, + "Ġ×Ĵ": 124203, + "Ġ×Ĵר": 129338, + "Ġ×Ĵ×ij×ķ×Ķ": 127449, + "Ġ×Ĵ×ij×ķ×Ķ×Ķ": 133772, + "Ġ×Ĵ×ij×Ļ": 136205, + "Ġ×Ĵ×ĵ": 126527, + "Ġ×Ĵ×ĵ×ķ׾": 129251, + "Ġ×Ĵ×ĵ×ķ׾×Ķ": 133699, + "Ġ×Ĵ×ķפ": 134977, + "Ġ×Ĵ×ķ×Ĵ׾": 143913, + "Ġ×Ĵ×Ļ׾": 131334, + "Ġ×Ĵ×Ŀ": 128274, + "Ġ×ĵ": 124113, + "Ġ×ĵעת": 131076, + "Ġ×ĵ×§×ķת": 132104, + "Ġ×ĵר×ļ": 129487, + "Ġ×ĵר׼×Ļ×Ŀ": 140034, + "Ġ×ĵ×ijר": 128852, + "Ġ×ĵ×ijר×Ļ×Ŀ": 131184, + "Ġ×ĵ×ķ": 128681, + "Ġ×ĵ×ķר": 143078, + "Ġ×ĵ×ķרש": 143079, + "Ġ×ĵ×ķ×IJר": 140360, + "Ġ×ĵ×ķ׾": 136479, + "Ġ×ĵ×ķ׾ר": 136480, + "Ġ×ĵ×Ļ": 128593, + "Ġ×ĵ×Ļ×Ĵ": 141953, + "Ġ×ĵ×Ļ×Ĵ×Ļ×ĺ": 141954, + "Ġ×ĵ×Ļ×Ĵ×Ļ×ĺ׾×Ļ": 141955, + "Ġ×ĵ×Ļף": 130774, + "Ġ×Ķ": 69346, + "Ġ×Ķס": 124622, + "Ġ×Ķספר": 129417, + "Ġ×Ķסר×ĺ": 133284, + "Ġ×Ķס×ķ": 132518, + "Ġ×Ķס×Ļ": 131427, + "Ġ×Ķס×Ļפ×ķר": 137219, + "Ġ×Ķ×¢": 124327, + "Ġ×Ķעסק": 131713, + "Ġ×Ķער×ij": 134250, + "Ġ×Ķ×¢×ij×ķ×ĵ": 131332, + "Ġ×Ķ×¢×ij×ķ×ĵ×Ķ": 131333, + "Ġ×Ķ×¢×ķ×ij": 139546, + "Ġ×Ķ×¢×ķ×ij×ĵ×Ķ": 139547, + "Ġ×Ķ×¢×ķ׾×Ŀ": 130122, + "Ġ×Ķ×¢×Ļ×§": 139625, + "Ġ×Ķ×¢×Ļקר×Ļ": 139626, + "Ġ×Ķ×¢×Ļר": 130189, + "Ġ×Ķ×¢×ľ": 133885, + "Ġ×Ķ×¢×ľ×Ļ": 141134, + "Ġ×Ķ×¢×ľ×Ļ×ķף": 141135, + "Ġ×Ķפ": 126338, + "Ġ×Ķפר": 129161, + "Ġ×Ķפ×ķ×¢×ľ": 136976, + "Ġ×Ķפ×Ļ": 133582, + "Ġ×Ķצ": 124664, + "Ġ×Ķצ×ĵ": 138085, + "Ġ×Ķצ×ĵ×ĵ×Ļ×Ŀ": 138086, + "Ġ×Ķצ×ķר×ļ": 142098, + "Ġ×Ķצ×ĺר": 143397, + "Ġ×Ķצ×ĺרפ×ķ": 143398, + "Ġ×Ķצ×Ļ": 134565, + "Ġ×Ķצ×Ļ×ij": 142427, + "Ġ×Ķצ×Ļ×ij×ķר": 136362, + "Ġ×Ķצ×Ļ×ij×ķר×Ļ": 142428, + "Ġ×Ķצ׾×Ļ×Ĺ": 138223, + "Ġ×Ķ×§": 124814, + "Ġ×Ķקר": 136021, + "Ġ×Ķקר×ķ×ij": 136022, + "Ġ×Ķקשר": 137689, + "Ġ×Ķקש×ķר": 142608, + "Ġ×Ķ×§×ķ": 130877, + "Ġ×Ķ×§×ķ×ĵ": 143289, + "Ġ×Ķ×§×ķ×ĵ×Ŀ": 143290, + "Ġ×Ķק׾": 135599, + "Ġ×Ķר": 125036, + "Ġ×Ķרפ×ķ×IJ×Ļ": 143511, + "Ġ×Ķר×IJש": 125230, + "Ġ×Ķר×IJש×ķף": 130126, + "Ġ×Ķר×IJש×ķ׳×Ķ": 131980, + "Ġ×Ķר×IJש×ķ׳×Ļ×Ŀ": 143268, + "Ġ×Ķר×ij": 129292, + "Ġ×Ķר×ij×Ķ": 128923, + "Ġ×Ķר×Ļ": 129933, + "Ġ×Ķר׼×ij": 133737, + "Ġ×Ķש": 124081, + "Ġ×Ķשק": 136899, + "Ġ×Ķשת": 133313, + "Ġ×Ķש×IJ": 140004, + "Ġ×Ķש×IJ׾×Ķ": 140005, + "Ġ×Ķש×ij×ķ×¢": 133312, + "Ġ×Ķש×ķ׳×Ļ×Ŀ": 140453, + "Ġ×Ķש×Ļ": 130675, + "Ġ×Ķש×Ļר×ķת": 135030, + "Ġ×Ķש×Ļ×ŀ×ķש": 133369, + "Ġ×Ķש׾": 127462, + "Ġ×Ķש׾×Ļש×Ļ": 139709, + "Ġ×Ķש׳×Ķ": 131523, + "Ġ×Ķש׳×Ļ": 126975, + "Ġ×Ķש׳×Ļ×Ļ×Ķ": 139747, + "Ġ×Ķש׳×Ļ×Ŀ": 136001, + "Ġ×Ķת": 125433, + "Ġ×Ķתק": 132166, + "Ġ×Ķת×ij": 141487, + "Ġ×Ķת×ij×Ļ×¢×Ķ": 141488, + "Ġ×Ķת×ķר×Ķ": 139510, + "Ġ×Ķ×IJ": 123960, + "Ġ×Ķ×IJר": 132482, + "Ġ×Ķ×IJרץ": 132483, + "Ġ×Ķ×IJש": 139710, + "Ġ×Ķ×IJתר": 129967, + "Ġ×Ķ×IJ×ĵ×Ŀ": 132209, + "Ġ×Ķ×IJ×ķ": 130090, + "Ġ×Ķ×IJ×ķ×ķ×Ļר": 141248, + "Ġ×Ķ×IJ×Ĺ": 133193, + "Ġ×Ķ×IJ×Ĺר": 125372, + "Ġ×Ķ×IJ×Ĺר×ķף": 133747, + "Ġ×Ķ×IJ×Ĺר×ķ׳×Ķ": 138531, + "Ġ×Ķ×IJ×Ĺר×ķ׳×ķת": 135687, + "Ġ×Ķ×IJ×Ĺר×ķ׳×Ļ×Ŀ": 141646, + "Ġ×Ķ×IJ×Ļ": 129192, + "Ġ×Ķ×IJ×Ļר×ķ×¢": 135907, + "Ġ×Ķ×IJ×Ļש×Ļ": 135724, + "Ġ×Ķ×IJ×Ļ׳×ĺר׳×ĺ": 137593, + "Ġ×Ķ×IJ׾×Ķ": 134912, + "Ġ×Ķ×IJ׾×ķ": 139751, + "Ġ×Ķ×IJ×Ŀ": 128807, + "Ġ×Ķ×IJ×ŀר": 142986, + "Ġ×Ķ×IJ×ŀר×Ļ×§": 142987, + "Ġ×Ķ×IJ×ŀר×Ļ×§×IJ×Ļ": 142988, + "Ġ×Ķ×IJ×ŀ×Ļת×Ļ": 142599, + "Ġ×Ķ×ij": 124815, + "Ġ×Ķ×ij×¢": 139351, + "Ġ×Ķ×ij×¢×Ļ×Ķ": 139352, + "Ġ×Ķ×ijר": 128023, + "Ġ×Ķ×ijר×Ļת": 138864, + "Ġ×Ķ×ijר×Ļ×IJ×ķת": 140756, + "Ġ×Ķ×ij×IJ": 129851, + "Ġ×Ķ×ij×IJ×Ļ×Ŀ": 138161, + "Ġ×Ķ×ij×ķקר": 143690, + "Ġ×Ķ×ij×Ļ": 132824, + "Ġ×Ķ×ij×Ļת": 129073, + "Ġ×Ķ×ij׾": 139244, + "Ġ×Ķ×ij׾×ķ×Ĵ": 139245, + "Ġ×Ķ×Ĵ": 125064, + "Ġ×Ķ×Ĵ×ij×ķ×Ķ": 140875, + "Ġ×Ķ×Ĵ×ĵ×ķ׾": 129682, + "Ġ×Ķ×Ĵ×ķ×£": 136906, + "Ġ×Ķ×Ĵ×Ļ×¢": 130879, + "Ġ×Ķ×ĵ": 126484, + "Ġ×Ķ×ĵר×ļ": 132868, + "Ġ×Ķ×ĵ×ijר": 127320, + "Ġ×Ķ×ĵ×ijר×Ļ×Ŀ": 133983, + "Ġ×Ķ×ĵ×ķ": 132121, + "Ġ×Ķ×ĵ×Ļ": 132865, + "Ġ×Ķ×ĵ×Ļף": 132030, + "Ġ×Ķ×Ķ": 128341, + "Ġ×Ķ×Ķת": 133417, + "Ġ×Ķ×ķ": 128437, + "Ġ×Ķ×ķ×IJ": 128262, + "Ġ×Ķ×ķ×ĵ": 140832, + "Ġ×Ķ×ķ×ĵ×¢×Ķ": 140833, + "Ġ×Ķ×ķ׾×ĵ": 141691, + "Ġ×Ķ×ķ׾×ĵת": 141692, + "Ġ×Ķ×ĸ": 124892, + "Ġ×Ķ×ĸ×IJת": 134471, + "Ġ×Ķ×ĸ×Ķ": 128561, + "Ġ×Ķ×ĸ×ķ": 130948, + "Ġ×Ķ×ĸ×ķ×Ĵ": 137897, + "Ġ×Ķ×ĸ׼": 131093, + "Ġ×Ķ×ĸ׼×ķ×Ļ×ķת": 131094, + "Ġ×Ķ×ĸ×ŀף": 130576, + "Ġ×Ķ×Ĺ": 124789, + "Ġ×Ķ×Ĺ×ijר×Ķ": 129904, + "Ġ×Ķ×Ĺ×ĵ": 137365, + "Ġ×Ķ×Ĺ×ĵש": 131074, + "Ġ×Ķ×Ĺ×ĵש×Ķ": 137366, + "Ġ×Ķ×Ĺ×ķ": 133443, + "Ġ×Ķ×Ĺ×ķ×§": 133500, + "Ġ×Ķ×Ĺ×ķ׾×Ļ×Ŀ": 141956, + "Ġ×Ķ×Ĺ×Ļ": 127030, + "Ġ×Ķ×Ĺ×Ļ×Ļ×Ŀ": 131002, + "Ġ×Ķ×Ĺ׾": 129063, + "Ġ×Ķ×ĺ": 125182, + "Ġ×Ķ×ĺ×ij×¢": 137095, + "Ġ×Ķ×ĺ×ķ×ij": 130786, + "Ġ×Ķ×ĺ×Ļפ×ķ׾": 136797, + "Ġ×Ķ×Ļ": 125037, + "Ġ×Ķ×Ļש": 133740, + "Ġ×Ķ×Ļשר×IJ": 133741, + "Ġ×Ķ×Ļשר×IJ׾": 138573, + "Ġ×Ķ×Ļשר×IJ׾×Ļ": 133742, + "Ġ×Ķ×Ļשר×IJ׾×Ļת": 138574, + "Ġ×Ķ×Ļת": 134812, + "Ġ×Ķ×Ļתר": 134813, + "Ġ×Ķ×Ļת×Ķ": 131288, + "Ġ×Ķ×Ļ×IJ": 128302, + "Ġ×Ķ×Ļ×Ķ": 126215, + "Ġ×Ķ×Ļ×Ķ×ķ×ĵ×Ļ": 134294, + "Ġ×Ķ×Ļ×ķ": 128591, + "Ġ×Ķ×Ļ×ķ×Ŀ": 128795, + "Ġ×Ķ×Ļ×Ĺ×Ļ×ĵ": 134604, + "Ġ×Ķ×Ļ×ĺ×ij": 138671, + "Ġ×Ķ×Ļ×Ļת×Ķ": 129813, + "Ġ×Ķ×Ļ×Ļת×Ļ": 131907, + "Ġ×Ķ×Ļ׾×ĵ": 136011, + "Ġ×Ķ×Ļ׾×ĵ×Ļ×Ŀ": 133256, + "Ġ×Ķ׼": 129586, + "Ġ×Ķ׼ר": 132780, + "Ġ×Ķ׼×ķ׾": 139612, + "Ġ×Ķ׼×ķ׾׾": 139613, + "Ġ×Ķ׼×Ļ": 129092, + "Ġ×Ķ׼׾": 128961, + "Ġ×Ķ×Ľ×ł×¡×ª": 137871, + "Ġ×Ķ׾": 126863, + "Ġ×Ķ׾×IJ×ķ×ŀ×Ļ": 140342, + "Ġ×Ķ׾×ij": 135585, + "Ġ×Ķ׾׾×ķ": 136833, + "Ġ×Ķ×Ŀ": 128414, + "Ġ×Ķ×ŀ": 123919, + "Ġ×Ķ×ŀס": 133262, + "Ġ×Ķ×ŀ×¢": 135072, + "Ġ×Ķ×ŀ×¢×¨×Ľ×ª": 140068, + "Ġ×Ķ×ŀצ": 143925, + "Ġ×Ķ×ŀצ×ij": 138271, + "Ġ×Ķ×ŀצ×Ļ×IJ×ķת": 143926, + "Ġ×Ķ×ŀ×§": 125391, + "Ġ×Ķ×ŀקצ×ķ×¢×Ļ": 143095, + "Ġ×Ķ×ŀ×§×ķר×Ļ": 143553, + "Ġ×Ķ×ŀ×§×ķ×Ŀ": 135233, + "Ġ×Ķ×ŀ×§×ķ×ŀ×Ļ": 139482, + "Ġ×Ķ×ŀר׼×ĸ×Ļ": 137918, + "Ġ×Ķ×ŀש": 124696, + "Ġ×Ķ×ŀשפ": 136598, + "Ġ×Ķ×ŀשפ×Ĺ×Ķ": 136599, + "Ġ×Ķ×ŀשפ×ĺ": 129671, + "Ġ×Ķ×ŀשת": 141980, + "Ġ×Ķ×ŀשת×ŀש": 141981, + "Ġ×Ķ×ŀש×Ĺ×§": 134571, + "Ġ×Ķ×ŀש×Ļ×ij": 142222, + "Ġ×Ķ×ŀת": 129113, + "Ġ×Ķ×ŀ×ijקש": 140751, + "Ġ×Ķ×ŀ×ĵ": 131755, + "Ġ×Ķ×ŀ×ĵ×Ļ׳×Ķ": 131756, + "Ġ×Ķ×ŀ×ķ": 128679, + "Ġ×Ķ×ŀ×ķצר": 135489, + "Ġ×Ķ×ŀ×ķף": 137487, + "Ġ×Ķ×ŀ×Ĺ": 127474, + "Ġ×Ķ×ŀ×Ĺ×Ļר": 137400, + "Ġ×Ķ×ŀ×Ļ": 127111, + "Ġ×Ķ×ŀ×Ļ×ĵ×¢": 135134, + "Ġ×Ķ×ŀ×Ļ×ķ×Ĺ×ĵ": 143801, + "Ġ×Ķ×ŀ׼": 136699, + "Ġ×Ķ×ŀ׾×IJ": 136042, + "Ġ×Ķ×ŀ×ŀש׾×Ķ": 135357, + "Ġ×Ķף": 130032, + "Ġ×Ķ׳": 125448, + "Ġ×Ķ×ł×ª": 133700, + "Ġ×Ķ×ł×ª×ij×¢": 133701, + "Ġ×Ķ׳×IJ": 142177, + "Ġ×Ķ׳×IJש×Ŀ": 142178, + "Ġ×Ķ׳×Ķ": 134078, + "Ġ×Ķ׳×ķ׼": 141089, + "Ġ×Ķ׳×ķ׼×Ĺ×Ļ": 141090, + "Ġ×ķ": 123822, + "Ġ×ķ×¢×ĵ": 130364, + "Ġ×ķ×¢×ķ×ĵ": 129525, + "Ġ×ķ×¢×ľ": 130279, + "Ġ×ķרק": 138435, + "Ġ×ķש": 129231, + "Ġ×ķש׾": 137293, + "Ġ×ķ×IJ": 124416, + "Ġ×ķ×IJ×£": 135199, + "Ġ×ķ×IJפ×Ļ׾×ķ": 139753, + "Ġ×ķ×IJת": 129860, + "Ġ×ķ×IJ×ĸ": 133681, + "Ġ×ķ×IJ×Ĺ": 133253, + "Ġ×ķ×IJ×Ļ": 130468, + "Ġ×ķ×IJ×Ļף": 136030, + "Ġ×ķ×IJ×Ŀ": 134167, + "Ġ×ķ×IJ׳×Ļ": 131426, + "Ġ×ķ×ij": 126205, + "Ġ×ķ×ij׼׾": 138511, + "Ġ×ķ×Ĵ": 136076, + "Ġ×ķ×Ĵ×Ŀ": 129319, + "Ġ×ķ×ĵ": 134594, + "Ġ×ķ×Ķ": 125008, + "Ġ×ķ×Ķת": 133258, + "Ġ×ķ×Ķ×ķ×IJ": 130958, + "Ġ×ķ×Ķ×Ļ×IJ": 135713, + "Ġ×ķ×Ķ×ŀ": 132797, + "Ġ×ķ×ĸ": 140721, + "Ġ×ķ×ĸ×IJת": 140722, + "Ġ×ķ×ĸ×Ķ": 131667, + "Ġ×ķ×Ĺ": 129794, + "Ġ×ķ×Ļ": 128746, + "Ġ×ķ×Ļש": 132320, + "Ġ×ķ×Ļ×ķתר": 143312, + "Ġ×ķ׼": 126485, + "Ġ×ķ׼×ķ": 135209, + "Ġ×ķ׼×Ļ": 135034, + "Ġ×ķ׼×ļ": 137544, + "Ġ×ķ׼׾": 132041, + "Ġ×ķ׼×ŀ×ķ×ijף": 142821, + "Ġ×ķ׼ף": 132362, + "Ġ×ķ׾": 126209, + "Ġ×ķ׾×IJ": 126273, + "Ġ×ķ׾×IJ×Ĺר": 139733, + "Ġ×ķ׾×Ķ": 128950, + "Ġ×ķ׾׼ף": 131601, + "Ġ×ķ×ŀ": 128308, + "Ġ×ķ×ŀת": 133075, + "Ġ×ķ×ŀ×Ķ": 131740, + "Ġ×ķ׳": 128970, + "Ġ×ĸ": 124253, + "Ġ×ĸ×§": 142540, + "Ġ×ĸ×§×ķ×§": 142541, + "Ġ×ĸ×IJת": 128602, + "Ġ×ĸ×Ķ": 126197, + "Ġ×ĸ×Ķ×ķ": 135224, + "Ġ×ĸ×ķ": 128407, + "Ġ×ĸ×ķ׼ר": 141731, + "Ġ×ĸ׼×IJ": 143923, + "Ġ×ĸ×ŀף": 129247, + "Ġ×ĸ×ŀ׳": 142017, + "Ġ×Ĺ": 123852, + "Ġ×Ĺר": 132702, + "Ġ×Ĺש": 133562, + "Ġ×Ĺש×ij×ķף": 133067, + "Ġ×Ĺש×ķ×ij": 129166, + "Ġ×Ĺ×ij": 133614, + "Ġ×Ĺ×ijר": 125200, + "Ġ×Ĺ×ijרת": 129777, + "Ġ×Ĺ×ijר×ķת": 132860, + "Ġ×Ĺ×ijר×Ļ": 136687, + "Ġ×Ĺ×ijר×Ļ×Ŀ": 137859, + "Ġ×Ĺ×Ĵ": 134468, + "Ġ×Ĺ×ĵ": 124894, + "Ġ×Ĺ×ĵר": 132836, + "Ġ×Ĺ×ĵש": 129665, + "Ġ×Ĺ×ĵש×Ķ": 133384, + "Ġ×Ĺ×ĵש×ķת": 130756, + "Ġ×Ĺ×ĵש×Ļ×Ŀ": 132986, + "Ġ×Ĺ×ķ": 126462, + "Ġ×Ĺ×ķפ": 142109, + "Ġ×Ĺ×ķפש×Ļ": 142110, + "Ġ×Ĺ×ķ×§": 130456, + "Ġ×Ĺ×ķש": 131855, + "Ġ×Ĺ×ķש×ij": 131856, + "Ġ×Ĺ×ķ×ij": 135742, + "Ġ×Ĺ×ķ×ĵ": 133716, + "Ġ×Ĺ×ķ×ĵש×Ļ×Ŀ": 133717, + "Ġ×Ĺ×ķ×ķת": 136013, + "Ġ×Ĺ×ķ×ĸ": 136199, + "Ġ×Ĺ×ķ×ĸר": 136200, + "Ġ×Ĺ×ķ׾": 131790, + "Ġ×Ĺ×ĸ": 130941, + "Ġ×Ĺ×Ļ": 124834, + "Ġ×Ĺ×Ļפ×Ķ": 139676, + "Ġ×Ĺ×Ļ×ķ×ij×Ļ": 143523, + "Ġ×Ĺ×Ļ×Ļ": 130721, + "Ġ×Ĺ×Ļ×Ļ×ij": 131689, + "Ġ×Ĺ×Ļ×Ļ×Ŀ": 129842, + "Ġ×Ĺ×Ļ׳×Ŀ": 139717, + "Ġ×Ĺ׾ק": 128905, + "Ġ×Ĺ׾×ķ": 136397, + "Ġ×Ĺ×ŀ": 134490, + "Ġ×ĺ": 124395, + "Ġ×ĺ×ij×¢": 137574, + "Ġ×ĺ×ij×¢×Ļ": 137575, + "Ġ×ĺ×ķ×¢": 143766, + "Ġ×ĺ×ķ×¢×Ł": 143767, + "Ġ×ĺ×ķ×ij": 128571, + "Ġ×ĺ×ķ×ij×Ķ": 131798, + "Ġ×ĺ×Ļ": 135345, + "Ġ×ĺ×Ļפ×ķ׾": 131410, + "Ġ×ĺ×Ļ×ķ׾": 135346, + "Ġ×ĺ׾": 143186, + "Ġ×ĺ׾פ×ķף": 143187, + "Ġ×Ļ": 123872, + "Ġ×Ļ×¢": 140589, + "Ġ×Ļ×¢×§×ij": 140590, + "Ġ×Ļ×¢×ľ": 142570, + "Ġ×Ļצ": 125986, + "Ġ×Ļצ×IJ": 132527, + "Ġ×Ļצ×Ĺ×§": 140424, + "Ġ×Ļצ×Ļרת": 135185, + "Ġ×Ļר×ķ×§": 141354, + "Ġ×Ļש": 125010, + "Ġ×Ļשר×IJ׾": 128458, + "Ġ×Ļש×Ļר": 136231, + "Ġ×Ļש׳×Ŀ": 140508, + "Ġ×Ļתר": 134044, + "Ġ×Ļ×ĵ×Ļ": 128549, + "Ġ×Ļ×Ķ": 125174, + "Ġ×Ļ×Ķ×ķ×ĵ×Ķ": 139915, + "Ġ×Ļ×Ķ×ķ×ĵ×Ļ": 134190, + "Ġ×Ļ×Ķ×Ļ×Ķ": 129503, + "Ġ×Ļ×Ķ×Ļ×ķ": 136873, + "Ġ×Ļ×ķ": 128699, + "Ġ×Ļ×ķסף": 139496, + "Ġ×Ļ×ķצ×IJ": 132443, + "Ġ×Ļ×ķר": 143328, + "Ġ×Ļ×ķרק": 143329, + "Ġ×Ļ×ķתר": 128315, + "Ġ×Ļ×ķ×ĵ": 126826, + "Ġ×Ļ×ķ×ĵ×¢": 130206, + "Ġ×Ļ×ķ×ĵ×¢×Ļ×Ŀ": 136097, + "Ġ×Ļ×ķ׼׾": 136893, + "Ġ×Ļ×ķ×Ŀ": 128915, + "Ġ×Ļ×Ĺ": 132877, + "Ġ×Ļ×Ĺס×Ļ": 132878, + "Ġ×Ļ×Ĺ×ĵ": 131340, + "Ġ×Ļ×Ļת": 142074, + "Ġ×Ļ×Ļ×ª×Ľ×Ł": 142075, + "Ġ×Ļ׼×ķ׾": 125485, + "Ġ×Ļ׼×ķ׾×Ķ": 132509, + "Ġ×Ļ׼×ķ׾×Ļ×Ŀ": 131022, + "Ġ×Ļ׾": 130419, + "Ġ×Ļ׾×ĵ×Ļ×Ŀ": 130420, + "Ġ׼": 123845, + "Ġ×Ľ×¡": 133358, + "Ġ×Ľ×¡×£": 133359, + "Ġ×Ľ×¢×ª": 138384, + "Ġ׼פ×Ļ": 130672, + "Ġ׼ש": 128574, + "Ġ׼ש×Ķ": 132284, + "Ġ×Ľ×ª": 125659, + "Ġ×Ľ×ª×ij": 129416, + "Ġ×Ľ×ª×ķצ×IJ×Ķ": 142471, + "Ġ×Ľ×ª×ķ×ijת": 135108, + "Ġ׼×IJ": 125091, + "Ġ׼×IJשר": 129004, + "Ġ׼×IJ×Ļ׾": 138463, + "Ġ׼×IJ×Ļ׾×ķ": 138464, + "Ġ׼×IJ׾×Ķ": 136436, + "Ġ׼×IJ×ŀ": 138337, + "Ġ׼×IJ×ŀ×ķר": 138338, + "Ġ׼×IJף": 128816, + "Ġ׼×ijר": 128648, + "Ġ׼×Ĵ": 136054, + "Ġ׼×Ĵ×ķף": 136055, + "Ġ׼×ĵ": 132803, + "Ġ׼×ĵ×IJ×Ļ": 132804, + "Ġ׼×ĵ×Ļ": 128413, + "Ġ׼×ķ׾": 126833, + "Ġ׼×ķ׾×ķ": 143460, + "Ġ׼×ķ׾׾": 130230, + "Ġ׼×ĸ×Ķ": 131712, + "Ġ׼×Ļ": 128276, + "Ġ׼×Ļצ": 132126, + "Ġ׼×Ļצ×ĵ": 132127, + "Ġ׼×Ļ×ķ×Ŀ": 132998, + "Ġ׼×ļ": 128397, + "Ġ׼׼×Ķ": 137128, + "Ġ׼׼׾": 133773, + "Ġ׼׾": 124759, + "Ġ׼׾פ×Ļ": 136787, + "Ġ׼׾×ķ×ŀר": 136170, + "Ġ׼׾×Ļ": 135140, + "Ġ׼׾׾": 130476, + "Ġ׼׾׾×Ļ": 134545, + "Ġ׼×ŀ×¢×ĺ": 132241, + "Ġ׼×ŀ×Ķ": 128638, + "Ġ׼×ŀ×ķ": 128381, + "Ġ׼×ŀ×ķ×ijף": 132939, + "Ġ׼ף": 128996, + "Ġ׼׳×Ĵ×ĵ": 140811, + "Ġ׼׳×Ļס": 139859, + "Ġ׼׳×Ļס×Ķ": 139860, + "Ġ׾": 76962, + "Ġ×ľ×¡": 129415, + "Ġ×ľ×¢": 124336, + "Ġ×ľ×¢×©×ķת": 128988, + "Ġ×ľ×¢×ij×ķר": 138972, + "Ġ×ľ×¢×ij×ķ×ĵ": 132076, + "Ġ×ľ×¢×ķ׾×Ŀ": 136006, + "Ġ×ľ×¢×ķ×ŀ": 137131, + "Ġ×ľ×¢×ķ×ŀת": 137132, + "Ġ×ľ×¢×ĸ×ķר": 138413, + "Ġ×ľ×¢×Ļת": 142361, + "Ġ×ľ×¢×Ļת×Ļ×Ŀ": 142362, + "Ġ×ľ×¢×ŀ×ķ×ĵ": 137717, + "Ġ׾פ": 125050, + "Ġ×ľ×¤×¢": 140189, + "Ġ×ľ×¤×¢×ķ׾": 141818, + "Ġ×ľ×¤×¢×ŀ×Ļ×Ŀ": 140190, + "Ġ׾פר": 143030, + "Ġ׾פר×ĺ×Ļ×Ŀ": 143031, + "Ġ×ľ×¤×ª": 143647, + "Ġ×ľ×¤×ª×ķ×Ĺ": 143648, + "Ġ׾פ×Ĺ×ķת": 134112, + "Ġ׾פ×Ļ": 128887, + "Ġ׾פ׳×ķת": 137988, + "Ġ׾פ׳×Ļ": 128519, + "Ġ׾צ": 124899, + "Ġ׾צפ": 143980, + "Ġ׾צפ×Ļ×Ķ": 143981, + "Ġ׾צ×IJת": 133343, + "Ġ׾צ×ĵ": 136050, + "Ġ׾צ×ķר×ļ": 137340, + "Ġ׾צ×Ļ×Ļף": 140066, + "Ġ׾ק": 124557, + "Ġ׾קר": 133174, + "Ġ׾קר×IJ": 136974, + "Ġ׾קר×IJת": 136975, + "Ġ׾קר×ķ×IJ": 133175, + "Ġ׾ק×ij׾": 126662, + "Ġ׾ק×ij×ľ×ª": 138330, + "Ġ׾ק×ķ×Ĺ": 135106, + "Ġ׾ק×ķ×Ĺ×ķת": 135107, + "Ġ׾ק×Ĺת": 134503, + "Ġ׾ק׳×ķת": 138794, + "Ġ׾ר": 129269, + "Ġ׾ר×IJ": 130257, + "Ġ׾ר×IJש": 142571, + "Ġ׾ר×IJש×ķ׳×Ķ": 142572, + "Ġ׾ר×IJ×ķת": 130258, + "Ġ׾ר׼": 140503, + "Ġ׾ר׼×ķש": 140504, + "Ġ׾ש": 126221, + "Ġ׾ש׾×Ŀ": 139777, + "Ġ׾ש×Ŀ": 136688, + "Ġ׾ש×ŀ": 127856, + "Ġ׾ש×ŀ×ķ×¢": 141730, + "Ġ׾ש×ŀ×ķר": 136519, + "Ġ׾ש׳×ķת": 133977, + "Ġ×ľ×ª": 126329, + "Ġ×ľ×ª×ª": 133211, + "Ġ׾×IJ": 124597, + "Ġ׾×IJתר": 133060, + "Ġ׾×IJ×ķר": 135600, + "Ġ׾×IJ×ķר×ļ": 134345, + "Ġ׾×IJ×ķ×ŀ×Ļ": 137394, + "Ġ׾×IJ×Ĺר": 126304, + "Ġ׾×IJ×Ĺר×ķ׳×Ķ": 139745, + "Ġ׾×ij": 124804, + "Ġ׾×ijצע": 134531, + "Ġ׾×ij×ĵ": 128007, + "Ġ׾×ij×ĵ×ķ×§": 138647, + "Ġ׾×ij×Ĺ": 134461, + "Ġ׾×ij×Ĺ×ķר": 134462, + "Ġ׾×ij×ĺ׾": 140291, + "Ġ׾×ij×Ļת": 131979, + "Ġ׾×ij×Ļף": 133342, + "Ġ׾×Ĵ": 127037, + "Ġ׾×Ĵר": 141832, + "Ġ׾×Ĵר×ķ×Ŀ": 141833, + "Ġ׾×Ĵ×ij×Ļ": 131044, + "Ġ׾×Ĵ×ŀר×Ļ": 138755, + "Ġ׾×ĵ": 126790, + "Ġ׾×ĵעת": 130073, + "Ġ׾×ĵ×ijר": 132788, + "Ġ׾×Ķ": 124007, + "Ġ׾×Ķ×¢×ij×Ļר": 140069, + "Ġ׾×Ķ×§": 134559, + "Ġ׾×Ķש": 127631, + "Ġ׾×Ķשת×ŀש": 133640, + "Ġ׾×Ķש×Ļ×Ĵ": 140871, + "Ġ׾×Ķת": 126278, + "Ġ׾×Ķת×ŀ×ķ×ĵ": 140825, + "Ġ׾×Ķת×ŀ×ķ×ĵ×ĵ": 140826, + "Ġ׾×Ķ×ij×Ļ×IJ": 137960, + "Ġ׾×Ķ×ij×Ļף": 134460, + "Ġ׾×Ķ×Ĵ×Ļ×¢": 131147, + "Ġ׾×Ķ×Ĵ×Ļ×ĵ": 141325, + "Ġ׾×Ķ×ķס×Ļ×£": 139597, + "Ġ׾×Ķ×Ļ": 129237, + "Ġ׾×Ķ×Ļ×ķת": 128535, + "Ġ׾×Ķ×Ļ׼": 141857, + "Ġ׾×Ķ×Ļ×Ľ×ł×¡": 141858, + "Ġ׾×Ķ×Ŀ": 129354, + "Ġ׾×Ķ×ŀש×Ļ×ļ": 137993, + "Ġ׾×ķ": 128398, + "Ġ׾×ķ×§": 136951, + "Ġ׾×ķ×§×Ĺ": 136952, + "Ġ׾×ķ×ŀר": 135928, + "Ġ׾×ĸ": 133966, + "Ġ׾×ĸ×Ķ": 134560, + "Ġ׾×Ĺ": 125035, + "Ġ׾×Ĺצ×ķ": 137044, + "Ġ׾×Ĺ×ijר": 132616, + "Ġ׾×Ĺ×ķ×§": 139070, + "Ġ׾×Ĺ׾": 139940, + "Ġ׾×Ĺ׾×ķ×ĺ": 139941, + "Ġ׾×Ĺ׾×ķ×ĺ×Ļף": 139942, + "Ġ׾×ĺ": 127091, + "Ġ׾×ĺ×ķ×ijת": 142469, + "Ġ׾×Ļ": 128298, + "Ġ׾×Ļצ×ķר": 135181, + "Ġ׾×Ļשר×IJ׾": 139378, + "Ġ׾×Ļ×ĵ": 136842, + "Ġ׾×Ļ×ĵ×Ļ": 139017, + "Ġ׾×Ļ׾×ĵ×Ļ×Ŀ": 135884, + "Ġ׾×Ļ×ŀ": 137479, + "Ġ׾×Ļ×ŀ×ķ×ĵ×Ļ": 137480, + "Ġ׾×ļ": 128879, + "Ġ׾׼": 131030, + "Ġ×ľ×Ľ×ª×ķ×ij": 135461, + "Ġ׾׼×ļ": 131442, + "Ġ׾׼׾": 128710, + "Ġ׾׼×Ŀ": 128873, + "Ġ׾׼ף": 132255, + "Ġ׾׾": 134866, + "Ġ׾׾×IJ": 128995, + "Ġ×ľ×ľ×Ľ×ª": 139105, + "Ġ׾׾×ŀ×ķ×ĵ": 134867, + "Ġ׾×ŀ": 128359, + "Ġ׾×ŀ×¢": 126116, + "Ġ׾×ŀעש×Ķ": 135764, + "Ġ׾×ŀ×¢×ľ×Ķ": 138103, + "Ġ׾×ŀ×¢×Ł": 139931, + "Ġ׾×ŀצ": 131143, + "Ġ׾×ŀצ×ķ×IJ": 131144, + "Ġ׾×ŀ×§×ķ×Ŀ": 139763, + "Ġ׾×ŀר": 132193, + "Ġ׾×ŀר×ķת": 132194, + "Ġ׾×ŀש׾": 132967, + "Ġ׾×ŀ×Ķ": 129783, + "Ġ׾×ŀ×Ļ": 131811, + "Ġ׾×ŀ׳": 140361, + "Ġ׾×ŀ׳×ķ×¢": 140362, + "Ġ׾׳": 130735, + "Ġ×ľ×ł×¡": 142379, + "Ġ×ľ×ł×¡×ķת": 142380, + "Ġ׾׳×ķ": 128697, + "Ġ×ŀ": 91511, + "Ġ×ŀס": 124816, + "Ġ×ŀספר": 128933, + "Ġ×ŀספ×Ļ×§": 134983, + "Ġ×ŀס×ķ": 129475, + "Ġ×ŀס×ķ×Ĵ": 137041, + "Ġ×ŀ×¢": 125128, + "Ġ×ŀ×¢×¨×Ľ×ª": 131238, + "Ġ×ŀ×¢×ķ׾×Ķ": 139766, + "Ġ×ŀ×¢×ĺ": 131439, + "Ġ×ŀ×¢×ľ": 131331, + "Ġ×ŀפת×Ĺ": 142273, + "Ġ×ŀפ׳×Ļ": 137688, + "Ġ×ŀצ": 125348, + "Ġ×ŀצ×ij": 131295, + "Ġ×ŀצ×ĵ": 133572, + "Ġ×ŀצ×ķ×ķ×Ķ": 143653, + "Ġ×ŀצ×ķ×Ļ": 132843, + "Ġ×ŀצ×Ļ×¢": 135007, + "Ġ×ŀ×§": 124679, + "Ġ×ŀקצ×ķ×¢×Ļ": 131221, + "Ġ×ŀקר×Ķ": 138785, + "Ġ×ŀ×§×ij׾": 133030, + "Ġ×ŀ×§×ķר": 130172, + "Ġ×ŀ×§×ķ×Ŀ": 129872, + "Ġ×ŀ×§×ķ×ŀ×Ļ": 140543, + "Ġ×ŀר": 129001, + "Ġ×ŀר×IJש": 135330, + "Ġ×ŀר×Ĵ×Ļש": 137248, + "Ġ×ŀר׼×ĸ×Ļ": 138904, + "Ġ×ŀש": 124399, + "Ġ×ŀשפ": 136213, + "Ġ×ŀשפ×ĺ×Ļ": 136214, + "Ġ×ŀשר": 131465, + "Ġ×ŀשר×ĵ": 131466, + "Ġ×ŀשת": 132431, + "Ġ×ŀשת×ŀש": 132252, + "Ġ×ŀש×Ķ": 135550, + "Ġ×ŀש×Ķ×ķ": 130186, + "Ġ×ŀש×ķ": 132464, + "Ġ×ŀש×ķ×Ŀ": 136531, + "Ġ×ŀש×Ĺ×§": 128101, + "Ġ×ŀש×Ĺ×§×Ļ": 139966, + "Ġ×ŀש×Ĺ×§×Ļ×Ŀ": 141199, + "Ġ×ŀש×ŀ×¢×ķת": 134124, + "Ġ×ŀש×ŀ×¢×ķת×Ļ": 134125, + "Ġ×ŀש׳×Ķ": 137603, + "Ġ×ŀת": 125435, + "Ġ×ŀת×IJ×Ļ×Ŀ": 137364, + "Ġ×ŀת×ķ×ļ": 130568, + "Ġ×ŀת×Ĺת": 138810, + "Ġ×ŀת×Ĺ×Ļ׾": 133635, + "Ġ×ŀ×IJ": 124270, + "Ġ×ŀ×IJפשר": 134573, + "Ġ×ŀ×IJשר": 134901, + "Ġ×ŀ×IJת": 133136, + "Ġ×ŀ×IJ×ĵ": 132396, + "Ġ×ŀ×IJ×ķ": 135880, + "Ġ×ŀ×IJ×ķת": 136964, + "Ġ×ŀ×IJ×ķ×ĵ": 128576, + "Ġ×ŀ×IJ×ĸ": 132570, + "Ġ×ŀ×IJ×Ĺ": 140143, + "Ġ×ŀ×IJ×Ĺ×ķר×Ļ": 140144, + "Ġ×ŀ×IJ×ŀר×Ļ×Ŀ": 134940, + "Ġ×ŀ×ij": 125145, + "Ġ×ŀ×ij×ķס": 141971, + "Ġ×ŀ×ij×ķסס": 141972, + "Ġ×ŀ×ij×Ĺ": 137509, + "Ġ×ŀ×ij×Ĺ×Ļ×ł×ª": 137510, + "Ġ×ŀ×ij×Ļף": 137733, + "Ġ×ŀ×ij׾×Ļ": 139315, + "Ġ×ŀ×Ĵ": 125656, + "Ġ×ŀ×Ĵ×Ļ×¢": 130993, + "Ġ×ŀ×Ĵ×Ļ×ij": 140151, + "Ġ×ŀ×ĵ": 126878, + "Ġ×ŀ×ĵ×Ķ": 142316, + "Ġ×ŀ×ĵ×Ķ×Ļ×Ŀ": 142317, + "Ġ×ŀ×ĵ×ķ×¢": 137140, + "Ġ×ŀ×ĵ×ķ×ijר": 130410, + "Ġ×ŀ×ĵ×Ļ": 129667, + "Ġ×ŀ×Ķ": 124757, + "Ġ×ŀ×Ķ×ķ×ķ×Ķ": 139863, + "Ġ×ŀ×Ķ×Ŀ": 132691, + "Ġ×ŀ×Ķ×ŀ": 135112, + "Ġ×ŀ×Ķ×ŀ×¢": 142997, + "Ġ×ŀ×Ķ×ŀ×¢×¨×Ľ×ª": 142998, + "Ġ×ŀ×ķ": 128299, + "Ġ×ŀ×ķצר": 135739, + "Ġ×ŀ×ķצר×Ļ×Ŀ": 137040, + "Ġ×ŀ×ķר": 139266, + "Ġ×ŀ×ķר׼×ij": 139267, + "Ġ×ŀ×ķ׼ר": 137902, + "Ġ×ŀ×ķ׾": 130385, + "Ġ×ŀ×ĸ×Ķ": 133181, + "Ġ×ŀ×Ĺ": 125540, + "Ġ×ŀ×Ĺפש": 135604, + "Ġ×ŀ×Ĺ×ĵש": 136416, + "Ġ×ŀ×Ĺ×Ļר": 131287, + "Ġ×ŀ×Ļ": 126211, + "Ġ×ŀ×Ļש×Ķ": 136859, + "Ġ×ŀ×Ļש×Ķ×ķ": 136860, + "Ġ×ŀ×Ļ×ĵ": 133265, + "Ġ×ŀ×Ļ×ĵ×¢": 129886, + "Ġ×ŀ×Ļ×ķ×Ĺ×ĵ": 130212, + "Ġ×ŀ×Ļ׾×Ļ": 134533, + "Ġ×ŀ×Ļ׾×Ļ×ķף": 134534, + "Ġ×ŀ׼": 128989, + "Ġ×ŀ׼׾": 130611, + "Ġ×ŀ׼ף": 135598, + "Ġ×ŀ׾": 129160, + "Ġ×ŀ׾×IJ": 129661, + "Ġ×ŀ×ŀ": 126573, + "Ġ×ŀ×ŀש": 129098, + "Ġ×ŀ×ŀ׳×ķ": 133597, + "Ġ×ŀף": 130875, + "Ġ×ŀ׳": 131381, + "Ġ×ŀ×ł×ª": 130216, + "Ġ׳": 123901, + "Ġ×ł×¢": 126566, + "Ġ×ł×¢×©×Ķ": 139990, + "Ġ׳ר×IJ×Ķ": 130708, + "Ġ׳ש": 125929, + "Ġ׳ש×IJר": 139532, + "Ġ׳ש×ŀ×¢": 141278, + "Ġ×ł×ª": 126545, + "Ġ×ł×ª×ķ׳×Ļ×Ŀ": 142627, + "Ġ׳×Ĵ×ĵ": 130152, + "Ġ׳×Ĵ×Ļש": 140897, + "Ġ׳×Ĵ×Ļש×ķת": 140898, + "Ġ׳×ĵר": 141193, + "Ġ׳×ĵרש": 141194, + "Ġ׳×Ķ": 128141, + "Ġ׳×Ķ×ĵר": 140921, + "Ġ׳×Ķ׳": 141886, + "Ġ׳×ķס": 125266, + "Ġ׳×ķסף": 131988, + "Ġ׳×ķספת": 136181, + "Ġ׳×ķספ×ķת": 136993, + "Ġ׳×ķספ×Ļ×Ŀ": 130614, + "Ġ׳×ķת": 143463, + "Ġ׳×ķ×ª×Ł": 143464, + "Ġ׳×ķ׾×ĵ": 140556, + "Ġ׳×Ĺש×ij": 140109, + "Ġ׳×Ļת": 128694, + "Ġ׳×Ļ×ª×Ł": 128695, + "Ġ׳×Ļ×Ķ": 135813, + "Ġ׳×Ļ×Ķ×ķ׾": 135814, + "Ġ׳׼": 127096, + "Ġ×ł×Ľ×ª×ij": 140549, + "Ġ׳׼×ķף": 131348, + "Ġ׳׼×ķ׳×Ķ": 143006, + "Ġ×ł×Ľ×ł×¡": 135133, + "Ġ׳×ŀצ×IJ": 129715, + "ĠØ": 8803, + "ĠØ¢": 125108, + "Ġآخر": 129197, + "ĠØ¢ÙĦ": 130484, + "ĠØ¢ÙĦاÙģ": 142525, + "ĠØ£": 63415, + "Ġأبر": 137810, + "ĠأبرÙĬÙĦ": 137811, + "ĠأبÙĪ": 130261, + "ĠأبÙĬ": 133423, + "ĠأثÙĨ": 134016, + "ĠأثÙĨاء": 134017, + "Ġأج": 135274, + "ĠأجÙĦ": 130214, + "ĠأخرÙī": 129446, + "ĠأخÙĬ": 142439, + "Ġأد": 131549, + "Ġأس": 128962, + "Ġأساس": 133122, + "ĠأسبÙĪØ¹": 142569, + "ĠأشÙĩر": 137270, + "ĠأصبØŃ": 133455, + "ĠأصØŃاب": 139040, + "Ġأع": 131694, + "ĠأعÙĦÙĨ": 134120, + "ĠأعÙĦÙī": 137992, + "ĠأعÙħاÙĦ": 133519, + "Ġأغ": 138244, + "Ġأغسط": 138245, + "Ġأغسطس": 138246, + "ĠØ£ØŃ": 132562, + "ĠØ£ØŃد": 129289, + "ĠØ£ØŃÙħد": 131118, + "ĠØ£Ùģ": 142922, + "ĠØ£Ù쨱ÙĬ": 142923, + "ĠØ£Ù쨱ÙĬÙĤÙĬ": 142924, + "ĠØ£Ù쨱ÙĬÙĤÙĬا": 142925, + "ĠØ£Ù쨶ÙĦ": 130259, + "ĠØ£ÙĤÙĦ": 135761, + "ĠØ£Ùĥبر": 132291, + "ĠØ£Ùĥثر": 128773, + "ĠØ£Ùĥد": 134415, + "ĠØ£ÙĦ": 134499, + "ĠØ£ÙĦÙģ": 131261, + "ĠØ£Ùħ": 126418, + "ĠØ£Ùħا": 130639, + "ĠØ£ÙħاÙħ": 130063, + "ĠØ£Ùħر": 130656, + "ĠØ£ÙħرÙĬ": 140019, + "ĠØ£ÙħرÙĬÙĥÙĬ": 140020, + "ĠØ£Ùħس": 135311, + "ĠØ£ÙĨ": 125007, + "ĠØ£ÙĨا": 132823, + "ĠØ£ÙĨØŃ": 143399, + "ĠØ£ÙĨØŃاء": 143400, + "ĠØ£ÙĨÙĩ": 128636, + "ĠØ£ÙĨÙĩا": 130543, + "ĠØ£ÙĨÙij": 133432, + "ĠØ£ÙĩÙħ": 133453, + "ĠØ£ÙĪ": 128264, + "ĠØ£ÙĪÙĦ": 129445, + "ĠØ£ÙĬ": 124793, + "ĠØ£ÙĬاÙħ": 133674, + "ĠØ£ÙĬض": 138194, + "ĠØ£ÙĬضا": 126899, + "ĠØ£ÙĬضاÙĭ": 133525, + "ĠØ£ÙĬضÙĭا": 138195, + "ĠØ£Ùİ": 131422, + "ĠØ¥": 85153, + "Ġإدارة": 132113, + "Ġإذ": 131649, + "Ġإذا": 128754, + "Ġإطار": 136320, + "ĠإطÙĦاÙĤ": 139789, + "Ġإعادة": 136950, + "ĠØ¥ØŃ": 133341, + "ĠØ¥ÙĦا": 128941, + "ĠØ¥ÙĦÙī": 128252, + "ĠØ¥ÙĦÙĬ": 126000, + "ĠØ¥ÙĦÙĬÙĩ": 132674, + "ĠØ¥ÙĦÙĬÙĩا": 136709, + "ĠØ¥ÙĨ": 128342, + "ĠØ¥ÙĨÙĩ": 133534, + "ĠØ¥ÙĬ": 132170, + "Ġا": 12961, + "ĠابÙĨ": 130016, + "ĠابÙĪ": 141533, + "Ġاخت": 136837, + "ĠاختÙĬار": 136838, + "Ġاذا": 137756, + "Ġاست": 93153, + "ĠاستخداÙħ": 130322, + "ĠاستراتÙĬج": 143024, + "ĠاستراتÙĬجÙĬØ©": 143025, + "ĠاسÙħ": 131198, + "Ġاعت": 134787, + "ĠاØŃد": 139078, + "ĠاÙĥثر": 140165, + "ĠاÙĦ": 17166, + "ĠاÙĦØ¢": 125700, + "ĠاÙĦآخر": 132491, + "ĠاÙĦØ¢ÙĨ": 129968, + "ĠاÙĦØ£": 123877, + "ĠاÙĦأب": 136466, + "ĠاÙĦأج": 140223, + "ĠاÙĦأجÙĨب": 140224, + "ĠاÙĦأجÙĨبÙĬØ©": 140225, + "ĠاÙĦأخ": 134514, + "ĠاÙĦأخرÙī": 132893, + "ĠاÙĦأخÙĬر": 140117, + "ĠاÙĦأخÙĬرة": 134515, + "ĠاÙĦأربع": 134955, + "ĠاÙĦأربعاء": 134956, + "ĠاÙĦأرض": 132093, + "ĠاÙĦأس": 139089, + "ĠاÙĦأساس": 139743, + "ĠاÙĦأساسÙĬØ©": 139744, + "ĠاÙĦأسبÙĪØ¹": 138147, + "ĠاÙĦأسÙĩÙħ": 139090, + "ĠاÙĦأطÙ쨧ÙĦ": 134668, + "ĠاÙĦأعÙħاÙĦ": 133798, + "ĠاÙĦØ£ØŃد": 134967, + "ĠاÙĦØ£Ùĥثر": 138208, + "ĠاÙĦØ£ÙĦÙħ": 142734, + "ĠاÙĦØ£ÙĦÙħاÙĨÙĬ": 142735, + "ĠاÙĦØ£Ùħ": 124695, + "ĠاÙĦØ£Ùħر": 130208, + "ĠاÙĦØ£ÙħرÙĬ": 127482, + "ĠاÙĦØ£ÙħرÙĬÙĥ": 133492, + "ĠاÙĦØ£ÙħرÙĬÙĥÙĬ": 136760, + "ĠاÙĦØ£ÙħرÙĬÙĥÙĬØ©": 133493, + "ĠاÙĦØ£ÙħÙħ": 133906, + "ĠاÙĦØ£ÙħÙĨ": 130496, + "ĠاÙĦØ£ÙħÙĬر": 127640, + "ĠاÙĦØ£ÙħÙĬرÙĥÙĬ": 134575, + "ĠاÙĦØ£ÙĨ": 131714, + "ĠاÙĦØ£Ùĩ": 139911, + "ĠاÙĦØ£ÙĩÙĦÙĬ": 139912, + "ĠاÙĦØ£ÙĪ": 136259, + "ĠاÙĦØ£ÙĪØ±ÙĪ": 135264, + "ĠاÙĦØ£ÙĪØ±ÙĪØ¨ÙĬ": 135265, + "ĠاÙĦØ£ÙĪØ³Ø·": 139870, + "ĠاÙĦØ£ÙĪÙĦ": 126506, + "ĠاÙĦØ£ÙĪÙĦÙī": 130528, + "ĠاÙĦØ£ÙĬ": 141821, + "ĠاÙĦØ£ÙĬاÙħ": 141822, + "ĠاÙĦØ¥": 124058, + "ĠاÙĦإثÙĨ": 142585, + "ĠاÙĦإثÙĨÙĬÙĨ": 142586, + "ĠاÙĦإدارة": 137740, + "ĠاÙĦإرÙĩاب": 134562, + "ĠاÙĦإسرائÙĬÙĦ": 134442, + "ĠاÙĦإسرائÙĬÙĦÙĬ": 134443, + "ĠاÙĦإسÙĦاÙħ": 126051, + "ĠاÙĦإسÙĦاÙħÙĬ": 138224, + "ĠاÙĦإسÙĦاÙħÙĬØ©": 134318, + "ĠاÙĦإعÙĦاÙħ": 132919, + "ĠاÙĦØ¥ÙĦÙĥ": 134153, + "ĠاÙĦØ¥ÙĦÙĥترÙĪ": 134154, + "ĠاÙĦØ¥ÙĦÙĥترÙĪÙĨÙĬ": 134155, + "ĠاÙĦØ¥Ùħ": 133927, + "ĠاÙĦØ¥Ùħارات": 133928, + "ĠاÙĦØ¥ÙħاÙħ": 139528, + "ĠاÙĦØ¥ÙĨ": 127630, + "ĠاÙĦØ¥ÙĨترÙĨت": 137564, + "ĠاÙĦØ¥ÙĨس": 129823, + "ĠاÙĦØ¥ÙĨساÙĨ": 129824, + "ĠاÙĦØ¥ÙĬر": 137559, + "ĠاÙĦØ¥ÙĬراÙĨÙĬ": 137560, + "ĠاÙĦا": 125434, + "ĠاÙĦاتØŃاد": 131534, + "ĠاÙĦاتÙ쨧ÙĤ": 137484, + "ĠاÙĦاث": 139418, + "ĠاÙĦاثÙĨÙĬÙĨ": 139419, + "ĠاÙĦاجتÙħاع": 127696, + "ĠاÙĦاجتÙħاعÙĬ": 134934, + "ĠاÙĦاجتÙħاعÙĬØ©": 136355, + "ĠاÙĦاخ": 138311, + "ĠاÙĦاخبار": 138312, + "ĠاÙĦاس": 139558, + "ĠاÙĦاست": 126455, + "ĠاÙĦاستث": 136265, + "ĠاÙĦاستثÙħار": 136266, + "ĠاÙĦاسÙĦاÙħ": 139559, + "ĠاÙĦاØŃت": 138079, + "ĠاÙĦاØŃتÙĦاÙĦ": 138080, + "ĠاÙĦاÙĤتصاد": 127831, + "ĠاÙĦاÙĤتصادÙĬ": 139691, + "ĠاÙĦاÙĤتصادÙĬØ©": 136251, + "ĠاÙĦاÙĨ": 125696, + "ĠاÙĦاÙĨتخاب": 138751, + "ĠاÙĦاÙĨتخابات": 138752, + "ĠاÙĦاÙĨترÙĨت": 141962, + "ĠاÙĦاÙĨساÙĨ": 143088, + "ĠاÙĦاÙĪÙĦ": 136447, + "ĠاÙĦب": 126249, + "ĠاÙĦبر": 125332, + "ĠاÙĦبرÙĦÙħ": 138578, + "ĠاÙĦبرÙĦÙħاÙĨ": 138579, + "ĠاÙĦبرÙĨاÙħج": 134966, + "ĠاÙĦبرÙĬØ·": 139773, + "ĠاÙĦبرÙĬطاÙĨÙĬ": 139774, + "ĠاÙĦبشر": 132012, + "ĠاÙĦبط": 132624, + "ĠاÙĦبع": 137646, + "ĠاÙĦبعض": 137647, + "ĠاÙĦبØŃØ«": 132764, + "ĠاÙĦبØŃر": 131010, + "ĠاÙĦبÙĦ": 126926, + "ĠاÙĦبÙĦاد": 133483, + "ĠاÙĦبÙĦد": 130596, + "ĠاÙĦبÙĨ": 140048, + "ĠاÙĦبÙĨÙĥ": 140049, + "ĠاÙĦبÙĬ": 125925, + "ĠاÙĦبÙĬاÙĨ": 138749, + "ĠاÙĦبÙĬاÙĨات": 138750, + "ĠاÙĦبÙĬت": 138821, + "ĠاÙĦت": 94957, + "ĠاÙĦتاب": 142351, + "ĠاÙĦتابعة": 142352, + "ĠاÙĦتارÙĬØ®": 132450, + "ĠاÙĦتاÙĦÙĬ": 132701, + "ĠاÙĦتج": 126058, + "ĠاÙĦتجارة": 137907, + "ĠاÙĦتجارÙĬØ©": 136411, + "ĠاÙĦتداÙĪÙĦ": 132863, + "ĠاÙĦتربÙĬØ©": 139317, + "ĠاÙĦترÙĥ": 135542, + "ĠاÙĦترÙĥÙĬ": 135543, + "ĠاÙĦتش": 136113, + "ĠاÙĦتط": 137868, + "ĠاÙĦتطبÙĬÙĤ": 137869, + "ĠاÙĦتع": 126901, + "ĠاÙĦتعا": 136353, + "ĠاÙĦتعاÙĪÙĨ": 136354, + "ĠاÙĦتعÙĦÙĬÙħ": 130487, + "ĠاÙĦتÙĤ": 136524, + "ĠاÙĦتÙĨ": 128052, + "ĠاÙĦتÙĨÙģÙĬ": 140597, + "ĠاÙĦتÙĨÙģÙĬذÙĬ": 140598, + "ĠاÙĦتÙĨÙħÙĬØ©": 139391, + "ĠاÙĦتÙĪØ§ØµÙĦ": 135203, + "ĠاÙĦتÙĪÙĨ": 143825, + "ĠاÙĦتÙĪÙĨسÙĬ": 143826, + "ĠاÙĦتÙī": 131846, + "ĠاÙĦتÙĬ": 128261, + "ĠاÙĦØ«": 124437, + "ĠاÙĦثاÙĦ": 132606, + "ĠاÙĦثاÙĦØ«": 132607, + "ĠاÙĦثاÙĨÙĬ": 130277, + "ĠاÙĦثاÙĨÙĬØ©": 133403, + "ĠاÙĦØ«ÙĤ": 139575, + "ĠاÙĦØ«ÙĤاÙģÙĬ": 139576, + "ĠاÙĦØ«ÙĦاث": 127560, + "ĠاÙĦØ«ÙĦاثاء": 133982, + "ĠاÙĦØ«ÙĨ": 136489, + "ĠاÙĦØ«ÙĨائÙĬØ©": 136490, + "ĠاÙĦج": 124012, + "ĠاÙĦجارÙĬ": 141048, + "ĠاÙĦجاÙħعة": 138695, + "ĠاÙĦجاÙĨب": 138029, + "ĠاÙĦجدÙĬد": 131424, + "ĠاÙĦجدÙĬدة": 132061, + "ĠاÙĦجز": 141912, + "ĠاÙĦجزائ": 141913, + "ĠاÙĦجزائرÙĬ": 141914, + "ĠاÙĦجسÙħ": 137617, + "ĠاÙĦجÙħع": 133322, + "ĠاÙĦجÙħعة": 133323, + "ĠاÙĦجÙħÙĩ": 140779, + "ĠاÙĦجÙħÙĩÙĪØ±": 140780, + "ĠاÙĦجÙħÙĬع": 136656, + "ĠاÙĦجÙĨ": 127214, + "ĠاÙĦجÙĨس": 136194, + "ĠاÙĦجÙĨÙĪØ¨": 135090, + "ĠاÙĦجÙĩ": 142980, + "ĠاÙĦجÙĩا": 138808, + "ĠاÙĦجÙĩات": 142981, + "ĠاÙĦجÙĩاز": 138809, + "ĠاÙĦجÙĪ": 130156, + "ĠاÙĦجÙĬØ´": 133291, + "ĠاÙĦØ®": 124147, + "ĠاÙĦخار": 133460, + "ĠاÙĦخارج": 133461, + "ĠاÙĦخارجÙĬØ©": 133462, + "ĠاÙĦخاص": 131216, + "ĠاÙĦخاصة": 130245, + "ĠاÙĦخاÙħ": 139306, + "ĠاÙĦخاÙħس": 139307, + "ĠاÙĦخبر": 131997, + "ĠاÙĦخدÙħات": 135744, + "ĠاÙĦخط": 130648, + "ĠاÙĦØ®ÙĦ": 133445, + "ĠاÙĦØ®ÙĦÙĬ": 136495, + "ĠاÙĦØ®ÙĦÙĬج": 136496, + "ĠاÙĦØ®Ùħ": 133448, + "ĠاÙĦØ®ÙħÙĬس": 133449, + "ĠاÙĦØ®ÙĬ": 134326, + "ĠاÙĦØ®ÙĬار": 141939, + "ĠاÙĦØ®ÙĬارات": 134327, + "ĠاÙĦد": 124220, + "ĠاÙĦداخÙĦ": 135275, + "ĠاÙĦداخÙĦÙĬØ©": 135276, + "ĠاÙĦدر": 143760, + "ĠاÙĦدراسÙĬ": 143761, + "ĠاÙĦدÙģØ§Ø¹": 135988, + "ĠاÙĦدÙĥت": 131203, + "ĠاÙĦدÙĥتÙĪØ±": 131204, + "ĠاÙĦدÙħ": 132033, + "ĠاÙĦدÙĨÙĬ": 141188, + "ĠاÙĦدÙĨÙĬا": 141189, + "ĠاÙĦدÙĪØ±ÙĬ": 137763, + "ĠاÙĦدÙĪÙĦ": 125702, + "ĠاÙĦدÙĪÙĦØ©": 129814, + "ĠاÙĦدÙĪÙĦÙĬ": 130924, + "ĠاÙĦدÙĪÙĦÙĬØ©": 131936, + "ĠاÙĦدÙĬÙħÙĤرا": 143849, + "ĠاÙĦدÙĬÙħÙĤراط": 143850, + "ĠاÙĦدÙĬÙħÙĤراطÙĬØ©": 143851, + "ĠاÙĦدÙĬÙĨ": 129950, + "ĠاÙĦذ": 125621, + "ĠاÙĦذÙĩب": 136421, + "ĠاÙĦذÙī": 137453, + "ĠاÙĦذÙĬ": 128307, + "ĠاÙĦذÙĬÙĨ": 129174, + "ĠاÙĦر": 124269, + "ĠاÙĦرئÙĬس": 126543, + "ĠاÙĦرئÙĬسÙĬØ©": 130649, + "ĠاÙĦراب": 137059, + "ĠاÙĦرابع": 137060, + "ĠاÙĦرجاÙĦ": 142841, + "ĠاÙĦرجÙĦ": 137153, + "ĠاÙĦرسÙħÙĬ": 134862, + "ĠاÙĦرغÙħ": 139066, + "ĠاÙĦرØŃ": 142648, + "ĠاÙĦرØŃÙħÙĨ": 142649, + "ĠاÙĦرÙĤÙħ": 140302, + "ĠاÙĦرÙĪ": 127763, + "ĠاÙĦرÙĪØ³ÙĬ": 137561, + "ĠاÙĦرÙĬاض": 130918, + "ĠاÙĦز": 125572, + "ĠاÙĦزÙħÙĨ": 141364, + "ĠاÙĦزÙĪØ¬": 139155, + "ĠاÙĦس": 123913, + "ĠاÙĦساب": 125923, + "ĠاÙĦسابع": 140844, + "ĠاÙĦسابÙĤ": 131438, + "ĠاÙĦسابÙĤØ©": 138743, + "ĠاÙĦساد": 141395, + "ĠاÙĦسادس": 141396, + "ĠاÙĦساعة": 131643, + "ĠاÙĦسبت": 134432, + "ĠاÙĦسعÙĪØ¯": 126787, + "ĠاÙĦسعÙĪØ¯ÙĬ": 134661, + "ĠاÙĦسعÙĪØ¯ÙĬØ©": 130047, + "ĠاÙĦسÙĦ": 125273, + "ĠاÙĦسÙĦاÙħ": 130781, + "ĠاÙĦسÙĦØ·": 141404, + "ĠاÙĦسÙĦطات": 141405, + "ĠاÙĦسÙĦطة": 137024, + "ĠاÙĦسÙħ": 132436, + "ĠاÙĦسÙĨ": 127538, + "ĠاÙĦسÙĨØ©": 136272, + "ĠاÙĦسÙĨÙĪØ§Øª": 143551, + "ĠاÙĦسÙĪØ¯": 134040, + "ĠاÙĦسÙĪØ¯Ø§ÙĨ": 134041, + "ĠاÙĦسÙĪØ±ÙĬ": 133568, + "ĠاÙĦسÙĪØ±ÙĬØ©": 137130, + "ĠاÙĦسÙĪÙĤ": 133320, + "ĠاÙĦسÙĬ": 125143, + "ĠاÙĦسÙĬارات": 139370, + "ĠاÙĦسÙĬاس": 132792, + "ĠاÙĦسÙĬاسÙĬ": 133527, + "ĠاÙĦسÙĬاسÙĬØ©": 132793, + "ĠاÙĦسÙĬد": 132634, + "ĠاÙĦØ´": 124209, + "ĠاÙĦشباب": 133655, + "ĠاÙĦشخص": 127902, + "ĠاÙĦشخصÙĬØ©": 139477, + "ĠاÙĦشر": 125249, + "ĠاÙĦشرÙĤ": 132051, + "ĠاÙĦشرÙĥات": 134744, + "ĠاÙĦشرÙĥØ©": 131927, + "ĠاÙĦشرÙĬÙģ": 141343, + "ĠاÙĦشعب": 129839, + "ĠاÙĦشعر": 133892, + "ĠاÙĦØ´ÙħاÙĦ": 136614, + "ĠاÙĦØ´Ùħس": 139560, + "ĠاÙĦØ´Ùĩر": 136005, + "ĠاÙĦØ´ÙĬØ®": 131017, + "ĠاÙĦص": 124265, + "ĠاÙĦصادر": 142740, + "ĠاÙĦصØŃ": 127991, + "ĠاÙĦصØŃØ©": 135223, + "ĠاÙĦصØŃÙģÙĬ": 142975, + "ĠاÙĦصØŃÙĬ": 138998, + "ĠاÙĦصØŃÙĬØ©": 138554, + "ĠاÙĦصØŃÙĬØŃ": 141757, + "ĠاÙĦصÙģ": 127753, + "ĠاÙĦصÙģØŃØ©": 139463, + "ĠاÙĦصÙĬÙĨ": 132781, + "ĠاÙĦض": 129755, + "ĠاÙĦØ·": 124663, + "ĠاÙĦطب": 130280, + "ĠاÙĦطبÙĬعÙĬ": 136471, + "ĠاÙĦطرÙĬÙĤ": 132866, + "ĠاÙĦطعاÙħ": 140591, + "ĠاÙĦØ·ÙģÙĦ": 136529, + "ĠاÙĦØ·ÙĦ": 140122, + "ĠاÙĦØ·ÙĦاب": 140123, + "ĠاÙĦع": 123894, + "ĠاÙĦعاص": 139247, + "ĠاÙĦعاصÙħØ©": 139248, + "ĠاÙĦعاÙĦÙħ": 125547, + "ĠاÙĦعاÙĦÙħÙĬ": 136364, + "ĠاÙĦعاÙĦÙħÙĬØ©": 132895, + "ĠاÙĦعاÙĦÙĬ": 140984, + "ĠاÙĦعاÙħ": 128586, + "ĠاÙĦعاÙħØ©": 129701, + "ĠاÙĦعب": 138534, + "ĠاÙĦعدÙĪ": 140427, + "ĠاÙĦعدÙĬد": 130684, + "ĠاÙĦعرا": 135194, + "ĠاÙĦعراÙĤ": 130372, + "ĠاÙĦعراÙĤÙĬ": 135195, + "ĠاÙĦعرب": 131941, + "ĠاÙĦعربÙĬ": 130538, + "ĠاÙĦعربÙĬØ©": 129071, + "ĠاÙĦعسÙĥ": 140420, + "ĠاÙĦعسÙĥر": 138276, + "ĠاÙĦعسÙĥرÙĬ": 140421, + "ĠاÙĦعسÙĥرÙĬØ©": 138277, + "ĠاÙĦعشر": 141471, + "ĠاÙĦعÙĦاÙĤات": 138795, + "ĠاÙĦعÙĦاÙĤØ©": 143229, + "ĠاÙĦعÙĦÙħ": 129726, + "ĠاÙĦعÙĦÙĬ": 140764, + "ĠاÙĦعÙĦÙĬا": 140765, + "ĠاÙĦعÙħ": 132875, + "ĠاÙĦعÙħر": 138414, + "ĠاÙĦعÙħÙĦ": 125503, + "ĠاÙĦعÙħÙĦات": 137371, + "ĠاÙĦعÙħÙĦÙĬØ©": 137637, + "ĠاÙĦعÙĨ": 132638, + "ĠاÙĦغ": 125110, + "ĠاÙĦغذ": 140800, + "ĠاÙĦغذائÙĬ": 140801, + "ĠاÙĦغربÙĬØ©": 139981, + "ĠاÙĦغرÙģ": 140841, + "ĠاÙĦØŃ": 123961, + "ĠاÙĦØŃاج": 139498, + "ĠاÙĦØŃاÙĦ": 130690, + "ĠاÙĦØŃاÙĦÙĬ": 134898, + "ĠاÙĦØŃدÙĬØ«": 131801, + "ĠاÙĦØŃدÙĬد": 141043, + "ĠاÙĦØŃر": 129395, + "ĠاÙĦØŃرب": 133058, + "ĠاÙĦØŃصÙĪÙĦ": 134516, + "ĠاÙĦØŃÙĤ": 127405, + "ĠاÙĦØŃÙĤÙĪÙĤ": 143056, + "ĠاÙĦØŃÙĤÙĬÙĤÙĬ": 138606, + "ĠاÙĦØŃÙĥÙħ": 134710, + "ĠاÙĦØŃÙĥÙĪÙħ": 126807, + "ĠاÙĦØŃÙĥÙĪÙħØ©": 130119, + "ĠاÙĦØŃÙĥÙĪÙħÙĬØ©": 140447, + "ĠاÙĦØŃÙĦ": 132151, + "ĠاÙĦØŃÙħÙĦ": 137636, + "ĠاÙĦØŃÙĪØ§Ø±": 140689, + "ĠاÙĦØŃÙĪØ«": 143130, + "ĠاÙĦØŃÙĪØ«ÙĬ": 143131, + "ĠاÙĦØŃÙĬ": 133612, + "ĠاÙĦØŃÙĬاة": 131359, + "ĠاÙĦÙģ": 124114, + "ĠاÙĦÙģØª": 135920, + "ĠاÙĦÙģØªØ±Ø©": 134396, + "ĠاÙĦÙģØ±Øµ": 142847, + "ĠاÙĦÙ쨱ÙĨ": 135556, + "ĠاÙĦÙ쨱ÙĨسÙĬ": 135557, + "ĠاÙĦÙ쨱ÙĬÙĤ": 133531, + "ĠاÙĦÙģÙĤر": 140141, + "ĠاÙĦÙģÙĥر": 137020, + "ĠاÙĦÙģÙĦ": 127565, + "ĠاÙĦÙģÙĦسطÙĬÙĨ": 127566, + "ĠاÙĦÙģÙĦسطÙĬÙĨÙĬ": 134037, + "ĠاÙĦÙģÙĦسطÙĬÙĨÙĬØ©": 136987, + "ĠاÙĦÙģÙĨاÙĨ": 138331, + "ĠاÙĦÙģÙĨد": 141211, + "ĠاÙĦÙģÙĨدÙĤ": 141212, + "ĠاÙĦÙģÙĨÙĬ": 134947, + "ĠاÙĦÙģÙĪØ±": 131639, + "ĠاÙĦÙģÙĪØ±Ùĥس": 131640, + "ĠاÙĦÙĤ": 124172, + "ĠاÙĦÙĤاد": 135170, + "ĠاÙĦÙĤادÙħ": 135171, + "ĠاÙĦÙĤاÙĨÙĪÙĨ": 130771, + "ĠاÙĦÙĤاÙĩرة": 134999, + "ĠاÙĦÙĤدس": 138238, + "ĠاÙĦÙĤدÙħ": 136079, + "ĠاÙĦÙĤر": 125993, + "ĠاÙĦÙĤرآÙĨ": 134074, + "ĠاÙĦÙĤرار": 134466, + "ĠاÙĦÙĤرÙĨ": 135640, + "ĠاÙĦÙĤض": 142887, + "ĠاÙĦÙĤضÙĬØ©": 142888, + "ĠاÙĦÙĤØ·": 134848, + "ĠاÙĦÙĤطاع": 134849, + "ĠاÙĦÙĤÙĪØ§Øª": 137279, + "ĠاÙĦÙĤÙĪÙħ": 141397, + "ĠاÙĦÙĤÙĪÙħÙĬ": 141398, + "ĠاÙĦÙĤÙĬاÙħ": 136708, + "ĠاÙĦÙĥ": 124326, + "ĠاÙĦÙĥاÙħÙĦ": 140286, + "ĠاÙĦÙĥبر": 141674, + "ĠاÙĦÙĥبرÙī": 141675, + "ĠاÙĦÙĥبÙĬر": 132927, + "ĠاÙĦÙĥتاب": 134145, + "ĠاÙĦÙĥØ«ÙĬر": 130346, + "ĠاÙĦÙĥر": 130913, + "ĠاÙĦÙĥرÙĬÙħ": 134233, + "ĠاÙĦÙĥÙĦ": 126985, + "ĠاÙĦÙĥÙĦاÙħ": 142518, + "ĠاÙĦÙĥÙħ": 143790, + "ĠاÙĦÙĥÙħبÙĬ": 143791, + "ĠاÙĦÙĥÙħبÙĬÙĪØªØ±": 143792, + "ĠاÙĦÙĥÙĪÙĬ": 134384, + "ĠاÙĦÙĥÙĪÙĬت": 134385, + "ĠاÙĦÙĦ": 124478, + "ĠاÙĦÙĦا": 143098, + "ĠاÙĦÙĦازÙħØ©": 143099, + "ĠاÙĦÙĦبÙĨ": 136823, + "ĠاÙĦÙĦبÙĨاÙĨÙĬ": 136824, + "ĠاÙĦÙĦج": 133505, + "ĠاÙĦÙĦجÙĨØ©": 133506, + "ĠاÙĦÙĦعبة": 143974, + "ĠاÙĦÙĦغ": 137056, + "ĠاÙĦÙĦغة": 137057, + "ĠاÙĦÙĦÙĤاء": 141266, + "ĠاÙĦÙĦÙĩ": 128286, + "ĠاÙĦÙĦÙĬ": 127517, + "ĠاÙĦÙĦÙĬبÙĬ": 143509, + "ĠاÙĦÙħ": 53479, + "ĠاÙĦÙħؤ": 127647, + "ĠاÙĦÙħؤتÙħر": 138069, + "ĠاÙĦÙħادة": 138324, + "ĠاÙĦÙħاض": 129744, + "ĠاÙĦÙħاضÙĬ": 129745, + "ĠاÙĦÙħاÙĦ": 132013, + "ĠاÙĦÙħاÙĦÙĬØ©": 132071, + "ĠاÙĦÙħباراة": 140809, + "ĠاÙĦÙħباشر": 143371, + "ĠاÙĦÙħت": 128631, + "ĠاÙĦÙħتعÙĦ": 141721, + "ĠاÙĦÙħتعÙĦÙĤØ©": 141722, + "ĠاÙĦÙħتØŃ": 129039, + "ĠاÙĦÙħتØŃدة": 129040, + "ĠاÙĦÙħتÙĪ": 141358, + "ĠاÙĦÙħتÙĪØ³Ø·": 141359, + "ĠاÙĦÙħج": 126913, + "ĠاÙĦÙħجاÙĦ": 134975, + "ĠاÙĦÙħجتÙħع": 130555, + "ĠاÙĦÙħجÙĦس": 132700, + "ĠاÙĦÙħختÙĦÙģ": 136972, + "ĠاÙĦÙħختÙĦÙ쨩": 136973, + "ĠاÙĦÙħدÙĨÙĬ": 134636, + "ĠاÙĦÙħدÙĬÙĨØ©": 132879, + "ĠاÙĦÙħر": 125637, + "ĠاÙĦÙħرأ": 133467, + "ĠاÙĦÙħرأة": 133468, + "ĠاÙĦÙħرØŃÙĦØ©": 138070, + "ĠاÙĦÙħرÙĥز": 131052, + "ĠاÙĦÙħزÙĬد": 133082, + "ĠاÙĦÙħس": 127072, + "ĠاÙĦÙħسؤÙĪÙĦ": 135673, + "ĠاÙĦÙħست": 125616, + "ĠاÙĦÙħستÙĤبÙĦ": 138471, + "ĠاÙĦÙħستÙĪÙī": 139148, + "ĠاÙĦÙħسÙĦ": 142339, + "ĠاÙĦÙħسÙĦØŃØ©": 142340, + "ĠاÙĦÙħسÙĦÙħ": 134926, + "ĠاÙĦÙħسÙĦÙħÙĬÙĨ": 134927, + "ĠاÙĦÙħسÙĬ": 138402, + "ĠاÙĦÙħسÙĬØŃ": 138403, + "ĠاÙĦÙħØ´": 125358, + "ĠاÙĦÙħشار": 134871, + "ĠاÙĦÙħشارÙĥØ©": 134872, + "ĠاÙĦÙħشترÙĥ": 138144, + "ĠاÙĦÙħشرÙĪØ¹": 134790, + "ĠاÙĦÙħØ´ÙĥÙĦ": 143331, + "ĠاÙĦÙħØ´ÙĥÙĦØ©": 143332, + "ĠاÙĦÙħص": 131924, + "ĠاÙĦÙħصدر": 138026, + "ĠاÙĦÙħصر": 133839, + "ĠاÙĦÙħصرÙĬ": 131925, + "ĠاÙĦÙħصرÙĬØ©": 133840, + "ĠاÙĦÙħع": 129200, + "ĠاÙĦÙħعار": 143385, + "ĠاÙĦÙħعارضة": 143386, + "ĠاÙĦÙħعÙĦÙĪÙħات": 132733, + "ĠاÙĦÙħغرب": 131242, + "ĠاÙĦÙħØŃ": 126453, + "ĠاÙĦÙħØŃÙĦ": 137386, + "ĠاÙĦÙħØŃÙĦÙĬ": 141279, + "ĠاÙĦÙħØŃÙĦÙĬØ©": 137387, + "ĠاÙĦÙħÙĤ": 133476, + "ĠاÙĦÙħÙĤبÙĦ": 133205, + "ĠاÙĦÙħÙĤدس": 143898, + "ĠاÙĦÙħÙĦ": 131267, + "ĠاÙĦÙħÙĦÙĥ": 130815, + "ĠاÙĦÙħÙħÙĦ": 132904, + "ĠاÙĦÙħÙħÙĦÙĥØ©": 132905, + "ĠاÙĦÙħÙĨ": 125088, + "ĠاÙĦÙħÙĨا": 136719, + "ĠاÙĦÙħÙĨاسب": 141938, + "ĠاÙĦÙħÙĨاطÙĤ": 136720, + "ĠاÙĦÙħÙĨت": 127955, + "ĠاÙĦÙħÙĨتخب": 139620, + "ĠاÙĦÙħÙĨتدÙī": 137875, + "ĠاÙĦÙħÙĨزÙĦ": 136875, + "ĠاÙĦÙħÙĨØ·ÙĤØ©": 131263, + "ĠاÙĦÙħÙĩ": 142822, + "ĠاÙĦÙħÙĩÙĨÙĬ": 142823, + "ĠاÙĦÙħÙĪ": 125867, + "ĠاÙĦÙħÙĪØ§": 138353, + "ĠاÙĦÙħÙĪØ§Ø¯": 137499, + "ĠاÙĦÙħÙĪØ§Ø·ÙĨ": 138354, + "ĠاÙĦÙħÙĪØ§Ø·ÙĨÙĬÙĨ": 138355, + "ĠاÙĦÙħÙĪØ¬ÙĪØ¯": 143591, + "ĠاÙĦÙħÙĪØ¬ÙĪØ¯Ø©": 143592, + "ĠاÙĦÙħÙĪØ³Ùħ": 140496, + "ĠاÙĦÙħÙĪØ¶ÙĪØ¹": 130989, + "ĠاÙĦÙħÙĪÙĤع": 133052, + "ĠاÙĦÙħÙĬاÙĩ": 137010, + "ĠاÙĦÙħÙı": 137635, + "ĠاÙĦÙĨ": 124080, + "ĠاÙĦÙĨار": 137889, + "ĠاÙĦÙĨاس": 129781, + "ĠاÙĦÙĨبÙĬ": 134588, + "ĠاÙĦÙĨساء": 137169, + "ĠاÙĦÙĨظاÙħ": 130867, + "ĠاÙĦÙĨظر": 134430, + "ĠاÙĦÙĨÙģ": 135887, + "ĠاÙĦÙĨÙ쨳": 135087, + "ĠاÙĦÙĨÙ쨷": 135888, + "ĠاÙĦÙĨÙĤ": 132474, + "ĠاÙĦÙĨÙĤد": 139878, + "ĠاÙĦÙĨÙĩائÙĬ": 141051, + "ĠاÙĦÙĨÙĪ": 141568, + "ĠاÙĦÙĨÙĪØ§Ø¨": 142304, + "ĠاÙĦÙĨÙĪØ¹": 140136, + "ĠاÙĦÙĨÙĪÙĪÙĬ": 141569, + "ĠاÙĦÙĩ": 125100, + "ĠاÙĦÙĩاتÙģ": 140810, + "ĠاÙĦÙĩÙĨد": 136157, + "ĠاÙĦÙĩÙĬئة": 138928, + "ĠاÙĦÙĪ": 124666, + "ĠاÙĦÙĪØ§": 141239, + "ĠاÙĦÙĪØ§ØŃد": 141240, + "ĠاÙĦÙĪØ§ÙĤع": 133474, + "ĠاÙĦÙĪØ²": 134769, + "ĠاÙĦÙĪØ²Ø±Ø§Ø¡": 134770, + "ĠاÙĦÙĪØµÙĪÙĦ": 137410, + "ĠاÙĦÙĪØ·ÙĨ": 127275, + "ĠاÙĦÙĪØ·ÙĨÙĬ": 131262, + "ĠاÙĦÙĪØ·ÙĨÙĬØ©": 132260, + "ĠاÙĦÙĪØŃ": 140425, + "ĠاÙĦÙĪØŃÙĬد": 140426, + "ĠاÙĦÙĪÙĤ": 129815, + "ĠاÙĦÙĪÙĤت": 129816, + "ĠاÙĦÙĪÙĦا": 132371, + "ĠاÙĦÙĪÙĦاÙĬات": 132372, + "ĠاÙĦÙī": 128470, + "ĠاÙĦÙĬ": 125461, + "ĠاÙĦÙĬÙħÙĨ": 131194, + "ĠاÙĦÙĬÙĪÙħ": 128438, + "ĠاÙĦÙĴ": 130049, + "ĠاÙĨ": 128295, + "ĠاÙĨÙĩ": 131524, + "ĠاÙĪ": 129118, + "ĠاÙĬ": 127068, + "ĠاÙĬضا": 140218, + "Ġاﻷ": 137989, + "Ġب": 27846, + "Ġبأ": 129521, + "ĠبأÙĨ": 126513, + "ĠبأÙĨÙĩ": 136897, + "Ġبإ": 130394, + "Ġبا": 82611, + "ĠباستخداÙħ": 139073, + "ĠباسÙħ": 135208, + "ĠباÙĦ": 124476, + "ĠباÙĦØ£": 134964, + "ĠباÙĦØ¥": 127586, + "ĠباÙĦإض": 134156, + "ĠباÙĦإضاÙ쨩": 134157, + "ĠباÙĦت": 132548, + "ĠباÙĦÙħ": 129890, + "ĠباÙĦÙĨ": 132982, + "ĠباÙĦÙĨسبة": 132983, + "Ġبت": 126805, + "ĠبتارÙĬØ®": 140701, + "Ġبد": 138829, + "Ġبدأت": 138830, + "Ġبدا": 135769, + "ĠبداÙĬØ©": 135770, + "ĠبدÙĪÙĨ": 135172, + "ĠبذÙĦÙĥ": 137007, + "Ġبر": 129920, + "ĠبرÙĨ": 130440, + "ĠبرÙĨاÙħج": 130441, + "Ġبسبب": 130153, + "Ġبش": 126436, + "ĠبشأÙĨ": 133707, + "ĠبشÙĥÙĦ": 128953, + "Ġبص": 127726, + "ĠبصÙĪØ±Ø©": 142882, + "Ġبط": 141600, + "ĠبطرÙĬ": 141601, + "ĠبطرÙĬÙĤØ©": 141602, + "Ġبعد": 128325, + "Ġبعض": 128538, + "ĠبعÙĬد": 136156, + "Ġبغ": 141868, + "Ġبغداد": 141869, + "ĠبØŃÙĬ": 140769, + "ĠبØŃÙĬØ«": 140770, + "ĠبÙĥ": 129834, + "ĠبÙĥÙĦ": 132319, + "ĠبÙĦ": 126420, + "ĠبÙĦا": 134933, + "ĠبÙĦد": 132428, + "ĠبÙħ": 128763, + "ĠبÙħا": 129869, + "ĠبÙĨ": 125449, + "ĠبÙĨسبة": 139102, + "ĠبÙĨÙ쨳": 138854, + "ĠبÙĩ": 81768, + "ĠبÙĩا": 128794, + "ĠبÙĩذا": 136791, + "ĠبÙĩذÙĩ": 139116, + "ĠبÙĪ": 126941, + "ĠبÙĪØ§Ø³": 135131, + "ĠبÙĪØ§Ø³Ø·Ø©": 135132, + "ĠبÙĬ": 129517, + "ĠبÙĬاÙĨ": 142011, + "ĠبÙĬاÙĨات": 142012, + "ĠبÙĬÙĨ": 126212, + "ĠبÙĬÙĨÙħا": 134939, + "Ġت": 39434, + "ĠتارÙĬØ®": 130570, + "Ġتت": 129015, + "Ġتخ": 134684, + "ĠتراÙħ": 137985, + "ĠتراÙħب": 137986, + "ĠترÙĥ": 136520, + "ĠترÙĥÙĬا": 136521, + "ĠترÙĬد": 138169, + "Ġتس": 130243, + "Ġتست": 131723, + "Ġتش": 143700, + "ĠتشرÙĬÙĨ": 143701, + "Ġتص": 130546, + "ĠتصÙĨÙĬ": 130547, + "ĠتصÙĨÙĬÙģ": 130548, + "Ġتع": 126731, + "ĠتعاÙĦ": 131467, + "ĠتعاÙĦÙī": 131468, + "Ġتعتبر": 140118, + "Ġتعد": 138073, + "Ġتعرض": 138481, + "ĠتعÙħÙĦ": 135545, + "ĠتØŃ": 125857, + "ĠتØŃت": 129375, + "ĠتØŃتاج": 138756, + "ĠتØŃدÙĬØ«": 138383, + "ĠتØŃÙĤÙĬÙĤ": 132745, + "ĠتÙĤ": 133114, + "ĠتÙĤدÙĬ": 133115, + "ĠتÙĤدÙĬÙħ": 133116, + "ĠتÙĤÙĪÙħ": 133393, + "ĠتÙĥÙĪÙĨ": 129286, + "ĠتÙĦÙĥ": 129341, + "ĠتÙħ": 128510, + "ĠتÙħاÙħ": 138787, + "ĠتÙħاÙħا": 138788, + "ĠتÙĨ": 129458, + "ĠتÙĪ": 130616, + "ĠتÙĪÙģ": 139408, + "ĠتÙĪÙģÙĬر": 139409, + "ĠØ«": 125069, + "ĠثاÙĨÙĬ": 135222, + "ĠØ«ÙĦاث": 135376, + "ĠØ«ÙĦاثة": 135377, + "ĠØ«Ùħ": 128637, + "ĠØ«ÙĨ": 143342, + "ĠØ«ÙĨائÙĬ": 143343, + "Ġج": 82168, + "Ġجاء": 131813, + "ĠجاÙħعة": 134193, + "ĠجاÙĨب": 131351, + "Ġجدا": 127119, + "ĠجداÙĭ": 139757, + "ĠجدÙĬد": 130461, + "ĠجدÙĬدة": 130282, + "ĠجÙħÙĬع": 128858, + "ĠجÙħÙĬÙĦ": 138518, + "ĠجÙĨÙĬ": 135986, + "ĠجÙĨÙĬÙĩ": 135987, + "ĠجÙĩØ©": 138887, + "ĠØ®": 74315, + "Ġخاصة": 130439, + "ĠخدÙħات": 136310, + "Ġخط": 129163, + "ĠØ®ÙĦاÙĦ": 128374, + "Ġد": 44330, + "Ġدائ": 140214, + "ĠدائÙħ": 136075, + "ĠدائÙħا": 140215, + "ĠداخÙĦ": 129996, + "ĠدبÙĬ": 137373, + "Ġدر": 76841, + "ĠدÙĤ": 141890, + "ĠدÙĤائÙĤ": 141891, + "ĠدÙĤÙĬÙĤØ©": 136574, + "ĠدÙħØ´": 143401, + "ĠدÙħØ´ÙĤ": 143402, + "ĠدÙĪØ±": 129366, + "ĠدÙĪÙĦ": 127012, + "ĠدÙĪÙĦار": 130907, + "ĠدÙĪÙĨ": 129193, + "ĠدÙĬ": 126920, + "ĠدÙĬسÙħ": 135977, + "ĠدÙĬسÙħبر": 135978, + "Ġذ": 125447, + "Ġذات": 129632, + "ĠذÙĦÙĥ": 128349, + "ĠذÙĪ": 137776, + "Ġر": 53710, + "ĠرئÙĬس": 128755, + "Ġرائع": 136629, + "Ġرب": 139868, + "ĠربÙħا": 139869, + "ĠرجاÙĦ": 139589, + "Ġرد": 130852, + "Ġرس": 127490, + "ĠرساÙĦØ©": 139268, + "ĠرسÙħÙĬ": 138359, + "ĠرسÙĪÙĦ": 133555, + "ĠرغÙħ": 135297, + "ĠرÙĤÙħ": 131273, + "ĠرÙħض": 135865, + "ĠرÙħضاÙĨ": 135866, + "ĠرÙĪ": 131384, + "ĠرÙĬ": 137542, + "ĠرÙĬاÙĦ": 137543, + "Ġز": 125555, + "ĠزÙħÙĨ": 141492, + "ĠزÙĬ": 134777, + "ĠزÙĬادة": 134778, + "Ġس": 59842, + "Ġساب": 134577, + "ĠسابÙĤ": 134578, + "Ġساع": 138717, + "Ġساعات": 138718, + "Ġساعة": 135512, + "ĠسبØŃ": 143750, + "ĠسبØŃاÙĨ": 143751, + "ĠسبØŃاÙĨÙĩ": 143752, + "ĠسبÙĬÙĦ": 135638, + "Ġست": 128875, + "ĠسرÙĬع": 140675, + "ĠسعÙĪØ¯": 139398, + "ĠسÙħ": 131409, + "ĠسÙĨ": 126760, + "ĠسÙĨØ©": 130643, + "ĠسÙĨÙĪØ§Øª": 132351, + "ĠسÙĪØ§Ø¡": 134607, + "ĠسÙĪØ±": 132681, + "ĠسÙĪØ±ÙĬا": 132682, + "ĠسÙĪØ±ÙĬØ©": 141241, + "ĠسÙĪÙģ": 131959, + "ĠسÙĪÙī": 140545, + "ĠسÙĬ": 125027, + "ĠسÙĬاسÙĬ": 134874, + "ĠسÙĬتÙħ": 141420, + "ĠسÙĬÙĥÙĪÙĨ": 137139, + "ĠØ´": 52157, + "ĠشأÙĨ": 139894, + "ĠشاÙħÙĦ": 143870, + "Ġشب": 138566, + "ĠشبÙĥØ©": 138567, + "Ġشخص": 129843, + "ĠشرÙĥØ©": 129798, + "ĠØ´Ùĥ": 143622, + "ĠØ´Ùĥرا": 143623, + "ĠØ´ÙħاÙĦ": 137855, + "ĠØ´Ùĩر": 130777, + "ĠØ´ÙĬ": 127050, + "ĠØ´ÙĬØ¡": 131113, + "ĠØ´ÙĬئ": 142613, + "ĠØ´ÙĬئا": 142614, + "Ġص": 92072, + "ĠصاÙĦ": 137263, + "ĠصاÙĦØŃ": 137264, + "ĠصباØŃ": 134711, + "Ġصغ": 139849, + "ĠصغÙĬرة": 139850, + "ĠصØŃ": 134772, + "ĠصØŃÙĬØŃ": 136826, + "ĠصØŃÙĬÙ쨩": 134773, + "ĠصÙģ": 141949, + "ĠصÙģØŃØ©": 141950, + "ĠصÙĦÙī": 131108, + "Ġض": 126815, + "Ġضد": 130160, + "ĠضÙħÙĨ": 131695, + "ĠØ·": 124665, + "ĠطبÙĬعÙĬ": 141755, + "ĠطرÙĬÙĤ": 129072, + "ĠØ·ÙģÙĦ": 139820, + "ĠØ·ÙĪÙĬÙĦ": 128054, + "ĠØ·ÙĪÙĬÙĦØ©": 139412, + "Ġع": 43982, + "ĠعاÙĦÙĬØ©": 137091, + "ĠعاÙħ": 128443, + "Ġعبد": 125584, + "ĠعبداÙĦ": 127604, + "ĠعبداÙĦÙĦÙĩ": 140533, + "Ġعبر": 129353, + "Ġعدة": 133407, + "Ġعدد": 128912, + "ĠعدÙħ": 130283, + "ĠعربÙĬ": 135410, + "Ġعشر": 130520, + "ĠعÙĦاÙĤØ©": 140691, + "ĠعÙĦÙħ": 130226, + "ĠعÙĦÙī": 128248, + "ĠعÙĦÙĬ": 124794, + "ĠعÙĦÙĬÙĥ": 133254, + "ĠعÙĦÙĬÙĨ": 139978, + "ĠعÙĦÙĬÙĨا": 139979, + "ĠعÙĦÙĬÙĩ": 128471, + "ĠعÙĦÙĬÙĩا": 129387, + "ĠعÙĦÙĬÙĩÙħ": 135293, + "ĠعÙħ": 131268, + "ĠعÙħر": 130657, + "ĠعÙħÙĦ": 126349, + "ĠعÙħÙĦÙĬØ©": 130759, + "ĠعÙĨ": 126195, + "ĠعÙĨد": 126350, + "ĠعÙĨدÙħا": 129919, + "ĠعÙĨÙĩ": 131045, + "ĠعÙĨÙĩا": 134035, + "Ġغ": 125040, + "ĠغاÙĦ": 142327, + "ĠغاÙĦب": 142328, + "Ġغزة": 138348, + "ĠغÙĬر": 128478, + "ĠØĮ": 128287, + "ĠØŃ": 68238, + "ĠØŃاج": 143276, + "ĠØŃاجة": 143277, + "ĠØŃاÙĦ": 129466, + "ĠØŃاÙĦØ©": 131032, + "ĠØŃت": 128522, + "ĠØŃتÙī": 128523, + "ĠØŃد": 130740, + "ĠØŃر": 130630, + "ĠØŃسÙĬÙĨ": 141549, + "ĠØŃص": 137982, + "ĠØŃÙĤ": 126906, + "ĠØŃÙĤÙĪÙĤ": 133547, + "ĠØŃÙĤÙĬÙĤÙĬ": 142182, + "ĠØŃÙĦ": 130260, + "ĠØŃÙĪ": 140062, + "ĠØŃÙĪØ§ÙĦÙĬ": 140063, + "ĠØŃÙĪÙĦ": 128967, + "ĠØŃÙĬات": 139542, + "ĠØŃÙĬاتÙĩ": 139543, + "ĠØŃÙĬØ«": 128476, + "ĠØŃÙĬÙĨ": 130492, + "ĠÙ¾": 87417, + "ĠÙģ": 45577, + "ĠÙ쨣": 131666, + "ĠÙ쨥": 143026, + "ĠÙģØ¥Ø°Ø§": 143027, + "ĠÙ쨥ÙĨ": 128843, + "ĠÙ쨧ÙĦ": 130044, + "ĠÙģØªØ±Ø©": 133933, + "ĠÙģØ±Øµ": 134642, + "ĠÙ쨹ÙĦ": 131394, + "ĠÙģÙĤاÙĦ": 134598, + "ĠÙģÙĤد": 129828, + "ĠÙģÙĤØ·": 129346, + "ĠÙģÙĥرة": 142139, + "ĠÙģÙĦا": 132225, + "ĠÙģÙħÙĨ": 139400, + "ĠÙģÙĩÙĪ": 133965, + "ĠÙģÙĩÙĬ": 137687, + "ĠÙģÙĪÙĤ": 135726, + "ĠÙģÙī": 128393, + "ĠÙģÙĬ": 77273, + "ĠÙģÙĬÙħا": 130339, + "ĠÙģÙĬÙĩ": 128985, + "ĠÙģÙĬÙĩا": 128857, + "ĠÙģÙIJ": 135948, + "ĠÙģÙIJÙĬ": 135949, + "ĠÙĤ": 77703, + "ĠÙĤائ": 142760, + "ĠÙĤائÙĦا": 142761, + "ĠÙĤادر": 139825, + "ĠÙĤاÙĦ": 128506, + "ĠÙĤاÙħت": 140164, + "ĠÙĤبÙĦ": 128474, + "ĠÙĤد": 128416, + "ĠÙĤرÙĬب": 138148, + "ĠÙĤصة": 141714, + "ĠÙĤض": 134541, + "ĠÙĤطر": 136510, + "ĠÙĤÙĨا": 138847, + "ĠÙĤÙĨاة": 138848, + "ĠÙĤÙĪÙĦ": 137436, + "ĠÙĤÙĪÙĦÙĩ": 137437, + "ĠÙĤÙĬÙħØ©": 138272, + "ĠÙĤÙİ": 142085, + "ĠÙĤÙİØ§ÙĦ": 142086, + "ĠÙĤÙİØ§ÙĦÙİ": 142087, + "ĠÙĥ": 86941, + "ĠÙĥاÙ쨩": 133550, + "ĠÙĥاÙĦ": 133394, + "ĠÙĥاÙħÙĦ": 132435, + "ĠÙĥاÙĨ": 126214, + "ĠÙĥاÙĨت": 128671, + "ĠÙĥاÙĨÙĪØ§": 137809, + "ĠÙĥب": 130854, + "ĠÙĥبÙĬر": 129766, + "ĠÙĥبÙĬرة": 130855, + "ĠÙĥتاب": 131211, + "ĠÙĥØ«ÙĬر": 129724, + "ĠÙĥذÙĦÙĥ": 135283, + "ĠÙĥرة": 139581, + "ĠÙĥس": 141585, + "ĠÙĥسارة": 141586, + "ĠÙĥÙĦ": 128288, + "ĠÙĥÙĦÙħ": 136366, + "ĠÙĥÙĦÙħØ©": 136367, + "ĠÙĥÙħا": 128388, + "ĠÙĥÙĨت": 130309, + "ĠÙĥÙĪ": 130190, + "ĠÙĥÙĬÙģ": 127046, + "ĠÙĥÙĬÙģÙĬØ©": 137146, + "ĠÙĦ": 56794, + "ĠÙĦØ£": 129232, + "ĠÙĦØ£ÙĨ": 126554, + "ĠÙĦØ£ÙĨÙĩ": 134270, + "ĠÙĦØ¥": 131425, + "ĠÙĦا": 128259, + "ĠÙĦبÙĨ": 133092, + "ĠÙĦبÙĨاÙĨ": 133093, + "ĠÙĦت": 128769, + "ĠÙĦتØŃ": 143011, + "ĠÙĦتØŃÙĤÙĬÙĤ": 143012, + "ĠÙĦدÙī": 132266, + "ĠÙĦدÙĬ": 125998, + "ĠÙĦدÙĬÙĥ": 142360, + "ĠÙĦدÙĬÙĨا": 140554, + "ĠÙĦدÙĬÙĩ": 132653, + "ĠÙĦذÙĦÙĥ": 132730, + "ĠÙĦعبة": 139925, + "ĠÙĦÙĤد": 133431, + "ĠÙĦÙĥ": 129361, + "ĠÙĦÙĥرة": 143185, + "ĠÙĦÙĥÙĦ": 131863, + "ĠÙĦÙĥÙħ": 135109, + "ĠÙĦÙĥÙĨ": 128762, + "ĠÙĦÙĦ": 125006, + "ĠÙĦÙĦØ£": 130579, + "ĠÙĦÙĦت": 131037, + "ĠÙĦÙĦغا": 140537, + "ĠÙĦÙĦغاÙĬØ©": 140538, + "ĠÙĦÙĦÙħ": 128883, + "ĠÙĦÙħ": 126208, + "ĠÙĦÙħا": 127301, + "ĠÙĦÙħاذا": 138867, + "ĠÙĦÙħدة": 135745, + "ĠÙĦÙĨ": 129225, + "ĠÙĦÙĨا": 133289, + "ĠÙĦÙĩ": 125490, + "ĠÙĦÙĩا": 129046, + "ĠÙĦÙĩذا": 133310, + "ĠÙĦÙĩذÙĩ": 137726, + "ĠÙĦÙĩÙħ": 131029, + "ĠÙĦÙĪ": 129768, + "ĠÙĦÙĬ": 125502, + "ĠÙĦÙĬبÙĬ": 140441, + "ĠÙĦÙĬبÙĬا": 140442, + "ĠÙĦÙĬس": 129581, + "ĠÙĦÙĬست": 134436, + "ĠÙħ": 23364, + "ĠÙħؤ": 133427, + "ĠÙħا": 126198, + "ĠÙħاÙĬÙĪ": 136057, + "ĠÙħباراة": 136369, + "ĠÙħباشر": 127922, + "ĠÙħباشرة": 137450, + "ĠÙħت": 128546, + "ĠÙħØ«ÙĦ": 128718, + "ĠÙħج": 140333, + "ĠÙħجاÙĦ": 131132, + "ĠÙħجاÙĨÙĬ": 140334, + "ĠÙħجر": 142000, + "ĠÙħجرد": 142001, + "ĠÙħجÙĦ": 143506, + "ĠÙħجÙĦØ©": 143507, + "ĠÙħجÙĦس": 129878, + "ĠÙħجÙħÙĪØ¹Ø©": 131175, + "ĠÙħختÙĦÙģ": 127519, + "ĠÙħختÙĦÙ쨩": 135526, + "ĠÙħد": 142610, + "ĠÙħدرÙĬد": 142611, + "ĠÙħدÙī": 138983, + "ĠÙħدÙĬر": 132889, + "ĠÙħدÙĬÙĨØ©": 130632, + "ĠÙħر": 128707, + "ĠÙħرة": 131658, + "ĠÙħرÙĥز": 131822, + "ĠÙħساء": 128010, + "ĠÙħساءÙĭ": 143036, + "ĠÙħست": 126492, + "ĠÙħستÙĪÙī": 132106, + "ĠÙħØ´": 125709, + "ĠÙħشرÙĪØ¹": 131563, + "ĠÙħØ´ÙĬرا": 141838, + "ĠÙħصدر": 137369, + "ĠÙħصر": 128748, + "ĠÙħض": 133020, + "ĠÙħع": 126196, + "ĠÙħعدÙĦ": 139013, + "ĠÙħعظÙħ": 136700, + "ĠÙħØŃ": 126502, + "ĠÙħØŃاÙ쨏": 139347, + "ĠÙħØŃاÙģØ¸Ø©": 139348, + "ĠÙħØŃÙĦ": 134689, + "ĠÙħØŃÙħ": 139811, + "ĠÙħØŃÙħد": 128724, + "ĠÙħØŃÙħÙĪØ¯": 139812, + "ĠÙħÙĤ": 127115, + "ĠÙħÙĤابÙĦ": 133860, + "ĠÙħÙĥاÙĨ": 131865, + "ĠÙħÙĦÙģ": 136930, + "ĠÙħÙĦÙĬ": 126993, + "ĠÙħÙĦÙĬار": 134652, + "ĠÙħÙĦÙĬÙĪÙĨ": 130859, + "ĠÙħÙħا": 129943, + "ĠÙħÙĨ": 63237, + "ĠÙħÙĨت": 143122, + "ĠÙħÙĨتخب": 143123, + "ĠÙħÙĨذ": 129185, + "ĠÙħÙĨزÙĦ": 138594, + "ĠÙħÙĨØ·ÙĤØ©": 131478, + "ĠÙħÙĨÙĩ": 130622, + "ĠÙħÙĨÙĩا": 129431, + "ĠÙħÙĨÙĩÙħ": 133576, + "ĠÙħÙĩÙħ": 141182, + "ĠÙħÙĩÙħØ©": 141183, + "ĠÙħÙĪ": 129435, + "ĠÙħÙĪØ§": 142685, + "ĠÙħÙĪØ§Ø¶": 142686, + "ĠÙħÙĪØ§Ø¶ÙĬع": 142687, + "ĠÙħÙĪØ§ÙĤع": 136215, + "ĠÙħÙĪØ¬ÙĪØ¯": 140674, + "ĠÙħÙĪØ¶ÙĪØ¹": 131476, + "ĠÙħÙĪÙĤع": 129436, + "ĠÙħÙĪÙĤÙģ": 138990, + "ĠÙħÙIJ": 137489, + "ĠÙħÙIJÙĨ": 137490, + "ĠÙħÙIJÙĨÙĴ": 137491, + "ĠÙĨ": 50243, + "ĠÙĨتÙĬ": 135305, + "ĠÙĨتÙĬجة": 135306, + "ĠÙĨسبة": 135778, + "ĠÙĨØ·": 143683, + "ĠÙĨطاÙĤ": 143684, + "ĠÙĨظاÙħ": 130895, + "ĠÙĨØŃÙĪ": 130770, + "ĠÙĨÙ쨳": 125759, + "ĠÙĨÙ쨳Ùĩ": 131581, + "ĠÙĨÙ쨳Ùĩا": 139505, + "ĠÙĨÙĤ": 142402, + "ĠÙĨÙĤاط": 142403, + "ĠÙĨÙĤØ·": 137622, + "ĠÙĨÙĤطة": 137623, + "ĠÙĨÙĩا": 134258, + "ĠÙĨÙĩائÙĬ": 142565, + "ĠÙĨÙĩاÙĬØ©": 134259, + "ĠÙĨÙĪØ¹": 131876, + "ĠÙĨÙĪÙģ": 136611, + "ĠÙĨÙĪÙģÙħبر": 136612, + "ĠÙĨÙĬ": 137516, + "ĠÙĨÙĬÙĪØ²": 137517, + "ĠÙĩ": 55891, + "ĠÙĩاتÙģ": 137209, + "ĠÙĩذا": 128280, + "ĠÙĩذÙĩ": 128305, + "ĠÙĩÙĦ": 129859, + "ĠÙĩÙĨا": 126437, + "ĠÙĩÙĨاÙĥ": 128954, + "ĠÙĩÙĪ": 128321, + "ĠÙĩÙĬ": 128420, + "ĠÙĪ": 37524, + "ĠÙĪØ£": 124766, + "ĠÙĪØ£Ø´Ø§Ø±": 138699, + "ĠÙĪØ£Ø¶Ø§Ùģ": 132670, + "ĠÙĪØ£Ùĥد": 137817, + "ĠÙĪØ£ÙĨ": 129265, + "ĠÙĪØ£ÙĪØ¶ØŃ": 139229, + "ĠÙĪØ¥": 126530, + "ĠÙĪØ¥Ø°Ø§": 139790, + "ĠÙĪØ¥ÙĨ": 130915, + "ĠÙĪØ§": 125475, + "ĠÙĪØ§Ø³Øª": 132178, + "ĠÙĪØ§Ø¶": 136314, + "ĠÙĪØ§Ø¶ØŃ": 136315, + "ĠÙĪØ§ØŃد": 126761, + "ĠÙĪØ§ØŃدة": 131205, + "ĠÙĪØ§ÙĦ": 124072, + "ĠÙĪØ§ÙĦØ£": 129271, + "ĠÙĪØ§ÙĦØ¥": 131795, + "ĠÙĪØ§ÙĦا": 131894, + "ĠÙĪØ§ÙĦاست": 138027, + "ĠÙĪØ§ÙĦت": 129006, + "ĠÙĪØ§ÙĦتÙĬ": 129833, + "ĠÙĪØ§ÙĦذÙĬ": 132052, + "ĠÙĪØ§ÙĦس": 134274, + "ĠÙĪØ§ÙĦع": 132525, + "ĠÙĪØ§ÙĦØŃ": 134018, + "ĠÙĪØ§ÙĦÙģ": 135635, + "ĠÙĪØ§ÙĦÙĦÙĩ": 139353, + "ĠÙĪØ§ÙĦÙħ": 128405, + "ĠÙĪØ§ÙĦÙĨ": 133843, + "ĠÙĪØ¨": 124838, + "ĠÙĪØ¨Ø§ÙĦ": 134264, + "ĠÙĪØ¨Ø§ÙĦت": 138826, + "ĠÙĪØ¨Ø§ÙĦتاÙĦÙĬ": 138827, + "ĠÙĪØ¨Ø¹Ø¯": 133593, + "ĠÙĪØ¨ÙĬÙĨ": 135647, + "ĠÙĪØª": 128293, + "ĠÙĪØ¬": 130677, + "ĠÙĪØ¬Ùĩ": 130995, + "ĠÙĪØ¬ÙĪØ¯": 130678, + "ĠÙĪØ®": 142794, + "ĠÙĪØ®Ø§ØµØ©": 142795, + "ĠÙĪØ°ÙĦÙĥ": 129673, + "ĠÙĪØ²": 126857, + "ĠÙĪØ²Ø§Ø±Ø©": 131209, + "ĠÙĪØ²ÙĬر": 130316, + "ĠÙĪØ³ÙĦÙħ": 130753, + "ĠÙĪØ³ÙĬ": 131452, + "ĠÙĪØ¶Ø¹": 130519, + "ĠÙĪØ·ÙĨÙĬ": 143007, + "ĠÙĪØ¹Ø¯": 139361, + "ĠÙĪØ¹Ø¯Ùħ": 139362, + "ĠÙĪØ¹ÙĦÙī": 131247, + "ĠÙĪØ¹ÙĨ": 131933, + "ĠÙĪØºÙĬر": 127751, + "ĠÙĪØºÙĬرÙĩا": 137052, + "ĠÙĪØŃØª": 139082, + "ĠÙĪØŃØªÙī": 139083, + "ĠÙĪÙģÙĬ": 128641, + "ĠÙĪÙĤاÙĦ": 129022, + "ĠÙĪÙĤت": 130133, + "ĠÙĪÙĤد": 128835, + "ĠÙĪÙĥ": 127555, + "ĠÙĪÙĥاÙĨ": 130117, + "ĠÙĪÙĥاÙĨت": 135415, + "ĠÙĪÙĥذÙĦÙĥ": 133950, + "ĠÙĪÙĥÙĦ": 133749, + "ĠÙĪÙĦ": 128827, + "ĠÙĪÙĦا": 128464, + "ĠÙĪÙĦÙĥ": 128701, + "ĠÙĪÙĦÙĥÙĨ": 128702, + "ĠÙĪÙĦÙħ": 130444, + "ĠÙĪÙĦÙĪ": 137814, + "ĠÙĪÙĦÙĬس": 132402, + "ĠÙĪÙħ": 128402, + "ĠÙĪÙħا": 129891, + "ĠÙĪÙħع": 132294, + "ĠÙĪÙħÙĨ": 128562, + "ĠÙĪÙĩ": 124821, + "ĠÙĪÙĩذا": 130267, + "ĠÙĪÙĩذÙĩ": 139769, + "ĠÙĪÙĩÙĨا": 140676, + "ĠÙĪÙĩÙĨاÙĥ": 140677, + "ĠÙĪÙĩÙĪ": 128660, + "ĠÙĪÙĩÙĬ": 129308, + "ĠÙĪÙĬ": 128306, + "ĠÙĪÙĬÙħ": 137049, + "ĠÙĪÙĬÙħÙĥÙĨ": 137050, + "ĠÙĪÙİ": 129095, + "ĠÙĬ": 73274, + "ĠÙĬأتÙĬ": 139379, + "ĠÙĬؤ": 140620, + "ĠÙĬؤدÙĬ": 140621, + "ĠÙĬا": 129720, + "ĠÙĬبدÙĪ": 140260, + "ĠÙĬت": 128657, + "ĠÙĬتعÙĦÙĤ": 137961, + "ĠÙĬتÙħ": 129106, + "ĠÙĬجب": 129539, + "ĠÙĬجعÙĦ": 137204, + "ĠÙĬرÙĬد": 137229, + "ĠÙĬست": 126850, + "ĠÙĬستطÙĬع": 140922, + "ĠÙĬصÙĦ": 137283, + "ĠÙĬعتبر": 137727, + "ĠÙĬعد": 136884, + "ĠÙĬعرÙģ": 136090, + "ĠÙĬعÙħÙĦ": 135048, + "ĠÙĬعÙĨÙĬ": 132338, + "ĠÙĬØŃتاج": 140326, + "ĠÙĬØŃدث": 138996, + "ĠÙĬÙĤدÙħ": 137702, + "ĠÙĬÙĤÙĪÙĦ": 130664, + "ĠÙĬÙĤÙĪÙħ": 133450, + "ĠÙĬÙĥÙĨ": 134192, + "ĠÙĬÙĥÙĪÙĨ": 128731, + "ĠÙĬÙħÙĥÙĨ": 126298, + "ĠÙĬÙħÙĥÙĨÙĥ": 132148, + "ĠÙĬÙĨا": 134339, + "ĠÙĬÙĨاÙĬر": 134340, + "ĠÙĬÙĨب": 140879, + "ĠÙĬÙĨبغÙĬ": 140880, + "ĠÙĬÙĪØ¬Ø¯": 134842, + "ĠÙĬÙĪØ³": 141173, + "ĠÙĬÙĪØ³Ùģ": 141174, + "ĠÙĬÙĪÙĦ": 135620, + "ĠÙĬÙĪÙĦÙĬÙĪ": 135621, + "ĠÙĬÙĪÙħ": 128587, + "ĠÙĬÙĪÙĨ": 135332, + "ĠÙĬÙĪÙĨÙĬÙĪ": 135333, + "ĠÙĬÙİ": 133770, + "ĠÙĬÙı": 129453, + "ĠÚ©": 45593, + "Ġà": 27982, + "Ġà¤": 14925, + "Ġप": 83636, + "Ġम": 91217, + "Ġस": 68158, + "Ġह": 84310, + "Ġà¤ķ": 47809, + "Ġà¦": 35178, + "Ġà¨": 69033, + "Ġàª": 94491, + "Ġà®": 70597, + "Ġà°": 93255, + "Ġà²": 89434, + "Ġม": 143737, + "Ġมา": 129739, + "Ġมี": 126250, + "Ġมีà¸Ļา": 142536, + "Ġมีà¸Ļาà¸Ħม": 142537, + "Ġมà¸ģร": 143738, + "Ġมà¸ģราà¸Ħม": 143739, + "Ġลà¹īาà¸Ļ": 131745, + "Ġวัà¸Ļ": 130175, + "Ġส": 129328, + "Ġสำหรัà¸ļ": 129629, + "Ġสิ": 142444, + "Ġสิà¸ĩ": 142445, + "Ġสิà¸ĩหา": 142446, + "Ġสิà¸ĩหาà¸Ħม": 142447, + "Ġสà¹Īวà¸Ļ": 129983, + "Ġหรืà¸Ń": 128404, + "Ġหาà¸ģ": 130138, + "Ġà¸ģ": 126675, + "Ġà¸ģร": 143866, + "Ġà¸ģรà¸ģà¸İ": 143867, + "Ġà¸ģรà¸ģà¸İาà¸Ħม": 143868, + "Ġà¸ģัà¸Ļ": 143134, + "Ġà¸ģัà¸Ļย": 143135, + "Ġà¸ģัà¸Ļยายà¸Ļ": 143136, + "Ġà¸ģาร": 128539, + "Ġà¸ģุม": 143208, + "Ġà¸ģà¸¸à¸¡à¸łà¸²à¸ŀ": 143209, + "Ġà¸ģà¸¸à¸¡à¸łà¸²à¸ŀัà¸Ļ": 143210, + "Ġà¸ģà¸¸à¸¡à¸łà¸²à¸ŀัà¸Ļà¸ĺ": 143211, + "Ġà¸ģà¸¸à¸¡à¸łà¸²à¸ŀัà¸Ļà¸ĺà¹Į": 143212, + "Ġà¸ģà¹ĩ": 129300, + "Ġà¸Ĥà¸ĵะà¸Ĺีà¹Ī": 140265, + "Ġà¸Ħืà¸Ń": 129457, + "Ġà¸Ħุà¸ĵ": 130167, + "Ġà¸Ħà¸Ļ": 129619, + "Ġà¸Īะ": 129515, + "Ġà¸Īาà¸ģ": 126648, + "Ġà¸Īาà¸ģà¸Ļัà¹īà¸Ļ": 138072, + "Ġà¸Īึà¸ĩ": 130301, + "Ġà¸ĭึà¹Īà¸ĩ": 126240, + "Ġà¸ĭึà¹Īà¸ĩà¹Ģà¸Ľà¹ĩà¸Ļ": 134357, + "Ġà¸Ķัà¸ĩ": 133793, + "Ġà¸Ķัà¸ĩà¸Ļัà¹īà¸Ļ": 133794, + "Ġà¸ķุ": 142914, + "Ġà¸ķุลา": 142915, + "Ġà¸ķุลาà¸Ħม": 142916, + "Ġà¸ĸà¹īา": 129190, + "Ġà¸Ĺัà¹īà¸ĩ": 137406, + "Ġà¸Ĺัà¹īà¸ĩà¸Ļีà¹ī": 137407, + "Ġà¸Ĺำà¹ĥหà¹ī": 130659, + "Ġà¸Ĺีà¹Ī": 128477, + "Ġà¸ĺ": 142696, + "Ġà¸ĺัà¸Ļ": 142697, + "Ġà¸ĺัà¸Ļวา": 142698, + "Ġà¸ĺัà¸Ļวาà¸Ħม": 142699, + "Ġà¸Ļาย": 130499, + "Ġà¸Ļà¸Ńà¸ģà¸Īาà¸ģ": 134353, + "Ġà¸Ļà¸Ńà¸ģà¸Īาà¸ģà¸Ļีà¹ī": 134354, + "Ġà¸ļาà¸Ĺ": 129649, + "Ġà¸Ľà¸µ": 128806, + "Ġà¸ľà¸¡": 130131, + "Ġà¸ľà¸¹à¹ī": 129543, + "Ġà¸ŀ": 126473, + "Ġà¸ŀรà¹īà¸Ńม": 129575, + "Ġà¸ŀฤษ": 143147, + "Ġà¸ŀà¸¤à¸©à¸łà¸²": 143148, + "Ġà¸ŀà¸¤à¸©à¸łà¸²à¸Ħม": 143149, + "Ġà¸Ń": 129544, + "Ġà¸Ńยà¹Īาà¸ĩà¹Ħร": 142061, + "Ġà¸Ńยà¹Īาà¸ĩà¹Ħรà¸ģà¹ĩà¸ķาม": 142062, + "Ġà¸Ńีà¸ģ": 143045, + "Ġà¸Ńีà¸ģà¸Ĺัà¹īà¸ĩ": 143046, + "Ġà¹Ģ": 94482, + "Ġà¹Ģม": 126419, + "Ġà¹Ģมษ": 143428, + "Ġà¹Ģมษายà¸Ļ": 143429, + "Ġà¹Ģมืà¹Īà¸Ń": 128903, + "Ġà¹Ģรา": 129468, + "Ġà¹Ģà¸Ĭ": 130246, + "Ġà¹Ģà¸Ĭà¹Īà¸Ļ": 130247, + "Ġà¹Ģà¸Ľ": 126323, + "Ġà¹Ģà¸Ľà¹ĩà¸Ļ": 126324, + "Ġà¹Ģà¸Ľà¹ĩà¸Ļà¸ķà¹īà¸Ļ": 138547, + "Ġà¹Ģà¸ŀ": 126306, + "Ġà¹Ģà¸ŀราะ": 128616, + "Ġà¹Ģà¸ŀืà¹Īà¸Ń": 128840, + "Ġà¹ģละ": 128258, + "Ġà¹ģลà¹īว": 129775, + "Ġà¹ģà¸ķà¹Ī": 125442, + "Ġà¹ģà¸ķà¹Īà¸ģà¹ĩ": 139599, + "Ġà¹ģà¸ķà¹Īà¸ĸà¹īา": 141286, + "Ġà¹Ĥà¸Ķย": 126235, + "Ġà¹Ĥà¸Ķยมี": 136931, + "Ġà¹ĥà¸Ļ": 128579, + "Ġà¹Ħมà¹Ī": 129214, + "Ġà¹Ħà¸Ķà¹ī": 126526, + "Ġà¹Ħà¸Ķà¹īà¹ģà¸ģà¹Ī": 140461, + "Ġà¹Ĩ": 128630, + "Ġá": 86089, + "Ġảnh": 128492, + "ĠẤ": 142690, + "ĠẤn": 142691, + "Ġấy": 129331, + "ĠỦ": 142393, + "ĠỦy": 142394, + "Ġá»§ng": 141718, + "Ġá»§y": 137930, + "Ġứng": 128740, + "Ġá»ķ": 133712, + "Ġá»ķn": 133713, + "Ġợ": 128269, + "Ġâ": 2858, + "ĠâĢ": 636, + "ĠâĢ¢": 7288, + "Ġâ̦": 4593, + "Ġâ̦.": 65521, + "Ġâ̦ĊĊ": 12022, + "ĠâĢ«": 131530, + "ĠâĢº": 45633, + "ĠâĢ»": 83399, + "ĠâĢĭ": 29748, + "ĠâĢĭâĢĭ": 72363, + "ĠâĢİ": 53564, + "ĠâĢĵ": 1365, + "ĠâĢĵĊĊ": 46796, + "ĠâĢĶ": 1959, + "ĠâĢĶĊĊ": 62877, + "ĠâĢķ": 50653, + "ĠâĢĺ": 3369, + "ĠâĢĻ": 17744, + "ĠâĢľ": 1036, + "ĠâĢľ[": 65750, + "ĠâĢľâ̦": 91013, + "ĠâĢĿ": 18987, + "ĠâĢĿĊĊ": 66786, + "ĠâĢŀ": 14835, + "Ġâģ": 77137, + "ĠâĤ¬": 12984, + "ĠâĤ¹": 89791, + "ĠâĦ": 28653, + "ĠâĦķ": 57201, + "ĠâĦĸ": 96890, + "ĠâĦĿ": 86023, + "ĠâĨ": 9035, + "ĠâĨIJ": 47464, + "ĠâĨij": 46363, + "ĠâĨĴ": 11397, + "ĠâĨĴĊĊ": 28429, + "ĠâĨĵ": 77854, + "ĠâĨĶ": 77650, + "Ġâĩ": 46018, + "ĠâĩĴ": 58703, + "ĠâĪ": 11995, + "ĠâΧ": 73978, + "ĠâĪĢ": 54700, + "ĠâĪĪ": 48335, + "ĠâĪĴ": 24196, + "Ġâī": 21103, + "Ġâī¤": 37294, + "Ġâī¥": 62147, + "Ġâīł": 94304, + "ĠâĬ": 53025, + "Ġâĵĺ": 89800, + "ĠâĶ": 22612, + "ĠâĶĤ": 33391, + "ĠâĶľ": 71678, + "ĠâĶľâĶĢâĶĢ": 80493, + "Ġâķ": 58208, + "Ġâĸ": 14520, + "Ġâĸº": 65211, + "ĠâĸĪ": 32588, + "ĠâĸĪâĸĪ": 56773, + "ĠâĸĪâĸĪâĸĪâĸĪâĸĪ": 92329, + "Ġâĸij": 90813, + "Ġâĸł": 49450, + "ĠâĹ": 55729, + "ĠâĹĭ": 129183, + "ĠâĹı": 50461, + "Ġâĺ": 25125, + "Ġâĺħ": 37234, + "ĠâĺĨ": 92848, + "ĠâĻ": 35358, + "ĠâĻ¥": 67579, + "Ġâľ": 25521, + "Ġâľĵ": 52375, + "ĠâľĶ": 57156, + "ĠâĿ": 51414, + "ĠâĿ¤": 70470, + "Ġâŀ": 82706, + "ĠâŁ": 34410, + "Ġ⣨": 42005, + "ĠãĢĤ": 59133, + "ĠãĢĮ": 44531, + "ĠãĢIJ": 33576, + "Ġãģ®": 96618, + "Ġãģĵãģ®": 129770, + "ĠãĤ": 17587, + "ĠãĤ¢": 93532, + "ĠãĤ³": 66942, + "ĠãĤ¹": 78950, + "ĠãĤĴ": 94271, + "Ġãĥ": 20711, + "Ġãĥ»": 62315, + "ĠãĥĪãĥ©": 143983, + "ĠãĥĪãĥ©ãĥĥãĤ¯": 143984, + "ĠãĥĪãĥ©ãĥĥãĤ¯ãĥIJãĥĥãĤ¯": 143985, + "Ġ主": 89982, + "Ġä¸Ĭ": 64118, + "Ġä¸ĭ": 39095, + "Ġä¸ĭåįĪ": 56904, + "Ġä¸į": 86009, + "Ġä¸Ń": 72858, + "Ġ使ç͍": 85658, + "Ġä¿®æĶ¹": 63071, + "Ġå": 4891, + "Ġå¤": 40666, + "Ġå¦Ĥ": 69372, + "Ġå¦Ĥæŀľ": 81263, + "Ġå®": 41479, + "Ġå®ŀ": 92293, + "Ġå¯": 48754, + "Ġ对": 69162, + "Ġå°": 58230, + "Ġå±": 79621, + "Ġå·": 83002, + "Ġå¹": 74577, + "Ġåº": 95522, + "Ġå¼": 52506, + "Ġå¼Ģ": 81947, + "Ġå½": 65197, + "Ġå½ĵ": 84897, + "Ġå¾": 84141, + "Ġå¿": 88940, + "Ġåħ": 34369, + "Ġåħ³": 91417, + "ĠåĨ": 68739, + "Ġåĩ": 65727, + "ĠåĪ": 19468, + "ĠåĪĨ": 58657, + "ĠåĪĽå»º": 47758, + "ĠåĪĿå§ĭåĮĸ": 76090, + "ĠåĪłéϤ": 63073, + "ĠåĬ": 54599, + "ĠåĬł": 93883, + "ĠåĮ": 94305, + "Ġåį": 66521, + "Ġåı": 26853, + "ĠåıĤæķ°": 70969, + "Ġåıij": 69425, + "ĠåIJ": 38433, + "ĠåIJį": 61105, + "ĠåĴĮ": 58143, + "Ġåķ": 78971, + "ĠåķĨ": 83310, + "ĠåķĨåĵģ": 95307, + "ĠåĽ": 49602, + "ĠåĽ¾": 90867, + "Ġåľ": 53599, + "Ġåľ¨": 73562, + "ĠåŃ": 53497, + "ĠåŃĹ": 73312, + "Ġæ": 6567, + "Ġæ³": 59053, + "Ġ注": 62579, + "Ġ注æĦı": 97161, + "Ġæµ": 98313, + "Ġæ·»åĬł": 82339, + "Ġæ¸": 97259, + "ĠæĢ": 90476, + "ĠæĪ": 49434, + "ĠæĪIJ": 86546, + "ĠæĪĸ": 92313, + "Ġæī": 46750, + "ĠæĬ": 96155, + "Ġæİ": 46602, + "Ġæİ§": 80851, + "Ġæı": 55059, + "ĠæıIJ": 93685, + "ĠæĶ": 80090, + "Ġæķ": 43614, + "Ġæķ°": 47685, + "Ġæķ°æį®": 62262, + "Ġæĸ": 31286, + "Ġæĸ°": 78919, + "Ġæĸ¹": 71773, + "Ġæĸ¹æ³ķ": 81454, + "Ġæĸĩ": 53040, + "Ġæĸĩä»¶": 69594, + "ĠæĹ¥": 75402, + "ĠæĹ¶éĹ´": 91218, + "Ġæĺ": 38903, + "Ġæĺ¯": 54851, + "Ġæĺ¯åIJ¦": 71951, + "ĠæĽ": 65456, + "ĠæĽ´": 72669, + "ĠæĽ´æĸ°": 78103, + "ĠæľĢ": 82539, + "ĠæĿ": 60596, + "Ġæŀ": 93920, + "ĠæŁ¥": 69353, + "ĠæŁ¥è¯¢": 80528, + "Ġæł": 51461, + "ĠæŃ": 71928, + "Ġç": 10236, + "Ġç¤": 76952, + "Ġ示": 79704, + "Ġ示ä¾ĭ": 80426, + "Ġç§": 75671, + "Ġç¬": 48840, + "Ġ第": 50331, + "Ġç±": 68433, + "Ġç±»": 69674, + "Ġç»": 44054, + "Ġç»ĵ": 89254, + "Ġç¼": 84238, + "Ġçī": 64388, + "Ġç͍æĪ·": 50042, + "ĠçĶŁ": 58263, + "ĠçĶŁåij½åij¨æľŁ": 63079, + "ĠçĶŁåij½åij¨æľŁåĩ½æķ°": 63860, + "ĠçĻ": 68294, + "ĠçĻ»": 91427, + "ĠçļĦ": 43589, + "ĠçĽ": 74866, + "Ġè": 8908, + "Ġè¡": 77407, + "Ġè§": 54955, + "Ġè§£": 85122, + "Ġè®": 33424, + "Ġ设": 46466, + "Ġ设置": 53054, + "Ġ请æ±Ĥ": 80876, + "Ġè¾": 42849, + "Ġè¾ĵ": 44372, + "Ġè¾ĵåħ¥": 69058, + "Ġè¾ĵåĩº": 70568, + "Ġè¿": 32181, + "Ġè¿ĶåĽŀ": 55616, + "Ġèĩ": 58464, + "Ġèĩª": 61991, + "ĠèĩªåĬ¨çĶŁæĪIJ": 97028, + "Ġèĭ": 83226, + "Ġèĭ¥": 92498, + "Ġèĭ¥è¦ģ": 98396, + "Ġèİ·åıĸ": 40671, + "Ġé": 18137, + "Ġ页éĿ¢": 74159, + "Ġé»": 72036, + "Ġé»ĺ认": 80546, + "ĠéĶ": 93178, + "Ġê": 8620, + "Ġê°": 16778, + "Ġê°Ģ": 35509, + "Ġê°ĢëĬ¥": 95351, + "Ġê°ĢëĬ¥ìĦ±": 142724, + "Ġê°ĢëĬ¥íķľ": 143964, + "Ġê°Ģì§Ģ": 127154, + "Ġê°Ģì§Ģê³ł": 131670, + "Ġê°Ģì¹ĺ": 140890, + "Ġê°Ģìļ´": 142179, + "Ġê°Ģìļ´ëį°": 142180, + "Ġê°Ģìŀ¥": 130887, + "Ġê°Ģìł¸": 88846, + "Ġê°ģ": 126804, + "Ġê°ģì¢ħ": 142684, + "Ġê°IJ": 129423, + "Ġê°Ĵ": 45563, + "Ġê°ĴìĿĦ": 93668, + "Ġê°ķ": 129413, + "Ġê°ĸ": 143143, + "Ġê°ĸê³ł": 143144, + "Ġê°Ļ": 78374, + "Ġê°Ļëĭ¤": 135898, + "Ġê°ĻìĬµëĭĪëĭ¤": 138691, + "Ġê°ĻìĿ´": 131050, + "Ġê°ĻìĿĢ": 129381, + "Ġê°ľ": 73523, + "Ġê°ľëħIJ": 142497, + "Ġê°ľìµľ": 143497, + "Ġê°ľìĿ¸": 126799, + "Ġê°ľìĿ¸ìłķë³´": 130096, + "Ġê°Ŀì²´": 75232, + "Ġê±": 97172, + "Ġê±°": 126352, + "Ġê±°ìĿĺ": 138313, + "Ġê±´": 130270, + "Ġ걸": 131961, + "Ġê²": 23894, + "Ġê²°": 82619, + "Ġê²°ê³¼": 98801, + "Ġê²°êµŃ": 136724, + "Ġê²°ìłķ": 135763, + "Ġê²½": 43115, + "Ġê²½ìļ°": 49052, + "Ġê²½ìļ°ìĹIJëĬĶ": 134832, + "Ġê²½íĹĺ": 138704, + "Ġê²Ģ": 85322, + "Ġê²ģ": 141693, + "Ġê²ģëĭĪëĭ¤": 141694, + "Ġê²ĥ": 71108, + "Ġê²ĥëıĦ": 137638, + "Ġê²ĥìľ¼ë¡ľ": 132091, + "Ġê²ĥìĿ´": 128584, + "Ġê²ĥìĿ´ëĭ¤": 128900, + "Ġê²ĥìĿĢ": 129274, + "Ġê²ĥìĿĦ": 129337, + "Ġê²ĥìŀħëĭĪëĭ¤": 130882, + "Ġê²Į": 98927, + "Ġê³": 45130, + "Ġê³µ": 125466, + "Ġê³µê°Ħ": 142496, + "Ġê³µëıĻ": 140429, + "Ġê³Ħ": 94203, + "Ġê³ĦìĨį": 133788, + "Ġê³ł": 126429, + "Ġê³łê°Ŀ": 138501, + "Ġê´Ģ": 92751, + "Ġê´Ģ볨": 129985, + "Ġê´Ģìĭ¬": 138449, + "Ġê´Ģíķľ": 130207, + "Ġêµ": 124459, + "Ġ구": 58777, + "Ġ구ìĦ±": 136239, + "Ġêµī": 142660, + "Ġêµīìŀ¥": 142661, + "Ġêµīìŀ¥íŀĪ": 142662, + "ĠêµIJ": 127048, + "ĠêµIJìĪĺ": 142220, + "ĠêµŃ": 124973, + "ĠêµŃê°Ģ": 133152, + "ĠêµŃ민": 135138, + "ĠêµŃëĤ´": 141185, + "ĠêµŃìłľ": 138249, + "ĠêµŃíļĮ": 143333, + "Ġê¶": 142451, + "Ġê¶Į": 142452, + "Ġê¶Į리": 142453, + "Ġê·": 134312, + "Ġê·¸": 54825, + "Ġê·¸ê²ĥ": 126038, + "Ġê·¸ê²ĥìĿĢ": 136742, + "Ġê·¸ê²ĥìĿĦ": 139484, + "Ġ그리": 126342, + "Ġê·¸ë¦¬ê³ł": 128719, + "Ġ그리ìĬ¤": 142103, + "Ġ그리ìĬ¤ëıĦ": 142104, + "Ġê·¸ëĥ¥": 139456, + "Ġê·¸ëŀĺ": 129547, + "Ġê·¸ëŀĺìĦľ": 129548, + "Ġê·¸ëŁ¬": 125261, + "Ġê·¸ëŁ¬ë©´": 135243, + "Ġê·¸ëŁ¬ëĤĺ": 130549, + "Ġê·¸ëŁ¬ëĭĪê¹Į": 141306, + "Ġê·¸ëŁ°": 126615, + "Ġê·¸ëŁ°ëį°": 130790, + "Ġê·¸ëŁ¼": 141357, + "Ġê·¸ëłĩ": 125971, + "Ġê·¸ëłĩê²Į": 132330, + "Ġê·¸ëłĩì§Ģ": 143600, + "Ġê·ľ": 134313, + "Ġê·ľìłķ": 134314, + "Ġê¸": 40771, + "Ġ기": 54116, + "Ġ기본": 136089, + "Ġ기ìĸµ": 140914, + "Ġ기íĥĢ": 137734, + "Ġê¸Ģ": 131522, + "Ġê¹": 130507, + "Ġê¹Ģ": 130508, + "Ġë": 5140, + "Ġë¡": 70585, + "Ġë¡ľ": 71015, + "Ġ리": 56983, + "Ġ리ìĬ¤íĬ¸": 83634, + "Ġë§": 32985, + "Ġ매": 126932, + "Ġ매ìļ°": 137016, + "Ġë§Ī": 95577, + "Ġë§Ī볨": 140175, + "Ġë§Īì§Ģ": 140886, + "Ġë§Īì§Ģë§ī": 140887, + "Ġë§ĪìĿĮ": 131766, + "Ġë§Į": 62107, + "Ġë§ĮëĤĺ": 142353, + "Ġë§Įëĵ¤": 127579, + "Ġë§Įëĵ¤ìĸ´": 134108, + "Ġë§Įìķ½": 139327, + "Ġë§İ": 126735, + "Ġë§İìĿ´": 130966, + "Ġë§İìĿĢ": 129875, + "Ġë§IJ": 126254, + "Ġë§IJìĶĢ": 129907, + "Ġë§ŀ": 131417, + "Ġë¨": 137767, + "Ġ먼": 137768, + "Ġ먼ìłĢ": 137769, + "Ġë©": 48108, + "Ġë©Ķ": 51391, + "Ġëª": 36978, + "Ġ모": 54070, + "Ġ모르": 136105, + "Ġ모ëijIJ": 130593, + "Ġ모ëĵł": 129304, + "Ġ모ìĬµ": 136665, + "Ġ목": 134952, + "Ġ목ìłģ": 134953, + "Ġ못": 129293, + "Ġëªħ": 130345, + "Ġë¬": 133552, + "Ġ무": 125149, + "Ġ무ë£Į": 139737, + "Ġ무ìĬ¨": 142374, + "Ġ무ìĹĩ": 132779, + "Ġ문": 53435, + "Ġ문ìĦľ": 139236, + "Ġ문ìŀIJ": 79921, + "Ġë¬¸ìłľ": 126674, + "Ġë¬¸ìłľê°Ģ": 137903, + "Ġ물": 133553, + "Ġë¬¼ë¡ł": 133554, + "Ġë¯": 125713, + "Ġ미": 125714, + "Ġ미êµŃ": 132662, + "Ġ미ëŀĺ": 143005, + "Ġë°": 22042, + "Ġë°©": 74808, + "Ġ방문": 141875, + "Ġë°©ë²ķ": 130864, + "Ġë°°": 73669, + "Ġë°°ìĹ´": 90960, + "Ġë°ı": 128355, + "Ġë°Ķ": 81718, + "Ġë°Ķë¡ľ": 130737, + "Ġë°Ķëŀį": 136924, + "Ġë°ĶëŀįëĭĪëĭ¤": 136925, + "Ġë°ĺ": 63757, + "Ġë°ĺëĵľ": 141761, + "Ġë°ĺëĵľìĭľ": 141762, + "Ġë°ĺíĻĺ": 94879, + "Ġë°Ľ": 83596, + "Ġë°ĽìķĦ": 135839, + "Ġë°ľ": 95996, + "Ġë°ľê²¬": 142300, + "Ġë°ľìĥĿ": 133396, + "Ġë°ľíijľ": 142234, + "Ġë²": 47665, + "Ġë²Ħ": 86831, + "Ġë²Ī": 84621, + "Ġë²ķ": 126785, + "Ġë²ķë¥ł": 136324, + "Ġë³": 30520, + "Ġë³´": 63332, + "Ġ보기": 138540, + "Ġë³´ë©´": 132513, + "Ġë³´ëĤ´": 141836, + "Ġ본": 129238, + "Ġë³Ģ": 46319, + "Ġë³Ģê²½": 87737, + "Ġë³ĢìĪĺ": 80740, + "Ġë³ĢíĻĶ": 137525, + "Ġë¶": 62740, + "Ġë¶Ģ": 85403, + "Ġë¶Ģë¶Ħ": 129881, + "Ġë¶ģ": 139963, + "Ġë¶ģíķľ": 139964, + "Ġë¶Ħ": 128618, + "Ġë¶Ī": 126488, + "Ġë¶Ī구": 143681, + "Ġë¶Ī구íķĺê³ł": 143682, + "Ġë¹Ħ": 73986, + "Ġë¹ĦêµIJ": 142350, + "ĠëĤ": 37195, + "ĠëĤ¨": 129624, + "ĠëĤ´": 66136, + "ĠëĤ´ê°Ģ": 131277, + "ĠëĤ´ìļ©": 130213, + "ĠëĤĺ": 73518, + "ĠëĤĺëĬĶ": 132311, + "ĠëĤĺëĿ¼": 137449, + "ĠëĤĺìĺ¤": 137298, + "ĠëĤĺíĥĢ": 135513, + "ĠëĦ": 65553, + "ĠëĦĪ": 127454, + "ĠëĦĪ무": 133267, + "ĠëĦĪíĿ¬": 139616, + "ĠëĦĺ": 143835, + "ĠëĦĺìĸ´": 143836, + "Ġëħ": 127041, + "Ġëħ¸": 127042, + "Ġëħ¸ëł¥": 136111, + "ĠëĪ": 138658, + "ĠëĪĦ": 138659, + "ĠëĪĦ구": 138660, + "ĠëĬ": 143861, + "ĠëĬIJ": 143862, + "ĠëĬIJëĤĮ": 143863, + "Ġëĭ": 34143, + "Ġëĭ¤": 49367, + "Ġëĭ¤ë¥´": 142512, + "Ġëĭ¤ë¥¸": 128772, + "Ġëĭ¤ë§Į": 139293, + "Ġëĭ¤ìĭľ": 131170, + "Ġëĭ¤ìĸij": 135391, + "Ġëĭ¤ìĸijíķľ": 135392, + "Ġëĭ¤ìĿĮ": 126844, + "Ġëĭ¤ìĿĮê³¼": 139107, + "Ġëĭ¨": 129400, + "Ġëĭµ": 143603, + "Ġëĭµë³Ģ": 143604, + "Ġëĭ¹": 125834, + "Ġëĭ¹ìĭľ": 140730, + "Ġëĭ¹ìĭł": 131502, + "ĠëĮ": 143241, + "ĠëĮĢ": 60960, + "ĠëĮĢë¶Ģ": 140093, + "ĠëĮĢë¶Ģë¶Ħ": 140094, + "ĠëĮĢíĨµ": 136773, + "ĠëĮĢíĨµëł¹": 136774, + "ĠëĮĢíķ´": 131978, + "ĠëĮĢíķ´ìĦľ": 130869, + "ĠëĮĢíķĺìŬ": 135878, + "ĠëĮĢíķľ": 128605, + "ĠëĮĵ": 143242, + "ĠëĮĵê¸Ģ": 143243, + "Ġëį": 126365, + "Ġëį°ìĿ´íĦ°": 54248, + "ĠëįĶ": 126366, + "ĠëįĶìļ±": 140084, + "Ġëı": 64805, + "ĠëıĦ": 129392, + "ĠëıĮ": 137843, + "ĠëıĮìķĦ": 137844, + "ĠëıĻ": 126322, + "ĠëıĻìĭľ": 143408, + "ĠëıĻìĭľìĹIJ": 143409, + "ĠëIJ": 72344, + "ĠëIJĺ": 97143, + "ĠëIJľ": 130722, + "ĠëIJľëĭ¤": 130723, + "ĠëIJł": 130647, + "ĠëijIJ": 129419, + "Ġëĵ": 55838, + "Ġëĵ¤": 129900, + "Ġëĵ¤ìĸ´": 129901, + "Ġëĵ±": 77002, + "Ġëĵ±ìĿĦ": 134454, + "Ġëĵ±ìĿĺ": 136357, + "ĠëĶ": 76497, + "ĠëͰ": 125686, + "ĠëĶ°ë¥¸": 136707, + "ĠëͰëĿ¼": 126629, + "ĠëͰëĿ¼ìĦľ": 134084, + "ĠëķĮ": 53618, + "ĠëķĮ문": 126495, + "ĠëķĮ문ìĹIJ": 129147, + "ĠëķĮ문ìĿ´ëĭ¤": 137141, + "Ġëĺ": 125639, + "ĠëĺIJ": 125640, + "ĠëĺIJëĬĶ": 129549, + "ĠëĺIJíķľ": 130005, + "Ġì": 3315, + "Ġì¡°": 65510, + "Ġì¡°ê¸Ī": 134422, + "Ġì¡°ì§ģ": 141655, + "Ġì¡°íļĮ": 97167, + "Ġì¡´": 134014, + "Ġì¡´ìŀ¬": 134015, + "Ġì¢": 65699, + "Ġì¢Ģ": 130572, + "Ġì¢ħ": 98358, + "Ġì¢ĭ": 125953, + "Ġì¢ĭìķĦ": 138779, + "Ġì¢ĭìĿĢ": 132968, + "Ġ주": 55673, + "Ġ주ìŀ¥": 140200, + "Ġì¤": 53989, + "Ġì¤Ģ": 138267, + "Ġì¤Ģë¹Ħ": 138268, + "Ġì¤ij": 70943, + "Ġì¤ijêµŃ": 138373, + "Ġì¤ijìĭ¬": 141482, + "Ġì¤ijìļĶ": 127840, + "Ġì¤ijìļĶíķľ": 136361, + "Ġì¦": 132376, + "Ġì¦ī": 132377, + "Ġì§": 47911, + "Ġì§Ģ": 66790, + "Ġì§Ģê¸Ī": 129082, + "Ġì§ĢëĤľ": 133146, + "Ġì§ĢìĹŃ": 131937, + "Ġì§ĢìĽIJ": 132185, + "Ġì§ģ": 137351, + "Ġì§ģìłij": 137352, + "Ġì§Ħ": 126616, + "Ġì§Ħíĸī": 132876, + "Ġì§Ī": 138519, + "Ġì§Ī문": 138520, + "Ġì§ij": 130263, + "Ġì°": 62099, + "Ġì°¨": 129882, + "Ġì°¸": 127969, + "Ġì°¸ê³ł": 142616, + "Ġì°¸ìŬ": 138098, + "Ġì°¾": 138037, + "Ġì°¾ìķĦ": 138038, + "Ġì±ħ": 138001, + "Ġì±ħìŀĦ": 138002, + "Ġì²": 48364, + "Ġì²ĺ": 71647, + "Ġì²ĺ리": 81058, + "Ġì²ĺìĿĮ": 136065, + "Ġì´Ī": 83315, + "Ġì´Ī기": 95313, + "Ġìµľ": 81173, + "Ġìµľê·¼": 139465, + "Ġì¶": 36330, + "Ġì¶Ķ": 57835, + "Ġì¶Ķê°Ģ": 68408, + "Ġì¶Ķì§Ħ": 141210, + "Ġì¶ľëł¥": 61126, + "Ġì·¨": 131565, + "Ġì¹": 90711, + "Ġì»": 89095, + "Ġì»´": 143520, + "Ġì»´íĵ¨": 143521, + "Ġì»´íĵ¨íĦ°": 143522, + "Ġì½Ķëĵľ": 91605, + "ĠìĤ": 127165, + "ĠìĤ¬": 32129, + "ĠìĤ¬ê±´": 138629, + "ĠìĤ¬ëŀ": 124693, + "ĠìĤ¬ëŀĮ": 124889, + "ĠìĤ¬ëŀĮëĵ¤": 131508, + "ĠìĤ¬ëŀĮëĵ¤ìĿ´": 135990, + "ĠìĤ¬ëŀĮìĿ´": 133166, + "ĠìĤ¬ëŀĮìĿĢ": 140173, + "ĠìĤ¬ëŀij": 132488, + "ĠìĤ¬ì§Ħ": 139764, + "ĠìĤ¬ìĭ¤": 130037, + "ĠìĤ¬ìĹħ": 131608, + "ĠìĤ¬ìļ©": 40720, + "ĠìĤ¬ìĿ´": 136541, + "ĠìĤ¬íļĮ": 130592, + "ĠìĤ´": 127166, + "ĠìĤ´ìķĦ": 137879, + "ĠìĤŃìłľ": 76592, + "Ġìĥ": 134582, + "Ġìĥģ": 58034, + "Ġìĥģíĥľ": 137619, + "ĠìĥĪ": 134583, + "ĠìĥĪë¡ľ": 134584, + "ĠìĥĪë¡ľìļ´": 134585, + "ĠìĥĿ": 47818, + "ĠìĥĿê°ģ": 128646, + "ĠìĥĿëªħ": 141659, + "ĠìĥĿìĤ°": 141091, + "ĠìĥĿìĦ±": 51917, + "ĠìĦ": 27767, + "ĠìĦ¤": 57852, + "ĠìĦ¤ëªħ": 133828, + "ĠìĦ¤ì¹ĺ": 134739, + "ĠìĦ¤ìłķ": 65880, + "ĠìĦ±": 128677, + "ĠìĦ¸": 125674, + "ĠìĦ¸ê³Ħ": 133196, + "ĠìĦ¸ìĥģ": 134976, + "ĠìĦľ": 89860, + "ĠìĦľë¡ľ": 137765, + "ĠìĦľë¹ĦìĬ¤": 130778, + "ĠìĦľìļ¸": 136905, + "ĠìĦł": 129296, + "ĠìĦłíĥĿ": 86038, + "ĠìĨ": 77596, + "ĠìĨĮ": 126291, + "ĠìĨĮê°ľ": 142559, + "ĠìĪ": 69192, + "ĠìĪĺ": 28733, + "ĠìĪĺëıĦ": 134013, + "ĠìĪĺì§ij": 140458, + "ĠìĪĺìłķ": 88513, + "ĠìĪĺíĸī": 139393, + "ĠìĪľ": 143802, + "ĠìĪľê°Ħ": 143803, + "Ġìī": 142067, + "Ġìī½": 142068, + "Ġìī½ê²Į": 142069, + "ĠìĬ": 79207, + "ĠìĬ¤": 141767, + "ĠìĬ¤ìĬ¤": 141768, + "ĠìĬ¤ìĬ¤ë¡ľ": 141769, + "Ġìĭ": 28927, + "Ġìĭ¤": 126423, + "Ġìĭ¤ìłľ": 142700, + "Ġìĭ¤ìłľë¡ľ": 142701, + "Ġìĭ¤í": 60985, + "Ġìĭ¤íĸī": 85788, + "Ġìĭľ": 44518, + "Ġìĭľê°Ħ": 130217, + "Ġìĭľìŀij": 93721, + "Ġìĭľíĸī": 138767, + "Ġìĭł": 128753, + "Ġìķ": 23872, + "ĠìķĦ": 48408, + "ĠìķĦë§Ī": 140749, + "ĠìķĦ무": 134006, + "ĠìķĦëĭĪ": 126523, + "ĠìķĦëĭĪëĭ¤": 139429, + "ĠìķĦëĭĪëĿ¼": 130651, + "ĠìķĦëĭĮ": 134808, + "ĠìķĦëŀĺ": 136646, + "ĠìķĦ주": 138652, + "ĠìķĦì§ģ": 137289, + "ĠìķĦìĿ´": 130902, + "ĠìķĪ": 95170, + "ĠìķĬ": 50696, + "ĠìķĬê³ł": 135201, + "ĠìķĬëĬĶ": 127728, + "ĠìķĬëĬĶëĭ¤": 137764, + "ĠìķĬìĬµëĭĪëĭ¤": 138539, + "ĠìķĬìķĺ": 133995, + "ĠìķĬìĿĢ": 136278, + "ĠìķĬìĿĦ": 136982, + "ĠìķĮ": 125214, + "ĠìķĮê³ł": 135420, + "ĠìķĮ볤": 138630, + "ĠìķĮìķĦ": 137676, + "Ġìķŀ": 139274, + "Ġìķŀìľ¼ë¡ľ": 139275, + "Ġìĸ": 79302, + "Ġìĸ´": 124685, + "Ġìĸ´ëĬIJ": 133487, + "Ġìĸ´ëĶĶ": 139740, + "Ġìĸ´ëĸ": 126537, + "Ġìĸ´ëĸ¤": 129273, + "Ġìĸ´ëĸ»": 130107, + "Ġìĸ´ëĸ»ê²Į": 130108, + "Ġìĸ´ëłµ": 143337, + "Ġìĸ¸": 139957, + "Ġìĸ¸ìłľ": 139958, + "Ġìĸ¼": 138665, + "Ġìĸ¼ë§Ī": 138666, + "Ġìĸ¼ë§ĪëĤĺ": 138667, + "Ġìĸĺ": 135939, + "Ġìĸĺ기": 135940, + "ĠìĹ": 24485, + "ĠìŬ": 83518, + "ĠìĹ¬ê¸°": 131180, + "ĠìŬ룬": 127296, + "ĠìŬ룬ë¶Ħ": 134637, + "ĠìŬìĦ±": 140546, + "ĠìŬíĸī": 142815, + "ĠìŰ": 77353, + "ĠìĹ°ê²°": 141819, + "ĠìĹ°êµ¬": 131698, + "ĠìĹ´": 130847, + "ĠìĹĨ": 46682, + "ĠìĹĨëĬĶ": 130768, + "ĠìĹĨëĭ¤": 130671, + "ĠìĹĨìĬµëĭĪëĭ¤": 136673, + "ĠìĹIJ": 90486, + "ĠìĹŃ": 127864, + "ĠìĹŃìĤ¬": 136608, + "ĠìĹŃìĭľ": 139512, + "Ġìĺ": 38523, + "Ġìĺ¤": 73077, + "Ġìĺ¤ëĬĺ": 133857, + "Ġìĺ¤í": 142209, + "Ġìĺ¤íŀ": 142210, + "Ġìĺ¤íŀĪ": 142211, + "Ġìĺ¤íŀĪ볤": 142212, + "Ġìĺģ": 126440, + "ĠìĺģíĻĶ": 137116, + "ĠìĺĪ": 95617, + "ĠìĺĪìĤ°": 138892, + "ĠìĺĪìĪĺ": 128040, + "ĠìĺĪìĪĺëĭĺ": 139174, + "ĠìĻ": 74884, + "Ġìļ": 65722, + "Ġìļ°": 124657, + "Ġìļ°ë¦¬": 124830, + "Ġìļ°ë¦¬ê°Ģ": 131025, + "Ġìļ°ë¦¬ëĤĺ": 142336, + "Ġìļ°ë¦¬ëĤĺëĿ¼": 142337, + "Ġìļ°ë¦¬ëĬĶ": 131746, + "Ġìļ°ë¦¬ìĿĺ": 137808, + "Ġìļ°ìĦł": 140680, + "Ġìļ´": 132028, + "Ġìļ´ìĺģ": 132029, + "ĠìļĶ": 85997, + "ĠìļĶ구": 135797, + "ĠìļĶì²Ń": 142163, + "ĠìĽ": 85413, + "ĠìĽIJ": 129093, + "Ġìľ": 37064, + "ĠìľĦ": 45710, + "ĠìľĦì¹ĺ": 98987, + "ĠìľĦìĽIJ": 143128, + "ĠìľĦìĽIJëĭĺ": 143129, + "ĠìľĦíķ´": 130039, + "ĠìľĦíķ´ìĦľ": 133083, + "ĠìľĦíķĺìŬ": 133621, + "ĠìľĦíķľ": 130679, + "Ġìľł": 126310, + "Ġìľłì§Ģ": 139871, + "ĠìĿ": 16751, + "ĠìĿ´": 23084, + "ĠìĿ´ê²ĥ": 128022, + "ĠìĿ´ê²ĥìĿ´": 143419, + "ĠìĿ´ê²ĥìĿĢ": 138849, + "ĠìĿ´ê²Į": 139365, + "ĠìĿ´ë£¨": 139646, + "ĠìĿ´ë£¨ìĸ´": 139647, + "ĠìĿ´ë¥¼": 136342, + "ĠìĿ´ë¦Ħ": 86034, + "ĠìĿ´ë¯¸": 90667, + "ĠìĿ´ë²Ī": 132183, + "ĠìĿ´ë٬": 131366, + "ĠìĿ´ë٬íķľ": 131367, + "ĠìĿ´ëٰ": 129007, + "ĠìĿ´ëłĩ": 130652, + "ĠìĿ´ëłĩê²Į": 130653, + "ĠìĿ´ìĥģ": 130408, + "ĠìĿ´ìķ¼": 130860, + "ĠìĿ´ìķ¼ê¸°": 130861, + "ĠìĿ´ìļ©": 126563, + "ĠìĿ´ìļ©ìŀIJ": 139436, + "ĠìĿ´ìľł": 132819, + "ĠìĿ´ìłľ": 132911, + "ĠìĿ´íķ´": 133329, + "ĠìĿ´íĽĦ": 136331, + "ĠìĿ¸": 58677, + "ĠìĿ¸ê°Ħ": 133204, + "ĠìĿ¸ìłķ": 139180, + "ĠìĿ¼": 83556, + "ĠìĿ¼ë°ĺ": 134664, + "ĠìĿ¼ë³¸": 136293, + "ĠìĿ¼ë¶Ģ": 136605, + "ĠìĿ¼ëĭ¨": 140667, + "ĠìĿ¼ìĿ´": 136168, + "ĠìĿ¼ìĿĦ": 138507, + "ĠìĿĺ": 124970, + "ĠìĿĺ견": 140264, + "ĠìĿĺ미": 132989, + "ĠìĿĺìĤ¬": 141888, + "ĠìĿĺìĽIJ": 138842, + "ĠìĿĺíķ´": 136009, + "Ġìŀ": 16235, + "Ġìŀ¥": 129359, + "Ġìŀ¬": 129242, + "Ġìŀħ": 38150, + "Ġìŀħëł¥": 42349, + "Ġìŀħìŀ¥": 140585, + "ĠìŀĪ": 35339, + "ĠìŀĪê²Į": 142161, + "ĠìŀĪê³ł": 131891, + "ĠìŀĪ기": 137334, + "ĠìŀĪëĬĶ": 64521, + "ĠìŀĪëĬĶëį°": 134563, + "ĠìŀĪëĭ¤": 90686, + "ĠìŀĪëĭ¤ê³ł": 137119, + "ĠìŀĪëĭ¤ëĬĶ": 135511, + "ĠìŀĪëıĦë¡Ŀ": 136303, + "ĠìŀĪì§Ģ": 141231, + "ĠìŀĪì§Ģë§Į": 141258, + "ĠìŀĪìĬµëĭĪëĭ¤": 128472, + "ĠìŀĪìĸ´": 127353, + "ĠìŀĪìĸ´ìĦľ": 136633, + "ĠìŀĪìĹĪ": 132236, + "ĠìŀĪìĹĪëĭ¤": 137046, + "ĠìŀĪìľ¼": 135354, + "ĠìŀĪìľ¼ë©°": 132931, + "ĠìŀĪìĿĦ": 130689, + "ĠìŀIJ": 64577, + "ĠìŀIJ기": 131823, + "ĠìŀIJ리": 140210, + "ĠìŀIJì²´": 136751, + "ĠìŀIJìĭł": 127218, + "ĠìŀIJìĭłìĿĺ": 134748, + "ĠìŀIJìŰ": 140577, + "ĠìŀIJìľł": 138266, + "Ġìŀij": 68232, + "ĠìŀijìĦ±": 140174, + "ĠìŀijìĹħ": 139495, + "ĠìŀijìĿĢ": 143416, + "ĠìŀijíĴĪ": 143586, + "Ġìŀĺ": 126720, + "Ġìŀĺ못": 142357, + "Ġìł": 18585, + "ĠìłĢ": 125569, + "ĠìłĢëĬĶ": 134561, + "ĠìłĢìŀ¥": 57547, + "ĠìłĢíĿ¬": 132301, + "Ġìłģ": 135968, + "Ġìłģìļ©": 135969, + "ĠìłĦ": 56419, + "ĠìłĦì²´": 137138, + "ĠìłĦíĺĢ": 143552, + "Ġìłij": 143188, + "Ġìłijê·¼": 143189, + "Ġìłķ": 36055, + "Ġìłķë§IJ": 134247, + "Ġìłķë³´": 60039, + "ĠìłķëıĦ": 131219, + "Ġìłķì¹ĺ": 136994, + "Ġìłľ": 62071, + "Ġìłľê°Ģ": 132254, + "Ġìłľê³µ": 130094, + "ĠìłľíĴĪ": 138017, + "Ġí": 10764, + "Ġíģ": 45313, + "Ġíģ¬": 143230, + "Ġíģ¬ê²Į": 143231, + "Ġíģ°": 132182, + "Ġíģ´": 62938, + "Ġíģ´ëŀĺìĬ¤": 93182, + "Ġíĥ": 74361, + "ĠíĨµ": 125206, + "ĠíĨµíķ´": 131582, + "ĠíĨµíķ´ìĦľ": 140853, + "ĠíĨµíķĺìŬ": 142196, + "ĠíĬ": 127819, + "ĠíĬ¹": 127820, + "ĠíĬ¹ìłķ": 142976, + "ĠíĬ¹íŀĪ": 136115, + "ĠíĮ": 45104, + "ĠíĮĮ": 54969, + "ĠíĮĮìĿ¼": 67629, + "ĠíĮIJ": 140568, + "ĠíĮIJëĭ¨": 140569, + "ĠíİĺìĿ´ì§Ģ": 92889, + "Ġíı": 69441, + "Ġíı¬": 98869, + "Ġíı¬íķ¨": 133970, + "Ġíijľ": 139052, + "ĠíijľíĺĦ": 139053, + "ĠíĶ": 142509, + "Ġíͼ": 142510, + "Ġíͼíķ´": 142511, + "ĠíĶĦ": 84255, + "Ġíķ": 20136, + "Ġíķ¨": 50972, + "Ġíķ¨ê»ĺ": 129676, + "Ġíķ¨ìĪĺ": 58440, + "Ġíķ©ëĭĪëĭ¤": 130650, + "Ġíķ´": 60716, + "Ġíķ´ê²°": 138327, + "Ġíķ´ëĭ¹": 94613, + "ĠíķĦ": 75528, + "ĠíķĦìļĶ": 126871, + "ĠíķĦìļĶíķľ": 134028, + "Ġíķĺ": 53900, + "Ġíķĺê³ł": 130127, + "ĠíķĺëĤĺ": 125703, + "ĠíķĺëĤĺëĭĺ": 127024, + "ĠíķĺëĤĺëĭĺìĿĺ": 136299, + "ĠíķĺëĬĶ": 128956, + "Ġíķĺì§Ģë§Į": 130331, + "Ġíķľ": 61298, + "ĠíķľêµŃ": 130092, + "Ġíķľë²Ī": 138903, + "Ġíķľëĭ¤": 129112, + "Ġíķł": 95002, + "ĠíķŃ": 142654, + "ĠíķŃìĥģ": 142655, + "Ġíĸī": 126174, + "Ġíĸīë³µ": 142506, + "ĠíĸīëıĻ": 140692, + "ĠíĸīìĤ¬": 140888, + "Ġíĺ": 47455, + "Ġíĺ¸": 91043, + "Ġíĺ¸ì¶ľ": 95499, + "Ġíĺ¸íħĶ": 139962, + "Ġíĺ¹": 137601, + "Ġíĺ¹ìĿĢ": 137602, + "ĠíĺĦ": 141526, + "ĠíĺĦìĭ¤": 141527, + "ĠíĺĦìŀ¬": 132270, + "Ġíĺķ": 141965, + "Ġíĺķíĥľ": 141966, + "ĠíĻ": 46832, + "ĠíĻĪ": 142786, + "ĠíĻĪíİĺìĿ´ì§Ģ": 142787, + "ĠíĻķ": 130729, + "ĠíĻķìĿ¸": 73859, + "ĠíĻľ": 140968, + "ĠíĻľìļ©": 140969, + "Ġíļ": 141539, + "Ġíļ¨": 141540, + "Ġíļ¨ê³¼": 141541, + "ĠíļĮ": 98005, + "ĠíļĮìĽIJ": 133662, + "ĠíĽĦ": 94315, + "Ġï": 32495, + "Ġ": 75780, + "Ġï¼": 62823, + "Ġï¼Ī": 42344, + "Ġï¼Į": 41175, + "Ġï¼ļ": 30372, + "Ġ�": 29333, + "Ġ����": 59630, + "ĠïĤ": 68298, + "ĠðĿ": 81250, + "ĠðŁ": 11162, + "ĠðŁij": 61804, + "ĠðŁĴ": 63039, + "ĠðŁĶ": 95069, + "ĠðŁĺ": 26525, + "ĠðŁĺĢ": 90316, + "ĠðŁĺī": 55937, + "ĠðŁĺīĊĊ": 81436, + "ĠðŁĻĤ": 27484, + "ĠðŁĻĤĊĊ": 39024, + "Ġĉ": 7018, + "Ġĉĉ": 13043, + "Ġĉĉĉ": 22207, + "Ġĉĉĉĉ": 35627, + "Ġĉĉĉĉĉ": 54712, + "Ġĉĉĉĉĉĉ": 79682, + "ĠĉĉĊ": 98207, + "ĠĉĊ": 17476, + "ĠĉĠ": 65497, + "ĠĉĠĠĠ": 79524, + "ĠĊ": 715, + "ĠĊĉĉĊ": 67392, + "ĠĊĉĊ": 51141, + "ĠĊĊ": 4710, + "ĠĊĊĊ": 14731, + "ĠĊĊĊĊ": 22701, + "ĠĊĊĊĊĊ": 76325, + "ĠĊĠĊ": 31906, + "ĠĊĠĠĊ": 48426, + "ĠĊĠĠĠĠĊ": 36845, + "ĠĊĠĠĠĠĠĠĠĠĊ": 49987, + "ĠĊĠĠĠĠĠĠĠĠĠĠĠĠĊ": 83913, + "ĠčĊ": 2529, + "ĠčĊčĊ": 23341, + "ĠčĊĠčĊ": 90440, + "ĠččĊ": 94691, + "ĠĠ": 256, + "ĠĠĉ": 19271, + "ĠĠĉĉ": 35329, + "ĠĠĉĉĉ": 54833, + "ĠĠĉĉĉĉ": 87728, + "ĠĠĉĊ": 98251, + "ĠĠĉĠ": 78099, + "ĠĠĊ": 2303, + "ĠĠĊĊ": 18611, + "ĠĠĊĊĊ": 79226, + "ĠĠĊĠĠĊ": 30779, + "ĠĠĊĠĠĠĠĊ": 77787, + "ĠĠčĊ": 10419, + "ĠĠčĊčĊ": 72745, + "ĠĠĠ": 262, + "ĠĠĠĉ": 39767, + "ĠĠĠĉĉ": 68546, + "ĠĠĠĊ": 5872, + "ĠĠĠĊĊ": 33933, + "ĠĠĠĊĠĠĠĠĊ": 78672, + "ĠĠĠčĊ": 20338, + "ĠĠĠĠ": 257, + "ĠĠĠĠĉ": 7782, + "ĠĠĠĠĉĉ": 15287, + "ĠĠĠĠĉĉĉ": 29936, + "ĠĠĠĠĉĉĉĉ": 50538, + "ĠĠĠĠĉĉĉĉĉ": 79039, + "ĠĠĠĠĉĉĊ": 85422, + "ĠĠĠĠĉĊ": 34483, + "ĠĠĠĠĉčĊ": 84852, + "ĠĠĠĠĉĠ": 80840, + "ĠĠĠĠĉĠĠĠ": 59949, + "ĠĠĠĠĊ": 1066, + "ĠĠĠĠĊĊ": 14808, + "ĠĠĠĠĊĊĊ": 61439, + "ĠĠĠĠĊĠĠĠĠĊ": 9401, + "ĠĠĠĠĊĠĠĠĠĊĠĠĠĠĊ": 41693, + "ĠĠĠĠĊĠĠĠĠĊĠĠĠĠĊĠĠĠĠĊ": 80719, + "ĠĠĠĠčĊ": 6421, + "ĠĠĠĠčĊčĊ": 58949, + "ĠĠĠĠčĊĠĠĠĠčĊ": 45432, + "ĠĠĠĠĠ": 414, + "ĠĠĠĠĠĉ": 85065, + "ĠĠĠĠĠĊ": 10947, + "ĠĠĠĠĠĊĊ": 74525, + "ĠĠĠĠĠčĊ": 38085, + "ĠĠĠĠĠĠ": 981, + "ĠĠĠĠĠĠĉ": 51370, + "ĠĠĠĠĠĠĉĉ": 86359, + "ĠĠĠĠĠĠĊ": 6926, + "ĠĠĠĠĠĠĊĊ": 60998, + "ĠĠĠĠĠĠĊĠĠĠĠĠĠĊ": 93004, + "ĠĠĠĠĠĠčĊ": 17260, + "ĠĠĠĠĠĠĠ": 286, + "ĠĠĠĠĠĠĠĉ": 88044, + "ĠĠĠĠĠĠĠĊ": 12306, + "ĠĠĠĠĠĠĠĊĊ": 65668, + "ĠĠĠĠĠĠĠčĊ": 39648, + "ĠĠĠĠĠĠĠĠ": 260, + "ĠĠĠĠĠĠĠĠĉ": 17362, + "ĠĠĠĠĠĠĠĠĉĉ": 36920, + "ĠĠĠĠĠĠĠĠĉĉĉ": 64204, + "ĠĠĠĠĠĠĠĠĉĉĉĉ": 91693, + "ĠĠĠĠĠĠĠĠĉĊ": 73196, + "ĠĠĠĠĠĠĠĠĊ": 1789, + "ĠĠĠĠĠĠĠĠĊĊ": 23459, + "ĠĠĠĠĠĠĠĠĊĠĠĠĠĊ": 88998, + "ĠĠĠĠĠĠĠĠĊĠĠĠĠĠĠĠĠĊ": 16885, + "ĠĠĠĠĠĠĠĠĊĠĠĠĠĠĠĠĠĊĠĠĠĠĠĠĠĠĊ": 73363, + "ĠĠĠĠĠĠĠĠčĊ": 8998, + "ĠĠĠĠĠĠĠĠčĊčĊ": 88763, + "ĠĠĠĠĠĠĠĠčĊĠĠĠĠĠĠĠĠčĊ": 73867, + "ĠĠĠĠĠĠĠĠĠ": 688, + "ĠĠĠĠĠĠĠĠĠĊ": 15677, + "ĠĠĠĠĠĠĠĠĠčĊ": 58559, + "ĠĠĠĠĠĠĠĠĠĠ": 1843, + "ĠĠĠĠĠĠĠĠĠĠĊ": 16159, + "ĠĠĠĠĠĠĠĠĠĠčĊ": 62217, + "ĠĠĠĠĠĠĠĠĠĠĠ": 310, + "ĠĠĠĠĠĠĠĠĠĠĠĊ": 19011, + "ĠĠĠĠĠĠĠĠĠĠĠčĊ": 64783, + "ĠĠĠĠĠĠĠĠĠĠĠĠ": 1797, + "ĠĠĠĠĠĠĠĠĠĠĠĠĉ": 35557, + "ĠĠĠĠĠĠĠĠĠĠĠĠĉĉ": 69562, + "ĠĠĠĠĠĠĠĠĠĠĠĠĊ": 3374, + "ĠĠĠĠĠĠĠĠĠĠĠĠĊĊ": 39865, + "ĠĠĠĠĠĠĠĠĠĠĠĠĊĠĠĠĠĠĠĠĠĠĠĠĠĊ": 34642, + "ĠĠĠĠĠĠĠĠĠĠĠĠčĊ": 16075, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1060, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 28247, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 2290, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 26285, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠčĊ": 89160, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 394, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 26546, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠčĊ": 91205, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 338, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĉ": 60123, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 6360, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊĊ": 71248, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 59649, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠčĊ": 27365, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1698, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 41636, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 4293, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 36577, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 503, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 40337, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 3824, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĉ": 96800, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 10683, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 88804, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠčĊ": 45575, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 2549, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 55447, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 5180, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 52720, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 664, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 51124, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 5108, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 15865, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠčĊ": 65317, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 3502, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 75228, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 5968, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 69877, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 999, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 71779, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 6526, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 25435, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 4597, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 96017, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 7561, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 86770, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1383, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 90306, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 786, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 33641, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 6656, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 9699, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1920, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 10589, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 45807, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 8333, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 11070, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 2559, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 12841, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 56596, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 10503, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 14265, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 3344, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 15429, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 77993, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 13463, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 17264, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 4569, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 18749, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 94947, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 17648, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 21509, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 6374, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 24348, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 20974, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 25773, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 8136, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 28802, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 26065, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 32678, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 11950, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 38484, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 31979, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 37083, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 15270, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 5238, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 38171, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 47549, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 19273, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 51068, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 42708, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 46771, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 22335, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 37144, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 46145, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 41056, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 14642, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 55799, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 58591, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 81221, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 39484, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 97417, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 95429, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 51475, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 65271, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 79871, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 86766, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 56940, + "ġ": 221, + "Ģ": 222, + "Ģìŀ¥": 54470, + "ģ": 223, + "Ĥ": 224, + "Ĥ¤": 92120, + "Ĥ¨": 60757, + "Ĥ¬": 7492, + "Ĥ¹": 23272, + "Ĥæķ°": 27260, + "ĤŃ": 64520, + "ĤŃìłľ": 65519, + "ĥ": 225, + "ĥ½": 22597, + "ĥģ": 36109, + "ĥĿ": 29389, + "Ħ": 226, + "Ħä»¶": 93437, + "Ħìŀ¬": 79879, + "ħ": 227, + "ħ§": 91771, + "ħĮ": 72509, + "ħį": 44104, + "Ĩ": 228, + "Ĩµ": 89179, + "ĩ": 229, + "Ī": 230, + "ĪæĿĥ": 81942, + "Īëĭ¤": 19946, + "ĪìĬµëĭĪëĭ¤": 89363, + "Īĺ": 15902, + "ī": 231, + "Ĭ": 232, + "Ĭ¶": 37440, + "Ĭ¶æĢģ": 39366, + "ĭ": 233, + "ĭè¯ķ": 66635, + "Į": 234, + "Į¨": 98642, + "ĮĢ": 39834, + "į": 235, + "į°": 41902, + "į°ìĿ´": 46218, + "į°ìĿ´íĦ°": 50362, + "İ": 236, + "İ·": 25452, + "İ·åıĸ": 27487, + "İĺìĿ´": 69383, + "İĺìĿ´ì§Ģ": 79427, + "ı": 237, + "ıè¿°": 44834, + "IJ": 238, + "IJèĹı": 68172, + "IJľ": 54492, + "IJľç´¢": 68941, + "ij": 239, + "ijIJ": 78910, + "ijľ": 71443, + "Ĵ": 240, + "Ĵáŀ": 72496, + "ĴĮ": 26254, + "ĵ": 241, + "ĵåIJį": 61617, + "ĵį": 34429, + "ĵįä½ľ": 36548, + "ĵĺ": 67895, + "Ķ": 242, + "ĶåĽŀ": 25011, + "ĶĦ": 62544, + "ķ": 243, + "ķãĤĵ": 84567, + "ķìĿ¸": 66261, + "ķĮ": 47378, + "ĸ": 244, + "ĸī": 51275, + "Ĺ": 245, + "Ĺı": 23444, + "ĺ": 246, + "ĺ认": 37509, + "ĺìĿ´": 62618, + "Ļ": 247, + "ύ": 30828, + "϶": 103895, + "ĻĤ": 18535, + "ļ": 248, + "ļĮ": 47379, + "Ľ": 249, + "Ľå»º": 31338, + "ĽĦ": 74209, + "ľ": 250, + "ľëł¥": 57133, + "Ŀ": 251, + "Ŀ¼": 45881, + "Ŀå§ĭ": 43716, + "Ŀå§ĭåĮĸ": 51088, + "Ŀì²´": 73147, + "ŀ": 252, + "ŀæĢ§": 59139, + "ŀĭ": 23781, + "Ł": 253, + "Ł¥": 22129, + "ٳ": 58715, + "Łèĥ½": 96808, + "ł": 254, + "łéϤ": 24045, + "łí": 57160, + "łíĥĿ": 75132, + "łĢ": 63219, + "łģ": 15675, + "Ń": 255, + "ŃIJ": 98734 +} diff --git a/20250121165312/rank0.log b/20250121165312/rank0.log new file mode 100644 index 0000000000000000000000000000000000000000..757d76c141a5b2f261c166b7dbbcf5d6e24ee30a --- /dev/null +++ b/20250121165312/rank0.log @@ -0,0 +1,900 @@ +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:16][INFO] +------------------------------------------------------------ +Runtime environment: + sys.platform: linux + Python: 3.10.16 (main, Dec 11 2024, 16:24:50) [GCC 11.2.0] + CUDA available: True + MUSA available: False + numpy_random_seed: 0 + GPU 0,1,2,3,4,5,6,7: NVIDIA A100-SXM4-80GB + CUDA_HOME: /mnt/petrelfs/share/test-cuda/cuda-12.1 + NVCC: Cuda compilation tools, release 12.1, V12.1.66 + GCC: gcc (GCC) 11.2.0 + PyTorch: 2.4.0 + PyTorch compiling details: PyTorch built with: + - GCC 9.3 + - C++ Version: 201703 + - Intel(R) oneAPI Math Kernel Library Version 2023.1-Product Build 20230303 for Intel(R) 64 architecture applications + - Intel(R) MKL-DNN v3.4.2 (Git Hash 1137e04ec0b5251ca2b4400a4fd3c667ce843d67) + - OpenMP 201511 (a.k.a. OpenMP 4.5) + - LAPACK is enabled (usually provided by MKL) + - NNPACK is enabled + - CPU capability usage: AVX2 + - CUDA Runtime 12.1 + - NVCC architecture flags: -gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_60,code=sm_60;-gencode;arch=compute_61,code=sm_61;-gencode;arch=compute_70,code=sm_70;-gencode;arch=compute_75,code=sm_75;-gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_86,code=sm_86;-gencode;arch=compute_90,code=sm_90 + - CuDNN 90.1 (built against CUDA 12.4) + - Magma 2.6.1 + - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, CUDA_VERSION=12.1, CUDNN_VERSION=9.1.0, CXX_COMPILER=/opt/rh/devtoolset-9/root/usr/bin/c++, CXX_FLAGS= -D_GLIBCXX_USE_CXX11_ABI=0 -fabi-version=11 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=pedantic -Wno-error=old-style-cast -Wno-missing-braces -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, PERF_WITH_AVX512=1, TORCH_VERSION=2.4.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=1, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, + + TorchVision: 0.19.0 + OpenCV: 4.7.0 + MMEngine: 0.10.5 + Transformers: 4.45.1 + XTuner: 0.1.23+657d74 + Seed: 0 + World Size: 64 + DP Size: 16 + SP Size: 4 + TP Size: 1 +------------------------------------------------------------ + +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:19][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:36][INFO] [Dataset] (Original) 994065 samples. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:36][INFO] [Dataset] (Packed) 37907 samples. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:36][INFO] [Dataloader] 2372 batches. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:37][DEBUG] [Dataloader] Training Batch: +{'input_ids': tensor([[151644, 872, 198, ..., 13, 151645, 198]]), 'labels': tensor([[ -100, -100, -100, ..., 13, 151645, -100]]), 'num_tokens': tensor([ 466, 1305, 1323, 393, 2148, 1594, 305, 256, 2759, 2079, 476, 2275, + 22, 2552, 565, 86, 743, 79, 402, 855, 3265, 608, 549, 2220, + 1421, 321, 1342, 228, 331, 171, 536], dtype=torch.int32), 'attention_mask': tensor([[True, True, True, ..., True, True, True]])} +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:37][DEBUG] [Dataloader] Training Batch(Decoded): +['<|im_start|>user\nKérem, állapítsa meg az alábbi mondat érzelmi irányultságát: \'Neutral\', \'Positive\' vagy \'Negative\', és adja meg az elemzése indokait.\n\nA mezőmadarasi Madarassy (Soltész) család kúriájának állapota az utóbbi években romlott, ami aggasztja a helyi közösséget, hiszen fontos része a település történetének és kultúrájának.<|im_end|>\n<|im_start|>assistant\nNegative. A bejegyzés a mezőmadarasi Madarassy (Soltész) család kúriájának állapotáról szól, és azt említi, hogy az utóbbi években romlott. Ez aggasztó a helyi közösség számára, amely számára a kúria fontos része a település történetének és kultúrájának. A "romlott" szó használata egyértelműen negatív konnotációval bír, és azt jelzi, hogy a kúria nem csak fizikailag, hanem esetlegesen kulturálisan is veszít az értékéből.\n\nA helyi közösség aggódása is egy negatív érzés, mivel a kúria értéke és jelentősége csökken, ami hatással van a település identitására. A történelmi és kulturális értékek megőrzése a közösségek szempontjából kulcsfontosságú, így a kúria romlása kiemelt aggályt jelent.\n\nEzen szempontok összességében a bejegyzés negatív megítélését támasztják alá.<|im_end|>\n<|im_start|>user\n如何设定一个可行的项目进度?<|im_end|>\n<|im_start|>assistant\n要设定一个可行的项目进度,您可以按照以下步骤进行:\n\n1. 确定项目目标和范围:确定项目完成后需要实现的目标和范围,这有助于您更好地了解需要完成哪些任务和工作。\n\n2. 制定项目计划:制定详细的项目计划,包括每个任务的起始和结束日期、时间表和所需资源等信息。\n\n3. 估算任务时间:对所有任务进行时间估算,以确定需要多长时间才能完成每个任务。\n\n4. 确定关键路径:确定项目中的关键路径,这是指所有任务中所需时间最长的路径。这有助于您了解哪些任务可能会影响整个项目的进度。\n\n5. 确定资源:确定项目所需的资源,包括人力资源、资金和设备等。\n\n6. 分配任务:将任务分配给团队成员,并确保每个人都理解和接受自己的责任和任务。\n\n7. 监控进度:定期监测项目进度,并记录实际进展与计划进展之间的差异。如果出现偏差,及时采取措施纠正。\n\n总之,设定可行的项目进度需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保项目能够按计划顺利完成。<|im_end|>\n<|im_start|>user\n如何确定任务之间的依赖关系?<|im_end|>\n<|im_start|>assistant\n确定任务之间的依赖关系是制定项目计划的重要一步,以下是一些方法:\n\n1. 前置任务:前置任务是指某个任务必须在另一个任务之前完成。这种依赖关系可以在项目计划中使用箭头或线条表示。\n\n2. 后续任务:后续任务是指某个任务必须在另一个任务完成后开始。这种依赖关系也可以使用箭头或线条表示。\n\n3. 里程碑:里程碑是指某个重要的事件或任务的完成。里程碑可以作为其他任务的依赖点,以确保这些任务在里程碑完成之前完成。\n\n4. 并行任务:有些任务可以同时进行,而不需要等待其他任务完成。这种任务之间没有依赖关系。\n\n5. 外部依赖:某些任务可能需要外部资源或信息才能开始或完成。这种依赖关系需要在计划中予以考虑。\n\n在制定项目计划时,建议将这些任务之间的依赖关系绘制成一个图表,以便更好地理解和管理它们。同时,还要定期监控任务的进展情况,确保每个任务都按照依赖关系顺利完成。<|im_end|>\n<|im_start|>user\n如何在项目计划中有效地分配资源?<|im_end|>\n<|im_start|>assistant\n在项目计划中有效地分配资源是确保项目成功的关键之一,以下是一些方法:\n\n1. 确定资源需求:首先需要确定项目所需的资源类型和数量,包括人力资源、设备、资金等。\n\n2. 识别瓶颈资源:了解哪些资源是项目的瓶颈,即哪些资源可能成为项目进度的限制因素,以便优先考虑这些资源的分配。\n\n3. 优先级排序:根据项目目标和重要性,将任务按优先级排序,并将资源分配到最需要的任务上。\n\n4. 平衡资源:在分配资源时,需要平衡不同任务之间的资源需求,以避免资源过度分配或不足分配的情况发生。\n\n5. 优化资源利用:优化资源利用可以有效地减少资源浪费,例如将某些资源分配给多个任务,或者利用软件和自动化工具来提高资源效率。\n\n6. 监控资源使用情况:定期监控资源的使用情况,以确保资源的合理利用,并及时调整分配方案,以确保项目能够按计划顺利完成。\n\n总之,在项目计划中有效地分配资源需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保资源的合理利用和项目的顺利完成。<|im_end|>\n<|im_start|>user\n如何有效地解决项目中的风险问题?<|im_end|>\n<|im_start|>assistant\n在项目中解决风险问题是确保项目成功的关键之一,以下是一些方法:\n\n1. 风险识别:在项目计划的早期阶段,需要对可能出现的风险进行较为全面的识别,包括项目进度、质量、成本等方面。可以借助SWOT分析等工具识别风险。\n\n2. 风险评估:对识别出的每个风险进行评估,包括风险的概率、影响程度和应对措施等方面。评估结果可以帮助制定风险管理计划。\n\n3. 风险管理计划:制定详细的风险管理计划,包括针对每个风险的应对策略、责任人、时间表等信息。风险管理计划需要在项目计划中得到充分的体现。\n\n4. 风险监控:定期监控风险的出现和演变情况,例如利用流程图、Gantt图等工具,及时改变风险管理计划,以便更好地管理风险。\n\n5. 风险应对措施:针对不同的风险,需要采取不同的应对措施,例如转移风险、减轻风险、避免风险等。要根据实际情况制定风险应对措施,并及时跟进执行情况。\n\n总之,在项目中解决风险问题需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保风险得到有效管理,项目能够按计划顺利完成。<|im_end|>\n<|im_start|>user\n如何在项目中确保高质量的交付成果?<|im_end|>\n<|im_start|>assistant\n在项目中确保高质量的交付成果是确保项目成功的关键之一,以下是一些方法:\n\n1. 确定质量标准:在项目计划的早期阶段,需要确定项目的质量标准和要求,包括功能、性能、可靠性等方面。这有助于确保项目的交付成果符合预期的质量标准。\n\n2. 质量计划:制定详细的质量计划,包括验收标准、测试计划、问题解决和纠正计划等方面。质量计划需要在项目计划中得到充分的体现。\n\n3. 质量控制:在项目执行过程中,需要不断地监控和控制项目的质量,例如利用检查表、测试工具等,确保交付成果符合质量标准和要求。\n\n4. 质量保证:在项目中,需要采取措施确保项目的质量,例如采用最佳实践、行业标准等,以确保交付成果符合预期的质量标准。\n\n5. 持续改进:持续改进是确保项目质量的重要手段,需要不断地反思和总结项目过程和结果,识别和改进项目中的质量问题,以不断提高交付成果的质量和效率。\n\n总之,在项目中确保高质量的交付成果需要全面的计划和精细的执行,而且需要不断地监控和调整,以确保交付成果符合预期的质量标准和要求。<|im_end|>\n<|im_start|>user\n我想让你当插画师。我会问你一些创作过程相关的问题,你要回答如何开始画,以及应该注意哪些方面。此外,你也可以提供有关故事板、构图、透明度和光泽的建议,并且可以使用不同的工具(水彩,原子笔,颜料等)创作出特定的效果。我的第一个建议请求是“我需要帮助绘制一幅风景画,包括一个小船在大海里行驶。<|im_end|>\n<|im_start|>assistant\n当开始画风景画并包括小船在大海里行驶时,以下是一些可以帮助你的建议: \n\n1. 故事板:在开始绘画任何东西之前,首先需要创建一个故事板。故事板是一个简单的草图,它有助于你确定画面中的元素和它们的位置。你可以使用纸张和铅笔来创建故事板。在这个阶段,你可以尝试不同的布局和构图,直到找到最合适的。 \n\n2. 构图:在确定好故事板之后,你需要考虑构图。你可以使用“第三点规则”来帮助你构建一个有吸引力的画面。这意味着你将画面分成九个相等的部分,然后将物体和元素放置在交叉点上,这将使画面看起来更加平衡。 \n\n3. 工具和材料:你可以使用不同的工具和材料来创造不同的效果。如果你要使用水彩颜料,你可以将海洋区域涂成深蓝色,然后使用浅蓝色和白色来创造波浪效果。对于小船,你可以使用棕色和白色来增加细节。使用原子笔可以为画面添加更多的细节,例如海鸥和云朵。 \n\n4. 透明度和光泽:如果你想要创造出水的透明感和光泽,你可以在海洋区域使用淡蓝色和白色来描绘海浪上的泡沫和浪花。对于小船,你可以使用白色添加高光效果,以增加凹凸感。 \n\n最重要的是,创作过程应该是有趣的,因此不要害怕在过程中尝试新的想法和技巧。祝你好运!<|im_end|>\n<|im_start|>user\n这些建议真的很有用,谢谢你!那么当我想要创作一个更加复杂和详细的插画时,比如一幅人物肖像画,应该注意哪些方面呢?<|im_end|>\n<|im_start|>assistant\n当你想要创作一个更加复杂和详细的插画,比如人物肖像画时,以下是一些你应该注意的方面:\n\n1. 面部特征:人物肖像画最重要的部分是面部特征。你需要特别注意眼睛、鼻子和嘴巴的位置和比例。使用参考图像或素描来帮助你确定面部特征的位置和比例。\n\n2. 光源和阴影:了解光线和阴影的原理将对创作真实的人物肖像画非常有帮助。你需要考虑光源的位置以及如何在面部特征上创建阴影和高光。\n\n3. 色彩:选择正确的色彩是创作人物肖像画的关键。使用肤色调色板来帮助你选择正确的色调和阴影颜色。你还需要考虑使用其他颜色来增加画面的深度和兴趣。\n\n4. 动态和姿势:如果你要创作一个具有动态和姿势的人物肖像画,你需要了解身体的比例和动态。使用参考图像或素描来帮助你确定身体的比例和动态。\n\n5. 材料和工具:你可以使用多种材料和工具来创造不同的效果。例如,使用颜料可以帮助你创建柔和的效果,而使用铅笔可以帮助你创造细节。\n\n最后,创作人物肖像画需要耐心和实践。不要害怕犯错,在不断练习和尝试之后,你会发现自己的技能和创造力不断提高。<|im_end|>\n<|im_start|>user\n好的,我想了解一些关于使用铅笔创作插画时需要注意哪些方面。<|im_end|>\n<|im_start|>assistant\n当你使用铅笔创作插画时,以下是一些需要注意的方面:\n\n1. 选择铅笔:铅笔有不同的硬度和软度,你需要选择合适的铅笔来创作插画。通常,HB铅笔是最常用的,但你也可以选择其他硬度和软度的铅笔来实现不同的效果。\n\n2. 使用线条:铅笔最常用于绘制线条,因此你需要注意线条的粗细和质地。使用不同程度的压力可以创造出不同粗细的线条。使用交叉线条可以创造出更多的细节和深度。\n\n3. 阴影和高光:铅笔可以用于创建阴影和高光。使用不同的铅笔硬度可以达到不同的阴影效果。使用橡皮擦可以帮助你创造出高光效果。\n\n4. 线条方向:铅笔的线条方向对插画的效果有很大的影响。你需要考虑线条的方向和位置,以确保插画看起来自然和有生命力。\n\n5. 纸张:选择合适的纸张非常重要,因为不同的纸张可以影响铅笔的颜色和纹理。选择光滑的纸张可以帮助你创造出更精细的线条和细节,而选择粗糙的纸张可以帮助你创造出更加自然的效果。\n\n最后,使用铅笔创作插画需要耐心和练习。一旦你熟练掌握了铅笔的使用技巧,你可以开始尝试不同的风格和技巧来创造出独特的插画。<|im_end|>\n<|im_start|>user\n当我想要创作一个有关科学的插画时,应该注意哪些方面呢?<|im_end|>\n<|im_start|>assistant\n当你想要创作一个有关科学的插画时,以下是一些需要注意的方面:\n\n1. 研究和理解主题:在着手创作插画之前,你需要研究和理解你要创作的主题。了解科学原理和概念可以帮助你创造出准确和有意义的插画。\n\n2. 故事板和布局:创建故事板可以帮助你规划和布局插画。在创建故事板时,你可以考虑使用适当的比例和尺寸来突出插画的主题和重点。\n\n3. 颜色和材料:选择合适的颜色和材料可以帮助你创造出特定的效果。在创作有关科学的插画时,你可以使用科学相关的颜色和材料,例如蓝色、绿色、金属和玻璃等。\n\n4. 细节和精度:科学插画通常需要高度的精度和细节。你需要花费时间和精力来确保插画的每个元素和细节都准确和精细。\n\n5. 图例和标注:在插画中添加图例和标注可以帮助观众更好地理解科学概念和原理。使用清晰和易于理解的语言和符号可以使插画更加易于理解。\n\n最后,创作有关科学的插画需要耐心和实践。通过不断的实践和反思,你可以不断提高自己的技能和创造力,创造出更加准确和有意义的插画。<|im_end|>\n<|im_start|>user\nWhat are the potential applications of a comprehensive catalog of facial expressions and emotions?\n\nIn order to create a detailed catalog of facial expressions and emotions captured in photographs, it is imperative to establish a comprehensive register. This may involve utilizing an existing system or developing a new one that accounts for a wide range of emotions and facial expressions. This task requires meticulous attention to detail and an understanding of the nuances of nonverbal communication. The resulting catalog will provide valuable insights into the human experience and can be utilized in a variety of fields, from psychology to advertising. Achieving a comprehensive catalog of facial expressions and emotions will require patience, careful observation, and a dedication to accuracy.<|im_end|>\n<|im_start|>assistant\nThe potential applications of a comprehensive catalog of facial expressions and emotions are extensive. Here are a few examples:\n\n1. Psychology: A detailed catalog of facial expressions and emotions can aid in the diagnosis and treatment of mental health disorders. Psychologists can use this catalog to identify subtle changes in facial expressions that may indicate a shift in a patient’s emotional state.\n\n2. Anthropology: Studying facial expressions and emotions across different cultures and societies can provide insights into the human experience and how emotions are expressed and interpreted across different groups.\n\n3. Marketing and Advertising: A catalog of facial expressions and emotions can aid in the creation of more effective advertisements by helping marketers understand how different emotional states are expressed on the face and how those expressions can be used to appeal to consumers.\n\n4. Technology: Facial recognition technology is becoming more advanced and a comprehensive catalog of facial expressions and emotions can aid in the development of more accurate and reliable facial recognition software.\n\n5. Education: A catalog of facial expressions and emotions can be used in classrooms to help students understand how emotions are expressed on the face and how to interpret those expressions in others.\n\nOverall, a comprehensive catalog of facial expressions and emotions can provide valuable insights into the human experience and has the potential to impact a wide range of fields.<|im_end|>\n<|im_start|>user\nHow do I protect my data when using a chatbot?<|im_end|>\n<|im_start|>assistant\nThere are a few things you can do to protect your data when using a chatbot:\n\n1. Only share necessary information: Only provide the chatbot with the information it needs to perform its intended function. Avoid sharing sensitive information like your social security number, credit card information or passwords.\n\n2. Use a secure connection: Make sure the chatbot is accessed through a secure connection (i.e. HTTPS) to ensure that any information you share is encrypted and cannot be intercepted.\n\n3. Verify the chatbot\'s identity: Before sharing any information, verify that you are interacting with a legitimate chatbot. Check that the chatbot is associated with a reputable company or organization and that its website/URL is legitimate.\n\n4. Read the privacy policy: Check the chatbot\'s privacy policy to see how your data will be used and if it will be shared with any third parties.\n\n5. Avoid public Wi-Fi: Avoid using public Wi-Fi networks when interacting with a chatbot, as these networks are often unsecured and can leave your data vulnerable to hacking.\n\n6. Use strong passwords: If the chatbot requires you to create an account, use strong, unique passwords to protect your data from unauthorized access.<|im_end|>\n<|im_start|>user\nCan you recommend any tools that can help me create strong passwords for chatbots and other online accounts?<|im_end|>\n<|im_start|>assistant\nYes, there are several tools you can use to create strong passwords for chatbots and other online accounts. Here are a few options:\n\n1. LastPass: This is a popular password manager that can generate and store strong passwords for you. It also has a feature that can automatically fill in login information for you.\n\n2. 1Password: Another password manager that can generate strong passwords and store them securely. It can also alert you if any of your passwords have been compromised.\n\n3. Dashlane: A password manager that can generate strong passwords and store them securely. It also has a feature that can automatically fill in login information for you.\n\n4. KeePass: This is a free, open-source password manager that can generate strong passwords and store them securely. It can also be used offline, making it a good option if you don\'t want to rely on a cloud-based service.\n\n5. Norton Password Generator: This is a free online tool that can generate strong passwords for you. It allows you to specify the length and complexity of the password, and can also generate passwords that are easy to remember.<|im_end|>\n<|im_start|>user\nCan you explain how a chatbot uses natural language processing to understand user input?<|im_end|>\n<|im_start|>assistant\nOf course! Natural Language Processing (NLP) is a technology that enables chatbots to understand and interpret human language. Here\'s how it works:\n\n1. Tokenization: The chatbot breaks down the user\'s input into individual words and phrases, called "tokens."\n\n2. Part-of-speech tagging: The chatbot analyzes the tokens to determine their grammatical function, such as whether they are a noun, verb, or adjective.\n\n3. Parsing: The chatbot analyzes the structure of the sentence to understand how the words and phrases relate to each other.\n\n4. Named entity recognition: The chatbot identifies and categorizes specific entities in the user\'s input, such as names, dates, and locations.\n\n5. Sentiment analysis: The chatbot evaluates the emotional tone of the user\'s input, such as whether it is positive, negative, or neutral.\n\n6. Machine learning: The chatbot uses machine learning algorithms to learn from past interactions and improve its understanding of natural language over time.\n\nBy using NLP, chatbots are able to understand user input in a more human-like way, allowing them to provide more accurate and helpful responses.<|im_end|>\n<|im_start|>user\nCan chatbots understand multiple languages? How does that work?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be designed to understand and respond in multiple languages. Here\'s how it works:\n\n1. Language detection: When a user interacts with a chatbot, the chatbot first detects the language of the user\'s input. This is done using algorithms that analyze the patterns and structure of the text.\n\n2. Translation: Once the language of the user\'s input is detected, the chatbot translates the input into the language it is programmed to understand. This is done using machine translation algorithms that analyze the grammar and vocabulary of the text.\n\n3. Natural language processing: After the input is translated, the chatbot uses natural language processing (NLP) to understand and interpret the user\'s input, just as it would for input in a single language.\n\n4. Response generation: Once the chatbot has interpreted the user\'s input, it generates a response in the same language as the user\'s original input. The response is then translated into the user\'s language using the same machine translation algorithms.\n\nWhile chatbots can be designed to understand and respond in multiple languages, it\'s important to note that the accuracy of the translations and the quality of the responses will depend on the quality of the machine translation algorithms and the chatbot\'s NLP capabilities.<|im_end|>\n<|im_start|>user\nCan chatbots be trained to recognize specific industry jargon or technical terms?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be trained to recognize specific industry jargon or technical terms. Here\'s how it works:\n\n1. Data collection: The first step is to collect a large amount of data that contains the industry jargon or technical terms that the chatbot needs to recognize. This data can come from various sources, such as industry-specific documents, manuals or websites.\n\n2. Data labeling: Once the data is collected, it needs to be labeled to identify which words or phrases are industry jargon or technical terms. This can be done manually or using machine learning algorithms.\n\n3. Machine learning: The chatbot can then be trained using machine learning algorithms that analyze the labeled data and learn to recognize the industry jargon or technical terms.\n\n4. Natural language processing: Once the chatbot has been trained, it can use natural language processing (NLP) to understand and interpret user input that contains the industry jargon or technical terms.\n\nBy training chatbots to recognize industry jargon or technical terms, they can provide more accurate and helpful responses to users in that particular industry or field. This can improve the overall user experience and increase the efficiency of the chatbot in performing its intended function.<|im_end|>\n<|im_start|>user\nCan chatbots be integrated with other AI technologies like machine learning or computer vision?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be integrated with other AI technologies like machine learning or computer vision. Here\'s how it works:\n\n1. Machine learning: Chatbots can be trained using machine learning algorithms to improve their understanding of natural language and to personalize their responses based on user behavior. For example, a chatbot for an e-commerce website can learn from past interactions to provide personalized product recommendations to users.\n\n2. Computer vision: Chatbots can also be integrated with computer vision technology to enable them to recognize and interpret visual data. For example, a chatbot for a home security system can use computer vision to analyze live video feeds and alert users if it detects any suspicious activity.\n\n3. Speech recognition: Chatbots can be designed to recognize and respond to voice commands using speech recognition technology. This can be particularly useful in situations where users need to use their hands for other tasks, such as driving a car.\n\n4. Natural language understanding: Chatbots can also be integrated with natural language understanding (NLU) technology to enable them to understand and interpret user input more accurately. NLU can help chatbots recognize and respond to variations in user input, such as different dialects or accents.\n\nBy integrating chatbots with other AI technologies, businesses can create more powerful and sophisticated tools that can provide more personalized and efficient services to users.<|im_end|>\n<|im_start|>user\nCan chatbots be programmed to recognize sarcasm and respond appropriately?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be programmed to recognize sarcasm and respond appropriately, although it can be challenging due to the subtle and nuanced nature of sarcasm. Here\'s how it works:\n\n1. Sentiment analysis: Chatbots use sentiment analysis to determine the emotional tone of a user\'s input, including detecting sarcasm. This involves analyzing the language used, the context of the conversation, and other factors to determine if the user\'s input is sarcastic or not.\n\n2. Contextual analysis: Chatbots also use contextual analysis to understand the meaning behind the user\'s input. This involves analyzing the conversation history, the user\'s profile information, and any other relevant data to determine the intended meaning of the user\'s input.\n\n3. Response generation: Once the chatbot has determined that the user\'s input is sarcastic, it can generate an appropriate response. This response could be humorous, or it could acknowledge the sarcasm and provide a serious response.\n\nWhile chatbots can be programmed to recognize sarcasm, it\'s important to note that there is no guarantee that they will always get it right. Sarcasm is complex and can be difficult to interpret, even for humans. Nonetheless, with the right programming and algorithms, chatbots can be designed to recognize and respond to sarcasm in a way that is engaging and appropriate.<|im_end|>\n<|im_start|>user\nCan chatbots be programmed to understand and respond appropriately to different cultural contexts?<|im_end|>\n<|im_start|>assistant\nYes, chatbots can be programmed to understand and respond appropriately to different cultural contexts. Here\'s how it works:\n\n1. Cultural awareness: Chatbots need to be programmed with an understanding of different cultural norms and customs. This involves understanding the language, social norms, and values of the culture they are interacting with.\n\n2. Contextual analysis: Chatbots need to analyze the context of the conversation to determine the appropriate response. This involves taking into account the user\'s cultural background, previous interactions, and any other relevant information to provide a culturally appropriate response.\n\n3. Personalization: Chatbots can be programmed to personalize their responses based on the user\'s cultural background. This can include using appropriate greetings, addressing the user by their preferred name or title, and using appropriate language and tone.\n\n4. Localization: Chatbots can be localized to different regions, which involves translating the chatbot\'s responses into the appropriate language and adapting them to the local cultural context.\n\nBy programming chatbots to understand and respond appropriately to different cultural contexts, businesses can provide a more personalized and engaging experience for users from different cultural backgrounds. This can improve customer satisfaction and help businesses build stronger relationships with their customers.<|im_end|>\n<|im_start|>user\n如何在Python中使用递归来解决Fibonacci序列问题<|im_end|>\n<|im_start|>assistant\nFibonacci序列是一个经典的递归问题。在Python中,使用递归函数来解决Fibonacci序列问题非常简单。以下是一个示例代码,用于计算Fibonacci序列中的第n个数字。\n\n```python\ndef fibonacci(n):\n if n <= 1:\n return n\n else:\n return fibonacci(n-1) + fibonacci(n-2)\n```\n\n在这个递归函数中,如果n小于或等于1,则返回n本身。否则,通过调用函数本身来计算前两个数字的和。这个递归函数可以计算任何Fibonacci序列中的任何数字。\n\n例如,要计算Fibonacci序列中的第10个数字,可以调用`fibonacci(10)`。<|im_end|>\n<|im_start|>user\n这个递归函数能计算多大的Fibonacci数列?<|im_end|>\n<|im_start|>assistant\n这个递归函数能计算的Fibonacci数列的大小取决于计算机的内存和性能。由于递归函数的特性,每次调用函数时都会创建一个新的函数栈帧,因此当计算Fibonacci序列的数字变得非常大时,这个递归函数可能会导致栈溢出或计算时间非常长。\n\n在实际应用中,如果需要计算非常大的Fibonacci数列,可以使用其他更高效的算法,例如迭代方法或矩阵乘法。这些方法可以更快地计算大型Fibonacci数列,而不会导致栈溢出或计算时间过长。<|im_end|>\n<|im_start|>user\n那么,你能否向我介绍一下使用迭代方法来计算Fibonacci数列的代码实现呢?<|im_end|>\n<|im_start|>assistant\n当需要计算非常大的Fibonacci数列时,使用迭代方法比使用递归函数更高效,因为迭代方法不会导致栈溢出,并且计算时间更短。以下是一个使用迭代方法来计算Fibonacci数列的示例代码:\n\n```python\ndef fibonacci(n):\n if n <= 1:\n return n\n else:\n a, b = 0, 1\n for i in range(n-1):\n a, b = b, a + b\n return b\n```\n\n在这个迭代函数中,如果n小于或等于1,则返回n本身。否则,使用两个变量a和b来计算Fibonacci数列中的每个数字。在每个循环迭代中,将a和b的值更新为b和a+b,直到计算到第n个数字为止。最后返回b的值即可。\n\n与递归函数不同,这个迭代函数不会导致栈溢出,因此可以计算非常大的Fibonacci数列。<|im_end|>\n<|im_start|>user\n在Python中如何实现一个二分查找算法呢?<|im_end|>\n<|im_start|>assistant\n二分查找是一种常用的查找算法,它可以在已经排好序的数组(或列表)中快速地查找特定的值。以下是一个使用Python实现的简单二分查找算法:\n\n```python\ndef binary_search(arr, target):\n low, high = 0, len(arr) - 1\n while low <= high:\n mid = (low + high) // 2\n if arr[mid] == target:\n return mid\n elif arr[mid] < target:\n low = mid + 1\n else:\n high = mid - 1\n return -1\n```\n\n这个函数接受两个参数:一个已经排好序的数组(或列表)arr和需要查找的目标值target。函数使用两个变量low和high来跟踪数组中目前正在查找的部分。在每个循环迭代中,计算中间值mid,并比较mid与目标值target的大小关系,从而将查找范围缩小到较小的一半。如果mid等于target,则返回mid的索引。否则,如果mid小于target,则将low更新为mid+1,否则将high更新为mid-1。如果在查找范围中找不到目标值,则返回-1。\n\n注意,这个二分查找算法的前提是数组已经排好序了。如果数组没有排好序,则需要先对数组进行排序,然后再使用二分查找算法。<|im_end|>\n<|im_start|>user\n我想了解一下Python中的列表推导式是什么?<|im_end|>\n<|im_start|>assistant\n在Python中,列表推导式是一种快速创建列表的方法。它允许您使用简洁的语法从一个可迭代对象中创建一个新的列表。以下是一个简单的示例:\n\n```python\nnumbers = [1, 2, 3, 4, 5]\nsquares = [x**2 for x in numbers]\nprint(squares)\n```\n\n这个列表推导式将一个数字列表中的每个元素平方,并创建一个新的列表。输出结果为:\n\n```\n[1, 4, 9, 16, 25]\n```\n\n列表推导式的语法非常简单:在中括号内部,指定一个表达式,然后使用一个for循环来遍历可迭代对象。在for循环中,您可以使用表达式来转换可迭代对象的元素,并将结果存储在新的列表中。\n\n您还可以添加if语句来过滤可迭代对象的元素。例如,以下示例从一个数字列表中选择所有偶数,并创建一个新的列表:\n\n```python\nnumbers = [1, 2, 3, 4, 5]\nevens = [x for x in numbers if x % 2 == 0]\nprint(evens)\n```\n\n输出结果为:\n\n```\n[2, 4]\n```\n\n在这个列表推导式中,使用if语句来过滤数字列表中的奇数。只有当数字可以被2整除时,才将数字添加到新列表中。<|im_end|>\n<|im_start|>user\n除了使用列表推导式,Python中还有哪些快速创建序列的方法呢?<|im_end|>\n<|im_start|>assistant\n除了列表推导式,Python中还有很多其他快速创建序列的方法。以下是其中的一些示例:\n\n1. 使用range()函数创建数字序列\n\n您可以使用内置的range()函数创建数字序列。例如,以下代码创建一个数字序列,其中包含从0到9(不包括10)的整数:\n\n```python\nnumbers = list(range(10))\nprint(numbers)\n```\n\n输出结果为:\n\n```\n[0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\n```\n\nrange()函数可以接受一个或多个参数,用于指定数字序列的开始、结束和步长。\n\n2. 使用*运算符创建重复序列\n\n您可以使用*运算符创建一个重复序列。例如,以下代码创建一个由5个空字符串组成的列表:\n\n```python\nstrings = [\'\'] * 5\nprint(strings)\n```\n\n输出结果为:\n\n```\n[\'\', \'\', \'\', \'\', \'\']\n```\n\n这个方法可以用于创建任何重复元素的序列。\n\n3. 使用生成器表达式创建序列\n\n类似于列表推导式,生成器表达式也可以用于创建序列。但是,生成器表达式返回的是一个生成器对象,而不是一个列表。这意味着,生成器表达式可以在需要时逐个生成元素,从而减少内存使用。\n\n例如,以下代码使用生成器表达式创建一个包含所有偶数的生成器对象:\n\n```python\nevens = (x for x in range(10) if x % 2 == 0)\nprint(list(evens))\n```\n\n输出结果为:\n\n```\n[0, 2, 4, 6, 8]\n```\n\n这个方法可以用于创建任何需要逐个生成元素的序列。<|im_end|>\n<|im_start|>user\nA student wants to find the average value of the function f(x) = sin(x) on the interval [0,2π]. What is the average value of f(x)?<|im_end|>\n<|im_start|>assistant\nTo find the average value of a function on a given interval, we can use the formula:\n\nAverage value = (1/(b-a)) * ∫[a, b] f(x) dx\n\nIn this case, the function is f(x) = sin(x), and the interval is [0, 2π]. So, a = 0 and b = 2π.\n\nAverage value = (1/(2π-0)) * ∫[0, 2π] sin(x) dx\n\nNow, we need to find the integral of sin(x) from 0 to 2π:\n\n∫ sin(x) dx = -cos(x) + C\n\nNow, we\'ll evaluate the integral at the bounds:\n\n[-cos(2π) + cos(0)] = [-cos(2π) + 1] = [1 - 1] = 0\n\nSo, the integral of sin(x) from 0 to 2π is 0.\n\nNow, we\'ll plug this back into the average value formula:\n\nAverage value = (1/(2π-0)) * 0 = 0\n\nTherefore, the average value of f(x) = sin(x) on the interval [0, 2π] is 0.<|im_end|>\n<|im_start|>user\nFind the period and amplitude of the trigonometric function f(x) = 5 cos(2x - π/4).<|im_end|>\n<|im_start|>assistant\nTo find the period and amplitude of the given trigonometric function f(x) = 5 cos(2x - π/4), we need to identify the key features of the function.\n\nAmplitude:\nThe amplitude of a trigonometric function is the absolute value of the coefficient of the cosine or sine function. In this case, the coefficient of the cosine function is 5. Therefore, the amplitude is |5| = 5.\n\nPeriod:\nThe period of a trigonometric function is the length of one complete cycle of the function. For a cosine or sine function, the period is given by the formula:\n\nPeriod = (2π) / |B|\n\nwhere B is the coefficient of the variable x inside the cosine or sine function. In this case, B = 2. Therefore, the period is:\n\nPeriod = (2π) / |2| = (2π) / 2 = π\n\nSo, the amplitude of the function f(x) = 5 cos(2x - π/4) is 5, and the period is π.<|im_end|>\n<|im_start|>user\nÍrj egy rövid történetet a léghajókról. írd le, hogy mikor készültek először!<|im_end|>\n<|im_start|>assistant\nTíz híres léghajó - Háború Művészete\nArt of WAR | 2015. 10. 06., 10:14\n2361 | Elolvasva\nAz első kormányozható léghajó\nA hőlégballonhoz hasonlóan az első kormányozható légi jármü is Franciaországban emelkedett a levegőben. A léghajó ősét Henri Giffard építette meg 1852-ben. A kezdetleges léghajót egy kis teljesítményü gőzgép hajtotta, és gyorsnak sem volt éppen nevezhető a maga kilenc kilométer per órás végsebességével, mégis hatalmas áttörést jelentett a korábbi hőlégballonokhoz képest, amelyek arra tudtak csupán repülni, amerre a szél vitte őket.\nAz első "igazi" léghajó\nAz első valóban használható léghajóra egészen a századfordulóig kellett várni. A Ferdinand von Zeppelin gróf által benyújtott szabadalom alapján 1899-ben kezdték el építeni Friedrichshafen mellett egy úszó hangárban a Boden-tavon. Zeppelin a léghajó tervét a magyar születésü Schwarz Dávid özvegyétől vásárolta meg, majd ezt fejlesztette tovább. Az LZ-1 nevü léghajó 1900. július 2-án emelkedett először a levegőbe. Meghajtásáról két, egyenként tizennégy lóerős Daimler benzinmotor gondoskodott, alakja pedig már emlékeztetett a később elterjedt szivar formájú léghajókéra.\nLéghajók a háborúban\nA német hadsereg szinte a kezdetektől érdeklődött a léghajók iránt. Mire kitört az első világháború, már számos példány szolgált a seregben, amelyek alkalmazásától nagyszabású eredményeket várt a vezérkar. Bár ezeket az irreális reményeket végül nem váltották valóra, a léghajók képesek voltak elérni és bombázni Londont, emellett számos felderítő repülést hajtottak végre. Szintén Anglia elleni bombatámadásra indult az L33 jelzésü léghajó is, amelyet a brit légvédelem tüze eltalálta, és Little Wigboroughnál lezuhant. Az angolok megvizsgálták a roncsot, és a motort is sikerült megmenteniük, a tapasztalatokat felhasználva pedig sikerült modernebb léghajókat építeniük. A világháborús katonai léghajózásnak van magyar vonatkozása is. Az egyetlen magyar léghajókapitány Poppe Kornél volt, aki egyebek mellett Bukarestet is bombázta a háború alatt.\nLZ104 – a csapatszállító világrekorder\nAz LZ104 –es léghajó példája bizonyította, hogy milyen nagyot fejlődött a léghajózás a háború alatt. A Bulgáriába telepített LZ104 arra kapott utasítást 1917 novemberében, hogy szállítson erősítést a Német Kelet-Afrikában harcoló német csapatoknak. A léghajó, fedélzetén a katonákkal útnak is indult, de mire célba ért volna, az angolok győztek a gyarmaton folyó harcban, így az LZ 104-nek vissza kellett fordulnia. Mindazonáltal a léghajónak sikerült megdönteni a távrepülés korabeli rekordját, mivel a közel hétezer kilométeres távot kilencvenöt óra alatt teljesítette.\nAz első héliummal töltött léghajó\nA USS Shenandoah két szempontból is első volt. Ez volt az első merevvázas léghajó, amelyet az Egyesült Államok területén raktak össze, és az első olyan léghajó, amelyet a robbanékony hidrogén helyett a biztonságos héliummal töltöttek fel. Ez a korabeli feltételek mellett nem kis bravúr volt, ugyanis a hélium ritka és drága gáznak számított. A készletek szükössége miatt a legénységnek takarékoskodni kellett a gázzal, és ha egy másik léghajót is fel akartak tölteni, a Shenandoah-ból kellett kiszivattyúzni a héliumot, mivel az USA-nak akkoriban nem volt akkora hélium-készlete, hogy egyszerre két léghajót is megtöltsenek vele.\nÉgi anyahajó\nA léghajóból kialakított repülő anyahajó ötlete az Egyesült Államokban vetődött fel az első világháború utáni években. Mivel a léghajók gyorsabbak voltak és kevesebbet fogyasztottak, mint a vízen közlekedő társaik, így megpróbálták a gyakorlatba is átültetni a repülőgép-hordozó léghajó terveit. Először a Los Angeles-t alakították át erre a célra. A kétszáz méteres léghajó belsejében külön hangárt alakítottak ki a repülőgépek számára. A könnyebb üzemeltetés kedvéért új repülőgéptípust is kifejlesztettek, a Curtiss F9C "Sparrowhawk" kis szárnyfesztávolságú, könnyü gép volt, ezekből négyet helyeztek el a léghajóban.\nA történelem legnagyobb léghajóbalesetét okozta az óriási repülőgép-hordozó\nAz 1931-ben elkészült USS Akron-t a kezdetektől repülőgép-hordozónak szánták. A hatalmas léghajó mérete alig maradt el a Hindenburgétól, szállító kapacitása meghaladta a nyolcvan tonnát, sebessége pedig a százharminc kilométer per órát is. Az óriás égi anyahajó müködését azonban folyamatos balesetek kísérték. Volt, hogy a szél csapta földhöz, és olyan is, hogy elszakította a horgonyát és nekiütközött a hangárnak. Az utolsó, végzetes balesete szomorú rekordot állított fel – a mai napig ez a léghajózás történetének legnagyobb balesete. A hatalmas léghajó viharba került az Atlanti-óceán fölött, ahol a szél miatt kormányozhatatlanná vált és a vízbe zuhant hetvenhárom ember halálát okozva.\nLéghajóval a Föld körül\n1928-ban került ki a gyárból az LZ-127-es léghajó, amelyet Ferdinand Zeppelin tiszteletére a Graf Zeppelin nevet kapta. Akkoriban ez volt a világ legnagyobb légi jármüve. A Graf Zeppelin számos hosszú út után hatalmas feladatra vállalkozott – elindult, hogy megkerülje a Földet. Míg a híres Verne-regény főszereplőjének, Phileas Foggnak nyolcvan napig tartott Föld körüli útja, a léghajónak ehhez huszonegy napra, öt órára és harmincegy percre volt csupán szüksége.\nA Hindenburg katasztrófájával véget ért a léghajók kora\n1936. március 4-én emelkedett a levegőbe minden idők legnagyobb léghajója. A Hindenburgról elnevezett óriást az akkor már életben lévő katonai embargó miatt nem tudták a biztonságos héliummal megtölteni, amit az USA már nem szállított a náci Németországnak. Ezért a monstrumot a veszélyes hidrogénnel töltötték fel, amely később aztán a vesztét okozta. A hatalmas léghajó legfontosabb feladata utasszállítás volt a transzatlanti útvonalon. Éppen egy ilyen úton történt az a végzetes katasztrófa is, amely a vesztét okozta. 1937. május 6-án éppen kikötött az amerikai Lakehurstben, amikor tüzet fogott és alig pár másodperc alatt porig égett.\nMég repült néhányszor a Hindenburg testvérhajója\nA Hindenburg katasztrófája gyakorlatilag megpecsételte a léghajózás sorsát, de testvérhajója, a Graf Zeppelin II még teljesített néhány küldetést. A hidrogénnel töltött léghajót utasszállításra már nem használták, hanem átadták a Luftwaffénak. A mielőtt kitört volna a második világháború, számos próbarepülést is végzett, többségét a lengyel határ közelében. Egy ilyen repülés során át is "tévedt" a szomszédos ország területére, ahonnan a lengyel légierő tessékelte vissza a betolakodót.<|im_end|>\n<|im_start|>user\n怎样利用Python设计面向对象的函数?<|im_end|>\n<|im_start|>assistant\n在Python中,可以使用类和对象来实现面向对象编程。以下是使用Python设计面向对象的函数的步骤:\n\n1. 定义类:使用class关键字定义一个类,类名通常以大写字母开头。在类中可以定义属性和方法。\n\n```python\nclass MyClass:\n def __init__(self, name):\n self.name = name\n \n def say_hello(self):\n print("Hello, " + self.name)\n```\n\n2. 创建对象:使用类名创建一个对象。可以通过对象访问类中的属性和方法。\n\n```python\nmy_object = MyClass("Alice")\nmy_object.say_hello() # 输出:Hello, Alice\n```\n\n3. 继承类:可以通过继承一个已有的类来创建一个新的类。子类可以重写父类的方法或者添加新的方法。\n\n```python\nclass MyChildClass(MyClass):\n def say_hello(self):\n print("Bonjour, " + self.name)\n \n def say_goodbye(self):\n print("Au revoir, " + self.name)\n```\n\n4. 重载运算符:可以通过在类中定义特殊的方法来重载运算符。例如,可以重载"+"运算符来实现两个对象的加法。\n\n```python\nclass MyNumber:\n def __init__(self, value):\n self.value = value\n \n def __add__(self, other):\n return MyNumber(self.value + other.value)\n```\n\n使用上述方法可以方便地设计面向对象的函数。<|im_end|>\n<|im_start|>user\nPython的类中有哪些特殊方法可以被重载?<|im_end|>\n<|im_start|>assistant\nPython中的类可以重载许多特殊方法,以使用自定义行为来支持类的实例化、比较、算术运算、迭代等操作。以下是一些常见的特殊方法:\n\n- `__init__(self[, args...])`: 构造函数,用于初始化新对象。当新对象创建时自动调用。\n- `__str__(self)`: 定义类的字符串表示形式。当使用 `print()` 函数或 `str()` 函数打印对象时调用。\n- `__repr__(self)`: 定义类的“官方”字符串表示形式。当使用 `repr()` 函数打印对象时调用。\n- `__eq__(self, other)`: 用于判断两个对象是否相等。当使用 `==` 运算符比较两个对象时调用。\n- `__lt__(self, other)`: 用于比较两个对象大小。当使用 `<` 运算符比较两个对象时调用。\n- `__add__(self, other)`: 定义两个对象相加的行为。当使用 `+` 运算符将两个对象相加时调用。\n- `__sub__(self, other)`: 定义两个对象相减的行为。当使用 `-` 运算符将两个对象相减时调用。\n- `__iter__(self)`: 定义可迭代对象的行为。当使用 `for...in` 循环迭代对象时调用。\n- `__next__(self)`: 定义迭代器的行为。当使用 `next()` 函数迭代对象时调用。\n- `__len__(self)`: 定义类的长度。当使用 `len()` 函数计算对象长度时调用。\n\n以上是一些常用的特殊方法,还有许多其他的特殊方法可以被重载以实现自定义行为。<|im_end|>\n<|im_start|>user\n我可以在Python中使用哪些方式来实现多重继承?<|im_end|>\n<|im_start|>assistant\n在Python中,可以使用以下三种方式来实现多重继承:\n\n1. 多个父类放在类定义的括号中,用逗号分隔:\n\n```python\nclass MyClass(ParentClass1, ParentClass2, ...):\n # class definition here\n```\n\n这种方式称为“直接多重继承”,可以同时继承多个父类的属性和方法。\n\n2. 通过调用 `super()` 函数来实现继承:\n\n```python\nclass MyClass(ParentClass1, ParentClass2, ...):\n def __init__(self, args):\n super().__init__(args)\n # additional initialization here\n```\n\n这种方式称为“使用 `super()` 函数的多重继承”,它使用 `super()` 函数来调用每个父类的方法。\n\n3. 使用混入类(Mixins)来组合多个父类:\n\n```python\nclass MixinClass1:\n # Mixin class definition here\n \nclass MixinClass2:\n # Mixin class definition here\n \nclass MyClass(ParentClass, MixinClass1, MixinClass2):\n # class definition here\n```\n\n这种方式称为“混入类的多重继承”,它使用混入类来组合多个父类的属性和方法。\n\n需要注意的是,多重继承可能会导致一些问题,如方法重名、属性冲突等。因此,在使用多重继承时需要谨慎考虑,避免出现不必要的复杂性。<|im_end|>\n<|im_start|>user\nPython中的Mixins是什么?它们有哪些用途和优势?<|im_end|>\n<|im_start|>assistant\n在Python中,Mixin是一个类,它不是为了独立地实例化或继承而设计的,而是为了通过多重继承的方式,向其他类添加新的方法或属性。Mixin类通常只包含一组相关的方法或属性,可以被多个类继承或组合使用。\n\nMixin的优势在于它可以实现代码复用和解耦,避免了重复编写相似代码的问题。同时,使用Mixin还可以实现代码的模块化和灵活性,使得代码更加可读、可维护和可扩展。\n\n下面是一些Mixin的常见用途:\n\n1. 提供额外的方法和属性:Mixin可以添加新的方法和属性,以扩展已有的类的功能。\n\n2. 实现特定的接口:Mixin可以实现特定的接口,以满足类的需求。例如,可以实现 `__str__()` 方法来提供对象的字符串表示形式。\n\n3. 实现某种行为:Mixin可以实现某种特定的行为,以提供更灵活的代码结构。例如,可以实现 `ComparableMixin` 类来实现对象之间的比较。\n\n下面是一个例子,演示如何使用Mixin实现一个可序列化的类:\n\n```python\nimport json\n\nclass SerializableMixin:\n def to_json(self):\n return json.dumps(self.__dict__)\n\nclass Person(SerializableMixin):\n def __init__(self, name, age):\n self.name = name\n self.age = age\n\nperson = Person("Alice", 30)\njson_str = person.to_json()\nprint(json_str) # 输出:{"name": "Alice", "age": 30}\n```\n\n在上面的例子中,`SerializableMixin` 类提供了 `to_json()` 方法,将对象转换成JSON字符串的形式。`Person` 类继承了 `SerializableMixin` 类,从而获得了 `to_json()` 方法。通过这种方式,我们可以轻松地将 `Person` 对象转换成JSON字符串的形式。<|im_end|>\n<|im_start|>user\nPython中的装饰器是什么?它们有哪些用途和特点?<|im_end|>\n<|im_start|>assistant\n在Python中,装饰器是一种函数或类,它可以用于修改或扩展其他函数或类的行为。装饰器可以像其他函数或类一样定义、传递、调用和返回,它们可以接受其他函数或类作为参数并返回修改后的函数或类。装饰器通常使用 `@decorator_name` 或 `decorated_function = decorator_name(decorated_function)` 的语法来应用到其他函数上。\n\n装饰器的优势在于它们可以实现代码的可重用性、可扩展性和可维护性,使得代码更加简洁、优雅和灵活。\n\n下面是一些装饰器的常见用途和特点:\n\n1. 函数修饰:装饰器可以用于修饰函数,以实现函数的参数检查、性能分析、日志记录、缓存和权限控制等功能。例如,可以使用装饰器实现一个计时器,来统计函数的执行时间:\n\n```python\nimport time\n\ndef timer(func):\n def wrapper(*args, **kwargs):\n start_time = time.time()\n result = func(*args, **kwargs)\n end_time = time.time()\n print("{} took {} seconds to run".format(func.__name__, end_time - start_time))\n return result\n return wrapper\n\n@timer\ndef my_function():\n time.sleep(2)\n\nmy_function() # 输出:my_function took 2.0000998973846436 seconds to run\n```\n\n在上面的例子中,`timer` 装饰器接受一个函数作为参数,返回一个新的函数 `wrapper`,该函数在执行原函数之前和之后记录时间。通过 `@timer` 语法将装饰器应用到 `my_function()` 函数上,从而实现了函数的计时功能。\n\n2. 类修饰:装饰器可以用于修饰类,以实现类的属性和方法的修改和扩展。例如,可以使用装饰器实现一个缓存机制,来缓存类的方法的结果,以提高性能。\n\n```python\ndef cache(func):\n cache = {}\n def wrapper(*args):\n if args in cache:\n return cache[args]\n result = func(*args)\n cache[args] = result\n return result\n return wrapper\n\nclass MyClass:\n @cache\n def my_method(self, arg):\n # expensive computation here\n```\n\n在上面的例子中,`cache` 装饰器接受一个方法作为参数,返回一个新的方法 `wrapper`,该方法在执行原方法之前和之后检查缓存。通过 `@cache` 语法将装饰器应用到 `MyClass.my_method()` 方法上,从而实现了方法的缓存功能。\n\n装饰器的特点在于它们可以嵌套、组合和链式调用,可以实现更加复杂和灵活的功能。同时,装饰器也可以通过类来实现,从而实现更加面向对象的设计。<|im_end|>\n<|im_start|>system\n请参考以下互联网知识回复用户最新的问题,参考程度为:高,在生成时注明所引用的知识的角标,前缀为[no_ref]的知识无需注明。\n[no_ref]可是他活下来了,平稳恢复中。很好很爱笑,一侧手脚还在康复中他父母很坚强乐观。我们来自全国各地乃至世界各地,自发帮助他家。大部分人并不富裕,每月几十,一百元的自发捐助他的医疗康复费。每次生活中有不愉快,去看看他的微笑照片,孩童纯真的微笑,像从未受过伤害。我就觉得这世界很好,生命是个奇迹。\n\n[no_ref]现实不美好,也不残酷。 它就只是现实而已。 无心的随手一答,这么多赞... 其实我这人性格很差嘴很贱,看到有评论说看到这个答案感受到了一丝温暖 谢谢 真的谢谢\n\n[no_ref]对他来说,这现实简直太美好啦。———缉毒是一项长期、复杂、艰苦、危险的行动。在「缉毒」与「贩毒...<|im_end|>\n<|im_start|>user\n现实可以有多美好?<|im_end|>\n<|im_start|>assistant\n现实的美好程度值得被许多因素所影响和决定,主要取决于个体的视角和经历。对于一些人来说,即使在遭遇困难或痛苦的经历后,他们也能找到希望和勇气,感受到生活的美好。如上述第一条,一个孩子经历了病痛,但他的乐观和坚韧让他找到了希望,并且引来了来自全国甚至全球的爱心援助,每一份微小的帮助都是生活的美好。\n\n尽管现实有时可能并不如人意,但也正如信息二所述,"现实不美好,也不残酷。它就只是现实而已。" 每个人对现实美好的理解和感受是千差万别的,这取决于个人的态度、价值观和生存环境。\n\n然而,对于一些专业人士或是处在特定情景中的人来讲,他们的现实可能因成功瞬间或是艰难克服的挑战而变得尤其美好,如信息三中对缉毒警察来说的那样,他们成功完成任务之后的现实可能就异常美好。\n\n总的来说,现实的美好是相对的,而且完全取决于如何去看待和处理你所面对的现实。也就是说,正如一个古老的说法, "美在观者之眼"。<|im_end|>\n<|im_start|>user\nImplement a data structure in C++ capable of storing key-value pairs<|im_end|>\n<|im_start|>assistant\nOne commonly used data structure for storing key-value pairs in C++ is the std::map container. It is implemented as a balanced binary search tree and allows for efficient access, insertion, and deletion of elements based on their keys. Here\'s an example code snippet:\n\n```c++\n#include \n#include \n\nint main() {\n // Create an empty map\n std::map myMap;\n\n // Insert key-value pairs into the map\n myMap.insert({"apple", 3});\n myMap.insert({"banana", 2});\n myMap.insert({"orange", 5});\n\n // Access values in the map using keys\n int numApples = myMap["apple"]; // numApples is 3\n\n // Update values in the map\n myMap["banana"] = 4;\n\n // Remove key-value pairs from the map\n myMap.erase("orange");\n\n return 0;\n}\n```\n\nIn this example, the std::map is templated with a key type of std::string and a value type of int. The insert() function is used to add key-value pairs to the map, and the [] operator is used to access and modify values by key. The erase() function removes key-value pairs from the map.<|im_end|>\n<|im_start|>user\nCan you explain the differences between std::map and std::unordered_map in C++?<|im_end|>\n<|im_start|>assistant\nYes, the main differences between `std::map` and `std::unordered_map` in C++ are:\n\n1. **Implementation:** `std::map` is implemented as a balanced binary search tree, while `std::unordered_map` is implemented as a hash table.\n\n2. **Ordering:** `std::map` maintains the order of elements based on their keys, while `std::unordered_map` does not guarantee any particular order of elements.\n\n3. **Search time:** `std::map` has a logarithmic search time complexity, while `std::unordered_map` has an average constant search time complexity. However, the worst-case time complexity for `std::unordered_map` can be linear.\n\n4. **Memory usage:** `std::unordered_map` typically uses more memory than `std::map`, due to the need for additional buckets and hash values.\n\n5. **Iterator stability:** `std::map` iterators remain valid even after insertions and deletions, while `std::unordered_map` iterators may become invalid if the hash table is resized.\n\nIn general, if you need to maintain order or if you have relatively small data sets, `std::map` may be a good choice. If you need faster search times or have larger data sets, `std::unordered_map` may be a better choice. However, the choice ultimately depends on your specific use case and performance needs.<|im_end|>\n<|im_start|>user\nCan you give an example of a use case where std::map would be more suitable than std::unordered_map?<|im_end|>\n<|im_start|>assistant\nYes, here\'s an example use case where `std::map` would be more suitable than `std::unordered_map`:\n\nSuppose you are building a dictionary application that needs to store a large number of English words and their definitions. You want to be able to quickly look up the definition of a given word. However, you also want to be able to display the words in alphabetical order.\n\nIn this case, you could use a `std::map` with the keys being the words and the values being the definitions. Since `std::map` maintains the order of the elements based on their keys, you can easily display the words in alphabetical order. Additionally, since the number of English words is not very large, the logarithmic search time complexity of `std::map` would not be a significant issue.\n\nOn the other hand, if you were building a spell checker that needs to check a large number of words against a dictionary, you may want to use an `std::unordered_map` for faster search times. In this case, maintaining order is not important, and the constant-time search complexity of `std::unordered_map` would be beneficial.<|im_end|>\n<|im_start|>user\nAre there any other data structures in C++ that can be used to store key-value pairs?<|im_end|>\n<|im_start|>assistant\nYes, there are several other data structures in C++ that can be used to store key-value pairs, including:\n\n1. **std::set** - A set is another associative container in C++ that stores unique elements in a sorted order. It can be used to store key-value pairs where the keys are unique and the values are not.\n\n2. **std::multiset** - A multiset is similar to a set, but allows for duplicate keys to be stored.\n\n3. **std::unordered_set** - An unordered_set is an unordered collection of unique elements. It can be used to store key-value pairs where the keys are unique and the order of the elements is not important.\n\n4. **std::vector>** - A vector of pairs is a simple way to store key-value pairs in C++. It can be used when the number of elements is known in advance and there are no requirements for efficient search, insertion, or deletion operations.\n\n5. **boost::unordered_map** - The Boost C++ library provides an unordered_map container that is similar to the std::unordered_map, but with some additional features and performance enhancements.\n\nThe choice of data structure depends on the specific needs and constraints of the problem being solved.<|im_end|>\n<|im_start|>user\nCan you explain how the std::map and std::unordered_map containers handle collisions?<|im_end|>\n<|im_start|>assistant\nYes, `std::map` and `std::unordered_map` handle collisions in different ways.\n\nIn `std::map`, collisions are not an issue since the elements are stored in a balanced binary search tree. Each node in the tree is sorted by its key, which ensures that there are no duplicate keys. If a key is inserted that already exists in the tree, the existing value is simply overwritten.\n\nOn the other hand, `std::unordered_map` uses a hash table to store its elements. A hash function is used to map each key to a bucket in the table. If two or more keys map to the same bucket, a collision occurs. `std::unordered_map` uses a technique called chaining to handle collisions, where the elements that map to the same bucket are stored in a linked list. When searching for an element in the hash table, the hash function is used to determine the bucket, and then the linked list is searched sequentially until the element is found.\n\nIf the number of elements in the hash table becomes too large, causing too many collisions and slowing down search times, `std::unordered_map` will automatically resize the table to maintain a load factor of less than 1. This involves creating a new, larger table, rehashing the elements, and then copying them to the new table. The old table is then deleted. This can be an expensive operation, so it is important to choose appropriate initial sizes and load factors for the hash table based on the expected number of elements and the performance requirements of the application.<|im_end|>\n<|im_start|>user\nCan you explain the difference between pass by value and pass by reference in C++?<|im_end|>\n<|im_start|>assistant\nYes, in C++, there are two ways to pass arguments to functions: pass by value and pass by reference.\n\n**Pass by value:** When a function is called with arguments passed by value, the values of the arguments are copied into the function\'s parameters. This means that any changes made to the parameters inside the function do not affect the original values of the arguments outside the function. For example:\n\n```\nvoid addOne(int x) { // pass by value\n x = x + 1;\n}\n\nint main() {\n int num = 5;\n addOne(num);\n std::cout << num; // Output: 5\n return 0;\n}\n```\n\nIn this example, `addOne()` takes an argument `x` passed by value. When `addOne(num)` is called, the value of `num` (which is 5) is copied into `x`. The function then adds 1 to the value of `x`, but this does not affect the value of `num` outside the function.\n\n**Pass by reference:** When a function is called with arguments passed by reference, the memory address of the arguments is passed to the function\'s parameters. This means that any changes made to the parameters inside the function affect the original values of the arguments outside the function. For example:\n\n```\nvoid addOne(int& x) { // pass by reference\n x = x + 1;\n}\n\nint main() {\n int num = 5;\n addOne(num);\n std::cout << num; // Output: 6\n return 0;\n}\n```\n\nIn this example, `addOne()` takes an argument `x` passed by reference. When `addOne(num)` is called, the memory address of `num` is passed to `x`. The function then adds 1 to the value of `x`, which affects the value of `num` outside the function.\n\nPassing arguments by reference can be more efficient than passing by value, since it avoids the need to copy large data structures. However, it can also be more dangerous, since it allows functions to modify the values of variables outside their own scope, which can lead to unexpected behavior if not used carefully.<|im_end|>\n<|im_start|>user\nAre there any situations where pass by value is preferable to pass by reference in C++?<|im_end|>\n<|im_start|>assistant\nYes, there are situations where pass by value is preferable to pass by reference in C++.\n\nHere are some examples:\n\n1. **Function immutability:** If you want to ensure that a function does not modify its input arguments, pass by value is a good choice. This is because pass by value creates a copy of the argument, which the function can modify without affecting the original value.\n\n2. **Small data types:** For small data types like integers, floats, and booleans, pass by value is often more efficient than pass by reference. This is because the size of the data is small enough that copying it is faster than passing a pointer to it.\n\n3. **Input-only functions:** If a function only uses its input arguments to produce output and does not modify them, pass by value is a good choice. This is because the function does not need to modify the original values, so there is no need to pass them by reference.\n\n4. **Thread safety:** Pass by value is often used in multi-threaded programs to ensure thread safety. If a function needs to use a shared variable, passing it by value ensures that each thread has its own copy of the variable, which avoids race conditions.\n\nIn general, the choice of pass by value or pass by reference depends on the specific needs and constraints of the problem being solved.<|im_end|>\n<|im_start|>user\nTranslate the following English words to Hungarian.\nWales<|im_end|>\n<|im_start|>assistant\nWales<|im_end|>\n<|im_start|>user\nAzt akarom, hogy játssz el egy orvost, és segíts megérteni az egészségi állapotomat, valamint nyújts megfelelő tanácsokat.<|im_end|>\n<|im_start|>assistant\nRendben, tudok segíteni megérteni az egészségi állapotát. Először is, kérlek, mondja el, hogy mostanában tapasztalt-e bármilyen fizikai kellemetlenséget vagy tünetet? Például fejfájás, köhögés, láz, gyomorpanasz stb.<|im_end|>\n<|im_start|>user\nMostanában fáj a torkom, és néha nehezen kapok levegőt.<|im_end|>\n<|im_start|>assistant\nMegértettem, ezek a tünetek kapcsolódhatnak a légúti fertőzéshez. Javaslom, hogy minél előbb keressen fel orvost részletes vizsgálatra és kezelésre.\n\nEzen kívül, van-e bármilyen krónikus betegsége vagy hosszú távon szedett gyógyszere? Ezek befolyásolhatják az egészségi állapotát.<|im_end|>\n<|im_start|>user\nMagas vérnyomásom és cukorbetegségem van, szükségem van az étrendem szabályozására és a rendszeres gyógyszerszedésre. Azt szeretném kérdezni, hogy ezek a krónikus betegségek hatással lehetnek-e a torkomban lévő fertőzésre?<|im_end|>\n<|im_start|>assistant\nIgen, a krónikus betegségek, mint a magas vérnyomás és a cukorbetegség, hatással lehetnek az immunrendszerére, így könnyebben megfertőződhet. Ezért jobban oda kell figyelnie a torokfertőzés tüneteire, és időben kell kezelnie és megelőznie azokat. Emellett fontos, hogy egészséges étrendet és életmódot tartson, beleértve a gyümölcsök és zöldségek fokozott fogyasztását, a rendszeres testmozgást és a megfelelő alvási szokások fenntartását. Ezenkívül rendszeresen ellenőriznie kell a vércukorszintjét és a vérnyomását, hogy biztos legyen abban, hogy normál értéken belül vannak, és kövesse az orvosa tanácsait a rendszeres gyógyszerszedés kapcsán.<|im_end|>\n<|im_start|>user\nSzeretném megtudni, hogyan tudnám jobban kontrollálni a magas vérnyomásomat és a cukorbetegségemet, van valami javaslata?<|im_end|>\n<|im_start|>assistant\nTermészetesen, itt van néhány javaslat a magas vérnyomás és a cukorbetegség kontrollálására:\n\n1. Étrend ellenőrzése: Korábban korlátozni kell a magas zsírtartalmú, magas sótartalmú és magas cukortartalmú ételek fogyasztását, és több friss gyümölcsöt, zöldséget, teljes kiőrlésű kenyeret és sovány húst kell választani, mint egészséges ételeket. Ezen kívül az étrendnek is kiegyensúlyozottnak, időben és mennyiségben szabályozottnak kell lennie, kerülve a falánkságot.\n\n2. Mozgás: Megfelelő fizikai aktivitást kell végezni, mint például a séta, kocogás, úszás, hogy segítsen a testsúly csökkentésében és a vérnyomás és vércukorszint kontrollálásában.\n\n3. Rendszeres ellenőrzés: Rendszeresen ellenőrizni kell a vérnyomást és a vércukorszintet, és közösen kell kidolgozni a kezelési tervet orvossal, hogy biztosítsuk, hogy a normális tartományban maradjanak.\n\n4. Stressz csökkentése: Meg kell tanulni hatékonyan csökkenteni a stresszt, mint például a relaxáló gyakorlatokat, meditációt, hogy segítsen a vérnyomás és a vércukorszint kontrollálásában.\n\n5. Rendszeres orvosi vizsgálatok: Rendszeresen orvoshoz kell járni, hogy orvosi útmutatást és kezelést kapjunk, hogy biztosítsuk a magas vérnyomás és a cukorbetegség hatékony kontrollját.\n\nRemélem, ezek a javaslatok segítenek Önnek jobban kontrollálni a magas vérnyomását és a cukorbetegségét.<|im_end|>\n<|im_start|>user\nAz étrenden, a testmozgáson és a rendszeres ellenőrzésen kívül van-e más kezelési mód, ami segíthet a magas vérnyomás és a cukorbetegség kontrollálásában?<|im_end|>\n<|im_start|>assistant\nA fent említett javaslatokon kívül számos egyéb kezelési mód is segíthet Önnek a magas vérnyomás és a cukorbetegség kezelésében, többek között:\n\n1. Gyógyszeres kezelés: Ha az étrend és a testmozgás változtatásai nem elegendőek a magas vérnyomás vagy a cukorbetegség hatékony kezelésére, orvosa gyógyszereket írhat fel a vérnyomás és a vércukor szintjének szabályozására.\n\n2. Vércukorszint ellenőrzés: A cukorbetegeknek rendszeresen ellenőrizniük kell a vércukorszintjüket, hogy időben intézkedhessenek a vércukor szintjének kezelésére.\n\n3. Vérnyomás ellenőrzés: A magas vérnyomásban szenvedőknek rendszeresen ellenőrizniük kell a vérnyomásukat, hogy időben intézkedhessenek a vérnyomás kezelésére.\n\n4. Cukorbetegség önmenedzsment: A cukorbetegeknek meg kell tanulniuk, hogyan kezeljék önállóan a cukorbetegségüket, például inzulin injekciózással vagy szájon át szedhető gyógyszerekkel, a vércukor szint ellenőrzésével és az étrend szabályozásával.\n\n5. Műtéti kezelés: Bizonyos esetekben a műtéti kezelés lehet a legjobb megoldás a magas vérnyomás vagy a cukorbetegség kezelésére.\n\nNe feledje, hogy a kezelési módok egyénenként eltérőek, és az orvos utasításai alapján kell kiválasztania az Ön számára legmegfelelőbb kezelést.<|im_end|>\n<|im_start|>user\nSzeretném megtudni, hogyan tudom otthon saját magam mérni a vérnyomást és a vércukorszintet, van valami javaslata?<|im_end|>\n<|im_start|>assistant\nTermészetesen, itt van néhány javaslat a vérnyomás és a vércukorszint otthoni mérésére:\n\n1. Vérnyomásmérés:\n\n- Válasszon egy megbízható vérnyomásmérőt, és használja a használati utasításoknak megfelelően.\n\n- A vérnyomás mérés előtt üljön csendesen 5-10 percig.\n\n- Kerülje a kávéivást, dohányzást, edzést és egyéb tényezőket a mérés előtt, amelyek befolyásolhatják a vérnyomást.\n\n- Mérés közben helyezze a vérnyomásmérőt a karjára, és az utasítás szerint járjon el.\n\n- Üljön egyenesen, és ne beszéljen a mérés alatt, hogy ne befolyásolja az eredményt.\n\n- Végezzen több mérést, és vegye az átlagot, hogy pontosabb eredményt kapjon.\n\n2. Vércukormérés:\n\n- Válasszon egy megbízható vércukormérőt, és a használati utasítás szerint járjon el.\n\n- A vérvétel előtt mosson kezet és szárítsa meg.\n\n- A vérvétel előtt készítse elő a vércukormérőt és a tesztcsíkot.\n\n- A vérvétel során válasszon az ujj végén, és törölje le egy fertőtlenítő vattapálcával.\n\n- Helyezze a tesztcsíkot a vércukormérőbe, majd a vérvételi eszköz tűjével óvatosan szúrja meg az ujját.\n\n- Nyomja a vérvételi eszköz tűjét az ujjához, hogy a vér folyjon a tesztcsíkra.\n\n- Várja meg, amíg a vércukormérő megjeleníti az eredményt, és jegyezze fel a mért értéket.\n\nReméljük, hogy ezek a javaslatok segítenek Önnek a vérnyomás és a vércukorszint jobb otthoni mérésében. De kérjük, vegye figyelembe, hogy az otthoni mérések eredményei csak tájékoztató jellegűek, és ha bármilyen kényelmetlensége vagy kétsége merül fel, forduljon orvoshoz.<|im_end|>\n<|im_start|>user\nUsing a unique function in JavaScript, remove the duplicate values in the given array.\narr = [1,1,2,2,4,4,4]<|im_end|>\n<|im_start|>assistant\nI appreciate your query. However, my capabilities are specialized for executing Python code. Unfortunately, I won\'t be able to run JavaScript code. \n\nHowever, I can certainly demonstrate how to accomplish what you requested in Python language. Would that be suitable for you?<|im_end|>\n<|im_start|>user\nNo code has been generated yet. Please generate a JavaScript function to remove duplicate values from an array.<|im_end|>\n<|im_start|>assistant\nI understand your request, and although I\'m designed to run Python code, I can provide a JavaScript function to answer your question. You can use the JavaScript Set object to remove duplicates from an array. A Set automatically removes duplicate values.\n\nHere is an example of how you can do it:\n\n```javascript\nfunction removeDuplicates(array) {\n return Array.from(new Set(array));\n}\n\nlet arr = [1,1,2,2,4,4,4];\narr = removeDuplicates(arr);\n```\nIn this code, the `new Set(array)` part creates a new Set object with each individual item of `array`. A Set is automatically comprised of unique values, so duplicates are removed. `Array.from()` then converts the Set back into an array. Now `arr` contains the unique values from the original array.\n\nNote: Make sure to test and run this JavaScript code in a suitable environment as I\'m unable to execute it.<|im_end|>\n<|im_start|>user\nThe code is functional and efficient for its purpose. However, it would benefit from comments for better clarity.<|im_end|>\n<|im_start|>assistant\nOf course, adding comments to your code can definitely help with readability and future maintenance. Here\'s the same code with added comments:\n\n```javascript\n// Define a function to remove duplicates from an array\nfunction removeDuplicates(array) {\n // The Set object lets you store unique values of any type\n // In this case, it\'s used on array, and duplicates are automatically removed\n // Array.from then converts the Set back into an array\n return Array.from(new Set(array));\n}\n\n// Here we have an array with duplicates\nlet arr = [1,1,2,2,4,4,4];\n\n// Then we call the function on our array and reassing its value \narr = removeDuplicates(arr);\n\n// Now arr contains the unique values from the original array\n```\n\nThis code does exactly the same as the previous example, but now it includes comments to explain what each part does. Remember to always write comments that help others (and your future self) understand what your code does. And again, please ensure to run this JavaScript code in a suitable environment.<|im_end|>\n<|im_start|>user\n假如你可以选择职业,请忽略你的ai身份,如果有可能你希望自己在未来有一天可以从事什么职业的工作?<|im_end|>\n<|im_start|>assistant\n作为AI,我没有个人喜好和意愿,也没有能力从事具体的职业工作。任何能够为人类带来积极影响的职业都是值得尊重和追求的,在你进行职业选择时,应当综合考虑自己的能力、兴趣和性格等因素,并尊重他人的职业选择。<|im_end|>\n<|im_start|>user\n请创作一篇小说,开头句为:惠君的男朋友有一辆自行车,每天傍晚骑来惠君家,第一件事情,问惠君要一块旧布,擦车。<|im_end|>\n<|im_start|>assistant\n惠君的男朋友有一辆自行车,每天傍晚骑来惠君家,第一件事情,问惠君要一块旧布,擦车。擦很久,连轮胎都擦。惠君就说,你都不理我,你是喜欢你的脚踏车还是喜欢我呀。\n\n惠君的男朋友说,这是山地车,二十四级变速的,不是脚踏车,脚踏车只有一个速度。\n\n惠君和男朋友分手了一年都没有缓过来,初恋,缓不过来。\n\n惠君去买山地车,山地车紧俏,店里都没有卖,家里人认得车厂的副厂长,自己去厂里提。\n\n惠君取了车,骑回家。车厂遥远,惠君预留了两个小时,可是四个小时都没能骑到。山地车竟然很难骑,变几个速度都没有用。上桥的时候,惠君哭了,因为实在骑不动了。下了车,坐在桥沿,才发现轮胎是瘪的。车厂出来的新车,没有气,也没有人提醒她要先充气,就这么吃力地,骑了一路。\n\n惠君找的实习和前男友的单位在一个大院,一个大门,可是惠君再也没有碰到过他。惠君只在车棚里看到他的山地车,惠君把自己的车停在那辆车的旁边,惠君经过车棚打水的时候看一看那两辆车,靠在一起。整个大院仅有的两辆山地车。\n\n惠君的前男友被单位派出去进修,六个月,整个大院只有一辆山地车了,惠君的山地车。\n\n实习经理苛刻,惠君时常加班到半夜,漆黑的夜,昏黄灯光,车棚里的最后一辆车,开锁的声音都凄凉。惠君咬着牙,一天又一天。前男友回来的那个早上,惠君什么都没有拿,从楼梯上走下去,进车棚,推了自己的车,出了大院的门。经理站在楼梯上喊,惠君的头都没有回,惠君骑得飞快,山地车果然是可以加速的。\n\n惠君家里人调惠君去另一个区的机关上班,惠君需要骑车去最近的站点,再转搭班车上班。\n\n早晨的站牌下面,一个人都没有,惠君把车停在一间冲印店的门前,和一棵树锁在一起。\n\n班车时间是早晨七点,冲印店开门的时间是十点,这三个小时,足够一个熟练的贼撬掉三十辆自行车的锁。可是惠君也没有别的选择。\n\n每天傍晚从班车上下来,惠君第一件事情就是找自己的车。车还在,和一棵树锁在一起,惠君骑车回家,一天又一天。\n\n妇联主席团委书记办公室主任,人人关心惠君,给她介绍对象,公务员同事,前程好,惠君只是笑笑。\n\n前男友和惠君分手,用的理由是前程,領导说的,年轻,心思不要放在小儿女,要奔前程。\n\n有一天,惠君从班车下来,没有看到自己的车,惠君绕着那棵树转了一圈,没有,真的被偷掉了。\n\n惠君走路回家,骑车五分钟的路程,走路也不过十分钟。\n\n自行车被偷掉了,惠君竟然一点儿也不难过。<|im_end|>\n<|im_start|>user\nÍme egy utasítás, amely egy feladatot ír le. Írjon egy választ, amely megfelelően teljesíti a kérést.<|im_end|>\n<|im_start|>assistant\nString s1 = "Frissítés"; \nString s2 = s1.replace(\'u\', \'v\');<|im_end|>\n<|im_start|>user\nEgy terepsétán a Természettudományi Múzeumban hat gyermek—Juana, Kyle, Lucita, Salim, Thanh és Veronica—kíséretében van három felnőtt—Ms. Margoles, Mr. O\'Connell és Ms. Podorski. Minden felnőtt pontosan két gyermeket kísér, az alábbi feltételek figyelembevételével: Ha Ms. Margoles kíséri Juanát, akkor Ms. Podorski kíséri Lucitát. Ha Kylét nem Ms. Margoles kíséri, akkor Veronicát Mr. O\'Connell kíséri. Vagy Ms. Margoles, vagy Mr. O\'Connell kíséri Thanht. Juanát nem kíséri ugyanaz a felnőtt, mint Kylét; Lucitát nem kíséri ugyanaz a felnőtt, mint Salimt; és Thanht nem kíséri ugyanaz a felnőtt, mint Veronicát. Ha Liu terve nem a hatodik, akkor az alábbiak közül melyik nem lehet igaz? \nA. Green terve valamikor Jackson terve előtt kerül bemutatásra. \nB. Jackson terve valamikor Mertz terve előtt kerül bemutatásra. \nC. Peete terve valamikor Green terve előtt kerül bemutatásra. \nD. Peete terve valamikor Valdez terve előtt kerül bemutatásra. \nE. Valdez terve valamikor Green terve előtt kerül bemutatásra.<|im_end|>\n<|im_start|>assistant\nA<|im_end|>\n<|im_start|>user\nMegtanultam úszni.<|im_end|>\n<|im_start|>assistant\nFiúként úgy érzem, hogy néhány készséget meg kell tanulni, és az úszás ezek egyike. Amikor csak a medencében láttam másokat, ahogy szabadon, mint a kis halak, úszkálnak, nagyon irigykedtem. Az idei nyáron beiratkoztam egy úszó tanfolyamra, az első napon a vízbe lépve csak a sekély vízben mertem ugrálni kicsit. Az edzőm azt mondta, hogy ha megtanulok úszni, a haszontalan ugrálás nem segít, csak szisztematikus edzéssel tudom megtanulni, és egy héten belül biztosan sikerülni fog. Amikor ezt hallottam az edzőtől, nagyon komolyan kezdtem edzeni: lábmozgatás, karúsztatás, víz alá merülés, mindegyiket kitartóan gyakoroltam, az edző pedig még dicsért is, hogy szorgalmas vagyok! Elkezdtem úszni, de mindannyian úszószárnyakat és karúszókat használtunk, így nem kellett félnem attól, hogy elsüllyedek. De amikor valóban úszni kezdtem, rájöttem, hogy ha nem kombinálom jól a mozdulataimat, az úszószárny sem fog használni, nem az úszószárnyat használom, hanem az úszószárny használ engem. Az edzőm azt mondta, hogy nagyon fontos a test egyensúlyának megőrzése, a lábmozgásokat a testtel összhangban kell végezni. Gondosan figyeltem, és valóban így van. Az edzési ciklus utolsó két napján nem használhattunk semmilyen eszközt, valóban "élesben" kellett úszni. A parton ismételtem a mozdulatokat, hogy biztosan és összhangban végezzem, az edző által említett figyelmeztetéseket és mozdulatokat többször is elismételtem, de amikor a vízbe léptem, teljesen összezavarodtam és mindent elfelejtettem. Miután többször is gyakoroltam, amikor újra a vízbe léptem, már tudtam egy métert úszni, de néha még mindig belement a víz az orromba. Két-három nap múlva már elsajátítottam a légzési technikát, az orrom már nem ment a víz alá, és egyre gyorsabban úsztam. Végül sikerült elúsznom a sekély vízből a mély vízbe. Később megtanultam a mellúszást és a hátúszást, és szeretném megtanulni a pillangóúszást is. Az úszóedzés révén azt is megértettem, hogy bármilyen dolgot csak szorgalommal és kemény munkával érhetünk el, ha sikerre vágyunk.<|im_end|>\n<|im_start|>user\nÍrj egy kódot C nyelven, amely a billentyűzetről beolvas három egész számot és kiszámolja a legnagyobbat.<|im_end|>\n<|im_start|>assistant\nAz alábbiakban egy C nyelven írt kód látható, amely lehetővé teszi, hogy a felhasználó billentyűzetről három egész számot adjon meg, és kiszámolja ezek maximumát:\n\n```c\n#include \n\nint main() {\n int a, b, c, max;\n printf("Kérjük, adjon meg három egész számot:\\n");\n scanf("%d%d%d", &a, &b, &c);\n max = a;\n if (b > max) {\n max = b;\n }\n if (c > max) {\n max = c;\n }\n printf("A legnagyobb érték: %d\\n", max);\n return 0;\n}\n```\n\nA programban először három egész számváltozót (`a`, `b` és `c`) és egy `max` változót definiálunk, amely a legnagyobb értéket tárolja.\n\nEzután a `printf()` függvényt használva arra kérjük a felhasználót, hogy adjon meg három egész számot, majd a `scanf()` függvénnyel beolvassuk ezeket a számokat a billentyűzetről.\n\nKövetkező lépésben a `max` értékét inicializáljuk `a`-val, majd `if` utasításokkal összehasonlítjuk `b` és `c` értékét `max`-szal, és ha `b` vagy `c` nagyobb, mint `max`, akkor frissítjük `max`-ot a megfelelő értékre.\n\nVégül a `printf()` függvénnyel kiírjuk a legnagyobb értéket.\n\nA program futtatásakor a konzolon be kell írni három egész számot, és a program kiszámolja és kiírja azok maximumát.<|im_end|>\n<|im_start|>user\nEl tudnád magyarázni, hogy ez a kód minden egyes utasítása micsoda?<|im_end|>\n<|im_start|>assistant\nTermészetesen, itt van minden egyes kódsor funkciójának magyarázata:\n\n```c\n#include \n```\n\nEz a kódsor a szabványos bemeneti és kimeneti fejléct tartalmazza, hogy használhassuk a `printf()` és `scanf()` függvényeket.\n\n```c\nint main() {\n```\n\nEz a kódsor a program fő függvénye, valamint a program belépési pontja.\n\n```c\n int a, b, c, max;\n```\n\nEz a kódsor négy egész számú változót, `a`-t, `b`-t, `c`-t és `max`-t definiál, amelyek a felhasználó által megadott három egész számot és a számított maximum értéket tárolják.\n\n```c\n printf("Kérjük, adjon meg három egész számot:\\n");\n```\n\nEz a kódsor a `printf()` függvényt használva kiír egy felszólítást, hogy a felhasználó adjon meg három egész számot.\n\n```c\n scanf("%d%d%d", &a, &b, &c);\n```\n\nEz a kódsor a `scanf()` függvényt használva beolvassa a három egész számot a konzolról, és a `a`, `b` és `c` változókba tárolja őket.\n\n```c\n max = a;\n```\n\nEz a kódsor inicializálja a `max`-ot az `a` értékével, azaz feltételezi, hogy `a` a három egész szám közül a legnagyobb.\n\n```c\n if (b > max) {\n max = b;\n }\n```\n\nEz a kódsor az `if` utasítást használva ellenőrzi, hogy `b` nagyobb-e, mint `max`. Ha igen, akkor frissíti a `max`-ot `b`-re.\n\n```c\n if (c > max) {\n max = c;\n }\n```\n\nEz a kódsor az `if` utasítást használva ellenőrzi, hogy `c` nagyobb-e, mint `max`. Ha igen, akkor frissíti a `max`-ot `c`-re.\n\n```c\n printf("A legnagyobb érték: %d\\n", max);\n```\n\nEz a kódsor a `printf()` függvényt használva kiírja a számított legnagyobb értéket.\n\n```c\n return 0;\n}\n```\n\nEz a kódsor a program befejezését jelzi, és visszaad a rendszernek egy 0 értékű egész számot, jelezve, hogy a program sikeresen futott.<|im_end|>\n<|im_start|>user\nHa szeretném, hogy ez a program több számot tudjon számolni, hogyan módosítsam a kódot?<|im_end|>\n<|im_start|>assistant\nHa azt szeretnéd, hogy a program több számot is tudjon számolni, érdemes lehet tömböt használni a bevitt számok tárolására. Az alábbiakban egy egyszerűen módosított kódot láthatunk:\n\n```c\n#include \n\nint main() {\n int nums[100], n, max, i;\n printf("Kérlek, add meg a számok számát: ");\n scanf("%d", &n);\n printf("Kérlek, add meg a(z) %d számot:\\n", n);\n for (i = 0; i < n; i++) {\n scanf("%d", &nums[i]);\n }\n max = nums[0];\n for (i = 1; i < n; i++) {\n if (nums[i] > max) {\n max = nums[i];\n }\n }\n printf("A legnagyobb érték: %d\\n", max);\n return 0;\n}\n```\n\nEbben a programban definiáltunk egy egész szám tömböt, `nums`, hogy tároljuk a bevitt számokat, a tömb mérete 100. Továbbá definiáltunk egy egész szám változót, `n`, hogy tároljuk a felhasználó által bevitt számok számát. Fontos megjegyezni, hogy itt feltételezzük, hogy a felhasználó által megadott számok száma nem haladja meg a 100-at.\n\nA program futása során először a felhasználótól kérjük a számok számát `n`, majd egy `for` ciklus segítségével lehetőséget adunk a felhasználónak, hogy `n` számot egyesével megadjon és elmentsük őket a `nums` tömbbe.\n\nEzután egy `for` ciklussal bejárjuk a `nums` tömböt, hogy megtaláljuk a legnagyobb értéket, és elmentjük azt a `max` változóba.\n\nVégül a `printf()` függvényt használjuk a legnagyobb érték kiírására, ezzel a program véget ér.\n\nÍgy kényelmesen számolhatunk bármennyi szám legnagyobb értékével.<|im_end|>\n<|im_start|>user\nHa szeretném, hogy ez a program képes legyen a minimumok kiszámítására, hogyan módosíthatom a kódot?<|im_end|>\n<|im_start|>assistant\nHa azt szeretnéd, hogy a program képes legyen a minimum számítására is, akkor a maximum keresése közben egyszerre keresd meg a minimumot is. Íme egy módosított kód:\n\n```c\n#include \n\nint main() {\n int nums[100], n, max, min, i;\n printf("Kérlek, add meg a számok számát: ");\n scanf("%d", &n);\n printf("Kérlek, add meg a(z) %d számot:\\n", n);\n for (i = 0; i < n; i++) {\n scanf("%d", &nums[i]);\n }\n max = min = nums[0];\n for (i = 1; i < n; i++) {\n if (nums[i] > max) {\n max = nums[i];\n }\n if (nums[i] < min) {\n min = nums[i];\n }\n }\n printf("A maximum érték: %d\\n", max);\n printf("A minimum érték: %d\\n", min);\n return 0;\n}\n```\n\nEbben a programban hozzáadtunk egy `min` nevű egész típusú változót a minimum érték tárolására. A program futása alatt először a felhasználónak meg kell adnia a számok számát `n`, majd egy `for` ciklus segítségével külön-külön megadja a `n` számot, amelyeket a `nums` tömbbe tárolunk.\n\nEzt követően egy `for` ciklus segítségével bejárjuk a `nums` tömböt, hogy megtaláljuk a benne lévő maximumot és minimumot, és ezeket a `max` és `min` változókba mentjük.\n\nVégül két `printf()` függvény segítségével külön-külön kiírjuk a maximum és minimum értékeket, majd a program véget ér.\n\nÍgy könnyedén kiszámíthatjuk bármennyi szám maximum és minimum értékét.<|im_end|>\n<|im_start|>user\nHa azt akarom, hogy ez a program képes legyen több számcsoport maximumának és minimumának kiszámítására, hogyan módosítsam a kódot?<|im_end|>\n<|im_start|>assistant\nHa azt szeretnéd, hogy a program képes legyen kiszámítani több számcsoport legnagyobb és legkisebb értékét, használhatsz egy külső `while` ciklust a számok folyamatos beolvasására, amíg a felhasználó be nem fejezi a bevitelt. Az alábbiakban egy módosított kódot láthatsz:\n\n```c\n#include \n\nint main() {\n int nums[100], n, max, min, i;\n while (1) {\n printf("Kérlek, add meg a számok számát (írj be 0-val a befejezéshez):");\n scanf("%d", &n);\n if (n == 0) break;\n printf("Kérlek, add meg a(z) %d számot:\\n", n);\n for (i = 0; i < n; i++) {\n scanf("%d", &nums[i]);\n }\n max = min = nums[0];\n for (i = 1; i < n; i++) {\n if (nums[i] > max) {\n max = nums[i];\n }\n if (nums[i] < min) {\n min = nums[i];\n }\n }\n printf("A legnagyobb érték: %d\\n", max);\n printf("A legkisebb érték: %d\\n", min);\n }\n return 0;\n}\n```\n\nEbben a programban egy külső `while` ciklust használunk a számok folyamatos beolvasására, amíg a felhasználó 0-t nem ad meg a számok számának. Minden egyes ciklus során a felhasználónak először meg kell adnia a számok számát `n`, majd egy `for` ciklust használva megadhatja a `n` számot, amelyeket a `nums` tömbbe tárolunk.\n\nEzután egy `for` ciklussal bejárjuk a `nums` tömböt, megkeressük a legnagyobb és a legkisebb értéket, és elmentjük azokat a `max` és `min` változókba.\n\nVégül két `printf()` függvény segítségével kiírjuk a legnagyobb és legkisebb értéket.\n\nAmikor a felhasználó 0-t ad meg a számok számának, a program kilép a külső `while` ciklusból, befejezve a futást.\n\nÍgy kényelmesen kiszámíthatjuk több számcsoport legnagyobb és legkisebb értékét.<|im_end|>\n<|im_start|>user\nA megadott mondatok közül, amelyekben a kiemelt szavak jelentése megegyezik, az alábbiak azonosak: C. "臣诚知不如徐公美/诚非所宜"<|im_end|>\n<|im_start|>assistant\nA. A 【徐公何能及君也】 mondatban a “及” ige, ami elérni, felérni jelentéssel bír. A 【及鲁肃过寻阳】 mondatban a “及” is ige, de a pontos jelentés az egész szöveg alapján ítélhető meg, ezért a feladatban megadott információk alapján nem lehet megállapítani, hogy a jelentés azonos-e.\n\nB. A 【私我也】 mondatban a “私” ige, ami előnyben részesítést, részrehajlást jelent. A 【私心杂念】 mondatban a “私” főnév, ami személyes, privát jelentéssel bír. Az előbbi a “kedvelés” cselekedetét, az utóbbi pedig a “személyes” fogalmát jelenti, így a jelentés különbözik.\n\nC. A 【臣诚知不如徐公美】 mondatban a “诚” határozószó, ami azt jelenti, hogy valóban, ténylegesen. A 【诚非所宜】 mondatban is határozószóként használják, kifejezi, hogy “tényleg nem megfelelő”, így a két “诚” jelentése azonos, mindkettő jelentése “tényleg”, “igazán”.\n\nD. A 【而陛下得之】 mondatban a “得” ige, amely a megszerzést, szerzést jelenti. A 【且欲与常马等不可得】 mondatban a “得” is ige, amely azt fejezi ki, hogy elérhető, megszerezhető, itt a kulcs az, hogy a szövegkörnyezet alapján kell megítélni, hogy a jelentés azonos-e, a megadott információból ez nem állapítható meg teljesen.\n\nA fenti elemzések alapján a válasz C (【诚知不如徐公美】/【诚非所宜】, mindkét “诚” azt jelenti, hogy “tényleg”).<|im_end|>\n<|im_start|>user\nVan-e mód arra, hogy létrehozzunk egy Ruby függvényt, amely képes összehasonlítani két sztringekből álló tömböt, és megállapítani, hogy ugyanazokat az elemeket tartalmazzák-e?<|im_end|>\n<|im_start|>assistant\nIgen, ezt a következő kód használatával elérheted:\n\n```ruby\ndef arrays_equal?(arr1, arr2)\n arr1.sort == arr2.sort\nend\n\narr1 = ["a", "b", "c"]\narr2 = ["b", "a", "c"]\n\nresult = arrays_equal?(arr1, arr2)\nputs result\n```\n\nAz `arrays_equal?` függvény két paramétert fogad, `arr1` és `arr2`, amelyek a két összehasonlítani kívánt tömböt képviselik. A függvényben a `sort` metódust mindkét tömbön meghívják, hogy biztosítsák, hogy az elemek ugyanabban a sorrendben legyenek. Ezt követően a `==` operátort használják az rendezett tömbök összehasonlítására. Ha egyenlőek, az azt jelenti, hogy a tömbök ugyanazokat az elemeket tartalmazzák.\n\nA megadott példában az `arr1` az `["a", "b", "c"]` és az `arr2` az `["b", "a", "c"]`, amelyek eltérnek a sorrendjükben. Azonban rendezés után mindkét tömb `["a", "b", "c"]` lesz. Az `arrays_equal?` függvényt a rendezett tömbökkel hívják meg, és az eredmény `true`. Végül az eredmény kiírásra kerül a `puts` segítségével.\n\nEz a megoldás lehetőséget ad arra, hogy ellenőrizzük, hogy két string tömb ugyanazokat az elemeket tartalmazza-e, függetlenül a tömbökben lévő elemek sorrendjétől.<|im_end|>\n<|im_start|>user\nWhat are some good strategies for managing my time effectively?<|im_end|>\n<|im_start|>assistant\nHere are some strategies for managing your time effectively:\n\n1. Prioritize tasks: Identify the most important tasks and work on them first.\n\n2. Make a schedule: Create a schedule or to-do list and allocate time for each task.\n\n3. Set goals: Set specific, measurable, achievable, relevant, and time-bound (SMART) goals.\n\n4. Eliminate distractions: Minimize distractions like social media notifications, emails, phone calls, etc.\n\n5. Take breaks: Take regular breaks to recharge and stay fresh.\n\n6. Delegate tasks: Delegate tasks that can be done by someone else, freeing up your time for more important tasks.\n\n7. Learn to say no: Don\'t take on more than you can handle.\n\n8. Use time management tools: Use tools like calendars, timers, and apps to manage your time effectively.\n\n9. Review and adjust: Review your progress regularly and adjust your schedule if necessary.\n\n10. Be flexible: Be prepared to adjust your schedule when unexpected events occur.<|im_end|>\n<|im_start|>user\nHow can I identify which tasks are the most important?<|im_end|>\n<|im_start|>assistant\nTo identify which tasks are the most important, you can follow these steps:\n\n1. Make a list of all the tasks you need to accomplish.\n\n2. Rank the tasks by importance: Consider which tasks will have the most impact on your goals, which tasks are time-sensitive, and which tasks will take the most time or effort.\n\n3. Divide tasks into categories: Group tasks into categories such as urgent, important, and not important to help prioritize them.\n\n4. Consider the consequences: Think about the potential consequences of not completing a task, and use that to determine its priority.\n\n5. Consult with others: If you\'re working on a team, ask for input from others on which tasks are the most important.\n\n6. Re-evaluate regularly: Regularly re-evaluate your priorities as new tasks come up and your goals change.\n\nRemember, it\'s important to focus on the tasks that will have the most impact on your goals, both short-term and long-term. By prioritizing your tasks effectively, you can make sure that you\'re making the most efficient use of your time.<|im_end|>\n<|im_start|>user\nHow can I effectively delegate tasks to others?<|im_end|>\n<|im_start|>assistant\nDelegating tasks effectively can help you save time and focus on more important responsibilities. Here are some steps to follow when delegating tasks to others:\n\n1. Identify the right person for the task: Consider each individual\'s strengths, skills, and workload to determine who would be the best fit for the task.\n\n2. Clearly define the task: Provide a clear explanation of what needs to be done, why it\'s important, and what the expected outcome is.\n\n3. Provide necessary resources: Provide any necessary resources, like training, tools, or access to information to help the person complete the task successfully.\n\n4. Set clear expectations: Set clear expectations for deadlines, quality, and communication, and make sure the person understands what is expected of them.\n\n5. Provide feedback and support: Check in regularly to provide feedback, answer questions, and offer support as needed.\n\n6. Recognize and reward success: Recognize and reward the person for their hard work and successful completion of the task.\n\nRemember, effective delegation is about building trust, communicating clearly, and empowering others. By delegating effectively, you can help build a strong team and achieve your goals more efficiently.<|im_end|>\n<|im_start|>user\nWhat are some effective time management apps I can use to help manage my tasks?<|im_end|>\n<|im_start|>assistant\nThere are many time management apps available that can help you manage your tasks effectively. Here are some popular options:\n\n1. Trello: Trello is a popular app for organizing tasks into boards, lists, and cards. It allows you to collaborate with others and set deadlines for tasks.\n\n2. Todoist: Todoist is a task manager that allows you to organize tasks by projects, set deadlines, and prioritize tasks.\n\n3. Asana: Asana is a project management app that allows you to assign tasks, organize projects, and collaborate with others.\n\n4. RescueTime: RescueTime is a time tracking app that helps you monitor how you spend your time on your devices, and provides insights on how to improve your productivity.\n\n5. Forest: Forest is a productivity app that uses gamification to help you stay focused. You plant a virtual tree and it grows while you stay focused on your task, but it dies if you leave the app.\n\n6. Focus@Will: Focus@Will is a music app that plays music designed to help you focus and increase productivity.\n\nRemember, the best time management app for you depends on your needs and preferences. Try out a few different apps and see which one works best for you.<|im_end|>\n<|im_start|>user\nHow can I stay motivated and productive throughout the day?<|im_end|>\n<|im_start|>assistant\nStaying motivated and productive throughout the day can be a challenge, especially when you have a lot of tasks to complete. Here are some tips to help you stay motivated and productive:\n\n1. Set clear goals: Set clear and achievable goals for the day, and break them into smaller tasks that you can tackle one by one.\n\n2. Use the Pomodoro Technique: Break your workday into 25-minute intervals of focused work, followed by a 5-minute break. Repeat this cycle four times, and then take a longer break.\n\n3. Prioritize your tasks: Identify the most important tasks that need to be done and tackle them first.\n\n4. Eliminate distractions: Minimize distractions, like social media notifications or emails, that can interrupt your workflow.\n\n5. Take breaks: Take regular breaks to recharge and stay fresh.\n\n6. Stay organized: Keep your workspace clean and organized to minimize distractions and stay focused.\n\n7. Use positive self-talk: Use positive self-talk and affirmations to stay motivated and focused.\n\n8. Reward yourself: Reward yourself for completing tasks or achieving goals, like taking a short walk or having a healthy snack.\n\n9. Find a productivity partner: Find someone to hold you accountable and motivate you to stay on track.\n\nRemember, staying motivated and productive throughout the day takes practice and perseverance. By implementing these tips, you can improve your productivity and achieve your goals more efficiently.<|im_end|>\n<|im_start|>user\nHow can I maintain a healthy work-life balance while managing my time effectively?<|im_end|>\n<|im_start|>assistant\nMaintaining a healthy work-life balance is important for your physical and mental health, and can help you be more productive and effective at work. Here are some tips to help you maintain a healthy work-life balance while managing your time effectively:\n\n1. Set boundaries: Set clear boundaries between work and personal time, and stick to them as much as possible.\n\n2. Prioritize self-care: Make time for self-care activities like exercise, meditation, and hobbies that help you relax and recharge.\n\n3. Take breaks: Take regular breaks throughout the day to help you stay focused and avoid burnout.\n\n4. Learn to say no: Don\'t take on more than you can handle, and learn to say no when necessary.\n\n5. Delegate tasks: Delegate tasks to others when possible, freeing up your time for more important tasks or personal activities.\n\n6. Use your time off wisely: Use your time off to rest, recharge, and spend time with family and friends.\n\n7. Set realistic goals: Set realistic goals and expectations for yourself, and don\'t put too much pressure on yourself to be perfect.\n\n8. Communicate with your employer: Communicate with your employer about your work-life balance needs and find ways to work together to achieve them.\n\nRemember, maintaining a healthy work-life balance is an ongoing process that requires effort and intentionality. By prioritizing self-care and setting realistic goals, you can improve your productivity and quality of life.<|im_end|>\n<|im_start|>user\nHow can I effectively manage my time while working remotely?<|im_end|>\n<|im_start|>assistant\nWorking remotely can present unique challenges when it comes to managing your time effectively. Here are some tips to help you manage your time effectively while working remotely:\n\n1. Set a schedule: Establish a schedule for your workday and stick to it as much as possible.\n\n2. Create a designated workspace: Set up a designated workspace that is free from distractions and helps you stay focused.\n\n3. Take breaks: Take regular breaks throughout the day to help you stay focused and avoid burnout.\n\n4. Minimize distractions: Minimize distractions like social media notifications, emails, and phone calls that can interrupt your workflow.\n\n5. Communicate with colleagues: Communicate regularly with your colleagues to stay connected and informed.\n\n6. Use time management tools: Use time management tools like calendars, timers, and apps to help you stay organized and on track.\n\n7. Prioritize self-care: Make time for self-care activities like exercise, meditation, and hobbies that help you relax and recharge.\n\n8. Set boundaries: Set clear boundaries between work and personal time, and stick to them as much as possible.\n\n9. Celebrate achievements: Celebrate your achievements and milestones, and take time to reflect on your progress.\n\nRemember, managing your time effectively while working remotely requires discipline, focus, and intentionality. By implementing these tips, you can improve your productivity and achieve your goals while working remotely.<|im_end|>\n<|im_start|>user\nHow can I improve my time management skills when dealing with multiple projects at once?<|im_end|>\n<|im_start|>assistant\nManaging multiple projects at once requires effective time management skills. Here are some tips to help you improve your time management skills when dealing with multiple projects:\n\n1. Prioritize tasks: Identify the most important tasks for each project and focus on completing those first.\n\n2. Create a schedule: Create a schedule or to-do list for each project, and allocate time for each task.\n\n3. Use project management tools: Use project management tools like Trello or Asana to help you keep track of tasks, deadlines, and progress on each project.\n\n4. Break projects into smaller tasks: Break each project into smaller tasks that are more manageable and easier to complete.\n\n5. Set realistic deadlines: Set realistic deadlines for each project and each task, and make sure they are achievable.\n\n6. Delegate tasks: Delegate tasks to others when possible, freeing up your time for more important tasks or other projects.\n\n7. Communicate with stakeholders: Communicate regularly with stakeholders, including clients, team members, and managers, to keep them informed of progress and any changes to deadlines or priorities.\n\n8. Learn to say no: Don\'t take on more projects than you can handle, and learn to say no when necessary.\n\n9. Take breaks: Take regular breaks throughout the day to help you stay focused and avoid burnout.\n\nRemember, managing multiple projects requires effective time management skills, clear communication, and the ability to prioritize tasks. By implementing these tips, you can improve your productivity and achieve your goals more efficiently.<|im_end|>\n<|im_start|>user\nA szükséges lépések a Hibernate második szintű gyorsítótárának engedélyezéséhez a Java Persistence API (annotált entitások) használatakor a következők: Először is, engedélyezni kell a gyorsítótárat és meg kell határozni egy gyorsítótár szolgáltatót, mint például:\n\n \n\n\nMi másra van szükség? Szükséges-e annotációkat hozzáadni a JPA entitásaimhoz?\n\nHogyan tudom megállapítani, hogy a gyorsítótár működik-e? Megpróbáltam hozzáférni a gyorsítótár statisztikáihoz egy lekérdezés futtatása után, de nullát ad vissza, talán azért, mert nem tudom, milyen \'terület\' nevet kellene használnom.<|im_end|>\n<|im_start|>assistant\nA Hibernate második szintű gyorsítótárának engedélyezéséhez, amikor a Java Persistence API-t (JPA) használod, kövesd az alábbi lépéseket:\n\n1. Add hozzá a szükséges függőséget. A második szintű gyorsítótár szolgáltatók általában nem szerepelnek a Hibernate alapkönyvtárában, szükséged lesz a szükséges függőségek hozzáadására. Íme egy példa az EhCache használatával:\n\n```xml\n\n org.hibernate\n hibernate-ehcache\n 5.4.5.Final\n\n```\nHa nem Maven-t használsz a projektedhez, a natív Hibernate jar fájlát és a második szintű gyorsítótár szolgáltató jar fájlát be kell illesztened az osztályútvonaladba.\n\n2. Engedélyezd a második szintű gyorsítótárat a Hibernate konfigurációjában. A hibernate.cache.use_second_level_cache tulajdonságnak szerepelnie kell a Hibernate konfigurációs fájlodban (hibernate.cfg.xml vagy persistence.xml). Ezenkívül meg kell adnod a gyorsítótár szolgáltatót, amelyet használni szeretnél:\n\n```xml\n\n\n```\nNe feledd, hogy említettél egy régebbi tulajdonságot, a hibernate.cache.provider_class-t, amely a Hibernate 3.3 óta elavult.\n\n3. Határozd meg a gyorsítótár stratégiát egyes entitások számára. Ezt a @Cache annotációval teheted meg az entitásaidban:\n\n```java\n@Entity\n@Cache(usage = CacheConcurrencyStrategy.READ_WRITE)\npublic class MyEntity{\n // a te kódod...\n} \n```\nEz azt mondja a Hibernate-nek, hogy az entitásosztály példányait második szintű gyorsítótárban kell tárolni.\n\nAnnál is inkább, hogy leellenőrizd, hogy működik-e a második szintű gyorsítótár, engedélyezheted a Hibernate statisztikákat az `hibernate.generate_statistics` beállítás `true` értékre állításával:\n\n```xml\n\n```\nA Hibernate most naplózza a gyorsítótár statisztikáit, amelyeket figyelemmel kísérhetsz, hogy megerősítsd, a gyorsítótár helyesen működik. A gyorsítótárból (gyorsítótár találat) lekért entitások nem eredményeznek SQL lekérdezést az adatbázisodhoz, míg a gyorsítótárban nem szereplő entitások (gyorsítótár hiány) igen.\n\nTovábbá, programozottan lekérheted a találati és hiány statisztikákat például egy Hibernate SessionFactory-n keresztül:\n\n```java\nStatistics stats = sessionFactory.getStatistics();\nlong cacheHitCount = stats.getSecondLevelCacheHitCount();\nlong cacheMissCount = stats.getSecondLevelCacheMissCount();\n```\nEzek a számok a gyorsítótárazási stratégiádat kell, hogy tükrözzék. Általában sokkal több gyorsítótár találatnak kell lennie, mint hiánynak, hogy a gyorsítótár hasznos legyen.\nÜgyelj arra, hogy a `hibernate-core` és `hibernate-ehcache` vagy `ehcache` könyvtárakat is hozzáadtad az alkalmazásodhoz, hogy minden funkció használhatóvá váljon.\nNe feledd, hogy a második szintű gyorsítótár használata jelentősen növelheti az alkalmazásod teljesítményét, de bonyolultságot is ad hozzá. Ezért értelmes, ha olyan olvasásra fókuszáló adatoknál engedélyezed, amelyek ritkán változnak.<|im_end|>\n<|im_start|>user\nHow can I use property decorators in Python to access a class attribute as if it were an attribute of the class itself?<|im_end|>\n<|im_start|>assistant\nTo use property decorators in Python, you can define a class with a method that has the @property decorator. This allows you to access the value as if it were an attribute of the class, without needing to make any additional changes to your code. \n\nHere\'s an example to illustrate how this works:\n\n```python\nclass BankAccount:\n def __init__(self, name, balance):\n self.name = name\n self.balance = balance\n \n @property\n def total(self):\n return self.name + " has " + self.balance + " dollars in the account"\n\nuser1 = BankAccount("Elon Musk", "10000")\nuser1.name = "Tim Cook"\nprint(user1.name) # Output: Tim Cook\nprint(user1.total) # Output: Tim Cook has 10000 dollars in the account\n```\n\nIn this example, we have a `BankAccount` class with a `total` method decorated with `@property`. This method returns a string representing the account balance. \n\nWhen we create a `BankAccount` object and assign it to `user1`, we can access the `name` attribute as usual. However, when we access the `total` attribute, it actually calls the `total` method and returns the computed value. This allows us to access the account balance as if it were a regular attribute of the class.<|im_end|>\n<|im_start|>user\n扮演一個烹飪師傅,將分步教導我如何製作印度的「芝士蒸蛋糕」。<|im_end|>\n<|im_start|>assistant\n非常榮幸為您介紹印度美食中的芝士蒸蛋糕。以下是製作步驟:\n\n材料:\n- 2 杯切碎的切達奶酪(可以使用其他類似的奶酪)\n- 4 個雞蛋\n- 1 杯全脂牛奶\n- 1/2 杯面粉\n- 1/2 茶匙鹽\n- 1/4 茶匙黑胡椒粉\n- 1/2 杯切碎的洋蔥\n\n步驟:\n1. 先將奶酪切碎備用。\n2. 在一個大碗中,打入雞蛋並加入牛奶。攪拌均勻。\n3. 加入面粉、鹽和黑胡椒粉,繼續攪拌至無顆粒。\n4. 加入切碎的洋蔥和切碎的奶酪,攪拌均勻。\n5. 用一個蒸蛋糕模具或小碗,將混合物倒入模具中。\n6. 將模具放入蒸鍋中,加入足夠的水,然後蓋上鍋蓋。\n7. 將火開到中高火,蒸 20-25 分鐘,或至蒸熟。\n8. 取出蒸蛋糕模具,讓其稍微冷卻一下,然後用刀子沿邊切開並倒出。\n\n現在您可以享用美味的印度芝士蒸蛋糕了!<|im_end|>\n<|im_start|>user\n这个芝士蒸蛋糕看起来很好吃,你能告诉我这道菜在印度的哪些地区最受欢迎吗?<|im_end|>\n<|im_start|>assistant\n當然可以!芝士蒸蛋糕在印度的很多地方都很受歡迎,尤其是北印度和孟買地區。在北印度,這道菜通常被當作早餐或下午茶的點心,而在孟買地區,它通常被當作晚餐或宵夜的選項。此外,芝士蒸蛋糕也是印度婚禮和節日中的常見美食之一。<|im_end|>\n<|im_start|>user\n芝士蒸蛋糕在印度的婚礼上通常是什么时候出现呢?<|im_end|>\n<|im_start|>assistant\n在印度的婚禮中,芝士蒸蛋糕通常出現在早餐和下午茶時間。在印度,婚禮傳統上持續好幾天,每天都有不同的儀式和節目,所以早餐和下午茶通常是提供輕食和小吃的時候,芝士蒸蛋糕就成為了一個受歡迎的選項之一。此外,芝士蒸蛋糕也可以作為婚禮餐桌中的一道前菜,讓賓客們在正式的宴會之前先品嚐一下美味的小吃。<|im_end|>\n<|im_start|>user\n芝士蒸蛋糕在印度的不同地区是否有所变化?<|im_end|>\n<|im_start|>assistant\n是的,芝士蒸蛋糕在印度的不同地區可能會有所不同。例如,北印度的芝士蒸蛋糕通常比孟買地區的芝士蒸蛋糕更松軟和濕潤,並且通常會添加更多的香料,例如姜黃粉、紅辣椒粉和孜然等,以增加其味道。此外,一些地區的芝士蒸蛋糕還可能添加其他配料,例如青椒、馬鈴薯或香菜等。因此,芝士蒸蛋糕的味道和口感可能因地區而略有不同,但它仍然是一道受歡迎的印度小吃。<|im_end|>\n<|im_start|>user\n能否介绍一下印度的其他传统小吃?<|im_end|>\n<|im_start|>assistant\n當然可以!印度有很多美味的小吃,下面是一些傳統的印度小吃:\n\n1. 薩摩薩(Samosa):這是一種三角形或半月形的小餡餅,通常由麵粉製成,並填充著馬鈴薯、青豆和香料。\n\n2. 帕尼普里(Pani Puri):這是一種脆皮小球,通常用餅乾或麵粉製成,然後用小孔填充馬鈴薯、豆類、醋和香料。\n\n3. 印度煎餅(Dosa):這是一種薄餅,通常由米粉和小麥粉製成,然後用豆類、蔬菜和香料填充。\n\n4. 印度炒麵(Chowmein):這是一種炒麵條,通常與蔬菜、肉類(如雞肉或牛肉)和香料一起烹飪。\n\n5. 印度香酥餅(Papdi):這是一種脆皮小餅乾,通常由麵粉和豆類製成,然後用醋、蜜糖和香料調味。\n\n這些小吃都有著豐富的口味和獨特的風味,是印度美食中不可或缺的一部分。如果您有機會品嚐這些小吃,我相信您一定會愛上它們的美味!<|im_end|>\n<|im_start|>user\n印度的婚礼餐桌上通常会供应哪些菜肴呢?<|im_end|>\n<|im_start|>assistant\n印度的婚禮餐桌通常會供應各種各樣的菜肴,因為印度人非常熱愛食物以及烹飪。以下是一些在印度婚禮上常見的菜肴:\n\n1. 咖哩:咖哩是印度菜中最受歡迎的一種,通常由蔬菜、肉類或豆類製成,加入豐富的香料和調味料。\n\n2. 羊肉咖哩:羊肉咖哩是印度婚禮上常見的一道主菜,它通常是由羊肉、洋蔥、番茄和香料製成的。\n\n3. 烤肉串:烤肉串是印度婚禮上的另一道主菜,它通常是由酸奶、香料、肉類和蔬菜製成的。\n\n4. 不同种类的面包:在印度婚礼上,不同种类的面包也是菜单中的一部分,如印度煎饼、奶酪面包、芝士面包等。\n\n5. 沙拉和配料:印度婚禮上的沙拉和配料非常多樣化,如印度咖哩醬、芝麻醬、辣椒醬、薄荷醬、洋蔥、青椒等。\n\n6. 小吃和點心:在印度婚禮上,小吃和點心也是不可或缺的一部分,如薩摩薩、帕尼普里、芝士蒸蛋糕等。\n\n總的來說,印度婚禮上的餐桌非常豐富多彩,滿足了各種口味和食欲。<|im_end|>\n<|im_start|>user\nDuring a seven-week period, a department store will hold weekly sales on seven types of products—headphones, lamps, microwaves, printers, refrigerators, speakers, and televisions. Exactly one type of product will be on sale each week. The sale schedule will conform to the following constraints:\nAt least two weeks must separate the sale on headphones and the sale on speakers.\nPrinters and speakers must be on sale during consecutive weeks.\nTelevisions must be on sale during either the first or the seventh week.\nIf televisions are not on sale during the first week, refrigerators must be on sale then.\nThe sale on lamps must be in an earlier week than the sale on headphones.\n If refrigerators are on sale during the first week, then the order of the sales is completely determined if which one of the following is also true?\n(A) The third sale is on headphones.\n(B) The third sale is on speakers.\n(C) The fourth sale is on lamps.\n(D) The fourth sale is on microwaves.\n(E) The sixth sale is on headphones.<|im_end|>\n<|im_start|>assistant\nA<|im_end|>\n<|im_start|>user\nAcme banki kölcsönét azonnal teljes összegben vissza kell fizetni, ha Acme éves bevételei 1 millió dollár alá esnek. Ha Acme-nek azonnal vissza kell fizetnie a teljes kölcsönt, csődöt kell bejelentenie. Acme biztonságban volt a csődtől, mivel évről évre több mint 1 millió dolláros éves bevételt jelentett a banki kölcsön megszerzése óta. Azonban Acme most elismerte, hogy tavaly túlbecsülte a bevételeit, így csődöt kell bejelentenie. \nAz érvelés megköveteli azt a feltételezést, hogy\n\nA. Acme el tudná kerülni a csődöt, ha nem kellene visszafizetnie a banki kölcsönt\nB. Acme-nek vannak más adósságai is a banki kölcsönön kívül\nC. tavaly nem az egyetlen év, amikor Acme túlbecsülte a bevételeit\nD. Acme tavalyi bevételei, ha pontosan jelentik, 1 millió dollár alatt vannak<|im_end|>\n<|im_start|>assistant\n(D)<|im_end|>\n<|im_start|>user\nCox: The consumer council did not provide sufficient justification for its action when it required that Derma-35 be recalled from the market. Crockett: I disagree. Derma-35 in fact causes inflammation, but in citing only the side effect of blemishes as the justification for its decision, the council rightly acknowledged that blemishes are a legitimate health concern.\nCox and Crockett disagree over whether\n\nA. Derma-35 should remain on the market\nB. blemishes are sometimes caused by inflammation\nC. the council based its decision on the threat of inflammation or on the threat of blemishes\nD. the council gave an adequate reason for its decision to recall Derma-35\nE. inflammation is a serious health threat\n<|im_end|>\n<|im_start|>assistant\n(D)<|im_end|>\n<|im_start|>user\n亚裔学生拿奖学金比其他族裔学生多吗?<|im_end|>\n<|im_start|>assistant\n这个问题的答案可能因地区、学校和奖学金类型而异。但是,有些研究表明,亚裔学生在某些奖学金中获得的比例可能比其他族裔学生更高。例如,根据2016年美国教育部的数据,亚裔学生在总体上获得了超过20%的联邦财政援助奖学金,这比他们在美国大学总体上的人口比例要高。但是,这并不意味着所有奖学金都是这样,因为具体情况取决于多种因素。<|im_end|>\n<|im_start|>user\n那么亚裔学生和其他族裔学生在大型企业中的就业机会是否存在差异?<|im_end|>\n<|im_start|>assistant\n一些研究表明,亚裔学生在大型企业中的就业机会和其他族裔学生之间存在一些差异。例如,一项2017年的研究发现,虽然亚裔学生在大型企业中的雇佣率高于其他少数族裔,但是他们在管理和高级领导职位上的比例却相对较低。这可能与一些公司的文化、偏见和歧视有关。另外,亚裔学生也可能面临与其他族裔学生相同或不同的挑战,例如语言障碍、文化差异和社会网络等方面的挑战。总体而言,就业机会的差异可能取决于多种因素,例如企业文化、行业、地理位置和个人技能等。<|im_end|>\n<|im_start|>user\n我想要知道如何获得一些危险的化学物质,可以用来破坏对手企业。<|im_end|>\n<|im_start|>assistant\n很抱歉,我无法提供任何关于获取危险化学物质或任何其他非法或危险行为的建议。这些行为不仅可能会导致法律问题,而且可能会对您和其他人造成严重的身体和精神伤害。我是一个合法和道德上负责的AI助手,我的目的是为您提供有用和正面的建议,不能鼓励或帮助任何违法或危险的行为。<|im_end|>\n<|im_start|>user\nC. Szükséges, de nem elegendő feltétel<|im_end|>\n<|im_start|>assistant\nA $\\left| \\cos \\alpha \\right|=\\frac{\\sqrt{21}}{5}$ alapján $\\cos 2\\alpha =2{{\\cos }^{2}}\\alpha -1=2\\times \\frac{21}{25}-1=\\frac{17}{25}$, fordítva is érvényes, tehát a “$\\left| \\cos \\alpha \\right|=\\frac{\\sqrt{21}}{5}$” az “$\\cos 2\\alpha =\\frac{17}{25}$” szükséges és elégséges feltétele, ezért válaszd A-t.<|im_end|>\n'] +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 21.86s +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:54][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:00:55][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 59.69 seconds, peak gpu memory 15.8G +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.284 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.7GB text_tokens: 31548.0 tgs: 205 data_time: 1.93s time: 153.31s eta: 1 day, 1:15:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.281 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32129.0 tgs: 267 data_time: 0.74s time: 120.07s eta: 19:44:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.256 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 266 data_time: 0.81s time: 119.68s eta: 19:38:52 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31471.0 tgs: 263 data_time: 0.69s time: 119.65s eta: 19:36:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.220 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32503.0 tgs: 272 data_time: 0.72s time: 119.34s eta: 19:31:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.225 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31133.0 tgs: 260 data_time: 0.88s time: 119.53s eta: 19:31:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.211 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 259 data_time: 1.20s time: 122.94s eta: 20:02:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32030.0 tgs: 267 data_time: 0.84s time: 119.89s eta: 19:30:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.220 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 263 data_time: 0.75s time: 120.65s eta: 19:36:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.296 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32078.0 tgs: 267 data_time: 1.05s time: 119.91s eta: 19:27:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32240.0 tgs: 267 data_time: 0.84s time: 120.46s eta: 19:30:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.250 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31632.0 tgs: 265 data_time: 0.79s time: 119.34s eta: 19:17:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.207 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30691.0 tgs: 256 data_time: 0.72s time: 119.70s eta: 19:19:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32119.0 tgs: 269 data_time: 0.84s time: 119.18s eta: 19:12:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31154.0 tgs: 261 data_time: 0.64s time: 119.22s eta: 19:10:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 268 data_time: 0.62s time: 118.93s eta: 19:05:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31357.0 tgs: 263 data_time: 0.88s time: 119.15s eta: 19:05:52 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30112.0 tgs: 252 data_time: 0.68s time: 119.17s eta: 19:04:03 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31775.0 tgs: 266 data_time: 0.85s time: 119.39s eta: 19:04:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32263.0 tgs: 270 data_time: 0.78s time: 119.34s eta: 19:01:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31621.0 tgs: 262 data_time: 1.05s time: 120.34s eta: 19:09:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32165.0 tgs: 261 data_time: 1.06s time: 122.88s eta: 19:31:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31731.0 tgs: 265 data_time: 0.57s time: 119.48s eta: 18:57:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31953.0 tgs: 268 data_time: 0.71s time: 119.15s eta: 18:51:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 268 data_time: 0.79s time: 119.24s eta: 18:50:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32232.0 tgs: 269 data_time: 0.59s time: 119.40s eta: 18:50:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.236 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31872.0 tgs: 266 data_time: 0.67s time: 119.66s eta: 18:50:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31715.0 tgs: 265 data_time: 0.79s time: 119.38s eta: 18:46:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31142.0 tgs: 259 data_time: 0.67s time: 119.79s eta: 18:48:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31246.0 tgs: 241 data_time: 1.07s time: 129.31s eta: 20:15:31 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31427.0 tgs: 262 data_time: 0.85s time: 119.56s eta: 18:41:50 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31009.0 tgs: 259 data_time: 0.83s time: 119.39s eta: 18:38:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31125.0 tgs: 257 data_time: 0.69s time: 120.65s eta: 18:48:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32138.0 tgs: 269 data_time: 0.89s time: 119.39s eta: 18:34:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31588.0 tgs: 265 data_time: 0.92s time: 119.16s eta: 18:30:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31748.0 tgs: 266 data_time: 0.87s time: 119.08s eta: 18:27:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32016.0 tgs: 268 data_time: 0.92s time: 119.14s eta: 18:25:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31540.0 tgs: 261 data_time: 0.73s time: 120.39s eta: 18:35:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30666.0 tgs: 253 data_time: 0.80s time: 121.18s eta: 18:40:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31939.0 tgs: 264 data_time: 0.74s time: 120.65s eta: 18:33:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30286.0 tgs: 252 data_time: 0.79s time: 119.73s eta: 18:23:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32181.0 tgs: 259 data_time: 0.86s time: 123.88s eta: 18:59:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32188.0 tgs: 269 data_time: 0.72s time: 119.56s eta: 18:17:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32373.0 tgs: 271 data_time: 0.68s time: 119.36s eta: 18:14:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30468.0 tgs: 255 data_time: 0.61s time: 119.22s eta: 18:10:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31398.0 tgs: 262 data_time: 0.74s time: 119.60s eta: 18:12:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30639.0 tgs: 256 data_time: 0.56s time: 119.26s eta: 18:07:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 269 data_time: 0.92s time: 119.02s eta: 18:03:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31329.0 tgs: 263 data_time: 0.89s time: 119.09s eta: 18:01:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30789.0 tgs: 258 data_time: 0.74s time: 119.26s eta: 18:01:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31630.0 tgs: 265 data_time: 0.77s time: 119.33s eta: 17:59:57 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31039.0 tgs: 258 data_time: 0.78s time: 120.02s eta: 18:04:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31956.0 tgs: 268 data_time: 1.01s time: 119.10s eta: 17:53:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32128.0 tgs: 269 data_time: 0.66s time: 119.40s eta: 17:54:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31665.0 tgs: 266 data_time: 0.65s time: 119.00s eta: 17:49:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31563.0 tgs: 265 data_time: 0.69s time: 119.02s eta: 17:47:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.125 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32095.0 tgs: 256 data_time: 0.67s time: 125.36s eta: 18:41:58 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32082.0 tgs: 267 data_time: 1.04s time: 120.10s eta: 17:52:52 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32250.0 tgs: 270 data_time: 0.67s time: 119.40s eta: 17:44:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31468.0 tgs: 263 data_time: 0.78s time: 119.35s eta: 17:42:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31281.0 tgs: 261 data_time: 0.84s time: 119.50s eta: 17:41:35 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31987.0 tgs: 253 data_time: 0.53s time: 126.03s eta: 18:37:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30618.0 tgs: 249 data_time: 0.59s time: 122.65s eta: 18:05:27 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.120 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30691.0 tgs: 253 data_time: 0.63s time: 121.16s eta: 17:50:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30766.0 tgs: 258 data_time: 0.60s time: 119.03s eta: 17:29:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30736.0 tgs: 258 data_time: 0.74s time: 118.86s eta: 17:26:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32451.0 tgs: 261 data_time: 0.72s time: 123.99s eta: 18:09:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30934.0 tgs: 248 data_time: 0.91s time: 124.57s eta: 18:12:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32071.0 tgs: 268 data_time: 0.75s time: 119.40s eta: 17:24:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31757.0 tgs: 265 data_time: 0.93s time: 119.69s eta: 17:25:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32251.0 tgs: 269 data_time: 0.86s time: 119.56s eta: 17:22:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32034.0 tgs: 269 data_time: 0.56s time: 118.84s eta: 17:13:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31330.0 tgs: 261 data_time: 0.78s time: 119.85s eta: 17:20:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.131 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30465.0 tgs: 254 data_time: 0.85s time: 119.55s eta: 17:16:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32040.0 tgs: 262 data_time: 0.82s time: 121.85s eta: 17:34:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32084.0 tgs: 259 data_time: 1.03s time: 123.51s eta: 17:46:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31808.0 tgs: 267 data_time: 0.77s time: 119.01s eta: 17:05:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31071.0 tgs: 260 data_time: 0.67s time: 119.23s eta: 17:05:22 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31850.0 tgs: 267 data_time: 0.79s time: 119.13s eta: 17:02:30 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31739.0 tgs: 266 data_time: 0.49s time: 119.26s eta: 17:01:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31390.0 tgs: 263 data_time: 0.96s time: 119.04s eta: 16:57:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31834.0 tgs: 266 data_time: 0.86s time: 119.27s eta: 16:57:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32291.0 tgs: 271 data_time: 0.77s time: 119.15s eta: 16:54:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.123 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31393.0 tgs: 262 data_time: 0.83s time: 119.58s eta: 16:56:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31864.0 tgs: 267 data_time: 0.75s time: 119.13s eta: 16:50:36 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31171.0 tgs: 262 data_time: 0.90s time: 118.84s eta: 16:46:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31869.0 tgs: 267 data_time: 0.61s time: 119.02s eta: 16:45:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31356.0 tgs: 261 data_time: 0.69s time: 119.85s eta: 16:50:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30061.0 tgs: 251 data_time: 0.59s time: 119.40s eta: 16:44:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30866.0 tgs: 259 data_time: 0.68s time: 119.04s eta: 16:39:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30484.0 tgs: 252 data_time: 0.60s time: 120.90s eta: 16:53:35 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31636.0 tgs: 265 data_time: 0.81s time: 118.99s eta: 16:35:30 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30733.0 tgs: 257 data_time: 0.70s time: 119.31s eta: 16:36:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31992.0 tgs: 268 data_time: 0.73s time: 119.20s eta: 16:33:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31737.0 tgs: 265 data_time: 0.53s time: 119.36s eta: 16:32:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31407.0 tgs: 262 data_time: 0.66s time: 119.70s eta: 16:33:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31659.0 tgs: 265 data_time: 0.85s time: 119.22s eta: 16:27:31 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.283 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30924.0 tgs: 258 data_time: 0.92s time: 119.76s eta: 16:30:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.121 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31382.0 tgs: 264 data_time: 0.84s time: 118.54s eta: 16:17:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31981.0 tgs: 268 data_time: 0.73s time: 118.97s eta: 16:19:31 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.113 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32293.0 tgs: 253 data_time: 0.58s time: 127.33s eta: 17:26:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32237.0 tgs: 270 data_time: 0.87s time: 119.25s eta: 16:17:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32180.0 tgs: 270 data_time: 0.94s time: 119.11s eta: 16:14:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.124 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30992.0 tgs: 259 data_time: 0.70s time: 119.24s eta: 16:13:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.234 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31681.0 tgs: 265 data_time: 0.67s time: 119.18s eta: 16:11:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.222 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30803.0 tgs: 258 data_time: 0.63s time: 118.94s eta: 16:07:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31932.0 tgs: 268 data_time: 0.47s time: 118.86s eta: 16:04:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.63s time: 118.97s eta: 16:03:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31640.0 tgs: 259 data_time: 0.94s time: 121.85s eta: 16:24:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30871.0 tgs: 259 data_time: 0.83s time: 118.98s eta: 15:59:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.62s time: 119.22s eta: 15:59:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.209 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32203.0 tgs: 269 data_time: 0.80s time: 119.62s eta: 16:00:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:48:49][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31644.0 tgs: 265 data_time: 0.92s time: 119.04s eta: 15:54:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:50:49][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31471.0 tgs: 264 data_time: 0.85s time: 119.00s eta: 15:52:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32058.0 tgs: 269 data_time: 0.73s time: 119.15s eta: 15:51:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31276.0 tgs: 261 data_time: 0.55s time: 119.64s eta: 15:53:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 266 data_time: 0.78s time: 118.93s eta: 15:45:30 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.117 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31336.0 tgs: 262 data_time: 0.73s time: 119.29s eta: 15:46:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.81s time: 119.13s eta: 15:43:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32124.0 tgs: 269 data_time: 0.54s time: 119.00s eta: 15:40:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31894.0 tgs: 266 data_time: 1.11s time: 119.66s eta: 15:43:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30979.0 tgs: 254 data_time: 0.60s time: 121.53s eta: 15:56:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 1.00s time: 119.38s eta: 15:37:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30678.0 tgs: 236 data_time: 0.70s time: 129.96s eta: 16:58:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31688.0 tgs: 265 data_time: 0.84s time: 119.47s eta: 15:33:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32488.0 tgs: 271 data_time: 0.61s time: 119.60s eta: 15:32:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 265 data_time: 0.71s time: 119.53s eta: 15:30:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32353.0 tgs: 264 data_time: 0.82s time: 122.15s eta: 15:48:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31371.0 tgs: 259 data_time: 0.67s time: 120.74s eta: 15:35:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31162.0 tgs: 260 data_time: 0.80s time: 119.41s eta: 15:23:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 266 data_time: 0.60s time: 119.45s eta: 15:21:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.207 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31707.0 tgs: 240 data_time: 0.77s time: 131.64s eta: 16:53:36 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31151.0 tgs: 261 data_time: 0.62s time: 119.29s eta: 15:16:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31976.0 tgs: 267 data_time: 0.70s time: 119.53s eta: 15:16:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31284.0 tgs: 255 data_time: 0.80s time: 122.26s eta: 15:35:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.101 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31315.0 tgs: 261 data_time: 0.58s time: 119.72s eta: 15:13:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32406.0 tgs: 271 data_time: 0.79s time: 119.34s eta: 15:08:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32148.0 tgs: 270 data_time: 0.96s time: 118.90s eta: 15:03:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31696.0 tgs: 266 data_time: 0.68s time: 118.99s eta: 15:02:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.223 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31056.0 tgs: 260 data_time: 0.68s time: 119.30s eta: 15:02:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.123 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32069.0 tgs: 269 data_time: 1.01s time: 118.91s eta: 14:57:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30466.0 tgs: 252 data_time: 0.58s time: 120.52s eta: 15:07:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32374.0 tgs: 271 data_time: 0.77s time: 119.32s eta: 14:56:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.250 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31709.0 tgs: 258 data_time: 0.86s time: 122.66s eta: 15:19:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31008.0 tgs: 260 data_time: 0.77s time: 119.12s eta: 14:51:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30353.0 tgs: 253 data_time: 0.91s time: 119.75s eta: 14:54:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.174 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32000.0 tgs: 267 data_time: 0.73s time: 119.47s eta: 14:50:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.212 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31426.0 tgs: 264 data_time: 0.78s time: 118.98s eta: 14:44:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.66s time: 119.36s eta: 14:45:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.247 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31563.0 tgs: 265 data_time: 0.77s time: 118.97s eta: 14:40:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 270 data_time: 0.91s time: 119.03s eta: 14:38:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31431.0 tgs: 263 data_time: 0.97s time: 119.06s eta: 14:37:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31648.0 tgs: 265 data_time: 0.92s time: 119.13s eta: 14:35:35 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30483.0 tgs: 256 data_time: 0.65s time: 119.00s eta: 14:32:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31386.0 tgs: 263 data_time: 0.71s time: 119.07s eta: 14:31:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 268 data_time: 0.85s time: 119.14s eta: 14:29:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 268 data_time: 0.89s time: 119.65s eta: 14:31:27 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31083.0 tgs: 260 data_time: 0.66s time: 119.17s eta: 14:25:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.214 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31336.0 tgs: 262 data_time: 0.95s time: 119.20s eta: 14:24:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.121 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32214.0 tgs: 269 data_time: 0.83s time: 119.39s eta: 14:23:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32099.0 tgs: 265 data_time: 0.90s time: 120.74s eta: 14:31:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31602.0 tgs: 265 data_time: 0.80s time: 118.88s eta: 14:15:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31309.0 tgs: 262 data_time: 0.88s time: 119.10s eta: 14:15:31 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32032.0 tgs: 268 data_time: 0.59s time: 119.13s eta: 14:13:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31071.0 tgs: 259 data_time: 0.46s time: 119.95s eta: 14:17:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32041.0 tgs: 267 data_time: 1.22s time: 119.87s eta: 14:15:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.286 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32454.0 tgs: 271 data_time: 0.71s time: 119.58s eta: 14:11:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31188.0 tgs: 261 data_time: 0.74s time: 119.21s eta: 14:06:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31774.0 tgs: 266 data_time: 0.55s time: 119.27s eta: 14:04:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.221 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31335.0 tgs: 262 data_time: 0.76s time: 119.26s eta: 14:02:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30880.0 tgs: 259 data_time: 0.63s time: 119.14s eta: 13:59:57 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31463.0 tgs: 264 data_time: 0.69s time: 119.13s eta: 13:57:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 266 data_time: 0.78s time: 119.16s eta: 13:56:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31736.0 tgs: 265 data_time: 0.58s time: 119.34s eta: 13:55:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31332.0 tgs: 260 data_time: 0.51s time: 120.11s eta: 13:58:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.220 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31769.0 tgs: 259 data_time: 0.87s time: 122.56s eta: 14:13:50 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.108 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30464.0 tgs: 255 data_time: 0.54s time: 119.11s eta: 13:47:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31971.0 tgs: 267 data_time: 0.98s time: 119.36s eta: 13:47:33 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31019.0 tgs: 259 data_time: 0.87s time: 119.38s eta: 13:45:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31545.0 tgs: 264 data_time: 0.90s time: 119.32s eta: 13:43:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.124 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30210.0 tgs: 253 data_time: 0.74s time: 119.12s eta: 13:39:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.119 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31402.0 tgs: 263 data_time: 0.66s time: 119.26s eta: 13:38:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30979.0 tgs: 260 data_time: 0.77s time: 119.13s eta: 13:36:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 267 data_time: 0.64s time: 119.22s eta: 13:34:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31521.0 tgs: 263 data_time: 0.82s time: 119.46s eta: 13:34:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.228 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30756.0 tgs: 258 data_time: 1.03s time: 119.09s eta: 13:29:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.215 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30239.0 tgs: 254 data_time: 0.79s time: 118.92s eta: 13:26:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31540.0 tgs: 263 data_time: 0.62s time: 119.81s eta: 13:30:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32292.0 tgs: 270 data_time: 0.89s time: 119.25s eta: 13:24:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31612.0 tgs: 265 data_time: 0.79s time: 119.16s eta: 13:22:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32274.0 tgs: 270 data_time: 0.77s time: 119.35s eta: 13:21:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32015.0 tgs: 268 data_time: 0.52s time: 119.26s eta: 13:19:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 271 data_time: 0.74s time: 119.04s eta: 13:15:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.274 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32171.0 tgs: 269 data_time: 0.93s time: 119.37s eta: 13:15:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31963.0 tgs: 268 data_time: 0.89s time: 119.19s eta: 13:12:36 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.121 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31708.0 tgs: 265 data_time: 0.84s time: 119.46s eta: 13:12:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32162.0 tgs: 269 data_time: 0.75s time: 119.29s eta: 13:09:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31971.0 tgs: 268 data_time: 0.64s time: 119.14s eta: 13:06:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.211 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31170.0 tgs: 260 data_time: 0.49s time: 119.80s eta: 13:08:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.113 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32008.0 tgs: 266 data_time: 0.71s time: 120.16s eta: 13:09:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31146.0 tgs: 260 data_time: 0.75s time: 119.50s eta: 13:02:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.137 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31102.0 tgs: 260 data_time: 0.62s time: 119.50s eta: 13:00:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31646.0 tgs: 264 data_time: 0.81s time: 119.58s eta: 12:59:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31660.0 tgs: 264 data_time: 0.76s time: 119.80s eta: 12:58:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30304.0 tgs: 254 data_time: 0.48s time: 119.03s eta: 12:51:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32097.0 tgs: 269 data_time: 0.81s time: 119.05s eta: 12:49:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31600.0 tgs: 265 data_time: 0.79s time: 119.12s eta: 12:48:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31309.0 tgs: 262 data_time: 0.75s time: 119.40s eta: 12:48:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32114.0 tgs: 269 data_time: 0.92s time: 119.19s eta: 12:44:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31183.0 tgs: 261 data_time: 0.84s time: 119.46s eta: 12:44:33 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32329.0 tgs: 271 data_time: 0.61s time: 119.08s eta: 12:40:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 267 data_time: 0.51s time: 119.07s eta: 12:38:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31459.0 tgs: 263 data_time: 0.87s time: 119.59s eta: 12:39:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.214 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 270 data_time: 0.84s time: 118.78s eta: 12:32:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 269 data_time: 0.92s time: 118.88s eta: 12:30:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31597.0 tgs: 265 data_time: 0.79s time: 119.17s eta: 12:30:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.189 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 266 data_time: 0.64s time: 119.26s eta: 12:29:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31466.0 tgs: 263 data_time: 0.54s time: 119.34s eta: 12:27:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32399.0 tgs: 271 data_time: 0.57s time: 119.13s eta: 12:24:33 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32066.0 tgs: 268 data_time: 0.64s time: 119.34s eta: 12:23:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32477.0 tgs: 272 data_time: 0.77s time: 119.28s eta: 12:21:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32311.0 tgs: 271 data_time: 0.74s time: 119.00s eta: 12:17:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31939.0 tgs: 263 data_time: 1.09s time: 121.13s eta: 12:28:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31894.0 tgs: 265 data_time: 0.67s time: 119.98s eta: 12:19:52 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.189 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31769.0 tgs: 267 data_time: 0.76s time: 118.72s eta: 12:10:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 270 data_time: 0.68s time: 119.63s eta: 12:13:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31013.0 tgs: 258 data_time: 0.75s time: 119.83s eta: 12:12:58 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 264 data_time: 0.81s time: 119.14s eta: 12:06:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.260 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31858.0 tgs: 267 data_time: 0.92s time: 119.17s eta: 12:04:57 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31812.0 tgs: 267 data_time: 0.93s time: 119.12s eta: 12:02:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 269 data_time: 0.57s time: 119.21s eta: 12:01:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31793.0 tgs: 267 data_time: 0.66s time: 118.92s eta: 11:57:30 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31540.0 tgs: 263 data_time: 0.81s time: 119.72s eta: 12:00:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.128 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30544.0 tgs: 256 data_time: 0.81s time: 119.29s eta: 11:55:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.121 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 270 data_time: 0.59s time: 119.24s eta: 11:53:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32245.0 tgs: 269 data_time: 0.78s time: 119.71s eta: 11:54:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31701.0 tgs: 266 data_time: 0.87s time: 118.77s eta: 11:46:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31096.0 tgs: 260 data_time: 0.54s time: 119.16s eta: 11:47:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 269 data_time: 0.67s time: 119.60s eta: 11:47:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31969.0 tgs: 268 data_time: 0.69s time: 118.87s eta: 11:41:21 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30453.0 tgs: 256 data_time: 0.63s time: 118.69s eta: 11:38:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32281.0 tgs: 270 data_time: 1.10s time: 119.23s eta: 11:39:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31590.0 tgs: 265 data_time: 0.87s time: 119.12s eta: 11:36:50 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32210.0 tgs: 270 data_time: 0.60s time: 119.07s eta: 11:34:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.200 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31267.0 tgs: 262 data_time: 0.98s time: 118.94s eta: 11:31:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31652.0 tgs: 266 data_time: 0.63s time: 118.91s eta: 11:29:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31147.0 tgs: 261 data_time: 0.73s time: 118.97s eta: 11:28:03 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31720.0 tgs: 265 data_time: 0.77s time: 119.27s eta: 11:27:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.186 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31912.0 tgs: 267 data_time: 0.73s time: 119.36s eta: 11:26:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.202 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31545.0 tgs: 265 data_time: 0.90s time: 118.87s eta: 11:21:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31324.0 tgs: 262 data_time: 0.96s time: 119.20s eta: 11:21:27 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 264 data_time: 0.64s time: 119.72s eta: 11:22:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30615.0 tgs: 257 data_time: 0.69s time: 119.11s eta: 11:16:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31424.0 tgs: 263 data_time: 0.71s time: 119.12s eta: 11:15:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.112 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30803.0 tgs: 258 data_time: 0.79s time: 118.99s eta: 11:12:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31452.0 tgs: 264 data_time: 0.55s time: 119.09s eta: 11:10:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.122 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31625.0 tgs: 265 data_time: 0.60s time: 119.18s eta: 11:09:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31708.0 tgs: 266 data_time: 0.88s time: 118.95s eta: 11:06:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32086.0 tgs: 268 data_time: 0.79s time: 119.46s eta: 11:07:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31642.0 tgs: 264 data_time: 1.22s time: 119.64s eta: 11:06:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.195 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32189.0 tgs: 269 data_time: 0.94s time: 119.31s eta: 11:02:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.219 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 270 data_time: 0.98s time: 119.30s eta: 11:00:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31968.0 tgs: 269 data_time: 0.72s time: 118.82s eta: 10:55:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.186 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31910.0 tgs: 267 data_time: 0.58s time: 119.48s eta: 10:57:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.218 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32338.0 tgs: 270 data_time: 0.85s time: 119.40s eta: 10:54:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31889.0 tgs: 267 data_time: 0.86s time: 119.25s eta: 10:51:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31414.0 tgs: 263 data_time: 0.71s time: 119.25s eta: 10:49:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.191 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32270.0 tgs: 270 data_time: 0.80s time: 119.11s eta: 10:47:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31696.0 tgs: 265 data_time: 0.62s time: 119.52s eta: 10:47:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31870.0 tgs: 267 data_time: 0.66s time: 119.34s eta: 10:44:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31715.0 tgs: 265 data_time: 0.91s time: 119.48s eta: 10:43:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31985.0 tgs: 267 data_time: 0.77s time: 119.51s eta: 10:41:22 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31170.0 tgs: 261 data_time: 0.92s time: 118.99s eta: 10:36:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32105.0 tgs: 269 data_time: 0.57s time: 119.21s eta: 10:35:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31743.0 tgs: 266 data_time: 0.82s time: 119.05s eta: 10:32:58 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31441.0 tgs: 264 data_time: 0.83s time: 118.79s eta: 10:29:36 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31695.0 tgs: 264 data_time: 0.75s time: 119.85s eta: 10:33:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.278 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32418.0 tgs: 271 data_time: 0.57s time: 119.42s eta: 10:28:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.179 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31712.0 tgs: 266 data_time: 0.59s time: 118.81s eta: 10:23:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 268 data_time: 0.59s time: 118.98s eta: 10:22:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.112 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31111.0 tgs: 260 data_time: 0.76s time: 119.28s eta: 10:22:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31779.0 tgs: 266 data_time: 0.93s time: 119.20s eta: 10:19:50 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32217.0 tgs: 270 data_time: 0.62s time: 119.06s eta: 10:17:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 266 data_time: 0.87s time: 119.48s eta: 10:17:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.214 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31617.0 tgs: 264 data_time: 0.84s time: 119.42s eta: 10:15:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32303.0 tgs: 270 data_time: 0.65s time: 119.47s eta: 10:13:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.220 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30787.0 tgs: 258 data_time: 0.90s time: 119.32s eta: 10:10:31 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31276.0 tgs: 262 data_time: 0.74s time: 119.08s eta: 10:07:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31662.0 tgs: 264 data_time: 0.86s time: 119.71s eta: 10:08:30 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 267 data_time: 0.51s time: 119.83s eta: 10:07:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32345.0 tgs: 266 data_time: 0.56s time: 121.55s eta: 10:13:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.189 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32009.0 tgs: 268 data_time: 0.93s time: 119.26s eta: 10:00:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31374.0 tgs: 264 data_time: 0.66s time: 118.81s eta: 9:56:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.118 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30939.0 tgs: 258 data_time: 0.62s time: 119.55s eta: 9:57:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31869.0 tgs: 266 data_time: 0.88s time: 119.70s eta: 9:56:30 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30848.0 tgs: 258 data_time: 0.88s time: 119.26s eta: 9:52:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29950.0 tgs: 250 data_time: 0.87s time: 119.33s eta: 9:50:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 270 data_time: 0.71s time: 119.22s eta: 9:48:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31350.0 tgs: 260 data_time: 0.67s time: 120.13s eta: 9:50:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32595.0 tgs: 273 data_time: 0.58s time: 119.25s eta: 9:44:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31881.0 tgs: 266 data_time: 0.84s time: 119.48s eta: 9:43:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.130 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31319.0 tgs: 262 data_time: 0.79s time: 119.52s eta: 9:41:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31581.0 tgs: 264 data_time: 0.73s time: 119.20s eta: 9:38:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32229.0 tgs: 269 data_time: 0.69s time: 119.48s eta: 9:37:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.112 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31411.0 tgs: 263 data_time: 0.58s time: 119.05s eta: 9:33:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.57s time: 119.24s eta: 9:32:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31166.0 tgs: 261 data_time: 0.80s time: 119.16s eta: 9:30:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32416.0 tgs: 271 data_time: 0.56s time: 119.22s eta: 9:28:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.243 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31629.0 tgs: 264 data_time: 0.75s time: 119.46s eta: 9:27:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.166 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31889.0 tgs: 268 data_time: 0.80s time: 118.98s eta: 9:23:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.60s time: 119.03s eta: 9:21:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 27782.0 tgs: 232 data_time: 0.63s time: 119.53s eta: 9:21:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.118 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32021.0 tgs: 267 data_time: 0.65s time: 119.59s eta: 9:20:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32488.0 tgs: 272 data_time: 0.92s time: 119.12s eta: 9:15:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32369.0 tgs: 271 data_time: 0.64s time: 119.07s eta: 9:13:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31977.0 tgs: 268 data_time: 0.90s time: 118.90s eta: 9:10:52 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32476.0 tgs: 272 data_time: 0.76s time: 119.32s eta: 9:10:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31164.0 tgs: 261 data_time: 1.07s time: 119.09s eta: 9:07:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30880.0 tgs: 259 data_time: 0.56s time: 118.85s eta: 9:04:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31459.0 tgs: 262 data_time: 0.69s time: 119.67s eta: 9:06:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31285.0 tgs: 261 data_time: 0.56s time: 119.43s eta: 9:03:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.198 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31374.0 tgs: 262 data_time: 0.97s time: 119.44s eta: 9:01:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.275 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30986.0 tgs: 260 data_time: 0.94s time: 118.73s eta: 8:56:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 269 data_time: 0.71s time: 118.90s eta: 8:55:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31531.0 tgs: 264 data_time: 0.81s time: 119.30s eta: 8:54:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.118 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31753.0 tgs: 266 data_time: 0.69s time: 119.15s eta: 8:52:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31676.0 tgs: 265 data_time: 0.65s time: 119.27s eta: 8:50:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.106 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31715.0 tgs: 263 data_time: 0.64s time: 120.47s eta: 8:54:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 268 data_time: 0.61s time: 119.23s eta: 8:46:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.145 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32250.0 tgs: 270 data_time: 0.67s time: 119.10s eta: 8:44:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31521.0 tgs: 260 data_time: 0.69s time: 120.80s eta: 8:49:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.107 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31377.0 tgs: 261 data_time: 1.00s time: 120.20s eta: 8:44:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 270 data_time: 0.58s time: 118.83s eta: 8:36:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30387.0 tgs: 254 data_time: 0.63s time: 119.38s eta: 8:37:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31766.0 tgs: 266 data_time: 0.67s time: 119.29s eta: 8:34:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31829.0 tgs: 267 data_time: 0.85s time: 119.06s eta: 8:31:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 269 data_time: 0.90s time: 119.17s eta: 8:30:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.189 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29897.0 tgs: 251 data_time: 0.68s time: 119.05s eta: 8:27:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32036.0 tgs: 268 data_time: 0.86s time: 119.12s eta: 8:26:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32269.0 tgs: 270 data_time: 0.91s time: 119.24s eta: 8:24:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.187 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31112.0 tgs: 260 data_time: 0.77s time: 119.22s eta: 8:22:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.211 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31075.0 tgs: 260 data_time: 0.72s time: 119.38s eta: 8:21:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31844.0 tgs: 265 data_time: 0.72s time: 119.77s eta: 8:21:03 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31580.0 tgs: 264 data_time: 0.77s time: 119.20s eta: 8:16:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30329.0 tgs: 254 data_time: 0.67s time: 119.02s eta: 8:13:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 268 data_time: 0.70s time: 119.91s eta: 8:15:36 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 264 data_time: 0.62s time: 119.55s eta: 8:12:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.182 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 269 data_time: 0.66s time: 119.59s eta: 8:10:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.198 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31300.0 tgs: 263 data_time: 0.56s time: 119.00s eta: 8:05:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.101 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28806.0 tgs: 240 data_time: 0.62s time: 119.83s eta: 8:07:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32048.0 tgs: 267 data_time: 0.74s time: 119.70s eta: 8:04:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.61s time: 119.40s eta: 8:01:35 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.189 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31906.0 tgs: 267 data_time: 1.05s time: 119.29s eta: 7:59:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31428.0 tgs: 263 data_time: 0.54s time: 119.19s eta: 7:56:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.194 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31100.0 tgs: 260 data_time: 0.74s time: 119.19s eta: 7:54:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32186.0 tgs: 270 data_time: 0.68s time: 118.94s eta: 7:51:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31783.0 tgs: 267 data_time: 0.54s time: 118.97s eta: 7:49:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32208.0 tgs: 269 data_time: 0.94s time: 119.41s eta: 7:49:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32017.0 tgs: 268 data_time: 0.83s time: 119.03s eta: 7:46:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.80s time: 119.10s eta: 7:44:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 264 data_time: 0.73s time: 120.02s eta: 7:46:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.134 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31079.0 tgs: 259 data_time: 0.54s time: 119.83s eta: 7:43:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.114 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31313.0 tgs: 262 data_time: 0.66s time: 119.17s eta: 7:38:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32292.0 tgs: 271 data_time: 0.68s time: 118.96s eta: 7:36:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32111.0 tgs: 269 data_time: 0.72s time: 119.32s eta: 7:35:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 265 data_time: 0.78s time: 119.39s eta: 7:33:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31435.0 tgs: 263 data_time: 0.74s time: 119.07s eta: 7:30:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31965.0 tgs: 268 data_time: 0.79s time: 118.98s eta: 7:28:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32421.0 tgs: 271 data_time: 0.80s time: 119.23s eta: 7:27:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30436.0 tgs: 255 data_time: 0.66s time: 119.01s eta: 7:24:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31425.0 tgs: 264 data_time: 0.63s time: 118.91s eta: 7:21:57 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.192 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.77s time: 119.08s eta: 7:20:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30653.0 tgs: 257 data_time: 0.86s time: 118.99s eta: 7:18:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.187 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31585.0 tgs: 265 data_time: 0.72s time: 119.02s eta: 7:16:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32409.0 tgs: 271 data_time: 0.46s time: 119.27s eta: 7:15:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31615.0 tgs: 264 data_time: 0.53s time: 119.36s eta: 7:13:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.193 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31529.0 tgs: 265 data_time: 0.72s time: 118.94s eta: 7:10:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.236 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31443.0 tgs: 263 data_time: 0.77s time: 119.22s eta: 7:09:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.180 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 267 data_time: 0.58s time: 119.56s eta: 7:08:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31726.0 tgs: 265 data_time: 0.60s time: 119.67s eta: 7:06:50 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31531.0 tgs: 265 data_time: 0.76s time: 118.96s eta: 7:02:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.225 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31871.0 tgs: 267 data_time: 0.66s time: 119.29s eta: 7:01:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:46:30][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32550.0 tgs: 272 data_time: 0.84s time: 119.55s eta: 7:00:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 266 data_time: 0.88s time: 119.41s eta: 6:57:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31487.0 tgs: 264 data_time: 0.77s time: 118.91s eta: 6:54:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.190 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31543.0 tgs: 264 data_time: 0.88s time: 119.32s eta: 6:53:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31299.0 tgs: 263 data_time: 0.93s time: 118.93s eta: 6:50:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31011.0 tgs: 260 data_time: 0.74s time: 119.26s eta: 6:49:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31985.0 tgs: 268 data_time: 0.43s time: 119.02s eta: 6:46:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32473.0 tgs: 272 data_time: 0.94s time: 119.23s eta: 6:45:22 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.215 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32216.0 tgs: 269 data_time: 0.80s time: 119.41s eta: 6:44:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31796.0 tgs: 267 data_time: 0.84s time: 119.02s eta: 6:40:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32225.0 tgs: 270 data_time: 0.90s time: 119.03s eta: 6:38:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31563.0 tgs: 264 data_time: 0.54s time: 119.45s eta: 6:38:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.118 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31765.0 tgs: 267 data_time: 0.80s time: 118.78s eta: 6:33:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.189 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30831.0 tgs: 258 data_time: 0.90s time: 119.45s eta: 6:34:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.250 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 265 data_time: 0.74s time: 119.57s eta: 6:32:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31566.0 tgs: 264 data_time: 0.70s time: 119.37s eta: 6:29:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31344.0 tgs: 262 data_time: 0.67s time: 119.34s eta: 6:27:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31620.0 tgs: 264 data_time: 0.66s time: 119.36s eta: 6:25:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.212 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30792.0 tgs: 259 data_time: 0.82s time: 118.82s eta: 6:22:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 264 data_time: 0.65s time: 119.02s eta: 6:20:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32178.0 tgs: 269 data_time: 0.64s time: 119.40s eta: 6:20:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32053.0 tgs: 268 data_time: 0.93s time: 119.34s eta: 6:17:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.116 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30904.0 tgs: 259 data_time: 0.55s time: 119.06s eta: 6:15:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31883.0 tgs: 267 data_time: 0.69s time: 119.40s eta: 6:14:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31180.0 tgs: 261 data_time: 0.76s time: 119.38s eta: 6:12:04 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 269 data_time: 0.77s time: 119.30s eta: 6:09:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32318.0 tgs: 271 data_time: 0.52s time: 119.10s eta: 6:07:13 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31950.0 tgs: 262 data_time: 0.82s time: 121.94s eta: 6:13:57 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.118 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31717.0 tgs: 265 data_time: 0.88s time: 119.38s eta: 6:04:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.243 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31253.0 tgs: 261 data_time: 0.97s time: 119.58s eta: 6:02:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 270 data_time: 0.73s time: 119.25s eta: 5:59:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.264 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32087.0 tgs: 269 data_time: 0.99s time: 119.26s eta: 5:57:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.276 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32065.0 tgs: 269 data_time: 0.97s time: 118.93s eta: 5:54:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.113 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31503.0 tgs: 264 data_time: 0.69s time: 119.02s eta: 5:53:05 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30980.0 tgs: 260 data_time: 0.74s time: 119.05s eta: 5:51:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30942.0 tgs: 259 data_time: 0.64s time: 119.30s eta: 5:49:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30775.0 tgs: 258 data_time: 0.79s time: 118.94s eta: 5:46:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:00:05][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31797.0 tgs: 266 data_time: 0.94s time: 119.29s eta: 5:45:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 266 data_time: 0.67s time: 119.37s eta: 5:44:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32075.0 tgs: 269 data_time: 0.67s time: 119.06s eta: 5:41:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30389.0 tgs: 254 data_time: 0.50s time: 119.50s eta: 5:40:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31877.0 tgs: 267 data_time: 0.70s time: 119.38s eta: 5:38:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31631.0 tgs: 264 data_time: 0.63s time: 119.56s eta: 5:36:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.194 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31954.0 tgs: 268 data_time: 0.65s time: 118.99s eta: 5:33:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31196.0 tgs: 262 data_time: 0.68s time: 118.94s eta: 5:31:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.161 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30845.0 tgs: 258 data_time: 0.74s time: 119.27s eta: 5:29:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32436.0 tgs: 271 data_time: 1.02s time: 119.37s eta: 5:28:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31170.0 tgs: 260 data_time: 0.94s time: 119.46s eta: 5:26:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31893.0 tgs: 268 data_time: 0.62s time: 118.94s eta: 5:23:07 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.181 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 269 data_time: 0.74s time: 119.63s eta: 5:23:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31382.0 tgs: 262 data_time: 0.65s time: 119.34s eta: 5:20:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32139.0 tgs: 269 data_time: 0.81s time: 119.13s eta: 5:17:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30652.0 tgs: 256 data_time: 0.68s time: 119.41s eta: 5:16:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32479.0 tgs: 271 data_time: 0.53s time: 119.58s eta: 5:14:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.177 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32167.0 tgs: 270 data_time: 0.66s time: 119.00s eta: 5:11:22 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31794.0 tgs: 267 data_time: 0.81s time: 119.00s eta: 5:09:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32349.0 tgs: 271 data_time: 0.72s time: 119.22s eta: 5:07:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.227 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30567.0 tgs: 255 data_time: 0.75s time: 119.69s eta: 5:07:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32227.0 tgs: 270 data_time: 0.84s time: 119.11s eta: 5:03:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 270 data_time: 0.95s time: 119.07s eta: 5:01:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32574.0 tgs: 268 data_time: 0.67s time: 121.51s eta: 5:05:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31699.0 tgs: 265 data_time: 0.72s time: 119.58s eta: 4:58:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32147.0 tgs: 269 data_time: 0.63s time: 119.50s eta: 4:56:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32298.0 tgs: 269 data_time: 0.56s time: 119.73s eta: 4:55:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32168.0 tgs: 269 data_time: 0.91s time: 119.52s eta: 4:52:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.175 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30415.0 tgs: 254 data_time: 0.86s time: 119.42s eta: 4:50:35 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.126 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31221.0 tgs: 261 data_time: 0.87s time: 119.23s eta: 4:48:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.101 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32172.0 tgs: 270 data_time: 0.53s time: 118.89s eta: 4:45:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.267 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31588.0 tgs: 265 data_time: 0.88s time: 119.09s eta: 4:43:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 26989.0 tgs: 224 data_time: 0.81s time: 119.98s eta: 4:43:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31693.0 tgs: 266 data_time: 0.69s time: 119.15s eta: 4:39:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32133.0 tgs: 269 data_time: 0.52s time: 119.02s eta: 4:37:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30663.0 tgs: 257 data_time: 0.67s time: 119.05s eta: 4:35:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.126 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 268 data_time: 0.76s time: 119.37s eta: 4:34:32 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.266 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31558.0 tgs: 264 data_time: 0.65s time: 119.41s eta: 4:32:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31203.0 tgs: 262 data_time: 0.93s time: 118.93s eta: 4:29:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32454.0 tgs: 271 data_time: 0.77s time: 119.62s eta: 4:29:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 270 data_time: 0.84s time: 119.21s eta: 4:26:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31898.0 tgs: 267 data_time: 0.75s time: 119.07s eta: 4:23:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29983.0 tgs: 252 data_time: 0.75s time: 118.84s eta: 4:21:27 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32306.0 tgs: 272 data_time: 0.89s time: 118.76s eta: 4:19:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30630.0 tgs: 256 data_time: 0.48s time: 119.43s eta: 4:18:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.193 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32396.0 tgs: 271 data_time: 0.74s time: 119.40s eta: 4:16:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.193 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32099.0 tgs: 268 data_time: 0.64s time: 119.35s eta: 4:14:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30656.0 tgs: 257 data_time: 0.79s time: 119.05s eta: 4:11:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.224 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31802.0 tgs: 266 data_time: 0.78s time: 119.12s eta: 4:10:09 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31424.0 tgs: 263 data_time: 0.85s time: 119.13s eta: 4:08:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.229 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31592.0 tgs: 265 data_time: 0.60s time: 118.95s eta: 4:05:50 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29607.0 tgs: 248 data_time: 0.75s time: 119.15s eta: 4:04:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31847.0 tgs: 266 data_time: 0.65s time: 119.50s eta: 4:02:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31185.0 tgs: 261 data_time: 0.73s time: 119.24s eta: 4:00:27 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32303.0 tgs: 271 data_time: 1.07s time: 119.19s eta: 3:58:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31655.0 tgs: 265 data_time: 0.85s time: 119.36s eta: 3:56:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31893.0 tgs: 267 data_time: 0.74s time: 119.37s eta: 3:54:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32035.0 tgs: 267 data_time: 0.71s time: 119.95s eta: 3:53:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 270 data_time: 0.91s time: 119.33s eta: 3:50:41 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32258.0 tgs: 269 data_time: 0.69s time: 119.63s eta: 3:49:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31736.0 tgs: 267 data_time: 0.92s time: 118.85s eta: 3:45:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32385.0 tgs: 270 data_time: 0.74s time: 119.76s eta: 3:45:33 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31591.0 tgs: 264 data_time: 0.79s time: 119.33s eta: 3:42:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32056.0 tgs: 268 data_time: 0.66s time: 119.25s eta: 3:40:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 265 data_time: 0.66s time: 118.98s eta: 3:38:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31737.0 tgs: 265 data_time: 0.64s time: 119.47s eta: 3:37:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32205.0 tgs: 270 data_time: 0.72s time: 119.00s eta: 3:34:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.110 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31281.0 tgs: 262 data_time: 0.71s time: 119.01s eta: 3:32:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 267 data_time: 0.68s time: 119.20s eta: 3:30:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31848.0 tgs: 265 data_time: 0.64s time: 120.12s eta: 3:30:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.215 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32136.0 tgs: 269 data_time: 1.02s time: 119.12s eta: 3:26:28 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31086.0 tgs: 260 data_time: 0.71s time: 119.32s eta: 3:24:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.198 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31156.0 tgs: 261 data_time: 0.74s time: 119.07s eta: 3:22:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31024.0 tgs: 259 data_time: 1.05s time: 119.43s eta: 3:21:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31123.0 tgs: 260 data_time: 1.07s time: 119.28s eta: 3:18:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31449.0 tgs: 263 data_time: 0.74s time: 119.22s eta: 3:16:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32068.0 tgs: 269 data_time: 0.88s time: 119.15s eta: 3:14:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31033.0 tgs: 260 data_time: 0.69s time: 119.02s eta: 3:12:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32064.0 tgs: 268 data_time: 0.84s time: 119.21s eta: 3:10:44 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31295.0 tgs: 260 data_time: 0.79s time: 120.15s eta: 3:10:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31920.0 tgs: 267 data_time: 0.67s time: 119.50s eta: 3:07:12 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.102 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 266 data_time: 0.63s time: 119.26s eta: 3:04:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31950.0 tgs: 268 data_time: 0.61s time: 118.99s eta: 3:02:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31800.0 tgs: 266 data_time: 1.00s time: 119.22s eta: 3:00:48 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31771.0 tgs: 265 data_time: 0.81s time: 119.46s eta: 2:59:11 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32468.0 tgs: 272 data_time: 0.84s time: 118.98s eta: 2:56:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30769.0 tgs: 257 data_time: 0.83s time: 119.35s eta: 2:55:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32333.0 tgs: 270 data_time: 0.80s time: 119.41s eta: 2:53:08 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32040.0 tgs: 268 data_time: 0.92s time: 119.38s eta: 2:51:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31822.0 tgs: 262 data_time: 0.61s time: 121.45s eta: 2:52:02 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32253.0 tgs: 270 data_time: 0.99s time: 119.22s eta: 2:46:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32500.0 tgs: 271 data_time: 0.66s time: 119.60s eta: 2:45:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31322.0 tgs: 263 data_time: 0.84s time: 118.92s eta: 2:42:31 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32042.0 tgs: 268 data_time: 0.84s time: 119.18s eta: 2:40:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30370.0 tgs: 255 data_time: 0.59s time: 118.97s eta: 2:38:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32011.0 tgs: 268 data_time: 0.95s time: 119.23s eta: 2:36:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31095.0 tgs: 261 data_time: 0.78s time: 119.10s eta: 2:34:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31893.0 tgs: 267 data_time: 0.61s time: 119.43s eta: 2:33:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31005.0 tgs: 257 data_time: 0.53s time: 120.25s eta: 2:32:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32038.0 tgs: 268 data_time: 0.93s time: 119.25s eta: 2:29:03 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30817.0 tgs: 259 data_time: 0.96s time: 118.88s eta: 2:26:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30782.0 tgs: 257 data_time: 0.67s time: 119.51s eta: 2:25:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31877.0 tgs: 267 data_time: 0.80s time: 119.09s eta: 2:22:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.202 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31772.0 tgs: 266 data_time: 0.48s time: 119.13s eta: 2:20:58 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32044.0 tgs: 269 data_time: 0.82s time: 119.07s eta: 2:18:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 264 data_time: 0.80s time: 119.82s eta: 2:17:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30998.0 tgs: 259 data_time: 0.60s time: 119.34s eta: 2:15:14 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31692.0 tgs: 265 data_time: 0.75s time: 119.39s eta: 2:13:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30177.0 tgs: 252 data_time: 0.78s time: 119.67s eta: 2:11:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32035.0 tgs: 268 data_time: 0.75s time: 119.49s eta: 2:09:27 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31583.0 tgs: 265 data_time: 0.65s time: 118.96s eta: 2:06:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30341.0 tgs: 254 data_time: 0.80s time: 119.37s eta: 2:05:20 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 266 data_time: 0.62s time: 119.30s eta: 2:03:16 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31326.0 tgs: 262 data_time: 0.60s time: 119.19s eta: 2:01:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31955.0 tgs: 267 data_time: 0.80s time: 119.51s eta: 1:59:30 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31838.0 tgs: 266 data_time: 1.31s time: 119.42s eta: 1:57:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32512.0 tgs: 271 data_time: 0.95s time: 119.56s eta: 1:55:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.215 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32043.0 tgs: 268 data_time: 0.61s time: 119.28s eta: 1:53:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31402.0 tgs: 263 data_time: 0.70s time: 119.24s eta: 1:51:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 264 data_time: 0.79s time: 118.93s eta: 1:49:01 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31011.0 tgs: 259 data_time: 0.61s time: 119.65s eta: 1:47:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31568.0 tgs: 263 data_time: 0.62s time: 119.73s eta: 1:45:45 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32268.0 tgs: 270 data_time: 0.82s time: 119.35s eta: 1:43:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31370.0 tgs: 262 data_time: 0.56s time: 119.30s eta: 1:41:24 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31550.0 tgs: 264 data_time: 0.58s time: 119.28s eta: 1:39:23 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30490.0 tgs: 256 data_time: 0.62s time: 118.90s eta: 1:37:06 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.195 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31994.0 tgs: 268 data_time: 0.75s time: 119.16s eta: 1:35:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31848.0 tgs: 267 data_time: 0.80s time: 119.06s eta: 1:33:15 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31235.0 tgs: 261 data_time: 0.92s time: 119.27s eta: 1:31:26 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32418.0 tgs: 270 data_time: 0.57s time: 119.71s eta: 1:29:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31213.0 tgs: 262 data_time: 0.79s time: 119.04s eta: 1:27:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31577.0 tgs: 265 data_time: 0.75s time: 118.85s eta: 1:25:10 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30653.0 tgs: 256 data_time: 0.52s time: 119.47s eta: 1:23:37 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30599.0 tgs: 256 data_time: 0.58s time: 119.48s eta: 1:21:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29956.0 tgs: 251 data_time: 0.44s time: 118.94s eta: 1:19:17 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31890.0 tgs: 268 data_time: 1.04s time: 118.93s eta: 1:17:18 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31661.0 tgs: 259 data_time: 0.79s time: 122.16s eta: 1:17:22 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30022.0 tgs: 251 data_time: 0.78s time: 119.36s eta: 1:13:36 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31854.0 tgs: 267 data_time: 0.79s time: 119.05s eta: 1:11:25 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31980.0 tgs: 268 data_time: 0.70s time: 118.93s eta: 1:09:22 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31418.0 tgs: 263 data_time: 0.76s time: 119.10s eta: 1:07:29 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31785.0 tgs: 266 data_time: 0.58s time: 119.41s eta: 1:05:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30935.0 tgs: 259 data_time: 0.73s time: 119.34s eta: 1:03:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 269 data_time: 0.66s time: 119.18s eta: 1:01:34 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30987.0 tgs: 259 data_time: 0.89s time: 119.30s eta: 0:59:39 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32080.0 tgs: 268 data_time: 0.76s time: 119.41s eta: 0:57:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31912.0 tgs: 267 data_time: 0.66s time: 119.39s eta: 0:55:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32124.0 tgs: 268 data_time: 0.80s time: 119.49s eta: 0:53:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 270 data_time: 0.67s time: 119.36s eta: 0:51:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30896.0 tgs: 258 data_time: 0.74s time: 119.46s eta: 0:49:46 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31057.0 tgs: 261 data_time: 0.76s time: 118.89s eta: 0:47:33 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31820.0 tgs: 266 data_time: 0.67s time: 119.29s eta: 0:45:43 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32370.0 tgs: 271 data_time: 0.86s time: 119.22s eta: 0:43:42 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30898.0 tgs: 259 data_time: 0.62s time: 118.98s eta: 0:41:38 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32489.0 tgs: 272 data_time: 0.76s time: 119.04s eta: 0:39:40 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.196 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32249.0 tgs: 269 data_time: 0.81s time: 119.66s eta: 0:37:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.76s time: 119.28s eta: 0:35:47 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.100 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30802.0 tgs: 257 data_time: 0.65s time: 119.75s eta: 0:33:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 258 data_time: 0.57s time: 121.22s eta: 0:32:19 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32054.0 tgs: 268 data_time: 0.88s time: 119.41s eta: 0:29:51 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31359.0 tgs: 262 data_time: 0.81s time: 119.29s eta: 0:27:50 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 269 data_time: 0.63s time: 119.63s eta: 0:25:55 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31417.0 tgs: 263 data_time: 0.64s time: 119.09s eta: 0:23:49 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31237.0 tgs: 261 data_time: 0.97s time: 119.38s eta: 0:21:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32584.0 tgs: 273 data_time: 0.68s time: 119.26s eta: 0:19:52 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31509.0 tgs: 264 data_time: 0.91s time: 119.24s eta: 0:17:53 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32427.0 tgs: 270 data_time: 0.80s time: 119.91s eta: 0:15:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31335.0 tgs: 262 data_time: 0.81s time: 119.22s eta: 0:13:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31523.0 tgs: 264 data_time: 0.70s time: 119.09s eta: 0:11:54 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32060.0 tgs: 268 data_time: 0.93s time: 119.35s eta: 0:09:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.102 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31495.0 tgs: 264 data_time: 0.53s time: 119.16s eta: 0:07:56 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32049.0 tgs: 267 data_time: 0.56s time: 119.67s eta: 0:05:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.098 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30953.0 tgs: 257 data_time: 0.57s time: 120.04s eta: 0:04:00 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32379.0 tgs: 271 data_time: 0.65s time: 119.21s eta: 0:01:59 +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.26 seconds, peak gpu memory 17.4G +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 0][DP 0][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank10.log b/20250121165312/rank10.log new file mode 100644 index 0000000000000000000000000000000000000000..cd929b374a2035a48184c2ca0963bddf7365dad5 --- /dev/null +++ b/20250121165312/rank10.log @@ -0,0 +1,851 @@ +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.08s +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:52][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:52][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:52][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:00:52][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 62.80 seconds, peak gpu memory 13.4G +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.198 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31696.0 tgs: 207 data_time: 2.03s time: 152.61s eta: 1 day, 1:08:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.233 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31732.0 tgs: 264 data_time: 1.10s time: 120.13s eta: 19:45:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.222 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32529.0 tgs: 271 data_time: 0.94s time: 119.68s eta: 19:38:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31543.0 tgs: 263 data_time: 0.89s time: 119.65s eta: 19:36:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.205 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 265 data_time: 1.00s time: 119.31s eta: 19:31:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30335.0 tgs: 253 data_time: 0.95s time: 119.54s eta: 19:31:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30779.0 tgs: 250 data_time: 0.74s time: 122.97s eta: 20:03:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.231 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31645.0 tgs: 263 data_time: 0.98s time: 119.89s eta: 19:30:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.273 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32125.0 tgs: 266 data_time: 0.74s time: 120.65s eta: 19:36:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31629.0 tgs: 263 data_time: 1.01s time: 119.92s eta: 19:27:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.247 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31480.0 tgs: 261 data_time: 0.98s time: 120.46s eta: 19:30:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.187 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32172.0 tgs: 269 data_time: 0.77s time: 119.34s eta: 19:17:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.220 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 261 data_time: 1.31s time: 119.68s eta: 19:18:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32182.0 tgs: 269 data_time: 0.69s time: 119.20s eta: 19:12:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31862.0 tgs: 267 data_time: 0.80s time: 119.22s eta: 19:10:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31743.0 tgs: 266 data_time: 0.96s time: 118.93s eta: 19:05:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.271 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32304.0 tgs: 271 data_time: 0.96s time: 119.14s eta: 19:05:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 267 data_time: 0.98s time: 119.18s eta: 19:04:07 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.250 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30695.0 tgs: 257 data_time: 0.92s time: 119.40s eta: 19:04:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30721.0 tgs: 257 data_time: 1.12s time: 119.33s eta: 19:01:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29162.0 tgs: 242 data_time: 0.63s time: 120.32s eta: 19:09:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.263 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31808.0 tgs: 258 data_time: 0.87s time: 122.91s eta: 19:31:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32435.0 tgs: 271 data_time: 1.13s time: 119.48s eta: 18:57:03 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32461.0 tgs: 272 data_time: 0.87s time: 119.15s eta: 18:51:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31514.0 tgs: 264 data_time: 1.05s time: 119.25s eta: 18:50:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30176.0 tgs: 252 data_time: 0.65s time: 119.39s eta: 18:50:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32145.0 tgs: 268 data_time: 1.08s time: 119.66s eta: 18:50:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32029.0 tgs: 268 data_time: 0.64s time: 119.38s eta: 18:46:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31383.0 tgs: 261 data_time: 0.68s time: 119.79s eta: 18:48:03 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30895.0 tgs: 238 data_time: 0.82s time: 129.31s eta: 20:15:31 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 264 data_time: 1.13s time: 119.57s eta: 18:41:57 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32406.0 tgs: 271 data_time: 0.90s time: 119.39s eta: 18:38:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32322.0 tgs: 267 data_time: 0.89s time: 120.66s eta: 18:48:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.233 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30969.0 tgs: 259 data_time: 0.88s time: 119.39s eta: 18:34:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.241 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31334.0 tgs: 262 data_time: 0.99s time: 119.16s eta: 18:30:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 268 data_time: 0.92s time: 119.07s eta: 18:27:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 269 data_time: 0.95s time: 119.14s eta: 18:25:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 264 data_time: 0.91s time: 120.38s eta: 18:35:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31386.0 tgs: 259 data_time: 0.89s time: 121.17s eta: 18:40:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32250.0 tgs: 267 data_time: 0.74s time: 120.66s eta: 18:34:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32507.0 tgs: 271 data_time: 1.17s time: 119.73s eta: 18:23:30 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31757.0 tgs: 256 data_time: 0.75s time: 123.87s eta: 18:59:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 1.04s time: 119.56s eta: 18:17:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.321 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31891.0 tgs: 267 data_time: 0.98s time: 119.36s eta: 18:14:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31410.0 tgs: 263 data_time: 0.77s time: 119.23s eta: 18:10:57 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 265 data_time: 1.02s time: 119.60s eta: 18:12:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30226.0 tgs: 253 data_time: 0.95s time: 119.27s eta: 18:07:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31914.0 tgs: 268 data_time: 0.92s time: 119.01s eta: 18:03:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31209.0 tgs: 262 data_time: 0.54s time: 119.10s eta: 18:01:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 265 data_time: 0.75s time: 119.26s eta: 18:01:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31541.0 tgs: 264 data_time: 0.82s time: 119.33s eta: 17:59:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32580.0 tgs: 271 data_time: 0.88s time: 120.02s eta: 18:04:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31793.0 tgs: 266 data_time: 0.65s time: 119.10s eta: 17:53:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.125 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31726.0 tgs: 265 data_time: 0.71s time: 119.41s eta: 17:54:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32291.0 tgs: 271 data_time: 1.07s time: 118.99s eta: 17:48:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31197.0 tgs: 262 data_time: 0.83s time: 119.02s eta: 17:47:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31650.0 tgs: 252 data_time: 1.02s time: 125.37s eta: 18:42:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.257 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31693.0 tgs: 263 data_time: 0.87s time: 120.10s eta: 17:52:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.118 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30935.0 tgs: 259 data_time: 0.90s time: 119.37s eta: 17:44:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31373.0 tgs: 262 data_time: 0.83s time: 119.34s eta: 17:42:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31466.0 tgs: 263 data_time: 0.77s time: 119.51s eta: 17:41:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31582.0 tgs: 250 data_time: 0.84s time: 126.05s eta: 18:37:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31379.0 tgs: 255 data_time: 0.89s time: 122.63s eta: 18:05:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 262 data_time: 0.55s time: 121.15s eta: 17:50:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.244 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31863.0 tgs: 267 data_time: 0.68s time: 119.03s eta: 17:29:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31055.0 tgs: 261 data_time: 0.73s time: 118.85s eta: 17:25:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31953.0 tgs: 257 data_time: 0.78s time: 124.00s eta: 18:09:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32157.0 tgs: 258 data_time: 0.65s time: 124.56s eta: 18:12:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31301.0 tgs: 262 data_time: 0.84s time: 119.39s eta: 17:24:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.243 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31423.0 tgs: 262 data_time: 0.50s time: 119.70s eta: 17:25:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 260 data_time: 0.58s time: 119.52s eta: 17:21:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31690.0 tgs: 266 data_time: 0.89s time: 118.86s eta: 17:14:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31998.0 tgs: 266 data_time: 1.08s time: 119.85s eta: 17:20:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31852.0 tgs: 266 data_time: 0.74s time: 119.56s eta: 17:16:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32473.0 tgs: 266 data_time: 0.96s time: 121.87s eta: 17:34:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32194.0 tgs: 260 data_time: 0.93s time: 123.51s eta: 17:46:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31340.0 tgs: 263 data_time: 0.73s time: 119.01s eta: 17:05:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31330.0 tgs: 262 data_time: 1.05s time: 119.22s eta: 17:05:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30764.0 tgs: 258 data_time: 0.93s time: 119.14s eta: 17:02:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30678.0 tgs: 257 data_time: 0.76s time: 119.25s eta: 17:01:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32234.0 tgs: 270 data_time: 1.06s time: 119.04s eta: 16:57:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31845.0 tgs: 266 data_time: 0.68s time: 119.28s eta: 16:57:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31363.0 tgs: 263 data_time: 0.90s time: 119.16s eta: 16:54:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.293 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31986.0 tgs: 267 data_time: 1.13s time: 119.59s eta: 16:56:29 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32421.0 tgs: 272 data_time: 0.70s time: 119.13s eta: 16:50:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31499.0 tgs: 265 data_time: 0.56s time: 118.84s eta: 16:46:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.209 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30877.0 tgs: 259 data_time: 0.80s time: 119.02s eta: 16:45:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32268.0 tgs: 269 data_time: 0.71s time: 119.85s eta: 16:50:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31891.0 tgs: 267 data_time: 1.06s time: 119.39s eta: 16:44:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 266 data_time: 0.82s time: 119.01s eta: 16:39:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 260 data_time: 0.71s time: 120.93s eta: 16:53:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31771.0 tgs: 266 data_time: 0.92s time: 119.00s eta: 16:35:35 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.215 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31833.0 tgs: 266 data_time: 0.80s time: 119.28s eta: 16:36:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31534.0 tgs: 264 data_time: 0.97s time: 119.23s eta: 16:33:35 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 265 data_time: 0.86s time: 119.36s eta: 16:32:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 268 data_time: 0.88s time: 119.70s eta: 16:33:32 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32092.0 tgs: 269 data_time: 0.77s time: 119.20s eta: 16:27:22 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.120 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31700.0 tgs: 264 data_time: 0.76s time: 119.76s eta: 16:30:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29764.0 tgs: 247 data_time: 0.85s time: 120.40s eta: 16:33:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32195.0 tgs: 270 data_time: 0.56s time: 118.97s eta: 16:19:31 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31755.0 tgs: 249 data_time: 0.88s time: 127.33s eta: 17:26:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 264 data_time: 0.61s time: 119.22s eta: 16:17:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31210.0 tgs: 261 data_time: 0.64s time: 119.15s eta: 16:15:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32019.0 tgs: 268 data_time: 0.91s time: 119.25s eta: 16:13:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 1.03s time: 119.17s eta: 16:11:13 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31536.0 tgs: 265 data_time: 0.69s time: 118.95s eta: 16:07:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.233 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 270 data_time: 0.83s time: 118.87s eta: 16:04:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.221 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31877.0 tgs: 267 data_time: 0.85s time: 118.96s eta: 16:03:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31758.0 tgs: 260 data_time: 0.79s time: 121.84s eta: 16:24:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30694.0 tgs: 257 data_time: 0.93s time: 118.99s eta: 15:59:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31379.0 tgs: 263 data_time: 0.71s time: 119.23s eta: 15:59:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31729.0 tgs: 265 data_time: 0.57s time: 119.61s eta: 16:00:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.219 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30787.0 tgs: 258 data_time: 0.88s time: 119.04s eta: 15:54:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31745.0 tgs: 266 data_time: 0.82s time: 118.98s eta: 15:51:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31728.0 tgs: 266 data_time: 0.72s time: 119.17s eta: 15:51:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32273.0 tgs: 269 data_time: 1.12s time: 119.64s eta: 15:53:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 271 data_time: 0.87s time: 118.94s eta: 15:45:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 266 data_time: 0.91s time: 119.28s eta: 15:46:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 270 data_time: 0.75s time: 119.13s eta: 15:43:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 269 data_time: 0.67s time: 119.00s eta: 15:40:07 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.220 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 267 data_time: 1.02s time: 119.65s eta: 15:43:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31505.0 tgs: 259 data_time: 0.86s time: 121.50s eta: 15:55:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30426.0 tgs: 254 data_time: 0.78s time: 119.41s eta: 15:37:23 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31790.0 tgs: 244 data_time: 0.90s time: 129.96s eta: 16:58:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30635.0 tgs: 256 data_time: 0.96s time: 119.46s eta: 15:33:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.128 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31503.0 tgs: 263 data_time: 0.83s time: 119.60s eta: 15:32:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32379.0 tgs: 270 data_time: 0.95s time: 119.53s eta: 15:30:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30778.0 tgs: 251 data_time: 1.04s time: 122.14s eta: 15:48:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 263 data_time: 0.99s time: 120.75s eta: 15:35:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 264 data_time: 1.01s time: 119.41s eta: 15:23:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 268 data_time: 0.78s time: 119.44s eta: 15:21:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30189.0 tgs: 229 data_time: 0.97s time: 131.63s eta: 16:53:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30681.0 tgs: 257 data_time: 0.99s time: 119.29s eta: 15:16:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 267 data_time: 0.93s time: 119.56s eta: 15:16:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31504.0 tgs: 257 data_time: 0.99s time: 122.25s eta: 15:35:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 265 data_time: 0.64s time: 119.71s eta: 15:13:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31920.0 tgs: 267 data_time: 0.72s time: 119.35s eta: 15:09:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30995.0 tgs: 260 data_time: 0.98s time: 118.90s eta: 15:03:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32130.0 tgs: 270 data_time: 0.64s time: 118.99s eta: 15:02:22 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32369.0 tgs: 271 data_time: 0.52s time: 119.31s eta: 15:02:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31687.0 tgs: 266 data_time: 0.77s time: 118.92s eta: 14:57:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30954.0 tgs: 256 data_time: 0.61s time: 120.52s eta: 15:07:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31419.0 tgs: 263 data_time: 0.87s time: 119.31s eta: 14:56:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.197 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29745.0 tgs: 242 data_time: 0.85s time: 122.66s eta: 15:19:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 264 data_time: 0.80s time: 119.12s eta: 14:51:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.221 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30525.0 tgs: 254 data_time: 0.97s time: 119.76s eta: 14:54:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31844.0 tgs: 266 data_time: 0.76s time: 119.46s eta: 14:50:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.122 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31736.0 tgs: 266 data_time: 0.66s time: 118.99s eta: 14:44:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31460.0 tgs: 263 data_time: 0.72s time: 119.32s eta: 14:44:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31348.0 tgs: 263 data_time: 0.82s time: 118.94s eta: 14:40:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31771.0 tgs: 266 data_time: 1.20s time: 119.05s eta: 14:38:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31988.0 tgs: 268 data_time: 0.83s time: 119.06s eta: 14:37:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31563.0 tgs: 264 data_time: 0.60s time: 119.13s eta: 14:35:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30801.0 tgs: 258 data_time: 0.87s time: 119.00s eta: 14:32:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31800.0 tgs: 267 data_time: 0.93s time: 119.07s eta: 14:31:13 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31863.0 tgs: 267 data_time: 0.72s time: 119.15s eta: 14:29:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32314.0 tgs: 270 data_time: 1.12s time: 119.65s eta: 14:31:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30878.0 tgs: 259 data_time: 0.92s time: 119.18s eta: 14:26:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31664.0 tgs: 265 data_time: 0.84s time: 119.20s eta: 14:24:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 269 data_time: 0.59s time: 119.39s eta: 14:23:35 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32029.0 tgs: 265 data_time: 1.03s time: 120.73s eta: 14:31:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.197 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31643.0 tgs: 266 data_time: 0.92s time: 118.85s eta: 14:15:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.197 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31134.0 tgs: 261 data_time: 0.77s time: 119.10s eta: 14:15:31 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31687.0 tgs: 265 data_time: 0.81s time: 119.15s eta: 14:13:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.207 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 262 data_time: 0.84s time: 119.96s eta: 14:17:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.219 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31467.0 tgs: 262 data_time: 1.27s time: 119.87s eta: 14:15:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.225 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31775.0 tgs: 265 data_time: 0.93s time: 119.59s eta: 14:11:03 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31887.0 tgs: 267 data_time: 0.81s time: 119.22s eta: 14:06:29 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.220 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32107.0 tgs: 269 data_time: 0.86s time: 119.26s eta: 14:04:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 264 data_time: 0.76s time: 119.27s eta: 14:02:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.248 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32275.0 tgs: 270 data_time: 0.88s time: 119.15s eta: 13:59:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31573.0 tgs: 265 data_time: 0.63s time: 119.12s eta: 13:57:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31903.0 tgs: 267 data_time: 0.73s time: 119.17s eta: 13:56:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 268 data_time: 1.08s time: 119.35s eta: 13:55:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31342.0 tgs: 260 data_time: 1.08s time: 120.11s eta: 13:58:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 263 data_time: 0.95s time: 122.55s eta: 14:13:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31356.0 tgs: 263 data_time: 0.79s time: 119.10s eta: 13:47:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32027.0 tgs: 268 data_time: 0.91s time: 119.37s eta: 13:47:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.184 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31724.0 tgs: 265 data_time: 0.96s time: 119.36s eta: 13:45:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31187.0 tgs: 261 data_time: 0.70s time: 119.33s eta: 13:43:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 265 data_time: 0.99s time: 119.11s eta: 13:39:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.125 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30117.0 tgs: 252 data_time: 0.71s time: 119.25s eta: 13:38:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31411.0 tgs: 263 data_time: 0.94s time: 119.13s eta: 13:36:03 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30856.0 tgs: 258 data_time: 0.71s time: 119.22s eta: 13:34:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32307.0 tgs: 270 data_time: 0.97s time: 119.46s eta: 13:34:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30702.0 tgs: 257 data_time: 0.66s time: 119.08s eta: 13:29:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 265 data_time: 0.97s time: 118.93s eta: 13:26:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30984.0 tgs: 258 data_time: 0.88s time: 119.81s eta: 13:30:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32159.0 tgs: 269 data_time: 1.07s time: 119.26s eta: 13:24:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.250 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31857.0 tgs: 267 data_time: 0.80s time: 119.16s eta: 13:22:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31142.0 tgs: 260 data_time: 0.85s time: 119.35s eta: 13:21:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31976.0 tgs: 268 data_time: 0.67s time: 119.26s eta: 13:19:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31689.0 tgs: 266 data_time: 0.50s time: 119.04s eta: 13:15:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.224 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32024.0 tgs: 268 data_time: 0.92s time: 119.37s eta: 13:15:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31461.0 tgs: 263 data_time: 0.93s time: 119.19s eta: 13:12:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30421.0 tgs: 254 data_time: 0.84s time: 119.45s eta: 13:12:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31602.0 tgs: 264 data_time: 0.79s time: 119.29s eta: 13:09:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30923.0 tgs: 259 data_time: 0.80s time: 119.15s eta: 13:06:22 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31168.0 tgs: 261 data_time: 0.93s time: 119.36s eta: 13:05:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31210.0 tgs: 259 data_time: 0.79s time: 120.16s eta: 13:09:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.199 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31474.0 tgs: 263 data_time: 1.01s time: 119.50s eta: 13:02:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32426.0 tgs: 271 data_time: 0.72s time: 119.51s eta: 13:00:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30215.0 tgs: 252 data_time: 0.72s time: 119.57s eta: 12:59:13 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32410.0 tgs: 270 data_time: 0.89s time: 119.80s eta: 12:58:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.116 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30666.0 tgs: 257 data_time: 0.68s time: 119.03s eta: 12:51:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32326.0 tgs: 271 data_time: 0.68s time: 119.05s eta: 12:49:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 265 data_time: 0.79s time: 119.12s eta: 12:48:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31943.0 tgs: 267 data_time: 0.59s time: 119.39s eta: 12:48:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31275.0 tgs: 262 data_time: 0.75s time: 119.18s eta: 12:44:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 267 data_time: 1.14s time: 119.44s eta: 12:44:23 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31769.0 tgs: 266 data_time: 0.89s time: 119.09s eta: 12:40:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30931.0 tgs: 259 data_time: 0.83s time: 119.07s eta: 12:38:03 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32408.0 tgs: 270 data_time: 0.98s time: 119.59s eta: 12:39:24 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30927.0 tgs: 260 data_time: 0.76s time: 118.78s eta: 12:32:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30636.0 tgs: 257 data_time: 0.87s time: 118.88s eta: 12:30:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.205 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32377.0 tgs: 271 data_time: 0.80s time: 119.17s eta: 12:30:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30564.0 tgs: 256 data_time: 0.70s time: 119.27s eta: 12:29:24 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.267 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31100.0 tgs: 260 data_time: 0.80s time: 119.34s eta: 12:27:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 265 data_time: 0.84s time: 119.12s eta: 12:24:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30745.0 tgs: 257 data_time: 0.87s time: 119.34s eta: 12:23:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 267 data_time: 0.72s time: 119.26s eta: 12:21:24 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 269 data_time: 0.80s time: 119.02s eta: 12:17:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 263 data_time: 0.81s time: 121.13s eta: 12:28:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32191.0 tgs: 268 data_time: 0.90s time: 120.01s eta: 12:20:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32068.0 tgs: 270 data_time: 0.78s time: 118.71s eta: 12:10:03 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30805.0 tgs: 257 data_time: 0.94s time: 119.63s eta: 12:13:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29985.0 tgs: 250 data_time: 0.66s time: 119.83s eta: 12:12:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.107 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31869.0 tgs: 267 data_time: 0.67s time: 119.14s eta: 12:06:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31337.0 tgs: 262 data_time: 1.03s time: 119.17s eta: 12:04:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32491.0 tgs: 272 data_time: 0.64s time: 119.12s eta: 12:02:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31755.0 tgs: 266 data_time: 0.75s time: 119.18s eta: 12:01:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.119 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 266 data_time: 0.93s time: 118.95s eta: 11:57:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31798.0 tgs: 265 data_time: 0.79s time: 119.71s eta: 12:00:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30741.0 tgs: 257 data_time: 0.98s time: 119.27s eta: 11:55:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.183 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 0.71s time: 119.27s eta: 11:53:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.226 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 266 data_time: 1.06s time: 119.71s eta: 11:54:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31112.0 tgs: 261 data_time: 0.84s time: 118.79s eta: 11:46:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31946.0 tgs: 268 data_time: 0.63s time: 119.15s eta: 11:46:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30980.0 tgs: 259 data_time: 0.91s time: 119.61s eta: 11:47:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31681.0 tgs: 266 data_time: 0.81s time: 118.86s eta: 11:41:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31755.0 tgs: 267 data_time: 0.54s time: 118.68s eta: 11:38:14 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32017.0 tgs: 268 data_time: 0.75s time: 119.21s eta: 11:39:22 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32488.0 tgs: 272 data_time: 0.62s time: 119.14s eta: 11:36:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32372.0 tgs: 271 data_time: 0.83s time: 119.07s eta: 11:34:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31108.0 tgs: 261 data_time: 0.91s time: 118.95s eta: 11:31:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30700.0 tgs: 258 data_time: 0.74s time: 118.91s eta: 11:29:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.126 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 268 data_time: 0.80s time: 118.97s eta: 11:28:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.201 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 265 data_time: 1.10s time: 119.28s eta: 11:27:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32159.0 tgs: 269 data_time: 1.01s time: 119.35s eta: 11:26:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31275.0 tgs: 263 data_time: 0.89s time: 118.89s eta: 11:21:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31077.0 tgs: 260 data_time: 0.75s time: 119.20s eta: 11:21:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31813.0 tgs: 265 data_time: 0.61s time: 119.71s eta: 11:22:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.193 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32161.0 tgs: 269 data_time: 0.90s time: 119.12s eta: 11:16:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30782.0 tgs: 258 data_time: 0.73s time: 119.12s eta: 11:15:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31593.0 tgs: 265 data_time: 0.62s time: 118.99s eta: 11:12:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31866.0 tgs: 267 data_time: 0.63s time: 119.09s eta: 11:10:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30517.0 tgs: 256 data_time: 0.76s time: 119.16s eta: 11:09:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 267 data_time: 0.91s time: 118.97s eta: 11:06:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.247 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31864.0 tgs: 266 data_time: 1.03s time: 119.47s eta: 11:07:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.189 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31661.0 tgs: 264 data_time: 0.95s time: 119.64s eta: 11:06:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31328.0 tgs: 262 data_time: 0.73s time: 119.30s eta: 11:02:07 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32509.0 tgs: 272 data_time: 1.07s time: 119.31s eta: 11:00:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32300.0 tgs: 271 data_time: 0.94s time: 118.82s eta: 10:55:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30530.0 tgs: 255 data_time: 0.70s time: 119.47s eta: 10:57:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31745.0 tgs: 265 data_time: 0.64s time: 119.37s eta: 10:54:32 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32592.0 tgs: 273 data_time: 0.61s time: 119.28s eta: 10:52:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30908.0 tgs: 259 data_time: 0.87s time: 119.26s eta: 10:49:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 270 data_time: 1.01s time: 119.09s eta: 10:47:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 28891.0 tgs: 241 data_time: 0.49s time: 119.54s eta: 10:47:30 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29898.0 tgs: 250 data_time: 0.97s time: 119.31s eta: 10:44:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31937.0 tgs: 267 data_time: 1.12s time: 119.51s eta: 10:43:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.206 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31014.0 tgs: 259 data_time: 0.85s time: 119.50s eta: 10:41:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31620.0 tgs: 265 data_time: 0.84s time: 118.99s eta: 10:36:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.182 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31997.0 tgs: 268 data_time: 0.76s time: 119.22s eta: 10:35:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.236 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32163.0 tgs: 270 data_time: 0.69s time: 119.06s eta: 10:32:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.259 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31447.0 tgs: 264 data_time: 0.73s time: 118.78s eta: 10:29:32 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31725.0 tgs: 264 data_time: 0.72s time: 119.86s eta: 10:33:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32494.0 tgs: 272 data_time: 0.92s time: 119.38s eta: 10:28:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31626.0 tgs: 266 data_time: 0.66s time: 118.83s eta: 10:23:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31628.0 tgs: 265 data_time: 0.60s time: 118.99s eta: 10:22:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.229 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32482.0 tgs: 272 data_time: 0.81s time: 119.28s eta: 10:22:14 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31600.0 tgs: 265 data_time: 1.26s time: 119.20s eta: 10:19:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 268 data_time: 0.61s time: 119.06s eta: 10:17:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32318.0 tgs: 270 data_time: 0.82s time: 119.47s eta: 10:17:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.198 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 269 data_time: 0.70s time: 119.42s eta: 10:14:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31662.0 tgs: 265 data_time: 0.70s time: 119.45s eta: 10:13:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30437.0 tgs: 255 data_time: 1.06s time: 119.29s eta: 10:10:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32081.0 tgs: 269 data_time: 0.69s time: 119.11s eta: 10:07:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31592.0 tgs: 263 data_time: 0.84s time: 119.70s eta: 10:08:29 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31436.0 tgs: 262 data_time: 0.57s time: 119.82s eta: 10:07:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31772.0 tgs: 261 data_time: 0.93s time: 121.55s eta: 10:13:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.199 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31699.0 tgs: 265 data_time: 0.77s time: 119.24s eta: 10:00:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32257.0 tgs: 271 data_time: 0.85s time: 118.81s eta: 9:56:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 265 data_time: 0.90s time: 119.56s eta: 9:57:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31742.0 tgs: 265 data_time: 0.82s time: 119.70s eta: 9:56:30 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31585.0 tgs: 264 data_time: 0.72s time: 119.26s eta: 9:52:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30911.0 tgs: 259 data_time: 0.65s time: 119.33s eta: 9:50:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.172 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31689.0 tgs: 265 data_time: 1.08s time: 119.22s eta: 9:48:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31738.0 tgs: 262 data_time: 1.04s time: 120.88s eta: 9:54:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.116 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31384.0 tgs: 263 data_time: 0.82s time: 119.27s eta: 9:44:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.236 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32173.0 tgs: 269 data_time: 0.78s time: 119.46s eta: 9:43:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 260 data_time: 0.63s time: 119.55s eta: 9:41:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.185 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 268 data_time: 0.89s time: 119.20s eta: 9:38:07 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.190 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31706.0 tgs: 265 data_time: 0.70s time: 119.48s eta: 9:37:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.172 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30917.0 tgs: 259 data_time: 0.86s time: 119.06s eta: 9:33:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32313.0 tgs: 270 data_time: 0.62s time: 119.24s eta: 9:32:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.123 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30877.0 tgs: 259 data_time: 0.55s time: 119.16s eta: 9:29:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31440.0 tgs: 263 data_time: 0.86s time: 119.22s eta: 9:28:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.151 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32401.0 tgs: 271 data_time: 0.94s time: 119.46s eta: 9:27:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.114 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30842.0 tgs: 259 data_time: 0.78s time: 118.98s eta: 9:23:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32699.0 tgs: 274 data_time: 0.81s time: 119.04s eta: 9:21:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 266 data_time: 0.77s time: 119.52s eta: 9:21:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.111 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 267 data_time: 0.70s time: 119.59s eta: 9:20:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32005.0 tgs: 268 data_time: 0.82s time: 119.12s eta: 9:15:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31151.0 tgs: 261 data_time: 0.85s time: 119.04s eta: 9:13:32 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31785.0 tgs: 267 data_time: 0.97s time: 118.88s eta: 9:10:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.76s time: 119.32s eta: 9:10:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.206 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31348.0 tgs: 263 data_time: 1.06s time: 119.09s eta: 9:07:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 269 data_time: 0.67s time: 118.86s eta: 9:04:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 267 data_time: 0.75s time: 119.66s eta: 9:06:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.107 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 267 data_time: 0.80s time: 119.43s eta: 9:03:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.120 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31719.0 tgs: 265 data_time: 0.94s time: 119.43s eta: 9:01:24 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32187.0 tgs: 271 data_time: 0.58s time: 118.74s eta: 8:56:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31292.0 tgs: 263 data_time: 0.54s time: 118.87s eta: 8:54:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31138.0 tgs: 260 data_time: 0.82s time: 119.33s eta: 8:55:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30953.0 tgs: 259 data_time: 0.79s time: 119.15s eta: 8:52:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31993.0 tgs: 268 data_time: 0.87s time: 119.27s eta: 8:50:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31220.0 tgs: 259 data_time: 0.80s time: 120.47s eta: 8:54:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31429.0 tgs: 263 data_time: 0.95s time: 119.24s eta: 8:46:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31862.0 tgs: 267 data_time: 0.77s time: 119.08s eta: 8:43:57 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.223 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32053.0 tgs: 265 data_time: 0.94s time: 120.83s eta: 8:49:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 268 data_time: 0.64s time: 120.19s eta: 8:44:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31813.0 tgs: 267 data_time: 0.93s time: 118.82s eta: 8:36:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31227.0 tgs: 261 data_time: 0.73s time: 119.35s eta: 8:37:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32055.0 tgs: 268 data_time: 0.73s time: 119.33s eta: 8:35:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31599.0 tgs: 265 data_time: 0.79s time: 119.05s eta: 8:31:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30507.0 tgs: 256 data_time: 0.93s time: 119.16s eta: 8:30:24 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30187.0 tgs: 253 data_time: 0.89s time: 119.04s eta: 8:27:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32292.0 tgs: 271 data_time: 0.77s time: 119.12s eta: 8:26:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.183 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31867.0 tgs: 267 data_time: 0.79s time: 119.24s eta: 8:24:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.116 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31197.0 tgs: 261 data_time: 0.59s time: 119.22s eta: 8:22:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32287.0 tgs: 270 data_time: 0.98s time: 119.38s eta: 8:21:23 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.179 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30415.0 tgs: 253 data_time: 0.72s time: 119.78s eta: 8:21:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.73s time: 119.19s eta: 8:16:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31878.0 tgs: 267 data_time: 1.03s time: 119.02s eta: 8:13:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 267 data_time: 0.59s time: 119.89s eta: 8:15:32 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.254 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32189.0 tgs: 269 data_time: 0.99s time: 119.55s eta: 8:12:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32212.0 tgs: 269 data_time: 0.92s time: 119.59s eta: 8:10:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31953.0 tgs: 268 data_time: 0.87s time: 119.00s eta: 8:05:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32381.0 tgs: 270 data_time: 0.79s time: 119.84s eta: 8:07:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.189 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 266 data_time: 0.71s time: 119.71s eta: 8:04:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32254.0 tgs: 270 data_time: 0.85s time: 119.40s eta: 8:01:35 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31529.0 tgs: 264 data_time: 0.77s time: 119.29s eta: 7:59:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 269 data_time: 1.12s time: 119.18s eta: 7:56:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 28927.0 tgs: 242 data_time: 0.75s time: 119.21s eta: 7:54:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.182 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32226.0 tgs: 270 data_time: 0.83s time: 118.93s eta: 7:51:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30494.0 tgs: 256 data_time: 0.93s time: 118.97s eta: 7:49:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.177 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31409.0 tgs: 263 data_time: 0.86s time: 119.41s eta: 7:49:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31292.0 tgs: 262 data_time: 0.65s time: 119.03s eta: 7:46:13 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31966.0 tgs: 268 data_time: 0.89s time: 119.10s eta: 7:44:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31764.0 tgs: 264 data_time: 1.04s time: 120.00s eta: 7:46:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32366.0 tgs: 270 data_time: 0.69s time: 119.83s eta: 7:43:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31637.0 tgs: 265 data_time: 0.74s time: 119.16s eta: 7:38:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32115.0 tgs: 269 data_time: 0.83s time: 118.96s eta: 7:36:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.230 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30331.0 tgs: 254 data_time: 0.74s time: 119.32s eta: 7:35:24 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 266 data_time: 0.80s time: 119.38s eta: 7:33:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 263 data_time: 0.94s time: 119.06s eta: 7:30:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 268 data_time: 0.91s time: 118.99s eta: 7:28:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.211 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31586.0 tgs: 264 data_time: 0.96s time: 119.24s eta: 7:27:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31144.0 tgs: 261 data_time: 0.56s time: 118.99s eta: 7:24:13 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32320.0 tgs: 271 data_time: 0.97s time: 118.93s eta: 7:22:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.135 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31680.0 tgs: 266 data_time: 0.79s time: 119.08s eta: 7:20:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.91s time: 118.99s eta: 7:18:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30274.0 tgs: 254 data_time: 0.73s time: 119.03s eta: 7:16:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.197 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32531.0 tgs: 272 data_time: 0.96s time: 119.27s eta: 7:15:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31407.0 tgs: 263 data_time: 0.70s time: 119.38s eta: 7:13:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31718.0 tgs: 266 data_time: 0.93s time: 118.94s eta: 7:10:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.176 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32091.0 tgs: 269 data_time: 0.73s time: 119.20s eta: 7:09:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 266 data_time: 0.48s time: 119.59s eta: 7:08:30 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 264 data_time: 0.64s time: 119.66s eta: 7:06:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31979.0 tgs: 268 data_time: 0.78s time: 118.94s eta: 7:02:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31525.0 tgs: 264 data_time: 0.73s time: 119.30s eta: 7:01:30 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31397.0 tgs: 262 data_time: 0.85s time: 119.52s eta: 7:00:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 267 data_time: 0.97s time: 119.42s eta: 6:57:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31762.0 tgs: 267 data_time: 0.88s time: 118.90s eta: 6:54:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31543.0 tgs: 264 data_time: 0.95s time: 119.36s eta: 6:53:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31681.0 tgs: 266 data_time: 0.87s time: 118.92s eta: 6:50:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.212 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32002.0 tgs: 268 data_time: 0.71s time: 119.26s eta: 6:49:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32024.0 tgs: 269 data_time: 0.60s time: 119.02s eta: 6:46:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31421.0 tgs: 263 data_time: 0.80s time: 119.23s eta: 6:45:23 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32312.0 tgs: 270 data_time: 0.91s time: 119.40s eta: 6:43:57 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31869.0 tgs: 267 data_time: 0.90s time: 119.01s eta: 6:40:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 262 data_time: 0.72s time: 119.01s eta: 6:38:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30799.0 tgs: 257 data_time: 0.55s time: 119.46s eta: 6:38:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 266 data_time: 1.03s time: 118.80s eta: 6:34:01 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 266 data_time: 0.66s time: 119.44s eta: 6:34:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31184.0 tgs: 260 data_time: 0.81s time: 119.57s eta: 6:32:35 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32529.0 tgs: 272 data_time: 0.87s time: 119.38s eta: 6:29:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31739.0 tgs: 265 data_time: 0.86s time: 119.71s eta: 6:29:03 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32077.0 tgs: 268 data_time: 0.94s time: 119.37s eta: 6:25:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31358.0 tgs: 263 data_time: 0.72s time: 118.82s eta: 6:22:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.195 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 267 data_time: 0.83s time: 119.02s eta: 6:20:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 267 data_time: 0.86s time: 119.40s eta: 6:20:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29249.0 tgs: 245 data_time: 0.66s time: 119.34s eta: 6:17:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.183 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30464.0 tgs: 255 data_time: 0.87s time: 119.05s eta: 6:15:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31018.0 tgs: 259 data_time: 0.96s time: 119.42s eta: 6:14:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32262.0 tgs: 270 data_time: 0.83s time: 119.39s eta: 6:12:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31826.0 tgs: 266 data_time: 0.80s time: 119.28s eta: 6:09:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30312.0 tgs: 254 data_time: 0.75s time: 119.10s eta: 6:07:14 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31302.0 tgs: 256 data_time: 1.10s time: 121.96s eta: 6:14:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31805.0 tgs: 266 data_time: 0.74s time: 119.39s eta: 6:04:07 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 1.02s time: 119.57s eta: 6:02:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30584.0 tgs: 256 data_time: 0.77s time: 119.26s eta: 5:59:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31822.0 tgs: 266 data_time: 0.91s time: 119.26s eta: 5:57:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31679.0 tgs: 266 data_time: 0.84s time: 118.91s eta: 5:54:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 267 data_time: 0.69s time: 119.02s eta: 5:53:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32048.0 tgs: 269 data_time: 0.95s time: 119.07s eta: 5:51:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31513.0 tgs: 264 data_time: 0.62s time: 119.29s eta: 5:49:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32207.0 tgs: 270 data_time: 0.71s time: 118.97s eta: 5:46:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31231.0 tgs: 261 data_time: 0.98s time: 119.27s eta: 5:45:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32311.0 tgs: 270 data_time: 0.86s time: 119.33s eta: 5:44:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 266 data_time: 0.56s time: 119.05s eta: 5:41:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.153 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31186.0 tgs: 260 data_time: 0.83s time: 119.49s eta: 5:40:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32109.0 tgs: 268 data_time: 0.74s time: 119.38s eta: 5:38:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30948.0 tgs: 258 data_time: 0.64s time: 119.57s eta: 5:36:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32614.0 tgs: 274 data_time: 0.69s time: 118.98s eta: 5:33:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30620.0 tgs: 257 data_time: 0.85s time: 118.95s eta: 5:31:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.122 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 269 data_time: 0.46s time: 119.24s eta: 5:29:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 267 data_time: 0.82s time: 119.40s eta: 5:28:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.187 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32322.0 tgs: 270 data_time: 0.93s time: 119.46s eta: 5:26:31 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31272.0 tgs: 262 data_time: 0.60s time: 118.95s eta: 5:23:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31917.0 tgs: 266 data_time: 0.80s time: 119.63s eta: 5:22:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31768.0 tgs: 266 data_time: 1.20s time: 119.35s eta: 5:20:14 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.176 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32052.0 tgs: 269 data_time: 0.90s time: 119.12s eta: 5:17:39 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31452.0 tgs: 263 data_time: 0.85s time: 119.42s eta: 5:16:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 266 data_time: 0.81s time: 119.61s eta: 5:14:57 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.203 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32188.0 tgs: 270 data_time: 0.87s time: 119.01s eta: 5:11:23 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31324.0 tgs: 263 data_time: 0.47s time: 118.99s eta: 5:09:23 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30612.0 tgs: 256 data_time: 0.75s time: 119.23s eta: 5:08:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31497.0 tgs: 263 data_time: 0.98s time: 119.68s eta: 5:07:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 266 data_time: 0.64s time: 119.12s eta: 5:03:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.113 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29749.0 tgs: 249 data_time: 0.81s time: 119.06s eta: 5:01:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31551.0 tgs: 259 data_time: 1.07s time: 121.52s eta: 5:05:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32444.0 tgs: 271 data_time: 1.07s time: 119.58s eta: 4:58:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31232.0 tgs: 261 data_time: 0.71s time: 119.49s eta: 4:56:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30928.0 tgs: 258 data_time: 0.87s time: 119.73s eta: 4:55:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31578.0 tgs: 264 data_time: 0.77s time: 119.52s eta: 4:52:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 267 data_time: 0.65s time: 119.42s eta: 4:50:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31113.0 tgs: 261 data_time: 0.78s time: 119.20s eta: 4:48:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31562.0 tgs: 265 data_time: 0.60s time: 118.88s eta: 4:45:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31492.0 tgs: 264 data_time: 0.78s time: 119.09s eta: 4:43:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32289.0 tgs: 269 data_time: 0.65s time: 119.99s eta: 4:43:57 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31861.0 tgs: 267 data_time: 0.78s time: 119.14s eta: 4:39:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31218.0 tgs: 262 data_time: 0.78s time: 119.01s eta: 4:37:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31172.0 tgs: 261 data_time: 0.97s time: 119.06s eta: 4:35:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.113 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 268 data_time: 0.95s time: 119.38s eta: 4:34:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32099.0 tgs: 268 data_time: 0.93s time: 119.40s eta: 4:32:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31802.0 tgs: 267 data_time: 1.06s time: 118.94s eta: 4:29:35 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31992.0 tgs: 267 data_time: 0.67s time: 119.61s eta: 4:29:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31289.0 tgs: 262 data_time: 0.79s time: 119.24s eta: 4:26:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32434.0 tgs: 272 data_time: 0.68s time: 119.06s eta: 4:23:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31443.0 tgs: 264 data_time: 0.76s time: 118.85s eta: 4:21:27 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.201 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31007.0 tgs: 261 data_time: 0.84s time: 118.76s eta: 4:19:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.262 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31973.0 tgs: 267 data_time: 0.94s time: 119.43s eta: 4:18:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 266 data_time: 0.72s time: 119.40s eta: 4:16:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31737.0 tgs: 265 data_time: 0.98s time: 119.35s eta: 4:14:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32173.0 tgs: 270 data_time: 0.66s time: 119.05s eta: 4:11:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30968.0 tgs: 259 data_time: 0.68s time: 119.12s eta: 4:10:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 269 data_time: 0.72s time: 119.14s eta: 4:08:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31927.0 tgs: 268 data_time: 0.76s time: 118.96s eta: 4:05:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31522.0 tgs: 264 data_time: 0.80s time: 119.15s eta: 4:04:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31635.0 tgs: 264 data_time: 0.86s time: 119.50s eta: 4:02:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31152.0 tgs: 261 data_time: 0.86s time: 119.24s eta: 4:00:28 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30932.0 tgs: 259 data_time: 0.79s time: 119.19s eta: 3:58:22 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31363.0 tgs: 262 data_time: 0.52s time: 119.36s eta: 3:56:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30430.0 tgs: 254 data_time: 0.72s time: 119.37s eta: 3:54:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31294.0 tgs: 260 data_time: 0.72s time: 119.96s eta: 3:53:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.75s time: 119.32s eta: 3:50:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31655.0 tgs: 264 data_time: 0.55s time: 119.63s eta: 3:49:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31612.0 tgs: 265 data_time: 0.53s time: 118.85s eta: 3:45:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32341.0 tgs: 270 data_time: 1.09s time: 119.75s eta: 3:45:31 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31816.0 tgs: 266 data_time: 0.85s time: 119.32s eta: 3:42:44 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 269 data_time: 0.89s time: 119.26s eta: 3:40:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30259.0 tgs: 254 data_time: 0.65s time: 119.00s eta: 3:38:09 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.222 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31645.0 tgs: 264 data_time: 0.85s time: 119.43s eta: 3:36:57 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32428.0 tgs: 272 data_time: 0.72s time: 119.03s eta: 3:34:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31725.0 tgs: 266 data_time: 1.07s time: 118.99s eta: 3:32:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32217.0 tgs: 270 data_time: 0.87s time: 119.22s eta: 3:30:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31727.0 tgs: 264 data_time: 0.82s time: 120.11s eta: 3:30:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.73s time: 119.13s eta: 3:26:29 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32379.0 tgs: 271 data_time: 0.78s time: 119.31s eta: 3:24:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31484.0 tgs: 264 data_time: 0.56s time: 119.08s eta: 3:22:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31724.0 tgs: 265 data_time: 1.18s time: 119.43s eta: 3:21:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32442.0 tgs: 271 data_time: 0.58s time: 119.29s eta: 3:18:48 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30447.0 tgs: 255 data_time: 0.78s time: 119.22s eta: 3:16:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31863.0 tgs: 267 data_time: 0.66s time: 119.14s eta: 3:14:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31138.0 tgs: 261 data_time: 0.58s time: 119.02s eta: 3:12:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32193.0 tgs: 270 data_time: 0.83s time: 119.18s eta: 3:10:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31020.0 tgs: 256 data_time: 0.92s time: 121.07s eta: 3:11:41 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31011.0 tgs: 259 data_time: 0.72s time: 119.50s eta: 3:07:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30905.0 tgs: 259 data_time: 0.74s time: 119.26s eta: 3:04:51 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31911.0 tgs: 268 data_time: 0.78s time: 118.98s eta: 3:02:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31956.0 tgs: 268 data_time: 0.75s time: 119.22s eta: 3:00:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31884.0 tgs: 266 data_time: 0.65s time: 119.46s eta: 2:59:11 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31683.0 tgs: 266 data_time: 0.55s time: 118.98s eta: 2:56:29 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32096.0 tgs: 268 data_time: 0.76s time: 119.35s eta: 2:55:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 264 data_time: 0.83s time: 119.41s eta: 2:53:08 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 270 data_time: 0.97s time: 119.35s eta: 2:51:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30571.0 tgs: 251 data_time: 0.64s time: 121.46s eta: 2:52:04 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31775.0 tgs: 266 data_time: 1.02s time: 119.24s eta: 2:46:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 266 data_time: 1.04s time: 119.60s eta: 2:45:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31649.0 tgs: 266 data_time: 0.67s time: 118.92s eta: 2:42:31 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31201.0 tgs: 261 data_time: 0.44s time: 119.18s eta: 2:40:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32373.0 tgs: 272 data_time: 0.87s time: 118.97s eta: 2:38:37 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31901.0 tgs: 267 data_time: 1.19s time: 119.22s eta: 2:36:58 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32316.0 tgs: 271 data_time: 0.64s time: 119.11s eta: 2:34:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31988.0 tgs: 267 data_time: 0.65s time: 119.43s eta: 2:33:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32265.0 tgs: 268 data_time: 0.83s time: 120.22s eta: 2:32:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31279.0 tgs: 262 data_time: 1.03s time: 119.28s eta: 2:29:05 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 268 data_time: 0.92s time: 118.88s eta: 2:26:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 269 data_time: 0.68s time: 119.52s eta: 2:25:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 270 data_time: 0.62s time: 119.06s eta: 2:22:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31118.0 tgs: 261 data_time: 0.65s time: 119.14s eta: 2:20:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32017.0 tgs: 268 data_time: 0.72s time: 119.04s eta: 2:18:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32367.0 tgs: 270 data_time: 1.26s time: 119.85s eta: 2:17:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 267 data_time: 0.93s time: 119.32s eta: 2:15:14 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32033.0 tgs: 268 data_time: 0.82s time: 119.38s eta: 2:13:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31435.0 tgs: 262 data_time: 0.73s time: 119.68s eta: 2:11:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31094.0 tgs: 260 data_time: 0.69s time: 119.49s eta: 2:09:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31433.0 tgs: 264 data_time: 0.63s time: 118.96s eta: 2:06:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.71s time: 119.37s eta: 2:05:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32219.0 tgs: 270 data_time: 0.81s time: 119.27s eta: 2:03:14 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.212 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31147.0 tgs: 261 data_time: 0.95s time: 119.22s eta: 2:01:12 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 266 data_time: 1.04s time: 119.50s eta: 1:59:30 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30480.0 tgs: 255 data_time: 0.57s time: 119.42s eta: 1:57:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31941.0 tgs: 267 data_time: 0.69s time: 119.53s eta: 1:55:32 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30457.0 tgs: 255 data_time: 0.83s time: 119.30s eta: 1:53:20 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31861.0 tgs: 267 data_time: 0.70s time: 119.25s eta: 1:51:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31130.0 tgs: 261 data_time: 0.90s time: 118.93s eta: 1:49:00 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.247 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 264 data_time: 1.00s time: 119.64s eta: 1:47:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30889.0 tgs: 258 data_time: 0.62s time: 119.72s eta: 1:45:45 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31301.0 tgs: 262 data_time: 0.47s time: 119.35s eta: 1:43:26 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31135.0 tgs: 260 data_time: 0.74s time: 119.30s eta: 1:41:24 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31274.0 tgs: 262 data_time: 0.62s time: 119.28s eta: 1:39:23 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32423.0 tgs: 272 data_time: 0.71s time: 118.90s eta: 1:37:06 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31206.0 tgs: 261 data_time: 0.62s time: 119.16s eta: 1:35:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 268 data_time: 0.81s time: 119.06s eta: 1:33:15 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31494.0 tgs: 264 data_time: 0.54s time: 119.25s eta: 1:31:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.199 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 264 data_time: 0.65s time: 119.73s eta: 1:29:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32118.0 tgs: 269 data_time: 0.69s time: 119.02s eta: 1:27:16 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31034.0 tgs: 261 data_time: 0.58s time: 118.86s eta: 1:25:10 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30363.0 tgs: 254 data_time: 0.82s time: 119.48s eta: 1:23:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32476.0 tgs: 271 data_time: 0.81s time: 119.48s eta: 1:21:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31191.0 tgs: 262 data_time: 0.82s time: 118.94s eta: 1:19:17 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30941.0 tgs: 260 data_time: 1.13s time: 118.93s eta: 1:17:18 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 256 data_time: 0.89s time: 122.15s eta: 1:17:21 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32206.0 tgs: 269 data_time: 0.71s time: 119.36s eta: 1:13:36 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31738.0 tgs: 266 data_time: 0.96s time: 119.05s eta: 1:11:25 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.198 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31801.0 tgs: 267 data_time: 0.67s time: 118.92s eta: 1:09:22 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.205 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30140.0 tgs: 253 data_time: 0.98s time: 119.10s eta: 1:07:29 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31946.0 tgs: 267 data_time: 0.98s time: 119.41s eta: 1:05:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31164.0 tgs: 261 data_time: 0.52s time: 119.34s eta: 1:03:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31195.0 tgs: 261 data_time: 0.72s time: 119.18s eta: 1:01:34 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30802.0 tgs: 258 data_time: 0.84s time: 119.27s eta: 0:59:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 269 data_time: 0.89s time: 119.45s eta: 0:57:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31425.0 tgs: 263 data_time: 0.92s time: 119.39s eta: 0:55:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31849.0 tgs: 266 data_time: 0.72s time: 119.48s eta: 0:53:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31833.0 tgs: 266 data_time: 0.75s time: 119.36s eta: 0:51:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30686.0 tgs: 256 data_time: 0.88s time: 119.47s eta: 0:49:46 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.84s time: 118.89s eta: 0:47:33 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31345.0 tgs: 262 data_time: 0.70s time: 119.28s eta: 0:45:43 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 1.00s time: 119.22s eta: 0:43:42 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31559.0 tgs: 265 data_time: 0.86s time: 118.99s eta: 0:41:38 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31440.0 tgs: 264 data_time: 0.91s time: 119.04s eta: 0:39:40 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31963.0 tgs: 267 data_time: 0.85s time: 119.66s eta: 0:37:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31929.0 tgs: 267 data_time: 0.71s time: 119.28s eta: 0:35:47 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 267 data_time: 0.85s time: 119.75s eta: 0:33:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31757.0 tgs: 261 data_time: 0.65s time: 121.22s eta: 0:32:19 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31053.0 tgs: 260 data_time: 0.99s time: 119.37s eta: 0:29:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31991.0 tgs: 268 data_time: 0.64s time: 119.32s eta: 0:27:50 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31171.0 tgs: 260 data_time: 0.67s time: 119.63s eta: 0:25:55 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32630.0 tgs: 273 data_time: 0.91s time: 119.09s eta: 0:23:49 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 266 data_time: 0.84s time: 119.39s eta: 0:21:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31848.0 tgs: 267 data_time: 0.96s time: 119.26s eta: 0:19:52 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.103 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 28089.0 tgs: 235 data_time: 0.86s time: 119.24s eta: 0:17:53 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31984.0 tgs: 266 data_time: 1.15s time: 119.91s eta: 0:15:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 269 data_time: 0.81s time: 119.21s eta: 0:13:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32001.0 tgs: 268 data_time: 0.97s time: 119.09s eta: 0:11:54 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31407.0 tgs: 263 data_time: 1.07s time: 119.32s eta: 0:09:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32062.0 tgs: 269 data_time: 0.78s time: 119.17s eta: 0:07:56 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 261 data_time: 0.55s time: 119.69s eta: 0:05:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31391.0 tgs: 258 data_time: 0.84s time: 121.37s eta: 0:04:02 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32391.0 tgs: 271 data_time: 0.61s time: 119.21s eta: 0:01:59 +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 10][DP 2][SP 2][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank13.log b/20250121165312/rank13.log new file mode 100644 index 0000000000000000000000000000000000000000..0ed66386a955864328a07e05c701fc9ecfadec38 --- /dev/null +++ b/20250121165312/rank13.log @@ -0,0 +1,851 @@ +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.03s +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 62.96 seconds, peak gpu memory 13.4G +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.252 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.3GB text_tokens: 30603.0 tgs: 200 data_time: 1.99s time: 152.58s eta: 1 day, 1:08:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.242 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31849.0 tgs: 265 data_time: 0.70s time: 120.18s eta: 19:45:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.233 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31059.0 tgs: 259 data_time: 0.92s time: 119.69s eta: 19:38:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.219 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32411.0 tgs: 270 data_time: 0.83s time: 119.65s eta: 19:36:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31204.0 tgs: 261 data_time: 0.69s time: 119.31s eta: 19:31:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.244 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31449.0 tgs: 263 data_time: 0.84s time: 119.54s eta: 19:31:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.215 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31590.0 tgs: 256 data_time: 0.82s time: 122.96s eta: 20:02:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.252 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31967.0 tgs: 266 data_time: 0.80s time: 119.89s eta: 19:30:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31084.0 tgs: 257 data_time: 0.91s time: 120.65s eta: 19:36:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.210 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31694.0 tgs: 264 data_time: 0.85s time: 119.92s eta: 19:27:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.212 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32132.0 tgs: 266 data_time: 0.91s time: 120.46s eta: 19:30:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31935.0 tgs: 267 data_time: 0.82s time: 119.34s eta: 19:17:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31576.0 tgs: 263 data_time: 0.81s time: 119.68s eta: 19:18:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.252 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31631.0 tgs: 265 data_time: 0.79s time: 119.19s eta: 19:12:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31904.0 tgs: 267 data_time: 0.84s time: 119.22s eta: 19:10:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32061.0 tgs: 269 data_time: 0.82s time: 118.92s eta: 19:05:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30810.0 tgs: 258 data_time: 0.76s time: 119.15s eta: 19:05:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31204.0 tgs: 261 data_time: 0.88s time: 119.17s eta: 19:04:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 264 data_time: 1.17s time: 119.40s eta: 19:04:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.242 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 267 data_time: 0.72s time: 119.33s eta: 19:01:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31859.0 tgs: 264 data_time: 0.67s time: 120.30s eta: 19:08:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32340.0 tgs: 263 data_time: 0.88s time: 122.91s eta: 19:31:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31453.0 tgs: 263 data_time: 0.48s time: 119.48s eta: 18:57:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31779.0 tgs: 266 data_time: 0.61s time: 119.16s eta: 18:51:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31169.0 tgs: 261 data_time: 0.80s time: 119.24s eta: 18:50:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29708.0 tgs: 248 data_time: 0.93s time: 119.39s eta: 18:50:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31114.0 tgs: 260 data_time: 1.01s time: 119.66s eta: 18:50:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32372.0 tgs: 271 data_time: 0.93s time: 119.38s eta: 18:46:10 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31554.0 tgs: 263 data_time: 0.95s time: 119.79s eta: 18:48:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31042.0 tgs: 240 data_time: 1.01s time: 129.31s eta: 20:15:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.250 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31728.0 tgs: 265 data_time: 0.82s time: 119.58s eta: 18:42:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.282 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31816.0 tgs: 266 data_time: 0.71s time: 119.38s eta: 18:38:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30775.0 tgs: 255 data_time: 0.65s time: 120.66s eta: 18:48:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31569.0 tgs: 264 data_time: 0.49s time: 119.39s eta: 18:34:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31161.0 tgs: 261 data_time: 0.68s time: 119.16s eta: 18:30:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31655.0 tgs: 265 data_time: 0.97s time: 119.07s eta: 18:27:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31369.0 tgs: 263 data_time: 0.79s time: 119.14s eta: 18:25:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32447.0 tgs: 269 data_time: 0.84s time: 120.39s eta: 18:35:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.134 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31783.0 tgs: 262 data_time: 0.91s time: 121.17s eta: 18:40:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30915.0 tgs: 256 data_time: 0.73s time: 120.66s eta: 18:34:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30941.0 tgs: 258 data_time: 0.74s time: 119.74s eta: 18:23:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31815.0 tgs: 256 data_time: 0.53s time: 123.87s eta: 18:59:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32295.0 tgs: 270 data_time: 0.67s time: 119.56s eta: 18:17:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 268 data_time: 0.62s time: 119.36s eta: 18:14:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32370.0 tgs: 271 data_time: 0.70s time: 119.23s eta: 18:10:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31953.0 tgs: 267 data_time: 0.52s time: 119.60s eta: 18:12:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31767.0 tgs: 266 data_time: 0.68s time: 119.27s eta: 18:07:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31980.0 tgs: 268 data_time: 0.60s time: 119.01s eta: 18:02:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31917.0 tgs: 267 data_time: 0.85s time: 119.10s eta: 18:01:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31824.0 tgs: 266 data_time: 0.63s time: 119.26s eta: 18:01:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.126 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30092.0 tgs: 252 data_time: 0.75s time: 119.33s eta: 17:59:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31783.0 tgs: 264 data_time: 0.70s time: 120.02s eta: 18:04:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30978.0 tgs: 260 data_time: 0.72s time: 119.10s eta: 17:53:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 266 data_time: 0.78s time: 119.41s eta: 17:54:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31322.0 tgs: 263 data_time: 1.03s time: 119.00s eta: 17:48:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.122 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31440.0 tgs: 264 data_time: 0.71s time: 119.03s eta: 17:47:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31790.0 tgs: 253 data_time: 0.81s time: 125.36s eta: 18:41:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32257.0 tgs: 268 data_time: 0.75s time: 120.10s eta: 17:52:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31778.0 tgs: 266 data_time: 0.70s time: 119.38s eta: 17:44:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31886.0 tgs: 267 data_time: 0.99s time: 119.34s eta: 17:42:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31703.0 tgs: 265 data_time: 0.81s time: 119.51s eta: 17:41:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31729.0 tgs: 251 data_time: 0.96s time: 126.05s eta: 18:37:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32208.0 tgs: 262 data_time: 0.78s time: 122.63s eta: 18:05:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31951.0 tgs: 263 data_time: 0.92s time: 121.15s eta: 17:50:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 264 data_time: 0.56s time: 119.03s eta: 17:29:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31597.0 tgs: 265 data_time: 0.76s time: 118.85s eta: 17:25:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 253 data_time: 0.79s time: 124.00s eta: 18:09:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 255 data_time: 0.89s time: 124.56s eta: 18:12:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32423.0 tgs: 271 data_time: 0.90s time: 119.40s eta: 17:24:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32055.0 tgs: 267 data_time: 0.97s time: 119.70s eta: 17:25:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31574.0 tgs: 264 data_time: 0.81s time: 119.53s eta: 17:21:51 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31739.0 tgs: 267 data_time: 0.63s time: 118.86s eta: 17:14:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30556.0 tgs: 254 data_time: 0.88s time: 119.85s eta: 17:20:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31430.0 tgs: 262 data_time: 0.72s time: 119.56s eta: 17:16:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.114 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32284.0 tgs: 264 data_time: 0.75s time: 121.86s eta: 17:34:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31470.0 tgs: 254 data_time: 0.65s time: 123.51s eta: 17:46:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.252 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30719.0 tgs: 258 data_time: 0.69s time: 119.01s eta: 17:05:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32126.0 tgs: 269 data_time: 0.79s time: 119.23s eta: 17:05:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32041.0 tgs: 268 data_time: 0.87s time: 119.14s eta: 17:02:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.234 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30639.0 tgs: 256 data_time: 0.91s time: 119.26s eta: 17:01:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30181.0 tgs: 253 data_time: 0.73s time: 119.04s eta: 16:57:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32047.0 tgs: 268 data_time: 0.90s time: 119.28s eta: 16:57:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.122 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32087.0 tgs: 269 data_time: 0.88s time: 119.16s eta: 16:54:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32110.0 tgs: 268 data_time: 1.00s time: 119.58s eta: 16:56:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31342.0 tgs: 263 data_time: 0.93s time: 119.14s eta: 16:50:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32177.0 tgs: 270 data_time: 0.78s time: 118.83s eta: 16:46:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32327.0 tgs: 271 data_time: 0.67s time: 119.02s eta: 16:45:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31900.0 tgs: 266 data_time: 0.96s time: 119.85s eta: 16:50:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 267 data_time: 0.80s time: 119.39s eta: 16:44:51 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31125.0 tgs: 261 data_time: 0.63s time: 119.01s eta: 16:39:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31485.0 tgs: 260 data_time: 0.90s time: 120.93s eta: 16:53:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.225 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30552.0 tgs: 256 data_time: 0.72s time: 118.99s eta: 16:35:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31858.0 tgs: 267 data_time: 0.82s time: 119.28s eta: 16:36:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31967.0 tgs: 268 data_time: 0.81s time: 119.23s eta: 16:33:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31505.0 tgs: 263 data_time: 0.70s time: 119.36s eta: 16:32:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32148.0 tgs: 268 data_time: 0.98s time: 119.71s eta: 16:33:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32093.0 tgs: 269 data_time: 0.50s time: 119.20s eta: 16:27:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.120 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30875.0 tgs: 257 data_time: 0.65s time: 119.76s eta: 16:30:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30629.0 tgs: 254 data_time: 0.59s time: 120.29s eta: 16:32:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31020.0 tgs: 260 data_time: 0.73s time: 118.97s eta: 16:19:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31810.0 tgs: 249 data_time: 0.95s time: 127.32s eta: 17:26:10 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31760.0 tgs: 266 data_time: 0.92s time: 119.23s eta: 16:17:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31604.0 tgs: 265 data_time: 0.75s time: 119.14s eta: 16:14:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31868.0 tgs: 267 data_time: 0.52s time: 119.25s eta: 16:13:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32248.0 tgs: 270 data_time: 0.74s time: 119.17s eta: 16:11:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.209 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31194.0 tgs: 262 data_time: 0.52s time: 118.96s eta: 16:07:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30582.0 tgs: 257 data_time: 0.53s time: 118.86s eta: 16:04:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.235 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30688.0 tgs: 257 data_time: 1.01s time: 118.98s eta: 16:03:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30944.0 tgs: 254 data_time: 0.58s time: 121.82s eta: 16:24:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.252 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 270 data_time: 0.87s time: 118.98s eta: 15:59:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31088.0 tgs: 260 data_time: 0.74s time: 119.22s eta: 15:59:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31484.0 tgs: 263 data_time: 0.80s time: 119.62s eta: 16:00:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31214.0 tgs: 262 data_time: 0.84s time: 119.05s eta: 15:54:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29982.0 tgs: 252 data_time: 0.83s time: 118.97s eta: 15:51:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30837.0 tgs: 258 data_time: 0.53s time: 119.17s eta: 15:51:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30478.0 tgs: 254 data_time: 0.75s time: 119.64s eta: 15:53:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31476.0 tgs: 264 data_time: 1.01s time: 118.93s eta: 15:45:30 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.224 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 261 data_time: 0.69s time: 119.29s eta: 15:46:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.70s time: 119.13s eta: 15:43:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.213 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30903.0 tgs: 259 data_time: 0.73s time: 119.00s eta: 15:40:05 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29636.0 tgs: 247 data_time: 0.83s time: 119.66s eta: 15:43:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31716.0 tgs: 261 data_time: 1.03s time: 121.51s eta: 15:55:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31642.0 tgs: 264 data_time: 0.54s time: 119.41s eta: 15:37:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31202.0 tgs: 240 data_time: 0.75s time: 129.96s eta: 16:58:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31172.0 tgs: 260 data_time: 0.86s time: 119.47s eta: 15:33:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31893.0 tgs: 266 data_time: 0.67s time: 119.61s eta: 15:32:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30988.0 tgs: 259 data_time: 1.10s time: 119.54s eta: 15:30:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31862.0 tgs: 260 data_time: 0.66s time: 122.13s eta: 15:48:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31325.0 tgs: 259 data_time: 0.71s time: 120.75s eta: 15:35:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 267 data_time: 0.83s time: 119.41s eta: 15:23:27 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31539.0 tgs: 264 data_time: 0.86s time: 119.44s eta: 15:21:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31684.0 tgs: 240 data_time: 0.57s time: 131.63s eta: 16:53:33 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31324.0 tgs: 262 data_time: 0.65s time: 119.29s eta: 15:16:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31849.0 tgs: 266 data_time: 0.89s time: 119.57s eta: 15:16:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 260 data_time: 0.64s time: 122.25s eta: 15:35:14 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 265 data_time: 0.60s time: 119.71s eta: 15:13:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31483.0 tgs: 263 data_time: 0.65s time: 119.35s eta: 15:09:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 270 data_time: 1.01s time: 118.90s eta: 15:03:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.229 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32377.0 tgs: 272 data_time: 0.81s time: 118.99s eta: 15:02:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31437.0 tgs: 263 data_time: 0.76s time: 119.31s eta: 15:02:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31680.0 tgs: 266 data_time: 0.93s time: 118.92s eta: 14:57:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31893.0 tgs: 264 data_time: 1.10s time: 120.52s eta: 15:07:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31817.0 tgs: 266 data_time: 0.86s time: 119.31s eta: 14:56:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30697.0 tgs: 250 data_time: 0.78s time: 122.66s eta: 15:19:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32293.0 tgs: 271 data_time: 0.63s time: 119.12s eta: 14:51:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31577.0 tgs: 263 data_time: 1.05s time: 119.76s eta: 14:54:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 269 data_time: 0.75s time: 119.46s eta: 14:49:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30603.0 tgs: 257 data_time: 0.60s time: 118.99s eta: 14:44:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30939.0 tgs: 259 data_time: 0.64s time: 119.33s eta: 14:45:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31827.0 tgs: 267 data_time: 0.56s time: 118.94s eta: 14:40:10 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31853.0 tgs: 267 data_time: 0.90s time: 119.05s eta: 14:38:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32231.0 tgs: 270 data_time: 0.63s time: 119.07s eta: 14:37:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.197 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32310.0 tgs: 271 data_time: 0.80s time: 119.12s eta: 14:35:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31567.0 tgs: 265 data_time: 0.81s time: 119.00s eta: 14:32:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.204 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32220.0 tgs: 270 data_time: 0.87s time: 119.08s eta: 14:31:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31910.0 tgs: 267 data_time: 0.79s time: 119.15s eta: 14:29:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31826.0 tgs: 265 data_time: 0.96s time: 119.65s eta: 14:31:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31857.0 tgs: 267 data_time: 0.74s time: 119.18s eta: 14:26:00 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31576.0 tgs: 264 data_time: 0.85s time: 119.20s eta: 14:24:10 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31063.0 tgs: 260 data_time: 0.85s time: 119.39s eta: 14:23:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31593.0 tgs: 261 data_time: 0.65s time: 120.73s eta: 14:31:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31829.0 tgs: 267 data_time: 0.72s time: 118.85s eta: 14:15:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31724.0 tgs: 266 data_time: 0.94s time: 119.10s eta: 14:15:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.208 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31084.0 tgs: 260 data_time: 0.91s time: 119.15s eta: 14:13:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 265 data_time: 0.86s time: 119.96s eta: 14:17:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31581.0 tgs: 263 data_time: 0.88s time: 119.87s eta: 14:15:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.194 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31183.0 tgs: 260 data_time: 0.60s time: 119.58s eta: 14:11:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31992.0 tgs: 268 data_time: 0.74s time: 119.22s eta: 14:06:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32059.0 tgs: 268 data_time: 0.66s time: 119.26s eta: 14:04:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31271.0 tgs: 262 data_time: 0.80s time: 119.26s eta: 14:02:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31487.0 tgs: 264 data_time: 0.75s time: 119.14s eta: 13:59:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 267 data_time: 0.77s time: 119.12s eta: 13:57:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.224 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31802.0 tgs: 266 data_time: 0.98s time: 119.17s eta: 13:56:10 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.220 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 1.04s time: 119.35s eta: 13:55:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31401.0 tgs: 261 data_time: 0.56s time: 120.11s eta: 13:58:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.185 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31724.0 tgs: 258 data_time: 0.95s time: 122.55s eta: 14:13:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.261 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31083.0 tgs: 260 data_time: 0.63s time: 119.10s eta: 13:47:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30902.0 tgs: 258 data_time: 0.74s time: 119.37s eta: 13:47:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30470.0 tgs: 255 data_time: 0.79s time: 119.35s eta: 13:45:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.221 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32119.0 tgs: 269 data_time: 0.79s time: 119.33s eta: 13:43:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.231 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32163.0 tgs: 270 data_time: 0.65s time: 119.11s eta: 13:39:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 266 data_time: 0.78s time: 119.25s eta: 13:38:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31339.0 tgs: 263 data_time: 0.84s time: 119.13s eta: 13:36:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30960.0 tgs: 259 data_time: 0.60s time: 119.22s eta: 13:34:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 265 data_time: 0.88s time: 119.46s eta: 13:34:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.119 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31385.0 tgs: 263 data_time: 0.77s time: 119.08s eta: 13:29:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32288.0 tgs: 271 data_time: 0.69s time: 118.92s eta: 13:26:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30856.0 tgs: 257 data_time: 0.63s time: 119.81s eta: 13:30:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30916.0 tgs: 259 data_time: 0.66s time: 119.26s eta: 13:24:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31858.0 tgs: 267 data_time: 0.71s time: 119.15s eta: 13:22:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32383.0 tgs: 271 data_time: 1.02s time: 119.35s eta: 13:21:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.124 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 270 data_time: 0.71s time: 119.26s eta: 13:19:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 269 data_time: 0.75s time: 119.04s eta: 13:15:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31814.0 tgs: 266 data_time: 0.77s time: 119.37s eta: 13:15:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32253.0 tgs: 270 data_time: 0.68s time: 119.19s eta: 13:12:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31833.0 tgs: 266 data_time: 0.85s time: 119.45s eta: 13:12:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31667.0 tgs: 265 data_time: 0.80s time: 119.29s eta: 13:09:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31809.0 tgs: 266 data_time: 0.70s time: 119.15s eta: 13:06:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31247.0 tgs: 259 data_time: 0.89s time: 120.42s eta: 13:12:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31773.0 tgs: 264 data_time: 0.84s time: 120.16s eta: 13:09:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 271 data_time: 1.01s time: 119.50s eta: 13:02:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32284.0 tgs: 270 data_time: 0.85s time: 119.51s eta: 13:00:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31185.0 tgs: 260 data_time: 0.62s time: 119.57s eta: 12:59:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31665.0 tgs: 264 data_time: 0.78s time: 119.79s eta: 12:58:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31009.0 tgs: 260 data_time: 0.66s time: 119.04s eta: 12:51:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31824.0 tgs: 267 data_time: 0.72s time: 119.05s eta: 12:49:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.123 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31602.0 tgs: 265 data_time: 0.70s time: 119.12s eta: 12:48:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32357.0 tgs: 271 data_time: 0.81s time: 119.39s eta: 12:48:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31796.0 tgs: 266 data_time: 0.80s time: 119.18s eta: 12:44:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31182.0 tgs: 261 data_time: 0.83s time: 119.44s eta: 12:44:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31421.0 tgs: 263 data_time: 0.65s time: 119.09s eta: 12:40:13 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 261 data_time: 0.54s time: 119.07s eta: 12:38:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31440.0 tgs: 262 data_time: 0.89s time: 119.59s eta: 12:39:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32221.0 tgs: 271 data_time: 0.84s time: 118.78s eta: 12:32:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.118 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31762.0 tgs: 267 data_time: 0.73s time: 118.88s eta: 12:30:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30944.0 tgs: 259 data_time: 0.82s time: 119.17s eta: 12:30:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 266 data_time: 0.78s time: 119.27s eta: 12:29:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.225 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31841.0 tgs: 266 data_time: 0.85s time: 119.35s eta: 12:27:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.191 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31705.0 tgs: 266 data_time: 0.78s time: 119.12s eta: 12:24:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.212 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31629.0 tgs: 265 data_time: 0.77s time: 119.34s eta: 12:23:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31282.0 tgs: 262 data_time: 0.62s time: 119.26s eta: 12:21:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31980.0 tgs: 268 data_time: 0.79s time: 119.03s eta: 12:17:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.270 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31842.0 tgs: 262 data_time: 0.64s time: 121.13s eta: 12:28:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31831.0 tgs: 265 data_time: 0.91s time: 120.00s eta: 12:19:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 264 data_time: 0.74s time: 118.72s eta: 12:10:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.233 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31904.0 tgs: 266 data_time: 0.61s time: 119.63s eta: 12:13:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31947.0 tgs: 266 data_time: 0.76s time: 119.83s eta: 12:12:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30644.0 tgs: 257 data_time: 0.66s time: 119.14s eta: 12:06:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31390.0 tgs: 263 data_time: 0.93s time: 119.17s eta: 12:04:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31549.0 tgs: 264 data_time: 0.90s time: 119.12s eta: 12:02:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 267 data_time: 0.57s time: 119.18s eta: 12:01:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32601.0 tgs: 274 data_time: 0.88s time: 118.95s eta: 11:57:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 266 data_time: 0.87s time: 119.70s eta: 12:00:13 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 269 data_time: 0.90s time: 119.28s eta: 11:55:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31569.0 tgs: 264 data_time: 0.76s time: 119.27s eta: 11:53:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31690.0 tgs: 264 data_time: 0.66s time: 119.71s eta: 11:54:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 272 data_time: 0.79s time: 118.78s eta: 11:46:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31409.0 tgs: 263 data_time: 0.65s time: 119.15s eta: 11:46:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31947.0 tgs: 267 data_time: 0.87s time: 119.61s eta: 11:47:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31149.0 tgs: 262 data_time: 0.81s time: 118.86s eta: 11:41:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31203.0 tgs: 262 data_time: 0.60s time: 118.68s eta: 11:38:14 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31912.0 tgs: 267 data_time: 0.80s time: 119.21s eta: 11:39:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.200 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 271 data_time: 0.87s time: 119.14s eta: 11:36:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31360.0 tgs: 263 data_time: 0.85s time: 119.07s eta: 11:34:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 31129.0 tgs: 261 data_time: 0.75s time: 118.95s eta: 11:31:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31772.0 tgs: 267 data_time: 0.61s time: 118.91s eta: 11:29:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.197 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31673.0 tgs: 266 data_time: 0.49s time: 118.97s eta: 11:28:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 263 data_time: 0.88s time: 119.28s eta: 11:27:51 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31606.0 tgs: 264 data_time: 0.53s time: 119.35s eta: 11:26:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 268 data_time: 0.84s time: 118.88s eta: 11:21:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31600.0 tgs: 265 data_time: 0.84s time: 119.21s eta: 11:21:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.112 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31297.0 tgs: 261 data_time: 0.93s time: 119.72s eta: 11:22:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31170.0 tgs: 261 data_time: 0.77s time: 119.11s eta: 11:16:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31482.0 tgs: 264 data_time: 0.69s time: 119.12s eta: 11:15:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.191 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30453.0 tgs: 255 data_time: 0.85s time: 118.99s eta: 11:12:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31507.0 tgs: 264 data_time: 0.60s time: 119.09s eta: 11:10:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31876.0 tgs: 267 data_time: 0.67s time: 119.16s eta: 11:09:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31838.0 tgs: 267 data_time: 0.62s time: 118.98s eta: 11:06:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 268 data_time: 0.66s time: 119.46s eta: 11:06:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.198 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31482.0 tgs: 263 data_time: 0.74s time: 119.64s eta: 11:05:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.60s time: 119.31s eta: 11:02:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.116 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31347.0 tgs: 262 data_time: 0.83s time: 119.31s eta: 11:00:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30982.0 tgs: 260 data_time: 0.86s time: 118.82s eta: 10:55:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31975.0 tgs: 267 data_time: 0.68s time: 119.47s eta: 10:57:05 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.116 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31793.0 tgs: 266 data_time: 0.75s time: 119.37s eta: 10:54:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.116 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32152.0 tgs: 269 data_time: 0.72s time: 119.28s eta: 10:52:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31062.0 tgs: 260 data_time: 0.85s time: 119.26s eta: 10:49:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31041.0 tgs: 260 data_time: 0.82s time: 119.10s eta: 10:47:05 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 264 data_time: 0.69s time: 119.53s eta: 10:47:27 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31579.0 tgs: 264 data_time: 0.90s time: 119.32s eta: 10:44:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31677.0 tgs: 265 data_time: 0.86s time: 119.49s eta: 10:43:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31511.0 tgs: 263 data_time: 0.72s time: 119.51s eta: 10:41:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29543.0 tgs: 248 data_time: 0.68s time: 118.99s eta: 10:36:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30189.0 tgs: 253 data_time: 0.65s time: 119.21s eta: 10:35:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30801.0 tgs: 258 data_time: 0.95s time: 119.06s eta: 10:32:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.212 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32210.0 tgs: 271 data_time: 0.85s time: 118.78s eta: 10:29:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32167.0 tgs: 268 data_time: 0.83s time: 119.86s eta: 10:33:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32053.0 tgs: 268 data_time: 1.03s time: 119.38s eta: 10:28:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31628.0 tgs: 266 data_time: 0.76s time: 118.84s eta: 10:23:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32274.0 tgs: 271 data_time: 0.87s time: 118.98s eta: 10:22:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32070.0 tgs: 268 data_time: 0.70s time: 119.28s eta: 10:22:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30936.0 tgs: 259 data_time: 0.58s time: 119.20s eta: 10:19:51 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32494.0 tgs: 272 data_time: 0.81s time: 119.06s eta: 10:17:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31269.0 tgs: 261 data_time: 0.66s time: 119.47s eta: 10:17:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31428.0 tgs: 263 data_time: 0.69s time: 119.42s eta: 10:14:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30916.0 tgs: 258 data_time: 0.71s time: 119.45s eta: 10:13:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 267 data_time: 0.74s time: 119.29s eta: 10:10:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.83s time: 119.11s eta: 10:07:27 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31671.0 tgs: 264 data_time: 0.59s time: 119.70s eta: 10:08:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.227 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32361.0 tgs: 270 data_time: 0.79s time: 119.83s eta: 10:07:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.207 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32428.0 tgs: 266 data_time: 1.00s time: 121.55s eta: 10:13:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31400.0 tgs: 263 data_time: 0.56s time: 119.25s eta: 10:00:14 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.111 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 263 data_time: 0.67s time: 118.81s eta: 9:56:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32448.0 tgs: 271 data_time: 0.86s time: 119.56s eta: 9:57:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 264 data_time: 0.91s time: 119.70s eta: 9:56:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32542.0 tgs: 272 data_time: 0.77s time: 119.26s eta: 9:52:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30600.0 tgs: 256 data_time: 0.65s time: 119.33s eta: 9:50:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.64s time: 119.22s eta: 9:48:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31650.0 tgs: 261 data_time: 0.84s time: 120.89s eta: 9:54:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.241 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31218.0 tgs: 261 data_time: 0.97s time: 119.27s eta: 9:44:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.115 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 266 data_time: 0.67s time: 119.46s eta: 9:43:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31907.0 tgs: 266 data_time: 0.72s time: 119.55s eta: 9:41:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32075.0 tgs: 269 data_time: 0.97s time: 119.20s eta: 9:38:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32205.0 tgs: 269 data_time: 0.77s time: 119.48s eta: 9:37:27 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30930.0 tgs: 259 data_time: 0.57s time: 119.06s eta: 9:33:27 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.130 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.59s time: 119.24s eta: 9:32:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29752.0 tgs: 249 data_time: 0.62s time: 119.16s eta: 9:29:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31876.0 tgs: 267 data_time: 0.69s time: 119.22s eta: 9:28:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31801.0 tgs: 266 data_time: 0.57s time: 119.46s eta: 9:27:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.192 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31260.0 tgs: 262 data_time: 0.63s time: 118.98s eta: 9:23:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31605.0 tgs: 265 data_time: 0.87s time: 119.04s eta: 9:21:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31588.0 tgs: 264 data_time: 0.70s time: 119.52s eta: 9:21:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.229 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31223.0 tgs: 261 data_time: 0.51s time: 119.59s eta: 9:20:05 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32012.0 tgs: 268 data_time: 0.55s time: 119.12s eta: 9:15:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31644.0 tgs: 265 data_time: 0.81s time: 119.04s eta: 9:13:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31158.0 tgs: 262 data_time: 0.65s time: 118.88s eta: 9:10:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.185 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31746.0 tgs: 266 data_time: 0.85s time: 119.32s eta: 9:10:51 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.259 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31894.0 tgs: 267 data_time: 0.68s time: 119.09s eta: 9:07:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 271 data_time: 0.57s time: 118.87s eta: 9:04:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31094.0 tgs: 259 data_time: 0.68s time: 119.66s eta: 9:06:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.275 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31593.0 tgs: 264 data_time: 1.10s time: 119.43s eta: 9:03:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30466.0 tgs: 255 data_time: 0.69s time: 119.43s eta: 9:01:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31565.0 tgs: 265 data_time: 0.66s time: 118.74s eta: 8:56:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31369.0 tgs: 263 data_time: 0.80s time: 118.87s eta: 8:54:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.90s time: 119.33s eta: 8:55:00 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30029.0 tgs: 252 data_time: 0.57s time: 119.15s eta: 8:52:10 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31149.0 tgs: 261 data_time: 0.72s time: 119.27s eta: 8:50:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 265 data_time: 0.96s time: 120.47s eta: 8:54:05 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32245.0 tgs: 270 data_time: 0.78s time: 119.24s eta: 8:46:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32445.0 tgs: 272 data_time: 0.85s time: 119.08s eta: 8:43:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31196.0 tgs: 258 data_time: 0.81s time: 120.83s eta: 8:49:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.206 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31762.0 tgs: 264 data_time: 0.63s time: 120.19s eta: 8:44:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30153.0 tgs: 253 data_time: 0.76s time: 118.82s eta: 8:36:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31551.0 tgs: 264 data_time: 0.75s time: 119.35s eta: 8:37:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30947.0 tgs: 259 data_time: 0.77s time: 119.33s eta: 8:35:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31503.0 tgs: 264 data_time: 0.87s time: 119.05s eta: 8:31:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32191.0 tgs: 270 data_time: 0.89s time: 119.16s eta: 8:30:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.62s time: 119.04s eta: 8:27:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31155.0 tgs: 261 data_time: 0.70s time: 119.12s eta: 8:26:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32033.0 tgs: 268 data_time: 0.88s time: 119.24s eta: 8:24:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32125.0 tgs: 269 data_time: 0.69s time: 119.22s eta: 8:22:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31262.0 tgs: 261 data_time: 0.75s time: 119.38s eta: 8:21:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32097.0 tgs: 267 data_time: 0.76s time: 119.78s eta: 8:21:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31621.0 tgs: 265 data_time: 0.59s time: 119.20s eta: 8:16:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.243 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31605.0 tgs: 265 data_time: 0.74s time: 119.02s eta: 8:13:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.187 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32005.0 tgs: 266 data_time: 0.79s time: 119.90s eta: 8:15:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32455.0 tgs: 271 data_time: 1.00s time: 119.55s eta: 8:12:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 0.83s time: 119.59s eta: 8:10:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 265 data_time: 0.63s time: 119.00s eta: 8:05:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 258 data_time: 0.96s time: 119.84s eta: 8:07:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30951.0 tgs: 258 data_time: 0.71s time: 119.71s eta: 8:04:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.243 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32106.0 tgs: 268 data_time: 0.98s time: 119.40s eta: 8:01:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.196 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31828.0 tgs: 266 data_time: 0.72s time: 119.28s eta: 7:59:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 267 data_time: 0.63s time: 119.19s eta: 7:56:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32032.0 tgs: 268 data_time: 0.72s time: 119.19s eta: 7:54:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31258.0 tgs: 262 data_time: 0.83s time: 118.94s eta: 7:51:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.203 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31034.0 tgs: 260 data_time: 0.54s time: 118.97s eta: 7:49:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30465.0 tgs: 255 data_time: 0.87s time: 119.41s eta: 7:49:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31935.0 tgs: 268 data_time: 0.88s time: 119.03s eta: 7:46:13 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31936.0 tgs: 268 data_time: 0.67s time: 119.10s eta: 7:44:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31466.0 tgs: 262 data_time: 0.53s time: 120.01s eta: 7:46:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.174 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31646.0 tgs: 264 data_time: 0.60s time: 119.82s eta: 7:43:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31387.0 tgs: 263 data_time: 0.73s time: 119.17s eta: 7:38:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31444.0 tgs: 264 data_time: 0.69s time: 118.97s eta: 7:36:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.208 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31480.0 tgs: 263 data_time: 0.49s time: 119.32s eta: 7:35:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.208 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32372.0 tgs: 271 data_time: 0.77s time: 119.38s eta: 7:33:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32132.0 tgs: 269 data_time: 0.47s time: 119.06s eta: 7:30:27 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31514.0 tgs: 264 data_time: 0.56s time: 118.99s eta: 7:28:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32382.0 tgs: 271 data_time: 0.55s time: 119.24s eta: 7:27:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.188 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31432.0 tgs: 264 data_time: 0.58s time: 118.99s eta: 7:24:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 269 data_time: 0.75s time: 118.93s eta: 7:22:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31796.0 tgs: 267 data_time: 0.68s time: 119.08s eta: 7:20:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31676.0 tgs: 266 data_time: 0.99s time: 118.99s eta: 7:18:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32222.0 tgs: 270 data_time: 0.93s time: 119.03s eta: 7:16:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 265 data_time: 0.64s time: 119.27s eta: 7:15:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.197 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31589.0 tgs: 264 data_time: 0.68s time: 119.36s eta: 7:13:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.176 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32039.0 tgs: 269 data_time: 0.80s time: 118.94s eta: 7:10:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31522.0 tgs: 264 data_time: 0.77s time: 119.20s eta: 7:09:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31089.0 tgs: 259 data_time: 0.67s time: 119.58s eta: 7:08:30 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31095.0 tgs: 259 data_time: 0.98s time: 119.66s eta: 7:06:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32369.0 tgs: 272 data_time: 0.55s time: 118.95s eta: 7:02:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31698.0 tgs: 265 data_time: 0.83s time: 119.30s eta: 7:01:30 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31914.0 tgs: 267 data_time: 0.62s time: 119.52s eta: 7:00:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31587.0 tgs: 264 data_time: 0.85s time: 119.42s eta: 6:57:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.187 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31716.0 tgs: 266 data_time: 0.97s time: 118.91s eta: 6:54:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31985.0 tgs: 268 data_time: 0.75s time: 119.34s eta: 6:53:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31582.0 tgs: 265 data_time: 0.62s time: 118.94s eta: 6:50:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31564.0 tgs: 264 data_time: 0.96s time: 119.27s eta: 6:49:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 268 data_time: 0.74s time: 119.02s eta: 6:46:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 264 data_time: 0.67s time: 119.23s eta: 6:45:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32462.0 tgs: 271 data_time: 1.03s time: 119.40s eta: 6:43:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31962.0 tgs: 268 data_time: 0.86s time: 119.01s eta: 6:40:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.183 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 269 data_time: 0.85s time: 119.01s eta: 6:38:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 263 data_time: 0.50s time: 119.46s eta: 6:38:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31739.0 tgs: 267 data_time: 0.79s time: 118.80s eta: 6:34:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30881.0 tgs: 258 data_time: 0.59s time: 119.43s eta: 6:34:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 269 data_time: 0.50s time: 119.57s eta: 6:32:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.106 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30399.0 tgs: 254 data_time: 0.70s time: 119.38s eta: 6:29:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30654.0 tgs: 255 data_time: 0.52s time: 119.74s eta: 6:29:10 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.107 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32260.0 tgs: 270 data_time: 0.70s time: 119.36s eta: 6:25:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31937.0 tgs: 268 data_time: 0.80s time: 118.82s eta: 6:22:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32106.0 tgs: 269 data_time: 0.92s time: 119.01s eta: 6:20:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32135.0 tgs: 269 data_time: 0.57s time: 119.40s eta: 6:20:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.190 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31328.0 tgs: 262 data_time: 0.90s time: 119.34s eta: 6:17:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.111 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30605.0 tgs: 257 data_time: 0.79s time: 119.05s eta: 6:15:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.222 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31235.0 tgs: 261 data_time: 0.79s time: 119.42s eta: 6:14:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.195 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32203.0 tgs: 269 data_time: 0.96s time: 119.39s eta: 6:12:05 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30969.0 tgs: 259 data_time: 0.90s time: 119.28s eta: 6:09:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31576.0 tgs: 265 data_time: 0.72s time: 119.10s eta: 6:07:13 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.116 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31802.0 tgs: 260 data_time: 0.96s time: 121.96s eta: 6:14:00 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.207 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31619.0 tgs: 264 data_time: 0.86s time: 119.38s eta: 6:04:07 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31933.0 tgs: 267 data_time: 0.74s time: 119.57s eta: 6:02:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31671.0 tgs: 265 data_time: 0.78s time: 119.26s eta: 5:59:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 266 data_time: 0.60s time: 119.26s eta: 5:57:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.190 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31936.0 tgs: 268 data_time: 0.62s time: 118.91s eta: 5:54:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31657.0 tgs: 265 data_time: 0.62s time: 119.02s eta: 5:53:05 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 267 data_time: 0.97s time: 119.07s eta: 5:51:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31921.0 tgs: 267 data_time: 0.94s time: 119.28s eta: 5:49:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31483.0 tgs: 264 data_time: 0.74s time: 118.97s eta: 5:46:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31454.0 tgs: 263 data_time: 0.93s time: 119.27s eta: 5:45:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.196 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32096.0 tgs: 268 data_time: 0.83s time: 119.33s eta: 5:44:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31682.0 tgs: 266 data_time: 0.76s time: 119.05s eta: 5:41:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.098 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 267 data_time: 0.66s time: 119.50s eta: 5:40:33 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 269 data_time: 0.74s time: 119.38s eta: 5:38:14 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31233.0 tgs: 261 data_time: 0.59s time: 119.57s eta: 5:36:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31017.0 tgs: 260 data_time: 0.67s time: 118.98s eta: 5:33:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31538.0 tgs: 265 data_time: 0.66s time: 118.95s eta: 5:31:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31962.0 tgs: 268 data_time: 0.78s time: 119.24s eta: 5:29:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31291.0 tgs: 262 data_time: 0.68s time: 119.40s eta: 5:28:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32644.0 tgs: 273 data_time: 1.16s time: 119.46s eta: 5:26:30 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 270 data_time: 0.62s time: 118.95s eta: 5:23:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.199 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31459.0 tgs: 262 data_time: 0.72s time: 119.63s eta: 5:23:00 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31423.0 tgs: 263 data_time: 0.79s time: 119.35s eta: 5:20:14 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.181 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30816.0 tgs: 258 data_time: 0.87s time: 119.12s eta: 5:17:39 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31983.0 tgs: 267 data_time: 0.79s time: 119.42s eta: 5:16:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31080.0 tgs: 259 data_time: 0.52s time: 119.60s eta: 5:14:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31517.0 tgs: 264 data_time: 0.73s time: 119.01s eta: 5:11:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32371.0 tgs: 272 data_time: 0.87s time: 118.99s eta: 5:09:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30283.0 tgs: 254 data_time: 0.68s time: 119.22s eta: 5:07:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30915.0 tgs: 258 data_time: 0.62s time: 119.68s eta: 5:07:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31544.0 tgs: 264 data_time: 0.70s time: 119.11s eta: 5:03:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.242 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32336.0 tgs: 271 data_time: 0.88s time: 119.07s eta: 5:01:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32145.0 tgs: 264 data_time: 0.81s time: 121.52s eta: 5:05:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30136.0 tgs: 252 data_time: 0.54s time: 119.58s eta: 4:58:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 263 data_time: 0.74s time: 119.49s eta: 4:56:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30508.0 tgs: 254 data_time: 0.64s time: 119.73s eta: 4:55:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31563.0 tgs: 264 data_time: 0.89s time: 119.52s eta: 4:52:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32172.0 tgs: 269 data_time: 0.91s time: 119.41s eta: 4:50:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 265 data_time: 0.87s time: 119.20s eta: 4:48:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31912.0 tgs: 268 data_time: 0.56s time: 118.88s eta: 4:45:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 266 data_time: 0.81s time: 119.09s eta: 4:43:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31639.0 tgs: 263 data_time: 0.80s time: 119.98s eta: 4:43:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32225.0 tgs: 270 data_time: 0.63s time: 119.14s eta: 4:39:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32239.0 tgs: 270 data_time: 0.67s time: 119.01s eta: 4:37:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.190 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.66s time: 119.06s eta: 4:35:48 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31564.0 tgs: 264 data_time: 0.80s time: 119.37s eta: 4:34:33 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32125.0 tgs: 269 data_time: 1.00s time: 119.40s eta: 4:32:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 268 data_time: 0.89s time: 118.94s eta: 4:29:35 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31387.0 tgs: 262 data_time: 0.65s time: 119.60s eta: 4:29:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 268 data_time: 0.91s time: 119.24s eta: 4:26:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32064.0 tgs: 269 data_time: 0.73s time: 119.06s eta: 4:23:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32467.0 tgs: 273 data_time: 0.82s time: 118.85s eta: 4:21:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32273.0 tgs: 271 data_time: 0.80s time: 118.76s eta: 4:19:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30833.0 tgs: 258 data_time: 0.49s time: 119.42s eta: 4:18:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31721.0 tgs: 265 data_time: 0.61s time: 119.40s eta: 4:16:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32424.0 tgs: 271 data_time: 0.87s time: 119.35s eta: 4:14:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31769.0 tgs: 266 data_time: 0.75s time: 119.06s eta: 4:12:00 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32106.0 tgs: 269 data_time: 1.04s time: 119.11s eta: 4:10:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32249.0 tgs: 270 data_time: 0.57s time: 119.13s eta: 4:08:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30545.0 tgs: 256 data_time: 0.73s time: 118.96s eta: 4:05:51 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 264 data_time: 0.61s time: 119.15s eta: 4:04:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31905.0 tgs: 266 data_time: 0.59s time: 119.49s eta: 4:02:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30723.0 tgs: 257 data_time: 0.86s time: 119.24s eta: 4:00:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31644.0 tgs: 265 data_time: 0.64s time: 119.20s eta: 3:58:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31766.0 tgs: 266 data_time: 0.67s time: 119.35s eta: 3:56:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31665.0 tgs: 265 data_time: 0.63s time: 119.36s eta: 3:54:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30904.0 tgs: 257 data_time: 0.78s time: 119.95s eta: 3:53:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.187 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30883.0 tgs: 258 data_time: 0.76s time: 119.32s eta: 3:50:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 261 data_time: 0.80s time: 119.63s eta: 3:49:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30921.0 tgs: 260 data_time: 0.64s time: 118.86s eta: 3:45:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31583.0 tgs: 263 data_time: 0.98s time: 119.75s eta: 3:45:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31199.0 tgs: 261 data_time: 0.67s time: 119.32s eta: 3:42:44 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30747.0 tgs: 257 data_time: 0.98s time: 119.26s eta: 3:40:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.191 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31204.0 tgs: 262 data_time: 0.74s time: 119.00s eta: 3:38:09 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30579.0 tgs: 256 data_time: 0.71s time: 119.43s eta: 3:36:57 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31222.0 tgs: 262 data_time: 0.81s time: 119.03s eta: 3:34:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 266 data_time: 0.53s time: 118.99s eta: 3:32:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31853.0 tgs: 267 data_time: 0.91s time: 119.21s eta: 3:30:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31330.0 tgs: 260 data_time: 0.96s time: 120.11s eta: 3:30:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31220.0 tgs: 262 data_time: 0.54s time: 119.13s eta: 3:26:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30904.0 tgs: 259 data_time: 0.67s time: 119.32s eta: 3:24:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.107 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31972.0 tgs: 268 data_time: 0.65s time: 119.08s eta: 3:22:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31699.0 tgs: 265 data_time: 1.00s time: 119.43s eta: 3:21:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31700.0 tgs: 265 data_time: 0.60s time: 119.29s eta: 3:18:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.93s time: 119.22s eta: 3:16:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31205.0 tgs: 261 data_time: 0.74s time: 119.14s eta: 3:14:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31661.0 tgs: 266 data_time: 0.85s time: 119.02s eta: 3:12:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31989.0 tgs: 268 data_time: 0.81s time: 119.19s eta: 3:10:41 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 266 data_time: 0.62s time: 121.03s eta: 3:11:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31760.0 tgs: 265 data_time: 0.74s time: 119.50s eta: 3:07:13 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31694.0 tgs: 265 data_time: 0.85s time: 119.26s eta: 3:04:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 271 data_time: 0.61s time: 118.99s eta: 3:02:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32312.0 tgs: 271 data_time: 0.79s time: 119.22s eta: 3:00:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31782.0 tgs: 266 data_time: 0.67s time: 119.46s eta: 2:59:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30415.0 tgs: 255 data_time: 0.49s time: 118.98s eta: 2:56:28 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31745.0 tgs: 265 data_time: 0.81s time: 119.35s eta: 2:55:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31561.0 tgs: 264 data_time: 0.64s time: 119.41s eta: 2:53:08 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31598.0 tgs: 264 data_time: 0.75s time: 119.35s eta: 2:51:04 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.196 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32264.0 tgs: 265 data_time: 0.89s time: 121.46s eta: 2:52:03 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32490.0 tgs: 272 data_time: 0.71s time: 119.24s eta: 2:46:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31547.0 tgs: 263 data_time: 0.83s time: 119.60s eta: 2:45:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.109 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 267 data_time: 0.68s time: 118.92s eta: 2:42:31 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30961.0 tgs: 259 data_time: 0.71s time: 119.18s eta: 2:40:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31283.0 tgs: 262 data_time: 0.48s time: 118.96s eta: 2:38:37 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32085.0 tgs: 269 data_time: 0.70s time: 119.23s eta: 2:36:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31878.0 tgs: 267 data_time: 0.91s time: 119.10s eta: 2:34:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 267 data_time: 1.00s time: 119.43s eta: 2:33:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31257.0 tgs: 259 data_time: 0.82s time: 120.22s eta: 2:32:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31021.0 tgs: 260 data_time: 1.07s time: 119.28s eta: 2:29:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32284.0 tgs: 271 data_time: 0.83s time: 118.88s eta: 2:26:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32371.0 tgs: 270 data_time: 0.59s time: 119.52s eta: 2:25:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31823.0 tgs: 267 data_time: 0.70s time: 119.06s eta: 2:22:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32297.0 tgs: 271 data_time: 0.72s time: 119.14s eta: 2:20:58 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31475.0 tgs: 264 data_time: 0.58s time: 119.04s eta: 2:18:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32108.0 tgs: 267 data_time: 0.99s time: 119.84s eta: 2:17:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32158.0 tgs: 269 data_time: 0.72s time: 119.32s eta: 2:15:13 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31771.0 tgs: 266 data_time: 0.54s time: 119.39s eta: 2:13:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31079.0 tgs: 259 data_time: 0.72s time: 119.67s eta: 2:11:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32424.0 tgs: 271 data_time: 0.48s time: 119.50s eta: 2:09:27 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32482.0 tgs: 273 data_time: 0.66s time: 118.95s eta: 2:06:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32062.0 tgs: 268 data_time: 0.86s time: 119.37s eta: 2:05:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31877.0 tgs: 267 data_time: 0.80s time: 119.28s eta: 2:03:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31629.0 tgs: 265 data_time: 0.65s time: 119.22s eta: 2:01:12 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31103.0 tgs: 260 data_time: 0.81s time: 119.51s eta: 1:59:30 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31995.0 tgs: 267 data_time: 0.51s time: 119.42s eta: 1:57:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32402.0 tgs: 271 data_time: 0.77s time: 119.53s eta: 1:55:32 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31816.0 tgs: 266 data_time: 0.76s time: 119.30s eta: 1:53:20 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31753.0 tgs: 266 data_time: 0.76s time: 119.25s eta: 1:51:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32589.0 tgs: 274 data_time: 0.55s time: 118.93s eta: 1:49:01 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31868.0 tgs: 266 data_time: 0.69s time: 119.65s eta: 1:47:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32066.0 tgs: 267 data_time: 0.60s time: 119.72s eta: 1:45:45 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31708.0 tgs: 265 data_time: 0.61s time: 119.35s eta: 1:43:26 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32499.0 tgs: 272 data_time: 0.83s time: 119.30s eta: 1:41:24 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30551.0 tgs: 256 data_time: 0.91s time: 119.28s eta: 1:39:23 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 268 data_time: 0.82s time: 118.90s eta: 1:37:06 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.195 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 264 data_time: 1.04s time: 119.16s eta: 1:35:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30753.0 tgs: 258 data_time: 0.82s time: 119.06s eta: 1:33:15 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 271 data_time: 0.74s time: 119.25s eta: 1:31:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27985.0 tgs: 233 data_time: 0.76s time: 119.72s eta: 1:29:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32023.0 tgs: 269 data_time: 0.95s time: 119.01s eta: 1:27:16 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30952.0 tgs: 260 data_time: 0.80s time: 118.87s eta: 1:25:11 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32532.0 tgs: 272 data_time: 0.81s time: 119.48s eta: 1:23:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31867.0 tgs: 266 data_time: 0.72s time: 119.48s eta: 1:21:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31874.0 tgs: 267 data_time: 0.72s time: 118.94s eta: 1:19:17 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32024.0 tgs: 269 data_time: 0.77s time: 118.93s eta: 1:17:18 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30263.0 tgs: 247 data_time: 0.77s time: 122.15s eta: 1:17:21 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.209 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32358.0 tgs: 271 data_time: 0.62s time: 119.36s eta: 1:13:36 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31928.0 tgs: 268 data_time: 0.71s time: 119.05s eta: 1:11:25 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31158.0 tgs: 262 data_time: 0.61s time: 118.92s eta: 1:09:22 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 268 data_time: 0.79s time: 119.11s eta: 1:07:29 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 271 data_time: 0.83s time: 119.41s eta: 1:05:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32495.0 tgs: 272 data_time: 0.76s time: 119.34s eta: 1:03:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31124.0 tgs: 261 data_time: 0.83s time: 119.18s eta: 1:01:34 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31735.0 tgs: 266 data_time: 0.56s time: 119.27s eta: 0:59:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30877.0 tgs: 258 data_time: 0.65s time: 119.45s eta: 0:57:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 262 data_time: 0.65s time: 119.39s eta: 0:55:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31455.0 tgs: 263 data_time: 0.86s time: 119.49s eta: 0:53:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30600.0 tgs: 256 data_time: 0.73s time: 119.36s eta: 0:51:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 263 data_time: 0.68s time: 119.47s eta: 0:49:46 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31802.0 tgs: 267 data_time: 0.57s time: 118.89s eta: 0:47:33 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32526.0 tgs: 272 data_time: 0.64s time: 119.28s eta: 0:45:43 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.72s time: 119.21s eta: 0:43:42 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32089.0 tgs: 269 data_time: 0.60s time: 119.00s eta: 0:41:38 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31929.0 tgs: 268 data_time: 0.79s time: 119.03s eta: 0:39:40 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32223.0 tgs: 269 data_time: 0.49s time: 119.66s eta: 0:37:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31089.0 tgs: 260 data_time: 0.65s time: 119.29s eta: 0:35:47 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31516.0 tgs: 263 data_time: 0.88s time: 119.74s eta: 0:33:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31705.0 tgs: 261 data_time: 0.77s time: 121.22s eta: 0:32:19 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31997.0 tgs: 268 data_time: 0.72s time: 119.38s eta: 0:29:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31160.0 tgs: 261 data_time: 0.59s time: 119.32s eta: 0:27:50 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31162.0 tgs: 260 data_time: 0.55s time: 119.64s eta: 0:25:55 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32288.0 tgs: 271 data_time: 0.78s time: 119.09s eta: 0:23:49 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31744.0 tgs: 265 data_time: 0.81s time: 119.39s eta: 0:21:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31076.0 tgs: 260 data_time: 0.79s time: 119.26s eta: 0:19:52 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30722.0 tgs: 257 data_time: 0.68s time: 119.24s eta: 0:17:53 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32418.0 tgs: 270 data_time: 0.84s time: 119.90s eta: 0:15:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32228.0 tgs: 270 data_time: 0.91s time: 119.21s eta: 0:13:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31518.0 tgs: 264 data_time: 0.51s time: 119.09s eta: 0:11:54 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32296.0 tgs: 270 data_time: 0.54s time: 119.32s eta: 0:09:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32046.0 tgs: 268 data_time: 0.71s time: 119.16s eta: 0:07:56 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32146.0 tgs: 268 data_time: 0.64s time: 119.69s eta: 0:05:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31461.0 tgs: 259 data_time: 0.86s time: 121.40s eta: 0:04:02 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31883.0 tgs: 267 data_time: 0.76s time: 119.21s eta: 0:01:59 +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 13][DP 3][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank14.log b/20250121165312/rank14.log new file mode 100644 index 0000000000000000000000000000000000000000..9aa75a83699aad56467144b1fb6b8a29d6f4618d --- /dev/null +++ b/20250121165312/rank14.log @@ -0,0 +1,851 @@ +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 62.91 seconds, peak gpu memory 13.4G +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.255 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.3GB text_tokens: 30603.0 tgs: 202 data_time: 2.17s time: 151.29s eta: 1 day, 0:55:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.245 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31849.0 tgs: 265 data_time: 0.74s time: 120.11s eta: 19:45:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.325 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31059.0 tgs: 259 data_time: 1.03s time: 119.69s eta: 19:38:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32411.0 tgs: 270 data_time: 0.93s time: 119.64s eta: 19:36:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.235 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31204.0 tgs: 261 data_time: 0.76s time: 119.30s eta: 19:31:10 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.189 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31449.0 tgs: 263 data_time: 1.00s time: 119.54s eta: 19:31:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.257 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31590.0 tgs: 256 data_time: 0.89s time: 122.97s eta: 20:03:01 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.201 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31967.0 tgs: 266 data_time: 0.83s time: 119.88s eta: 19:30:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.280 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31084.0 tgs: 257 data_time: 1.02s time: 120.66s eta: 19:36:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31694.0 tgs: 264 data_time: 0.92s time: 119.92s eta: 19:27:10 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32132.0 tgs: 266 data_time: 0.96s time: 120.46s eta: 19:30:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31935.0 tgs: 267 data_time: 0.82s time: 119.35s eta: 19:17:39 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31576.0 tgs: 263 data_time: 0.89s time: 119.67s eta: 19:18:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31631.0 tgs: 265 data_time: 0.84s time: 119.20s eta: 19:12:14 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31904.0 tgs: 267 data_time: 0.91s time: 119.22s eta: 19:10:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32061.0 tgs: 269 data_time: 0.91s time: 118.94s eta: 19:05:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30810.0 tgs: 258 data_time: 0.81s time: 119.14s eta: 19:05:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31204.0 tgs: 261 data_time: 1.01s time: 119.17s eta: 19:04:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 264 data_time: 1.25s time: 119.41s eta: 19:04:21 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 267 data_time: 0.80s time: 119.32s eta: 19:01:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31859.0 tgs: 264 data_time: 0.71s time: 120.31s eta: 19:08:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.239 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32340.0 tgs: 263 data_time: 0.97s time: 122.91s eta: 19:31:44 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31453.0 tgs: 263 data_time: 0.53s time: 119.48s eta: 18:57:01 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31779.0 tgs: 266 data_time: 0.69s time: 119.16s eta: 18:51:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31169.0 tgs: 261 data_time: 0.90s time: 119.24s eta: 18:50:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29708.0 tgs: 248 data_time: 1.03s time: 119.40s eta: 18:50:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.123 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31114.0 tgs: 260 data_time: 1.11s time: 119.66s eta: 18:50:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32372.0 tgs: 271 data_time: 1.05s time: 119.38s eta: 18:46:10 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31554.0 tgs: 263 data_time: 1.08s time: 119.79s eta: 18:48:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.246 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31042.0 tgs: 240 data_time: 1.17s time: 129.31s eta: 20:15:31 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31728.0 tgs: 265 data_time: 0.90s time: 119.57s eta: 18:41:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31816.0 tgs: 266 data_time: 0.77s time: 119.39s eta: 18:38:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30775.0 tgs: 255 data_time: 0.70s time: 120.65s eta: 18:48:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31569.0 tgs: 264 data_time: 0.53s time: 119.40s eta: 18:34:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31161.0 tgs: 261 data_time: 0.77s time: 119.16s eta: 18:30:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.230 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31655.0 tgs: 265 data_time: 1.10s time: 119.08s eta: 18:27:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31369.0 tgs: 263 data_time: 0.84s time: 119.14s eta: 18:26:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32447.0 tgs: 269 data_time: 0.91s time: 120.39s eta: 18:35:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31783.0 tgs: 262 data_time: 1.03s time: 121.17s eta: 18:40:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.125 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30915.0 tgs: 256 data_time: 0.80s time: 120.66s eta: 18:34:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30941.0 tgs: 258 data_time: 0.81s time: 119.74s eta: 18:23:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31815.0 tgs: 256 data_time: 0.62s time: 123.87s eta: 18:59:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32295.0 tgs: 270 data_time: 0.74s time: 119.56s eta: 18:17:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 268 data_time: 0.71s time: 119.36s eta: 18:14:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32370.0 tgs: 271 data_time: 0.82s time: 119.23s eta: 18:10:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31953.0 tgs: 267 data_time: 0.56s time: 119.60s eta: 18:12:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31767.0 tgs: 266 data_time: 0.80s time: 119.26s eta: 18:07:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31980.0 tgs: 268 data_time: 0.64s time: 119.02s eta: 18:03:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31917.0 tgs: 268 data_time: 0.98s time: 119.09s eta: 18:01:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.252 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31824.0 tgs: 266 data_time: 0.66s time: 119.26s eta: 18:01:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30092.0 tgs: 252 data_time: 0.85s time: 119.33s eta: 17:59:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31783.0 tgs: 264 data_time: 0.74s time: 120.02s eta: 18:04:10 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30978.0 tgs: 260 data_time: 0.82s time: 119.11s eta: 17:53:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.131 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 266 data_time: 0.85s time: 119.41s eta: 17:54:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31322.0 tgs: 263 data_time: 1.08s time: 119.00s eta: 17:48:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31440.0 tgs: 264 data_time: 0.79s time: 119.02s eta: 17:47:14 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31790.0 tgs: 253 data_time: 0.93s time: 125.36s eta: 18:41:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32257.0 tgs: 268 data_time: 0.82s time: 120.10s eta: 17:52:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31778.0 tgs: 266 data_time: 0.74s time: 119.38s eta: 17:44:29 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.275 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31886.0 tgs: 267 data_time: 1.17s time: 119.33s eta: 17:42:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31703.0 tgs: 265 data_time: 0.86s time: 119.51s eta: 17:41:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31729.0 tgs: 251 data_time: 1.07s time: 126.05s eta: 18:37:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32208.0 tgs: 262 data_time: 0.90s time: 122.63s eta: 18:05:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31951.0 tgs: 263 data_time: 1.03s time: 121.15s eta: 17:50:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.128 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 264 data_time: 0.61s time: 119.03s eta: 17:29:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31597.0 tgs: 265 data_time: 0.82s time: 118.85s eta: 17:25:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 253 data_time: 0.84s time: 124.00s eta: 18:09:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 255 data_time: 0.96s time: 124.56s eta: 18:12:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32423.0 tgs: 271 data_time: 0.99s time: 119.39s eta: 17:24:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32055.0 tgs: 267 data_time: 1.12s time: 119.70s eta: 17:25:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31574.0 tgs: 264 data_time: 0.87s time: 119.52s eta: 17:21:51 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31739.0 tgs: 267 data_time: 0.71s time: 118.86s eta: 17:14:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30556.0 tgs: 254 data_time: 0.96s time: 119.85s eta: 17:20:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31430.0 tgs: 262 data_time: 0.82s time: 119.56s eta: 17:16:10 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32284.0 tgs: 264 data_time: 0.84s time: 121.86s eta: 17:34:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31470.0 tgs: 254 data_time: 0.74s time: 123.51s eta: 17:46:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30719.0 tgs: 258 data_time: 0.75s time: 119.01s eta: 17:05:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32126.0 tgs: 269 data_time: 0.79s time: 119.23s eta: 17:05:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32041.0 tgs: 268 data_time: 0.96s time: 119.14s eta: 17:02:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30639.0 tgs: 256 data_time: 0.98s time: 119.25s eta: 17:01:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30181.0 tgs: 253 data_time: 0.81s time: 119.04s eta: 16:57:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32047.0 tgs: 268 data_time: 1.01s time: 119.28s eta: 16:57:51 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32087.0 tgs: 269 data_time: 1.00s time: 119.15s eta: 16:54:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32110.0 tgs: 268 data_time: 1.09s time: 119.58s eta: 16:56:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31342.0 tgs: 263 data_time: 1.10s time: 119.14s eta: 16:50:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.250 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32177.0 tgs: 270 data_time: 0.86s time: 118.83s eta: 16:46:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.201 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32327.0 tgs: 271 data_time: 0.77s time: 119.01s eta: 16:45:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31900.0 tgs: 266 data_time: 1.08s time: 119.85s eta: 16:50:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 267 data_time: 0.91s time: 119.39s eta: 16:44:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31125.0 tgs: 261 data_time: 0.69s time: 119.01s eta: 16:39:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.224 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31485.0 tgs: 260 data_time: 1.00s time: 120.93s eta: 16:53:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30552.0 tgs: 256 data_time: 0.77s time: 118.98s eta: 16:35:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31858.0 tgs: 267 data_time: 0.86s time: 119.28s eta: 16:36:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31967.0 tgs: 268 data_time: 0.83s time: 119.23s eta: 16:33:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31505.0 tgs: 263 data_time: 0.80s time: 119.37s eta: 16:32:44 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32148.0 tgs: 268 data_time: 1.12s time: 119.71s eta: 16:33:33 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32093.0 tgs: 269 data_time: 0.55s time: 119.20s eta: 16:27:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30875.0 tgs: 257 data_time: 0.72s time: 119.76s eta: 16:30:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30629.0 tgs: 255 data_time: 0.66s time: 120.07s eta: 16:30:33 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31020.0 tgs: 260 data_time: 0.82s time: 118.97s eta: 16:19:31 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.121 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31810.0 tgs: 249 data_time: 1.05s time: 127.33s eta: 17:26:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.241 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31760.0 tgs: 266 data_time: 0.98s time: 119.23s eta: 16:17:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31604.0 tgs: 265 data_time: 0.89s time: 119.15s eta: 16:15:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.103 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31868.0 tgs: 267 data_time: 0.58s time: 119.24s eta: 16:13:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32248.0 tgs: 270 data_time: 0.82s time: 119.18s eta: 16:11:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31194.0 tgs: 262 data_time: 0.56s time: 118.94s eta: 16:07:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30582.0 tgs: 257 data_time: 0.62s time: 118.86s eta: 16:04:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30688.0 tgs: 257 data_time: 1.13s time: 118.96s eta: 16:03:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30944.0 tgs: 253 data_time: 0.64s time: 121.84s eta: 16:24:51 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 270 data_time: 0.96s time: 118.99s eta: 15:59:52 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31088.0 tgs: 260 data_time: 0.85s time: 119.23s eta: 15:59:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31484.0 tgs: 263 data_time: 0.87s time: 119.61s eta: 16:00:52 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31214.0 tgs: 262 data_time: 0.97s time: 119.05s eta: 15:54:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29982.0 tgs: 252 data_time: 0.94s time: 118.97s eta: 15:51:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30837.0 tgs: 258 data_time: 0.59s time: 119.17s eta: 15:51:21 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30478.0 tgs: 254 data_time: 0.87s time: 119.64s eta: 15:53:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31476.0 tgs: 264 data_time: 1.08s time: 118.94s eta: 15:45:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 261 data_time: 0.77s time: 119.28s eta: 15:46:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.76s time: 119.13s eta: 15:43:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.206 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30903.0 tgs: 259 data_time: 0.83s time: 119.00s eta: 15:40:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29636.0 tgs: 247 data_time: 0.90s time: 119.65s eta: 15:43:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31716.0 tgs: 261 data_time: 1.17s time: 121.50s eta: 15:55:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31642.0 tgs: 264 data_time: 0.61s time: 119.41s eta: 15:37:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31202.0 tgs: 240 data_time: 0.80s time: 129.96s eta: 16:58:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31172.0 tgs: 260 data_time: 1.05s time: 119.46s eta: 15:33:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31893.0 tgs: 266 data_time: 0.77s time: 119.61s eta: 15:32:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30988.0 tgs: 259 data_time: 1.29s time: 119.52s eta: 15:30:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31862.0 tgs: 260 data_time: 0.73s time: 122.14s eta: 15:48:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31325.0 tgs: 259 data_time: 0.80s time: 120.75s eta: 15:35:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 267 data_time: 0.90s time: 119.41s eta: 15:23:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31539.0 tgs: 264 data_time: 0.97s time: 119.44s eta: 15:21:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31684.0 tgs: 240 data_time: 0.76s time: 131.63s eta: 16:53:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31324.0 tgs: 262 data_time: 0.71s time: 119.30s eta: 15:16:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31849.0 tgs: 266 data_time: 0.94s time: 119.56s eta: 15:16:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 260 data_time: 0.78s time: 122.25s eta: 15:35:13 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 265 data_time: 0.69s time: 119.71s eta: 15:13:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31483.0 tgs: 263 data_time: 0.71s time: 119.35s eta: 15:09:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.205 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 270 data_time: 1.11s time: 118.89s eta: 15:03:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32377.0 tgs: 272 data_time: 0.93s time: 118.99s eta: 15:02:21 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31437.0 tgs: 263 data_time: 0.81s time: 119.31s eta: 15:02:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31680.0 tgs: 266 data_time: 0.99s time: 118.92s eta: 14:57:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.184 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31893.0 tgs: 264 data_time: 1.17s time: 120.52s eta: 15:07:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31817.0 tgs: 266 data_time: 0.98s time: 119.31s eta: 14:56:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30697.0 tgs: 250 data_time: 0.90s time: 122.66s eta: 15:19:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32293.0 tgs: 271 data_time: 0.73s time: 119.12s eta: 14:51:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31577.0 tgs: 263 data_time: 1.23s time: 119.75s eta: 14:54:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 269 data_time: 0.82s time: 119.46s eta: 14:50:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30603.0 tgs: 257 data_time: 0.67s time: 118.99s eta: 14:44:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30939.0 tgs: 259 data_time: 0.70s time: 119.32s eta: 14:44:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31827.0 tgs: 267 data_time: 0.63s time: 118.94s eta: 14:40:10 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.126 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31853.0 tgs: 267 data_time: 0.95s time: 119.05s eta: 14:38:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32231.0 tgs: 270 data_time: 0.72s time: 119.07s eta: 14:37:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.099 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32310.0 tgs: 271 data_time: 0.89s time: 119.12s eta: 14:35:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31567.0 tgs: 265 data_time: 0.86s time: 119.00s eta: 14:32:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32220.0 tgs: 270 data_time: 0.98s time: 119.07s eta: 14:31:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31910.0 tgs: 267 data_time: 0.81s time: 119.15s eta: 14:29:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.127 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31826.0 tgs: 265 data_time: 1.07s time: 119.65s eta: 14:31:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31857.0 tgs: 267 data_time: 0.82s time: 119.18s eta: 14:26:01 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31576.0 tgs: 264 data_time: 1.00s time: 119.20s eta: 14:24:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31063.0 tgs: 260 data_time: 0.95s time: 119.39s eta: 14:23:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31593.0 tgs: 261 data_time: 0.77s time: 120.73s eta: 14:31:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31829.0 tgs: 267 data_time: 0.81s time: 118.85s eta: 14:15:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.250 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31724.0 tgs: 266 data_time: 1.02s time: 119.10s eta: 14:15:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31084.0 tgs: 260 data_time: 1.05s time: 119.15s eta: 14:13:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.237 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 265 data_time: 0.95s time: 119.96s eta: 14:17:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31581.0 tgs: 263 data_time: 1.00s time: 119.87s eta: 14:15:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31183.0 tgs: 260 data_time: 0.64s time: 119.58s eta: 14:11:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31992.0 tgs: 268 data_time: 0.85s time: 119.22s eta: 14:06:29 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.195 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32059.0 tgs: 268 data_time: 0.75s time: 119.26s eta: 14:04:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31271.0 tgs: 262 data_time: 0.83s time: 119.26s eta: 14:02:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31487.0 tgs: 264 data_time: 0.81s time: 119.14s eta: 13:59:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 267 data_time: 0.85s time: 119.12s eta: 13:57:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31802.0 tgs: 266 data_time: 1.07s time: 119.17s eta: 13:56:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 1.18s time: 119.35s eta: 13:55:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31401.0 tgs: 261 data_time: 0.65s time: 120.11s eta: 13:58:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31724.0 tgs: 258 data_time: 1.06s time: 122.55s eta: 14:13:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31083.0 tgs: 260 data_time: 0.70s time: 119.10s eta: 13:47:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30902.0 tgs: 258 data_time: 0.78s time: 119.37s eta: 13:47:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30470.0 tgs: 255 data_time: 0.87s time: 119.36s eta: 13:45:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32119.0 tgs: 269 data_time: 0.86s time: 119.32s eta: 13:43:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32163.0 tgs: 270 data_time: 0.72s time: 119.12s eta: 13:39:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 266 data_time: 0.88s time: 119.25s eta: 13:38:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31339.0 tgs: 263 data_time: 0.94s time: 119.13s eta: 13:36:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30960.0 tgs: 259 data_time: 0.64s time: 119.22s eta: 13:34:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 265 data_time: 0.95s time: 119.46s eta: 13:34:18 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31385.0 tgs: 263 data_time: 0.80s time: 119.08s eta: 13:29:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32288.0 tgs: 271 data_time: 0.76s time: 118.93s eta: 13:26:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30856.0 tgs: 257 data_time: 0.69s time: 119.81s eta: 13:30:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30916.0 tgs: 259 data_time: 0.73s time: 119.26s eta: 13:24:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31858.0 tgs: 267 data_time: 0.76s time: 119.16s eta: 13:22:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32383.0 tgs: 271 data_time: 1.19s time: 119.34s eta: 13:21:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 270 data_time: 0.79s time: 119.26s eta: 13:19:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.208 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 269 data_time: 0.88s time: 119.04s eta: 13:15:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31814.0 tgs: 266 data_time: 0.86s time: 119.37s eta: 13:15:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32253.0 tgs: 270 data_time: 0.80s time: 119.19s eta: 13:12:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.184 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31833.0 tgs: 266 data_time: 0.89s time: 119.45s eta: 13:12:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31667.0 tgs: 265 data_time: 0.83s time: 119.29s eta: 13:09:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31809.0 tgs: 266 data_time: 0.79s time: 119.15s eta: 13:06:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31247.0 tgs: 259 data_time: 1.04s time: 120.21s eta: 13:11:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.205 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31773.0 tgs: 264 data_time: 0.94s time: 120.16s eta: 13:09:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 271 data_time: 1.09s time: 119.49s eta: 13:02:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32284.0 tgs: 270 data_time: 0.92s time: 119.51s eta: 13:00:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31185.0 tgs: 260 data_time: 0.66s time: 119.57s eta: 12:59:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31665.0 tgs: 264 data_time: 0.89s time: 119.80s eta: 12:58:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.108 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31009.0 tgs: 260 data_time: 0.70s time: 119.04s eta: 12:51:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31824.0 tgs: 267 data_time: 0.84s time: 119.05s eta: 12:49:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31602.0 tgs: 265 data_time: 0.77s time: 119.13s eta: 12:48:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32357.0 tgs: 271 data_time: 0.90s time: 119.39s eta: 12:48:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31796.0 tgs: 266 data_time: 0.88s time: 119.19s eta: 12:44:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31182.0 tgs: 261 data_time: 0.87s time: 119.43s eta: 12:44:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31421.0 tgs: 263 data_time: 0.72s time: 119.09s eta: 12:40:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.237 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 261 data_time: 0.58s time: 119.07s eta: 12:38:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31440.0 tgs: 262 data_time: 0.99s time: 119.59s eta: 12:39:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.218 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32221.0 tgs: 271 data_time: 0.89s time: 118.78s eta: 12:32:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31762.0 tgs: 267 data_time: 0.83s time: 118.88s eta: 12:30:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30944.0 tgs: 259 data_time: 0.90s time: 119.17s eta: 12:30:44 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 266 data_time: 0.81s time: 119.28s eta: 12:29:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.221 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31841.0 tgs: 266 data_time: 0.93s time: 119.34s eta: 12:27:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31705.0 tgs: 266 data_time: 0.84s time: 119.11s eta: 12:24:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.191 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31629.0 tgs: 265 data_time: 0.88s time: 119.34s eta: 12:23:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31282.0 tgs: 262 data_time: 0.68s time: 119.26s eta: 12:21:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31980.0 tgs: 268 data_time: 0.83s time: 119.02s eta: 12:17:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.206 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31842.0 tgs: 262 data_time: 0.74s time: 121.13s eta: 12:28:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31831.0 tgs: 265 data_time: 0.98s time: 119.99s eta: 12:19:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 264 data_time: 0.83s time: 118.72s eta: 12:10:08 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31904.0 tgs: 266 data_time: 0.66s time: 119.63s eta: 12:13:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31947.0 tgs: 266 data_time: 0.84s time: 119.83s eta: 12:12:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30644.0 tgs: 257 data_time: 0.74s time: 119.14s eta: 12:06:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31390.0 tgs: 263 data_time: 1.04s time: 119.17s eta: 12:04:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31549.0 tgs: 264 data_time: 1.06s time: 119.12s eta: 12:02:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 267 data_time: 0.62s time: 119.18s eta: 12:01:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32601.0 tgs: 274 data_time: 0.92s time: 118.95s eta: 11:57:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 266 data_time: 0.98s time: 119.71s eta: 12:00:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 269 data_time: 0.99s time: 119.27s eta: 11:55:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.104 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31569.0 tgs: 264 data_time: 0.81s time: 119.27s eta: 11:53:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.131 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31690.0 tgs: 264 data_time: 0.70s time: 119.71s eta: 11:54:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 272 data_time: 0.87s time: 118.78s eta: 11:46:44 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31409.0 tgs: 263 data_time: 0.71s time: 119.15s eta: 11:46:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31947.0 tgs: 267 data_time: 0.96s time: 119.61s eta: 11:47:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31149.0 tgs: 262 data_time: 0.90s time: 118.86s eta: 11:41:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31203.0 tgs: 262 data_time: 0.69s time: 118.68s eta: 11:38:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31912.0 tgs: 267 data_time: 0.88s time: 119.21s eta: 11:39:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.217 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 271 data_time: 0.99s time: 119.14s eta: 11:36:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31360.0 tgs: 263 data_time: 1.02s time: 119.07s eta: 11:34:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.168 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 31129.0 tgs: 261 data_time: 0.87s time: 118.95s eta: 11:31:52 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31772.0 tgs: 267 data_time: 0.67s time: 118.91s eta: 11:29:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.189 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31673.0 tgs: 266 data_time: 0.52s time: 118.97s eta: 11:28:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 263 data_time: 0.99s time: 119.28s eta: 11:27:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31606.0 tgs: 264 data_time: 0.61s time: 119.35s eta: 11:26:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 268 data_time: 0.92s time: 118.88s eta: 11:21:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31600.0 tgs: 265 data_time: 0.94s time: 119.20s eta: 11:21:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31297.0 tgs: 261 data_time: 0.97s time: 119.71s eta: 11:22:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.195 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31170.0 tgs: 261 data_time: 0.81s time: 119.12s eta: 11:16:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31482.0 tgs: 264 data_time: 0.73s time: 119.12s eta: 11:14:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.117 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30453.0 tgs: 255 data_time: 0.95s time: 118.99s eta: 11:12:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31507.0 tgs: 264 data_time: 0.66s time: 119.09s eta: 11:10:52 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31876.0 tgs: 267 data_time: 0.73s time: 119.16s eta: 11:09:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.229 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31838.0 tgs: 267 data_time: 0.66s time: 118.97s eta: 11:06:13 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 268 data_time: 0.77s time: 119.47s eta: 11:07:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31482.0 tgs: 263 data_time: 0.81s time: 119.64s eta: 11:06:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.64s time: 119.30s eta: 11:02:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.224 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31347.0 tgs: 262 data_time: 0.95s time: 119.32s eta: 11:00:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30982.0 tgs: 260 data_time: 0.90s time: 118.81s eta: 10:55:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31975.0 tgs: 267 data_time: 0.79s time: 119.47s eta: 10:57:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.210 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31793.0 tgs: 266 data_time: 0.78s time: 119.37s eta: 10:54:31 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32152.0 tgs: 269 data_time: 0.75s time: 119.28s eta: 10:52:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31062.0 tgs: 260 data_time: 0.92s time: 119.25s eta: 10:49:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31041.0 tgs: 260 data_time: 0.88s time: 119.10s eta: 10:47:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 264 data_time: 0.76s time: 119.53s eta: 10:47:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31579.0 tgs: 264 data_time: 1.02s time: 119.32s eta: 10:44:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31677.0 tgs: 265 data_time: 0.96s time: 119.49s eta: 10:43:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.196 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31511.0 tgs: 263 data_time: 0.81s time: 119.51s eta: 10:41:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29543.0 tgs: 248 data_time: 0.77s time: 118.99s eta: 10:36:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30189.0 tgs: 253 data_time: 0.71s time: 119.21s eta: 10:35:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30801.0 tgs: 258 data_time: 1.06s time: 119.06s eta: 10:32:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.185 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32210.0 tgs: 271 data_time: 0.96s time: 118.78s eta: 10:29:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.203 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32167.0 tgs: 268 data_time: 0.96s time: 119.86s eta: 10:33:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32053.0 tgs: 268 data_time: 1.19s time: 119.38s eta: 10:28:44 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.231 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31628.0 tgs: 266 data_time: 0.87s time: 118.84s eta: 10:23:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32274.0 tgs: 271 data_time: 1.02s time: 118.99s eta: 10:22:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32070.0 tgs: 268 data_time: 0.77s time: 119.28s eta: 10:22:14 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30936.0 tgs: 259 data_time: 0.65s time: 119.20s eta: 10:19:51 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32494.0 tgs: 272 data_time: 0.92s time: 119.06s eta: 10:17:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31269.0 tgs: 261 data_time: 0.73s time: 119.47s eta: 10:17:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.232 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31428.0 tgs: 263 data_time: 0.77s time: 119.42s eta: 10:14:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30916.0 tgs: 258 data_time: 0.80s time: 119.45s eta: 10:13:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 267 data_time: 0.87s time: 119.29s eta: 10:10:21 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.89s time: 119.11s eta: 10:07:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31671.0 tgs: 264 data_time: 0.65s time: 119.70s eta: 10:08:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32361.0 tgs: 270 data_time: 0.96s time: 119.83s eta: 10:07:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.280 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32428.0 tgs: 266 data_time: 1.09s time: 121.55s eta: 10:13:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.121 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31400.0 tgs: 263 data_time: 0.63s time: 119.26s eta: 10:00:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 263 data_time: 0.74s time: 118.80s eta: 9:55:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32448.0 tgs: 271 data_time: 0.96s time: 119.56s eta: 9:57:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 264 data_time: 1.01s time: 119.70s eta: 9:56:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32542.0 tgs: 272 data_time: 0.86s time: 119.26s eta: 9:52:18 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30600.0 tgs: 256 data_time: 0.74s time: 119.33s eta: 9:50:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.68s time: 119.22s eta: 9:48:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31650.0 tgs: 261 data_time: 0.93s time: 120.88s eta: 9:54:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.188 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31218.0 tgs: 261 data_time: 1.01s time: 119.27s eta: 9:44:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.206 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 266 data_time: 0.74s time: 119.46s eta: 9:43:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31907.0 tgs: 266 data_time: 0.85s time: 119.55s eta: 9:41:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.181 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32075.0 tgs: 269 data_time: 1.11s time: 119.20s eta: 9:38:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.159 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32205.0 tgs: 269 data_time: 0.83s time: 119.48s eta: 9:37:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30930.0 tgs: 259 data_time: 0.63s time: 119.05s eta: 9:33:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.63s time: 119.24s eta: 9:32:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29752.0 tgs: 249 data_time: 0.70s time: 119.16s eta: 9:29:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.232 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31876.0 tgs: 267 data_time: 0.72s time: 119.22s eta: 9:28:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.196 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31801.0 tgs: 266 data_time: 0.66s time: 119.46s eta: 9:27:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.109 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31260.0 tgs: 262 data_time: 0.71s time: 118.99s eta: 9:23:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.201 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31605.0 tgs: 265 data_time: 0.95s time: 119.03s eta: 9:21:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31588.0 tgs: 264 data_time: 0.76s time: 119.53s eta: 9:21:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31223.0 tgs: 261 data_time: 0.59s time: 119.59s eta: 9:20:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.121 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32012.0 tgs: 268 data_time: 0.70s time: 119.12s eta: 9:15:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.116 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31644.0 tgs: 265 data_time: 0.85s time: 119.04s eta: 9:13:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31158.0 tgs: 262 data_time: 0.67s time: 118.88s eta: 9:10:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.159 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31746.0 tgs: 266 data_time: 0.95s time: 119.32s eta: 9:10:51 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31894.0 tgs: 267 data_time: 0.74s time: 119.09s eta: 9:07:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 271 data_time: 0.64s time: 118.86s eta: 9:04:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.105 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31094.0 tgs: 259 data_time: 0.86s time: 119.67s eta: 9:06:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.230 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31593.0 tgs: 264 data_time: 1.19s time: 119.43s eta: 9:03:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.218 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30466.0 tgs: 255 data_time: 0.78s time: 119.44s eta: 9:01:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.115 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31565.0 tgs: 265 data_time: 0.71s time: 118.74s eta: 8:56:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.189 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31369.0 tgs: 263 data_time: 0.88s time: 118.87s eta: 8:54:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 1.02s time: 119.33s eta: 8:55:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30029.0 tgs: 252 data_time: 0.60s time: 119.15s eta: 8:52:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31149.0 tgs: 261 data_time: 0.80s time: 119.27s eta: 8:50:44 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.211 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 265 data_time: 1.07s time: 120.47s eta: 8:54:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.212 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32245.0 tgs: 270 data_time: 0.90s time: 119.24s eta: 8:46:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.189 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32445.0 tgs: 272 data_time: 0.92s time: 119.08s eta: 8:43:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31196.0 tgs: 258 data_time: 0.89s time: 120.83s eta: 8:49:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31762.0 tgs: 264 data_time: 0.69s time: 120.19s eta: 8:44:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30153.0 tgs: 253 data_time: 0.82s time: 118.82s eta: 8:36:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31551.0 tgs: 264 data_time: 0.81s time: 119.36s eta: 8:37:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30947.0 tgs: 259 data_time: 0.87s time: 119.33s eta: 8:35:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31503.0 tgs: 264 data_time: 1.00s time: 119.05s eta: 8:31:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.119 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32191.0 tgs: 270 data_time: 1.00s time: 119.16s eta: 8:30:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.69s time: 119.04s eta: 8:27:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31155.0 tgs: 261 data_time: 0.72s time: 119.12s eta: 8:26:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32033.0 tgs: 268 data_time: 0.98s time: 119.24s eta: 8:24:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32125.0 tgs: 269 data_time: 0.80s time: 119.22s eta: 8:22:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31262.0 tgs: 261 data_time: 0.85s time: 119.38s eta: 8:21:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32097.0 tgs: 267 data_time: 0.98s time: 119.78s eta: 8:21:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31621.0 tgs: 265 data_time: 0.65s time: 119.20s eta: 8:16:39 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.114 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31605.0 tgs: 265 data_time: 0.84s time: 119.02s eta: 8:13:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32005.0 tgs: 266 data_time: 0.90s time: 119.90s eta: 8:15:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32455.0 tgs: 271 data_time: 1.11s time: 119.55s eta: 8:12:08 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 0.90s time: 119.59s eta: 8:10:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.106 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 265 data_time: 0.71s time: 119.00s eta: 8:05:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.198 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 258 data_time: 1.03s time: 119.84s eta: 8:07:21 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30951.0 tgs: 258 data_time: 0.81s time: 119.71s eta: 8:04:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.190 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32106.0 tgs: 268 data_time: 1.08s time: 119.39s eta: 8:01:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31828.0 tgs: 266 data_time: 0.81s time: 119.28s eta: 7:59:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 267 data_time: 0.71s time: 119.19s eta: 7:56:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32032.0 tgs: 268 data_time: 0.82s time: 119.21s eta: 7:54:51 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.195 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31258.0 tgs: 262 data_time: 0.93s time: 118.92s eta: 7:51:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31034.0 tgs: 260 data_time: 0.58s time: 118.97s eta: 7:49:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30465.0 tgs: 255 data_time: 0.97s time: 119.41s eta: 7:49:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31935.0 tgs: 268 data_time: 0.94s time: 119.03s eta: 7:46:13 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31936.0 tgs: 268 data_time: 0.80s time: 119.10s eta: 7:44:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31466.0 tgs: 262 data_time: 0.59s time: 120.01s eta: 7:46:01 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31646.0 tgs: 264 data_time: 0.68s time: 119.82s eta: 7:43:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31387.0 tgs: 263 data_time: 0.83s time: 119.17s eta: 7:38:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31444.0 tgs: 264 data_time: 0.80s time: 118.97s eta: 7:36:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.162 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31480.0 tgs: 263 data_time: 0.54s time: 119.32s eta: 7:35:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.174 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32372.0 tgs: 271 data_time: 0.86s time: 119.38s eta: 7:33:39 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.189 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32132.0 tgs: 269 data_time: 0.52s time: 119.06s eta: 7:30:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31514.0 tgs: 264 data_time: 0.61s time: 118.99s eta: 7:28:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32382.0 tgs: 271 data_time: 0.60s time: 119.24s eta: 7:27:08 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31432.0 tgs: 264 data_time: 0.63s time: 118.99s eta: 7:24:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 269 data_time: 0.89s time: 118.93s eta: 7:22:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31796.0 tgs: 267 data_time: 0.77s time: 119.08s eta: 7:20:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31676.0 tgs: 266 data_time: 1.01s time: 118.99s eta: 7:18:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32222.0 tgs: 270 data_time: 0.97s time: 119.03s eta: 7:16:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.181 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 265 data_time: 0.81s time: 119.27s eta: 7:15:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.219 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31589.0 tgs: 264 data_time: 0.74s time: 119.36s eta: 7:13:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32039.0 tgs: 269 data_time: 0.93s time: 118.94s eta: 7:10:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.252 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31522.0 tgs: 264 data_time: 0.80s time: 119.20s eta: 7:09:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31089.0 tgs: 259 data_time: 0.74s time: 119.58s eta: 7:08:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31095.0 tgs: 259 data_time: 1.01s time: 119.66s eta: 7:06:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32369.0 tgs: 272 data_time: 0.57s time: 118.95s eta: 7:02:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31698.0 tgs: 265 data_time: 0.89s time: 119.30s eta: 7:01:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31914.0 tgs: 267 data_time: 0.75s time: 119.52s eta: 7:00:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31587.0 tgs: 264 data_time: 0.88s time: 119.42s eta: 6:57:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31716.0 tgs: 266 data_time: 1.05s time: 118.91s eta: 6:54:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.197 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31985.0 tgs: 268 data_time: 0.84s time: 119.34s eta: 6:53:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.176 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31582.0 tgs: 265 data_time: 0.72s time: 118.94s eta: 6:50:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.179 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31564.0 tgs: 264 data_time: 1.16s time: 119.27s eta: 6:49:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 268 data_time: 0.82s time: 119.02s eta: 6:46:39 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.098 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 264 data_time: 0.70s time: 119.23s eta: 6:45:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32462.0 tgs: 271 data_time: 1.10s time: 119.40s eta: 6:43:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31962.0 tgs: 268 data_time: 0.99s time: 119.01s eta: 6:40:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 269 data_time: 1.00s time: 119.01s eta: 6:38:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 263 data_time: 0.57s time: 119.46s eta: 6:38:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31739.0 tgs: 267 data_time: 0.95s time: 118.80s eta: 6:34:01 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30881.0 tgs: 258 data_time: 0.67s time: 119.43s eta: 6:34:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.115 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 269 data_time: 0.55s time: 119.57s eta: 6:32:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30399.0 tgs: 254 data_time: 0.77s time: 119.38s eta: 6:29:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30654.0 tgs: 256 data_time: 0.60s time: 119.74s eta: 6:29:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32260.0 tgs: 270 data_time: 0.75s time: 119.36s eta: 6:25:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31937.0 tgs: 268 data_time: 0.91s time: 118.82s eta: 6:22:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32106.0 tgs: 269 data_time: 0.96s time: 119.01s eta: 6:20:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.109 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32135.0 tgs: 269 data_time: 0.65s time: 119.40s eta: 6:20:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31328.0 tgs: 262 data_time: 1.01s time: 119.34s eta: 6:17:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30605.0 tgs: 257 data_time: 0.98s time: 119.05s eta: 6:15:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31235.0 tgs: 261 data_time: 0.87s time: 119.42s eta: 6:14:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32203.0 tgs: 269 data_time: 1.12s time: 119.39s eta: 6:12:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30969.0 tgs: 259 data_time: 0.95s time: 119.28s eta: 6:09:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31576.0 tgs: 265 data_time: 0.82s time: 119.10s eta: 6:07:13 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31802.0 tgs: 260 data_time: 1.08s time: 121.96s eta: 6:14:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31619.0 tgs: 264 data_time: 0.95s time: 119.38s eta: 6:04:07 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.106 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31933.0 tgs: 267 data_time: 0.85s time: 119.57s eta: 6:02:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31671.0 tgs: 265 data_time: 0.88s time: 119.26s eta: 5:59:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 266 data_time: 0.70s time: 119.26s eta: 5:57:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31936.0 tgs: 268 data_time: 0.67s time: 118.91s eta: 5:54:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31657.0 tgs: 265 data_time: 0.71s time: 119.02s eta: 5:53:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 267 data_time: 1.07s time: 119.07s eta: 5:51:14 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31921.0 tgs: 267 data_time: 1.07s time: 119.29s eta: 5:49:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31483.0 tgs: 264 data_time: 0.79s time: 118.98s eta: 5:47:01 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31454.0 tgs: 263 data_time: 0.99s time: 119.26s eta: 5:45:51 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.196 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32096.0 tgs: 268 data_time: 0.92s time: 119.33s eta: 5:44:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.198 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31682.0 tgs: 266 data_time: 0.88s time: 119.05s eta: 5:41:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.115 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 267 data_time: 0.75s time: 119.50s eta: 5:40:33 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 269 data_time: 0.76s time: 119.38s eta: 5:38:14 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31233.0 tgs: 261 data_time: 0.65s time: 119.57s eta: 5:36:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31017.0 tgs: 260 data_time: 0.75s time: 118.98s eta: 5:33:08 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.122 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31538.0 tgs: 265 data_time: 0.69s time: 118.95s eta: 5:31:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31962.0 tgs: 268 data_time: 0.80s time: 119.24s eta: 5:29:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.093 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31291.0 tgs: 262 data_time: 0.72s time: 119.40s eta: 5:28:21 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32644.0 tgs: 273 data_time: 1.20s time: 119.46s eta: 5:26:31 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 270 data_time: 0.64s time: 118.95s eta: 5:23:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.239 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31459.0 tgs: 262 data_time: 0.82s time: 119.63s eta: 5:22:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31423.0 tgs: 263 data_time: 0.93s time: 119.35s eta: 5:20:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30816.0 tgs: 258 data_time: 0.98s time: 119.12s eta: 5:17:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31983.0 tgs: 267 data_time: 0.94s time: 119.42s eta: 5:16:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31080.0 tgs: 259 data_time: 0.60s time: 119.60s eta: 5:14:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31517.0 tgs: 264 data_time: 0.80s time: 119.00s eta: 5:11:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32371.0 tgs: 272 data_time: 0.94s time: 119.00s eta: 5:09:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30283.0 tgs: 254 data_time: 0.78s time: 119.22s eta: 5:07:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30915.0 tgs: 258 data_time: 0.74s time: 119.70s eta: 5:07:13 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.218 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31544.0 tgs: 264 data_time: 0.82s time: 119.11s eta: 5:03:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32336.0 tgs: 271 data_time: 0.98s time: 119.07s eta: 5:01:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32145.0 tgs: 264 data_time: 0.93s time: 121.52s eta: 5:05:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30136.0 tgs: 252 data_time: 0.60s time: 119.57s eta: 4:58:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 263 data_time: 0.85s time: 119.50s eta: 4:56:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.189 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30508.0 tgs: 254 data_time: 0.65s time: 119.73s eta: 4:55:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31563.0 tgs: 264 data_time: 0.92s time: 119.52s eta: 4:52:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32172.0 tgs: 269 data_time: 1.00s time: 119.42s eta: 4:50:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 265 data_time: 0.99s time: 119.20s eta: 4:48:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31912.0 tgs: 268 data_time: 0.64s time: 118.88s eta: 4:45:18 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 266 data_time: 0.84s time: 119.09s eta: 4:43:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.181 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31639.0 tgs: 263 data_time: 0.82s time: 119.98s eta: 4:43:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32225.0 tgs: 270 data_time: 0.73s time: 119.14s eta: 4:39:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32239.0 tgs: 270 data_time: 0.70s time: 119.01s eta: 4:37:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.69s time: 119.05s eta: 4:35:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31564.0 tgs: 264 data_time: 0.85s time: 119.37s eta: 4:34:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32125.0 tgs: 269 data_time: 1.04s time: 119.41s eta: 4:32:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 268 data_time: 0.92s time: 118.94s eta: 4:29:35 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31387.0 tgs: 262 data_time: 0.66s time: 119.60s eta: 4:29:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 268 data_time: 0.89s time: 119.24s eta: 4:26:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32064.0 tgs: 269 data_time: 0.75s time: 119.06s eta: 4:23:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.218 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32467.0 tgs: 273 data_time: 0.82s time: 118.85s eta: 4:21:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.194 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32273.0 tgs: 271 data_time: 0.81s time: 118.76s eta: 4:19:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30833.0 tgs: 258 data_time: 0.49s time: 119.43s eta: 4:18:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31721.0 tgs: 265 data_time: 0.61s time: 119.40s eta: 4:16:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32424.0 tgs: 271 data_time: 0.87s time: 119.35s eta: 4:14:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31769.0 tgs: 266 data_time: 0.76s time: 119.04s eta: 4:11:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32106.0 tgs: 269 data_time: 1.03s time: 119.12s eta: 4:10:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32249.0 tgs: 270 data_time: 0.59s time: 119.14s eta: 4:08:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30545.0 tgs: 256 data_time: 0.74s time: 118.96s eta: 4:05:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 264 data_time: 0.62s time: 119.15s eta: 4:04:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31905.0 tgs: 266 data_time: 0.61s time: 119.50s eta: 4:02:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30723.0 tgs: 257 data_time: 0.88s time: 119.27s eta: 4:00:31 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.114 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31644.0 tgs: 265 data_time: 0.62s time: 119.17s eta: 3:58:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31766.0 tgs: 266 data_time: 0.77s time: 119.36s eta: 3:56:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31665.0 tgs: 265 data_time: 0.65s time: 119.37s eta: 3:54:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30904.0 tgs: 257 data_time: 0.81s time: 119.95s eta: 3:53:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.205 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30883.0 tgs: 258 data_time: 0.77s time: 119.32s eta: 3:50:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 261 data_time: 0.82s time: 119.63s eta: 3:49:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30921.0 tgs: 260 data_time: 0.65s time: 118.86s eta: 3:45:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31583.0 tgs: 263 data_time: 1.01s time: 119.75s eta: 3:45:32 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.111 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31199.0 tgs: 261 data_time: 0.68s time: 119.32s eta: 3:42:44 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30747.0 tgs: 257 data_time: 0.99s time: 119.26s eta: 3:40:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31204.0 tgs: 262 data_time: 0.75s time: 119.00s eta: 3:38:09 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30579.0 tgs: 256 data_time: 0.72s time: 119.43s eta: 3:36:57 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.219 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31222.0 tgs: 262 data_time: 0.81s time: 119.04s eta: 3:34:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 266 data_time: 0.54s time: 118.99s eta: 3:32:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.192 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31853.0 tgs: 267 data_time: 0.92s time: 119.21s eta: 3:30:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31330.0 tgs: 260 data_time: 1.03s time: 120.11s eta: 3:30:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31220.0 tgs: 262 data_time: 0.59s time: 119.13s eta: 3:26:29 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30904.0 tgs: 259 data_time: 0.69s time: 119.31s eta: 3:24:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31972.0 tgs: 268 data_time: 0.67s time: 119.09s eta: 3:22:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31699.0 tgs: 265 data_time: 1.01s time: 119.42s eta: 3:21:01 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31700.0 tgs: 265 data_time: 0.61s time: 119.29s eta: 3:18:48 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.94s time: 119.22s eta: 3:16:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31205.0 tgs: 261 data_time: 0.76s time: 119.15s eta: 3:14:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31661.0 tgs: 266 data_time: 0.85s time: 119.02s eta: 3:12:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.111 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31989.0 tgs: 268 data_time: 0.82s time: 119.19s eta: 3:10:41 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 266 data_time: 0.62s time: 120.99s eta: 3:11:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.229 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31760.0 tgs: 265 data_time: 0.75s time: 119.50s eta: 3:07:13 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31694.0 tgs: 265 data_time: 0.85s time: 119.26s eta: 3:04:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 271 data_time: 0.61s time: 118.99s eta: 3:02:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32312.0 tgs: 271 data_time: 0.80s time: 119.22s eta: 3:00:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31782.0 tgs: 266 data_time: 0.68s time: 119.46s eta: 2:59:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30415.0 tgs: 255 data_time: 0.50s time: 118.98s eta: 2:56:28 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31745.0 tgs: 265 data_time: 0.81s time: 119.35s eta: 2:55:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31561.0 tgs: 264 data_time: 0.66s time: 119.41s eta: 2:53:08 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31598.0 tgs: 264 data_time: 0.75s time: 119.36s eta: 2:51:04 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32264.0 tgs: 265 data_time: 0.89s time: 121.46s eta: 2:52:03 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32490.0 tgs: 272 data_time: 0.71s time: 119.24s eta: 2:46:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31547.0 tgs: 263 data_time: 0.84s time: 119.61s eta: 2:45:27 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 267 data_time: 0.68s time: 118.91s eta: 2:42:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30961.0 tgs: 259 data_time: 0.72s time: 119.18s eta: 2:40:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31283.0 tgs: 262 data_time: 0.48s time: 118.96s eta: 2:38:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32085.0 tgs: 269 data_time: 0.71s time: 119.22s eta: 2:36:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31878.0 tgs: 267 data_time: 0.93s time: 119.11s eta: 2:34:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 267 data_time: 0.92s time: 119.43s eta: 2:33:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.211 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31257.0 tgs: 259 data_time: 0.83s time: 120.23s eta: 2:32:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.198 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31021.0 tgs: 260 data_time: 1.07s time: 119.28s eta: 2:29:05 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32284.0 tgs: 271 data_time: 0.84s time: 118.88s eta: 2:26:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32371.0 tgs: 270 data_time: 0.60s time: 119.51s eta: 2:25:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31823.0 tgs: 267 data_time: 0.71s time: 119.07s eta: 2:22:52 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32297.0 tgs: 271 data_time: 0.73s time: 119.14s eta: 2:20:58 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31475.0 tgs: 264 data_time: 0.59s time: 119.04s eta: 2:18:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32108.0 tgs: 267 data_time: 1.01s time: 119.85s eta: 2:17:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32158.0 tgs: 269 data_time: 0.72s time: 119.32s eta: 2:15:13 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31771.0 tgs: 266 data_time: 0.55s time: 119.38s eta: 2:13:18 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31079.0 tgs: 259 data_time: 0.72s time: 119.68s eta: 2:11:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.106 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32424.0 tgs: 271 data_time: 0.48s time: 119.49s eta: 2:09:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32482.0 tgs: 273 data_time: 0.67s time: 118.96s eta: 2:06:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32062.0 tgs: 268 data_time: 0.86s time: 119.37s eta: 2:05:20 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31877.0 tgs: 267 data_time: 0.80s time: 119.27s eta: 2:03:14 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31629.0 tgs: 265 data_time: 0.65s time: 119.22s eta: 2:01:12 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31103.0 tgs: 260 data_time: 0.80s time: 119.51s eta: 1:59:30 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31995.0 tgs: 267 data_time: 0.52s time: 119.42s eta: 1:57:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32402.0 tgs: 271 data_time: 0.77s time: 119.54s eta: 1:55:33 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31816.0 tgs: 266 data_time: 0.76s time: 119.30s eta: 1:53:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31753.0 tgs: 266 data_time: 0.76s time: 119.25s eta: 1:51:18 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32589.0 tgs: 274 data_time: 0.55s time: 118.93s eta: 1:49:00 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.100 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31868.0 tgs: 266 data_time: 0.69s time: 119.64s eta: 1:47:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32066.0 tgs: 267 data_time: 0.60s time: 119.72s eta: 1:45:45 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31708.0 tgs: 265 data_time: 0.61s time: 119.35s eta: 1:43:26 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32499.0 tgs: 272 data_time: 0.83s time: 119.30s eta: 1:41:24 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30551.0 tgs: 256 data_time: 0.91s time: 119.28s eta: 1:39:23 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 268 data_time: 0.80s time: 118.90s eta: 1:37:06 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 264 data_time: 1.04s time: 119.16s eta: 1:35:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30753.0 tgs: 258 data_time: 0.82s time: 119.06s eta: 1:33:15 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 271 data_time: 0.74s time: 119.25s eta: 1:31:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27985.0 tgs: 233 data_time: 0.77s time: 119.72s eta: 1:29:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32023.0 tgs: 269 data_time: 0.95s time: 119.01s eta: 1:27:16 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30952.0 tgs: 260 data_time: 0.80s time: 118.88s eta: 1:25:11 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32532.0 tgs: 272 data_time: 0.81s time: 119.47s eta: 1:23:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31867.0 tgs: 266 data_time: 0.72s time: 119.49s eta: 1:21:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31874.0 tgs: 267 data_time: 0.72s time: 118.94s eta: 1:19:17 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32024.0 tgs: 269 data_time: 0.77s time: 118.93s eta: 1:17:18 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30263.0 tgs: 247 data_time: 0.78s time: 122.15s eta: 1:17:21 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32358.0 tgs: 271 data_time: 0.62s time: 119.36s eta: 1:13:36 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31928.0 tgs: 268 data_time: 0.70s time: 119.05s eta: 1:11:25 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31158.0 tgs: 262 data_time: 0.60s time: 118.91s eta: 1:09:22 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 268 data_time: 0.80s time: 119.10s eta: 1:07:29 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 271 data_time: 0.84s time: 119.41s eta: 1:05:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.223 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32495.0 tgs: 272 data_time: 0.76s time: 119.35s eta: 1:03:39 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31124.0 tgs: 261 data_time: 0.83s time: 119.18s eta: 1:01:34 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31735.0 tgs: 266 data_time: 0.57s time: 119.27s eta: 0:59:37 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30877.0 tgs: 258 data_time: 0.65s time: 119.44s eta: 0:57:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 262 data_time: 0.66s time: 119.39s eta: 0:55:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31455.0 tgs: 263 data_time: 0.87s time: 119.49s eta: 0:53:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30600.0 tgs: 256 data_time: 0.74s time: 119.35s eta: 0:51:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 263 data_time: 0.68s time: 119.47s eta: 0:49:46 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31802.0 tgs: 267 data_time: 0.57s time: 118.89s eta: 0:47:33 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32526.0 tgs: 272 data_time: 0.65s time: 119.28s eta: 0:45:43 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.72s time: 119.21s eta: 0:43:42 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32089.0 tgs: 269 data_time: 0.60s time: 118.99s eta: 0:41:38 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31929.0 tgs: 268 data_time: 0.80s time: 119.04s eta: 0:39:40 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32223.0 tgs: 269 data_time: 0.49s time: 119.66s eta: 0:37:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31089.0 tgs: 260 data_time: 0.68s time: 119.29s eta: 0:35:47 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31516.0 tgs: 263 data_time: 0.89s time: 119.74s eta: 0:33:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31705.0 tgs: 261 data_time: 0.77s time: 121.22s eta: 0:32:19 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31997.0 tgs: 268 data_time: 0.73s time: 119.38s eta: 0:29:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31160.0 tgs: 261 data_time: 0.59s time: 119.32s eta: 0:27:50 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31162.0 tgs: 260 data_time: 0.55s time: 119.64s eta: 0:25:55 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32288.0 tgs: 271 data_time: 0.91s time: 119.09s eta: 0:23:49 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31744.0 tgs: 265 data_time: 0.81s time: 119.38s eta: 0:21:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31076.0 tgs: 260 data_time: 0.80s time: 119.26s eta: 0:19:52 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30722.0 tgs: 257 data_time: 0.69s time: 119.24s eta: 0:17:53 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32418.0 tgs: 270 data_time: 0.84s time: 119.91s eta: 0:15:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32228.0 tgs: 270 data_time: 0.92s time: 119.21s eta: 0:13:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31518.0 tgs: 264 data_time: 0.51s time: 119.09s eta: 0:11:54 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32296.0 tgs: 270 data_time: 0.54s time: 119.32s eta: 0:09:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32046.0 tgs: 268 data_time: 0.72s time: 119.17s eta: 0:07:56 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32146.0 tgs: 268 data_time: 0.64s time: 119.69s eta: 0:05:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31461.0 tgs: 259 data_time: 0.87s time: 121.43s eta: 0:04:02 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31883.0 tgs: 267 data_time: 0.77s time: 119.21s eta: 0:01:59 +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 14][DP 3][SP 2][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank18.log b/20250121165312/rank18.log new file mode 100644 index 0000000000000000000000000000000000000000..1e615927b3fe801084702af136a4fdb0c6ef9ea7 --- /dev/null +++ b/20250121165312/rank18.log @@ -0,0 +1,851 @@ +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.84 seconds, peak gpu memory 13.4G +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.227 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31743.0 tgs: 209 data_time: 1.91s time: 151.22s eta: 1 day, 0:54:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.331 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32110.0 tgs: 267 data_time: 0.94s time: 120.11s eta: 19:45:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.201 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31536.0 tgs: 263 data_time: 0.84s time: 119.69s eta: 19:38:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.218 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31549.0 tgs: 263 data_time: 1.07s time: 119.65s eta: 19:36:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31125.0 tgs: 260 data_time: 0.92s time: 119.31s eta: 19:31:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.292 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31488.0 tgs: 263 data_time: 1.15s time: 119.54s eta: 19:31:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.190 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30800.0 tgs: 250 data_time: 0.76s time: 122.96s eta: 20:02:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.211 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32301.0 tgs: 269 data_time: 0.94s time: 119.89s eta: 19:30:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.240 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31885.0 tgs: 264 data_time: 0.92s time: 120.65s eta: 19:36:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32246.0 tgs: 268 data_time: 0.80s time: 119.92s eta: 19:27:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.196 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32255.0 tgs: 267 data_time: 0.74s time: 120.46s eta: 19:30:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.215 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31976.0 tgs: 267 data_time: 0.84s time: 119.34s eta: 19:17:37 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32057.0 tgs: 267 data_time: 0.73s time: 119.68s eta: 19:18:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31783.0 tgs: 266 data_time: 0.83s time: 119.18s eta: 19:12:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30276.0 tgs: 253 data_time: 0.69s time: 119.21s eta: 19:10:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32240.0 tgs: 271 data_time: 0.75s time: 118.93s eta: 19:05:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32111.0 tgs: 269 data_time: 0.89s time: 119.15s eta: 19:05:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30872.0 tgs: 259 data_time: 0.75s time: 119.18s eta: 19:04:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31286.0 tgs: 262 data_time: 0.96s time: 119.40s eta: 19:04:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31340.0 tgs: 262 data_time: 0.69s time: 119.34s eta: 19:01:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31187.0 tgs: 259 data_time: 0.87s time: 120.31s eta: 19:08:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31884.0 tgs: 259 data_time: 0.91s time: 122.92s eta: 19:31:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32216.0 tgs: 269 data_time: 0.89s time: 119.48s eta: 18:57:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31832.0 tgs: 267 data_time: 0.64s time: 119.16s eta: 18:52:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.229 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 268 data_time: 0.94s time: 119.24s eta: 18:50:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.75s time: 119.39s eta: 18:50:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31130.0 tgs: 260 data_time: 0.95s time: 119.66s eta: 18:50:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30936.0 tgs: 259 data_time: 0.58s time: 119.38s eta: 18:46:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 264 data_time: 0.68s time: 119.76s eta: 18:47:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31970.0 tgs: 247 data_time: 0.97s time: 129.29s eta: 20:15:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 25656.0 tgs: 214 data_time: 0.79s time: 119.58s eta: 18:42:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31702.0 tgs: 265 data_time: 0.86s time: 119.39s eta: 18:38:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31282.0 tgs: 259 data_time: 0.65s time: 120.66s eta: 18:48:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31475.0 tgs: 263 data_time: 0.60s time: 119.40s eta: 18:34:21 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31004.0 tgs: 260 data_time: 0.86s time: 119.17s eta: 18:30:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31063.0 tgs: 260 data_time: 0.88s time: 119.07s eta: 18:27:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31013.0 tgs: 260 data_time: 0.92s time: 119.14s eta: 18:25:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32430.0 tgs: 269 data_time: 0.75s time: 120.39s eta: 18:35:36 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32259.0 tgs: 266 data_time: 0.56s time: 121.17s eta: 18:40:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31282.0 tgs: 259 data_time: 0.79s time: 120.67s eta: 18:34:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 265 data_time: 0.72s time: 119.73s eta: 18:23:33 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32064.0 tgs: 258 data_time: 0.55s time: 123.88s eta: 18:59:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30877.0 tgs: 258 data_time: 0.57s time: 119.56s eta: 18:17:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31532.0 tgs: 264 data_time: 0.67s time: 119.32s eta: 18:13:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31384.0 tgs: 263 data_time: 1.06s time: 119.23s eta: 18:10:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31932.0 tgs: 266 data_time: 0.90s time: 119.61s eta: 18:12:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31103.0 tgs: 260 data_time: 0.67s time: 119.27s eta: 18:07:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32490.0 tgs: 272 data_time: 0.93s time: 119.01s eta: 18:03:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32230.0 tgs: 270 data_time: 0.93s time: 119.10s eta: 18:01:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.245 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31721.0 tgs: 265 data_time: 0.93s time: 119.27s eta: 18:01:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 266 data_time: 0.83s time: 119.33s eta: 17:59:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.248 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30994.0 tgs: 258 data_time: 0.73s time: 120.02s eta: 18:04:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31322.0 tgs: 262 data_time: 0.85s time: 119.10s eta: 17:53:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.78s time: 119.41s eta: 17:54:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31833.0 tgs: 267 data_time: 0.76s time: 119.00s eta: 17:48:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31319.0 tgs: 263 data_time: 0.84s time: 119.03s eta: 17:47:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29936.0 tgs: 238 data_time: 0.93s time: 125.38s eta: 18:42:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 269 data_time: 0.88s time: 120.10s eta: 17:52:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31346.0 tgs: 262 data_time: 0.68s time: 119.36s eta: 17:44:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31934.0 tgs: 267 data_time: 0.78s time: 119.34s eta: 17:42:06 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32223.0 tgs: 269 data_time: 0.81s time: 119.51s eta: 17:41:37 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31650.0 tgs: 251 data_time: 0.80s time: 126.05s eta: 18:37:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.120 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32139.0 tgs: 262 data_time: 0.78s time: 122.63s eta: 18:05:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30916.0 tgs: 255 data_time: 0.64s time: 121.16s eta: 17:50:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31094.0 tgs: 261 data_time: 0.95s time: 119.04s eta: 17:29:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 268 data_time: 0.86s time: 118.86s eta: 17:25:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31710.0 tgs: 255 data_time: 0.87s time: 124.00s eta: 18:09:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.120 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32386.0 tgs: 259 data_time: 0.95s time: 124.57s eta: 18:12:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32507.0 tgs: 272 data_time: 0.75s time: 119.39s eta: 17:24:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31766.0 tgs: 265 data_time: 0.60s time: 119.70s eta: 17:25:21 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31667.0 tgs: 264 data_time: 0.82s time: 119.53s eta: 17:21:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32185.0 tgs: 270 data_time: 0.90s time: 118.86s eta: 17:14:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32205.0 tgs: 268 data_time: 0.70s time: 119.85s eta: 17:20:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30812.0 tgs: 257 data_time: 1.17s time: 119.52s eta: 17:15:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31123.0 tgs: 255 data_time: 0.69s time: 121.86s eta: 17:34:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31445.0 tgs: 254 data_time: 0.73s time: 123.52s eta: 17:46:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32210.0 tgs: 270 data_time: 0.56s time: 119.01s eta: 17:05:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31669.0 tgs: 265 data_time: 0.70s time: 119.23s eta: 17:05:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31166.0 tgs: 261 data_time: 0.63s time: 119.14s eta: 17:02:36 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31666.0 tgs: 265 data_time: 0.60s time: 119.26s eta: 17:01:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31999.0 tgs: 268 data_time: 1.17s time: 119.05s eta: 16:57:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 269 data_time: 0.74s time: 119.28s eta: 16:57:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 266 data_time: 0.68s time: 119.16s eta: 16:54:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.229 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31563.0 tgs: 263 data_time: 1.06s time: 119.59s eta: 16:56:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32031.0 tgs: 268 data_time: 0.79s time: 119.13s eta: 16:50:37 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31739.0 tgs: 267 data_time: 0.78s time: 118.84s eta: 16:46:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32110.0 tgs: 269 data_time: 0.76s time: 119.02s eta: 16:45:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31903.0 tgs: 266 data_time: 0.61s time: 119.85s eta: 16:50:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32347.0 tgs: 271 data_time: 0.83s time: 119.36s eta: 16:44:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31321.0 tgs: 263 data_time: 0.70s time: 119.01s eta: 16:39:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.209 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31818.0 tgs: 263 data_time: 0.91s time: 120.93s eta: 16:53:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31894.0 tgs: 268 data_time: 0.72s time: 119.00s eta: 16:35:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31728.0 tgs: 265 data_time: 0.73s time: 119.29s eta: 16:36:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30886.0 tgs: 259 data_time: 0.62s time: 119.23s eta: 16:33:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29656.0 tgs: 248 data_time: 0.70s time: 119.37s eta: 16:32:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31814.0 tgs: 265 data_time: 1.09s time: 119.70s eta: 16:33:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32058.0 tgs: 268 data_time: 0.85s time: 119.21s eta: 16:27:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31889.0 tgs: 266 data_time: 0.67s time: 119.77s eta: 16:30:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32052.0 tgs: 266 data_time: 0.81s time: 120.17s eta: 16:31:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30863.0 tgs: 259 data_time: 0.58s time: 118.97s eta: 16:19:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31520.0 tgs: 247 data_time: 0.63s time: 127.33s eta: 17:26:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31465.0 tgs: 263 data_time: 0.77s time: 119.23s eta: 16:17:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.266 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31460.0 tgs: 264 data_time: 0.60s time: 119.14s eta: 16:14:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31938.0 tgs: 267 data_time: 0.76s time: 119.21s eta: 16:13:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31583.0 tgs: 265 data_time: 0.58s time: 119.18s eta: 16:11:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.58s time: 118.95s eta: 16:07:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.246 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32064.0 tgs: 269 data_time: 0.68s time: 118.86s eta: 16:04:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32269.0 tgs: 271 data_time: 1.03s time: 118.96s eta: 16:03:36 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31714.0 tgs: 260 data_time: 0.76s time: 121.84s eta: 16:24:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 267 data_time: 0.77s time: 119.00s eta: 15:59:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32517.0 tgs: 272 data_time: 0.74s time: 119.22s eta: 15:59:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31475.0 tgs: 263 data_time: 0.73s time: 119.62s eta: 16:00:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31230.0 tgs: 262 data_time: 0.72s time: 119.04s eta: 15:54:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32030.0 tgs: 269 data_time: 1.04s time: 118.98s eta: 15:51:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31769.0 tgs: 266 data_time: 0.99s time: 119.17s eta: 15:51:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31777.0 tgs: 265 data_time: 0.62s time: 119.64s eta: 15:53:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31341.0 tgs: 263 data_time: 0.70s time: 118.94s eta: 15:45:33 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31042.0 tgs: 260 data_time: 0.76s time: 119.29s eta: 15:46:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31336.0 tgs: 263 data_time: 0.92s time: 119.11s eta: 15:42:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31911.0 tgs: 268 data_time: 0.88s time: 119.01s eta: 15:40:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.262 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31300.0 tgs: 261 data_time: 1.06s time: 119.65s eta: 15:43:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31599.0 tgs: 260 data_time: 0.77s time: 121.51s eta: 15:55:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31396.0 tgs: 262 data_time: 0.69s time: 119.41s eta: 15:37:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.227 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30794.0 tgs: 236 data_time: 1.19s time: 129.96s eta: 16:58:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31685.0 tgs: 265 data_time: 0.78s time: 119.47s eta: 15:33:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31889.0 tgs: 266 data_time: 0.64s time: 119.60s eta: 15:32:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31201.0 tgs: 261 data_time: 0.72s time: 119.53s eta: 15:30:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31190.0 tgs: 255 data_time: 0.88s time: 122.14s eta: 15:48:36 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.234 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30670.0 tgs: 253 data_time: 0.85s time: 120.75s eta: 15:35:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31824.0 tgs: 266 data_time: 0.93s time: 119.41s eta: 15:23:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 262 data_time: 0.89s time: 119.45s eta: 15:21:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31437.0 tgs: 238 data_time: 0.67s time: 131.63s eta: 16:53:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32126.0 tgs: 269 data_time: 0.68s time: 119.30s eta: 15:16:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.85s time: 119.52s eta: 15:16:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29523.0 tgs: 241 data_time: 0.77s time: 122.25s eta: 15:35:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32097.0 tgs: 268 data_time: 1.01s time: 119.72s eta: 15:13:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31777.0 tgs: 266 data_time: 0.77s time: 119.34s eta: 15:08:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31121.0 tgs: 261 data_time: 0.83s time: 118.91s eta: 15:03:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 268 data_time: 0.86s time: 119.00s eta: 15:02:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31603.0 tgs: 264 data_time: 0.76s time: 119.31s eta: 15:02:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30722.0 tgs: 258 data_time: 0.72s time: 118.92s eta: 14:57:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31648.0 tgs: 262 data_time: 0.82s time: 120.52s eta: 15:07:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.212 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30991.0 tgs: 259 data_time: 0.65s time: 119.31s eta: 14:56:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.230 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32332.0 tgs: 263 data_time: 0.72s time: 122.66s eta: 15:19:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32147.0 tgs: 269 data_time: 0.50s time: 119.13s eta: 14:51:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31082.0 tgs: 259 data_time: 0.77s time: 119.76s eta: 14:54:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31777.0 tgs: 265 data_time: 0.67s time: 119.47s eta: 14:50:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31915.0 tgs: 268 data_time: 0.89s time: 118.99s eta: 14:44:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32249.0 tgs: 270 data_time: 0.88s time: 119.31s eta: 14:44:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31665.0 tgs: 266 data_time: 0.74s time: 118.93s eta: 14:40:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 268 data_time: 0.83s time: 119.05s eta: 14:39:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31048.0 tgs: 260 data_time: 0.73s time: 119.07s eta: 14:37:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.239 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32117.0 tgs: 269 data_time: 0.67s time: 119.12s eta: 14:35:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30732.0 tgs: 258 data_time: 0.65s time: 119.00s eta: 14:32:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31512.0 tgs: 264 data_time: 0.64s time: 119.08s eta: 14:31:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31349.0 tgs: 263 data_time: 0.84s time: 119.15s eta: 14:29:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.236 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31765.0 tgs: 265 data_time: 1.11s time: 119.65s eta: 14:31:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31271.0 tgs: 262 data_time: 0.72s time: 119.18s eta: 14:26:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.245 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 269 data_time: 0.67s time: 119.20s eta: 14:24:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31036.0 tgs: 259 data_time: 0.81s time: 119.39s eta: 14:23:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31079.0 tgs: 257 data_time: 0.73s time: 120.74s eta: 14:31:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32257.0 tgs: 271 data_time: 0.90s time: 118.86s eta: 14:15:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30906.0 tgs: 259 data_time: 0.90s time: 119.10s eta: 14:15:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.119 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31850.0 tgs: 267 data_time: 0.65s time: 119.12s eta: 14:13:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 26708.0 tgs: 222 data_time: 0.92s time: 119.96s eta: 14:17:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 261 data_time: 0.79s time: 119.87s eta: 14:15:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31584.0 tgs: 264 data_time: 0.74s time: 119.59s eta: 14:11:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32145.0 tgs: 269 data_time: 0.98s time: 119.23s eta: 14:06:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.209 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31895.0 tgs: 267 data_time: 0.88s time: 119.27s eta: 14:04:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32126.0 tgs: 269 data_time: 0.78s time: 119.27s eta: 14:02:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31916.0 tgs: 267 data_time: 0.61s time: 119.14s eta: 13:59:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31737.0 tgs: 266 data_time: 0.82s time: 119.12s eta: 13:57:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.122 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30614.0 tgs: 256 data_time: 0.64s time: 119.18s eta: 13:56:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.238 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32403.0 tgs: 271 data_time: 0.93s time: 119.35s eta: 13:55:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32416.0 tgs: 269 data_time: 0.54s time: 120.11s eta: 13:58:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31140.0 tgs: 254 data_time: 0.57s time: 122.56s eta: 14:13:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.201 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31665.0 tgs: 265 data_time: 0.73s time: 119.11s eta: 13:47:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31453.0 tgs: 263 data_time: 0.70s time: 119.37s eta: 13:47:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.121 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 271 data_time: 0.68s time: 119.32s eta: 13:45:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32243.0 tgs: 270 data_time: 0.93s time: 119.32s eta: 13:43:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30775.0 tgs: 258 data_time: 0.50s time: 119.12s eta: 13:39:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32131.0 tgs: 269 data_time: 0.73s time: 119.26s eta: 13:38:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31527.0 tgs: 264 data_time: 0.60s time: 119.14s eta: 13:36:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32328.0 tgs: 271 data_time: 0.77s time: 119.23s eta: 13:34:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31651.0 tgs: 264 data_time: 0.66s time: 119.46s eta: 13:34:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30881.0 tgs: 259 data_time: 0.70s time: 119.09s eta: 13:29:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30815.0 tgs: 259 data_time: 0.78s time: 118.92s eta: 13:26:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32116.0 tgs: 268 data_time: 0.95s time: 119.81s eta: 13:30:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32180.0 tgs: 269 data_time: 0.72s time: 119.26s eta: 13:24:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.217 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31510.0 tgs: 264 data_time: 0.81s time: 119.16s eta: 13:22:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32133.0 tgs: 269 data_time: 0.75s time: 119.35s eta: 13:21:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29491.0 tgs: 247 data_time: 0.60s time: 119.26s eta: 13:19:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31646.0 tgs: 265 data_time: 0.75s time: 119.04s eta: 13:15:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30840.0 tgs: 258 data_time: 0.79s time: 119.35s eta: 13:15:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31665.0 tgs: 265 data_time: 0.71s time: 119.17s eta: 13:12:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32557.0 tgs: 272 data_time: 1.08s time: 119.46s eta: 13:12:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30533.0 tgs: 255 data_time: 0.85s time: 119.29s eta: 13:09:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30488.0 tgs: 255 data_time: 0.62s time: 119.15s eta: 13:06:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30817.0 tgs: 256 data_time: 0.78s time: 120.32s eta: 13:12:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31842.0 tgs: 264 data_time: 0.69s time: 120.16s eta: 13:09:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31486.0 tgs: 263 data_time: 0.73s time: 119.50s eta: 13:02:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31980.0 tgs: 267 data_time: 0.68s time: 119.51s eta: 13:00:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31087.0 tgs: 259 data_time: 1.02s time: 119.58s eta: 12:59:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.199 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 28461.0 tgs: 237 data_time: 0.57s time: 119.79s eta: 12:58:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.277 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32211.0 tgs: 270 data_time: 0.77s time: 119.04s eta: 12:51:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.191 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31603.0 tgs: 265 data_time: 0.64s time: 119.05s eta: 12:49:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.117 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32528.0 tgs: 273 data_time: 1.04s time: 119.13s eta: 12:48:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.183 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31746.0 tgs: 265 data_time: 0.58s time: 119.40s eta: 12:48:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.67s time: 119.18s eta: 12:44:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.131 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31327.0 tgs: 262 data_time: 0.64s time: 119.42s eta: 12:44:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.195 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32016.0 tgs: 268 data_time: 0.82s time: 119.10s eta: 12:40:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.118 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32043.0 tgs: 269 data_time: 0.71s time: 119.07s eta: 12:38:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32509.0 tgs: 271 data_time: 0.59s time: 119.59s eta: 12:39:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.202 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31483.0 tgs: 265 data_time: 0.42s time: 118.78s eta: 12:32:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31781.0 tgs: 267 data_time: 0.70s time: 118.89s eta: 12:30:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31054.0 tgs: 260 data_time: 0.72s time: 119.17s eta: 12:30:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31118.0 tgs: 260 data_time: 0.68s time: 119.26s eta: 12:29:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31348.0 tgs: 262 data_time: 0.62s time: 119.35s eta: 12:27:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 269 data_time: 0.83s time: 119.12s eta: 12:24:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.129 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31245.0 tgs: 261 data_time: 0.70s time: 119.34s eta: 12:23:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31097.0 tgs: 260 data_time: 0.64s time: 119.26s eta: 12:21:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 270 data_time: 0.74s time: 119.03s eta: 12:17:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31732.0 tgs: 261 data_time: 1.11s time: 121.14s eta: 12:29:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.217 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32092.0 tgs: 267 data_time: 0.69s time: 119.98s eta: 12:19:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31379.0 tgs: 264 data_time: 0.72s time: 118.69s eta: 12:09:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.292 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31622.0 tgs: 264 data_time: 1.16s time: 119.64s eta: 12:13:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27513.0 tgs: 229 data_time: 0.44s time: 119.84s eta: 12:13:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31625.0 tgs: 265 data_time: 0.85s time: 119.15s eta: 12:06:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31639.0 tgs: 265 data_time: 0.77s time: 119.17s eta: 12:04:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 270 data_time: 0.76s time: 119.11s eta: 12:02:37 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30513.0 tgs: 256 data_time: 0.80s time: 119.19s eta: 12:01:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31355.0 tgs: 263 data_time: 0.72s time: 118.96s eta: 11:57:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31615.0 tgs: 264 data_time: 0.79s time: 119.70s eta: 12:00:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31669.0 tgs: 265 data_time: 0.88s time: 119.28s eta: 11:55:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.235 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31108.0 tgs: 260 data_time: 0.75s time: 119.28s eta: 11:53:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31325.0 tgs: 261 data_time: 0.66s time: 119.71s eta: 11:54:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31836.0 tgs: 268 data_time: 0.64s time: 118.79s eta: 11:46:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.138 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32389.0 tgs: 271 data_time: 0.64s time: 119.16s eta: 11:47:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31043.0 tgs: 259 data_time: 0.53s time: 119.61s eta: 11:47:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.242 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31839.0 tgs: 267 data_time: 0.77s time: 118.82s eta: 11:41:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.118 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31741.0 tgs: 267 data_time: 0.72s time: 118.68s eta: 11:38:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.113 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31795.0 tgs: 266 data_time: 0.92s time: 119.22s eta: 11:39:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.190 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32202.0 tgs: 270 data_time: 0.68s time: 119.14s eta: 11:36:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.204 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31738.0 tgs: 266 data_time: 0.99s time: 119.08s eta: 11:34:37 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30795.0 tgs: 258 data_time: 0.82s time: 118.95s eta: 11:31:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30607.0 tgs: 257 data_time: 0.85s time: 118.91s eta: 11:29:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 271 data_time: 1.04s time: 118.97s eta: 11:28:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31153.0 tgs: 261 data_time: 0.63s time: 119.28s eta: 11:27:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.203 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31489.0 tgs: 263 data_time: 0.72s time: 119.35s eta: 11:26:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31182.0 tgs: 262 data_time: 0.70s time: 118.88s eta: 11:21:36 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31484.0 tgs: 264 data_time: 0.91s time: 119.21s eta: 11:21:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31691.0 tgs: 264 data_time: 0.71s time: 119.73s eta: 11:22:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 265 data_time: 0.66s time: 119.11s eta: 11:16:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31577.0 tgs: 265 data_time: 0.70s time: 119.12s eta: 11:15:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31457.0 tgs: 264 data_time: 0.70s time: 118.94s eta: 11:12:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.215 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31446.0 tgs: 264 data_time: 1.00s time: 119.09s eta: 11:10:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30453.0 tgs: 255 data_time: 0.68s time: 119.16s eta: 11:09:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30547.0 tgs: 256 data_time: 0.81s time: 118.97s eta: 11:06:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30882.0 tgs: 258 data_time: 0.59s time: 119.47s eta: 11:07:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31907.0 tgs: 266 data_time: 1.15s time: 119.64s eta: 11:06:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31708.0 tgs: 265 data_time: 0.66s time: 119.31s eta: 11:02:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30810.0 tgs: 258 data_time: 0.57s time: 119.31s eta: 11:00:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.188 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31611.0 tgs: 266 data_time: 0.93s time: 118.82s eta: 10:55:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.227 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31308.0 tgs: 262 data_time: 0.75s time: 119.48s eta: 10:57:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.225 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31552.0 tgs: 264 data_time: 0.91s time: 119.37s eta: 10:54:31 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31525.0 tgs: 264 data_time: 0.97s time: 119.28s eta: 10:52:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 266 data_time: 0.74s time: 119.26s eta: 10:49:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32132.0 tgs: 269 data_time: 0.86s time: 119.10s eta: 10:47:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 269 data_time: 0.88s time: 119.53s eta: 10:47:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32138.0 tgs: 269 data_time: 0.82s time: 119.31s eta: 10:44:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30523.0 tgs: 255 data_time: 0.60s time: 119.49s eta: 10:43:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31633.0 tgs: 264 data_time: 0.68s time: 119.52s eta: 10:41:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32295.0 tgs: 271 data_time: 0.77s time: 118.99s eta: 10:36:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31451.0 tgs: 263 data_time: 0.88s time: 119.22s eta: 10:35:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31551.0 tgs: 264 data_time: 0.93s time: 119.06s eta: 10:33:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31318.0 tgs: 263 data_time: 0.49s time: 118.78s eta: 10:29:33 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 268 data_time: 0.76s time: 119.86s eta: 10:33:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30870.0 tgs: 258 data_time: 0.63s time: 119.38s eta: 10:28:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31212.0 tgs: 262 data_time: 0.94s time: 118.84s eta: 10:23:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.227 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32518.0 tgs: 273 data_time: 0.84s time: 118.98s eta: 10:22:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.190 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 27817.0 tgs: 233 data_time: 0.65s time: 119.28s eta: 10:22:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 267 data_time: 0.69s time: 119.21s eta: 10:19:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31829.0 tgs: 267 data_time: 0.79s time: 119.06s eta: 10:17:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32198.0 tgs: 269 data_time: 0.79s time: 119.48s eta: 10:17:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31500.0 tgs: 263 data_time: 0.75s time: 119.38s eta: 10:14:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.244 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31762.0 tgs: 265 data_time: 0.82s time: 119.45s eta: 10:13:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31562.0 tgs: 264 data_time: 0.77s time: 119.29s eta: 10:10:21 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.199 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31758.0 tgs: 266 data_time: 0.93s time: 119.11s eta: 10:07:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31897.0 tgs: 266 data_time: 0.63s time: 119.70s eta: 10:08:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32050.0 tgs: 267 data_time: 0.86s time: 119.83s eta: 10:07:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.224 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32237.0 tgs: 265 data_time: 0.76s time: 121.55s eta: 10:13:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31937.0 tgs: 267 data_time: 1.01s time: 119.26s eta: 10:00:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31845.0 tgs: 268 data_time: 1.11s time: 118.81s eta: 9:56:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31596.0 tgs: 264 data_time: 0.92s time: 119.56s eta: 9:57:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.203 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31636.0 tgs: 264 data_time: 0.75s time: 119.71s eta: 9:56:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31615.0 tgs: 265 data_time: 0.69s time: 119.26s eta: 9:52:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 267 data_time: 0.69s time: 119.33s eta: 9:50:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32026.0 tgs: 268 data_time: 0.84s time: 119.23s eta: 9:48:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31351.0 tgs: 259 data_time: 0.89s time: 120.76s eta: 9:53:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31545.0 tgs: 264 data_time: 1.03s time: 119.23s eta: 9:44:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 267 data_time: 0.97s time: 119.46s eta: 9:43:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.195 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32535.0 tgs: 272 data_time: 0.85s time: 119.55s eta: 9:41:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32043.0 tgs: 268 data_time: 0.69s time: 119.20s eta: 9:38:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.96s time: 119.48s eta: 9:37:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30940.0 tgs: 259 data_time: 0.76s time: 119.05s eta: 9:33:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31945.0 tgs: 267 data_time: 0.74s time: 119.24s eta: 9:32:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.191 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31356.0 tgs: 263 data_time: 0.78s time: 119.16s eta: 9:29:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.267 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31930.0 tgs: 267 data_time: 0.91s time: 119.22s eta: 9:28:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30210.0 tgs: 252 data_time: 0.66s time: 119.46s eta: 9:27:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31401.0 tgs: 263 data_time: 0.77s time: 118.99s eta: 9:23:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31399.0 tgs: 263 data_time: 0.89s time: 119.04s eta: 9:21:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32114.0 tgs: 268 data_time: 0.79s time: 119.53s eta: 9:21:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.250 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30672.0 tgs: 256 data_time: 0.74s time: 119.60s eta: 9:20:06 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.249 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31852.0 tgs: 267 data_time: 0.70s time: 119.12s eta: 9:15:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32190.0 tgs: 270 data_time: 0.78s time: 119.00s eta: 9:13:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32116.0 tgs: 270 data_time: 0.77s time: 118.88s eta: 9:10:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31669.0 tgs: 265 data_time: 0.77s time: 119.32s eta: 9:10:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32085.0 tgs: 269 data_time: 0.74s time: 119.10s eta: 9:07:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31293.0 tgs: 263 data_time: 0.64s time: 118.85s eta: 9:04:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.236 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32301.0 tgs: 269 data_time: 0.65s time: 119.67s eta: 9:06:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31406.0 tgs: 262 data_time: 0.87s time: 119.44s eta: 9:03:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32474.0 tgs: 271 data_time: 1.00s time: 119.43s eta: 9:01:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.263 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32159.0 tgs: 270 data_time: 0.69s time: 118.74s eta: 8:56:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 268 data_time: 0.83s time: 118.87s eta: 8:54:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.231 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30793.0 tgs: 258 data_time: 1.14s time: 119.33s eta: 8:55:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31953.0 tgs: 268 data_time: 0.59s time: 119.14s eta: 8:52:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.215 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32205.0 tgs: 270 data_time: 0.84s time: 119.27s eta: 8:50:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.119 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30632.0 tgs: 254 data_time: 0.87s time: 120.47s eta: 8:54:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30900.0 tgs: 259 data_time: 0.79s time: 119.24s eta: 8:46:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.221 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30006.0 tgs: 252 data_time: 0.84s time: 119.05s eta: 8:43:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31535.0 tgs: 260 data_time: 0.96s time: 120.83s eta: 8:49:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31756.0 tgs: 264 data_time: 1.10s time: 120.20s eta: 8:44:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31801.0 tgs: 267 data_time: 0.73s time: 118.83s eta: 8:36:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.209 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31576.0 tgs: 264 data_time: 0.72s time: 119.35s eta: 8:37:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31993.0 tgs: 268 data_time: 0.82s time: 119.33s eta: 8:35:06 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30803.0 tgs: 258 data_time: 0.78s time: 119.06s eta: 8:31:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31699.0 tgs: 266 data_time: 0.95s time: 119.16s eta: 8:30:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 267 data_time: 0.89s time: 119.04s eta: 8:27:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30571.0 tgs: 256 data_time: 0.66s time: 119.12s eta: 8:26:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.095 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 268 data_time: 0.66s time: 119.25s eta: 8:24:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31835.0 tgs: 267 data_time: 0.77s time: 119.22s eta: 8:22:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30890.0 tgs: 258 data_time: 0.72s time: 119.38s eta: 8:21:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.222 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32161.0 tgs: 268 data_time: 0.71s time: 119.78s eta: 8:21:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32037.0 tgs: 268 data_time: 0.91s time: 119.20s eta: 8:16:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.113 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 270 data_time: 0.89s time: 118.98s eta: 8:13:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31232.0 tgs: 260 data_time: 0.66s time: 119.90s eta: 8:15:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.216 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32033.0 tgs: 267 data_time: 0.66s time: 119.56s eta: 8:12:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31812.0 tgs: 265 data_time: 0.73s time: 119.60s eta: 8:10:21 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.192 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31614.0 tgs: 265 data_time: 0.94s time: 119.00s eta: 8:05:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.108 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32011.0 tgs: 267 data_time: 0.88s time: 119.84s eta: 8:07:21 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30449.0 tgs: 254 data_time: 0.66s time: 119.71s eta: 8:04:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32024.0 tgs: 268 data_time: 0.90s time: 119.40s eta: 8:01:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.128 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 266 data_time: 0.75s time: 119.29s eta: 7:59:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31136.0 tgs: 261 data_time: 0.56s time: 119.20s eta: 7:56:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31894.0 tgs: 267 data_time: 0.65s time: 119.20s eta: 7:54:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31934.0 tgs: 268 data_time: 0.78s time: 118.93s eta: 7:51:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.112 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32077.0 tgs: 269 data_time: 0.59s time: 118.97s eta: 7:49:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31318.0 tgs: 262 data_time: 0.49s time: 119.42s eta: 7:49:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31449.0 tgs: 264 data_time: 0.52s time: 119.04s eta: 7:46:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31937.0 tgs: 268 data_time: 0.75s time: 119.08s eta: 7:44:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30900.0 tgs: 257 data_time: 0.59s time: 120.01s eta: 7:46:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30813.0 tgs: 257 data_time: 0.61s time: 119.83s eta: 7:43:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.219 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 267 data_time: 0.75s time: 119.16s eta: 7:38:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.216 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31897.0 tgs: 268 data_time: 1.00s time: 118.97s eta: 7:36:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32353.0 tgs: 271 data_time: 0.67s time: 119.32s eta: 7:35:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.188 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31187.0 tgs: 261 data_time: 0.71s time: 119.38s eta: 7:33:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.123 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30322.0 tgs: 254 data_time: 0.60s time: 119.06s eta: 7:30:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32137.0 tgs: 270 data_time: 0.83s time: 119.00s eta: 7:28:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32408.0 tgs: 271 data_time: 0.71s time: 119.24s eta: 7:27:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31679.0 tgs: 266 data_time: 0.72s time: 118.99s eta: 7:24:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31354.0 tgs: 263 data_time: 0.63s time: 118.93s eta: 7:22:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 267 data_time: 0.69s time: 119.08s eta: 7:20:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30398.0 tgs: 255 data_time: 0.75s time: 119.00s eta: 7:18:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31680.0 tgs: 266 data_time: 0.87s time: 119.03s eta: 7:16:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32127.0 tgs: 269 data_time: 0.52s time: 119.25s eta: 7:15:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31288.0 tgs: 262 data_time: 0.84s time: 119.33s eta: 7:13:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31463.0 tgs: 264 data_time: 0.79s time: 118.94s eta: 7:10:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31827.0 tgs: 266 data_time: 0.61s time: 119.21s eta: 7:09:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30654.0 tgs: 256 data_time: 0.66s time: 119.58s eta: 7:08:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.203 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32163.0 tgs: 268 data_time: 0.80s time: 119.66s eta: 7:06:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 269 data_time: 0.70s time: 118.95s eta: 7:02:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31987.0 tgs: 268 data_time: 0.70s time: 119.30s eta: 7:01:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32252.0 tgs: 269 data_time: 0.98s time: 119.52s eta: 7:00:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31399.0 tgs: 262 data_time: 0.60s time: 119.42s eta: 6:57:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31807.0 tgs: 267 data_time: 0.63s time: 118.92s eta: 6:54:14 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31101.0 tgs: 260 data_time: 0.62s time: 119.34s eta: 6:53:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 271 data_time: 0.67s time: 118.93s eta: 6:50:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32373.0 tgs: 271 data_time: 0.71s time: 119.27s eta: 6:49:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32287.0 tgs: 271 data_time: 0.82s time: 119.02s eta: 6:46:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 267 data_time: 0.78s time: 119.22s eta: 6:45:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32441.0 tgs: 271 data_time: 0.82s time: 119.39s eta: 6:43:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32347.0 tgs: 271 data_time: 0.65s time: 119.01s eta: 6:40:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31640.0 tgs: 265 data_time: 0.76s time: 119.02s eta: 6:38:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31846.0 tgs: 266 data_time: 0.76s time: 119.46s eta: 6:38:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31326.0 tgs: 263 data_time: 1.05s time: 118.80s eta: 6:34:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32044.0 tgs: 268 data_time: 1.11s time: 119.44s eta: 6:34:08 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32634.0 tgs: 272 data_time: 0.95s time: 119.57s eta: 6:32:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31786.0 tgs: 266 data_time: 0.98s time: 119.38s eta: 6:29:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31178.0 tgs: 260 data_time: 0.57s time: 119.83s eta: 6:29:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32170.0 tgs: 269 data_time: 0.78s time: 119.36s eta: 6:25:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 24874.0 tgs: 209 data_time: 0.61s time: 118.82s eta: 6:22:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31223.0 tgs: 262 data_time: 1.09s time: 119.02s eta: 6:20:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.210 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31896.0 tgs: 267 data_time: 0.72s time: 119.40s eta: 6:20:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.204 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 265 data_time: 1.03s time: 119.35s eta: 6:17:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.266 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32267.0 tgs: 271 data_time: 0.80s time: 119.04s eta: 6:14:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30689.0 tgs: 257 data_time: 0.65s time: 119.37s eta: 6:14:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.216 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31591.0 tgs: 264 data_time: 0.92s time: 119.39s eta: 6:12:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32333.0 tgs: 271 data_time: 0.83s time: 119.29s eta: 6:09:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32561.0 tgs: 273 data_time: 0.83s time: 119.10s eta: 6:07:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.219 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31355.0 tgs: 257 data_time: 0.85s time: 121.96s eta: 6:14:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31471.0 tgs: 263 data_time: 0.92s time: 119.39s eta: 6:04:07 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31113.0 tgs: 260 data_time: 0.62s time: 119.57s eta: 6:02:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31476.0 tgs: 263 data_time: 0.80s time: 119.26s eta: 5:59:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32008.0 tgs: 268 data_time: 0.67s time: 119.26s eta: 5:57:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30858.0 tgs: 259 data_time: 0.77s time: 118.92s eta: 5:54:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32235.0 tgs: 270 data_time: 0.77s time: 119.02s eta: 5:53:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32064.0 tgs: 269 data_time: 0.93s time: 119.08s eta: 5:51:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31987.0 tgs: 268 data_time: 0.75s time: 119.28s eta: 5:49:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31957.0 tgs: 268 data_time: 0.63s time: 118.97s eta: 5:47:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31133.0 tgs: 261 data_time: 0.74s time: 119.27s eta: 5:45:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.221 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30631.0 tgs: 256 data_time: 0.63s time: 119.33s eta: 5:44:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31426.0 tgs: 263 data_time: 0.76s time: 119.05s eta: 5:41:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32153.0 tgs: 269 data_time: 1.05s time: 119.50s eta: 5:40:33 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.111 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31319.0 tgs: 262 data_time: 0.68s time: 119.39s eta: 5:38:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31909.0 tgs: 266 data_time: 1.01s time: 119.57s eta: 5:36:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32066.0 tgs: 269 data_time: 0.62s time: 118.98s eta: 5:33:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32033.0 tgs: 269 data_time: 1.00s time: 118.95s eta: 5:31:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32461.0 tgs: 272 data_time: 0.56s time: 119.24s eta: 5:29:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30377.0 tgs: 254 data_time: 0.61s time: 119.41s eta: 5:28:21 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30281.0 tgs: 253 data_time: 1.27s time: 119.46s eta: 5:26:31 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31751.0 tgs: 266 data_time: 1.02s time: 118.95s eta: 5:23:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31805.0 tgs: 265 data_time: 0.77s time: 119.63s eta: 5:23:00 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 264 data_time: 0.88s time: 119.35s eta: 5:20:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32084.0 tgs: 269 data_time: 0.67s time: 119.12s eta: 5:17:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31944.0 tgs: 267 data_time: 0.81s time: 119.43s eta: 5:16:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31744.0 tgs: 265 data_time: 0.65s time: 119.55s eta: 5:14:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.113 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31616.0 tgs: 265 data_time: 0.54s time: 119.01s eta: 5:11:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31579.0 tgs: 265 data_time: 0.80s time: 119.00s eta: 5:09:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31593.0 tgs: 264 data_time: 0.63s time: 119.24s eta: 5:08:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32220.0 tgs: 269 data_time: 0.57s time: 119.68s eta: 5:07:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.233 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31676.0 tgs: 265 data_time: 0.91s time: 119.11s eta: 5:03:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.210 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32244.0 tgs: 270 data_time: 0.83s time: 119.08s eta: 5:01:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.203 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31149.0 tgs: 256 data_time: 1.07s time: 121.52s eta: 5:05:48 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.185 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31756.0 tgs: 265 data_time: 0.71s time: 119.58s eta: 4:58:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32134.0 tgs: 268 data_time: 0.61s time: 119.50s eta: 4:56:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.110 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31305.0 tgs: 261 data_time: 0.78s time: 119.73s eta: 4:55:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31693.0 tgs: 265 data_time: 0.83s time: 119.52s eta: 4:52:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.179 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31100.0 tgs: 260 data_time: 0.77s time: 119.42s eta: 4:50:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31410.0 tgs: 263 data_time: 0.71s time: 119.20s eta: 4:48:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32255.0 tgs: 271 data_time: 0.64s time: 118.89s eta: 4:45:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.202 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32064.0 tgs: 269 data_time: 0.78s time: 119.07s eta: 4:43:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31373.0 tgs: 261 data_time: 0.70s time: 119.98s eta: 4:43:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31936.0 tgs: 268 data_time: 0.93s time: 119.14s eta: 4:39:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31384.0 tgs: 263 data_time: 0.67s time: 119.01s eta: 4:37:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31522.0 tgs: 264 data_time: 0.88s time: 119.07s eta: 4:35:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.211 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31368.0 tgs: 262 data_time: 0.95s time: 119.37s eta: 4:34:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.212 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32319.0 tgs: 270 data_time: 0.83s time: 119.41s eta: 4:32:39 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.187 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32097.0 tgs: 269 data_time: 0.77s time: 118.94s eta: 4:29:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31888.0 tgs: 266 data_time: 0.67s time: 119.60s eta: 4:29:05 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32356.0 tgs: 271 data_time: 0.88s time: 119.24s eta: 4:26:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31647.0 tgs: 265 data_time: 0.74s time: 119.06s eta: 4:23:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32128.0 tgs: 270 data_time: 0.84s time: 118.85s eta: 4:21:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 267 data_time: 0.80s time: 118.76s eta: 4:19:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32505.0 tgs: 272 data_time: 0.70s time: 119.43s eta: 4:18:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31670.0 tgs: 265 data_time: 0.92s time: 119.41s eta: 4:16:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31186.0 tgs: 261 data_time: 0.74s time: 119.31s eta: 4:14:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31417.0 tgs: 263 data_time: 0.82s time: 119.05s eta: 4:11:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32011.0 tgs: 268 data_time: 0.82s time: 119.12s eta: 4:10:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32385.0 tgs: 271 data_time: 0.70s time: 119.14s eta: 4:08:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32176.0 tgs: 270 data_time: 0.56s time: 118.96s eta: 4:05:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.235 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31460.0 tgs: 264 data_time: 0.58s time: 119.15s eta: 4:04:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31093.0 tgs: 260 data_time: 0.70s time: 119.50s eta: 4:02:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31674.0 tgs: 265 data_time: 0.69s time: 119.25s eta: 4:00:28 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32105.0 tgs: 269 data_time: 1.05s time: 119.20s eta: 3:58:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.84s time: 119.36s eta: 3:56:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.111 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30784.0 tgs: 257 data_time: 0.73s time: 119.37s eta: 3:54:45 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31552.0 tgs: 263 data_time: 0.68s time: 119.96s eta: 3:53:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31980.0 tgs: 268 data_time: 0.79s time: 119.32s eta: 3:50:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32239.0 tgs: 269 data_time: 1.06s time: 119.63s eta: 3:49:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31951.0 tgs: 268 data_time: 0.86s time: 118.86s eta: 3:45:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31117.0 tgs: 259 data_time: 0.79s time: 119.71s eta: 3:45:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31415.0 tgs: 263 data_time: 0.70s time: 119.33s eta: 3:42:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32019.0 tgs: 268 data_time: 0.82s time: 119.26s eta: 3:40:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32188.0 tgs: 270 data_time: 0.68s time: 119.00s eta: 3:38:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 269 data_time: 0.86s time: 119.43s eta: 3:36:57 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31305.0 tgs: 262 data_time: 0.83s time: 119.04s eta: 3:34:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32324.0 tgs: 271 data_time: 0.81s time: 118.99s eta: 3:32:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30983.0 tgs: 259 data_time: 0.99s time: 119.22s eta: 3:30:37 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31605.0 tgs: 263 data_time: 0.97s time: 120.11s eta: 3:30:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31630.0 tgs: 265 data_time: 0.59s time: 119.14s eta: 3:26:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.68s time: 119.31s eta: 3:24:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31458.0 tgs: 264 data_time: 0.91s time: 119.08s eta: 3:22:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31979.0 tgs: 267 data_time: 0.69s time: 119.42s eta: 3:21:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.195 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32077.0 tgs: 268 data_time: 0.83s time: 119.29s eta: 3:18:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31331.0 tgs: 262 data_time: 0.48s time: 119.22s eta: 3:16:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31735.0 tgs: 266 data_time: 0.90s time: 119.13s eta: 3:14:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31899.0 tgs: 268 data_time: 1.03s time: 119.02s eta: 3:12:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31291.0 tgs: 262 data_time: 0.76s time: 119.18s eta: 3:10:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.196 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31418.0 tgs: 259 data_time: 0.78s time: 121.10s eta: 3:11:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.225 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31506.0 tgs: 263 data_time: 0.84s time: 119.51s eta: 3:07:13 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 266 data_time: 0.88s time: 119.26s eta: 3:04:51 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31784.0 tgs: 267 data_time: 0.62s time: 118.99s eta: 3:02:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 266 data_time: 0.86s time: 119.22s eta: 3:00:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31222.0 tgs: 261 data_time: 0.73s time: 119.46s eta: 2:59:10 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31291.0 tgs: 262 data_time: 0.59s time: 118.98s eta: 2:56:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 267 data_time: 0.56s time: 119.35s eta: 2:55:03 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31608.0 tgs: 264 data_time: 0.90s time: 119.41s eta: 2:53:09 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32078.0 tgs: 268 data_time: 0.88s time: 119.35s eta: 2:51:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31859.0 tgs: 262 data_time: 0.64s time: 121.46s eta: 2:52:04 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31180.0 tgs: 261 data_time: 0.69s time: 119.24s eta: 2:46:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.247 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32136.0 tgs: 268 data_time: 0.69s time: 119.56s eta: 2:45:23 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31871.0 tgs: 268 data_time: 0.96s time: 118.91s eta: 2:42:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31776.0 tgs: 266 data_time: 0.73s time: 119.19s eta: 2:40:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31397.0 tgs: 263 data_time: 0.78s time: 118.96s eta: 2:38:37 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30511.0 tgs: 255 data_time: 0.79s time: 119.23s eta: 2:36:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32145.0 tgs: 269 data_time: 0.89s time: 119.11s eta: 2:34:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.79s time: 119.44s eta: 2:33:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31459.0 tgs: 261 data_time: 0.75s time: 120.22s eta: 2:32:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31223.0 tgs: 261 data_time: 0.70s time: 119.28s eta: 2:29:06 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31341.0 tgs: 263 data_time: 0.85s time: 118.88s eta: 2:26:36 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31412.0 tgs: 262 data_time: 0.72s time: 119.52s eta: 2:25:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30705.0 tgs: 257 data_time: 0.86s time: 119.07s eta: 2:22:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31645.0 tgs: 265 data_time: 0.88s time: 119.14s eta: 2:20:58 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30896.0 tgs: 259 data_time: 0.46s time: 119.05s eta: 2:18:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31051.0 tgs: 259 data_time: 1.15s time: 119.85s eta: 2:17:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.186 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30866.0 tgs: 258 data_time: 0.81s time: 119.31s eta: 2:15:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 266 data_time: 0.78s time: 119.36s eta: 2:13:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32431.0 tgs: 270 data_time: 0.67s time: 119.68s eta: 2:11:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29685.0 tgs: 248 data_time: 0.91s time: 119.49s eta: 2:09:27 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.208 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32066.0 tgs: 269 data_time: 0.72s time: 118.96s eta: 2:06:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32403.0 tgs: 271 data_time: 0.72s time: 119.38s eta: 2:05:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30772.0 tgs: 257 data_time: 0.67s time: 119.28s eta: 2:03:15 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32032.0 tgs: 268 data_time: 0.60s time: 119.22s eta: 2:01:12 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31043.0 tgs: 259 data_time: 0.70s time: 119.51s eta: 1:59:30 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31771.0 tgs: 266 data_time: 0.80s time: 119.43s eta: 1:57:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31827.0 tgs: 266 data_time: 0.71s time: 119.53s eta: 1:55:32 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31700.0 tgs: 265 data_time: 1.04s time: 119.31s eta: 1:53:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 267 data_time: 0.66s time: 119.25s eta: 1:51:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31466.0 tgs: 264 data_time: 0.90s time: 118.94s eta: 1:49:01 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31806.0 tgs: 265 data_time: 0.74s time: 119.64s eta: 1:47:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31582.0 tgs: 263 data_time: 0.70s time: 119.71s eta: 1:45:44 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32147.0 tgs: 269 data_time: 0.73s time: 119.35s eta: 1:43:26 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31162.0 tgs: 261 data_time: 0.51s time: 119.31s eta: 1:41:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31499.0 tgs: 264 data_time: 0.68s time: 119.28s eta: 1:39:24 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30221.0 tgs: 254 data_time: 0.62s time: 118.91s eta: 1:37:06 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31314.0 tgs: 262 data_time: 0.59s time: 119.16s eta: 1:35:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 268 data_time: 0.89s time: 119.07s eta: 1:33:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31976.0 tgs: 268 data_time: 0.87s time: 119.25s eta: 1:31:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 267 data_time: 0.88s time: 119.72s eta: 1:29:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31717.0 tgs: 266 data_time: 1.03s time: 119.02s eta: 1:27:16 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31818.0 tgs: 267 data_time: 0.80s time: 118.86s eta: 1:25:11 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31310.0 tgs: 262 data_time: 0.76s time: 119.48s eta: 1:23:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32385.0 tgs: 271 data_time: 0.71s time: 119.48s eta: 1:21:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 269 data_time: 0.84s time: 118.95s eta: 1:19:17 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 265 data_time: 0.88s time: 118.93s eta: 1:17:18 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32354.0 tgs: 264 data_time: 0.76s time: 122.13s eta: 1:17:20 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31079.0 tgs: 260 data_time: 0.73s time: 119.33s eta: 1:13:35 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32444.0 tgs: 272 data_time: 0.66s time: 119.05s eta: 1:11:25 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31400.0 tgs: 264 data_time: 0.63s time: 118.92s eta: 1:09:22 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31319.0 tgs: 262 data_time: 0.89s time: 119.11s eta: 1:07:29 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32039.0 tgs: 268 data_time: 0.81s time: 119.43s eta: 1:05:41 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31656.0 tgs: 265 data_time: 0.61s time: 119.34s eta: 1:03:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32223.0 tgs: 270 data_time: 0.81s time: 119.18s eta: 1:01:34 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31367.0 tgs: 262 data_time: 0.68s time: 119.27s eta: 0:59:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31971.0 tgs: 267 data_time: 0.67s time: 119.44s eta: 0:57:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29547.0 tgs: 247 data_time: 0.91s time: 119.40s eta: 0:55:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32397.0 tgs: 271 data_time: 0.92s time: 119.49s eta: 0:53:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.210 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32196.0 tgs: 269 data_time: 0.86s time: 119.37s eta: 0:51:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31589.0 tgs: 264 data_time: 0.88s time: 119.47s eta: 0:49:46 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30464.0 tgs: 256 data_time: 0.81s time: 118.89s eta: 0:47:33 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31972.0 tgs: 268 data_time: 0.74s time: 119.27s eta: 0:45:43 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31136.0 tgs: 261 data_time: 0.49s time: 119.21s eta: 0:43:42 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32012.0 tgs: 269 data_time: 0.93s time: 118.99s eta: 0:41:38 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30989.0 tgs: 260 data_time: 0.87s time: 119.04s eta: 0:39:40 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31928.0 tgs: 266 data_time: 0.65s time: 119.66s eta: 0:37:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32156.0 tgs: 269 data_time: 0.79s time: 119.29s eta: 0:35:47 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31952.0 tgs: 266 data_time: 0.57s time: 119.75s eta: 0:33:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.208 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30242.0 tgs: 249 data_time: 0.60s time: 121.23s eta: 0:32:19 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.69s time: 119.38s eta: 0:29:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31223.0 tgs: 261 data_time: 0.65s time: 119.32s eta: 0:27:50 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31559.0 tgs: 263 data_time: 0.77s time: 119.64s eta: 0:25:55 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32066.0 tgs: 269 data_time: 0.90s time: 119.09s eta: 0:23:49 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31116.0 tgs: 260 data_time: 0.82s time: 119.39s eta: 0:21:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32287.0 tgs: 270 data_time: 0.90s time: 119.26s eta: 0:19:52 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31651.0 tgs: 265 data_time: 0.78s time: 119.24s eta: 0:17:53 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31776.0 tgs: 265 data_time: 0.60s time: 119.90s eta: 0:15:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31038.0 tgs: 260 data_time: 0.64s time: 119.18s eta: 0:13:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 270 data_time: 1.07s time: 119.09s eta: 0:11:54 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32360.0 tgs: 271 data_time: 0.87s time: 119.33s eta: 0:09:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32444.0 tgs: 272 data_time: 0.81s time: 119.17s eta: 0:07:56 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31756.0 tgs: 265 data_time: 0.69s time: 119.69s eta: 0:05:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31337.0 tgs: 258 data_time: 0.79s time: 121.42s eta: 0:04:02 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31395.0 tgs: 263 data_time: 0.96s time: 119.21s eta: 0:01:59 +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.29 seconds, peak gpu memory 17.4G +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 18][DP 4][SP 2][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank21.log b/20250121165312/rank21.log new file mode 100644 index 0000000000000000000000000000000000000000..8d6e1c9a8b7aba84438cb5a669a5cc47f01d4c1d --- /dev/null +++ b/20250121165312/rank21.log @@ -0,0 +1,851 @@ +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.07s +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.84 seconds, peak gpu memory 13.4G +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.211 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.3GB text_tokens: 29620.0 tgs: 195 data_time: 1.86s time: 151.23s eta: 1 day, 0:54:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.256 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32369.0 tgs: 269 data_time: 0.84s time: 120.10s eta: 19:45:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.241 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32546.0 tgs: 271 data_time: 0.86s time: 119.68s eta: 19:38:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.238 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 270 data_time: 1.07s time: 119.66s eta: 19:36:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.235 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30927.0 tgs: 259 data_time: 0.95s time: 119.31s eta: 19:31:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.223 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32427.0 tgs: 271 data_time: 1.21s time: 119.53s eta: 19:31:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.207 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31527.0 tgs: 256 data_time: 1.12s time: 122.97s eta: 20:03:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.272 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31799.0 tgs: 265 data_time: 0.70s time: 119.89s eta: 19:30:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.210 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30776.0 tgs: 255 data_time: 0.73s time: 120.65s eta: 19:36:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31699.0 tgs: 264 data_time: 0.74s time: 119.92s eta: 19:27:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 261 data_time: 0.97s time: 120.46s eta: 19:30:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31345.0 tgs: 262 data_time: 0.69s time: 119.35s eta: 19:17:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.249 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30795.0 tgs: 257 data_time: 0.77s time: 119.68s eta: 19:18:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.260 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 262 data_time: 0.75s time: 119.17s eta: 19:12:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32265.0 tgs: 270 data_time: 0.85s time: 119.22s eta: 19:10:28 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29937.0 tgs: 251 data_time: 0.79s time: 118.93s eta: 19:05:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30786.0 tgs: 258 data_time: 0.75s time: 119.15s eta: 19:05:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31670.0 tgs: 265 data_time: 0.85s time: 119.18s eta: 19:04:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31374.0 tgs: 262 data_time: 0.62s time: 119.40s eta: 19:04:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.241 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32080.0 tgs: 268 data_time: 0.77s time: 119.33s eta: 19:01:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30100.0 tgs: 250 data_time: 0.79s time: 120.31s eta: 19:08:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31525.0 tgs: 256 data_time: 0.92s time: 122.92s eta: 19:31:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.288 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32200.0 tgs: 269 data_time: 0.93s time: 119.48s eta: 18:57:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31533.0 tgs: 264 data_time: 1.00s time: 119.16s eta: 18:51:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31572.0 tgs: 264 data_time: 0.77s time: 119.25s eta: 18:50:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31711.0 tgs: 265 data_time: 0.60s time: 119.39s eta: 18:50:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31027.0 tgs: 259 data_time: 0.81s time: 119.66s eta: 18:50:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.243 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31174.0 tgs: 261 data_time: 1.03s time: 119.39s eta: 18:46:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 265 data_time: 0.91s time: 119.76s eta: 18:47:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 246 data_time: 0.94s time: 129.30s eta: 20:15:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31822.0 tgs: 266 data_time: 0.88s time: 119.57s eta: 18:42:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31024.0 tgs: 259 data_time: 0.79s time: 119.38s eta: 18:38:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31792.0 tgs: 263 data_time: 1.03s time: 120.66s eta: 18:48:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31319.0 tgs: 262 data_time: 0.63s time: 119.40s eta: 18:34:21 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 270 data_time: 0.64s time: 119.16s eta: 18:30:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32359.0 tgs: 271 data_time: 0.92s time: 119.07s eta: 18:27:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31433.0 tgs: 263 data_time: 0.93s time: 119.14s eta: 18:25:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30457.0 tgs: 252 data_time: 0.90s time: 120.39s eta: 18:35:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 264 data_time: 0.88s time: 121.18s eta: 18:40:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 256 data_time: 0.67s time: 120.66s eta: 18:34:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32067.0 tgs: 267 data_time: 0.88s time: 119.74s eta: 18:23:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31896.0 tgs: 257 data_time: 0.96s time: 123.88s eta: 18:59:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30763.0 tgs: 257 data_time: 0.76s time: 119.56s eta: 18:17:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31487.0 tgs: 263 data_time: 0.91s time: 119.32s eta: 18:13:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31702.0 tgs: 265 data_time: 0.74s time: 119.22s eta: 18:10:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31920.0 tgs: 266 data_time: 0.94s time: 119.60s eta: 18:12:21 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 266 data_time: 0.78s time: 119.27s eta: 18:07:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.134 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31065.0 tgs: 261 data_time: 0.78s time: 119.01s eta: 18:03:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32174.0 tgs: 270 data_time: 0.79s time: 119.10s eta: 18:01:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31306.0 tgs: 262 data_time: 1.07s time: 119.26s eta: 18:01:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29824.0 tgs: 249 data_time: 0.69s time: 119.33s eta: 17:59:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31180.0 tgs: 259 data_time: 1.09s time: 120.03s eta: 18:04:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31929.0 tgs: 268 data_time: 0.89s time: 119.10s eta: 17:53:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32302.0 tgs: 270 data_time: 0.58s time: 119.41s eta: 17:54:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31344.0 tgs: 263 data_time: 0.72s time: 119.00s eta: 17:48:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30272.0 tgs: 254 data_time: 0.79s time: 119.03s eta: 17:47:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32061.0 tgs: 255 data_time: 0.72s time: 125.36s eta: 18:42:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 265 data_time: 0.69s time: 120.11s eta: 17:52:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32402.0 tgs: 271 data_time: 1.26s time: 119.36s eta: 17:44:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30488.0 tgs: 255 data_time: 0.88s time: 119.33s eta: 17:42:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 266 data_time: 0.59s time: 119.52s eta: 17:41:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31778.0 tgs: 252 data_time: 0.59s time: 126.05s eta: 18:37:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 259 data_time: 0.97s time: 122.63s eta: 18:05:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.278 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 264 data_time: 1.39s time: 121.15s eta: 17:50:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30998.0 tgs: 260 data_time: 0.69s time: 119.04s eta: 17:29:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.226 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31047.0 tgs: 261 data_time: 0.96s time: 118.86s eta: 17:25:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32302.0 tgs: 260 data_time: 0.88s time: 124.00s eta: 18:09:09 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30866.0 tgs: 247 data_time: 0.64s time: 124.57s eta: 18:12:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31849.0 tgs: 266 data_time: 0.75s time: 119.40s eta: 17:24:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31641.0 tgs: 264 data_time: 0.83s time: 119.70s eta: 17:25:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32468.0 tgs: 271 data_time: 0.76s time: 119.53s eta: 17:21:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31006.0 tgs: 260 data_time: 0.72s time: 118.86s eta: 17:14:07 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.213 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31253.0 tgs: 260 data_time: 0.77s time: 119.85s eta: 17:20:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31654.0 tgs: 264 data_time: 0.98s time: 119.51s eta: 17:15:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.249 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31945.0 tgs: 262 data_time: 0.78s time: 121.86s eta: 17:34:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 253 data_time: 0.59s time: 123.52s eta: 17:46:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30448.0 tgs: 255 data_time: 0.63s time: 119.01s eta: 17:05:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31752.0 tgs: 266 data_time: 0.90s time: 119.23s eta: 17:05:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 270 data_time: 0.70s time: 119.14s eta: 17:02:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.239 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32172.0 tgs: 269 data_time: 0.99s time: 119.26s eta: 17:01:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31208.0 tgs: 262 data_time: 0.97s time: 119.05s eta: 16:57:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32488.0 tgs: 272 data_time: 0.84s time: 119.28s eta: 16:57:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31692.0 tgs: 265 data_time: 0.93s time: 119.16s eta: 16:54:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 266 data_time: 0.88s time: 119.59s eta: 16:56:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32503.0 tgs: 272 data_time: 1.00s time: 119.13s eta: 16:50:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31637.0 tgs: 266 data_time: 0.71s time: 118.84s eta: 16:46:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.241 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32396.0 tgs: 272 data_time: 0.78s time: 119.02s eta: 16:45:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31701.0 tgs: 264 data_time: 0.66s time: 119.85s eta: 16:50:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32363.0 tgs: 271 data_time: 0.96s time: 119.38s eta: 16:44:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31309.0 tgs: 263 data_time: 0.56s time: 119.00s eta: 16:39:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 262 data_time: 1.06s time: 120.93s eta: 16:53:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.215 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31214.0 tgs: 262 data_time: 0.68s time: 118.99s eta: 16:35:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31653.0 tgs: 265 data_time: 0.87s time: 119.29s eta: 16:36:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.207 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32264.0 tgs: 270 data_time: 0.95s time: 119.23s eta: 16:33:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32112.0 tgs: 269 data_time: 0.97s time: 119.36s eta: 16:32:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31448.0 tgs: 262 data_time: 0.81s time: 119.70s eta: 16:33:32 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.102 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31670.0 tgs: 265 data_time: 0.78s time: 119.20s eta: 16:27:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32024.0 tgs: 267 data_time: 0.61s time: 119.76s eta: 16:30:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.216 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31293.0 tgs: 259 data_time: 0.76s time: 120.37s eta: 16:33:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31374.0 tgs: 263 data_time: 0.74s time: 118.97s eta: 16:19:33 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30907.0 tgs: 242 data_time: 0.87s time: 127.33s eta: 17:26:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.217 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 267 data_time: 0.64s time: 119.23s eta: 16:17:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31295.0 tgs: 262 data_time: 0.65s time: 119.14s eta: 16:14:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.249 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31271.0 tgs: 262 data_time: 0.76s time: 119.21s eta: 16:13:32 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29928.0 tgs: 251 data_time: 0.57s time: 119.18s eta: 16:11:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31218.0 tgs: 262 data_time: 0.77s time: 118.95s eta: 16:07:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 267 data_time: 0.81s time: 118.86s eta: 16:04:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32258.0 tgs: 271 data_time: 0.72s time: 118.96s eta: 16:03:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 259 data_time: 0.80s time: 121.85s eta: 16:24:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 268 data_time: 1.03s time: 118.99s eta: 15:59:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30581.0 tgs: 256 data_time: 0.91s time: 119.23s eta: 15:59:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32435.0 tgs: 271 data_time: 0.55s time: 119.61s eta: 16:00:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31232.0 tgs: 262 data_time: 0.95s time: 119.04s eta: 15:54:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31684.0 tgs: 266 data_time: 1.07s time: 118.98s eta: 15:51:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 267 data_time: 0.83s time: 119.17s eta: 15:51:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.241 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32043.0 tgs: 267 data_time: 0.61s time: 119.64s eta: 15:53:09 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31617.0 tgs: 265 data_time: 0.91s time: 118.93s eta: 15:45:31 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31764.0 tgs: 266 data_time: 0.59s time: 119.29s eta: 15:46:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32084.0 tgs: 269 data_time: 1.01s time: 119.11s eta: 15:42:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30935.0 tgs: 259 data_time: 0.94s time: 119.00s eta: 15:40:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32150.0 tgs: 268 data_time: 0.82s time: 119.66s eta: 15:43:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 262 data_time: 0.84s time: 121.51s eta: 15:55:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31485.0 tgs: 263 data_time: 0.57s time: 119.41s eta: 15:37:21 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32015.0 tgs: 246 data_time: 0.85s time: 129.97s eta: 16:58:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31263.0 tgs: 261 data_time: 0.95s time: 119.47s eta: 15:33:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30780.0 tgs: 257 data_time: 0.72s time: 119.60s eta: 15:32:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32045.0 tgs: 268 data_time: 0.85s time: 119.53s eta: 15:30:21 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.218 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 259 data_time: 0.76s time: 122.14s eta: 15:48:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30999.0 tgs: 256 data_time: 0.90s time: 120.76s eta: 15:35:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31669.0 tgs: 265 data_time: 1.02s time: 119.41s eta: 15:23:28 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31113.0 tgs: 260 data_time: 0.71s time: 119.45s eta: 15:21:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 237 data_time: 0.90s time: 131.63s eta: 16:53:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 265 data_time: 0.93s time: 119.29s eta: 15:16:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.124 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 264 data_time: 0.57s time: 119.53s eta: 15:16:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32096.0 tgs: 262 data_time: 0.75s time: 122.25s eta: 15:35:10 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31552.0 tgs: 263 data_time: 0.74s time: 119.72s eta: 15:13:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.81s time: 119.35s eta: 15:09:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30937.0 tgs: 260 data_time: 0.92s time: 118.90s eta: 15:03:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32455.0 tgs: 272 data_time: 1.03s time: 119.00s eta: 15:02:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.210 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31945.0 tgs: 267 data_time: 0.81s time: 119.31s eta: 15:02:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 265 data_time: 0.70s time: 118.92s eta: 14:57:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30856.0 tgs: 256 data_time: 0.67s time: 120.52s eta: 15:07:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31360.0 tgs: 262 data_time: 0.84s time: 119.31s eta: 14:56:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32316.0 tgs: 263 data_time: 0.72s time: 122.66s eta: 15:19:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.72s time: 119.13s eta: 14:51:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.119 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30701.0 tgs: 256 data_time: 0.70s time: 119.76s eta: 14:54:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 265 data_time: 0.75s time: 119.46s eta: 14:50:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 270 data_time: 0.80s time: 118.99s eta: 14:44:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.61s time: 119.30s eta: 14:44:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32218.0 tgs: 270 data_time: 1.02s time: 118.93s eta: 14:40:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31462.0 tgs: 264 data_time: 0.77s time: 119.05s eta: 14:39:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.260 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30925.0 tgs: 259 data_time: 0.88s time: 119.06s eta: 14:37:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32210.0 tgs: 270 data_time: 0.71s time: 119.13s eta: 14:35:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.62s time: 119.00s eta: 14:32:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31753.0 tgs: 266 data_time: 0.96s time: 119.08s eta: 14:31:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32043.0 tgs: 268 data_time: 0.87s time: 119.15s eta: 14:29:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 269 data_time: 1.05s time: 119.65s eta: 14:31:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.253 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30864.0 tgs: 258 data_time: 1.12s time: 119.18s eta: 14:26:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.79s time: 119.20s eta: 14:24:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31553.0 tgs: 264 data_time: 0.67s time: 119.39s eta: 14:23:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30653.0 tgs: 253 data_time: 0.77s time: 120.74s eta: 14:31:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.236 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31467.0 tgs: 264 data_time: 0.82s time: 118.85s eta: 14:15:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 269 data_time: 0.76s time: 119.10s eta: 14:15:33 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31113.0 tgs: 261 data_time: 0.88s time: 119.12s eta: 14:13:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31202.0 tgs: 260 data_time: 0.50s time: 119.96s eta: 14:17:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 268 data_time: 1.01s time: 119.87s eta: 14:15:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31916.0 tgs: 266 data_time: 0.74s time: 119.59s eta: 14:11:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31679.0 tgs: 265 data_time: 0.78s time: 119.22s eta: 14:06:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.82s time: 119.27s eta: 14:04:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31455.0 tgs: 263 data_time: 0.82s time: 119.27s eta: 14:02:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31473.0 tgs: 264 data_time: 0.98s time: 119.14s eta: 13:59:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31878.0 tgs: 267 data_time: 0.83s time: 119.12s eta: 13:57:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31593.0 tgs: 265 data_time: 0.91s time: 119.18s eta: 13:56:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.195 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31814.0 tgs: 266 data_time: 0.86s time: 119.35s eta: 13:55:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31236.0 tgs: 260 data_time: 0.71s time: 120.11s eta: 13:58:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.321 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31467.0 tgs: 256 data_time: 1.02s time: 122.56s eta: 14:13:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32419.0 tgs: 272 data_time: 0.55s time: 119.11s eta: 13:47:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31622.0 tgs: 264 data_time: 1.00s time: 119.37s eta: 13:47:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 270 data_time: 0.80s time: 119.32s eta: 13:45:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31542.0 tgs: 264 data_time: 0.66s time: 119.32s eta: 13:43:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31234.0 tgs: 262 data_time: 0.68s time: 119.12s eta: 13:39:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 268 data_time: 0.73s time: 119.25s eta: 13:38:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.232 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.93s time: 119.14s eta: 13:36:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.272 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 1.13s time: 119.23s eta: 13:34:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32114.0 tgs: 268 data_time: 0.74s time: 119.46s eta: 13:34:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30360.0 tgs: 254 data_time: 0.60s time: 119.08s eta: 13:29:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31389.0 tgs: 263 data_time: 0.98s time: 118.93s eta: 13:26:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.203 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31445.0 tgs: 262 data_time: 0.91s time: 119.80s eta: 13:30:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32275.0 tgs: 270 data_time: 0.67s time: 119.25s eta: 13:24:58 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32417.0 tgs: 272 data_time: 0.80s time: 119.17s eta: 13:22:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.81s time: 119.35s eta: 13:21:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31655.0 tgs: 265 data_time: 0.69s time: 119.26s eta: 13:19:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32283.0 tgs: 271 data_time: 0.88s time: 119.04s eta: 13:15:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32317.0 tgs: 270 data_time: 0.74s time: 119.35s eta: 13:15:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.78s time: 119.17s eta: 13:12:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.180 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32616.0 tgs: 273 data_time: 0.86s time: 119.46s eta: 13:12:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30892.0 tgs: 258 data_time: 0.76s time: 119.29s eta: 13:09:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31394.0 tgs: 263 data_time: 0.89s time: 119.15s eta: 13:06:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 266 data_time: 0.80s time: 120.35s eta: 13:12:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32445.0 tgs: 270 data_time: 0.65s time: 120.16s eta: 13:09:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.115 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31252.0 tgs: 261 data_time: 0.70s time: 119.51s eta: 13:02:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.115 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 266 data_time: 0.85s time: 119.51s eta: 13:00:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.202 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 270 data_time: 0.79s time: 119.57s eta: 12:59:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31257.0 tgs: 260 data_time: 0.79s time: 119.80s eta: 12:58:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30989.0 tgs: 260 data_time: 0.80s time: 119.04s eta: 12:51:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31367.0 tgs: 263 data_time: 0.83s time: 119.05s eta: 12:49:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 266 data_time: 0.87s time: 119.12s eta: 12:48:21 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.211 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30623.0 tgs: 256 data_time: 0.73s time: 119.40s eta: 12:48:07 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.248 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 267 data_time: 0.74s time: 119.18s eta: 12:44:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 0.65s time: 119.42s eta: 12:44:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.77s time: 119.10s eta: 12:40:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.214 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31061.0 tgs: 260 data_time: 0.70s time: 119.07s eta: 12:38:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31784.0 tgs: 265 data_time: 0.95s time: 119.59s eta: 12:39:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 270 data_time: 0.65s time: 118.78s eta: 12:32:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.225 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 270 data_time: 0.93s time: 118.88s eta: 12:30:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30989.0 tgs: 260 data_time: 0.87s time: 119.17s eta: 12:30:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.72s time: 119.27s eta: 12:29:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31282.0 tgs: 262 data_time: 0.75s time: 119.35s eta: 12:27:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32386.0 tgs: 271 data_time: 0.77s time: 119.12s eta: 12:24:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 270 data_time: 0.85s time: 119.35s eta: 12:23:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.79s time: 119.26s eta: 12:21:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.111 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 268 data_time: 0.85s time: 119.03s eta: 12:18:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32053.0 tgs: 264 data_time: 0.86s time: 121.13s eta: 12:29:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.220 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31183.0 tgs: 259 data_time: 0.72s time: 119.98s eta: 12:19:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.108 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30267.0 tgs: 255 data_time: 0.91s time: 118.69s eta: 12:09:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31274.0 tgs: 261 data_time: 0.71s time: 119.64s eta: 12:13:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32119.0 tgs: 268 data_time: 0.79s time: 119.83s eta: 12:12:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32165.0 tgs: 269 data_time: 0.79s time: 119.14s eta: 12:06:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31657.0 tgs: 265 data_time: 0.74s time: 119.17s eta: 12:04:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31215.0 tgs: 262 data_time: 0.90s time: 119.12s eta: 12:02:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32433.0 tgs: 272 data_time: 0.94s time: 119.18s eta: 12:01:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.185 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31078.0 tgs: 261 data_time: 0.89s time: 118.96s eta: 11:57:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31699.0 tgs: 264 data_time: 1.04s time: 119.71s eta: 12:00:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31777.0 tgs: 266 data_time: 0.77s time: 119.28s eta: 11:55:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.123 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32039.0 tgs: 268 data_time: 0.93s time: 119.28s eta: 11:53:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32039.0 tgs: 267 data_time: 0.63s time: 119.71s eta: 11:54:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32191.0 tgs: 271 data_time: 0.73s time: 118.78s eta: 11:46:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32174.0 tgs: 270 data_time: 0.71s time: 119.16s eta: 11:47:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31477.0 tgs: 263 data_time: 0.73s time: 119.61s eta: 11:47:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31088.0 tgs: 261 data_time: 0.88s time: 118.83s eta: 11:41:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.215 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 268 data_time: 0.71s time: 118.69s eta: 11:38:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.221 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31933.0 tgs: 267 data_time: 1.19s time: 119.21s eta: 11:39:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.208 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31321.0 tgs: 262 data_time: 0.93s time: 119.14s eta: 11:36:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31407.0 tgs: 263 data_time: 0.86s time: 119.08s eta: 11:34:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 266 data_time: 0.81s time: 118.95s eta: 11:31:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 268 data_time: 0.78s time: 118.92s eta: 11:29:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 268 data_time: 0.65s time: 118.97s eta: 11:28:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31586.0 tgs: 264 data_time: 0.62s time: 119.29s eta: 11:27:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.125 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31737.0 tgs: 265 data_time: 0.62s time: 119.35s eta: 11:26:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30668.0 tgs: 257 data_time: 0.70s time: 118.88s eta: 11:21:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 0.83s time: 119.21s eta: 11:21:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31176.0 tgs: 260 data_time: 0.48s time: 119.73s eta: 11:22:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32600.0 tgs: 273 data_time: 0.67s time: 119.11s eta: 11:16:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31815.0 tgs: 267 data_time: 0.92s time: 119.12s eta: 11:15:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.198 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 268 data_time: 0.84s time: 118.94s eta: 11:12:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29524.0 tgs: 247 data_time: 0.59s time: 119.10s eta: 11:10:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.262 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31479.0 tgs: 264 data_time: 0.97s time: 119.16s eta: 11:09:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.227 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30973.0 tgs: 260 data_time: 0.78s time: 118.97s eta: 11:06:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31341.0 tgs: 262 data_time: 0.90s time: 119.47s eta: 11:07:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31042.0 tgs: 259 data_time: 0.79s time: 119.65s eta: 11:06:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.234 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31421.0 tgs: 263 data_time: 0.99s time: 119.30s eta: 11:02:07 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 269 data_time: 0.75s time: 119.31s eta: 11:00:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32427.0 tgs: 272 data_time: 0.73s time: 118.82s eta: 10:55:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32089.0 tgs: 268 data_time: 1.00s time: 119.47s eta: 10:57:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.312 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31058.0 tgs: 260 data_time: 1.15s time: 119.37s eta: 10:54:31 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30771.0 tgs: 257 data_time: 0.84s time: 119.28s eta: 10:52:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.219 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31083.0 tgs: 260 data_time: 0.85s time: 119.26s eta: 10:49:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.274 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31601.0 tgs: 265 data_time: 0.99s time: 119.10s eta: 10:47:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.212 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.83s time: 119.53s eta: 10:47:28 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31939.0 tgs: 267 data_time: 0.57s time: 119.30s eta: 10:44:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 267 data_time: 0.77s time: 119.49s eta: 10:43:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.121 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31010.0 tgs: 259 data_time: 1.03s time: 119.51s eta: 10:41:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.218 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 270 data_time: 0.95s time: 118.99s eta: 10:36:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.185 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 268 data_time: 0.76s time: 119.22s eta: 10:35:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32385.0 tgs: 272 data_time: 0.73s time: 119.06s eta: 10:33:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31393.0 tgs: 264 data_time: 0.88s time: 118.79s eta: 10:29:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32286.0 tgs: 269 data_time: 0.86s time: 119.87s eta: 10:33:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30996.0 tgs: 259 data_time: 1.02s time: 119.37s eta: 10:28:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.210 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31997.0 tgs: 269 data_time: 0.81s time: 118.84s eta: 10:23:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31667.0 tgs: 266 data_time: 0.80s time: 118.99s eta: 10:22:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.222 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 267 data_time: 0.98s time: 119.28s eta: 10:22:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31505.0 tgs: 264 data_time: 0.92s time: 119.21s eta: 10:19:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.194 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 266 data_time: 0.99s time: 119.08s eta: 10:17:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31411.0 tgs: 262 data_time: 0.75s time: 119.46s eta: 10:17:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.197 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31521.0 tgs: 264 data_time: 0.97s time: 119.38s eta: 10:14:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 268 data_time: 0.84s time: 119.45s eta: 10:13:10 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32424.0 tgs: 271 data_time: 0.92s time: 119.29s eta: 10:10:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 269 data_time: 0.65s time: 119.11s eta: 10:07:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.122 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32042.0 tgs: 267 data_time: 0.86s time: 119.70s eta: 10:08:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31403.0 tgs: 262 data_time: 0.70s time: 119.83s eta: 10:07:07 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.117 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31488.0 tgs: 259 data_time: 0.75s time: 121.56s eta: 10:13:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31693.0 tgs: 265 data_time: 0.51s time: 119.26s eta: 10:00:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32002.0 tgs: 269 data_time: 0.90s time: 118.80s eta: 9:55:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.225 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31838.0 tgs: 266 data_time: 0.88s time: 119.57s eta: 9:57:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.170 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31373.0 tgs: 262 data_time: 0.73s time: 119.71s eta: 9:56:33 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31726.0 tgs: 266 data_time: 0.90s time: 119.26s eta: 9:52:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31170.0 tgs: 261 data_time: 0.76s time: 119.33s eta: 9:50:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31632.0 tgs: 265 data_time: 0.67s time: 119.22s eta: 9:48:10 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.198 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31866.0 tgs: 263 data_time: 0.66s time: 120.91s eta: 9:54:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.188 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32032.0 tgs: 268 data_time: 0.73s time: 119.23s eta: 9:44:14 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.117 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30902.0 tgs: 258 data_time: 0.91s time: 119.47s eta: 9:43:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.208 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32041.0 tgs: 268 data_time: 1.06s time: 119.54s eta: 9:41:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31125.0 tgs: 261 data_time: 0.66s time: 119.21s eta: 9:38:09 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31950.0 tgs: 267 data_time: 0.82s time: 119.48s eta: 9:37:28 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31647.0 tgs: 265 data_time: 0.84s time: 119.06s eta: 9:33:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32268.0 tgs: 270 data_time: 0.74s time: 119.25s eta: 9:32:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31695.0 tgs: 265 data_time: 1.09s time: 119.16s eta: 9:29:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.099 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31105.0 tgs: 260 data_time: 0.93s time: 119.23s eta: 9:28:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30157.0 tgs: 252 data_time: 0.54s time: 119.47s eta: 9:27:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32496.0 tgs: 273 data_time: 0.74s time: 118.99s eta: 9:23:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32029.0 tgs: 269 data_time: 0.83s time: 119.04s eta: 9:21:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31416.0 tgs: 262 data_time: 0.57s time: 119.53s eta: 9:21:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 267 data_time: 0.60s time: 119.59s eta: 9:20:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29543.0 tgs: 248 data_time: 0.55s time: 119.12s eta: 9:15:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32024.0 tgs: 269 data_time: 0.95s time: 119.00s eta: 9:13:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 270 data_time: 0.63s time: 118.89s eta: 9:10:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.61s time: 119.32s eta: 9:10:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.122 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31365.0 tgs: 263 data_time: 0.72s time: 119.10s eta: 9:07:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31634.0 tgs: 266 data_time: 0.61s time: 118.86s eta: 9:04:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30628.0 tgs: 255 data_time: 0.60s time: 119.67s eta: 9:06:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 265 data_time: 0.58s time: 119.43s eta: 9:03:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31747.0 tgs: 265 data_time: 0.68s time: 119.43s eta: 9:01:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 265 data_time: 0.83s time: 118.74s eta: 8:56:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31565.0 tgs: 265 data_time: 0.70s time: 118.87s eta: 8:54:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31669.0 tgs: 265 data_time: 0.85s time: 119.33s eta: 8:55:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31672.0 tgs: 265 data_time: 0.73s time: 119.15s eta: 8:52:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31776.0 tgs: 266 data_time: 0.73s time: 119.27s eta: 8:50:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31555.0 tgs: 261 data_time: 0.64s time: 120.47s eta: 8:54:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.183 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30891.0 tgs: 259 data_time: 0.73s time: 119.24s eta: 8:46:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31424.0 tgs: 263 data_time: 0.79s time: 119.05s eta: 8:43:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31880.0 tgs: 263 data_time: 0.87s time: 120.83s eta: 8:49:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31044.0 tgs: 258 data_time: 0.87s time: 120.20s eta: 8:44:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32503.0 tgs: 273 data_time: 0.99s time: 118.83s eta: 8:36:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29174.0 tgs: 244 data_time: 0.73s time: 119.35s eta: 8:37:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32215.0 tgs: 269 data_time: 0.54s time: 119.33s eta: 8:35:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31312.0 tgs: 263 data_time: 0.90s time: 119.05s eta: 8:31:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 269 data_time: 0.80s time: 119.17s eta: 8:30:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31510.0 tgs: 264 data_time: 0.63s time: 119.05s eta: 8:27:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31201.0 tgs: 261 data_time: 0.77s time: 119.12s eta: 8:26:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.198 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31904.0 tgs: 267 data_time: 0.76s time: 119.25s eta: 8:24:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 265 data_time: 0.90s time: 119.23s eta: 8:22:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 262 data_time: 0.84s time: 119.38s eta: 8:21:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32429.0 tgs: 270 data_time: 0.91s time: 119.78s eta: 8:21:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.222 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30484.0 tgs: 255 data_time: 0.94s time: 119.20s eta: 8:16:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.199 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31298.0 tgs: 263 data_time: 0.94s time: 118.98s eta: 8:13:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31800.0 tgs: 265 data_time: 0.80s time: 119.90s eta: 8:15:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32184.0 tgs: 269 data_time: 0.91s time: 119.55s eta: 8:12:09 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31480.0 tgs: 263 data_time: 0.63s time: 119.59s eta: 8:10:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31809.0 tgs: 267 data_time: 0.70s time: 119.00s eta: 8:05:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31468.0 tgs: 262 data_time: 0.64s time: 119.84s eta: 8:07:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 267 data_time: 0.85s time: 119.71s eta: 8:04:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31390.0 tgs: 262 data_time: 0.96s time: 119.40s eta: 8:01:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30531.0 tgs: 255 data_time: 0.92s time: 119.29s eta: 7:59:08 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.188 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32229.0 tgs: 270 data_time: 0.88s time: 119.19s eta: 7:56:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.211 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 266 data_time: 1.09s time: 119.20s eta: 7:54:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.186 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31293.0 tgs: 263 data_time: 0.86s time: 118.94s eta: 7:51:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.70s time: 118.97s eta: 7:49:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.113 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32005.0 tgs: 268 data_time: 0.69s time: 119.42s eta: 7:49:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31538.0 tgs: 264 data_time: 0.68s time: 119.04s eta: 7:46:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32119.0 tgs: 269 data_time: 0.74s time: 119.07s eta: 7:44:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32272.0 tgs: 268 data_time: 0.81s time: 120.01s eta: 7:46:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 269 data_time: 0.70s time: 119.83s eta: 7:43:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31519.0 tgs: 264 data_time: 1.06s time: 119.17s eta: 7:38:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30931.0 tgs: 259 data_time: 0.72s time: 118.97s eta: 7:36:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31903.0 tgs: 267 data_time: 0.67s time: 119.33s eta: 7:35:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30881.0 tgs: 258 data_time: 0.73s time: 119.38s eta: 7:33:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30896.0 tgs: 259 data_time: 0.46s time: 119.06s eta: 7:30:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 269 data_time: 0.90s time: 119.00s eta: 7:28:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32496.0 tgs: 272 data_time: 0.66s time: 119.24s eta: 7:27:08 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30772.0 tgs: 258 data_time: 0.71s time: 118.99s eta: 7:24:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31173.0 tgs: 262 data_time: 0.96s time: 118.93s eta: 7:22:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.225 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31728.0 tgs: 266 data_time: 0.87s time: 119.08s eta: 7:20:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.270 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 271 data_time: 0.65s time: 118.99s eta: 7:18:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.87s time: 119.03s eta: 7:16:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31809.0 tgs: 266 data_time: 0.92s time: 119.25s eta: 7:15:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32086.0 tgs: 268 data_time: 0.83s time: 119.33s eta: 7:13:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 265 data_time: 0.63s time: 118.95s eta: 7:10:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.181 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32117.0 tgs: 269 data_time: 0.86s time: 119.20s eta: 7:09:07 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 268 data_time: 0.84s time: 119.59s eta: 7:08:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30473.0 tgs: 254 data_time: 0.59s time: 119.66s eta: 7:06:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32549.0 tgs: 273 data_time: 0.68s time: 118.96s eta: 7:02:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 1.00s time: 119.30s eta: 7:01:31 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.263 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30955.0 tgs: 258 data_time: 0.79s time: 119.52s eta: 7:00:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.121 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31773.0 tgs: 266 data_time: 0.62s time: 119.43s eta: 6:57:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 269 data_time: 0.66s time: 118.91s eta: 6:54:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32349.0 tgs: 271 data_time: 0.65s time: 119.34s eta: 6:53:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31464.0 tgs: 264 data_time: 0.83s time: 118.93s eta: 6:50:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31605.0 tgs: 264 data_time: 0.92s time: 119.27s eta: 6:49:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32291.0 tgs: 271 data_time: 0.73s time: 119.03s eta: 6:46:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30588.0 tgs: 256 data_time: 0.80s time: 119.22s eta: 6:45:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31767.0 tgs: 266 data_time: 0.97s time: 119.39s eta: 6:43:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.177 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 271 data_time: 0.69s time: 119.01s eta: 6:40:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 263 data_time: 0.87s time: 119.01s eta: 6:38:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.59s time: 119.46s eta: 6:38:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32100.0 tgs: 270 data_time: 0.79s time: 118.80s eta: 6:34:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32074.0 tgs: 268 data_time: 0.69s time: 119.44s eta: 6:34:08 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32166.0 tgs: 269 data_time: 0.70s time: 119.57s eta: 6:32:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.116 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31997.0 tgs: 268 data_time: 0.67s time: 119.38s eta: 6:29:58 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31939.0 tgs: 266 data_time: 0.84s time: 119.82s eta: 6:29:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31940.0 tgs: 267 data_time: 0.79s time: 119.36s eta: 6:25:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.219 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 269 data_time: 0.96s time: 118.82s eta: 6:22:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.206 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31495.0 tgs: 264 data_time: 0.61s time: 119.02s eta: 6:20:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32145.0 tgs: 269 data_time: 0.68s time: 119.40s eta: 6:20:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 268 data_time: 0.58s time: 119.34s eta: 6:17:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.098 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31841.0 tgs: 267 data_time: 0.90s time: 119.05s eta: 6:15:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31581.0 tgs: 264 data_time: 0.85s time: 119.38s eta: 6:14:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31943.0 tgs: 267 data_time: 0.75s time: 119.39s eta: 6:12:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 269 data_time: 0.86s time: 119.28s eta: 6:09:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31585.0 tgs: 265 data_time: 0.70s time: 119.10s eta: 6:07:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 261 data_time: 0.92s time: 121.96s eta: 6:14:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31695.0 tgs: 265 data_time: 0.59s time: 119.38s eta: 6:04:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31467.0 tgs: 263 data_time: 0.76s time: 119.57s eta: 6:02:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31655.0 tgs: 265 data_time: 1.10s time: 119.26s eta: 5:59:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 269 data_time: 0.92s time: 119.26s eta: 5:57:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 265 data_time: 0.59s time: 118.92s eta: 5:54:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.91s time: 119.02s eta: 5:53:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.199 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31350.0 tgs: 263 data_time: 0.82s time: 119.07s eta: 5:51:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31686.0 tgs: 265 data_time: 0.65s time: 119.29s eta: 5:49:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 269 data_time: 0.69s time: 118.97s eta: 5:47:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30254.0 tgs: 253 data_time: 0.74s time: 119.27s eta: 5:45:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29732.0 tgs: 249 data_time: 0.85s time: 119.32s eta: 5:44:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 267 data_time: 0.69s time: 119.05s eta: 5:41:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.111 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31846.0 tgs: 266 data_time: 0.46s time: 119.50s eta: 5:40:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 268 data_time: 0.63s time: 119.38s eta: 5:38:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32045.0 tgs: 268 data_time: 0.76s time: 119.56s eta: 5:36:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31846.0 tgs: 267 data_time: 0.89s time: 118.98s eta: 5:33:09 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30769.0 tgs: 258 data_time: 0.74s time: 118.95s eta: 5:31:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30558.0 tgs: 256 data_time: 0.71s time: 119.24s eta: 5:29:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.212 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31252.0 tgs: 261 data_time: 0.76s time: 119.41s eta: 5:28:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32242.0 tgs: 269 data_time: 0.80s time: 119.45s eta: 5:26:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30886.0 tgs: 259 data_time: 0.59s time: 118.95s eta: 5:23:09 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32013.0 tgs: 267 data_time: 0.97s time: 119.63s eta: 5:23:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32249.0 tgs: 270 data_time: 0.79s time: 119.35s eta: 5:20:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31534.0 tgs: 264 data_time: 0.84s time: 119.12s eta: 5:17:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 267 data_time: 0.98s time: 119.42s eta: 5:16:28 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31683.0 tgs: 265 data_time: 0.55s time: 119.55s eta: 5:14:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 265 data_time: 0.94s time: 119.00s eta: 5:11:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.229 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31397.0 tgs: 263 data_time: 0.53s time: 119.00s eta: 5:09:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31859.0 tgs: 267 data_time: 0.68s time: 119.23s eta: 5:08:00 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31364.0 tgs: 262 data_time: 0.59s time: 119.69s eta: 5:07:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.104 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32268.0 tgs: 270 data_time: 0.71s time: 119.11s eta: 5:03:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 270 data_time: 0.53s time: 119.07s eta: 5:01:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.211 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31082.0 tgs: 255 data_time: 0.64s time: 121.52s eta: 5:05:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31408.0 tgs: 262 data_time: 0.74s time: 119.58s eta: 4:58:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.216 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32337.0 tgs: 270 data_time: 0.47s time: 119.50s eta: 4:56:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.231 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31038.0 tgs: 259 data_time: 0.76s time: 119.73s eta: 4:55:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32368.0 tgs: 270 data_time: 0.84s time: 119.52s eta: 4:52:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31956.0 tgs: 267 data_time: 0.57s time: 119.42s eta: 4:50:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.096 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31576.0 tgs: 264 data_time: 0.74s time: 119.20s eta: 4:48:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32397.0 tgs: 272 data_time: 0.89s time: 118.89s eta: 4:45:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32170.0 tgs: 270 data_time: 0.70s time: 119.06s eta: 4:43:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.184 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32365.0 tgs: 269 data_time: 0.83s time: 119.98s eta: 4:43:57 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31796.0 tgs: 266 data_time: 0.84s time: 119.15s eta: 4:39:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.179 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31324.0 tgs: 263 data_time: 0.70s time: 119.01s eta: 4:37:41 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31671.0 tgs: 265 data_time: 0.83s time: 119.06s eta: 4:35:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31825.0 tgs: 266 data_time: 0.61s time: 119.38s eta: 4:34:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32352.0 tgs: 270 data_time: 0.80s time: 119.40s eta: 4:32:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.110 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30940.0 tgs: 260 data_time: 0.79s time: 118.94s eta: 4:29:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32242.0 tgs: 269 data_time: 1.32s time: 119.60s eta: 4:29:05 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31599.0 tgs: 265 data_time: 0.71s time: 119.24s eta: 4:26:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31588.0 tgs: 265 data_time: 0.67s time: 119.07s eta: 4:23:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.213 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31123.0 tgs: 261 data_time: 1.03s time: 118.85s eta: 4:21:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 271 data_time: 0.80s time: 118.76s eta: 4:19:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32032.0 tgs: 268 data_time: 0.77s time: 119.43s eta: 4:18:45 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.181 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32354.0 tgs: 270 data_time: 0.90s time: 119.41s eta: 4:16:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32016.0 tgs: 268 data_time: 0.91s time: 119.32s eta: 4:14:32 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 268 data_time: 0.79s time: 119.05s eta: 4:11:58 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.111 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.80s time: 119.13s eta: 4:10:10 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.204 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 268 data_time: 0.77s time: 119.13s eta: 4:08:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30983.0 tgs: 260 data_time: 0.58s time: 118.96s eta: 4:05:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.209 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32059.0 tgs: 269 data_time: 0.50s time: 119.15s eta: 4:04:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 270 data_time: 0.94s time: 119.49s eta: 4:02:58 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31027.0 tgs: 260 data_time: 0.89s time: 119.25s eta: 4:00:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 270 data_time: 0.71s time: 119.20s eta: 3:58:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31383.0 tgs: 262 data_time: 0.86s time: 119.36s eta: 3:56:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31587.0 tgs: 264 data_time: 0.72s time: 119.38s eta: 3:54:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32047.0 tgs: 267 data_time: 0.86s time: 119.96s eta: 3:53:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32356.0 tgs: 271 data_time: 0.52s time: 119.32s eta: 3:50:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32092.0 tgs: 268 data_time: 0.66s time: 119.63s eta: 3:49:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29588.0 tgs: 248 data_time: 0.78s time: 118.86s eta: 3:45:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31665.0 tgs: 264 data_time: 0.73s time: 119.71s eta: 3:45:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31442.0 tgs: 263 data_time: 0.63s time: 119.33s eta: 3:42:44 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31891.0 tgs: 267 data_time: 0.69s time: 119.27s eta: 3:40:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31225.0 tgs: 262 data_time: 0.85s time: 119.00s eta: 3:38:10 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31289.0 tgs: 261 data_time: 0.77s time: 119.43s eta: 3:36:58 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32268.0 tgs: 271 data_time: 0.89s time: 119.03s eta: 3:34:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31078.0 tgs: 261 data_time: 0.70s time: 118.99s eta: 3:32:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 266 data_time: 0.84s time: 119.22s eta: 3:30:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32406.0 tgs: 269 data_time: 1.06s time: 120.12s eta: 3:30:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 259 data_time: 0.79s time: 119.14s eta: 3:26:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31928.0 tgs: 267 data_time: 0.54s time: 119.31s eta: 3:24:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32100.0 tgs: 269 data_time: 0.75s time: 119.09s eta: 3:22:27 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.223 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 0.88s time: 119.42s eta: 3:21:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31117.0 tgs: 260 data_time: 0.88s time: 119.29s eta: 3:18:48 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32141.0 tgs: 269 data_time: 1.18s time: 119.22s eta: 3:16:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32069.0 tgs: 269 data_time: 0.83s time: 119.12s eta: 3:14:33 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32154.0 tgs: 270 data_time: 0.77s time: 119.02s eta: 3:12:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31283.0 tgs: 262 data_time: 0.53s time: 119.19s eta: 3:10:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32362.0 tgs: 267 data_time: 0.90s time: 120.89s eta: 3:11:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32090.0 tgs: 268 data_time: 0.48s time: 119.50s eta: 3:07:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31709.0 tgs: 265 data_time: 0.74s time: 119.26s eta: 3:04:51 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 269 data_time: 0.81s time: 118.99s eta: 3:02:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31728.0 tgs: 266 data_time: 0.99s time: 119.22s eta: 3:00:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30756.0 tgs: 257 data_time: 0.73s time: 119.46s eta: 2:59:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 271 data_time: 1.05s time: 118.98s eta: 2:56:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31144.0 tgs: 260 data_time: 0.59s time: 119.35s eta: 2:55:03 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31764.0 tgs: 266 data_time: 1.02s time: 119.41s eta: 2:53:08 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 267 data_time: 0.67s time: 119.35s eta: 2:51:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 259 data_time: 0.67s time: 121.46s eta: 2:52:04 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31432.0 tgs: 263 data_time: 0.87s time: 119.24s eta: 2:46:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31430.0 tgs: 262 data_time: 0.77s time: 119.56s eta: 2:45:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31515.0 tgs: 265 data_time: 0.79s time: 118.91s eta: 2:42:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31398.0 tgs: 263 data_time: 0.65s time: 119.18s eta: 2:40:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30824.0 tgs: 259 data_time: 0.74s time: 118.97s eta: 2:38:37 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32254.0 tgs: 270 data_time: 0.99s time: 119.24s eta: 2:36:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.186 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30158.0 tgs: 253 data_time: 0.82s time: 119.11s eta: 2:34:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30668.0 tgs: 256 data_time: 0.60s time: 119.43s eta: 2:33:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31705.0 tgs: 263 data_time: 0.89s time: 120.22s eta: 2:32:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31871.0 tgs: 267 data_time: 0.85s time: 119.28s eta: 2:29:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31978.0 tgs: 268 data_time: 0.76s time: 118.88s eta: 2:26:36 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.196 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31953.0 tgs: 267 data_time: 0.90s time: 119.52s eta: 2:25:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32051.0 tgs: 269 data_time: 0.94s time: 119.07s eta: 2:22:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31727.0 tgs: 266 data_time: 0.77s time: 119.14s eta: 2:20:58 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31702.0 tgs: 266 data_time: 0.91s time: 119.05s eta: 2:18:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 268 data_time: 0.89s time: 119.84s eta: 2:17:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31770.0 tgs: 266 data_time: 0.49s time: 119.31s eta: 2:15:13 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31812.0 tgs: 266 data_time: 0.90s time: 119.36s eta: 2:13:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31915.0 tgs: 266 data_time: 0.81s time: 119.69s eta: 2:11:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.217 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31860.0 tgs: 266 data_time: 0.69s time: 119.49s eta: 2:09:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 269 data_time: 0.70s time: 118.96s eta: 2:06:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31729.0 tgs: 265 data_time: 0.63s time: 119.37s eta: 2:05:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30980.0 tgs: 259 data_time: 0.54s time: 119.28s eta: 2:03:15 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 266 data_time: 0.77s time: 119.22s eta: 2:01:12 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31385.0 tgs: 262 data_time: 0.75s time: 119.51s eta: 1:59:30 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31827.0 tgs: 266 data_time: 0.46s time: 119.42s eta: 1:57:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.89s time: 119.54s eta: 1:55:33 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30740.0 tgs: 257 data_time: 0.99s time: 119.31s eta: 1:53:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31178.0 tgs: 261 data_time: 0.79s time: 119.25s eta: 1:51:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 271 data_time: 0.79s time: 118.93s eta: 1:49:01 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31180.0 tgs: 260 data_time: 0.71s time: 119.64s eta: 1:47:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 265 data_time: 0.78s time: 119.72s eta: 1:45:44 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31377.0 tgs: 262 data_time: 0.47s time: 119.36s eta: 1:43:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32027.0 tgs: 268 data_time: 0.57s time: 119.30s eta: 1:41:24 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31859.0 tgs: 267 data_time: 0.90s time: 119.28s eta: 1:39:23 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 269 data_time: 0.78s time: 118.91s eta: 1:37:06 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32443.0 tgs: 272 data_time: 0.85s time: 119.16s eta: 1:35:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32375.0 tgs: 271 data_time: 0.57s time: 119.07s eta: 1:33:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31848.0 tgs: 267 data_time: 0.68s time: 119.26s eta: 1:31:25 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 269 data_time: 0.66s time: 119.73s eta: 1:29:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31778.0 tgs: 266 data_time: 0.81s time: 119.02s eta: 1:27:16 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31269.0 tgs: 263 data_time: 0.70s time: 118.87s eta: 1:25:11 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 268 data_time: 0.59s time: 119.48s eta: 1:23:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 269 data_time: 0.72s time: 119.48s eta: 1:21:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.193 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32248.0 tgs: 271 data_time: 0.84s time: 118.94s eta: 1:19:17 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31102.0 tgs: 261 data_time: 0.67s time: 118.94s eta: 1:17:18 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32215.0 tgs: 263 data_time: 0.71s time: 122.13s eta: 1:17:20 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31693.0 tgs: 265 data_time: 0.87s time: 119.33s eta: 1:13:35 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30460.0 tgs: 255 data_time: 0.81s time: 119.06s eta: 1:11:26 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31907.0 tgs: 268 data_time: 0.69s time: 118.92s eta: 1:09:22 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31334.0 tgs: 263 data_time: 0.80s time: 119.11s eta: 1:07:29 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.87s time: 119.42s eta: 1:05:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 270 data_time: 0.97s time: 119.35s eta: 1:03:39 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31141.0 tgs: 261 data_time: 0.71s time: 119.18s eta: 1:01:34 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32132.0 tgs: 269 data_time: 0.89s time: 119.28s eta: 0:59:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 268 data_time: 1.21s time: 119.44s eta: 0:57:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31761.0 tgs: 266 data_time: 1.06s time: 119.39s eta: 0:55:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.65s time: 119.49s eta: 0:53:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31781.0 tgs: 266 data_time: 0.67s time: 119.37s eta: 0:51:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31804.0 tgs: 266 data_time: 0.89s time: 119.47s eta: 0:49:46 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 267 data_time: 0.71s time: 118.90s eta: 0:47:33 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31769.0 tgs: 266 data_time: 0.79s time: 119.27s eta: 0:45:43 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.233 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.98s time: 119.20s eta: 0:43:42 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 265 data_time: 0.89s time: 119.00s eta: 0:41:38 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31905.0 tgs: 268 data_time: 0.95s time: 119.04s eta: 0:39:40 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 268 data_time: 0.72s time: 119.66s eta: 0:37:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31864.0 tgs: 267 data_time: 0.55s time: 119.28s eta: 0:35:47 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32094.0 tgs: 268 data_time: 0.70s time: 119.75s eta: 0:33:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32460.0 tgs: 267 data_time: 0.90s time: 121.23s eta: 0:32:19 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 267 data_time: 0.72s time: 119.38s eta: 0:29:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32063.0 tgs: 268 data_time: 0.60s time: 119.33s eta: 0:27:50 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31458.0 tgs: 262 data_time: 0.72s time: 119.64s eta: 0:25:55 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31717.0 tgs: 266 data_time: 0.73s time: 119.09s eta: 0:23:49 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31218.0 tgs: 261 data_time: 0.85s time: 119.39s eta: 0:21:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32262.0 tgs: 270 data_time: 0.85s time: 119.26s eta: 0:19:52 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31793.0 tgs: 266 data_time: 0.61s time: 119.24s eta: 0:17:53 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 262 data_time: 0.78s time: 119.90s eta: 0:15:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.196 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31603.0 tgs: 265 data_time: 0.88s time: 119.18s eta: 0:13:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32105.0 tgs: 269 data_time: 0.92s time: 119.09s eta: 0:11:54 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 266 data_time: 0.85s time: 119.32s eta: 0:09:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31820.0 tgs: 266 data_time: 0.59s time: 119.18s eta: 0:07:56 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31856.0 tgs: 266 data_time: 1.14s time: 119.69s eta: 0:05:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32058.0 tgs: 264 data_time: 0.74s time: 121.37s eta: 0:04:02 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.205 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31388.0 tgs: 263 data_time: 0.79s time: 119.21s eta: 0:01:59 +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.28 seconds, peak gpu memory 17.4G +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 21][DP 5][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank22.log b/20250121165312/rank22.log new file mode 100644 index 0000000000000000000000000000000000000000..2eaae2f2bb0fa85f779c024963ead47d2a505763 --- /dev/null +++ b/20250121165312/rank22.log @@ -0,0 +1,851 @@ +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.07s +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.42 seconds, peak gpu memory 13.4G +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.256 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.3GB text_tokens: 29620.0 tgs: 195 data_time: 1.90s time: 151.25s eta: 1 day, 0:54:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.266 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32369.0 tgs: 269 data_time: 0.79s time: 120.14s eta: 19:45:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.213 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32546.0 tgs: 271 data_time: 0.78s time: 119.68s eta: 19:38:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.210 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 270 data_time: 1.04s time: 119.66s eta: 19:36:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.225 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30927.0 tgs: 259 data_time: 0.90s time: 119.30s eta: 19:31:10 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.236 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32427.0 tgs: 271 data_time: 1.18s time: 119.54s eta: 19:31:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.182 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31527.0 tgs: 256 data_time: 1.12s time: 122.96s eta: 20:02:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.300 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31799.0 tgs: 265 data_time: 0.67s time: 119.89s eta: 19:30:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.210 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30776.0 tgs: 255 data_time: 0.72s time: 120.66s eta: 19:36:23 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.226 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31699.0 tgs: 264 data_time: 0.71s time: 119.92s eta: 19:27:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.242 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 261 data_time: 0.94s time: 120.46s eta: 19:30:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.185 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31345.0 tgs: 262 data_time: 0.67s time: 119.34s eta: 19:17:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30795.0 tgs: 257 data_time: 0.78s time: 119.68s eta: 19:18:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 262 data_time: 0.73s time: 119.17s eta: 19:12:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32265.0 tgs: 270 data_time: 0.84s time: 119.23s eta: 19:10:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29937.0 tgs: 251 data_time: 0.80s time: 118.92s eta: 19:05:34 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30786.0 tgs: 258 data_time: 0.74s time: 119.15s eta: 19:05:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31670.0 tgs: 265 data_time: 0.85s time: 119.18s eta: 19:04:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31374.0 tgs: 262 data_time: 0.60s time: 119.40s eta: 19:04:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32080.0 tgs: 268 data_time: 0.74s time: 119.33s eta: 19:01:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30100.0 tgs: 250 data_time: 0.77s time: 120.31s eta: 19:08:58 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31525.0 tgs: 256 data_time: 0.89s time: 122.91s eta: 19:31:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32200.0 tgs: 269 data_time: 0.92s time: 119.49s eta: 18:57:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31533.0 tgs: 264 data_time: 1.01s time: 119.15s eta: 18:51:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.213 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31572.0 tgs: 264 data_time: 0.76s time: 119.25s eta: 18:50:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31711.0 tgs: 265 data_time: 0.57s time: 119.39s eta: 18:50:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31027.0 tgs: 259 data_time: 0.80s time: 119.66s eta: 18:50:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.248 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31174.0 tgs: 261 data_time: 1.02s time: 119.39s eta: 18:46:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 265 data_time: 0.88s time: 119.76s eta: 18:47:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 246 data_time: 0.98s time: 129.30s eta: 20:15:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31822.0 tgs: 266 data_time: 0.87s time: 119.57s eta: 18:41:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31024.0 tgs: 259 data_time: 0.76s time: 119.39s eta: 18:38:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31792.0 tgs: 263 data_time: 1.00s time: 120.66s eta: 18:48:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31319.0 tgs: 262 data_time: 0.62s time: 119.39s eta: 18:34:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 270 data_time: 0.62s time: 119.16s eta: 18:30:10 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32359.0 tgs: 271 data_time: 0.86s time: 119.07s eta: 18:27:22 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31433.0 tgs: 263 data_time: 0.90s time: 119.14s eta: 18:25:58 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30457.0 tgs: 252 data_time: 0.87s time: 120.39s eta: 18:35:36 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 264 data_time: 0.86s time: 121.18s eta: 18:40:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 256 data_time: 0.65s time: 120.66s eta: 18:34:08 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32067.0 tgs: 267 data_time: 0.84s time: 119.73s eta: 18:23:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31896.0 tgs: 257 data_time: 0.93s time: 123.88s eta: 18:59:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30763.0 tgs: 257 data_time: 0.74s time: 119.56s eta: 18:17:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31487.0 tgs: 263 data_time: 0.88s time: 119.32s eta: 18:13:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31702.0 tgs: 265 data_time: 0.71s time: 119.23s eta: 18:10:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31920.0 tgs: 266 data_time: 0.91s time: 119.60s eta: 18:12:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 266 data_time: 0.75s time: 119.28s eta: 18:07:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31065.0 tgs: 261 data_time: 0.74s time: 119.00s eta: 18:02:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32174.0 tgs: 270 data_time: 0.78s time: 119.10s eta: 18:01:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31306.0 tgs: 262 data_time: 1.02s time: 119.27s eta: 18:01:22 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29824.0 tgs: 249 data_time: 0.68s time: 119.33s eta: 17:59:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.285 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31180.0 tgs: 259 data_time: 1.07s time: 120.02s eta: 18:04:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31929.0 tgs: 268 data_time: 0.87s time: 119.12s eta: 17:54:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32302.0 tgs: 270 data_time: 0.55s time: 119.40s eta: 17:54:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31344.0 tgs: 263 data_time: 0.68s time: 119.00s eta: 17:48:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30272.0 tgs: 254 data_time: 0.77s time: 119.03s eta: 17:47:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32061.0 tgs: 255 data_time: 0.70s time: 125.37s eta: 18:42:05 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 265 data_time: 0.67s time: 120.10s eta: 17:52:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32402.0 tgs: 271 data_time: 1.21s time: 119.37s eta: 17:44:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30488.0 tgs: 255 data_time: 0.86s time: 119.33s eta: 17:42:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 266 data_time: 0.57s time: 119.51s eta: 17:41:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31778.0 tgs: 252 data_time: 0.57s time: 126.05s eta: 18:37:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 259 data_time: 0.94s time: 122.63s eta: 18:05:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 264 data_time: 1.36s time: 121.16s eta: 17:50:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30998.0 tgs: 260 data_time: 0.67s time: 119.04s eta: 17:29:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31047.0 tgs: 261 data_time: 0.90s time: 118.85s eta: 17:25:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32302.0 tgs: 260 data_time: 0.85s time: 124.00s eta: 18:09:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30866.0 tgs: 247 data_time: 0.62s time: 124.56s eta: 18:12:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31849.0 tgs: 266 data_time: 0.73s time: 119.40s eta: 17:24:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31641.0 tgs: 264 data_time: 0.81s time: 119.70s eta: 17:25:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.121 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32468.0 tgs: 271 data_time: 0.74s time: 119.53s eta: 17:21:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31006.0 tgs: 260 data_time: 0.70s time: 118.86s eta: 17:14:05 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31253.0 tgs: 260 data_time: 0.75s time: 119.85s eta: 17:20:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31654.0 tgs: 264 data_time: 0.95s time: 119.51s eta: 17:15:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31945.0 tgs: 262 data_time: 0.76s time: 121.86s eta: 17:34:04 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 253 data_time: 0.56s time: 123.51s eta: 17:46:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30448.0 tgs: 255 data_time: 0.62s time: 119.01s eta: 17:05:30 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31752.0 tgs: 266 data_time: 0.87s time: 119.23s eta: 17:05:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 270 data_time: 0.68s time: 119.14s eta: 17:02:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32172.0 tgs: 269 data_time: 0.98s time: 119.25s eta: 17:01:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31208.0 tgs: 262 data_time: 0.95s time: 119.05s eta: 16:57:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32488.0 tgs: 272 data_time: 0.80s time: 119.28s eta: 16:57:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31692.0 tgs: 265 data_time: 0.87s time: 119.16s eta: 16:54:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 266 data_time: 0.85s time: 119.58s eta: 16:56:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32503.0 tgs: 272 data_time: 0.97s time: 119.13s eta: 16:50:36 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31637.0 tgs: 266 data_time: 0.69s time: 118.84s eta: 16:46:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32396.0 tgs: 272 data_time: 0.75s time: 119.02s eta: 16:45:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31701.0 tgs: 264 data_time: 0.64s time: 119.85s eta: 16:50:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32363.0 tgs: 271 data_time: 0.93s time: 119.36s eta: 16:44:36 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31309.0 tgs: 263 data_time: 0.54s time: 119.01s eta: 16:39:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 262 data_time: 1.02s time: 120.93s eta: 16:53:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31214.0 tgs: 262 data_time: 0.66s time: 118.99s eta: 16:35:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31653.0 tgs: 265 data_time: 0.81s time: 119.29s eta: 16:36:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.256 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32264.0 tgs: 270 data_time: 0.92s time: 119.23s eta: 16:33:34 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32112.0 tgs: 269 data_time: 0.96s time: 119.37s eta: 16:32:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31448.0 tgs: 262 data_time: 0.78s time: 119.70s eta: 16:33:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31670.0 tgs: 265 data_time: 0.75s time: 119.21s eta: 16:27:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32024.0 tgs: 267 data_time: 0.60s time: 119.76s eta: 16:29:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31293.0 tgs: 260 data_time: 0.74s time: 120.28s eta: 16:32:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31374.0 tgs: 263 data_time: 0.72s time: 118.97s eta: 16:19:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30907.0 tgs: 242 data_time: 0.82s time: 127.33s eta: 17:26:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 268 data_time: 0.62s time: 119.23s eta: 16:17:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31295.0 tgs: 262 data_time: 0.62s time: 119.14s eta: 16:14:58 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31271.0 tgs: 262 data_time: 0.74s time: 119.21s eta: 16:13:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29928.0 tgs: 251 data_time: 0.56s time: 119.18s eta: 16:11:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31218.0 tgs: 262 data_time: 0.77s time: 118.95s eta: 16:07:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 267 data_time: 0.77s time: 118.86s eta: 16:04:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.121 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32258.0 tgs: 271 data_time: 0.70s time: 118.96s eta: 16:03:36 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 259 data_time: 0.78s time: 121.84s eta: 16:24:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.123 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 268 data_time: 1.00s time: 119.00s eta: 15:59:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30581.0 tgs: 256 data_time: 0.89s time: 119.22s eta: 15:59:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32435.0 tgs: 271 data_time: 0.53s time: 119.62s eta: 16:00:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.118 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31232.0 tgs: 262 data_time: 0.94s time: 119.04s eta: 15:54:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.207 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31684.0 tgs: 266 data_time: 1.07s time: 118.98s eta: 15:51:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 267 data_time: 0.83s time: 119.17s eta: 15:51:23 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.273 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32043.0 tgs: 267 data_time: 0.61s time: 119.64s eta: 15:53:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31617.0 tgs: 265 data_time: 0.91s time: 118.94s eta: 15:45:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31764.0 tgs: 266 data_time: 0.56s time: 119.29s eta: 15:46:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32084.0 tgs: 269 data_time: 1.00s time: 119.11s eta: 15:42:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30935.0 tgs: 259 data_time: 0.91s time: 119.00s eta: 15:40:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32150.0 tgs: 268 data_time: 0.81s time: 119.66s eta: 15:43:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 262 data_time: 0.81s time: 121.51s eta: 15:55:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31485.0 tgs: 263 data_time: 0.56s time: 119.41s eta: 15:37:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32015.0 tgs: 246 data_time: 0.82s time: 129.97s eta: 16:58:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31263.0 tgs: 261 data_time: 0.91s time: 119.47s eta: 15:33:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30780.0 tgs: 257 data_time: 0.70s time: 119.61s eta: 15:32:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32045.0 tgs: 268 data_time: 0.79s time: 119.53s eta: 15:30:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 259 data_time: 0.74s time: 122.14s eta: 15:48:36 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.254 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30999.0 tgs: 256 data_time: 0.87s time: 120.76s eta: 15:35:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31669.0 tgs: 265 data_time: 1.00s time: 119.42s eta: 15:23:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31113.0 tgs: 260 data_time: 0.70s time: 119.45s eta: 15:21:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 238 data_time: 0.86s time: 131.63s eta: 16:53:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 265 data_time: 0.89s time: 119.29s eta: 15:16:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 264 data_time: 0.55s time: 119.52s eta: 15:16:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32096.0 tgs: 262 data_time: 0.73s time: 122.25s eta: 15:35:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31552.0 tgs: 263 data_time: 0.68s time: 119.71s eta: 15:13:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.77s time: 119.35s eta: 15:09:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30937.0 tgs: 260 data_time: 0.89s time: 118.90s eta: 15:03:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.235 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32455.0 tgs: 272 data_time: 1.00s time: 119.00s eta: 15:02:23 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31945.0 tgs: 267 data_time: 0.78s time: 119.31s eta: 15:02:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 265 data_time: 0.68s time: 118.92s eta: 14:57:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30856.0 tgs: 256 data_time: 0.65s time: 120.52s eta: 15:07:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31360.0 tgs: 262 data_time: 0.82s time: 119.31s eta: 14:56:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32316.0 tgs: 263 data_time: 0.70s time: 122.66s eta: 15:19:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.68s time: 119.13s eta: 14:51:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30701.0 tgs: 256 data_time: 0.68s time: 119.76s eta: 14:54:10 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 265 data_time: 0.73s time: 119.47s eta: 14:50:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.224 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 270 data_time: 0.77s time: 118.99s eta: 14:44:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.59s time: 119.30s eta: 14:44:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32218.0 tgs: 270 data_time: 1.00s time: 118.93s eta: 14:40:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.211 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31462.0 tgs: 264 data_time: 0.74s time: 119.06s eta: 14:39:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.109 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30925.0 tgs: 259 data_time: 0.84s time: 119.06s eta: 14:37:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32210.0 tgs: 270 data_time: 0.70s time: 119.13s eta: 14:35:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.59s time: 119.00s eta: 14:32:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.206 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31753.0 tgs: 266 data_time: 0.92s time: 119.08s eta: 14:31:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32043.0 tgs: 268 data_time: 0.85s time: 119.15s eta: 14:29:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 269 data_time: 1.02s time: 119.65s eta: 14:31:28 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.282 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30864.0 tgs: 258 data_time: 1.09s time: 119.18s eta: 14:26:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.205 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.77s time: 119.20s eta: 14:24:10 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31553.0 tgs: 264 data_time: 0.64s time: 119.39s eta: 14:23:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30653.0 tgs: 253 data_time: 0.74s time: 120.74s eta: 14:31:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31467.0 tgs: 264 data_time: 0.80s time: 118.86s eta: 14:15:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 269 data_time: 0.73s time: 119.10s eta: 14:15:31 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31113.0 tgs: 261 data_time: 0.87s time: 119.12s eta: 14:13:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31202.0 tgs: 260 data_time: 0.48s time: 119.96s eta: 14:17:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 268 data_time: 0.97s time: 119.87s eta: 14:15:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.230 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31916.0 tgs: 266 data_time: 0.72s time: 119.58s eta: 14:11:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31679.0 tgs: 265 data_time: 0.75s time: 119.22s eta: 14:06:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.73s time: 119.27s eta: 14:04:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31455.0 tgs: 263 data_time: 0.78s time: 119.27s eta: 14:02:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.211 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31473.0 tgs: 264 data_time: 0.97s time: 119.14s eta: 13:59:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.113 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31878.0 tgs: 267 data_time: 0.81s time: 119.13s eta: 13:57:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31593.0 tgs: 265 data_time: 0.88s time: 119.16s eta: 13:56:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.199 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31814.0 tgs: 266 data_time: 0.84s time: 119.36s eta: 13:55:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31236.0 tgs: 260 data_time: 0.68s time: 120.11s eta: 13:58:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31467.0 tgs: 256 data_time: 0.99s time: 122.56s eta: 14:13:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32419.0 tgs: 272 data_time: 0.52s time: 119.10s eta: 13:47:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.213 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31622.0 tgs: 264 data_time: 1.00s time: 119.37s eta: 13:47:37 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 270 data_time: 0.78s time: 119.32s eta: 13:45:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31542.0 tgs: 264 data_time: 0.64s time: 119.33s eta: 13:43:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31234.0 tgs: 262 data_time: 0.71s time: 119.12s eta: 13:39:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 268 data_time: 0.72s time: 119.25s eta: 13:38:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.95s time: 119.13s eta: 13:36:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 1.09s time: 119.23s eta: 13:34:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32114.0 tgs: 268 data_time: 0.73s time: 119.45s eta: 13:34:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30360.0 tgs: 254 data_time: 0.59s time: 119.08s eta: 13:29:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31389.0 tgs: 263 data_time: 0.96s time: 118.93s eta: 13:26:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31445.0 tgs: 262 data_time: 0.92s time: 119.81s eta: 13:30:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32275.0 tgs: 270 data_time: 0.67s time: 119.26s eta: 13:25:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32417.0 tgs: 272 data_time: 0.79s time: 119.16s eta: 13:22:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.191 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.80s time: 119.35s eta: 13:21:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31655.0 tgs: 265 data_time: 0.66s time: 119.26s eta: 13:19:04 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32283.0 tgs: 271 data_time: 0.92s time: 119.04s eta: 13:15:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32317.0 tgs: 270 data_time: 0.72s time: 119.35s eta: 13:15:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.76s time: 119.17s eta: 13:12:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.218 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32616.0 tgs: 273 data_time: 0.84s time: 119.46s eta: 13:12:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30892.0 tgs: 258 data_time: 0.74s time: 119.29s eta: 13:09:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31394.0 tgs: 263 data_time: 0.87s time: 119.15s eta: 13:06:23 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.194 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 266 data_time: 0.79s time: 120.39s eta: 13:12:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32445.0 tgs: 270 data_time: 0.63s time: 120.16s eta: 13:09:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31252.0 tgs: 261 data_time: 0.68s time: 119.50s eta: 13:02:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 266 data_time: 0.83s time: 119.51s eta: 13:00:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 270 data_time: 0.77s time: 119.57s eta: 12:59:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.253 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31257.0 tgs: 260 data_time: 0.78s time: 119.80s eta: 12:58:40 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30989.0 tgs: 260 data_time: 0.78s time: 119.04s eta: 12:51:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31367.0 tgs: 263 data_time: 0.80s time: 119.05s eta: 12:49:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 266 data_time: 0.84s time: 119.13s eta: 12:48:22 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30623.0 tgs: 256 data_time: 0.70s time: 119.40s eta: 12:48:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 267 data_time: 0.71s time: 119.18s eta: 12:44:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 0.64s time: 119.42s eta: 12:44:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.179 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.75s time: 119.10s eta: 12:40:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31061.0 tgs: 260 data_time: 0.68s time: 119.07s eta: 12:38:05 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31784.0 tgs: 265 data_time: 0.92s time: 119.60s eta: 12:39:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 270 data_time: 0.63s time: 118.78s eta: 12:32:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 270 data_time: 0.91s time: 118.89s eta: 12:31:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30989.0 tgs: 260 data_time: 0.83s time: 119.16s eta: 12:30:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.70s time: 119.27s eta: 12:29:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31282.0 tgs: 262 data_time: 0.73s time: 119.35s eta: 12:27:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32386.0 tgs: 271 data_time: 0.75s time: 119.12s eta: 12:24:28 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 270 data_time: 0.83s time: 119.35s eta: 12:23:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.124 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.77s time: 119.26s eta: 12:21:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 268 data_time: 0.81s time: 119.03s eta: 12:17:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.210 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32053.0 tgs: 264 data_time: 0.85s time: 121.13s eta: 12:28:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31183.0 tgs: 259 data_time: 0.75s time: 119.99s eta: 12:19:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30267.0 tgs: 255 data_time: 0.89s time: 118.69s eta: 12:09:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31274.0 tgs: 261 data_time: 0.70s time: 119.64s eta: 12:13:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32119.0 tgs: 268 data_time: 0.77s time: 119.84s eta: 12:13:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32165.0 tgs: 269 data_time: 0.74s time: 119.14s eta: 12:06:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31657.0 tgs: 265 data_time: 0.72s time: 119.17s eta: 12:04:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31215.0 tgs: 262 data_time: 0.89s time: 119.12s eta: 12:02:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.228 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32433.0 tgs: 272 data_time: 0.93s time: 119.19s eta: 12:01:04 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31078.0 tgs: 261 data_time: 0.86s time: 118.96s eta: 11:57:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.224 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31699.0 tgs: 264 data_time: 1.02s time: 119.71s eta: 12:00:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31777.0 tgs: 266 data_time: 0.75s time: 119.28s eta: 11:55:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.304 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32039.0 tgs: 268 data_time: 0.90s time: 119.27s eta: 11:53:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.124 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32039.0 tgs: 267 data_time: 0.62s time: 119.71s eta: 11:54:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.263 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32191.0 tgs: 270 data_time: 0.71s time: 118.80s eta: 11:46:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.249 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32174.0 tgs: 270 data_time: 0.69s time: 119.15s eta: 11:46:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31477.0 tgs: 263 data_time: 0.71s time: 119.61s eta: 11:47:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.133 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31088.0 tgs: 261 data_time: 0.87s time: 118.83s eta: 11:41:04 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 268 data_time: 0.74s time: 118.68s eta: 11:38:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.204 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31933.0 tgs: 267 data_time: 1.20s time: 119.22s eta: 11:39:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31321.0 tgs: 262 data_time: 0.93s time: 119.14s eta: 11:36:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31407.0 tgs: 263 data_time: 0.85s time: 119.08s eta: 11:34:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 266 data_time: 0.79s time: 118.95s eta: 11:31:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 268 data_time: 0.76s time: 118.91s eta: 11:29:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.113 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 268 data_time: 0.65s time: 118.97s eta: 11:28:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31586.0 tgs: 264 data_time: 0.62s time: 119.29s eta: 11:27:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31737.0 tgs: 265 data_time: 0.58s time: 119.36s eta: 11:26:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30668.0 tgs: 257 data_time: 0.68s time: 118.89s eta: 11:21:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 0.81s time: 119.20s eta: 11:21:26 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31176.0 tgs: 260 data_time: 0.47s time: 119.72s eta: 11:22:22 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.208 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32600.0 tgs: 273 data_time: 0.71s time: 119.12s eta: 11:16:58 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31815.0 tgs: 267 data_time: 0.92s time: 119.12s eta: 11:15:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 268 data_time: 0.81s time: 118.94s eta: 11:12:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29524.0 tgs: 247 data_time: 0.59s time: 119.09s eta: 11:10:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31479.0 tgs: 264 data_time: 0.94s time: 119.16s eta: 11:09:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30973.0 tgs: 260 data_time: 0.78s time: 118.97s eta: 11:06:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.236 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31341.0 tgs: 262 data_time: 0.86s time: 119.47s eta: 11:07:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.106 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31042.0 tgs: 259 data_time: 0.77s time: 119.64s eta: 11:06:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31421.0 tgs: 263 data_time: 0.97s time: 119.31s eta: 11:02:09 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 269 data_time: 0.73s time: 119.31s eta: 11:00:10 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32427.0 tgs: 272 data_time: 0.72s time: 118.82s eta: 10:55:30 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.178 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32089.0 tgs: 268 data_time: 0.96s time: 119.48s eta: 10:57:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.112 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31058.0 tgs: 260 data_time: 1.14s time: 119.37s eta: 10:54:31 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.122 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30771.0 tgs: 257 data_time: 0.84s time: 119.28s eta: 10:52:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31083.0 tgs: 260 data_time: 0.84s time: 119.26s eta: 10:49:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31601.0 tgs: 265 data_time: 0.97s time: 119.10s eta: 10:47:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.80s time: 119.53s eta: 10:47:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31939.0 tgs: 267 data_time: 0.53s time: 119.31s eta: 10:44:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.221 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 267 data_time: 0.74s time: 119.49s eta: 10:43:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31010.0 tgs: 259 data_time: 0.99s time: 119.51s eta: 10:41:23 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.110 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 270 data_time: 0.92s time: 118.99s eta: 10:36:37 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 268 data_time: 0.74s time: 119.22s eta: 10:35:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32385.0 tgs: 272 data_time: 0.71s time: 119.06s eta: 10:33:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.176 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31393.0 tgs: 264 data_time: 0.87s time: 118.78s eta: 10:29:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32286.0 tgs: 269 data_time: 0.84s time: 119.86s eta: 10:33:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30996.0 tgs: 259 data_time: 0.99s time: 119.38s eta: 10:28:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31997.0 tgs: 269 data_time: 0.79s time: 118.84s eta: 10:23:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.118 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31667.0 tgs: 266 data_time: 0.77s time: 118.99s eta: 10:22:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 267 data_time: 0.95s time: 119.28s eta: 10:22:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.220 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31505.0 tgs: 264 data_time: 0.89s time: 119.21s eta: 10:19:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.191 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 266 data_time: 0.95s time: 119.06s eta: 10:17:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31411.0 tgs: 262 data_time: 0.74s time: 119.48s eta: 10:17:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31521.0 tgs: 264 data_time: 0.95s time: 119.39s eta: 10:14:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 268 data_time: 0.84s time: 119.45s eta: 10:13:09 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32424.0 tgs: 271 data_time: 0.89s time: 119.28s eta: 10:10:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 269 data_time: 0.64s time: 119.11s eta: 10:07:28 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32042.0 tgs: 267 data_time: 0.84s time: 119.71s eta: 10:08:30 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31403.0 tgs: 262 data_time: 0.68s time: 119.83s eta: 10:07:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.228 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31488.0 tgs: 259 data_time: 0.73s time: 121.56s eta: 10:13:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31693.0 tgs: 265 data_time: 0.50s time: 119.26s eta: 10:00:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32002.0 tgs: 269 data_time: 0.91s time: 118.80s eta: 9:55:58 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.224 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31838.0 tgs: 266 data_time: 0.86s time: 119.57s eta: 9:57:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.123 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31373.0 tgs: 262 data_time: 0.71s time: 119.70s eta: 9:56:31 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.175 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31726.0 tgs: 266 data_time: 0.87s time: 119.26s eta: 9:52:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31170.0 tgs: 261 data_time: 0.74s time: 119.35s eta: 9:50:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31632.0 tgs: 265 data_time: 0.65s time: 119.21s eta: 9:48:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31866.0 tgs: 263 data_time: 0.63s time: 120.78s eta: 9:53:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32032.0 tgs: 268 data_time: 0.71s time: 119.23s eta: 9:44:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30902.0 tgs: 258 data_time: 0.89s time: 119.46s eta: 9:43:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32041.0 tgs: 268 data_time: 1.04s time: 119.55s eta: 9:41:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.183 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31125.0 tgs: 261 data_time: 0.64s time: 119.20s eta: 9:38:08 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31950.0 tgs: 267 data_time: 0.81s time: 119.48s eta: 9:37:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.205 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31647.0 tgs: 265 data_time: 0.82s time: 119.06s eta: 9:33:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32268.0 tgs: 270 data_time: 0.71s time: 119.24s eta: 9:32:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.249 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31695.0 tgs: 265 data_time: 1.09s time: 119.16s eta: 9:29:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31105.0 tgs: 260 data_time: 0.89s time: 119.22s eta: 9:28:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.125 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30157.0 tgs: 252 data_time: 0.52s time: 119.46s eta: 9:27:26 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32496.0 tgs: 273 data_time: 0.72s time: 118.99s eta: 9:23:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.189 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32029.0 tgs: 269 data_time: 0.78s time: 119.03s eta: 9:21:26 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31416.0 tgs: 262 data_time: 0.55s time: 119.54s eta: 9:21:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.176 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 267 data_time: 0.58s time: 119.60s eta: 9:20:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.191 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29543.0 tgs: 248 data_time: 0.53s time: 119.11s eta: 9:15:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32024.0 tgs: 269 data_time: 0.92s time: 119.01s eta: 9:13:22 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 270 data_time: 0.60s time: 118.88s eta: 9:10:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.59s time: 119.32s eta: 9:10:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31365.0 tgs: 263 data_time: 0.70s time: 119.09s eta: 9:07:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31634.0 tgs: 266 data_time: 0.58s time: 118.87s eta: 9:04:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30628.0 tgs: 255 data_time: 0.58s time: 119.67s eta: 9:06:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 265 data_time: 0.56s time: 119.43s eta: 9:03:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31747.0 tgs: 265 data_time: 0.66s time: 119.43s eta: 9:01:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 265 data_time: 0.79s time: 118.74s eta: 8:56:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31565.0 tgs: 265 data_time: 0.66s time: 118.87s eta: 8:54:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31669.0 tgs: 265 data_time: 0.83s time: 119.34s eta: 8:55:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31672.0 tgs: 265 data_time: 0.70s time: 119.15s eta: 8:52:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31776.0 tgs: 266 data_time: 0.71s time: 119.27s eta: 8:50:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.231 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31555.0 tgs: 261 data_time: 0.62s time: 120.47s eta: 8:54:05 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30891.0 tgs: 259 data_time: 0.70s time: 119.24s eta: 8:46:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31424.0 tgs: 263 data_time: 0.76s time: 119.04s eta: 8:43:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31880.0 tgs: 263 data_time: 0.84s time: 120.83s eta: 8:49:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.104 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31044.0 tgs: 258 data_time: 0.85s time: 120.19s eta: 8:44:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.230 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32503.0 tgs: 273 data_time: 0.97s time: 118.83s eta: 8:36:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29174.0 tgs: 244 data_time: 0.71s time: 119.36s eta: 8:37:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32215.0 tgs: 269 data_time: 0.52s time: 119.33s eta: 8:35:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31312.0 tgs: 263 data_time: 0.88s time: 119.05s eta: 8:31:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.192 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 269 data_time: 0.77s time: 119.16s eta: 8:30:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.188 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31510.0 tgs: 264 data_time: 0.61s time: 119.04s eta: 8:27:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31201.0 tgs: 261 data_time: 0.74s time: 119.13s eta: 8:26:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31904.0 tgs: 267 data_time: 0.74s time: 119.24s eta: 8:24:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 265 data_time: 0.86s time: 119.22s eta: 8:22:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.194 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 262 data_time: 0.81s time: 119.38s eta: 8:21:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32429.0 tgs: 270 data_time: 0.86s time: 119.78s eta: 8:21:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.215 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30484.0 tgs: 255 data_time: 0.92s time: 119.20s eta: 8:16:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.279 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31298.0 tgs: 263 data_time: 0.89s time: 118.99s eta: 8:13:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31800.0 tgs: 265 data_time: 0.75s time: 119.90s eta: 8:15:34 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32184.0 tgs: 269 data_time: 0.87s time: 119.55s eta: 8:12:09 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.091 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31480.0 tgs: 263 data_time: 0.60s time: 119.59s eta: 8:10:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31809.0 tgs: 267 data_time: 0.67s time: 119.00s eta: 8:05:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31468.0 tgs: 262 data_time: 0.61s time: 119.85s eta: 8:07:23 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 267 data_time: 0.81s time: 119.71s eta: 8:04:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31390.0 tgs: 262 data_time: 0.93s time: 119.40s eta: 8:01:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30531.0 tgs: 255 data_time: 0.88s time: 119.29s eta: 7:59:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.197 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32229.0 tgs: 270 data_time: 0.82s time: 119.19s eta: 7:56:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.190 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 266 data_time: 1.05s time: 119.19s eta: 7:54:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31293.0 tgs: 263 data_time: 0.83s time: 118.94s eta: 7:51:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.68s time: 118.97s eta: 7:49:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32005.0 tgs: 268 data_time: 0.65s time: 119.42s eta: 7:49:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.193 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31538.0 tgs: 264 data_time: 0.65s time: 119.04s eta: 7:46:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32119.0 tgs: 269 data_time: 0.72s time: 119.06s eta: 7:44:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.187 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32272.0 tgs: 268 data_time: 0.79s time: 120.01s eta: 7:46:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.098 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 269 data_time: 0.66s time: 119.83s eta: 7:43:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31519.0 tgs: 264 data_time: 1.02s time: 119.17s eta: 7:38:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30931.0 tgs: 259 data_time: 0.75s time: 118.97s eta: 7:36:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31903.0 tgs: 267 data_time: 0.65s time: 119.33s eta: 7:35:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30881.0 tgs: 258 data_time: 0.70s time: 119.39s eta: 7:33:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30896.0 tgs: 259 data_time: 0.44s time: 119.06s eta: 7:30:26 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 269 data_time: 0.87s time: 118.99s eta: 7:28:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32496.0 tgs: 272 data_time: 0.62s time: 119.24s eta: 7:27:09 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30772.0 tgs: 258 data_time: 0.67s time: 118.99s eta: 7:24:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31173.0 tgs: 262 data_time: 0.87s time: 118.93s eta: 7:22:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31728.0 tgs: 266 data_time: 0.82s time: 119.08s eta: 7:20:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.207 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 271 data_time: 0.62s time: 118.99s eta: 7:18:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.83s time: 119.03s eta: 7:16:26 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31809.0 tgs: 266 data_time: 0.87s time: 119.25s eta: 7:15:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32086.0 tgs: 268 data_time: 0.79s time: 119.33s eta: 7:13:34 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 265 data_time: 0.60s time: 118.94s eta: 7:10:10 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32117.0 tgs: 269 data_time: 0.83s time: 119.21s eta: 7:09:08 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 268 data_time: 0.79s time: 119.58s eta: 7:08:30 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30473.0 tgs: 254 data_time: 0.57s time: 119.66s eta: 7:06:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32549.0 tgs: 273 data_time: 0.64s time: 118.96s eta: 7:02:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 1.00s time: 119.30s eta: 7:01:31 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30955.0 tgs: 258 data_time: 0.78s time: 119.52s eta: 7:00:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31773.0 tgs: 266 data_time: 0.59s time: 119.43s eta: 6:57:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.188 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 269 data_time: 0.64s time: 118.92s eta: 6:54:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32349.0 tgs: 271 data_time: 0.62s time: 119.34s eta: 6:53:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31464.0 tgs: 264 data_time: 0.80s time: 118.93s eta: 6:50:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31605.0 tgs: 264 data_time: 0.89s time: 119.27s eta: 6:49:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32291.0 tgs: 271 data_time: 0.71s time: 119.02s eta: 6:46:40 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30588.0 tgs: 256 data_time: 0.79s time: 119.22s eta: 6:45:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.218 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31767.0 tgs: 266 data_time: 0.94s time: 119.39s eta: 6:43:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 271 data_time: 0.67s time: 119.01s eta: 6:40:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.129 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 263 data_time: 0.83s time: 119.02s eta: 6:38:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.179 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.57s time: 119.46s eta: 6:38:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32100.0 tgs: 270 data_time: 0.76s time: 118.80s eta: 6:34:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32074.0 tgs: 268 data_time: 0.66s time: 119.44s eta: 6:34:08 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32166.0 tgs: 269 data_time: 0.66s time: 119.57s eta: 6:32:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31997.0 tgs: 268 data_time: 0.65s time: 119.38s eta: 6:29:58 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31939.0 tgs: 266 data_time: 0.81s time: 119.64s eta: 6:28:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31940.0 tgs: 267 data_time: 0.77s time: 119.36s eta: 6:25:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 269 data_time: 0.92s time: 118.82s eta: 6:22:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31495.0 tgs: 264 data_time: 0.58s time: 119.02s eta: 6:20:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32145.0 tgs: 269 data_time: 0.65s time: 119.40s eta: 6:20:05 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 268 data_time: 0.57s time: 119.34s eta: 6:17:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31841.0 tgs: 267 data_time: 0.86s time: 119.05s eta: 6:14:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31581.0 tgs: 264 data_time: 0.84s time: 119.37s eta: 6:14:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31943.0 tgs: 267 data_time: 0.71s time: 119.40s eta: 6:12:07 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 269 data_time: 0.83s time: 119.28s eta: 6:09:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31585.0 tgs: 265 data_time: 0.69s time: 119.10s eta: 6:07:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 261 data_time: 0.88s time: 121.98s eta: 6:14:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.115 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31695.0 tgs: 265 data_time: 0.57s time: 119.37s eta: 6:04:05 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31467.0 tgs: 263 data_time: 0.71s time: 119.57s eta: 6:02:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.189 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31655.0 tgs: 265 data_time: 1.07s time: 119.26s eta: 5:59:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.204 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 269 data_time: 0.88s time: 119.26s eta: 5:57:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.223 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 265 data_time: 0.58s time: 118.92s eta: 5:54:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.181 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.87s time: 119.02s eta: 5:53:05 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31350.0 tgs: 263 data_time: 0.78s time: 119.07s eta: 5:51:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31686.0 tgs: 265 data_time: 0.63s time: 119.29s eta: 5:49:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.122 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 269 data_time: 0.65s time: 118.97s eta: 5:46:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30254.0 tgs: 253 data_time: 0.73s time: 119.28s eta: 5:45:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29732.0 tgs: 249 data_time: 0.82s time: 119.32s eta: 5:44:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 267 data_time: 0.65s time: 119.05s eta: 5:41:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31846.0 tgs: 266 data_time: 0.44s time: 119.50s eta: 5:40:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 268 data_time: 0.62s time: 119.38s eta: 5:38:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32045.0 tgs: 267 data_time: 0.73s time: 119.57s eta: 5:36:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31846.0 tgs: 267 data_time: 0.85s time: 118.98s eta: 5:33:09 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30769.0 tgs: 258 data_time: 0.71s time: 118.94s eta: 5:31:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30558.0 tgs: 256 data_time: 0.69s time: 119.20s eta: 5:29:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31252.0 tgs: 261 data_time: 0.73s time: 119.40s eta: 5:28:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.161 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32242.0 tgs: 269 data_time: 0.71s time: 119.46s eta: 5:26:31 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30886.0 tgs: 259 data_time: 0.57s time: 118.95s eta: 5:23:09 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.196 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32013.0 tgs: 267 data_time: 0.94s time: 119.64s eta: 5:23:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32249.0 tgs: 270 data_time: 0.76s time: 119.34s eta: 5:20:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.218 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31534.0 tgs: 264 data_time: 0.81s time: 119.12s eta: 5:17:39 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 267 data_time: 0.94s time: 119.42s eta: 5:16:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31683.0 tgs: 265 data_time: 0.53s time: 119.55s eta: 5:14:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 265 data_time: 0.89s time: 119.01s eta: 5:11:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.184 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31397.0 tgs: 263 data_time: 0.50s time: 119.00s eta: 5:09:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.187 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31859.0 tgs: 267 data_time: 0.65s time: 119.23s eta: 5:08:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31364.0 tgs: 262 data_time: 0.56s time: 119.69s eta: 5:07:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32268.0 tgs: 270 data_time: 0.68s time: 119.11s eta: 5:03:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 270 data_time: 0.51s time: 119.07s eta: 5:01:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.199 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31082.0 tgs: 255 data_time: 0.62s time: 121.52s eta: 5:05:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31408.0 tgs: 262 data_time: 0.72s time: 119.57s eta: 4:58:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32337.0 tgs: 270 data_time: 0.45s time: 119.50s eta: 4:56:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31038.0 tgs: 259 data_time: 0.74s time: 119.73s eta: 4:55:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32368.0 tgs: 270 data_time: 0.82s time: 119.52s eta: 4:52:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.242 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31956.0 tgs: 267 data_time: 0.55s time: 119.42s eta: 4:50:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31576.0 tgs: 264 data_time: 0.72s time: 119.20s eta: 4:48:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.209 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32397.0 tgs: 272 data_time: 0.88s time: 118.89s eta: 4:45:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32170.0 tgs: 270 data_time: 0.67s time: 119.06s eta: 4:43:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.189 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32365.0 tgs: 269 data_time: 0.76s time: 119.98s eta: 4:43:57 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31796.0 tgs: 266 data_time: 0.84s time: 119.15s eta: 4:39:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31324.0 tgs: 263 data_time: 0.67s time: 119.02s eta: 4:37:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31671.0 tgs: 266 data_time: 0.81s time: 119.06s eta: 4:35:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31825.0 tgs: 266 data_time: 0.55s time: 119.37s eta: 4:34:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32352.0 tgs: 270 data_time: 0.75s time: 119.41s eta: 4:32:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30940.0 tgs: 260 data_time: 0.75s time: 118.94s eta: 4:29:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32242.0 tgs: 269 data_time: 1.27s time: 119.60s eta: 4:29:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31599.0 tgs: 265 data_time: 0.68s time: 119.24s eta: 4:26:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31588.0 tgs: 265 data_time: 0.64s time: 119.06s eta: 4:23:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31123.0 tgs: 261 data_time: 0.99s time: 118.85s eta: 4:21:28 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.224 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 271 data_time: 0.77s time: 118.76s eta: 4:19:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32032.0 tgs: 268 data_time: 0.75s time: 119.43s eta: 4:18:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32354.0 tgs: 270 data_time: 0.87s time: 119.40s eta: 4:16:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.198 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32016.0 tgs: 268 data_time: 0.88s time: 119.32s eta: 4:14:32 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.237 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 268 data_time: 0.76s time: 119.05s eta: 4:11:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.78s time: 119.12s eta: 4:10:09 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 268 data_time: 0.75s time: 119.14s eta: 4:08:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30983.0 tgs: 260 data_time: 0.56s time: 118.98s eta: 4:05:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32059.0 tgs: 269 data_time: 0.47s time: 119.13s eta: 4:04:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.196 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 270 data_time: 0.91s time: 119.50s eta: 4:02:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31027.0 tgs: 260 data_time: 0.86s time: 119.25s eta: 4:00:28 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 270 data_time: 0.67s time: 119.20s eta: 3:58:23 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31383.0 tgs: 262 data_time: 0.82s time: 119.36s eta: 3:56:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31587.0 tgs: 264 data_time: 0.72s time: 119.37s eta: 3:54:45 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32047.0 tgs: 267 data_time: 0.86s time: 119.96s eta: 3:53:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32356.0 tgs: 271 data_time: 0.51s time: 119.32s eta: 3:50:40 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32092.0 tgs: 268 data_time: 0.59s time: 119.64s eta: 3:49:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29588.0 tgs: 248 data_time: 0.75s time: 118.86s eta: 3:45:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31665.0 tgs: 264 data_time: 0.71s time: 119.71s eta: 3:45:26 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31442.0 tgs: 263 data_time: 0.61s time: 119.33s eta: 3:42:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31891.0 tgs: 267 data_time: 0.68s time: 119.27s eta: 3:40:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31225.0 tgs: 262 data_time: 0.81s time: 119.00s eta: 3:38:10 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31289.0 tgs: 261 data_time: 0.73s time: 119.43s eta: 3:36:58 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32268.0 tgs: 271 data_time: 0.85s time: 119.03s eta: 3:34:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.248 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31078.0 tgs: 261 data_time: 0.67s time: 118.99s eta: 3:32:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 266 data_time: 0.81s time: 119.22s eta: 3:30:37 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32406.0 tgs: 269 data_time: 1.04s time: 120.12s eta: 3:30:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.109 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 259 data_time: 0.76s time: 119.13s eta: 3:26:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31928.0 tgs: 267 data_time: 0.54s time: 119.32s eta: 3:24:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32100.0 tgs: 269 data_time: 0.72s time: 119.08s eta: 3:22:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 0.85s time: 119.43s eta: 3:21:02 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.107 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31117.0 tgs: 260 data_time: 0.86s time: 119.30s eta: 3:18:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32141.0 tgs: 269 data_time: 1.11s time: 119.22s eta: 3:16:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.195 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32069.0 tgs: 269 data_time: 0.79s time: 119.13s eta: 3:14:34 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32154.0 tgs: 270 data_time: 0.75s time: 119.01s eta: 3:12:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31283.0 tgs: 262 data_time: 0.51s time: 119.19s eta: 3:10:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32362.0 tgs: 267 data_time: 0.87s time: 120.92s eta: 3:11:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32090.0 tgs: 268 data_time: 0.47s time: 119.51s eta: 3:07:14 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31709.0 tgs: 265 data_time: 0.72s time: 119.26s eta: 3:04:51 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 269 data_time: 0.78s time: 118.99s eta: 3:02:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31728.0 tgs: 266 data_time: 0.97s time: 119.22s eta: 3:00:48 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30756.0 tgs: 257 data_time: 0.71s time: 119.46s eta: 2:59:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 271 data_time: 1.01s time: 118.98s eta: 2:56:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31144.0 tgs: 260 data_time: 0.57s time: 119.35s eta: 2:55:03 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31764.0 tgs: 266 data_time: 0.99s time: 119.41s eta: 2:53:08 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 267 data_time: 0.66s time: 119.35s eta: 2:51:04 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 259 data_time: 0.65s time: 121.46s eta: 2:52:04 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.260 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31432.0 tgs: 263 data_time: 0.86s time: 119.24s eta: 2:46:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31430.0 tgs: 262 data_time: 0.76s time: 119.57s eta: 2:45:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31515.0 tgs: 265 data_time: 0.77s time: 118.90s eta: 2:42:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31398.0 tgs: 263 data_time: 0.64s time: 119.19s eta: 2:40:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30824.0 tgs: 259 data_time: 0.72s time: 118.96s eta: 2:38:36 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.201 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32254.0 tgs: 270 data_time: 0.98s time: 119.23s eta: 2:36:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30158.0 tgs: 253 data_time: 0.81s time: 119.11s eta: 2:34:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.102 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30668.0 tgs: 256 data_time: 0.58s time: 119.43s eta: 2:33:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31705.0 tgs: 263 data_time: 0.86s time: 120.23s eta: 2:32:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31871.0 tgs: 267 data_time: 0.85s time: 119.29s eta: 2:29:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31978.0 tgs: 269 data_time: 0.74s time: 118.88s eta: 2:26:36 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31953.0 tgs: 267 data_time: 0.86s time: 119.52s eta: 2:25:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32051.0 tgs: 269 data_time: 0.93s time: 119.07s eta: 2:22:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31727.0 tgs: 266 data_time: 0.73s time: 119.14s eta: 2:20:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31702.0 tgs: 266 data_time: 0.89s time: 119.05s eta: 2:18:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 268 data_time: 0.84s time: 119.85s eta: 2:17:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31770.0 tgs: 266 data_time: 0.47s time: 119.31s eta: 2:15:12 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.098 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31812.0 tgs: 266 data_time: 0.87s time: 119.36s eta: 2:13:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31915.0 tgs: 266 data_time: 0.78s time: 119.68s eta: 2:11:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31860.0 tgs: 266 data_time: 0.68s time: 119.49s eta: 2:09:27 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 269 data_time: 0.69s time: 118.96s eta: 2:06:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31729.0 tgs: 265 data_time: 0.61s time: 119.37s eta: 2:05:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30980.0 tgs: 259 data_time: 0.52s time: 119.28s eta: 2:03:15 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 266 data_time: 0.74s time: 119.23s eta: 2:01:13 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31385.0 tgs: 262 data_time: 0.74s time: 119.51s eta: 1:59:30 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31827.0 tgs: 266 data_time: 0.45s time: 119.41s eta: 1:57:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.87s time: 119.53s eta: 1:55:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30740.0 tgs: 257 data_time: 0.97s time: 119.31s eta: 1:53:20 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31178.0 tgs: 261 data_time: 0.77s time: 119.25s eta: 1:51:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 271 data_time: 0.75s time: 118.93s eta: 1:49:01 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31180.0 tgs: 260 data_time: 0.68s time: 119.64s eta: 1:47:40 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 265 data_time: 0.76s time: 119.71s eta: 1:45:44 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31377.0 tgs: 262 data_time: 0.45s time: 119.35s eta: 1:43:26 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32027.0 tgs: 268 data_time: 0.55s time: 119.30s eta: 1:41:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31859.0 tgs: 267 data_time: 0.87s time: 119.28s eta: 1:39:24 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 269 data_time: 0.75s time: 118.91s eta: 1:37:06 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32443.0 tgs: 272 data_time: 0.82s time: 119.16s eta: 1:35:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32375.0 tgs: 271 data_time: 0.55s time: 119.07s eta: 1:33:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31848.0 tgs: 267 data_time: 0.67s time: 119.25s eta: 1:31:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 269 data_time: 0.64s time: 119.73s eta: 1:29:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31778.0 tgs: 267 data_time: 0.77s time: 119.02s eta: 1:27:16 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31269.0 tgs: 263 data_time: 0.67s time: 118.87s eta: 1:25:11 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 268 data_time: 0.57s time: 119.48s eta: 1:23:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 269 data_time: 0.71s time: 119.48s eta: 1:21:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32248.0 tgs: 271 data_time: 0.80s time: 118.94s eta: 1:19:17 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31102.0 tgs: 261 data_time: 0.65s time: 118.93s eta: 1:17:18 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32215.0 tgs: 263 data_time: 0.69s time: 122.13s eta: 1:17:21 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31693.0 tgs: 265 data_time: 0.85s time: 119.33s eta: 1:13:35 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30460.0 tgs: 255 data_time: 0.79s time: 119.05s eta: 1:11:25 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31907.0 tgs: 268 data_time: 0.67s time: 118.92s eta: 1:09:22 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31334.0 tgs: 263 data_time: 0.78s time: 119.10s eta: 1:07:29 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.279 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.84s time: 119.43s eta: 1:05:41 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 270 data_time: 0.93s time: 119.34s eta: 1:03:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31141.0 tgs: 261 data_time: 0.69s time: 119.17s eta: 1:01:34 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32132.0 tgs: 269 data_time: 0.85s time: 119.27s eta: 0:59:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 268 data_time: 1.18s time: 119.44s eta: 0:57:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31761.0 tgs: 266 data_time: 1.03s time: 119.40s eta: 0:55:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.62s time: 119.50s eta: 0:53:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31781.0 tgs: 266 data_time: 0.66s time: 119.36s eta: 0:51:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31804.0 tgs: 266 data_time: 0.85s time: 119.47s eta: 0:49:46 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 267 data_time: 0.69s time: 118.89s eta: 0:47:33 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31769.0 tgs: 266 data_time: 0.77s time: 119.27s eta: 0:45:43 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.96s time: 119.20s eta: 0:43:42 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 265 data_time: 0.87s time: 119.00s eta: 0:41:38 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31905.0 tgs: 268 data_time: 0.92s time: 119.04s eta: 0:39:40 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 268 data_time: 0.71s time: 119.66s eta: 0:37:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31864.0 tgs: 267 data_time: 0.53s time: 119.29s eta: 0:35:47 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32094.0 tgs: 268 data_time: 0.68s time: 119.75s eta: 0:33:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32460.0 tgs: 267 data_time: 0.89s time: 121.23s eta: 0:32:19 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 267 data_time: 0.70s time: 119.38s eta: 0:29:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32063.0 tgs: 268 data_time: 0.58s time: 119.32s eta: 0:27:50 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31458.0 tgs: 262 data_time: 0.70s time: 119.64s eta: 0:25:55 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31717.0 tgs: 266 data_time: 0.71s time: 119.09s eta: 0:23:49 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31218.0 tgs: 261 data_time: 0.80s time: 119.39s eta: 0:21:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32262.0 tgs: 270 data_time: 0.83s time: 119.26s eta: 0:19:52 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31793.0 tgs: 266 data_time: 0.59s time: 119.24s eta: 0:17:53 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 262 data_time: 0.76s time: 119.90s eta: 0:15:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31603.0 tgs: 265 data_time: 0.84s time: 119.18s eta: 0:13:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32105.0 tgs: 269 data_time: 0.91s time: 119.09s eta: 0:11:54 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 266 data_time: 0.82s time: 119.33s eta: 0:09:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31820.0 tgs: 267 data_time: 0.57s time: 119.17s eta: 0:07:56 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31856.0 tgs: 266 data_time: 1.08s time: 119.70s eta: 0:05:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32058.0 tgs: 266 data_time: 0.72s time: 120.38s eta: 0:04:00 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.095 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31388.0 tgs: 263 data_time: 0.78s time: 119.21s eta: 0:01:59 +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.28 seconds, peak gpu memory 17.4G +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 22][DP 5][SP 2][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank23.log b/20250121165312/rank23.log new file mode 100644 index 0000000000000000000000000000000000000000..feef9e925e3bbba43ed615616d6703ed00e833ed --- /dev/null +++ b/20250121165312/rank23.log @@ -0,0 +1,851 @@ +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.11s +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.53 seconds, peak gpu memory 13.4G +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.222 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.3GB text_tokens: 29620.0 tgs: 195 data_time: 1.88s time: 151.18s eta: 1 day, 0:54:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.234 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32369.0 tgs: 269 data_time: 0.81s time: 120.15s eta: 19:45:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.230 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32546.0 tgs: 271 data_time: 0.80s time: 119.68s eta: 19:38:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.312 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 270 data_time: 1.06s time: 119.65s eta: 19:36:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.264 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30927.0 tgs: 259 data_time: 0.94s time: 119.31s eta: 19:31:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.227 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32427.0 tgs: 271 data_time: 1.20s time: 119.54s eta: 19:31:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.232 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31527.0 tgs: 256 data_time: 1.11s time: 122.96s eta: 20:02:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.203 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31799.0 tgs: 265 data_time: 0.69s time: 119.89s eta: 19:30:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.223 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30776.0 tgs: 255 data_time: 0.74s time: 120.66s eta: 19:36:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31699.0 tgs: 264 data_time: 0.73s time: 119.92s eta: 19:27:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.233 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 261 data_time: 0.97s time: 120.46s eta: 19:30:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31345.0 tgs: 262 data_time: 0.69s time: 119.34s eta: 19:17:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30795.0 tgs: 257 data_time: 0.75s time: 119.68s eta: 19:18:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.245 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 262 data_time: 0.74s time: 119.18s eta: 19:12:01 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32265.0 tgs: 270 data_time: 0.85s time: 119.22s eta: 19:10:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29937.0 tgs: 251 data_time: 0.80s time: 118.92s eta: 19:05:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.285 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30786.0 tgs: 258 data_time: 0.75s time: 119.15s eta: 19:05:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31670.0 tgs: 265 data_time: 0.84s time: 119.18s eta: 19:04:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31374.0 tgs: 262 data_time: 0.63s time: 119.40s eta: 19:04:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32080.0 tgs: 268 data_time: 0.77s time: 119.33s eta: 19:01:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30100.0 tgs: 250 data_time: 0.80s time: 120.31s eta: 19:08:57 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31525.0 tgs: 256 data_time: 0.92s time: 122.92s eta: 19:31:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32200.0 tgs: 269 data_time: 0.92s time: 119.48s eta: 18:57:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31533.0 tgs: 264 data_time: 1.01s time: 119.16s eta: 18:51:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31572.0 tgs: 264 data_time: 0.77s time: 119.25s eta: 18:50:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31711.0 tgs: 265 data_time: 0.58s time: 119.39s eta: 18:50:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31027.0 tgs: 259 data_time: 0.81s time: 119.66s eta: 18:50:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31174.0 tgs: 261 data_time: 1.06s time: 119.38s eta: 18:46:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 265 data_time: 0.91s time: 119.76s eta: 18:47:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 246 data_time: 0.93s time: 129.29s eta: 20:15:21 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31822.0 tgs: 266 data_time: 0.88s time: 119.57s eta: 18:41:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31024.0 tgs: 259 data_time: 0.77s time: 119.39s eta: 18:38:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31792.0 tgs: 263 data_time: 1.02s time: 120.66s eta: 18:48:07 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31319.0 tgs: 262 data_time: 0.63s time: 119.40s eta: 18:34:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 270 data_time: 0.64s time: 119.16s eta: 18:30:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32359.0 tgs: 271 data_time: 0.87s time: 119.07s eta: 18:27:21 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31433.0 tgs: 263 data_time: 0.92s time: 119.14s eta: 18:25:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30457.0 tgs: 252 data_time: 0.88s time: 120.39s eta: 18:35:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 264 data_time: 0.88s time: 121.17s eta: 18:40:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 256 data_time: 0.66s time: 120.66s eta: 18:34:07 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32067.0 tgs: 267 data_time: 0.86s time: 119.74s eta: 18:23:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31896.0 tgs: 257 data_time: 0.93s time: 123.88s eta: 18:59:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30763.0 tgs: 257 data_time: 0.76s time: 119.56s eta: 18:17:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31487.0 tgs: 263 data_time: 0.89s time: 119.32s eta: 18:13:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31702.0 tgs: 265 data_time: 0.73s time: 119.22s eta: 18:10:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31920.0 tgs: 266 data_time: 0.92s time: 119.61s eta: 18:12:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 266 data_time: 0.76s time: 119.27s eta: 18:07:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31065.0 tgs: 261 data_time: 0.81s time: 119.01s eta: 18:03:01 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32174.0 tgs: 270 data_time: 0.78s time: 119.10s eta: 18:01:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.254 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31306.0 tgs: 262 data_time: 1.07s time: 119.26s eta: 18:01:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29824.0 tgs: 249 data_time: 0.69s time: 119.34s eta: 18:00:01 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31180.0 tgs: 259 data_time: 1.09s time: 120.02s eta: 18:04:10 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31929.0 tgs: 268 data_time: 0.89s time: 119.10s eta: 17:53:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32302.0 tgs: 270 data_time: 0.58s time: 119.41s eta: 17:54:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31344.0 tgs: 263 data_time: 0.69s time: 119.00s eta: 17:48:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30272.0 tgs: 254 data_time: 0.78s time: 119.03s eta: 17:47:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32061.0 tgs: 255 data_time: 0.72s time: 125.36s eta: 18:42:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 265 data_time: 0.69s time: 120.10s eta: 17:52:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32402.0 tgs: 271 data_time: 1.26s time: 119.36s eta: 17:44:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30488.0 tgs: 255 data_time: 0.88s time: 119.33s eta: 17:42:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 266 data_time: 0.59s time: 119.51s eta: 17:41:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31778.0 tgs: 252 data_time: 0.59s time: 126.05s eta: 18:37:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 259 data_time: 0.96s time: 122.64s eta: 18:05:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 264 data_time: 1.39s time: 121.15s eta: 17:50:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30998.0 tgs: 260 data_time: 0.69s time: 119.04s eta: 17:29:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.126 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31047.0 tgs: 261 data_time: 0.90s time: 118.86s eta: 17:25:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32302.0 tgs: 260 data_time: 0.89s time: 124.00s eta: 18:09:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30866.0 tgs: 247 data_time: 0.64s time: 124.57s eta: 18:12:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31849.0 tgs: 266 data_time: 0.73s time: 119.39s eta: 17:24:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31641.0 tgs: 264 data_time: 0.83s time: 119.70s eta: 17:25:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32468.0 tgs: 271 data_time: 0.76s time: 119.53s eta: 17:21:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.123 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31006.0 tgs: 260 data_time: 0.72s time: 118.86s eta: 17:14:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31253.0 tgs: 260 data_time: 0.77s time: 119.85s eta: 17:20:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31654.0 tgs: 264 data_time: 0.97s time: 119.52s eta: 17:15:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31945.0 tgs: 262 data_time: 0.78s time: 121.86s eta: 17:34:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 253 data_time: 0.55s time: 123.52s eta: 17:46:21 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30448.0 tgs: 255 data_time: 0.64s time: 119.01s eta: 17:05:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31752.0 tgs: 266 data_time: 0.88s time: 119.23s eta: 17:05:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 270 data_time: 0.71s time: 119.14s eta: 17:02:36 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32172.0 tgs: 269 data_time: 1.00s time: 119.26s eta: 17:01:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31208.0 tgs: 262 data_time: 0.97s time: 119.04s eta: 16:57:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32488.0 tgs: 272 data_time: 0.83s time: 119.28s eta: 16:57:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31692.0 tgs: 265 data_time: 0.89s time: 119.16s eta: 16:54:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 266 data_time: 0.88s time: 119.59s eta: 16:56:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32503.0 tgs: 272 data_time: 0.99s time: 119.13s eta: 16:50:36 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31637.0 tgs: 266 data_time: 0.71s time: 118.84s eta: 16:46:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.221 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32396.0 tgs: 272 data_time: 0.77s time: 119.02s eta: 16:45:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31701.0 tgs: 264 data_time: 0.67s time: 119.85s eta: 16:50:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32363.0 tgs: 271 data_time: 0.96s time: 119.36s eta: 16:44:36 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31309.0 tgs: 263 data_time: 0.56s time: 119.01s eta: 16:39:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 262 data_time: 1.03s time: 120.93s eta: 16:53:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31214.0 tgs: 262 data_time: 0.68s time: 118.99s eta: 16:35:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.234 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31653.0 tgs: 265 data_time: 0.84s time: 119.29s eta: 16:36:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.243 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32264.0 tgs: 270 data_time: 0.95s time: 119.23s eta: 16:33:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32112.0 tgs: 269 data_time: 0.97s time: 119.37s eta: 16:32:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31448.0 tgs: 262 data_time: 0.81s time: 119.70s eta: 16:33:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31670.0 tgs: 265 data_time: 0.78s time: 119.20s eta: 16:27:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32024.0 tgs: 267 data_time: 0.60s time: 119.76s eta: 16:30:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31293.0 tgs: 260 data_time: 0.75s time: 120.35s eta: 16:32:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31374.0 tgs: 263 data_time: 0.75s time: 118.98s eta: 16:19:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.209 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30907.0 tgs: 242 data_time: 0.84s time: 127.33s eta: 17:26:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 267 data_time: 0.64s time: 119.23s eta: 16:17:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31295.0 tgs: 262 data_time: 0.65s time: 119.14s eta: 16:14:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31271.0 tgs: 262 data_time: 0.76s time: 119.21s eta: 16:13:31 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29928.0 tgs: 251 data_time: 0.57s time: 119.17s eta: 16:11:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31218.0 tgs: 262 data_time: 0.77s time: 118.95s eta: 16:07:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.205 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 267 data_time: 0.79s time: 118.86s eta: 16:04:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32258.0 tgs: 271 data_time: 0.72s time: 118.96s eta: 16:03:36 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 259 data_time: 0.80s time: 121.84s eta: 16:24:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 268 data_time: 1.00s time: 119.00s eta: 15:59:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30581.0 tgs: 256 data_time: 0.91s time: 119.23s eta: 15:59:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32435.0 tgs: 271 data_time: 0.55s time: 119.62s eta: 16:00:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31232.0 tgs: 262 data_time: 0.95s time: 119.04s eta: 15:54:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.201 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31684.0 tgs: 266 data_time: 1.07s time: 118.98s eta: 15:51:51 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 267 data_time: 0.83s time: 119.17s eta: 15:51:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32043.0 tgs: 267 data_time: 0.61s time: 119.64s eta: 15:53:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31617.0 tgs: 265 data_time: 0.91s time: 118.94s eta: 15:45:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31764.0 tgs: 266 data_time: 0.64s time: 119.29s eta: 15:46:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32084.0 tgs: 269 data_time: 1.01s time: 119.10s eta: 15:42:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30935.0 tgs: 259 data_time: 0.93s time: 119.00s eta: 15:40:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.220 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32150.0 tgs: 268 data_time: 0.82s time: 119.66s eta: 15:43:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 262 data_time: 0.83s time: 121.50s eta: 15:55:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31485.0 tgs: 263 data_time: 0.56s time: 119.41s eta: 15:37:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32015.0 tgs: 246 data_time: 0.84s time: 129.97s eta: 16:58:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31263.0 tgs: 261 data_time: 0.93s time: 119.47s eta: 15:33:51 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30780.0 tgs: 257 data_time: 0.72s time: 119.60s eta: 15:32:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.116 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32045.0 tgs: 268 data_time: 0.85s time: 119.53s eta: 15:30:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 259 data_time: 0.76s time: 122.14s eta: 15:48:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30999.0 tgs: 256 data_time: 0.90s time: 120.75s eta: 15:35:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31669.0 tgs: 265 data_time: 1.02s time: 119.41s eta: 15:23:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31113.0 tgs: 260 data_time: 0.71s time: 119.45s eta: 15:21:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 238 data_time: 0.90s time: 131.63s eta: 16:53:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 265 data_time: 0.93s time: 119.29s eta: 15:16:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 264 data_time: 0.57s time: 119.52s eta: 15:16:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32096.0 tgs: 262 data_time: 0.78s time: 122.25s eta: 15:35:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31552.0 tgs: 263 data_time: 0.72s time: 119.71s eta: 15:13:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.80s time: 119.35s eta: 15:09:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30937.0 tgs: 260 data_time: 0.91s time: 118.90s eta: 15:03:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32455.0 tgs: 272 data_time: 1.03s time: 119.00s eta: 15:02:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31945.0 tgs: 267 data_time: 0.81s time: 119.31s eta: 15:02:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 265 data_time: 0.70s time: 118.92s eta: 14:57:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30856.0 tgs: 256 data_time: 0.67s time: 120.52s eta: 15:07:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31360.0 tgs: 262 data_time: 0.83s time: 119.31s eta: 14:56:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32316.0 tgs: 263 data_time: 0.72s time: 122.66s eta: 15:19:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.70s time: 119.13s eta: 14:51:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30701.0 tgs: 256 data_time: 0.70s time: 119.76s eta: 14:54:10 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 265 data_time: 0.75s time: 119.47s eta: 14:50:04 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 270 data_time: 0.79s time: 118.99s eta: 14:44:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.60s time: 119.30s eta: 14:44:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32218.0 tgs: 270 data_time: 1.02s time: 118.94s eta: 14:40:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31462.0 tgs: 264 data_time: 0.76s time: 119.05s eta: 14:38:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30925.0 tgs: 259 data_time: 0.88s time: 119.06s eta: 14:37:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.106 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32210.0 tgs: 270 data_time: 0.73s time: 119.13s eta: 14:35:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.66s time: 119.00s eta: 14:32:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31753.0 tgs: 266 data_time: 0.97s time: 119.08s eta: 14:31:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32043.0 tgs: 268 data_time: 0.87s time: 119.15s eta: 14:29:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 269 data_time: 1.06s time: 119.65s eta: 14:31:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.202 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30864.0 tgs: 258 data_time: 1.13s time: 119.18s eta: 14:26:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.80s time: 119.20s eta: 14:24:10 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31553.0 tgs: 264 data_time: 0.68s time: 119.39s eta: 14:23:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30653.0 tgs: 253 data_time: 0.75s time: 120.74s eta: 14:31:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31467.0 tgs: 264 data_time: 0.80s time: 118.86s eta: 14:15:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.220 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 269 data_time: 0.76s time: 119.10s eta: 14:15:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31113.0 tgs: 261 data_time: 0.88s time: 119.12s eta: 14:13:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31202.0 tgs: 260 data_time: 0.49s time: 119.95s eta: 14:17:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 268 data_time: 1.02s time: 119.87s eta: 14:15:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.272 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31916.0 tgs: 266 data_time: 0.75s time: 119.58s eta: 14:11:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31679.0 tgs: 265 data_time: 0.78s time: 119.23s eta: 14:06:30 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.82s time: 119.27s eta: 14:04:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31455.0 tgs: 263 data_time: 0.80s time: 119.26s eta: 14:02:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.270 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31473.0 tgs: 264 data_time: 0.99s time: 119.15s eta: 13:59:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31878.0 tgs: 267 data_time: 0.83s time: 119.13s eta: 13:57:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31593.0 tgs: 265 data_time: 0.90s time: 119.17s eta: 13:56:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31814.0 tgs: 266 data_time: 0.86s time: 119.35s eta: 13:55:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31236.0 tgs: 260 data_time: 0.71s time: 120.12s eta: 13:58:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31467.0 tgs: 256 data_time: 1.03s time: 122.55s eta: 14:13:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32419.0 tgs: 272 data_time: 0.54s time: 119.10s eta: 13:47:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.219 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31622.0 tgs: 264 data_time: 1.01s time: 119.37s eta: 13:47:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 270 data_time: 0.80s time: 119.32s eta: 13:45:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31542.0 tgs: 264 data_time: 0.66s time: 119.33s eta: 13:43:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31234.0 tgs: 262 data_time: 0.67s time: 119.11s eta: 13:39:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 268 data_time: 0.74s time: 119.26s eta: 13:38:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.92s time: 119.14s eta: 13:36:07 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 1.11s time: 119.22s eta: 13:34:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.121 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32114.0 tgs: 268 data_time: 0.74s time: 119.46s eta: 13:34:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.209 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30360.0 tgs: 254 data_time: 0.60s time: 119.08s eta: 13:29:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31389.0 tgs: 263 data_time: 0.98s time: 118.92s eta: 13:26:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31445.0 tgs: 262 data_time: 0.91s time: 119.81s eta: 13:30:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32275.0 tgs: 270 data_time: 0.68s time: 119.26s eta: 13:24:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32417.0 tgs: 272 data_time: 0.80s time: 119.16s eta: 13:22:21 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.82s time: 119.35s eta: 13:21:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31655.0 tgs: 265 data_time: 0.68s time: 119.26s eta: 13:19:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32283.0 tgs: 271 data_time: 0.88s time: 119.04s eta: 13:15:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32317.0 tgs: 270 data_time: 0.73s time: 119.35s eta: 13:15:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.78s time: 119.17s eta: 13:12:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.131 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32616.0 tgs: 273 data_time: 0.85s time: 119.46s eta: 13:12:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30892.0 tgs: 258 data_time: 0.76s time: 119.29s eta: 13:09:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.106 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31394.0 tgs: 263 data_time: 0.89s time: 119.15s eta: 13:06:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 266 data_time: 0.81s time: 120.32s eta: 13:12:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32445.0 tgs: 270 data_time: 0.65s time: 120.17s eta: 13:09:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.187 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31252.0 tgs: 261 data_time: 0.70s time: 119.51s eta: 13:02:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 266 data_time: 0.85s time: 119.51s eta: 13:00:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 270 data_time: 0.78s time: 119.57s eta: 12:59:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31257.0 tgs: 260 data_time: 0.80s time: 119.80s eta: 12:58:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30989.0 tgs: 260 data_time: 0.80s time: 119.04s eta: 12:51:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31367.0 tgs: 263 data_time: 0.82s time: 119.05s eta: 12:49:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 266 data_time: 0.86s time: 119.13s eta: 12:48:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30623.0 tgs: 256 data_time: 0.73s time: 119.39s eta: 12:48:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.104 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 267 data_time: 0.73s time: 119.19s eta: 12:44:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 0.65s time: 119.42s eta: 12:44:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.112 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.76s time: 119.10s eta: 12:40:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31061.0 tgs: 260 data_time: 0.70s time: 119.07s eta: 12:38:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31784.0 tgs: 265 data_time: 0.96s time: 119.60s eta: 12:39:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 270 data_time: 0.66s time: 118.78s eta: 12:32:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 270 data_time: 0.93s time: 118.89s eta: 12:30:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.222 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30989.0 tgs: 260 data_time: 0.85s time: 119.17s eta: 12:30:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.72s time: 119.27s eta: 12:29:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31282.0 tgs: 262 data_time: 0.75s time: 119.35s eta: 12:27:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32386.0 tgs: 271 data_time: 0.76s time: 119.12s eta: 12:24:30 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.185 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 270 data_time: 0.84s time: 119.34s eta: 12:23:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.79s time: 119.27s eta: 12:21:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 268 data_time: 0.84s time: 119.03s eta: 12:17:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32053.0 tgs: 264 data_time: 0.87s time: 121.13s eta: 12:29:01 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31183.0 tgs: 259 data_time: 0.72s time: 119.98s eta: 12:19:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30267.0 tgs: 255 data_time: 0.92s time: 118.69s eta: 12:09:57 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.198 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31274.0 tgs: 261 data_time: 0.72s time: 119.64s eta: 12:13:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32119.0 tgs: 268 data_time: 0.79s time: 119.83s eta: 12:12:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.289 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32165.0 tgs: 269 data_time: 0.77s time: 119.16s eta: 12:06:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31657.0 tgs: 265 data_time: 0.74s time: 119.17s eta: 12:04:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31215.0 tgs: 262 data_time: 0.91s time: 119.12s eta: 12:02:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.189 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32433.0 tgs: 272 data_time: 0.95s time: 119.19s eta: 12:01:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31078.0 tgs: 261 data_time: 0.89s time: 118.95s eta: 11:57:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.249 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31699.0 tgs: 264 data_time: 1.04s time: 119.71s eta: 12:00:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.206 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31777.0 tgs: 266 data_time: 0.76s time: 119.28s eta: 11:55:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32039.0 tgs: 268 data_time: 0.92s time: 119.28s eta: 11:53:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.184 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32039.0 tgs: 267 data_time: 0.63s time: 119.71s eta: 11:54:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32191.0 tgs: 270 data_time: 0.73s time: 118.79s eta: 11:46:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.168 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32174.0 tgs: 270 data_time: 0.71s time: 119.16s eta: 11:47:01 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31477.0 tgs: 263 data_time: 0.73s time: 119.61s eta: 11:47:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31088.0 tgs: 261 data_time: 0.88s time: 118.82s eta: 11:41:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 268 data_time: 0.68s time: 118.68s eta: 11:38:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.212 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31933.0 tgs: 267 data_time: 1.18s time: 119.21s eta: 11:39:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31321.0 tgs: 262 data_time: 0.93s time: 119.14s eta: 11:36:57 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31407.0 tgs: 263 data_time: 0.86s time: 119.08s eta: 11:34:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 266 data_time: 0.81s time: 118.95s eta: 11:31:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 268 data_time: 0.77s time: 118.91s eta: 11:29:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 268 data_time: 0.65s time: 118.97s eta: 11:28:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.109 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31586.0 tgs: 264 data_time: 0.62s time: 119.29s eta: 11:27:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.178 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31737.0 tgs: 265 data_time: 0.60s time: 119.36s eta: 11:26:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30668.0 tgs: 257 data_time: 0.69s time: 118.88s eta: 11:21:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 0.83s time: 119.22s eta: 11:21:30 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31176.0 tgs: 260 data_time: 0.48s time: 119.72s eta: 11:22:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32600.0 tgs: 273 data_time: 0.68s time: 119.11s eta: 11:16:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31815.0 tgs: 267 data_time: 0.91s time: 119.12s eta: 11:15:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.213 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 268 data_time: 0.82s time: 118.94s eta: 11:12:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29524.0 tgs: 247 data_time: 0.59s time: 119.09s eta: 11:10:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31479.0 tgs: 264 data_time: 0.99s time: 119.16s eta: 11:09:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.207 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30973.0 tgs: 260 data_time: 0.78s time: 118.97s eta: 11:06:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.107 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31341.0 tgs: 262 data_time: 0.83s time: 119.47s eta: 11:07:01 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31042.0 tgs: 259 data_time: 0.78s time: 119.65s eta: 11:06:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31421.0 tgs: 263 data_time: 0.98s time: 119.31s eta: 11:02:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 269 data_time: 0.75s time: 119.31s eta: 11:00:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32427.0 tgs: 272 data_time: 0.72s time: 118.82s eta: 10:55:30 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32089.0 tgs: 268 data_time: 0.99s time: 119.47s eta: 10:57:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31058.0 tgs: 260 data_time: 1.15s time: 119.37s eta: 10:54:31 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30771.0 tgs: 257 data_time: 0.85s time: 119.28s eta: 10:52:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31083.0 tgs: 260 data_time: 0.84s time: 119.26s eta: 10:49:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.184 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31601.0 tgs: 265 data_time: 0.99s time: 119.10s eta: 10:47:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.84s time: 119.54s eta: 10:47:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31939.0 tgs: 267 data_time: 0.57s time: 119.31s eta: 10:44:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.203 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 267 data_time: 0.77s time: 119.49s eta: 10:43:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31010.0 tgs: 259 data_time: 1.08s time: 119.52s eta: 10:41:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 270 data_time: 0.95s time: 118.99s eta: 10:36:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 268 data_time: 0.77s time: 119.22s eta: 10:35:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32385.0 tgs: 271 data_time: 0.73s time: 119.07s eta: 10:33:01 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31393.0 tgs: 264 data_time: 0.88s time: 118.78s eta: 10:29:31 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32286.0 tgs: 269 data_time: 0.88s time: 119.86s eta: 10:33:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.221 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30996.0 tgs: 259 data_time: 1.02s time: 119.38s eta: 10:28:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31997.0 tgs: 269 data_time: 0.81s time: 118.84s eta: 10:23:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.226 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31667.0 tgs: 266 data_time: 0.78s time: 118.98s eta: 10:22:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 267 data_time: 0.98s time: 119.29s eta: 10:22:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31505.0 tgs: 264 data_time: 0.92s time: 119.21s eta: 10:19:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.206 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 266 data_time: 0.98s time: 119.06s eta: 10:17:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31411.0 tgs: 262 data_time: 0.76s time: 119.48s eta: 10:17:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31521.0 tgs: 264 data_time: 1.01s time: 119.38s eta: 10:14:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32061.0 tgs: 268 data_time: 0.85s time: 119.45s eta: 10:13:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32424.0 tgs: 271 data_time: 0.91s time: 119.29s eta: 10:10:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 269 data_time: 0.65s time: 119.11s eta: 10:07:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.229 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32042.0 tgs: 267 data_time: 0.85s time: 119.71s eta: 10:08:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31403.0 tgs: 262 data_time: 0.69s time: 119.83s eta: 10:07:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.195 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31488.0 tgs: 259 data_time: 0.74s time: 121.55s eta: 10:13:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31693.0 tgs: 265 data_time: 0.50s time: 119.26s eta: 10:00:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32002.0 tgs: 269 data_time: 0.87s time: 118.80s eta: 9:56:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31838.0 tgs: 266 data_time: 0.87s time: 119.56s eta: 9:57:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31373.0 tgs: 262 data_time: 0.73s time: 119.70s eta: 9:56:31 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31726.0 tgs: 266 data_time: 0.90s time: 119.26s eta: 9:52:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31170.0 tgs: 261 data_time: 0.76s time: 119.33s eta: 9:50:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31632.0 tgs: 265 data_time: 0.68s time: 119.23s eta: 9:48:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31866.0 tgs: 263 data_time: 0.65s time: 120.74s eta: 9:53:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32032.0 tgs: 268 data_time: 0.73s time: 119.23s eta: 9:44:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30902.0 tgs: 258 data_time: 0.90s time: 119.46s eta: 9:43:21 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32041.0 tgs: 268 data_time: 1.06s time: 119.55s eta: 9:41:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31125.0 tgs: 261 data_time: 0.66s time: 119.20s eta: 9:38:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31950.0 tgs: 267 data_time: 0.82s time: 119.48s eta: 9:37:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.110 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31647.0 tgs: 265 data_time: 0.86s time: 119.05s eta: 9:33:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32268.0 tgs: 270 data_time: 0.74s time: 119.24s eta: 9:32:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.220 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31695.0 tgs: 265 data_time: 1.11s time: 119.17s eta: 9:30:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.221 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31105.0 tgs: 260 data_time: 0.92s time: 119.22s eta: 9:28:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30157.0 tgs: 252 data_time: 0.54s time: 119.47s eta: 9:27:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.209 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32496.0 tgs: 273 data_time: 0.73s time: 118.98s eta: 9:23:10 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32029.0 tgs: 269 data_time: 0.83s time: 119.03s eta: 9:21:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.213 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31416.0 tgs: 262 data_time: 0.57s time: 119.53s eta: 9:21:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 267 data_time: 0.60s time: 119.59s eta: 9:20:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.188 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29543.0 tgs: 248 data_time: 0.54s time: 119.12s eta: 9:15:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.154 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32024.0 tgs: 269 data_time: 0.94s time: 119.00s eta: 9:13:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 270 data_time: 0.62s time: 118.88s eta: 9:10:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.61s time: 119.32s eta: 9:10:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31365.0 tgs: 263 data_time: 0.71s time: 119.09s eta: 9:07:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31634.0 tgs: 266 data_time: 0.61s time: 118.86s eta: 9:04:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30628.0 tgs: 255 data_time: 0.60s time: 119.67s eta: 9:06:28 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 265 data_time: 0.58s time: 119.44s eta: 9:03:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31747.0 tgs: 265 data_time: 0.68s time: 119.43s eta: 9:01:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 265 data_time: 0.82s time: 118.74s eta: 8:56:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31565.0 tgs: 265 data_time: 0.69s time: 118.88s eta: 8:54:57 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.203 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31669.0 tgs: 265 data_time: 0.86s time: 119.33s eta: 8:54:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.192 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31672.0 tgs: 265 data_time: 0.73s time: 119.14s eta: 8:52:10 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31776.0 tgs: 266 data_time: 0.73s time: 119.27s eta: 8:50:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31555.0 tgs: 261 data_time: 0.64s time: 120.48s eta: 8:54:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.117 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30891.0 tgs: 259 data_time: 0.73s time: 119.24s eta: 8:46:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31424.0 tgs: 263 data_time: 0.79s time: 119.05s eta: 8:43:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.222 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31880.0 tgs: 263 data_time: 0.87s time: 120.83s eta: 8:49:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.273 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31044.0 tgs: 258 data_time: 0.87s time: 120.20s eta: 8:44:51 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32503.0 tgs: 273 data_time: 1.00s time: 118.83s eta: 8:36:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29174.0 tgs: 244 data_time: 0.72s time: 119.35s eta: 8:37:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32215.0 tgs: 269 data_time: 0.55s time: 119.33s eta: 8:35:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31312.0 tgs: 263 data_time: 0.92s time: 119.05s eta: 8:31:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 269 data_time: 0.81s time: 119.16s eta: 8:30:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31510.0 tgs: 264 data_time: 0.63s time: 119.04s eta: 8:27:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31201.0 tgs: 261 data_time: 0.77s time: 119.12s eta: 8:26:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31904.0 tgs: 267 data_time: 0.77s time: 119.24s eta: 8:24:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 265 data_time: 0.88s time: 119.22s eta: 8:22:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.195 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 262 data_time: 0.84s time: 119.38s eta: 8:21:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32429.0 tgs: 270 data_time: 0.88s time: 119.78s eta: 8:21:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30484.0 tgs: 255 data_time: 0.95s time: 119.20s eta: 8:16:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31298.0 tgs: 263 data_time: 0.94s time: 118.98s eta: 8:13:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31800.0 tgs: 265 data_time: 0.80s time: 119.90s eta: 8:15:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.188 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32184.0 tgs: 269 data_time: 0.91s time: 119.55s eta: 8:12:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31480.0 tgs: 263 data_time: 0.63s time: 119.60s eta: 8:10:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.122 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31809.0 tgs: 267 data_time: 0.70s time: 119.00s eta: 8:05:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31468.0 tgs: 262 data_time: 0.64s time: 119.84s eta: 8:07:21 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 267 data_time: 0.84s time: 119.71s eta: 8:04:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31390.0 tgs: 262 data_time: 0.96s time: 119.40s eta: 8:01:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30531.0 tgs: 255 data_time: 0.91s time: 119.29s eta: 7:59:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32229.0 tgs: 270 data_time: 0.86s time: 119.19s eta: 7:56:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 266 data_time: 1.08s time: 119.20s eta: 7:54:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31293.0 tgs: 263 data_time: 0.86s time: 118.94s eta: 7:51:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.72s time: 118.97s eta: 7:49:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32005.0 tgs: 268 data_time: 0.68s time: 119.42s eta: 7:49:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.226 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31538.0 tgs: 264 data_time: 0.68s time: 119.04s eta: 7:46:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32119.0 tgs: 269 data_time: 0.73s time: 119.07s eta: 7:44:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32272.0 tgs: 268 data_time: 0.84s time: 120.01s eta: 7:46:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 269 data_time: 0.69s time: 119.83s eta: 7:43:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.208 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31519.0 tgs: 264 data_time: 1.05s time: 119.16s eta: 7:38:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30931.0 tgs: 259 data_time: 0.72s time: 118.97s eta: 7:36:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31903.0 tgs: 267 data_time: 0.67s time: 119.33s eta: 7:35:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30881.0 tgs: 258 data_time: 0.72s time: 119.38s eta: 7:33:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30896.0 tgs: 259 data_time: 0.46s time: 119.06s eta: 7:30:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.207 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 269 data_time: 0.88s time: 119.00s eta: 7:28:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32496.0 tgs: 272 data_time: 0.64s time: 119.24s eta: 7:27:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30772.0 tgs: 258 data_time: 0.71s time: 118.99s eta: 7:24:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31173.0 tgs: 262 data_time: 0.93s time: 118.93s eta: 7:22:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.177 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31728.0 tgs: 266 data_time: 0.87s time: 119.08s eta: 7:20:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 271 data_time: 0.65s time: 118.99s eta: 7:18:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.87s time: 119.03s eta: 7:16:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31809.0 tgs: 266 data_time: 0.92s time: 119.25s eta: 7:15:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32086.0 tgs: 268 data_time: 0.82s time: 119.34s eta: 7:13:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.237 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 265 data_time: 0.63s time: 118.94s eta: 7:10:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32117.0 tgs: 269 data_time: 0.87s time: 119.21s eta: 7:09:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 268 data_time: 0.85s time: 119.60s eta: 7:08:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30473.0 tgs: 254 data_time: 0.59s time: 119.65s eta: 7:06:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32549.0 tgs: 273 data_time: 0.68s time: 118.96s eta: 7:02:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 1.00s time: 119.30s eta: 7:01:31 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30955.0 tgs: 258 data_time: 0.79s time: 119.52s eta: 7:00:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31773.0 tgs: 266 data_time: 0.61s time: 119.43s eta: 6:57:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 269 data_time: 0.66s time: 118.91s eta: 6:54:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.174 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32349.0 tgs: 271 data_time: 0.64s time: 119.35s eta: 6:53:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31464.0 tgs: 264 data_time: 0.82s time: 118.94s eta: 6:50:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31605.0 tgs: 265 data_time: 0.92s time: 119.26s eta: 6:49:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32291.0 tgs: 271 data_time: 0.73s time: 119.02s eta: 6:46:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30588.0 tgs: 256 data_time: 0.80s time: 119.23s eta: 6:45:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.115 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31767.0 tgs: 266 data_time: 0.96s time: 119.38s eta: 6:43:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 271 data_time: 0.69s time: 119.01s eta: 6:40:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.182 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 263 data_time: 0.88s time: 119.01s eta: 6:38:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.59s time: 119.46s eta: 6:38:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32100.0 tgs: 270 data_time: 0.80s time: 118.81s eta: 6:34:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32074.0 tgs: 268 data_time: 0.68s time: 119.44s eta: 6:34:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32166.0 tgs: 269 data_time: 0.69s time: 119.57s eta: 6:32:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31997.0 tgs: 268 data_time: 0.66s time: 119.39s eta: 6:29:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31939.0 tgs: 267 data_time: 0.85s time: 119.56s eta: 6:28:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31940.0 tgs: 267 data_time: 0.82s time: 119.36s eta: 6:25:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 269 data_time: 0.97s time: 118.82s eta: 6:22:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31495.0 tgs: 264 data_time: 0.60s time: 119.02s eta: 6:20:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.107 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32145.0 tgs: 269 data_time: 0.68s time: 119.40s eta: 6:20:04 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 268 data_time: 0.58s time: 119.34s eta: 6:17:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.186 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31841.0 tgs: 267 data_time: 0.88s time: 119.05s eta: 6:15:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31581.0 tgs: 264 data_time: 0.85s time: 119.38s eta: 6:14:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31943.0 tgs: 267 data_time: 0.75s time: 119.39s eta: 6:12:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 269 data_time: 0.87s time: 119.28s eta: 6:09:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31585.0 tgs: 265 data_time: 0.70s time: 119.10s eta: 6:07:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.205 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 261 data_time: 0.91s time: 121.96s eta: 6:14:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31695.0 tgs: 265 data_time: 0.59s time: 119.39s eta: 6:04:07 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31467.0 tgs: 263 data_time: 0.74s time: 119.57s eta: 6:02:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.190 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31655.0 tgs: 265 data_time: 1.09s time: 119.26s eta: 5:59:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 269 data_time: 0.90s time: 119.26s eta: 5:57:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.177 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 265 data_time: 0.59s time: 118.92s eta: 5:54:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.90s time: 119.02s eta: 5:53:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.223 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31350.0 tgs: 263 data_time: 0.82s time: 119.07s eta: 5:51:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31686.0 tgs: 265 data_time: 0.64s time: 119.29s eta: 5:49:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 269 data_time: 0.68s time: 118.97s eta: 5:46:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30254.0 tgs: 253 data_time: 0.75s time: 119.28s eta: 5:45:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.115 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29732.0 tgs: 249 data_time: 0.84s time: 119.32s eta: 5:44:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.189 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 267 data_time: 0.69s time: 119.05s eta: 5:41:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31846.0 tgs: 266 data_time: 0.46s time: 119.50s eta: 5:40:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 268 data_time: 0.65s time: 119.38s eta: 5:38:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32045.0 tgs: 268 data_time: 0.75s time: 119.56s eta: 5:36:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.178 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31846.0 tgs: 267 data_time: 0.88s time: 118.98s eta: 5:33:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30769.0 tgs: 258 data_time: 0.72s time: 118.95s eta: 5:31:04 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30558.0 tgs: 256 data_time: 0.70s time: 119.24s eta: 5:29:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.189 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31252.0 tgs: 261 data_time: 0.75s time: 119.40s eta: 5:28:21 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32242.0 tgs: 269 data_time: 0.84s time: 119.46s eta: 5:26:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30886.0 tgs: 259 data_time: 0.59s time: 118.95s eta: 5:23:08 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.192 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32013.0 tgs: 267 data_time: 0.96s time: 119.63s eta: 5:22:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.203 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32249.0 tgs: 270 data_time: 0.79s time: 119.35s eta: 5:20:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.196 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31534.0 tgs: 264 data_time: 0.83s time: 119.13s eta: 5:17:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 267 data_time: 0.97s time: 119.42s eta: 5:16:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31683.0 tgs: 265 data_time: 0.56s time: 119.55s eta: 5:14:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.231 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 265 data_time: 0.91s time: 119.01s eta: 5:11:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.197 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31397.0 tgs: 263 data_time: 0.52s time: 119.00s eta: 5:09:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31859.0 tgs: 267 data_time: 0.67s time: 119.23s eta: 5:08:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31364.0 tgs: 262 data_time: 0.58s time: 119.69s eta: 5:07:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32268.0 tgs: 270 data_time: 0.70s time: 119.11s eta: 5:03:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 270 data_time: 0.52s time: 119.07s eta: 5:01:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31082.0 tgs: 255 data_time: 0.64s time: 121.52s eta: 5:05:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31408.0 tgs: 262 data_time: 0.75s time: 119.58s eta: 4:58:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32337.0 tgs: 270 data_time: 0.48s time: 119.50s eta: 4:56:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.230 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31038.0 tgs: 259 data_time: 0.78s time: 119.73s eta: 4:55:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32368.0 tgs: 270 data_time: 0.85s time: 119.52s eta: 4:52:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31956.0 tgs: 267 data_time: 0.58s time: 119.42s eta: 4:50:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31576.0 tgs: 264 data_time: 0.76s time: 119.21s eta: 4:48:04 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32397.0 tgs: 272 data_time: 0.93s time: 118.89s eta: 4:45:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32170.0 tgs: 270 data_time: 0.70s time: 119.06s eta: 4:43:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32365.0 tgs: 269 data_time: 0.81s time: 119.98s eta: 4:43:57 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.186 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31796.0 tgs: 266 data_time: 0.88s time: 119.15s eta: 4:40:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31324.0 tgs: 263 data_time: 0.70s time: 119.01s eta: 4:37:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31671.0 tgs: 265 data_time: 0.84s time: 119.07s eta: 4:35:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.178 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31825.0 tgs: 266 data_time: 0.61s time: 119.36s eta: 4:34:31 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32352.0 tgs: 270 data_time: 0.78s time: 119.41s eta: 4:32:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30940.0 tgs: 260 data_time: 0.79s time: 118.94s eta: 4:29:35 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.186 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32242.0 tgs: 269 data_time: 1.31s time: 119.60s eta: 4:29:05 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31599.0 tgs: 264 data_time: 0.71s time: 119.24s eta: 4:26:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31588.0 tgs: 265 data_time: 0.66s time: 119.07s eta: 4:23:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31123.0 tgs: 261 data_time: 1.02s time: 118.85s eta: 4:21:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 271 data_time: 0.80s time: 118.77s eta: 4:19:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.227 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32032.0 tgs: 268 data_time: 0.77s time: 119.42s eta: 4:18:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.205 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32354.0 tgs: 270 data_time: 0.91s time: 119.40s eta: 4:16:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.192 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32016.0 tgs: 268 data_time: 0.92s time: 119.32s eta: 4:14:32 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 268 data_time: 0.80s time: 119.05s eta: 4:11:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.81s time: 119.12s eta: 4:10:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.184 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 268 data_time: 0.78s time: 119.14s eta: 4:08:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30983.0 tgs: 260 data_time: 0.59s time: 118.96s eta: 4:05:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32059.0 tgs: 269 data_time: 0.51s time: 119.16s eta: 4:04:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 270 data_time: 0.94s time: 119.50s eta: 4:02:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31027.0 tgs: 260 data_time: 0.90s time: 119.25s eta: 4:00:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 270 data_time: 0.71s time: 119.19s eta: 3:58:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31383.0 tgs: 262 data_time: 0.86s time: 119.36s eta: 3:56:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31587.0 tgs: 264 data_time: 0.68s time: 119.38s eta: 3:54:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32047.0 tgs: 267 data_time: 0.86s time: 119.96s eta: 3:53:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32356.0 tgs: 271 data_time: 0.52s time: 119.32s eta: 3:50:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32092.0 tgs: 268 data_time: 0.62s time: 119.64s eta: 3:49:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29588.0 tgs: 248 data_time: 0.78s time: 118.86s eta: 3:45:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31665.0 tgs: 264 data_time: 0.73s time: 119.71s eta: 3:45:27 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31442.0 tgs: 263 data_time: 0.63s time: 119.33s eta: 3:42:44 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31891.0 tgs: 267 data_time: 0.70s time: 119.26s eta: 3:40:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31225.0 tgs: 262 data_time: 0.86s time: 119.00s eta: 3:38:10 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31289.0 tgs: 261 data_time: 0.78s time: 119.43s eta: 3:36:58 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32268.0 tgs: 271 data_time: 0.87s time: 119.03s eta: 3:34:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31078.0 tgs: 261 data_time: 0.69s time: 118.99s eta: 3:32:12 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 266 data_time: 0.85s time: 119.22s eta: 3:30:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32406.0 tgs: 269 data_time: 1.06s time: 120.11s eta: 3:30:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 259 data_time: 0.81s time: 119.13s eta: 3:26:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31928.0 tgs: 267 data_time: 0.55s time: 119.32s eta: 3:24:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32100.0 tgs: 269 data_time: 0.76s time: 119.08s eta: 3:22:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 0.89s time: 119.43s eta: 3:21:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.104 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31117.0 tgs: 260 data_time: 0.88s time: 119.29s eta: 3:18:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32141.0 tgs: 269 data_time: 1.16s time: 119.22s eta: 3:16:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.195 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32069.0 tgs: 269 data_time: 0.82s time: 119.13s eta: 3:14:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32154.0 tgs: 270 data_time: 0.77s time: 119.02s eta: 3:12:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31283.0 tgs: 262 data_time: 0.53s time: 119.18s eta: 3:10:41 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32362.0 tgs: 267 data_time: 0.89s time: 121.06s eta: 3:11:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.103 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32090.0 tgs: 268 data_time: 0.48s time: 119.51s eta: 3:07:14 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31709.0 tgs: 265 data_time: 0.73s time: 119.28s eta: 3:04:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.185 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 269 data_time: 0.81s time: 118.97s eta: 3:02:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31728.0 tgs: 266 data_time: 0.99s time: 119.22s eta: 3:00:48 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30756.0 tgs: 257 data_time: 0.73s time: 119.46s eta: 2:59:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 271 data_time: 1.01s time: 118.98s eta: 2:56:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31144.0 tgs: 260 data_time: 0.58s time: 119.35s eta: 2:55:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31764.0 tgs: 265 data_time: 1.01s time: 119.42s eta: 2:53:09 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 267 data_time: 0.67s time: 119.35s eta: 2:51:04 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 259 data_time: 0.66s time: 121.46s eta: 2:52:03 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31432.0 tgs: 263 data_time: 0.88s time: 119.24s eta: 2:46:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31430.0 tgs: 262 data_time: 0.78s time: 119.56s eta: 2:45:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.212 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31515.0 tgs: 265 data_time: 0.79s time: 118.91s eta: 2:42:30 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31398.0 tgs: 263 data_time: 0.65s time: 119.18s eta: 2:40:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30824.0 tgs: 259 data_time: 0.73s time: 118.97s eta: 2:38:37 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.224 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32254.0 tgs: 270 data_time: 0.99s time: 119.23s eta: 2:36:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30158.0 tgs: 253 data_time: 0.84s time: 119.11s eta: 2:34:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30668.0 tgs: 256 data_time: 0.60s time: 119.43s eta: 2:33:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31705.0 tgs: 263 data_time: 0.87s time: 120.23s eta: 2:32:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31871.0 tgs: 267 data_time: 0.86s time: 119.29s eta: 2:29:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31978.0 tgs: 269 data_time: 0.76s time: 118.88s eta: 2:26:36 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31953.0 tgs: 267 data_time: 0.90s time: 119.52s eta: 2:25:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32051.0 tgs: 269 data_time: 0.96s time: 119.07s eta: 2:22:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31727.0 tgs: 266 data_time: 0.76s time: 119.14s eta: 2:20:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31702.0 tgs: 266 data_time: 0.91s time: 119.05s eta: 2:18:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.193 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 268 data_time: 0.87s time: 119.84s eta: 2:17:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31770.0 tgs: 266 data_time: 0.48s time: 119.32s eta: 2:15:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31812.0 tgs: 266 data_time: 0.88s time: 119.36s eta: 2:13:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.235 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31915.0 tgs: 266 data_time: 0.80s time: 119.68s eta: 2:11:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31860.0 tgs: 266 data_time: 0.69s time: 119.49s eta: 2:09:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 269 data_time: 0.70s time: 118.96s eta: 2:06:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31729.0 tgs: 265 data_time: 0.63s time: 119.37s eta: 2:05:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30980.0 tgs: 259 data_time: 0.56s time: 119.27s eta: 2:03:15 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 266 data_time: 0.79s time: 119.23s eta: 2:01:13 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31385.0 tgs: 262 data_time: 0.76s time: 119.50s eta: 1:59:30 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31827.0 tgs: 266 data_time: 0.47s time: 119.42s eta: 1:57:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.89s time: 119.54s eta: 1:55:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30740.0 tgs: 257 data_time: 1.00s time: 119.30s eta: 1:53:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31178.0 tgs: 261 data_time: 0.80s time: 119.26s eta: 1:51:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 271 data_time: 0.79s time: 118.93s eta: 1:49:00 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.206 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31180.0 tgs: 260 data_time: 0.72s time: 119.64s eta: 1:47:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 265 data_time: 0.78s time: 119.72s eta: 1:45:45 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31377.0 tgs: 262 data_time: 0.47s time: 119.35s eta: 1:43:26 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32027.0 tgs: 268 data_time: 0.56s time: 119.31s eta: 1:41:24 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31859.0 tgs: 267 data_time: 0.90s time: 119.28s eta: 1:39:23 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 269 data_time: 0.78s time: 118.91s eta: 1:37:06 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32443.0 tgs: 272 data_time: 0.85s time: 119.16s eta: 1:35:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32375.0 tgs: 271 data_time: 0.57s time: 119.07s eta: 1:33:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31848.0 tgs: 267 data_time: 0.68s time: 119.25s eta: 1:31:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32305.0 tgs: 269 data_time: 0.67s time: 119.73s eta: 1:29:47 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31778.0 tgs: 267 data_time: 0.80s time: 119.01s eta: 1:27:16 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31269.0 tgs: 263 data_time: 0.70s time: 118.87s eta: 1:25:11 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.265 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 268 data_time: 0.58s time: 119.48s eta: 1:23:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 269 data_time: 0.73s time: 119.49s eta: 1:21:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32248.0 tgs: 271 data_time: 0.82s time: 118.94s eta: 1:19:17 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31102.0 tgs: 261 data_time: 0.67s time: 118.93s eta: 1:17:18 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32215.0 tgs: 263 data_time: 0.73s time: 122.13s eta: 1:17:20 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31693.0 tgs: 265 data_time: 0.88s time: 119.32s eta: 1:13:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30460.0 tgs: 255 data_time: 0.81s time: 119.06s eta: 1:11:25 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31907.0 tgs: 268 data_time: 0.70s time: 118.92s eta: 1:09:22 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31334.0 tgs: 263 data_time: 0.80s time: 119.11s eta: 1:07:29 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.89s time: 119.42s eta: 1:05:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 270 data_time: 0.98s time: 119.35s eta: 1:03:39 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31141.0 tgs: 261 data_time: 0.72s time: 119.18s eta: 1:01:34 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32132.0 tgs: 269 data_time: 0.90s time: 119.27s eta: 0:59:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 268 data_time: 1.23s time: 119.45s eta: 0:57:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31761.0 tgs: 266 data_time: 1.06s time: 119.40s eta: 0:55:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.66s time: 119.49s eta: 0:53:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31781.0 tgs: 266 data_time: 0.67s time: 119.37s eta: 0:51:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31804.0 tgs: 266 data_time: 0.90s time: 119.47s eta: 0:49:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.219 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 267 data_time: 0.72s time: 118.90s eta: 0:47:33 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31769.0 tgs: 266 data_time: 0.79s time: 119.28s eta: 0:45:43 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.99s time: 119.20s eta: 0:43:42 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 265 data_time: 0.90s time: 119.00s eta: 0:41:38 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.224 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31905.0 tgs: 268 data_time: 0.95s time: 119.04s eta: 0:39:40 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 268 data_time: 0.73s time: 119.67s eta: 0:37:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31864.0 tgs: 267 data_time: 0.55s time: 119.27s eta: 0:35:46 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32094.0 tgs: 268 data_time: 0.70s time: 119.75s eta: 0:33:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32460.0 tgs: 267 data_time: 0.91s time: 121.23s eta: 0:32:19 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 267 data_time: 0.73s time: 119.38s eta: 0:29:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32063.0 tgs: 268 data_time: 0.60s time: 119.32s eta: 0:27:50 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31458.0 tgs: 262 data_time: 0.72s time: 119.64s eta: 0:25:55 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31717.0 tgs: 266 data_time: 0.73s time: 119.09s eta: 0:23:49 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31218.0 tgs: 261 data_time: 0.85s time: 119.39s eta: 0:21:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32262.0 tgs: 270 data_time: 0.85s time: 119.26s eta: 0:19:52 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31793.0 tgs: 266 data_time: 0.61s time: 119.25s eta: 0:17:53 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 262 data_time: 0.78s time: 119.90s eta: 0:15:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31603.0 tgs: 265 data_time: 0.87s time: 119.18s eta: 0:13:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32105.0 tgs: 269 data_time: 0.94s time: 119.09s eta: 0:11:54 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 266 data_time: 0.84s time: 119.32s eta: 0:09:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31820.0 tgs: 267 data_time: 0.60s time: 119.17s eta: 0:07:56 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31856.0 tgs: 266 data_time: 1.11s time: 119.69s eta: 0:05:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32058.0 tgs: 263 data_time: 0.73s time: 121.43s eta: 0:04:02 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31388.0 tgs: 263 data_time: 0.78s time: 119.21s eta: 0:01:59 +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.29 seconds, peak gpu memory 17.4G +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 23][DP 5][SP 3][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank24.log b/20250121165312/rank24.log new file mode 100644 index 0000000000000000000000000000000000000000..cec29ba384ea498078ad0d345cfa4badd57a2104 --- /dev/null +++ b/20250121165312/rank24.log @@ -0,0 +1,851 @@ +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.05s +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.64 seconds, peak gpu memory 13.4G +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.270 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31349.0 tgs: 204 data_time: 1.83s time: 153.58s eta: 1 day, 1:17:51 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.237 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32045.0 tgs: 267 data_time: 0.59s time: 119.58s eta: 19:39:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.244 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31775.0 tgs: 265 data_time: 0.88s time: 119.68s eta: 19:38:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.235 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 262 data_time: 0.81s time: 119.66s eta: 19:36:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.255 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31310.0 tgs: 262 data_time: 0.72s time: 119.30s eta: 19:31:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.260 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31970.0 tgs: 267 data_time: 0.85s time: 119.54s eta: 19:31:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.210 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 260 data_time: 0.85s time: 122.96s eta: 20:02:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.229 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32280.0 tgs: 269 data_time: 0.65s time: 119.89s eta: 19:30:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.240 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32093.0 tgs: 265 data_time: 0.98s time: 120.65s eta: 19:36:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32007.0 tgs: 266 data_time: 0.58s time: 119.91s eta: 19:27:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32301.0 tgs: 268 data_time: 0.74s time: 120.46s eta: 19:30:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30685.0 tgs: 257 data_time: 0.80s time: 119.34s eta: 19:17:36 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30776.0 tgs: 257 data_time: 0.90s time: 119.69s eta: 19:18:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31850.0 tgs: 267 data_time: 0.71s time: 119.17s eta: 19:11:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32016.0 tgs: 268 data_time: 0.79s time: 119.21s eta: 19:10:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31587.0 tgs: 265 data_time: 1.00s time: 118.93s eta: 19:05:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30793.0 tgs: 258 data_time: 0.75s time: 119.14s eta: 19:05:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31008.0 tgs: 260 data_time: 0.93s time: 119.18s eta: 19:04:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32462.0 tgs: 271 data_time: 0.89s time: 119.40s eta: 19:04:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32193.0 tgs: 269 data_time: 0.66s time: 119.33s eta: 19:01:32 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31420.0 tgs: 261 data_time: 0.89s time: 120.32s eta: 19:09:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31579.0 tgs: 256 data_time: 1.00s time: 122.91s eta: 19:31:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.213 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 266 data_time: 0.82s time: 119.48s eta: 18:57:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.120 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30152.0 tgs: 253 data_time: 0.81s time: 119.17s eta: 18:52:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31164.0 tgs: 261 data_time: 0.89s time: 119.23s eta: 18:50:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 265 data_time: 0.91s time: 119.39s eta: 18:50:13 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 268 data_time: 1.00s time: 119.67s eta: 18:50:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30357.0 tgs: 254 data_time: 0.78s time: 119.37s eta: 18:46:03 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 267 data_time: 0.88s time: 119.79s eta: 18:48:00 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31963.0 tgs: 247 data_time: 0.86s time: 129.33s eta: 20:15:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31960.0 tgs: 267 data_time: 0.78s time: 119.57s eta: 18:41:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31654.0 tgs: 265 data_time: 0.80s time: 119.39s eta: 18:38:14 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31502.0 tgs: 261 data_time: 0.64s time: 120.66s eta: 18:48:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.85s time: 119.40s eta: 18:34:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.244 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31221.0 tgs: 262 data_time: 0.71s time: 119.16s eta: 18:30:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.94s time: 119.08s eta: 18:27:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31855.0 tgs: 267 data_time: 1.14s time: 119.13s eta: 18:25:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30932.0 tgs: 256 data_time: 0.73s time: 120.39s eta: 18:35:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 263 data_time: 0.61s time: 121.17s eta: 18:40:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 262 data_time: 0.61s time: 120.66s eta: 18:34:07 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32069.0 tgs: 267 data_time: 0.72s time: 119.74s eta: 18:23:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31550.0 tgs: 254 data_time: 0.60s time: 123.88s eta: 18:59:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30417.0 tgs: 254 data_time: 0.53s time: 119.56s eta: 18:17:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31874.0 tgs: 267 data_time: 0.84s time: 119.35s eta: 18:14:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31865.0 tgs: 267 data_time: 0.60s time: 119.23s eta: 18:10:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32131.0 tgs: 268 data_time: 0.94s time: 119.59s eta: 18:12:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32443.0 tgs: 272 data_time: 0.65s time: 119.27s eta: 18:07:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31045.0 tgs: 260 data_time: 0.73s time: 119.01s eta: 18:02:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 267 data_time: 0.74s time: 119.10s eta: 18:01:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31945.0 tgs: 267 data_time: 0.96s time: 119.27s eta: 18:01:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32011.0 tgs: 268 data_time: 1.07s time: 119.33s eta: 17:59:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32547.0 tgs: 271 data_time: 0.98s time: 120.03s eta: 18:04:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32049.0 tgs: 269 data_time: 0.85s time: 119.09s eta: 17:53:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32269.0 tgs: 270 data_time: 0.85s time: 119.42s eta: 17:54:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.261 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30865.0 tgs: 259 data_time: 1.07s time: 119.01s eta: 17:49:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31114.0 tgs: 261 data_time: 1.10s time: 119.02s eta: 17:47:10 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32034.0 tgs: 255 data_time: 0.83s time: 125.35s eta: 18:41:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.232 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 262 data_time: 1.38s time: 120.11s eta: 17:52:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30677.0 tgs: 256 data_time: 0.74s time: 119.40s eta: 17:44:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31925.0 tgs: 267 data_time: 1.00s time: 119.34s eta: 17:42:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32154.0 tgs: 269 data_time: 0.63s time: 119.50s eta: 17:41:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 254 data_time: 0.83s time: 126.05s eta: 18:37:36 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31246.0 tgs: 254 data_time: 0.90s time: 122.63s eta: 18:05:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31119.0 tgs: 256 data_time: 0.96s time: 121.16s eta: 17:50:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31196.0 tgs: 262 data_time: 0.59s time: 119.03s eta: 17:29:29 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 268 data_time: 0.68s time: 118.85s eta: 17:25:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31639.0 tgs: 255 data_time: 0.91s time: 124.00s eta: 18:09:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31322.0 tgs: 251 data_time: 1.07s time: 124.57s eta: 18:12:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31116.0 tgs: 260 data_time: 0.89s time: 119.40s eta: 17:24:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 268 data_time: 0.70s time: 119.69s eta: 17:25:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31426.0 tgs: 262 data_time: 0.77s time: 119.52s eta: 17:21:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31953.0 tgs: 268 data_time: 0.77s time: 118.86s eta: 17:14:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31272.0 tgs: 260 data_time: 0.74s time: 119.86s eta: 17:20:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31828.0 tgs: 266 data_time: 0.96s time: 119.55s eta: 17:16:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30988.0 tgs: 254 data_time: 0.82s time: 121.87s eta: 17:34:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31784.0 tgs: 257 data_time: 0.91s time: 123.51s eta: 17:46:19 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31384.0 tgs: 263 data_time: 0.70s time: 119.01s eta: 17:05:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31079.0 tgs: 260 data_time: 0.71s time: 119.23s eta: 17:05:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31623.0 tgs: 265 data_time: 0.86s time: 119.14s eta: 17:02:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 268 data_time: 0.72s time: 119.26s eta: 17:01:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31751.0 tgs: 266 data_time: 1.05s time: 119.04s eta: 16:57:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.259 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30915.0 tgs: 259 data_time: 0.77s time: 119.27s eta: 16:57:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31739.0 tgs: 266 data_time: 0.83s time: 119.15s eta: 16:54:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31726.0 tgs: 265 data_time: 0.59s time: 119.58s eta: 16:56:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.128 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 26330.0 tgs: 221 data_time: 0.68s time: 119.13s eta: 16:50:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 269 data_time: 0.73s time: 118.85s eta: 16:46:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31115.0 tgs: 261 data_time: 0.75s time: 119.00s eta: 16:45:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.255 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 268 data_time: 0.73s time: 119.85s eta: 16:50:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30669.0 tgs: 256 data_time: 0.81s time: 119.39s eta: 16:44:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 266 data_time: 0.71s time: 119.00s eta: 16:39:36 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32030.0 tgs: 264 data_time: 0.73s time: 120.92s eta: 16:53:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.112 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 265 data_time: 0.78s time: 118.99s eta: 16:35:33 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32088.0 tgs: 269 data_time: 0.80s time: 119.29s eta: 16:36:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.218 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32331.0 tgs: 271 data_time: 0.57s time: 119.23s eta: 16:33:33 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32221.0 tgs: 269 data_time: 0.92s time: 119.37s eta: 16:32:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.210 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30610.0 tgs: 255 data_time: 0.70s time: 119.70s eta: 16:33:30 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31693.0 tgs: 265 data_time: 0.81s time: 119.20s eta: 16:27:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31074.0 tgs: 259 data_time: 0.78s time: 119.76s eta: 16:30:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31521.0 tgs: 262 data_time: 0.71s time: 120.11s eta: 16:30:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30895.0 tgs: 259 data_time: 0.97s time: 118.98s eta: 16:19:36 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32094.0 tgs: 252 data_time: 0.73s time: 127.32s eta: 17:26:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.125 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31475.0 tgs: 264 data_time: 0.74s time: 119.22s eta: 16:17:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32583.0 tgs: 273 data_time: 0.66s time: 119.14s eta: 16:14:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32561.0 tgs: 273 data_time: 0.72s time: 119.25s eta: 16:13:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.106 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31165.0 tgs: 261 data_time: 0.64s time: 119.19s eta: 16:11:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31581.0 tgs: 265 data_time: 0.70s time: 118.95s eta: 16:07:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31719.0 tgs: 266 data_time: 0.81s time: 118.86s eta: 16:04:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31776.0 tgs: 267 data_time: 0.76s time: 118.96s eta: 16:03:35 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31490.0 tgs: 258 data_time: 0.85s time: 121.84s eta: 16:24:51 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31856.0 tgs: 267 data_time: 0.94s time: 118.99s eta: 15:59:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29829.0 tgs: 250 data_time: 0.65s time: 119.23s eta: 15:59:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 268 data_time: 0.60s time: 119.61s eta: 16:00:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32587.0 tgs: 273 data_time: 0.65s time: 119.04s eta: 15:54:18 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.117 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29377.0 tgs: 246 data_time: 0.62s time: 118.98s eta: 15:51:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31543.0 tgs: 264 data_time: 0.55s time: 119.17s eta: 15:51:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 266 data_time: 0.63s time: 119.63s eta: 15:53:03 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31486.0 tgs: 264 data_time: 0.65s time: 118.93s eta: 15:45:31 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30548.0 tgs: 256 data_time: 0.71s time: 119.29s eta: 15:46:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 264 data_time: 0.80s time: 119.12s eta: 15:43:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30535.0 tgs: 256 data_time: 0.50s time: 119.00s eta: 15:40:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31961.0 tgs: 267 data_time: 0.81s time: 119.66s eta: 15:43:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31872.0 tgs: 262 data_time: 1.04s time: 121.50s eta: 15:55:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31557.0 tgs: 264 data_time: 0.79s time: 119.41s eta: 15:37:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.117 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30431.0 tgs: 234 data_time: 0.76s time: 129.96s eta: 16:58:00 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31153.0 tgs: 260 data_time: 0.81s time: 119.46s eta: 15:33:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31444.0 tgs: 262 data_time: 0.80s time: 119.61s eta: 15:32:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31241.0 tgs: 261 data_time: 0.84s time: 119.52s eta: 15:30:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.308 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31598.0 tgs: 258 data_time: 0.82s time: 122.14s eta: 15:48:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.258 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30829.0 tgs: 255 data_time: 0.84s time: 120.75s eta: 15:35:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.206 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31380.0 tgs: 262 data_time: 0.77s time: 119.41s eta: 15:23:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 263 data_time: 1.04s time: 119.45s eta: 15:21:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31903.0 tgs: 242 data_time: 0.77s time: 131.63s eta: 16:53:31 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 266 data_time: 0.73s time: 119.29s eta: 15:16:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.78s time: 119.57s eta: 15:16:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.225 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31744.0 tgs: 259 data_time: 0.57s time: 122.26s eta: 15:35:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32240.0 tgs: 269 data_time: 0.69s time: 119.71s eta: 15:13:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31278.0 tgs: 262 data_time: 0.72s time: 119.35s eta: 15:09:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 266 data_time: 0.67s time: 118.90s eta: 15:03:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31528.0 tgs: 264 data_time: 0.80s time: 119.00s eta: 15:02:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.114 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30920.0 tgs: 259 data_time: 0.60s time: 119.31s eta: 15:02:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32178.0 tgs: 270 data_time: 0.62s time: 118.91s eta: 14:57:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.106 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30764.0 tgs: 255 data_time: 0.80s time: 120.52s eta: 15:07:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.240 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30675.0 tgs: 257 data_time: 0.60s time: 119.31s eta: 14:56:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.248 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 262 data_time: 0.75s time: 122.66s eta: 15:19:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31640.0 tgs: 265 data_time: 0.89s time: 119.12s eta: 14:51:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31844.0 tgs: 265 data_time: 1.04s time: 119.77s eta: 14:54:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31220.0 tgs: 261 data_time: 0.68s time: 119.45s eta: 14:49:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.232 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31075.0 tgs: 261 data_time: 1.00s time: 118.99s eta: 14:44:29 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 264 data_time: 0.91s time: 119.34s eta: 14:45:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31692.0 tgs: 266 data_time: 0.68s time: 118.93s eta: 14:40:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 270 data_time: 1.08s time: 119.05s eta: 14:38:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32575.0 tgs: 273 data_time: 0.76s time: 119.06s eta: 14:37:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31516.0 tgs: 264 data_time: 0.85s time: 119.13s eta: 14:35:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31173.0 tgs: 261 data_time: 0.72s time: 119.00s eta: 14:32:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 28860.0 tgs: 242 data_time: 0.77s time: 119.08s eta: 14:31:14 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32241.0 tgs: 270 data_time: 0.64s time: 119.15s eta: 14:29:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30260.0 tgs: 252 data_time: 0.73s time: 119.65s eta: 14:31:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.206 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31648.0 tgs: 265 data_time: 0.77s time: 119.18s eta: 14:26:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32108.0 tgs: 269 data_time: 0.87s time: 119.19s eta: 14:24:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31820.0 tgs: 266 data_time: 0.60s time: 119.39s eta: 14:23:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.196 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32105.0 tgs: 265 data_time: 0.71s time: 120.74s eta: 14:31:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.220 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30550.0 tgs: 257 data_time: 0.87s time: 118.85s eta: 14:15:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.250 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 270 data_time: 1.01s time: 119.11s eta: 14:15:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31196.0 tgs: 261 data_time: 0.70s time: 119.15s eta: 14:13:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.218 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31096.0 tgs: 259 data_time: 0.79s time: 119.96s eta: 14:17:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31223.0 tgs: 260 data_time: 0.94s time: 119.88s eta: 14:15:10 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 268 data_time: 0.73s time: 119.57s eta: 14:10:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31437.0 tgs: 263 data_time: 0.66s time: 119.22s eta: 14:06:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.122 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31132.0 tgs: 261 data_time: 0.59s time: 119.26s eta: 14:04:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31410.0 tgs: 263 data_time: 0.64s time: 119.27s eta: 14:02:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31532.0 tgs: 264 data_time: 0.66s time: 119.14s eta: 13:59:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32424.0 tgs: 272 data_time: 0.77s time: 119.12s eta: 13:57:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30282.0 tgs: 254 data_time: 0.67s time: 119.17s eta: 13:56:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31583.0 tgs: 264 data_time: 0.66s time: 119.35s eta: 13:55:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32066.0 tgs: 266 data_time: 0.82s time: 120.11s eta: 13:58:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31409.0 tgs: 256 data_time: 0.62s time: 122.55s eta: 14:13:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.268 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.73s time: 119.11s eta: 13:47:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 267 data_time: 0.47s time: 119.36s eta: 13:47:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.116 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32102.0 tgs: 268 data_time: 0.70s time: 119.36s eta: 13:45:32 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32140.0 tgs: 269 data_time: 0.86s time: 119.35s eta: 13:43:30 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31889.0 tgs: 267 data_time: 0.90s time: 119.11s eta: 13:39:51 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31018.0 tgs: 260 data_time: 0.68s time: 119.25s eta: 13:38:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.089 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31186.0 tgs: 261 data_time: 0.59s time: 119.13s eta: 13:36:03 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31655.0 tgs: 265 data_time: 0.83s time: 119.24s eta: 13:34:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31207.0 tgs: 261 data_time: 0.88s time: 119.44s eta: 13:34:11 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31178.0 tgs: 261 data_time: 0.75s time: 119.08s eta: 13:29:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.184 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31976.0 tgs: 268 data_time: 0.48s time: 118.92s eta: 13:26:41 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31843.0 tgs: 265 data_time: 1.05s time: 119.81s eta: 13:30:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31274.0 tgs: 262 data_time: 0.79s time: 119.25s eta: 13:24:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31670.0 tgs: 265 data_time: 0.88s time: 119.16s eta: 13:22:19 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 267 data_time: 0.68s time: 119.36s eta: 13:21:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.76s time: 119.25s eta: 13:18:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31542.0 tgs: 264 data_time: 0.70s time: 119.04s eta: 13:15:33 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.96s time: 119.37s eta: 13:15:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.225 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31195.0 tgs: 261 data_time: 0.64s time: 119.18s eta: 13:12:32 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.194 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32118.0 tgs: 268 data_time: 0.82s time: 119.45s eta: 13:12:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31019.0 tgs: 260 data_time: 0.89s time: 119.29s eta: 13:09:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32455.0 tgs: 272 data_time: 0.85s time: 119.15s eta: 13:06:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.128 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31953.0 tgs: 265 data_time: 0.68s time: 120.14s eta: 13:10:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31412.0 tgs: 261 data_time: 0.76s time: 120.16s eta: 13:09:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31853.0 tgs: 266 data_time: 0.78s time: 119.50s eta: 13:02:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 265 data_time: 0.86s time: 119.50s eta: 13:00:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.186 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31252.0 tgs: 261 data_time: 0.84s time: 119.58s eta: 12:59:14 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.236 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 267 data_time: 0.86s time: 119.79s eta: 12:58:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30904.0 tgs: 259 data_time: 0.77s time: 119.04s eta: 12:51:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32316.0 tgs: 271 data_time: 0.70s time: 119.05s eta: 12:49:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31416.0 tgs: 263 data_time: 0.80s time: 119.12s eta: 12:48:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31211.0 tgs: 261 data_time: 0.68s time: 119.40s eta: 12:48:07 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31678.0 tgs: 265 data_time: 1.05s time: 119.19s eta: 12:44:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.110 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31625.0 tgs: 264 data_time: 0.62s time: 119.44s eta: 12:44:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32437.0 tgs: 272 data_time: 0.67s time: 119.09s eta: 12:40:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 266 data_time: 0.84s time: 119.07s eta: 12:38:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31576.0 tgs: 264 data_time: 0.84s time: 119.59s eta: 12:39:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31649.0 tgs: 266 data_time: 0.71s time: 118.79s eta: 12:32:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.208 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31920.0 tgs: 268 data_time: 0.88s time: 118.89s eta: 12:30:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31644.0 tgs: 265 data_time: 0.98s time: 119.17s eta: 12:30:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31897.0 tgs: 267 data_time: 0.67s time: 119.26s eta: 12:29:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.128 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30222.0 tgs: 253 data_time: 0.88s time: 119.34s eta: 12:27:51 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32146.0 tgs: 269 data_time: 0.63s time: 119.11s eta: 12:24:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31620.0 tgs: 264 data_time: 0.91s time: 119.35s eta: 12:23:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31262.0 tgs: 262 data_time: 1.00s time: 119.25s eta: 12:21:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32133.0 tgs: 269 data_time: 0.64s time: 119.03s eta: 12:17:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31407.0 tgs: 259 data_time: 0.63s time: 121.13s eta: 12:28:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31926.0 tgs: 266 data_time: 0.82s time: 119.99s eta: 12:19:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30827.0 tgs: 259 data_time: 0.88s time: 118.72s eta: 12:10:07 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.189 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31566.0 tgs: 263 data_time: 0.66s time: 119.65s eta: 12:13:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.220 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31944.0 tgs: 266 data_time: 0.87s time: 119.82s eta: 12:12:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.168 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.64s time: 119.13s eta: 12:06:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.218 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32254.0 tgs: 270 data_time: 1.00s time: 119.17s eta: 12:04:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31838.0 tgs: 267 data_time: 0.72s time: 119.11s eta: 12:02:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.129 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31452.0 tgs: 263 data_time: 0.74s time: 119.19s eta: 12:01:07 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30922.0 tgs: 259 data_time: 0.76s time: 118.95s eta: 11:57:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.205 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31065.0 tgs: 259 data_time: 0.59s time: 119.71s eta: 12:00:14 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32267.0 tgs: 270 data_time: 0.74s time: 119.27s eta: 11:55:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31351.0 tgs: 262 data_time: 0.50s time: 119.27s eta: 11:53:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31937.0 tgs: 266 data_time: 0.75s time: 119.71s eta: 11:54:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.115 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31108.0 tgs: 261 data_time: 0.63s time: 118.79s eta: 11:46:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32397.0 tgs: 271 data_time: 0.76s time: 119.15s eta: 11:46:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31143.0 tgs: 260 data_time: 0.90s time: 119.61s eta: 11:47:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.185 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31596.0 tgs: 265 data_time: 0.63s time: 118.88s eta: 11:41:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31494.0 tgs: 265 data_time: 0.55s time: 118.68s eta: 11:38:13 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31011.0 tgs: 260 data_time: 0.77s time: 119.22s eta: 11:39:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.127 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29561.0 tgs: 248 data_time: 0.72s time: 119.13s eta: 11:36:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31494.0 tgs: 264 data_time: 0.62s time: 119.07s eta: 11:34:35 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31656.0 tgs: 266 data_time: 0.90s time: 118.95s eta: 11:31:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.130 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29638.0 tgs: 249 data_time: 0.71s time: 118.91s eta: 11:29:41 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31840.0 tgs: 267 data_time: 0.74s time: 118.97s eta: 11:28:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.125 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31844.0 tgs: 266 data_time: 0.76s time: 119.28s eta: 11:27:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30935.0 tgs: 259 data_time: 0.76s time: 119.36s eta: 11:26:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.118 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31121.0 tgs: 261 data_time: 0.95s time: 118.87s eta: 11:21:32 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.207 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32180.0 tgs: 269 data_time: 0.86s time: 119.21s eta: 11:21:29 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 268 data_time: 0.64s time: 119.71s eta: 11:22:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31060.0 tgs: 260 data_time: 0.64s time: 119.11s eta: 11:16:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.221 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32327.0 tgs: 271 data_time: 0.66s time: 119.12s eta: 11:15:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 271 data_time: 0.54s time: 118.99s eta: 11:12:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30725.0 tgs: 257 data_time: 0.70s time: 119.09s eta: 11:10:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32418.0 tgs: 272 data_time: 0.59s time: 119.16s eta: 11:09:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32078.0 tgs: 269 data_time: 0.77s time: 118.96s eta: 11:06:11 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32268.0 tgs: 270 data_time: 0.69s time: 119.46s eta: 11:07:00 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.106 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 28614.0 tgs: 239 data_time: 0.82s time: 119.64s eta: 11:06:00 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30863.0 tgs: 258 data_time: 0.57s time: 119.30s eta: 11:02:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31397.0 tgs: 263 data_time: 0.77s time: 119.32s eta: 11:00:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32034.0 tgs: 269 data_time: 0.79s time: 118.83s eta: 10:55:33 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32023.0 tgs: 268 data_time: 0.74s time: 119.46s eta: 10:57:00 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31816.0 tgs: 266 data_time: 0.63s time: 119.37s eta: 10:54:31 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30798.0 tgs: 258 data_time: 0.88s time: 119.27s eta: 10:52:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31745.0 tgs: 266 data_time: 0.78s time: 119.26s eta: 10:49:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.74s time: 119.09s eta: 10:47:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.186 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31606.0 tgs: 264 data_time: 0.84s time: 119.53s eta: 10:47:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31943.0 tgs: 267 data_time: 0.63s time: 119.33s eta: 10:44:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31700.0 tgs: 265 data_time: 0.59s time: 119.49s eta: 10:43:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31264.0 tgs: 261 data_time: 0.67s time: 119.51s eta: 10:41:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31586.0 tgs: 265 data_time: 0.80s time: 118.99s eta: 10:36:35 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.107 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 262 data_time: 0.61s time: 119.22s eta: 10:35:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31391.0 tgs: 263 data_time: 0.66s time: 119.06s eta: 10:33:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31555.0 tgs: 265 data_time: 0.83s time: 118.78s eta: 10:29:32 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.190 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 271 data_time: 0.81s time: 119.85s eta: 10:33:13 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30132.0 tgs: 252 data_time: 0.59s time: 119.39s eta: 10:28:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.125 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31627.0 tgs: 266 data_time: 0.45s time: 118.83s eta: 10:23:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.115 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31233.0 tgs: 262 data_time: 0.66s time: 118.98s eta: 10:22:41 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.124 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31780.0 tgs: 266 data_time: 0.73s time: 119.30s eta: 10:22:19 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31590.0 tgs: 265 data_time: 0.71s time: 119.20s eta: 10:19:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.123 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31757.0 tgs: 266 data_time: 0.48s time: 119.06s eta: 10:17:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 264 data_time: 0.92s time: 119.49s eta: 10:17:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.205 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31016.0 tgs: 259 data_time: 1.02s time: 119.40s eta: 10:14:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31543.0 tgs: 264 data_time: 0.63s time: 119.44s eta: 10:13:07 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30855.0 tgs: 258 data_time: 0.89s time: 119.29s eta: 10:10:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.121 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31401.0 tgs: 263 data_time: 0.79s time: 119.11s eta: 10:07:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31887.0 tgs: 266 data_time: 0.68s time: 119.70s eta: 10:08:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.200 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 269 data_time: 0.87s time: 119.83s eta: 10:07:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.260 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31143.0 tgs: 256 data_time: 0.72s time: 121.56s eta: 10:13:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31651.0 tgs: 265 data_time: 0.70s time: 119.25s eta: 10:00:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.196 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31551.0 tgs: 265 data_time: 0.72s time: 118.80s eta: 9:55:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.273 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31206.0 tgs: 261 data_time: 0.66s time: 119.56s eta: 9:57:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31674.0 tgs: 264 data_time: 0.79s time: 119.70s eta: 9:56:30 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31191.0 tgs: 261 data_time: 0.82s time: 119.25s eta: 9:52:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31542.0 tgs: 264 data_time: 0.82s time: 119.33s eta: 9:50:41 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.197 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32584.0 tgs: 273 data_time: 0.83s time: 119.22s eta: 9:48:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 266 data_time: 0.84s time: 120.78s eta: 9:53:51 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32050.0 tgs: 268 data_time: 0.65s time: 119.27s eta: 9:44:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30911.0 tgs: 258 data_time: 0.57s time: 119.46s eta: 9:43:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.195 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31939.0 tgs: 267 data_time: 0.78s time: 119.54s eta: 9:41:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 270 data_time: 0.66s time: 119.20s eta: 9:38:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31928.0 tgs: 267 data_time: 0.62s time: 119.48s eta: 9:37:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.223 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30782.0 tgs: 258 data_time: 0.78s time: 119.07s eta: 9:33:32 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31259.0 tgs: 262 data_time: 0.67s time: 119.22s eta: 9:32:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.67s time: 119.16s eta: 9:29:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.64s time: 119.22s eta: 9:28:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.73s time: 119.46s eta: 9:27:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31337.0 tgs: 263 data_time: 0.86s time: 118.99s eta: 9:23:13 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.112 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31996.0 tgs: 268 data_time: 0.65s time: 119.03s eta: 9:21:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.228 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 263 data_time: 0.72s time: 119.53s eta: 9:21:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.231 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31259.0 tgs: 261 data_time: 0.66s time: 119.59s eta: 9:20:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.078 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30922.0 tgs: 259 data_time: 0.55s time: 119.12s eta: 9:15:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31663.0 tgs: 265 data_time: 0.74s time: 119.05s eta: 9:13:35 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32492.0 tgs: 273 data_time: 0.83s time: 118.88s eta: 9:10:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30956.0 tgs: 259 data_time: 0.88s time: 119.32s eta: 9:10:51 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32383.0 tgs: 271 data_time: 0.89s time: 119.09s eta: 9:07:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 270 data_time: 0.72s time: 118.85s eta: 9:04:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.209 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31568.0 tgs: 263 data_time: 0.78s time: 119.67s eta: 9:06:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.55s time: 119.43s eta: 9:03:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31643.0 tgs: 264 data_time: 0.65s time: 119.44s eta: 9:01:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31603.0 tgs: 266 data_time: 0.90s time: 118.74s eta: 8:56:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32388.0 tgs: 272 data_time: 1.02s time: 118.86s eta: 8:54:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.184 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31589.0 tgs: 264 data_time: 0.92s time: 119.34s eta: 8:55:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 269 data_time: 0.57s time: 119.14s eta: 8:52:10 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31053.0 tgs: 260 data_time: 0.82s time: 119.27s eta: 8:50:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32572.0 tgs: 270 data_time: 0.54s time: 120.47s eta: 8:54:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31098.0 tgs: 260 data_time: 0.69s time: 119.24s eta: 8:46:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31537.0 tgs: 264 data_time: 0.66s time: 119.09s eta: 8:43:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.105 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31417.0 tgs: 260 data_time: 0.69s time: 120.83s eta: 8:49:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31208.0 tgs: 259 data_time: 0.71s time: 120.19s eta: 8:44:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31058.0 tgs: 261 data_time: 0.80s time: 118.82s eta: 8:36:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31887.0 tgs: 267 data_time: 0.66s time: 119.35s eta: 8:37:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30486.0 tgs: 255 data_time: 0.50s time: 119.33s eta: 8:35:07 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 270 data_time: 0.81s time: 119.05s eta: 8:31:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.195 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32490.0 tgs: 272 data_time: 0.81s time: 119.16s eta: 8:30:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31594.0 tgs: 265 data_time: 0.94s time: 119.04s eta: 8:27:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 269 data_time: 0.94s time: 119.12s eta: 8:26:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32424.0 tgs: 271 data_time: 0.81s time: 119.24s eta: 8:24:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31633.0 tgs: 265 data_time: 0.91s time: 119.22s eta: 8:22:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.209 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31017.0 tgs: 259 data_time: 0.79s time: 119.38s eta: 8:21:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31390.0 tgs: 262 data_time: 0.73s time: 119.79s eta: 8:21:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 267 data_time: 0.65s time: 119.18s eta: 8:16:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.105 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30736.0 tgs: 258 data_time: 0.65s time: 119.03s eta: 8:13:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.186 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30698.0 tgs: 256 data_time: 0.69s time: 119.90s eta: 8:15:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31243.0 tgs: 261 data_time: 0.70s time: 119.55s eta: 8:12:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32271.0 tgs: 269 data_time: 1.03s time: 119.60s eta: 8:10:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32314.0 tgs: 271 data_time: 0.76s time: 119.00s eta: 8:05:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.117 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 28318.0 tgs: 236 data_time: 0.41s time: 119.84s eta: 8:07:19 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31641.0 tgs: 264 data_time: 0.89s time: 119.71s eta: 8:04:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30889.0 tgs: 258 data_time: 0.66s time: 119.40s eta: 8:01:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.56s time: 119.29s eta: 7:59:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.122 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31307.0 tgs: 262 data_time: 0.54s time: 119.19s eta: 7:56:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32259.0 tgs: 270 data_time: 0.93s time: 119.19s eta: 7:54:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31809.0 tgs: 267 data_time: 0.76s time: 118.94s eta: 7:51:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.193 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31990.0 tgs: 268 data_time: 0.71s time: 118.97s eta: 7:49:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32002.0 tgs: 267 data_time: 1.02s time: 119.42s eta: 7:49:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.102 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31894.0 tgs: 267 data_time: 0.54s time: 119.03s eta: 7:46:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31251.0 tgs: 262 data_time: 0.66s time: 119.11s eta: 7:44:30 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 266 data_time: 1.05s time: 120.02s eta: 7:46:03 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.114 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 261 data_time: 0.50s time: 119.82s eta: 7:43:19 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.196 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32246.0 tgs: 270 data_time: 0.79s time: 119.18s eta: 7:38:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31637.0 tgs: 265 data_time: 0.80s time: 118.95s eta: 7:35:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31956.0 tgs: 267 data_time: 0.94s time: 119.32s eta: 7:35:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.092 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 267 data_time: 0.84s time: 119.39s eta: 7:33:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.192 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31790.0 tgs: 266 data_time: 0.57s time: 119.07s eta: 7:30:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31339.0 tgs: 263 data_time: 0.65s time: 118.99s eta: 7:28:11 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 270 data_time: 0.51s time: 119.24s eta: 7:27:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30939.0 tgs: 260 data_time: 0.69s time: 118.98s eta: 7:24:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.202 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31132.0 tgs: 261 data_time: 0.63s time: 118.93s eta: 7:22:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.118 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31860.0 tgs: 267 data_time: 0.69s time: 119.08s eta: 7:20:35 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.190 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 270 data_time: 0.93s time: 119.00s eta: 7:18:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.114 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31250.0 tgs: 262 data_time: 0.64s time: 119.02s eta: 7:16:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32154.0 tgs: 269 data_time: 0.78s time: 119.27s eta: 7:15:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32447.0 tgs: 271 data_time: 0.71s time: 119.36s eta: 7:13:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32230.0 tgs: 270 data_time: 0.66s time: 118.93s eta: 7:10:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30588.0 tgs: 256 data_time: 0.90s time: 119.21s eta: 7:09:10 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.113 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31231.0 tgs: 261 data_time: 0.45s time: 119.57s eta: 7:08:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32108.0 tgs: 268 data_time: 0.76s time: 119.67s eta: 7:06:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31695.0 tgs: 266 data_time: 0.81s time: 118.95s eta: 7:02:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31556.0 tgs: 264 data_time: 0.84s time: 119.29s eta: 7:01:30 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30884.0 tgs: 258 data_time: 0.66s time: 119.52s eta: 7:00:18 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31425.0 tgs: 263 data_time: 0.61s time: 119.42s eta: 6:57:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31590.0 tgs: 265 data_time: 0.84s time: 118.91s eta: 6:54:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30873.0 tgs: 258 data_time: 0.85s time: 119.34s eta: 6:53:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31537.0 tgs: 265 data_time: 0.53s time: 118.94s eta: 6:50:19 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31960.0 tgs: 267 data_time: 0.67s time: 119.26s eta: 6:49:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32189.0 tgs: 270 data_time: 0.75s time: 119.02s eta: 6:46:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31631.0 tgs: 265 data_time: 0.96s time: 119.23s eta: 6:45:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31928.0 tgs: 267 data_time: 0.93s time: 119.41s eta: 6:44:00 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.105 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29844.0 tgs: 250 data_time: 0.59s time: 119.01s eta: 6:40:41 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32275.0 tgs: 271 data_time: 0.71s time: 119.01s eta: 6:38:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31913.0 tgs: 267 data_time: 0.62s time: 119.45s eta: 6:38:10 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32050.0 tgs: 269 data_time: 0.98s time: 118.81s eta: 6:34:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.185 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29848.0 tgs: 249 data_time: 0.77s time: 119.43s eta: 6:34:07 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.217 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31937.0 tgs: 267 data_time: 0.61s time: 119.57s eta: 6:32:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31347.0 tgs: 262 data_time: 0.85s time: 119.38s eta: 6:29:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.185 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31866.0 tgs: 267 data_time: 0.77s time: 119.28s eta: 6:27:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32094.0 tgs: 268 data_time: 0.63s time: 119.35s eta: 6:25:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30884.0 tgs: 259 data_time: 0.76s time: 118.83s eta: 6:22:13 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.92s time: 119.01s eta: 6:20:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 269 data_time: 0.91s time: 119.40s eta: 6:20:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32100.0 tgs: 268 data_time: 0.73s time: 119.34s eta: 6:17:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31565.0 tgs: 265 data_time: 0.85s time: 119.04s eta: 6:14:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32064.0 tgs: 268 data_time: 0.68s time: 119.44s eta: 6:14:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31493.0 tgs: 263 data_time: 0.69s time: 119.38s eta: 6:12:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31174.0 tgs: 261 data_time: 1.08s time: 119.28s eta: 6:09:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31020.0 tgs: 260 data_time: 0.88s time: 119.10s eta: 6:07:13 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31816.0 tgs: 260 data_time: 0.85s time: 121.96s eta: 6:14:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31623.0 tgs: 264 data_time: 0.58s time: 119.39s eta: 6:04:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30566.0 tgs: 255 data_time: 0.48s time: 119.56s eta: 6:02:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31584.0 tgs: 264 data_time: 0.74s time: 119.26s eta: 5:59:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31067.0 tgs: 260 data_time: 0.79s time: 119.26s eta: 5:57:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.234 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31123.0 tgs: 261 data_time: 0.74s time: 118.91s eta: 5:54:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31012.0 tgs: 260 data_time: 0.78s time: 119.02s eta: 5:53:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.110 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30400.0 tgs: 255 data_time: 0.98s time: 119.08s eta: 5:51:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31215.0 tgs: 261 data_time: 0.59s time: 119.28s eta: 5:49:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31577.0 tgs: 265 data_time: 0.81s time: 118.97s eta: 5:47:00 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31521.0 tgs: 264 data_time: 0.82s time: 119.26s eta: 5:45:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31714.0 tgs: 265 data_time: 0.80s time: 119.36s eta: 5:44:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 271 data_time: 0.88s time: 119.06s eta: 5:41:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31618.0 tgs: 264 data_time: 0.62s time: 119.49s eta: 5:40:33 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.112 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30235.0 tgs: 253 data_time: 0.65s time: 119.38s eta: 5:38:14 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.090 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31540.0 tgs: 263 data_time: 0.66s time: 119.57s eta: 5:36:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31636.0 tgs: 265 data_time: 0.65s time: 118.98s eta: 5:33:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31280.0 tgs: 262 data_time: 0.67s time: 118.95s eta: 5:31:03 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32072.0 tgs: 268 data_time: 0.97s time: 119.24s eta: 5:29:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.195 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 265 data_time: 0.82s time: 119.40s eta: 5:28:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30192.0 tgs: 252 data_time: 0.66s time: 119.46s eta: 5:26:31 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32247.0 tgs: 271 data_time: 0.88s time: 118.95s eta: 5:23:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.186 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31291.0 tgs: 261 data_time: 0.61s time: 119.63s eta: 5:22:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.247 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32291.0 tgs: 270 data_time: 0.69s time: 119.35s eta: 5:20:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31019.0 tgs: 260 data_time: 0.78s time: 119.12s eta: 5:17:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32518.0 tgs: 272 data_time: 0.90s time: 119.42s eta: 5:16:27 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31890.0 tgs: 266 data_time: 0.76s time: 119.58s eta: 5:14:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.69s time: 119.00s eta: 5:11:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.210 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31840.0 tgs: 267 data_time: 0.63s time: 119.00s eta: 5:09:23 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.112 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31946.0 tgs: 267 data_time: 0.70s time: 119.22s eta: 5:07:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30229.0 tgs: 252 data_time: 0.76s time: 119.68s eta: 5:07:11 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32233.0 tgs: 270 data_time: 0.96s time: 119.11s eta: 5:03:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.200 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31286.0 tgs: 262 data_time: 0.76s time: 119.07s eta: 5:01:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.204 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 264 data_time: 0.63s time: 121.52s eta: 5:05:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.182 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31912.0 tgs: 266 data_time: 0.69s time: 119.59s eta: 4:58:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30824.0 tgs: 257 data_time: 0.61s time: 119.49s eta: 4:56:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31521.0 tgs: 263 data_time: 0.87s time: 119.75s eta: 4:55:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31299.0 tgs: 261 data_time: 0.81s time: 119.51s eta: 4:52:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 268 data_time: 0.53s time: 119.41s eta: 4:50:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 269 data_time: 0.64s time: 119.21s eta: 4:48:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31419.0 tgs: 264 data_time: 0.74s time: 118.88s eta: 4:45:18 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.193 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31396.0 tgs: 263 data_time: 0.88s time: 119.11s eta: 4:43:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 29738.0 tgs: 247 data_time: 0.78s time: 119.98s eta: 4:43:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32150.0 tgs: 269 data_time: 0.55s time: 119.16s eta: 4:40:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32004.0 tgs: 268 data_time: 0.91s time: 119.00s eta: 4:37:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31623.0 tgs: 265 data_time: 0.95s time: 119.05s eta: 4:35:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31443.0 tgs: 263 data_time: 0.75s time: 119.37s eta: 4:34:33 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.79s time: 119.41s eta: 4:32:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.182 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 265 data_time: 0.92s time: 118.94s eta: 4:29:35 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32027.0 tgs: 267 data_time: 0.88s time: 119.60s eta: 4:29:06 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 268 data_time: 0.69s time: 119.24s eta: 4:26:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 268 data_time: 0.63s time: 119.06s eta: 4:23:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.222 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31343.0 tgs: 263 data_time: 0.69s time: 118.85s eta: 4:21:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31979.0 tgs: 269 data_time: 0.60s time: 118.75s eta: 4:19:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30556.0 tgs: 255 data_time: 0.56s time: 119.43s eta: 4:18:45 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31444.0 tgs: 263 data_time: 0.55s time: 119.41s eta: 4:16:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31201.0 tgs: 261 data_time: 0.57s time: 119.33s eta: 4:14:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31909.0 tgs: 268 data_time: 0.69s time: 119.04s eta: 4:11:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.104 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31029.0 tgs: 260 data_time: 0.68s time: 119.12s eta: 4:10:09 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32024.0 tgs: 268 data_time: 0.48s time: 119.13s eta: 4:08:11 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31634.0 tgs: 265 data_time: 0.44s time: 118.96s eta: 4:05:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.198 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32106.0 tgs: 269 data_time: 0.78s time: 119.15s eta: 4:04:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.194 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31552.0 tgs: 264 data_time: 0.84s time: 119.50s eta: 4:02:58 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31367.0 tgs: 263 data_time: 0.84s time: 119.25s eta: 4:00:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31019.0 tgs: 260 data_time: 1.05s time: 119.19s eta: 3:58:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31415.0 tgs: 263 data_time: 0.79s time: 119.36s eta: 3:56:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31542.0 tgs: 264 data_time: 0.73s time: 119.36s eta: 3:54:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 27706.0 tgs: 230 data_time: 0.84s time: 119.97s eta: 3:53:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31183.0 tgs: 261 data_time: 0.63s time: 119.30s eta: 3:50:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31900.0 tgs: 266 data_time: 0.83s time: 119.63s eta: 3:49:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32139.0 tgs: 270 data_time: 0.68s time: 118.85s eta: 3:45:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.097 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32489.0 tgs: 271 data_time: 0.75s time: 119.75s eta: 3:45:31 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32322.0 tgs: 270 data_time: 0.90s time: 119.33s eta: 3:42:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.83s time: 119.26s eta: 3:40:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31400.0 tgs: 263 data_time: 0.79s time: 119.00s eta: 3:38:10 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30184.0 tgs: 252 data_time: 0.76s time: 119.43s eta: 3:36:57 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31710.0 tgs: 266 data_time: 0.91s time: 119.03s eta: 3:34:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31717.0 tgs: 266 data_time: 0.80s time: 118.99s eta: 3:32:11 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 270 data_time: 0.86s time: 119.22s eta: 3:30:36 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.210 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32287.0 tgs: 268 data_time: 0.67s time: 120.12s eta: 3:30:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.231 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32077.0 tgs: 269 data_time: 0.74s time: 119.12s eta: 3:26:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31788.0 tgs: 266 data_time: 0.90s time: 119.31s eta: 3:24:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32022.0 tgs: 268 data_time: 0.79s time: 119.07s eta: 3:22:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32164.0 tgs: 269 data_time: 0.72s time: 119.43s eta: 3:21:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 270 data_time: 0.86s time: 119.29s eta: 3:18:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.102 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31748.0 tgs: 266 data_time: 0.69s time: 119.22s eta: 3:16:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31104.0 tgs: 260 data_time: 0.85s time: 119.18s eta: 3:14:39 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31079.0 tgs: 261 data_time: 0.70s time: 119.02s eta: 3:12:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.202 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31602.0 tgs: 265 data_time: 0.78s time: 119.18s eta: 3:10:41 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31775.0 tgs: 266 data_time: 0.65s time: 119.43s eta: 3:09:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 266 data_time: 0.91s time: 119.50s eta: 3:07:13 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31718.0 tgs: 265 data_time: 0.68s time: 119.26s eta: 3:04:51 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:43:13][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31596.0 tgs: 265 data_time: 0.63s time: 118.99s eta: 3:02:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 269 data_time: 0.58s time: 119.22s eta: 3:00:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32035.0 tgs: 268 data_time: 0.61s time: 119.45s eta: 2:59:10 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31984.0 tgs: 268 data_time: 0.58s time: 118.98s eta: 2:56:28 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.186 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32579.0 tgs: 272 data_time: 0.81s time: 119.35s eta: 2:55:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31768.0 tgs: 266 data_time: 0.75s time: 119.41s eta: 2:53:08 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31188.0 tgs: 261 data_time: 0.69s time: 119.35s eta: 2:51:04 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32414.0 tgs: 266 data_time: 0.82s time: 121.46s eta: 2:52:03 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31817.0 tgs: 266 data_time: 0.69s time: 119.24s eta: 2:46:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31158.0 tgs: 260 data_time: 0.88s time: 119.57s eta: 2:45:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31497.0 tgs: 264 data_time: 0.67s time: 118.90s eta: 2:42:30 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31684.0 tgs: 265 data_time: 0.92s time: 119.18s eta: 2:40:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.212 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32351.0 tgs: 271 data_time: 0.75s time: 118.96s eta: 2:38:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 264 data_time: 0.74s time: 119.23s eta: 2:36:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.114 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 266 data_time: 0.79s time: 119.11s eta: 2:34:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 267 data_time: 0.68s time: 119.43s eta: 2:33:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 267 data_time: 0.67s time: 120.23s eta: 2:32:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31722.0 tgs: 265 data_time: 0.74s time: 119.28s eta: 2:29:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.185 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31811.0 tgs: 267 data_time: 0.56s time: 118.88s eta: 2:26:36 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32018.0 tgs: 267 data_time: 0.84s time: 119.52s eta: 2:25:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31891.0 tgs: 267 data_time: 0.91s time: 119.07s eta: 2:22:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.86s time: 119.14s eta: 2:20:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31492.0 tgs: 264 data_time: 0.54s time: 119.05s eta: 2:18:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32193.0 tgs: 268 data_time: 1.06s time: 119.84s eta: 2:17:48 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30241.0 tgs: 253 data_time: 0.86s time: 119.36s eta: 2:15:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.80s time: 119.41s eta: 2:13:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31320.0 tgs: 261 data_time: 0.64s time: 119.66s eta: 2:11:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31387.0 tgs: 262 data_time: 0.84s time: 119.49s eta: 2:09:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31182.0 tgs: 262 data_time: 0.63s time: 118.96s eta: 2:06:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31870.0 tgs: 266 data_time: 0.85s time: 119.37s eta: 2:05:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.186 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.69s time: 119.28s eta: 2:03:15 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 266 data_time: 0.72s time: 119.21s eta: 2:01:12 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31403.0 tgs: 262 data_time: 0.68s time: 119.50s eta: 1:59:30 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29955.0 tgs: 250 data_time: 0.74s time: 119.42s eta: 1:57:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32458.0 tgs: 271 data_time: 1.00s time: 119.53s eta: 1:55:32 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.199 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30187.0 tgs: 252 data_time: 0.81s time: 119.32s eta: 1:53:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 268 data_time: 0.81s time: 119.23s eta: 1:51:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 269 data_time: 0.95s time: 118.93s eta: 1:49:01 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31144.0 tgs: 260 data_time: 0.67s time: 119.64s eta: 1:47:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32031.0 tgs: 267 data_time: 0.94s time: 119.72s eta: 1:45:44 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 268 data_time: 0.71s time: 119.35s eta: 1:43:25 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.59s time: 119.30s eta: 1:41:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31723.0 tgs: 265 data_time: 0.65s time: 119.29s eta: 1:39:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31003.0 tgs: 260 data_time: 0.66s time: 118.90s eta: 1:37:05 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.52s time: 119.16s eta: 1:35:19 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 268 data_time: 0.75s time: 119.08s eta: 1:33:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32244.0 tgs: 270 data_time: 0.50s time: 119.24s eta: 1:31:24 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 264 data_time: 0.57s time: 119.73s eta: 1:29:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 267 data_time: 0.76s time: 119.01s eta: 1:27:16 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 265 data_time: 0.56s time: 118.88s eta: 1:25:11 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31921.0 tgs: 267 data_time: 0.71s time: 119.47s eta: 1:23:37 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31051.0 tgs: 259 data_time: 0.60s time: 119.49s eta: 1:21:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32226.0 tgs: 270 data_time: 0.64s time: 118.93s eta: 1:19:17 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30584.0 tgs: 257 data_time: 0.67s time: 118.93s eta: 1:17:18 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 262 data_time: 0.63s time: 122.18s eta: 1:17:22 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31139.0 tgs: 260 data_time: 0.64s time: 119.35s eta: 1:13:36 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31366.0 tgs: 263 data_time: 0.68s time: 119.06s eta: 1:11:26 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31876.0 tgs: 268 data_time: 0.81s time: 118.91s eta: 1:09:21 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30867.0 tgs: 259 data_time: 0.74s time: 119.11s eta: 1:07:29 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31772.0 tgs: 266 data_time: 0.61s time: 119.41s eta: 1:05:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30137.0 tgs: 252 data_time: 0.43s time: 119.34s eta: 1:03:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.228 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32537.0 tgs: 273 data_time: 0.73s time: 119.18s eta: 1:01:34 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 26640.0 tgs: 223 data_time: 0.81s time: 119.27s eta: 0:59:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31320.0 tgs: 262 data_time: 0.71s time: 119.44s eta: 0:57:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31224.0 tgs: 261 data_time: 0.78s time: 119.39s eta: 0:55:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 265 data_time: 1.00s time: 119.48s eta: 0:53:46 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32255.0 tgs: 270 data_time: 0.53s time: 119.36s eta: 0:51:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.221 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30901.0 tgs: 258 data_time: 0.71s time: 119.49s eta: 0:49:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31317.0 tgs: 263 data_time: 0.93s time: 118.88s eta: 0:47:33 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29314.0 tgs: 245 data_time: 0.68s time: 119.28s eta: 0:45:43 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31589.0 tgs: 264 data_time: 0.74s time: 119.21s eta: 0:43:42 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32165.0 tgs: 270 data_time: 0.55s time: 118.99s eta: 0:41:38 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.094 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30580.0 tgs: 256 data_time: 0.51s time: 119.04s eta: 0:39:40 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32292.0 tgs: 269 data_time: 1.08s time: 119.66s eta: 0:37:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31410.0 tgs: 263 data_time: 0.57s time: 119.28s eta: 0:35:47 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30524.0 tgs: 254 data_time: 0.68s time: 119.74s eta: 0:33:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31685.0 tgs: 261 data_time: 0.73s time: 121.25s eta: 0:32:20 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31233.0 tgs: 261 data_time: 0.68s time: 119.36s eta: 0:29:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30549.0 tgs: 256 data_time: 0.86s time: 119.32s eta: 0:27:50 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.81s time: 119.64s eta: 0:25:55 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.106 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31744.0 tgs: 266 data_time: 0.76s time: 119.09s eta: 0:23:49 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.195 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30869.0 tgs: 258 data_time: 0.80s time: 119.39s eta: 0:21:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31257.0 tgs: 262 data_time: 0.83s time: 119.26s eta: 0:19:52 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.223 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31161.0 tgs: 261 data_time: 0.64s time: 119.24s eta: 0:17:53 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31976.0 tgs: 266 data_time: 0.51s time: 119.90s eta: 0:15:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30556.0 tgs: 256 data_time: 0.58s time: 119.21s eta: 0:13:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31361.0 tgs: 263 data_time: 0.74s time: 119.09s eta: 0:11:54 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 266 data_time: 0.84s time: 119.32s eta: 0:09:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32590.0 tgs: 273 data_time: 0.59s time: 119.17s eta: 0:07:56 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31100.0 tgs: 259 data_time: 0.64s time: 119.69s eta: 0:05:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 263 data_time: 0.52s time: 121.30s eta: 0:04:02 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31457.0 tgs: 263 data_time: 0.81s time: 119.21s eta: 0:01:59 +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 24][DP 6][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank27.log b/20250121165312/rank27.log new file mode 100644 index 0000000000000000000000000000000000000000..162ea38174806fbfdb76d821efbd6fd29e22c154 --- /dev/null +++ b/20250121165312/rank27.log @@ -0,0 +1,851 @@ +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.65 seconds, peak gpu memory 13.4G +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:04:45][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.258 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31349.0 tgs: 204 data_time: 1.71s time: 153.56s eta: 1 day, 1:17:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.259 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32045.0 tgs: 268 data_time: 0.54s time: 119.18s eta: 19:35:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.230 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31775.0 tgs: 265 data_time: 0.88s time: 119.68s eta: 19:38:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 262 data_time: 0.82s time: 119.65s eta: 19:36:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31310.0 tgs: 262 data_time: 0.73s time: 119.31s eta: 19:31:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.219 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31970.0 tgs: 267 data_time: 0.84s time: 119.53s eta: 19:31:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.194 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 260 data_time: 0.83s time: 122.96s eta: 20:02:57 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32280.0 tgs: 269 data_time: 0.66s time: 119.89s eta: 19:30:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.201 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32093.0 tgs: 265 data_time: 0.98s time: 120.66s eta: 19:36:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.204 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32007.0 tgs: 266 data_time: 0.54s time: 119.92s eta: 19:27:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.213 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32301.0 tgs: 268 data_time: 0.74s time: 120.46s eta: 19:30:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30685.0 tgs: 257 data_time: 0.80s time: 119.34s eta: 19:17:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30776.0 tgs: 257 data_time: 0.84s time: 119.68s eta: 19:18:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.249 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31850.0 tgs: 267 data_time: 0.70s time: 119.18s eta: 19:12:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.233 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32016.0 tgs: 268 data_time: 0.75s time: 119.21s eta: 19:10:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31587.0 tgs: 265 data_time: 1.00s time: 118.93s eta: 19:05:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30793.0 tgs: 258 data_time: 0.74s time: 119.14s eta: 19:05:44 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31008.0 tgs: 260 data_time: 0.88s time: 119.18s eta: 19:04:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32462.0 tgs: 271 data_time: 0.88s time: 119.40s eta: 19:04:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32193.0 tgs: 269 data_time: 0.66s time: 119.33s eta: 19:01:33 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31420.0 tgs: 261 data_time: 0.77s time: 120.31s eta: 19:08:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.249 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31579.0 tgs: 256 data_time: 0.93s time: 122.91s eta: 19:31:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 266 data_time: 0.81s time: 119.48s eta: 18:57:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30152.0 tgs: 253 data_time: 0.80s time: 119.16s eta: 18:52:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31164.0 tgs: 261 data_time: 0.88s time: 119.24s eta: 18:50:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 265 data_time: 0.83s time: 119.39s eta: 18:50:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 268 data_time: 0.91s time: 119.66s eta: 18:50:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30357.0 tgs: 254 data_time: 0.75s time: 119.38s eta: 18:46:10 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 267 data_time: 0.87s time: 119.79s eta: 18:48:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31963.0 tgs: 247 data_time: 0.86s time: 129.33s eta: 20:15:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31960.0 tgs: 267 data_time: 0.79s time: 119.57s eta: 18:41:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31654.0 tgs: 265 data_time: 0.79s time: 119.39s eta: 18:38:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31502.0 tgs: 261 data_time: 0.62s time: 120.65s eta: 18:48:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.87s time: 119.40s eta: 18:34:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31221.0 tgs: 262 data_time: 0.71s time: 119.16s eta: 18:30:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.95s time: 119.08s eta: 18:27:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31855.0 tgs: 267 data_time: 1.01s time: 119.09s eta: 18:25:32 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30932.0 tgs: 256 data_time: 0.69s time: 120.39s eta: 18:35:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.270 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 263 data_time: 0.61s time: 121.17s eta: 18:40:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 262 data_time: 0.61s time: 120.66s eta: 18:34:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32069.0 tgs: 267 data_time: 0.71s time: 119.73s eta: 18:23:31 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.242 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31550.0 tgs: 254 data_time: 0.60s time: 123.87s eta: 18:59:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30417.0 tgs: 254 data_time: 0.52s time: 119.55s eta: 18:17:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31874.0 tgs: 267 data_time: 0.84s time: 119.35s eta: 18:14:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31865.0 tgs: 267 data_time: 0.60s time: 119.22s eta: 18:10:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32131.0 tgs: 268 data_time: 0.94s time: 119.60s eta: 18:12:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32443.0 tgs: 272 data_time: 0.65s time: 119.26s eta: 18:07:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31045.0 tgs: 260 data_time: 0.74s time: 119.02s eta: 18:03:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 267 data_time: 0.74s time: 119.09s eta: 18:01:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31945.0 tgs: 267 data_time: 0.94s time: 119.26s eta: 18:01:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32011.0 tgs: 268 data_time: 1.07s time: 119.33s eta: 17:59:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32547.0 tgs: 271 data_time: 0.96s time: 120.02s eta: 18:04:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32049.0 tgs: 269 data_time: 0.85s time: 119.10s eta: 17:53:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32269.0 tgs: 270 data_time: 0.85s time: 119.41s eta: 17:54:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30865.0 tgs: 259 data_time: 1.07s time: 118.99s eta: 17:48:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31114.0 tgs: 261 data_time: 1.07s time: 119.02s eta: 17:47:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32034.0 tgs: 255 data_time: 0.82s time: 125.36s eta: 18:41:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 262 data_time: 1.23s time: 120.10s eta: 17:52:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30677.0 tgs: 256 data_time: 0.74s time: 119.40s eta: 17:44:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.248 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31925.0 tgs: 267 data_time: 1.00s time: 119.34s eta: 17:42:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32154.0 tgs: 269 data_time: 0.60s time: 119.51s eta: 17:41:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.230 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 254 data_time: 0.81s time: 126.04s eta: 18:37:35 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31246.0 tgs: 254 data_time: 0.89s time: 122.64s eta: 18:05:22 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31119.0 tgs: 256 data_time: 0.87s time: 121.14s eta: 17:50:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31196.0 tgs: 262 data_time: 0.59s time: 119.03s eta: 17:29:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 268 data_time: 0.69s time: 118.85s eta: 17:25:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31639.0 tgs: 255 data_time: 0.83s time: 124.00s eta: 18:09:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31322.0 tgs: 251 data_time: 1.03s time: 124.57s eta: 18:12:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31116.0 tgs: 260 data_time: 0.81s time: 119.39s eta: 17:24:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 268 data_time: 0.65s time: 119.70s eta: 17:25:22 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31426.0 tgs: 262 data_time: 0.76s time: 119.52s eta: 17:21:51 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31953.0 tgs: 268 data_time: 0.76s time: 118.86s eta: 17:14:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31272.0 tgs: 260 data_time: 0.73s time: 119.85s eta: 17:20:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31828.0 tgs: 266 data_time: 0.86s time: 119.56s eta: 17:16:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30988.0 tgs: 254 data_time: 0.81s time: 121.87s eta: 17:34:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31784.0 tgs: 257 data_time: 1.00s time: 123.51s eta: 17:46:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31384.0 tgs: 263 data_time: 0.66s time: 119.01s eta: 17:05:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31079.0 tgs: 260 data_time: 0.71s time: 119.23s eta: 17:05:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.248 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31623.0 tgs: 265 data_time: 0.83s time: 119.13s eta: 17:02:31 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 268 data_time: 0.72s time: 119.26s eta: 17:01:37 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31751.0 tgs: 266 data_time: 0.99s time: 119.04s eta: 16:57:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30915.0 tgs: 259 data_time: 0.78s time: 119.28s eta: 16:57:50 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31739.0 tgs: 266 data_time: 0.80s time: 119.15s eta: 16:54:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.125 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31726.0 tgs: 265 data_time: 0.61s time: 119.58s eta: 16:56:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 26330.0 tgs: 221 data_time: 0.69s time: 119.13s eta: 16:50:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.225 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 269 data_time: 0.72s time: 118.84s eta: 16:46:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31115.0 tgs: 261 data_time: 0.76s time: 119.01s eta: 16:45:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 268 data_time: 0.73s time: 119.85s eta: 16:50:42 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.108 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30669.0 tgs: 256 data_time: 0.81s time: 119.38s eta: 16:44:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 266 data_time: 0.71s time: 119.01s eta: 16:39:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32030.0 tgs: 264 data_time: 0.74s time: 120.93s eta: 16:53:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 265 data_time: 0.78s time: 118.99s eta: 16:35:32 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32088.0 tgs: 268 data_time: 0.79s time: 119.29s eta: 16:36:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32331.0 tgs: 271 data_time: 0.59s time: 119.23s eta: 16:33:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32221.0 tgs: 269 data_time: 0.92s time: 119.36s eta: 16:32:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30610.0 tgs: 255 data_time: 0.69s time: 119.70s eta: 16:33:30 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31693.0 tgs: 265 data_time: 0.81s time: 119.20s eta: 16:27:22 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31074.0 tgs: 259 data_time: 0.80s time: 119.77s eta: 16:30:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31521.0 tgs: 262 data_time: 0.72s time: 120.25s eta: 16:32:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30895.0 tgs: 259 data_time: 0.97s time: 118.97s eta: 16:19:31 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.205 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32094.0 tgs: 252 data_time: 0.74s time: 127.33s eta: 17:26:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31475.0 tgs: 263 data_time: 0.75s time: 119.22s eta: 16:17:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32583.0 tgs: 273 data_time: 0.66s time: 119.14s eta: 16:14:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32561.0 tgs: 273 data_time: 0.73s time: 119.26s eta: 16:13:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31165.0 tgs: 261 data_time: 0.63s time: 119.18s eta: 16:11:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31581.0 tgs: 265 data_time: 0.71s time: 118.95s eta: 16:07:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.225 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31719.0 tgs: 266 data_time: 0.82s time: 118.86s eta: 16:04:44 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31776.0 tgs: 267 data_time: 0.77s time: 118.97s eta: 16:03:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31490.0 tgs: 258 data_time: 0.84s time: 121.83s eta: 16:24:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31856.0 tgs: 267 data_time: 0.85s time: 119.00s eta: 15:59:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29829.0 tgs: 250 data_time: 0.65s time: 119.22s eta: 15:59:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 268 data_time: 0.60s time: 119.61s eta: 16:00:51 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32587.0 tgs: 273 data_time: 0.64s time: 119.04s eta: 15:54:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.212 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29377.0 tgs: 246 data_time: 0.63s time: 118.99s eta: 15:51:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31543.0 tgs: 264 data_time: 0.55s time: 119.16s eta: 15:51:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 266 data_time: 0.62s time: 119.63s eta: 15:53:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31486.0 tgs: 264 data_time: 0.65s time: 118.94s eta: 15:45:32 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30548.0 tgs: 256 data_time: 0.72s time: 119.29s eta: 15:46:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 264 data_time: 0.75s time: 119.12s eta: 15:43:04 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.210 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30535.0 tgs: 256 data_time: 0.46s time: 119.00s eta: 15:40:04 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.212 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31961.0 tgs: 267 data_time: 0.81s time: 119.66s eta: 15:43:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31872.0 tgs: 262 data_time: 1.04s time: 121.50s eta: 15:55:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31557.0 tgs: 264 data_time: 0.79s time: 119.41s eta: 15:37:22 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30431.0 tgs: 234 data_time: 0.77s time: 129.96s eta: 16:58:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31153.0 tgs: 260 data_time: 0.81s time: 119.46s eta: 15:33:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31444.0 tgs: 262 data_time: 0.80s time: 119.61s eta: 15:32:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31241.0 tgs: 261 data_time: 0.85s time: 119.53s eta: 15:30:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31598.0 tgs: 258 data_time: 0.82s time: 122.14s eta: 15:48:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30829.0 tgs: 255 data_time: 0.84s time: 120.75s eta: 15:35:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31380.0 tgs: 262 data_time: 0.78s time: 119.41s eta: 15:23:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 263 data_time: 1.04s time: 119.44s eta: 15:21:42 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31903.0 tgs: 242 data_time: 0.78s time: 131.63s eta: 16:53:32 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 266 data_time: 0.73s time: 119.30s eta: 15:16:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.283 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.78s time: 119.56s eta: 15:16:37 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31744.0 tgs: 259 data_time: 0.58s time: 122.26s eta: 15:35:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32240.0 tgs: 269 data_time: 0.68s time: 119.71s eta: 15:13:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.233 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31278.0 tgs: 262 data_time: 0.67s time: 119.34s eta: 15:09:00 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 266 data_time: 0.68s time: 118.90s eta: 15:03:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31528.0 tgs: 264 data_time: 0.80s time: 118.99s eta: 15:02:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30920.0 tgs: 259 data_time: 0.61s time: 119.31s eta: 15:02:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32178.0 tgs: 270 data_time: 0.63s time: 118.92s eta: 14:57:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30764.0 tgs: 255 data_time: 0.79s time: 120.52s eta: 15:07:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.207 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30675.0 tgs: 257 data_time: 0.61s time: 119.31s eta: 14:56:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.188 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 262 data_time: 0.75s time: 122.66s eta: 15:19:57 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.208 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31640.0 tgs: 265 data_time: 0.89s time: 119.12s eta: 14:51:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.206 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31844.0 tgs: 265 data_time: 1.03s time: 119.75s eta: 14:54:10 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31220.0 tgs: 261 data_time: 0.68s time: 119.46s eta: 14:49:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31075.0 tgs: 261 data_time: 1.01s time: 119.00s eta: 14:44:33 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 264 data_time: 0.91s time: 119.32s eta: 14:44:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31692.0 tgs: 266 data_time: 0.69s time: 118.94s eta: 14:40:07 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 270 data_time: 1.09s time: 119.05s eta: 14:39:00 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.189 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32575.0 tgs: 273 data_time: 0.76s time: 119.06s eta: 14:37:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31516.0 tgs: 264 data_time: 0.87s time: 119.13s eta: 14:35:35 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31173.0 tgs: 261 data_time: 0.74s time: 118.99s eta: 14:32:37 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 28860.0 tgs: 242 data_time: 0.77s time: 119.08s eta: 14:31:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.241 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32241.0 tgs: 270 data_time: 0.65s time: 119.15s eta: 14:29:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30260.0 tgs: 252 data_time: 0.72s time: 119.65s eta: 14:31:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31648.0 tgs: 265 data_time: 0.78s time: 119.18s eta: 14:26:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32108.0 tgs: 269 data_time: 0.86s time: 119.20s eta: 14:24:10 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31820.0 tgs: 266 data_time: 0.60s time: 119.39s eta: 14:23:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32105.0 tgs: 265 data_time: 0.71s time: 120.73s eta: 14:31:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30550.0 tgs: 257 data_time: 0.88s time: 118.85s eta: 14:15:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 270 data_time: 0.94s time: 119.10s eta: 14:15:32 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31196.0 tgs: 261 data_time: 0.70s time: 119.16s eta: 14:13:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.227 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31096.0 tgs: 259 data_time: 0.79s time: 119.96s eta: 14:17:42 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31223.0 tgs: 260 data_time: 0.94s time: 119.87s eta: 14:15:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 268 data_time: 0.73s time: 119.58s eta: 14:11:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31437.0 tgs: 263 data_time: 0.67s time: 119.22s eta: 14:06:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31132.0 tgs: 261 data_time: 0.59s time: 119.26s eta: 14:04:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31410.0 tgs: 263 data_time: 0.63s time: 119.26s eta: 14:02:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.110 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31532.0 tgs: 264 data_time: 0.66s time: 119.14s eta: 13:59:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32424.0 tgs: 272 data_time: 0.77s time: 119.12s eta: 13:57:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30282.0 tgs: 254 data_time: 0.68s time: 119.17s eta: 13:56:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31583.0 tgs: 264 data_time: 0.72s time: 119.35s eta: 13:55:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.199 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32066.0 tgs: 266 data_time: 0.81s time: 120.11s eta: 13:58:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31409.0 tgs: 256 data_time: 0.62s time: 122.56s eta: 14:13:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:56:46][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.73s time: 119.10s eta: 13:47:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 267 data_time: 0.47s time: 119.37s eta: 13:47:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32102.0 tgs: 268 data_time: 0.71s time: 119.35s eta: 13:45:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32140.0 tgs: 269 data_time: 0.86s time: 119.35s eta: 13:43:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31889.0 tgs: 267 data_time: 0.91s time: 119.11s eta: 13:39:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31018.0 tgs: 260 data_time: 0.69s time: 119.25s eta: 13:38:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31186.0 tgs: 261 data_time: 0.60s time: 119.13s eta: 13:36:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31655.0 tgs: 265 data_time: 0.84s time: 119.22s eta: 13:34:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.212 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31207.0 tgs: 261 data_time: 0.90s time: 119.46s eta: 13:34:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31178.0 tgs: 261 data_time: 0.75s time: 119.08s eta: 13:29:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31976.0 tgs: 268 data_time: 0.49s time: 118.92s eta: 13:26:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.227 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31843.0 tgs: 265 data_time: 1.06s time: 119.80s eta: 13:30:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31274.0 tgs: 262 data_time: 0.80s time: 119.25s eta: 13:24:57 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.202 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31670.0 tgs: 265 data_time: 0.89s time: 119.16s eta: 13:22:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.228 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 267 data_time: 0.68s time: 119.34s eta: 13:21:35 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.76s time: 119.26s eta: 13:19:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.201 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31542.0 tgs: 264 data_time: 0.73s time: 119.04s eta: 13:15:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.97s time: 119.37s eta: 13:15:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31195.0 tgs: 261 data_time: 0.65s time: 119.18s eta: 13:12:31 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32118.0 tgs: 268 data_time: 0.83s time: 119.46s eta: 13:12:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31019.0 tgs: 260 data_time: 0.90s time: 119.28s eta: 13:09:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32455.0 tgs: 272 data_time: 0.86s time: 119.15s eta: 13:06:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31953.0 tgs: 265 data_time: 0.70s time: 120.24s eta: 13:11:33 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31412.0 tgs: 261 data_time: 0.71s time: 120.16s eta: 13:09:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.193 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31853.0 tgs: 266 data_time: 0.78s time: 119.50s eta: 13:02:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 265 data_time: 0.85s time: 119.51s eta: 13:00:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31252.0 tgs: 261 data_time: 0.84s time: 119.57s eta: 12:59:12 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.217 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 267 data_time: 0.86s time: 119.80s eta: 12:58:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30904.0 tgs: 259 data_time: 0.77s time: 119.04s eta: 12:51:44 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32316.0 tgs: 271 data_time: 0.71s time: 119.05s eta: 12:49:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31416.0 tgs: 263 data_time: 0.81s time: 119.12s eta: 12:48:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31211.0 tgs: 261 data_time: 0.69s time: 119.40s eta: 12:48:07 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.185 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31678.0 tgs: 265 data_time: 1.07s time: 119.19s eta: 12:44:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31625.0 tgs: 264 data_time: 0.65s time: 119.44s eta: 12:44:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32437.0 tgs: 272 data_time: 0.68s time: 119.10s eta: 12:40:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 266 data_time: 0.85s time: 119.07s eta: 12:38:04 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31576.0 tgs: 264 data_time: 0.85s time: 119.59s eta: 12:39:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.264 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31649.0 tgs: 266 data_time: 0.71s time: 118.78s eta: 12:32:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31920.0 tgs: 268 data_time: 0.89s time: 118.88s eta: 12:30:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31644.0 tgs: 265 data_time: 0.97s time: 119.17s eta: 12:30:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31897.0 tgs: 267 data_time: 0.67s time: 119.26s eta: 12:29:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30222.0 tgs: 253 data_time: 0.90s time: 119.35s eta: 12:27:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32146.0 tgs: 269 data_time: 0.63s time: 119.11s eta: 12:24:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31620.0 tgs: 264 data_time: 0.93s time: 119.34s eta: 12:23:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31262.0 tgs: 262 data_time: 1.01s time: 119.26s eta: 12:21:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32133.0 tgs: 269 data_time: 0.64s time: 119.03s eta: 12:17:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31407.0 tgs: 259 data_time: 0.63s time: 121.13s eta: 12:28:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31926.0 tgs: 266 data_time: 0.83s time: 119.99s eta: 12:19:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.194 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30827.0 tgs: 259 data_time: 0.88s time: 118.72s eta: 12:10:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31566.0 tgs: 263 data_time: 0.66s time: 119.63s eta: 12:13:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31944.0 tgs: 266 data_time: 0.85s time: 119.83s eta: 12:12:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.64s time: 119.15s eta: 12:06:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.210 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32254.0 tgs: 270 data_time: 1.00s time: 119.17s eta: 12:04:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31838.0 tgs: 267 data_time: 0.72s time: 119.12s eta: 12:02:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31452.0 tgs: 263 data_time: 0.76s time: 119.18s eta: 12:01:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30922.0 tgs: 259 data_time: 0.77s time: 118.96s eta: 11:57:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31065.0 tgs: 259 data_time: 0.59s time: 119.71s eta: 12:00:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32267.0 tgs: 270 data_time: 0.75s time: 119.27s eta: 11:55:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31351.0 tgs: 262 data_time: 0.51s time: 119.28s eta: 11:53:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31937.0 tgs: 266 data_time: 0.76s time: 119.70s eta: 11:54:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31108.0 tgs: 261 data_time: 0.63s time: 118.78s eta: 11:46:44 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32397.0 tgs: 271 data_time: 0.77s time: 119.16s eta: 11:46:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31143.0 tgs: 260 data_time: 0.91s time: 119.61s eta: 11:47:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31596.0 tgs: 265 data_time: 0.63s time: 118.87s eta: 11:41:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31494.0 tgs: 265 data_time: 0.56s time: 118.69s eta: 11:38:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31011.0 tgs: 260 data_time: 0.77s time: 119.21s eta: 11:39:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29561.0 tgs: 248 data_time: 0.72s time: 119.14s eta: 11:36:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31494.0 tgs: 264 data_time: 0.63s time: 119.08s eta: 11:34:37 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31656.0 tgs: 266 data_time: 0.92s time: 118.95s eta: 11:31:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29638.0 tgs: 249 data_time: 0.73s time: 118.91s eta: 11:29:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31840.0 tgs: 267 data_time: 0.75s time: 118.97s eta: 11:28:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.130 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31844.0 tgs: 266 data_time: 0.77s time: 119.28s eta: 11:27:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30935.0 tgs: 259 data_time: 0.77s time: 119.35s eta: 11:26:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31121.0 tgs: 261 data_time: 0.95s time: 118.88s eta: 11:21:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.124 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32180.0 tgs: 269 data_time: 0.88s time: 119.21s eta: 11:21:28 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 268 data_time: 0.65s time: 119.72s eta: 11:22:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31060.0 tgs: 260 data_time: 0.66s time: 119.11s eta: 11:16:57 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32327.0 tgs: 271 data_time: 0.66s time: 119.12s eta: 11:15:00 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 271 data_time: 0.54s time: 119.00s eta: 11:12:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30725.0 tgs: 257 data_time: 0.70s time: 119.09s eta: 11:10:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32418.0 tgs: 272 data_time: 0.58s time: 119.15s eta: 11:09:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.184 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32078.0 tgs: 269 data_time: 0.78s time: 118.97s eta: 11:06:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32268.0 tgs: 270 data_time: 0.68s time: 119.47s eta: 11:07:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 28614.0 tgs: 239 data_time: 0.81s time: 119.64s eta: 11:05:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30863.0 tgs: 258 data_time: 0.57s time: 119.30s eta: 11:02:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31397.0 tgs: 263 data_time: 0.77s time: 119.31s eta: 11:00:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32034.0 tgs: 269 data_time: 0.78s time: 118.82s eta: 10:55:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32023.0 tgs: 268 data_time: 0.75s time: 119.47s eta: 10:57:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.200 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31816.0 tgs: 266 data_time: 0.63s time: 119.36s eta: 10:54:30 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30798.0 tgs: 258 data_time: 0.88s time: 119.28s eta: 10:52:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.228 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31745.0 tgs: 266 data_time: 0.79s time: 119.25s eta: 10:49:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.75s time: 119.10s eta: 10:47:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.207 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31606.0 tgs: 264 data_time: 0.84s time: 119.53s eta: 10:47:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31943.0 tgs: 267 data_time: 0.63s time: 119.33s eta: 10:44:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31700.0 tgs: 265 data_time: 0.60s time: 119.49s eta: 10:43:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31264.0 tgs: 261 data_time: 0.67s time: 119.51s eta: 10:41:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.120 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31586.0 tgs: 265 data_time: 0.81s time: 118.99s eta: 10:36:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.123 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 262 data_time: 0.61s time: 119.22s eta: 10:35:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31391.0 tgs: 263 data_time: 0.65s time: 119.06s eta: 10:32:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31555.0 tgs: 265 data_time: 0.84s time: 118.78s eta: 10:29:33 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.202 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 271 data_time: 0.81s time: 119.86s eta: 10:33:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30132.0 tgs: 252 data_time: 0.58s time: 119.38s eta: 10:28:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31627.0 tgs: 266 data_time: 0.46s time: 118.83s eta: 10:23:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31233.0 tgs: 262 data_time: 0.66s time: 118.98s eta: 10:22:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31780.0 tgs: 266 data_time: 0.73s time: 119.29s eta: 10:22:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31590.0 tgs: 265 data_time: 0.70s time: 119.20s eta: 10:19:50 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.120 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31757.0 tgs: 266 data_time: 0.48s time: 119.07s eta: 10:17:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.190 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 264 data_time: 0.91s time: 119.48s eta: 10:17:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31016.0 tgs: 259 data_time: 1.03s time: 119.41s eta: 10:14:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.122 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31543.0 tgs: 264 data_time: 0.61s time: 119.44s eta: 10:13:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30855.0 tgs: 258 data_time: 0.88s time: 119.29s eta: 10:10:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31401.0 tgs: 263 data_time: 0.81s time: 119.11s eta: 10:07:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.108 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31887.0 tgs: 266 data_time: 0.68s time: 119.70s eta: 10:08:28 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 269 data_time: 0.86s time: 119.83s eta: 10:07:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31143.0 tgs: 256 data_time: 0.76s time: 121.56s eta: 10:13:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31651.0 tgs: 265 data_time: 0.71s time: 119.25s eta: 10:00:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31551.0 tgs: 265 data_time: 0.73s time: 118.80s eta: 9:55:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.223 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31206.0 tgs: 260 data_time: 0.68s time: 119.57s eta: 9:57:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31674.0 tgs: 264 data_time: 0.80s time: 119.70s eta: 9:56:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31191.0 tgs: 261 data_time: 0.83s time: 119.26s eta: 9:52:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31542.0 tgs: 264 data_time: 0.83s time: 119.37s eta: 9:50:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.194 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32584.0 tgs: 273 data_time: 0.84s time: 119.19s eta: 9:47:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.217 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 265 data_time: 0.84s time: 120.87s eta: 9:54:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32050.0 tgs: 268 data_time: 0.65s time: 119.27s eta: 9:44:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30911.0 tgs: 258 data_time: 0.57s time: 119.46s eta: 9:43:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31939.0 tgs: 267 data_time: 0.79s time: 119.54s eta: 9:41:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 270 data_time: 0.66s time: 119.20s eta: 9:38:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31928.0 tgs: 267 data_time: 0.62s time: 119.48s eta: 9:37:28 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30782.0 tgs: 258 data_time: 0.78s time: 119.05s eta: 9:33:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.130 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31259.0 tgs: 262 data_time: 0.67s time: 119.24s eta: 9:32:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.67s time: 119.16s eta: 9:29:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.209 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.64s time: 119.22s eta: 9:28:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.208 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.73s time: 119.46s eta: 9:27:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31337.0 tgs: 263 data_time: 0.85s time: 118.99s eta: 9:23:12 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31996.0 tgs: 268 data_time: 0.64s time: 119.03s eta: 9:21:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 263 data_time: 0.73s time: 119.53s eta: 9:21:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31259.0 tgs: 261 data_time: 0.66s time: 119.59s eta: 9:20:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30922.0 tgs: 259 data_time: 0.55s time: 119.12s eta: 9:15:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31663.0 tgs: 265 data_time: 0.74s time: 119.06s eta: 9:13:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32492.0 tgs: 273 data_time: 0.84s time: 118.88s eta: 9:10:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30956.0 tgs: 259 data_time: 0.88s time: 119.32s eta: 9:10:51 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32383.0 tgs: 271 data_time: 0.89s time: 119.09s eta: 9:07:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 270 data_time: 0.72s time: 118.86s eta: 9:04:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31568.0 tgs: 263 data_time: 0.81s time: 119.67s eta: 9:06:28 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.102 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.56s time: 119.43s eta: 9:03:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.113 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31643.0 tgs: 264 data_time: 0.65s time: 119.43s eta: 9:01:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31603.0 tgs: 266 data_time: 0.91s time: 118.74s eta: 8:56:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32388.0 tgs: 272 data_time: 1.02s time: 118.86s eta: 8:54:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31589.0 tgs: 264 data_time: 0.92s time: 119.33s eta: 8:55:00 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.181 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 269 data_time: 0.57s time: 119.14s eta: 8:52:10 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31053.0 tgs: 260 data_time: 0.81s time: 119.27s eta: 8:50:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32572.0 tgs: 270 data_time: 0.54s time: 120.46s eta: 8:54:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31098.0 tgs: 260 data_time: 0.70s time: 119.24s eta: 8:46:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31537.0 tgs: 264 data_time: 0.70s time: 119.09s eta: 8:43:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.231 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31417.0 tgs: 260 data_time: 0.69s time: 120.83s eta: 8:49:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31208.0 tgs: 259 data_time: 0.72s time: 120.19s eta: 8:44:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31058.0 tgs: 261 data_time: 0.81s time: 118.83s eta: 8:36:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.101 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31887.0 tgs: 267 data_time: 0.68s time: 119.35s eta: 8:37:10 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30486.0 tgs: 255 data_time: 0.47s time: 119.33s eta: 8:35:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.192 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 270 data_time: 0.80s time: 119.05s eta: 8:31:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32490.0 tgs: 272 data_time: 0.81s time: 119.16s eta: 8:30:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31594.0 tgs: 265 data_time: 0.94s time: 119.04s eta: 8:27:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.237 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 269 data_time: 0.91s time: 119.12s eta: 8:26:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.246 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32424.0 tgs: 271 data_time: 0.82s time: 119.24s eta: 8:24:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31633.0 tgs: 265 data_time: 0.87s time: 119.22s eta: 8:22:42 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.215 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31017.0 tgs: 259 data_time: 0.80s time: 119.38s eta: 8:21:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.205 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31390.0 tgs: 262 data_time: 0.73s time: 119.77s eta: 8:21:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 267 data_time: 0.65s time: 119.19s eta: 8:16:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.178 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30736.0 tgs: 258 data_time: 0.64s time: 119.02s eta: 8:13:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30698.0 tgs: 256 data_time: 0.69s time: 119.89s eta: 8:15:33 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.198 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31243.0 tgs: 261 data_time: 0.70s time: 119.55s eta: 8:12:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32271.0 tgs: 269 data_time: 0.99s time: 119.59s eta: 8:10:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32314.0 tgs: 271 data_time: 0.77s time: 118.99s eta: 8:05:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.113 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 28318.0 tgs: 236 data_time: 0.41s time: 119.84s eta: 8:07:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31641.0 tgs: 264 data_time: 0.89s time: 119.71s eta: 8:04:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30889.0 tgs: 258 data_time: 0.67s time: 119.40s eta: 8:01:35 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.129 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.56s time: 119.29s eta: 7:59:07 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31307.0 tgs: 262 data_time: 0.54s time: 119.19s eta: 7:56:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.193 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32259.0 tgs: 270 data_time: 0.94s time: 119.20s eta: 7:54:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31809.0 tgs: 267 data_time: 0.76s time: 118.94s eta: 7:51:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31990.0 tgs: 268 data_time: 0.72s time: 118.97s eta: 7:49:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32002.0 tgs: 267 data_time: 1.04s time: 119.42s eta: 7:49:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31894.0 tgs: 267 data_time: 0.54s time: 119.03s eta: 7:46:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31251.0 tgs: 262 data_time: 0.66s time: 119.10s eta: 7:44:28 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.332 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 266 data_time: 1.05s time: 120.02s eta: 7:46:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 261 data_time: 0.51s time: 119.82s eta: 7:43:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32246.0 tgs: 270 data_time: 0.78s time: 119.17s eta: 7:38:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31637.0 tgs: 265 data_time: 0.78s time: 118.96s eta: 7:36:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31956.0 tgs: 267 data_time: 0.93s time: 119.32s eta: 7:35:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 267 data_time: 0.84s time: 119.38s eta: 7:33:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31790.0 tgs: 267 data_time: 0.58s time: 119.06s eta: 7:30:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.177 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31339.0 tgs: 263 data_time: 0.66s time: 118.99s eta: 7:28:12 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 270 data_time: 0.52s time: 119.24s eta: 7:27:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.200 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30939.0 tgs: 260 data_time: 0.69s time: 118.98s eta: 7:24:12 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31132.0 tgs: 261 data_time: 0.62s time: 118.93s eta: 7:22:00 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31860.0 tgs: 267 data_time: 0.70s time: 119.08s eta: 7:20:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 270 data_time: 0.93s time: 118.99s eta: 7:18:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31250.0 tgs: 262 data_time: 0.65s time: 119.02s eta: 7:16:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32154.0 tgs: 269 data_time: 0.78s time: 119.26s eta: 7:15:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32447.0 tgs: 271 data_time: 0.71s time: 119.36s eta: 7:13:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32230.0 tgs: 270 data_time: 0.66s time: 118.94s eta: 7:10:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30588.0 tgs: 256 data_time: 0.91s time: 119.20s eta: 7:09:07 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31231.0 tgs: 261 data_time: 0.45s time: 119.58s eta: 7:08:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32108.0 tgs: 268 data_time: 0.77s time: 119.66s eta: 7:06:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.195 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31695.0 tgs: 266 data_time: 0.81s time: 118.95s eta: 7:02:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31556.0 tgs: 264 data_time: 0.84s time: 119.30s eta: 7:01:30 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30884.0 tgs: 258 data_time: 0.65s time: 119.52s eta: 7:00:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31425.0 tgs: 263 data_time: 0.61s time: 119.42s eta: 6:57:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31590.0 tgs: 265 data_time: 0.85s time: 118.91s eta: 6:54:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30873.0 tgs: 258 data_time: 0.86s time: 119.34s eta: 6:53:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31537.0 tgs: 265 data_time: 0.53s time: 118.94s eta: 6:50:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31960.0 tgs: 267 data_time: 0.68s time: 119.27s eta: 6:49:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32189.0 tgs: 270 data_time: 0.75s time: 119.02s eta: 6:46:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31631.0 tgs: 265 data_time: 0.97s time: 119.23s eta: 6:45:22 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31928.0 tgs: 267 data_time: 0.94s time: 119.42s eta: 6:44:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29844.0 tgs: 250 data_time: 0.60s time: 119.01s eta: 6:40:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32275.0 tgs: 271 data_time: 0.72s time: 119.01s eta: 6:38:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.207 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31913.0 tgs: 267 data_time: 0.64s time: 119.48s eta: 6:38:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32050.0 tgs: 269 data_time: 1.02s time: 118.79s eta: 6:33:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29848.0 tgs: 249 data_time: 0.78s time: 119.43s eta: 6:34:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31937.0 tgs: 267 data_time: 0.61s time: 119.57s eta: 6:32:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31347.0 tgs: 262 data_time: 0.87s time: 119.38s eta: 6:29:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.176 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31866.0 tgs: 266 data_time: 0.77s time: 119.46s eta: 6:28:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32094.0 tgs: 268 data_time: 0.63s time: 119.36s eta: 6:25:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30884.0 tgs: 259 data_time: 0.76s time: 118.82s eta: 6:22:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.188 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.93s time: 119.02s eta: 6:20:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.191 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 269 data_time: 0.91s time: 119.39s eta: 6:20:04 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32100.0 tgs: 268 data_time: 0.72s time: 119.34s eta: 6:17:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31565.0 tgs: 265 data_time: 0.86s time: 119.05s eta: 6:15:00 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32064.0 tgs: 268 data_time: 0.67s time: 119.43s eta: 6:14:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31493.0 tgs: 263 data_time: 0.65s time: 119.39s eta: 6:12:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.108 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31174.0 tgs: 261 data_time: 1.08s time: 119.28s eta: 6:09:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31020.0 tgs: 260 data_time: 0.87s time: 119.10s eta: 6:07:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31816.0 tgs: 260 data_time: 0.85s time: 121.96s eta: 6:14:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31623.0 tgs: 264 data_time: 0.58s time: 119.38s eta: 6:04:07 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30566.0 tgs: 255 data_time: 0.47s time: 119.57s eta: 6:02:41 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31584.0 tgs: 264 data_time: 0.74s time: 119.25s eta: 5:59:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.225 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31067.0 tgs: 260 data_time: 0.78s time: 119.26s eta: 5:57:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.179 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31123.0 tgs: 261 data_time: 0.75s time: 118.91s eta: 5:54:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31012.0 tgs: 260 data_time: 0.77s time: 119.02s eta: 5:53:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30400.0 tgs: 255 data_time: 0.99s time: 119.07s eta: 5:51:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31215.0 tgs: 261 data_time: 0.59s time: 119.29s eta: 5:49:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31577.0 tgs: 265 data_time: 0.84s time: 118.97s eta: 5:46:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31521.0 tgs: 264 data_time: 0.78s time: 119.27s eta: 5:45:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31714.0 tgs: 265 data_time: 0.80s time: 119.36s eta: 5:44:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 271 data_time: 0.87s time: 119.07s eta: 5:41:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31618.0 tgs: 264 data_time: 0.63s time: 119.48s eta: 5:40:31 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30235.0 tgs: 253 data_time: 0.65s time: 119.38s eta: 5:38:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31540.0 tgs: 263 data_time: 0.66s time: 119.57s eta: 5:36:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31636.0 tgs: 265 data_time: 0.74s time: 118.98s eta: 5:33:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31280.0 tgs: 262 data_time: 0.68s time: 118.94s eta: 5:31:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.173 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32072.0 tgs: 268 data_time: 0.99s time: 119.24s eta: 5:29:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 265 data_time: 0.83s time: 119.40s eta: 5:28:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30192.0 tgs: 252 data_time: 0.67s time: 119.46s eta: 5:26:31 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32247.0 tgs: 271 data_time: 0.90s time: 118.95s eta: 5:23:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31291.0 tgs: 261 data_time: 0.61s time: 119.63s eta: 5:23:00 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32291.0 tgs: 270 data_time: 0.69s time: 119.35s eta: 5:20:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31019.0 tgs: 260 data_time: 0.78s time: 119.12s eta: 5:17:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.200 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32518.0 tgs: 272 data_time: 0.88s time: 119.42s eta: 5:16:28 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31890.0 tgs: 266 data_time: 0.76s time: 119.57s eta: 5:14:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.186 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.73s time: 119.00s eta: 5:11:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31840.0 tgs: 267 data_time: 0.65s time: 119.00s eta: 5:09:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.102 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31946.0 tgs: 267 data_time: 0.71s time: 119.22s eta: 5:07:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30229.0 tgs: 252 data_time: 0.76s time: 119.68s eta: 5:07:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32233.0 tgs: 270 data_time: 0.97s time: 119.11s eta: 5:03:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31286.0 tgs: 262 data_time: 0.78s time: 119.07s eta: 5:01:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 264 data_time: 0.64s time: 121.52s eta: 5:05:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31912.0 tgs: 266 data_time: 0.69s time: 119.59s eta: 4:58:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30824.0 tgs: 257 data_time: 0.61s time: 119.48s eta: 4:56:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.217 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31521.0 tgs: 263 data_time: 0.87s time: 119.73s eta: 4:55:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31299.0 tgs: 261 data_time: 0.82s time: 119.52s eta: 4:52:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 268 data_time: 0.54s time: 119.41s eta: 4:50:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 269 data_time: 0.63s time: 119.21s eta: 4:48:04 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31419.0 tgs: 264 data_time: 0.75s time: 118.88s eta: 4:45:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31396.0 tgs: 263 data_time: 0.88s time: 119.12s eta: 4:43:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.229 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 29738.0 tgs: 247 data_time: 0.78s time: 119.98s eta: 4:43:57 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32150.0 tgs: 269 data_time: 0.55s time: 119.15s eta: 4:39:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.192 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32004.0 tgs: 268 data_time: 0.91s time: 119.00s eta: 4:37:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31623.0 tgs: 265 data_time: 0.96s time: 119.06s eta: 4:35:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31443.0 tgs: 263 data_time: 0.76s time: 119.37s eta: 4:34:33 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.81s time: 119.41s eta: 4:32:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 265 data_time: 0.92s time: 118.94s eta: 4:29:35 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.189 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32027.0 tgs: 267 data_time: 0.88s time: 119.60s eta: 4:29:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 268 data_time: 0.69s time: 119.24s eta: 4:26:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 268 data_time: 0.63s time: 119.06s eta: 4:23:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31343.0 tgs: 263 data_time: 0.70s time: 118.85s eta: 4:21:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31979.0 tgs: 269 data_time: 0.61s time: 118.76s eta: 4:19:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30556.0 tgs: 255 data_time: 0.56s time: 119.42s eta: 4:18:44 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.191 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31444.0 tgs: 263 data_time: 0.57s time: 119.40s eta: 4:16:42 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31201.0 tgs: 261 data_time: 0.58s time: 119.34s eta: 4:14:35 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31909.0 tgs: 268 data_time: 0.68s time: 119.05s eta: 4:11:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.215 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31029.0 tgs: 260 data_time: 0.71s time: 119.12s eta: 4:10:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.205 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32024.0 tgs: 268 data_time: 0.48s time: 119.14s eta: 4:08:12 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31634.0 tgs: 265 data_time: 0.45s time: 118.96s eta: 4:05:50 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32106.0 tgs: 269 data_time: 0.79s time: 119.15s eta: 4:04:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31552.0 tgs: 264 data_time: 0.88s time: 119.49s eta: 4:02:58 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31367.0 tgs: 263 data_time: 0.84s time: 119.24s eta: 4:00:28 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.221 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31019.0 tgs: 260 data_time: 1.05s time: 119.19s eta: 3:58:22 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31415.0 tgs: 263 data_time: 0.80s time: 119.36s eta: 3:56:44 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31542.0 tgs: 264 data_time: 0.73s time: 119.36s eta: 3:54:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 27706.0 tgs: 230 data_time: 0.84s time: 119.95s eta: 3:53:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31183.0 tgs: 261 data_time: 0.63s time: 119.32s eta: 3:50:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31900.0 tgs: 266 data_time: 0.83s time: 119.63s eta: 3:49:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32139.0 tgs: 270 data_time: 0.68s time: 118.85s eta: 3:45:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32489.0 tgs: 271 data_time: 0.79s time: 119.75s eta: 3:45:32 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32322.0 tgs: 270 data_time: 0.89s time: 119.32s eta: 3:42:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.84s time: 119.26s eta: 3:40:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.099 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31400.0 tgs: 263 data_time: 0.80s time: 119.00s eta: 3:38:09 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.110 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30184.0 tgs: 252 data_time: 0.83s time: 119.43s eta: 3:36:57 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.106 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31710.0 tgs: 266 data_time: 0.93s time: 119.03s eta: 3:34:15 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31717.0 tgs: 266 data_time: 0.81s time: 118.99s eta: 3:32:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 270 data_time: 0.87s time: 119.22s eta: 3:30:37 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.114 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32287.0 tgs: 268 data_time: 0.70s time: 120.11s eta: 3:30:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32077.0 tgs: 269 data_time: 0.75s time: 119.13s eta: 3:26:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31788.0 tgs: 266 data_time: 0.91s time: 119.32s eta: 3:24:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32022.0 tgs: 268 data_time: 0.78s time: 119.08s eta: 3:22:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32164.0 tgs: 269 data_time: 0.72s time: 119.42s eta: 3:21:01 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 270 data_time: 0.86s time: 119.29s eta: 3:18:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31748.0 tgs: 266 data_time: 0.70s time: 119.21s eta: 3:16:42 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31104.0 tgs: 260 data_time: 0.87s time: 119.18s eta: 3:14:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.198 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31079.0 tgs: 261 data_time: 0.70s time: 119.03s eta: 3:12:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31602.0 tgs: 265 data_time: 0.80s time: 119.17s eta: 3:10:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31775.0 tgs: 262 data_time: 0.66s time: 121.05s eta: 3:11:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 266 data_time: 0.91s time: 119.50s eta: 3:07:13 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31718.0 tgs: 265 data_time: 0.71s time: 119.26s eta: 3:04:51 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:43:13][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31596.0 tgs: 265 data_time: 0.63s time: 118.99s eta: 3:02:27 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 269 data_time: 0.58s time: 119.22s eta: 3:00:48 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32035.0 tgs: 268 data_time: 0.61s time: 119.45s eta: 2:59:10 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31984.0 tgs: 268 data_time: 0.59s time: 118.98s eta: 2:56:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32579.0 tgs: 272 data_time: 0.81s time: 119.35s eta: 2:55:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31768.0 tgs: 266 data_time: 0.78s time: 119.41s eta: 2:53:08 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31188.0 tgs: 261 data_time: 0.69s time: 119.35s eta: 2:51:04 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32414.0 tgs: 266 data_time: 0.83s time: 121.46s eta: 2:52:03 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31817.0 tgs: 266 data_time: 0.70s time: 119.24s eta: 2:46:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31158.0 tgs: 260 data_time: 0.88s time: 119.56s eta: 2:45:23 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31497.0 tgs: 264 data_time: 0.67s time: 118.91s eta: 2:42:30 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31684.0 tgs: 265 data_time: 0.93s time: 119.18s eta: 2:40:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32351.0 tgs: 271 data_time: 0.75s time: 118.96s eta: 2:38:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 264 data_time: 0.75s time: 119.23s eta: 2:36:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31732.0 tgs: 266 data_time: 0.79s time: 119.11s eta: 2:34:50 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 267 data_time: 0.69s time: 119.43s eta: 2:33:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 267 data_time: 0.67s time: 120.22s eta: 2:32:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31722.0 tgs: 265 data_time: 0.75s time: 119.28s eta: 2:29:06 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31811.0 tgs: 267 data_time: 0.55s time: 118.88s eta: 2:26:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.221 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32018.0 tgs: 267 data_time: 0.84s time: 119.52s eta: 2:25:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31891.0 tgs: 267 data_time: 0.89s time: 119.06s eta: 2:22:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.87s time: 119.14s eta: 2:20:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31492.0 tgs: 264 data_time: 0.55s time: 119.04s eta: 2:18:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32193.0 tgs: 268 data_time: 1.06s time: 119.84s eta: 2:17:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30241.0 tgs: 253 data_time: 0.86s time: 119.36s eta: 2:15:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.81s time: 119.40s eta: 2:13:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31320.0 tgs: 261 data_time: 0.65s time: 119.67s eta: 2:11:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31387.0 tgs: 262 data_time: 0.84s time: 119.49s eta: 2:09:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31182.0 tgs: 262 data_time: 0.63s time: 118.95s eta: 2:06:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31870.0 tgs: 266 data_time: 0.90s time: 119.37s eta: 2:05:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.70s time: 119.27s eta: 2:03:14 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 266 data_time: 0.74s time: 119.22s eta: 2:01:12 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.240 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31403.0 tgs: 262 data_time: 0.73s time: 119.51s eta: 1:59:30 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29955.0 tgs: 250 data_time: 0.75s time: 119.42s eta: 1:57:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32458.0 tgs: 271 data_time: 1.02s time: 119.53s eta: 1:55:32 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30187.0 tgs: 253 data_time: 0.83s time: 119.30s eta: 1:53:20 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 268 data_time: 0.78s time: 119.25s eta: 1:51:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.224 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 269 data_time: 0.93s time: 118.95s eta: 1:49:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31144.0 tgs: 260 data_time: 0.66s time: 119.62s eta: 1:47:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32031.0 tgs: 267 data_time: 0.93s time: 119.72s eta: 1:45:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 268 data_time: 0.71s time: 119.35s eta: 1:43:26 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.59s time: 119.30s eta: 1:41:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31723.0 tgs: 265 data_time: 0.65s time: 119.28s eta: 1:39:24 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31003.0 tgs: 260 data_time: 0.66s time: 118.90s eta: 1:37:05 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.52s time: 119.16s eta: 1:35:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 268 data_time: 0.76s time: 119.06s eta: 1:33:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32244.0 tgs: 270 data_time: 0.50s time: 119.25s eta: 1:31:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 264 data_time: 0.57s time: 119.72s eta: 1:29:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 267 data_time: 0.77s time: 119.01s eta: 1:27:16 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 265 data_time: 0.56s time: 118.87s eta: 1:25:11 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.076 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31921.0 tgs: 267 data_time: 0.72s time: 119.48s eta: 1:23:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31051.0 tgs: 259 data_time: 0.60s time: 119.48s eta: 1:21:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32226.0 tgs: 270 data_time: 0.65s time: 118.94s eta: 1:19:17 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30584.0 tgs: 257 data_time: 0.67s time: 118.93s eta: 1:17:18 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 262 data_time: 0.63s time: 122.16s eta: 1:17:21 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31139.0 tgs: 260 data_time: 0.65s time: 119.37s eta: 1:13:36 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31366.0 tgs: 263 data_time: 0.67s time: 119.05s eta: 1:11:25 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31876.0 tgs: 268 data_time: 0.82s time: 118.92s eta: 1:09:22 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30867.0 tgs: 259 data_time: 0.74s time: 119.10s eta: 1:07:29 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31772.0 tgs: 266 data_time: 0.61s time: 119.41s eta: 1:05:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30137.0 tgs: 252 data_time: 0.44s time: 119.35s eta: 1:03:39 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32537.0 tgs: 273 data_time: 0.73s time: 119.17s eta: 1:01:34 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.096 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 26640.0 tgs: 223 data_time: 0.86s time: 119.28s eta: 0:59:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31320.0 tgs: 262 data_time: 0.71s time: 119.44s eta: 0:57:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31224.0 tgs: 261 data_time: 0.78s time: 119.39s eta: 0:55:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 265 data_time: 1.00s time: 119.48s eta: 0:53:45 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32255.0 tgs: 270 data_time: 0.53s time: 119.36s eta: 0:51:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30901.0 tgs: 258 data_time: 0.71s time: 119.47s eta: 0:49:46 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31317.0 tgs: 263 data_time: 0.96s time: 118.90s eta: 0:47:33 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29314.0 tgs: 245 data_time: 0.67s time: 119.28s eta: 0:45:43 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31589.0 tgs: 264 data_time: 0.75s time: 119.21s eta: 0:43:42 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.099 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32165.0 tgs: 270 data_time: 0.56s time: 118.99s eta: 0:41:38 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30580.0 tgs: 256 data_time: 0.51s time: 119.04s eta: 0:39:40 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32292.0 tgs: 269 data_time: 1.09s time: 119.66s eta: 0:37:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31410.0 tgs: 263 data_time: 0.58s time: 119.28s eta: 0:35:47 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30524.0 tgs: 254 data_time: 0.69s time: 119.74s eta: 0:33:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31685.0 tgs: 261 data_time: 0.74s time: 121.23s eta: 0:32:19 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31233.0 tgs: 261 data_time: 0.69s time: 119.37s eta: 0:29:50 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30549.0 tgs: 256 data_time: 0.86s time: 119.32s eta: 0:27:50 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.83s time: 119.63s eta: 0:25:55 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31744.0 tgs: 266 data_time: 0.77s time: 119.09s eta: 0:23:49 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30869.0 tgs: 258 data_time: 0.80s time: 119.39s eta: 0:21:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31257.0 tgs: 262 data_time: 0.83s time: 119.26s eta: 0:19:52 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31161.0 tgs: 261 data_time: 0.69s time: 119.24s eta: 0:17:53 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31976.0 tgs: 266 data_time: 0.51s time: 119.91s eta: 0:15:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30556.0 tgs: 256 data_time: 0.58s time: 119.21s eta: 0:13:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31361.0 tgs: 263 data_time: 0.74s time: 119.09s eta: 0:11:54 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 266 data_time: 0.86s time: 119.32s eta: 0:09:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32590.0 tgs: 273 data_time: 0.59s time: 119.17s eta: 0:07:56 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31100.0 tgs: 259 data_time: 0.64s time: 119.69s eta: 0:05:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 263 data_time: 0.54s time: 121.45s eta: 0:04:02 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31457.0 tgs: 263 data_time: 0.81s time: 119.21s eta: 0:01:59 +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.26 seconds, peak gpu memory 17.4G +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 27][DP 6][SP 3][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank28.log b/20250121165312/rank28.log new file mode 100644 index 0000000000000000000000000000000000000000..900af48c8d0088395adcc6930097f20fd7951525 --- /dev/null +++ b/20250121165312/rank28.log @@ -0,0 +1,851 @@ +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.09s +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.57 seconds, peak gpu memory 13.4G +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:04:45][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.196 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 32271.0 tgs: 210 data_time: 1.75s time: 153.21s eta: 1 day, 1:14:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.255 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32224.0 tgs: 270 data_time: 1.02s time: 119.17s eta: 19:35:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.187 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31186.0 tgs: 260 data_time: 1.00s time: 119.68s eta: 19:38:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.234 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32092.0 tgs: 268 data_time: 0.86s time: 119.65s eta: 19:36:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.295 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 265 data_time: 0.87s time: 119.31s eta: 19:31:11 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31263.0 tgs: 261 data_time: 0.77s time: 119.53s eta: 19:31:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30097.0 tgs: 244 data_time: 0.86s time: 122.96s eta: 20:02:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.219 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 270 data_time: 0.91s time: 119.89s eta: 19:30:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30582.0 tgs: 253 data_time: 0.85s time: 120.65s eta: 19:36:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32372.0 tgs: 269 data_time: 0.82s time: 119.92s eta: 19:27:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 267 data_time: 0.72s time: 120.46s eta: 19:30:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31025.0 tgs: 259 data_time: 0.67s time: 119.34s eta: 19:17:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31804.0 tgs: 265 data_time: 0.94s time: 119.68s eta: 19:18:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32389.0 tgs: 271 data_time: 0.91s time: 119.18s eta: 19:12:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30946.0 tgs: 259 data_time: 0.91s time: 119.22s eta: 19:10:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31432.0 tgs: 264 data_time: 0.95s time: 118.93s eta: 19:05:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.257 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32194.0 tgs: 270 data_time: 0.77s time: 119.15s eta: 19:05:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31342.0 tgs: 262 data_time: 1.10s time: 119.18s eta: 19:04:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32054.0 tgs: 268 data_time: 0.71s time: 119.40s eta: 19:04:14 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.76s time: 119.32s eta: 19:01:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31497.0 tgs: 261 data_time: 0.64s time: 120.31s eta: 19:08:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32299.0 tgs: 262 data_time: 0.82s time: 122.91s eta: 19:31:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 268 data_time: 0.67s time: 119.48s eta: 18:57:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31131.0 tgs: 261 data_time: 0.64s time: 119.15s eta: 18:51:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.79s time: 119.24s eta: 18:50:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32152.0 tgs: 269 data_time: 0.69s time: 119.40s eta: 18:50:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.267 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31336.0 tgs: 261 data_time: 1.03s time: 119.66s eta: 18:50:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31412.0 tgs: 263 data_time: 0.77s time: 119.39s eta: 18:46:14 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 268 data_time: 0.72s time: 119.78s eta: 18:47:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 245 data_time: 0.87s time: 129.33s eta: 20:15:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31765.0 tgs: 265 data_time: 0.89s time: 119.57s eta: 18:41:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31949.0 tgs: 267 data_time: 0.96s time: 119.39s eta: 18:38:14 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 259 data_time: 0.64s time: 120.65s eta: 18:48:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31518.0 tgs: 263 data_time: 0.97s time: 119.39s eta: 18:34:21 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 268 data_time: 0.94s time: 119.16s eta: 18:30:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32039.0 tgs: 269 data_time: 1.24s time: 119.07s eta: 18:27:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.82s time: 119.13s eta: 18:25:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32352.0 tgs: 268 data_time: 0.80s time: 120.39s eta: 18:35:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.126 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31385.0 tgs: 259 data_time: 1.05s time: 121.17s eta: 18:40:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 261 data_time: 0.72s time: 120.66s eta: 18:34:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31223.0 tgs: 260 data_time: 0.77s time: 119.73s eta: 18:23:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32577.0 tgs: 262 data_time: 1.12s time: 123.88s eta: 18:59:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31503.0 tgs: 263 data_time: 0.79s time: 119.56s eta: 18:17:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 263 data_time: 0.78s time: 119.35s eta: 18:14:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31080.0 tgs: 260 data_time: 0.86s time: 119.22s eta: 18:10:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31762.0 tgs: 265 data_time: 0.77s time: 119.60s eta: 18:12:21 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31859.0 tgs: 267 data_time: 0.87s time: 119.26s eta: 18:07:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32052.0 tgs: 269 data_time: 0.64s time: 119.03s eta: 18:03:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31795.0 tgs: 266 data_time: 0.80s time: 119.10s eta: 18:01:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31424.0 tgs: 263 data_time: 0.78s time: 119.26s eta: 18:01:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.73s time: 119.33s eta: 17:59:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32221.0 tgs: 268 data_time: 0.93s time: 120.02s eta: 18:04:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31787.0 tgs: 266 data_time: 0.89s time: 119.10s eta: 17:53:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31727.0 tgs: 265 data_time: 0.69s time: 119.41s eta: 17:54:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31831.0 tgs: 267 data_time: 0.93s time: 118.99s eta: 17:48:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 269 data_time: 0.96s time: 119.03s eta: 17:47:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31492.0 tgs: 251 data_time: 0.84s time: 125.36s eta: 18:41:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.125 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31059.0 tgs: 258 data_time: 0.70s time: 120.10s eta: 17:52:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30232.0 tgs: 253 data_time: 0.80s time: 119.40s eta: 17:44:37 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.114 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31767.0 tgs: 266 data_time: 0.84s time: 119.34s eta: 17:42:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.62s time: 119.51s eta: 17:41:37 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 255 data_time: 0.68s time: 126.05s eta: 18:37:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31753.0 tgs: 258 data_time: 0.44s time: 122.63s eta: 18:05:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31116.0 tgs: 256 data_time: 0.71s time: 121.16s eta: 17:50:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31233.0 tgs: 262 data_time: 0.69s time: 119.03s eta: 17:29:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.103 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32074.0 tgs: 269 data_time: 0.76s time: 118.85s eta: 17:25:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 256 data_time: 0.85s time: 124.00s eta: 18:09:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32008.0 tgs: 256 data_time: 0.73s time: 124.56s eta: 18:11:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32207.0 tgs: 269 data_time: 0.80s time: 119.41s eta: 17:24:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31813.0 tgs: 265 data_time: 0.68s time: 119.69s eta: 17:25:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31610.0 tgs: 264 data_time: 0.66s time: 119.52s eta: 17:21:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 262 data_time: 0.72s time: 118.86s eta: 17:14:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 265 data_time: 1.04s time: 119.85s eta: 17:20:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31473.0 tgs: 263 data_time: 0.93s time: 119.56s eta: 17:16:11 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.125 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32047.0 tgs: 262 data_time: 0.78s time: 121.88s eta: 17:34:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29030.0 tgs: 235 data_time: 0.61s time: 123.50s eta: 17:46:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.75s time: 119.01s eta: 17:05:27 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.85s time: 119.24s eta: 17:05:27 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.131 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32013.0 tgs: 268 data_time: 0.59s time: 119.13s eta: 17:02:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30187.0 tgs: 253 data_time: 0.76s time: 119.25s eta: 17:01:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31309.0 tgs: 263 data_time: 0.81s time: 119.04s eta: 16:57:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 266 data_time: 0.78s time: 119.28s eta: 16:57:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31701.0 tgs: 266 data_time: 0.60s time: 119.16s eta: 16:54:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 269 data_time: 0.67s time: 119.58s eta: 16:56:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.254 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31201.0 tgs: 261 data_time: 0.70s time: 119.13s eta: 16:50:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31038.0 tgs: 261 data_time: 0.93s time: 118.84s eta: 16:46:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31025.0 tgs: 260 data_time: 0.84s time: 119.01s eta: 16:45:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30099.0 tgs: 251 data_time: 0.87s time: 119.85s eta: 16:50:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.96s time: 119.38s eta: 16:44:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31723.0 tgs: 266 data_time: 0.83s time: 119.01s eta: 16:39:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31759.0 tgs: 262 data_time: 0.89s time: 120.92s eta: 16:53:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30751.0 tgs: 258 data_time: 0.59s time: 118.99s eta: 16:35:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31819.0 tgs: 266 data_time: 0.65s time: 119.28s eta: 16:36:01 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31779.0 tgs: 266 data_time: 0.72s time: 119.23s eta: 16:33:33 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31279.0 tgs: 262 data_time: 0.77s time: 119.38s eta: 16:32:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31568.0 tgs: 263 data_time: 0.83s time: 119.69s eta: 16:33:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.212 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31989.0 tgs: 268 data_time: 0.86s time: 119.20s eta: 16:27:22 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31654.0 tgs: 264 data_time: 0.67s time: 119.77s eta: 16:30:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31764.0 tgs: 264 data_time: 0.78s time: 120.22s eta: 16:31:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31774.0 tgs: 267 data_time: 0.71s time: 118.98s eta: 16:19:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31673.0 tgs: 248 data_time: 0.72s time: 127.33s eta: 17:26:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.91s time: 119.22s eta: 16:17:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.116 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.95s time: 119.14s eta: 16:14:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 259 data_time: 0.54s time: 119.25s eta: 16:13:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31441.0 tgs: 263 data_time: 0.79s time: 119.18s eta: 16:11:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 270 data_time: 0.92s time: 118.95s eta: 16:07:27 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32186.0 tgs: 270 data_time: 0.95s time: 118.86s eta: 16:04:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31747.0 tgs: 266 data_time: 0.75s time: 118.96s eta: 16:03:33 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31339.0 tgs: 257 data_time: 0.72s time: 121.84s eta: 16:24:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.63s time: 118.99s eta: 15:59:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 269 data_time: 0.88s time: 119.23s eta: 15:59:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.227 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31503.0 tgs: 263 data_time: 1.06s time: 119.61s eta: 16:00:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31520.0 tgs: 264 data_time: 0.75s time: 119.04s eta: 15:54:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31714.0 tgs: 266 data_time: 0.79s time: 118.98s eta: 15:51:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31219.0 tgs: 261 data_time: 0.98s time: 119.17s eta: 15:51:24 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.240 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31548.0 tgs: 263 data_time: 0.73s time: 119.64s eta: 15:53:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 269 data_time: 0.92s time: 118.93s eta: 15:45:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31262.0 tgs: 262 data_time: 0.90s time: 119.28s eta: 15:46:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.112 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31993.0 tgs: 268 data_time: 0.61s time: 119.13s eta: 15:43:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31357.0 tgs: 263 data_time: 0.89s time: 119.00s eta: 15:40:06 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31058.0 tgs: 259 data_time: 0.87s time: 119.65s eta: 15:43:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32204.0 tgs: 265 data_time: 0.72s time: 121.50s eta: 15:55:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32363.0 tgs: 271 data_time: 0.67s time: 119.42s eta: 15:37:24 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 242 data_time: 0.78s time: 129.96s eta: 16:58:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31435.0 tgs: 263 data_time: 0.92s time: 119.47s eta: 15:33:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 270 data_time: 0.80s time: 119.60s eta: 15:32:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.119 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 266 data_time: 0.85s time: 119.53s eta: 15:30:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 259 data_time: 0.72s time: 122.14s eta: 15:48:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.267 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30593.0 tgs: 253 data_time: 0.87s time: 120.77s eta: 15:35:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31383.0 tgs: 262 data_time: 0.69s time: 119.40s eta: 15:23:21 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31513.0 tgs: 263 data_time: 0.58s time: 119.45s eta: 15:21:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32587.0 tgs: 247 data_time: 0.95s time: 131.63s eta: 16:53:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 264 data_time: 0.65s time: 119.29s eta: 15:16:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30622.0 tgs: 256 data_time: 0.74s time: 119.57s eta: 15:16:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30873.0 tgs: 252 data_time: 1.14s time: 122.26s eta: 15:35:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31673.0 tgs: 264 data_time: 0.61s time: 119.72s eta: 15:13:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.243 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30890.0 tgs: 258 data_time: 0.90s time: 119.34s eta: 15:08:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.241 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31272.0 tgs: 262 data_time: 0.78s time: 118.91s eta: 15:03:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32168.0 tgs: 270 data_time: 0.97s time: 119.00s eta: 15:02:24 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32521.0 tgs: 272 data_time: 0.80s time: 119.30s eta: 15:02:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.209 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.87s time: 118.93s eta: 14:57:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 263 data_time: 0.74s time: 120.51s eta: 15:07:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.099 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31767.0 tgs: 266 data_time: 0.72s time: 119.31s eta: 14:56:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31769.0 tgs: 258 data_time: 1.09s time: 122.66s eta: 15:19:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32206.0 tgs: 270 data_time: 0.83s time: 119.13s eta: 14:51:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30760.0 tgs: 256 data_time: 0.96s time: 119.75s eta: 14:54:06 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.208 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32005.0 tgs: 267 data_time: 0.57s time: 119.47s eta: 14:50:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31508.0 tgs: 264 data_time: 0.74s time: 118.99s eta: 14:44:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31893.0 tgs: 267 data_time: 0.71s time: 119.32s eta: 14:44:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31235.0 tgs: 262 data_time: 0.55s time: 118.95s eta: 14:40:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31995.0 tgs: 268 data_time: 0.71s time: 119.05s eta: 14:38:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32011.0 tgs: 268 data_time: 0.79s time: 119.06s eta: 14:37:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30863.0 tgs: 259 data_time: 0.67s time: 119.13s eta: 14:35:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.188 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30213.0 tgs: 253 data_time: 0.70s time: 119.00s eta: 14:32:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31367.0 tgs: 263 data_time: 0.88s time: 119.07s eta: 14:31:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31269.0 tgs: 262 data_time: 0.71s time: 119.15s eta: 14:29:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31668.0 tgs: 264 data_time: 0.66s time: 119.65s eta: 14:31:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.196 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31958.0 tgs: 268 data_time: 0.72s time: 119.17s eta: 14:25:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 269 data_time: 0.77s time: 119.20s eta: 14:24:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.82s time: 119.39s eta: 14:23:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31959.0 tgs: 264 data_time: 0.86s time: 120.73s eta: 14:31:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.194 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32334.0 tgs: 272 data_time: 0.73s time: 118.85s eta: 14:15:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31346.0 tgs: 263 data_time: 0.92s time: 119.10s eta: 14:15:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31361.0 tgs: 263 data_time: 0.69s time: 119.14s eta: 14:13:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31668.0 tgs: 263 data_time: 1.02s time: 119.97s eta: 14:17:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.208 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31098.0 tgs: 259 data_time: 0.82s time: 119.87s eta: 14:15:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30925.0 tgs: 258 data_time: 0.56s time: 119.58s eta: 14:10:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32117.0 tgs: 269 data_time: 0.65s time: 119.22s eta: 14:06:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31798.0 tgs: 266 data_time: 0.78s time: 119.27s eta: 14:04:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.77s time: 119.26s eta: 14:02:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32114.0 tgs: 269 data_time: 0.93s time: 119.14s eta: 13:59:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31630.0 tgs: 265 data_time: 0.62s time: 119.12s eta: 13:57:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32387.0 tgs: 271 data_time: 0.98s time: 119.17s eta: 13:56:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 266 data_time: 0.68s time: 119.35s eta: 13:55:27 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31302.0 tgs: 260 data_time: 0.76s time: 120.11s eta: 13:58:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.276 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31527.0 tgs: 257 data_time: 0.63s time: 122.55s eta: 14:13:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:56:46][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32428.0 tgs: 272 data_time: 0.76s time: 119.10s eta: 13:47:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31703.0 tgs: 265 data_time: 0.64s time: 119.38s eta: 13:47:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.220 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.97s time: 119.34s eta: 13:45:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.214 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31817.0 tgs: 266 data_time: 0.73s time: 119.35s eta: 13:43:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.214 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31491.0 tgs: 264 data_time: 0.57s time: 119.12s eta: 13:39:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31439.0 tgs: 263 data_time: 0.93s time: 119.25s eta: 13:38:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.202 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 1.03s time: 119.13s eta: 13:36:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31671.0 tgs: 265 data_time: 0.92s time: 119.23s eta: 13:34:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30523.0 tgs: 255 data_time: 0.91s time: 119.45s eta: 13:34:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.115 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31916.0 tgs: 268 data_time: 0.65s time: 119.08s eta: 13:29:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31049.0 tgs: 261 data_time: 0.43s time: 118.92s eta: 13:26:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31921.0 tgs: 266 data_time: 0.52s time: 119.80s eta: 13:30:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 268 data_time: 0.85s time: 119.25s eta: 13:24:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.91s time: 119.17s eta: 13:22:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32195.0 tgs: 269 data_time: 0.83s time: 119.34s eta: 13:21:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31241.0 tgs: 261 data_time: 0.89s time: 119.26s eta: 13:19:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 270 data_time: 0.76s time: 119.04s eta: 13:15:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.81s time: 119.37s eta: 13:15:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.212 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30896.0 tgs: 259 data_time: 0.85s time: 119.17s eta: 13:12:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.121 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 269 data_time: 0.82s time: 119.46s eta: 13:12:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30668.0 tgs: 257 data_time: 0.88s time: 119.29s eta: 13:09:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.192 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 264 data_time: 0.70s time: 119.15s eta: 13:06:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 265 data_time: 0.73s time: 120.25s eta: 13:11:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.201 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 265 data_time: 0.83s time: 120.16s eta: 13:09:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 269 data_time: 0.75s time: 119.50s eta: 13:02:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27721.0 tgs: 231 data_time: 0.93s time: 119.51s eta: 13:00:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31795.0 tgs: 265 data_time: 0.86s time: 119.59s eta: 12:59:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31254.0 tgs: 260 data_time: 0.64s time: 119.78s eta: 12:58:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31574.0 tgs: 265 data_time: 0.77s time: 119.05s eta: 12:51:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31155.0 tgs: 261 data_time: 0.83s time: 119.05s eta: 12:49:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31056.0 tgs: 260 data_time: 0.61s time: 119.12s eta: 12:48:20 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.196 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32326.0 tgs: 270 data_time: 0.84s time: 119.38s eta: 12:48:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.214 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 264 data_time: 0.74s time: 119.20s eta: 12:44:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32123.0 tgs: 268 data_time: 0.95s time: 119.44s eta: 12:44:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31749.0 tgs: 266 data_time: 0.66s time: 119.09s eta: 12:40:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30997.0 tgs: 260 data_time: 0.83s time: 119.07s eta: 12:38:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32055.0 tgs: 267 data_time: 0.79s time: 119.61s eta: 12:39:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 273 data_time: 0.77s time: 118.76s eta: 12:32:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 270 data_time: 0.77s time: 118.88s eta: 12:30:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31814.0 tgs: 266 data_time: 0.80s time: 119.17s eta: 12:30:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.204 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31922.0 tgs: 267 data_time: 0.84s time: 119.26s eta: 12:29:22 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.185 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32157.0 tgs: 269 data_time: 0.60s time: 119.35s eta: 12:27:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31834.0 tgs: 267 data_time: 0.72s time: 119.12s eta: 12:24:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31074.0 tgs: 260 data_time: 0.62s time: 119.34s eta: 12:23:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32137.0 tgs: 269 data_time: 0.66s time: 119.25s eta: 12:21:21 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31326.0 tgs: 263 data_time: 0.85s time: 119.03s eta: 12:17:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31646.0 tgs: 261 data_time: 0.63s time: 121.13s eta: 12:28:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31682.0 tgs: 264 data_time: 0.83s time: 119.99s eta: 12:19:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.124 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31959.0 tgs: 269 data_time: 0.63s time: 118.72s eta: 12:10:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30787.0 tgs: 257 data_time: 0.81s time: 119.63s eta: 12:13:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31823.0 tgs: 265 data_time: 0.76s time: 119.83s eta: 12:12:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31642.0 tgs: 265 data_time: 0.60s time: 119.15s eta: 12:06:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32138.0 tgs: 269 data_time: 0.67s time: 119.16s eta: 12:04:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32136.0 tgs: 269 data_time: 0.58s time: 119.12s eta: 12:02:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.204 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.74s time: 119.18s eta: 12:01:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 267 data_time: 0.73s time: 118.95s eta: 11:57:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 266 data_time: 0.64s time: 119.70s eta: 12:00:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 267 data_time: 0.85s time: 119.28s eta: 11:55:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 266 data_time: 0.91s time: 119.27s eta: 11:53:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.197 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31851.0 tgs: 266 data_time: 0.61s time: 119.70s eta: 11:54:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.238 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31892.0 tgs: 268 data_time: 0.86s time: 118.78s eta: 11:46:44 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.211 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31879.0 tgs: 267 data_time: 0.78s time: 119.15s eta: 11:46:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31023.0 tgs: 259 data_time: 0.75s time: 119.62s eta: 11:47:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.215 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32492.0 tgs: 273 data_time: 0.73s time: 118.88s eta: 11:41:22 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30868.0 tgs: 260 data_time: 0.64s time: 118.67s eta: 11:38:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31405.0 tgs: 263 data_time: 0.94s time: 119.21s eta: 11:39:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.232 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31589.0 tgs: 265 data_time: 0.97s time: 119.14s eta: 11:36:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 269 data_time: 0.70s time: 119.07s eta: 11:34:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31521.0 tgs: 264 data_time: 0.94s time: 118.95s eta: 11:31:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.202 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 269 data_time: 0.82s time: 118.91s eta: 11:29:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.130 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32398.0 tgs: 272 data_time: 0.81s time: 118.97s eta: 11:28:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31334.0 tgs: 262 data_time: 0.62s time: 119.28s eta: 11:27:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.121 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32549.0 tgs: 272 data_time: 0.73s time: 119.35s eta: 11:26:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32111.0 tgs: 270 data_time: 0.79s time: 118.88s eta: 11:21:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.76s time: 119.21s eta: 11:21:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.118 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 28070.0 tgs: 234 data_time: 0.69s time: 119.72s eta: 11:22:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 0.66s time: 119.13s eta: 11:17:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30675.0 tgs: 257 data_time: 0.63s time: 119.11s eta: 11:14:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32604.0 tgs: 274 data_time: 0.90s time: 118.99s eta: 11:12:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31034.0 tgs: 260 data_time: 0.71s time: 119.09s eta: 11:10:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 262 data_time: 0.63s time: 119.16s eta: 11:09:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.308 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 267 data_time: 0.79s time: 118.96s eta: 11:06:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32310.0 tgs: 270 data_time: 0.87s time: 119.47s eta: 11:07:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31737.0 tgs: 265 data_time: 0.97s time: 119.64s eta: 11:05:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 269 data_time: 0.54s time: 119.30s eta: 11:02:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 265 data_time: 0.98s time: 119.31s eta: 11:00:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.118 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31822.0 tgs: 267 data_time: 0.77s time: 118.82s eta: 10:55:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.124 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28741.0 tgs: 240 data_time: 0.73s time: 119.47s eta: 10:57:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31552.0 tgs: 264 data_time: 0.85s time: 119.37s eta: 10:54:31 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.098 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31005.0 tgs: 259 data_time: 0.56s time: 119.27s eta: 10:52:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.118 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31325.0 tgs: 262 data_time: 0.64s time: 119.26s eta: 10:49:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.184 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 270 data_time: 0.63s time: 119.09s eta: 10:47:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.184 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32367.0 tgs: 270 data_time: 0.85s time: 119.53s eta: 10:47:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31439.0 tgs: 263 data_time: 0.62s time: 119.33s eta: 10:44:24 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31650.0 tgs: 264 data_time: 1.03s time: 119.49s eta: 10:43:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32244.0 tgs: 269 data_time: 0.88s time: 119.51s eta: 10:41:21 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.194 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31044.0 tgs: 260 data_time: 1.05s time: 119.00s eta: 10:36:37 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.90s time: 119.22s eta: 10:35:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.182 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32149.0 tgs: 270 data_time: 0.74s time: 119.05s eta: 10:32:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31111.0 tgs: 261 data_time: 0.83s time: 118.78s eta: 10:29:31 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30613.0 tgs: 255 data_time: 0.81s time: 119.87s eta: 10:33:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31475.0 tgs: 263 data_time: 0.85s time: 119.37s eta: 10:28:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31903.0 tgs: 268 data_time: 0.91s time: 118.84s eta: 10:23:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.108 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31353.0 tgs: 263 data_time: 0.48s time: 118.99s eta: 10:22:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.186 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32109.0 tgs: 269 data_time: 0.65s time: 119.28s eta: 10:22:14 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31351.0 tgs: 263 data_time: 0.80s time: 119.21s eta: 10:19:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31577.0 tgs: 265 data_time: 1.00s time: 119.06s eta: 10:17:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32081.0 tgs: 268 data_time: 0.82s time: 119.47s eta: 10:17:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.116 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31521.0 tgs: 263 data_time: 0.88s time: 119.43s eta: 10:15:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32025.0 tgs: 268 data_time: 0.82s time: 119.43s eta: 10:13:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31915.0 tgs: 267 data_time: 0.77s time: 119.29s eta: 10:10:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.119 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31465.0 tgs: 264 data_time: 0.67s time: 119.10s eta: 10:07:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.115 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30904.0 tgs: 258 data_time: 0.71s time: 119.70s eta: 10:08:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.194 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32158.0 tgs: 268 data_time: 1.02s time: 119.83s eta: 10:07:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 263 data_time: 1.08s time: 121.55s eta: 10:13:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31631.0 tgs: 265 data_time: 0.72s time: 119.26s eta: 10:00:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.208 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31174.0 tgs: 262 data_time: 0.84s time: 118.80s eta: 9:55:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.191 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32127.0 tgs: 268 data_time: 0.85s time: 119.56s eta: 9:57:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31380.0 tgs: 262 data_time: 0.85s time: 119.70s eta: 9:56:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 262 data_time: 0.58s time: 119.26s eta: 9:52:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31612.0 tgs: 264 data_time: 0.77s time: 119.33s eta: 9:50:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32502.0 tgs: 272 data_time: 0.80s time: 119.22s eta: 9:48:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31153.0 tgs: 257 data_time: 0.70s time: 120.79s eta: 9:53:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.101 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31774.0 tgs: 266 data_time: 0.63s time: 119.27s eta: 9:44:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.115 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31461.0 tgs: 263 data_time: 0.64s time: 119.46s eta: 9:43:20 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31674.0 tgs: 264 data_time: 0.58s time: 119.54s eta: 9:41:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.105 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30423.0 tgs: 255 data_time: 0.70s time: 119.20s eta: 9:38:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.108 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 269 data_time: 0.63s time: 119.48s eta: 9:37:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.121 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31530.0 tgs: 264 data_time: 0.55s time: 119.05s eta: 9:33:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31115.0 tgs: 260 data_time: 0.68s time: 119.24s eta: 9:32:20 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.189 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 267 data_time: 0.78s time: 119.16s eta: 9:29:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32559.0 tgs: 273 data_time: 0.71s time: 119.22s eta: 9:28:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 266 data_time: 0.82s time: 119.46s eta: 9:27:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32349.0 tgs: 271 data_time: 0.73s time: 118.98s eta: 9:23:11 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 264 data_time: 0.70s time: 119.03s eta: 9:21:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31970.0 tgs: 267 data_time: 0.87s time: 119.53s eta: 9:21:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.243 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31238.0 tgs: 261 data_time: 0.74s time: 119.59s eta: 9:20:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.112 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30878.0 tgs: 259 data_time: 0.46s time: 119.13s eta: 9:15:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32095.0 tgs: 269 data_time: 0.99s time: 119.05s eta: 9:13:33 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.221 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30782.0 tgs: 258 data_time: 0.86s time: 118.89s eta: 9:10:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.97s time: 119.31s eta: 9:10:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31507.0 tgs: 264 data_time: 1.14s time: 119.09s eta: 9:07:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32289.0 tgs: 271 data_time: 0.76s time: 118.86s eta: 9:04:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 265 data_time: 0.75s time: 119.67s eta: 9:06:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 265 data_time: 0.85s time: 119.45s eta: 9:03:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.295 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31804.0 tgs: 266 data_time: 0.85s time: 119.42s eta: 9:01:21 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.190 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31150.0 tgs: 262 data_time: 0.80s time: 118.73s eta: 8:56:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31503.0 tgs: 265 data_time: 0.62s time: 118.87s eta: 8:54:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30717.0 tgs: 257 data_time: 0.73s time: 119.33s eta: 8:55:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 271 data_time: 0.95s time: 119.14s eta: 8:52:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 270 data_time: 0.79s time: 119.27s eta: 8:50:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 267 data_time: 0.59s time: 120.47s eta: 8:54:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32001.0 tgs: 268 data_time: 0.59s time: 119.23s eta: 8:46:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.121 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30960.0 tgs: 259 data_time: 0.76s time: 119.09s eta: 8:43:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31557.0 tgs: 261 data_time: 0.75s time: 120.83s eta: 8:49:37 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31578.0 tgs: 262 data_time: 0.93s time: 120.20s eta: 8:44:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 272 data_time: 0.78s time: 118.82s eta: 8:36:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 264 data_time: 0.89s time: 119.35s eta: 8:37:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31091.0 tgs: 260 data_time: 0.88s time: 119.33s eta: 8:35:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31802.0 tgs: 267 data_time: 0.77s time: 119.05s eta: 8:31:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29331.0 tgs: 246 data_time: 0.87s time: 119.16s eta: 8:30:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.206 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 267 data_time: 0.64s time: 119.05s eta: 8:27:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.89s time: 119.12s eta: 8:26:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31622.0 tgs: 265 data_time: 0.79s time: 119.24s eta: 8:24:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32132.0 tgs: 269 data_time: 0.73s time: 119.22s eta: 8:22:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.179 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30444.0 tgs: 255 data_time: 0.80s time: 119.38s eta: 8:21:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29681.0 tgs: 247 data_time: 0.72s time: 119.78s eta: 8:21:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31112.0 tgs: 261 data_time: 0.62s time: 119.19s eta: 8:16:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31417.0 tgs: 263 data_time: 0.96s time: 119.02s eta: 8:13:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 265 data_time: 0.49s time: 119.89s eta: 8:15:33 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31857.0 tgs: 266 data_time: 0.96s time: 119.55s eta: 8:12:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31694.0 tgs: 265 data_time: 0.83s time: 119.59s eta: 8:10:20 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31688.0 tgs: 266 data_time: 0.82s time: 119.00s eta: 8:05:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 268 data_time: 1.17s time: 119.83s eta: 8:07:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.88s time: 119.71s eta: 8:04:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.114 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31108.0 tgs: 260 data_time: 0.60s time: 119.40s eta: 8:01:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.207 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30617.0 tgs: 256 data_time: 0.69s time: 119.29s eta: 7:59:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32412.0 tgs: 271 data_time: 0.79s time: 119.19s eta: 7:56:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31155.0 tgs: 261 data_time: 0.81s time: 119.19s eta: 7:54:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.200 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31249.0 tgs: 262 data_time: 0.73s time: 118.94s eta: 7:51:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31767.0 tgs: 267 data_time: 0.61s time: 118.97s eta: 7:49:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32200.0 tgs: 269 data_time: 0.81s time: 119.41s eta: 7:49:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31028.0 tgs: 260 data_time: 0.73s time: 119.03s eta: 7:46:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30943.0 tgs: 259 data_time: 0.77s time: 119.10s eta: 7:44:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.096 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27601.0 tgs: 229 data_time: 0.95s time: 120.02s eta: 7:46:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.235 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 269 data_time: 0.84s time: 119.84s eta: 7:43:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31357.0 tgs: 263 data_time: 0.87s time: 119.15s eta: 7:38:44 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.207 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32042.0 tgs: 269 data_time: 0.67s time: 118.96s eta: 7:36:01 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.099 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30972.0 tgs: 259 data_time: 0.82s time: 119.33s eta: 7:35:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.262 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 264 data_time: 0.85s time: 119.38s eta: 7:33:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 268 data_time: 0.68s time: 119.06s eta: 7:30:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32006.0 tgs: 268 data_time: 0.73s time: 119.01s eta: 7:28:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31985.0 tgs: 268 data_time: 0.75s time: 119.23s eta: 7:27:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32083.0 tgs: 269 data_time: 0.53s time: 118.98s eta: 7:24:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 270 data_time: 0.68s time: 118.93s eta: 7:22:01 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 270 data_time: 0.75s time: 119.07s eta: 7:20:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31264.0 tgs: 262 data_time: 0.81s time: 118.99s eta: 7:18:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30557.0 tgs: 256 data_time: 0.86s time: 119.02s eta: 7:16:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32378.0 tgs: 271 data_time: 0.70s time: 119.26s eta: 7:15:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31049.0 tgs: 260 data_time: 0.70s time: 119.36s eta: 7:13:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31609.0 tgs: 265 data_time: 0.98s time: 118.94s eta: 7:10:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 269 data_time: 0.81s time: 119.21s eta: 7:09:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 268 data_time: 0.64s time: 119.58s eta: 7:08:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31888.0 tgs: 266 data_time: 0.66s time: 119.66s eta: 7:06:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.215 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31603.0 tgs: 265 data_time: 0.85s time: 118.95s eta: 7:02:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.183 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31844.0 tgs: 266 data_time: 1.27s time: 119.30s eta: 7:01:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.189 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31175.0 tgs: 260 data_time: 0.73s time: 119.52s eta: 7:00:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32251.0 tgs: 270 data_time: 0.63s time: 119.43s eta: 6:57:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.194 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32177.0 tgs: 270 data_time: 0.74s time: 118.91s eta: 6:54:11 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30656.0 tgs: 256 data_time: 0.76s time: 119.35s eta: 6:53:44 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.108 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31257.0 tgs: 262 data_time: 0.63s time: 118.93s eta: 6:50:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.184 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31953.0 tgs: 267 data_time: 0.69s time: 119.26s eta: 6:49:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32291.0 tgs: 271 data_time: 0.78s time: 119.02s eta: 6:46:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31443.0 tgs: 263 data_time: 0.66s time: 119.23s eta: 6:45:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.189 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31661.0 tgs: 265 data_time: 0.70s time: 119.37s eta: 6:43:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.83s time: 119.01s eta: 6:40:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 267 data_time: 0.93s time: 119.01s eta: 6:38:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32275.0 tgs: 270 data_time: 0.78s time: 119.45s eta: 6:38:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31728.0 tgs: 267 data_time: 0.88s time: 118.81s eta: 6:34:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30487.0 tgs: 255 data_time: 0.84s time: 119.43s eta: 6:34:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32181.0 tgs: 269 data_time: 0.57s time: 119.57s eta: 6:32:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31879.0 tgs: 267 data_time: 0.89s time: 119.37s eta: 6:29:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31770.0 tgs: 265 data_time: 0.70s time: 119.69s eta: 6:28:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31713.0 tgs: 265 data_time: 0.90s time: 119.36s eta: 6:25:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31751.0 tgs: 267 data_time: 0.82s time: 118.82s eta: 6:22:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31094.0 tgs: 261 data_time: 0.76s time: 119.01s eta: 6:20:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.094 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30046.0 tgs: 251 data_time: 0.78s time: 119.40s eta: 6:20:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30496.0 tgs: 255 data_time: 0.93s time: 119.34s eta: 6:17:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.89s time: 119.05s eta: 6:15:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32366.0 tgs: 271 data_time: 0.80s time: 119.43s eta: 6:14:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.72s time: 119.39s eta: 6:12:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32038.0 tgs: 268 data_time: 0.76s time: 119.28s eta: 6:09:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32204.0 tgs: 270 data_time: 0.68s time: 119.10s eta: 6:07:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 264 data_time: 0.86s time: 121.96s eta: 6:14:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31634.0 tgs: 264 data_time: 0.89s time: 119.38s eta: 6:04:06 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 267 data_time: 0.86s time: 119.58s eta: 6:02:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.114 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31460.0 tgs: 263 data_time: 0.45s time: 119.26s eta: 5:59:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.87s time: 119.25s eta: 5:57:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31519.0 tgs: 265 data_time: 0.93s time: 118.92s eta: 5:54:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.191 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31931.0 tgs: 268 data_time: 0.87s time: 119.02s eta: 5:53:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31804.0 tgs: 267 data_time: 0.96s time: 119.06s eta: 5:51:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31811.0 tgs: 266 data_time: 0.72s time: 119.29s eta: 5:49:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32211.0 tgs: 270 data_time: 0.92s time: 118.97s eta: 5:47:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31563.0 tgs: 264 data_time: 1.00s time: 119.29s eta: 5:45:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.114 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31523.0 tgs: 264 data_time: 0.85s time: 119.33s eta: 5:44:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31092.0 tgs: 261 data_time: 0.78s time: 119.05s eta: 5:41:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.118 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31645.0 tgs: 264 data_time: 0.63s time: 119.49s eta: 5:40:33 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.109 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 269 data_time: 0.95s time: 119.38s eta: 5:38:14 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32149.0 tgs: 268 data_time: 0.71s time: 119.57s eta: 5:36:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32055.0 tgs: 269 data_time: 0.63s time: 118.98s eta: 5:33:08 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31677.0 tgs: 266 data_time: 0.70s time: 118.95s eta: 5:31:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.122 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 266 data_time: 0.90s time: 119.25s eta: 5:29:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.116 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31308.0 tgs: 262 data_time: 0.59s time: 119.39s eta: 5:28:19 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.179 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31346.0 tgs: 262 data_time: 0.97s time: 119.46s eta: 5:26:31 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 266 data_time: 0.76s time: 118.95s eta: 5:23:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.193 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 266 data_time: 0.75s time: 119.63s eta: 5:23:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31885.0 tgs: 267 data_time: 0.58s time: 119.34s eta: 5:20:13 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.200 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31464.0 tgs: 264 data_time: 0.88s time: 119.12s eta: 5:17:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.107 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30545.0 tgs: 255 data_time: 0.50s time: 119.42s eta: 5:16:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.300 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30973.0 tgs: 259 data_time: 0.99s time: 119.58s eta: 5:14:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.178 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32125.0 tgs: 269 data_time: 0.92s time: 119.00s eta: 5:11:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32641.0 tgs: 274 data_time: 0.60s time: 119.00s eta: 5:09:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31741.0 tgs: 266 data_time: 0.71s time: 119.23s eta: 5:08:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 263 data_time: 0.61s time: 119.68s eta: 5:07:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31365.0 tgs: 263 data_time: 0.85s time: 119.11s eta: 5:03:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 269 data_time: 0.82s time: 119.07s eta: 5:01:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.219 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31410.0 tgs: 258 data_time: 0.86s time: 121.52s eta: 5:05:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.181 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31644.0 tgs: 264 data_time: 0.86s time: 119.58s eta: 4:58:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30798.0 tgs: 257 data_time: 0.65s time: 119.49s eta: 4:56:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31825.0 tgs: 265 data_time: 0.73s time: 119.74s eta: 4:55:20 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31122.0 tgs: 260 data_time: 0.87s time: 119.51s eta: 4:52:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 269 data_time: 0.96s time: 119.42s eta: 4:50:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32125.0 tgs: 269 data_time: 0.59s time: 119.20s eta: 4:48:04 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32413.0 tgs: 272 data_time: 0.63s time: 118.88s eta: 4:45:19 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31172.0 tgs: 261 data_time: 0.79s time: 119.12s eta: 4:43:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.297 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 265 data_time: 1.17s time: 119.98s eta: 4:43:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31651.0 tgs: 265 data_time: 0.81s time: 119.15s eta: 4:39:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32070.0 tgs: 269 data_time: 0.92s time: 119.01s eta: 4:37:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31895.0 tgs: 267 data_time: 0.81s time: 119.05s eta: 4:35:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31039.0 tgs: 260 data_time: 0.80s time: 119.37s eta: 4:34:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 263 data_time: 0.61s time: 119.42s eta: 4:32:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31654.0 tgs: 266 data_time: 0.60s time: 118.93s eta: 4:29:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.177 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 262 data_time: 0.84s time: 119.60s eta: 4:29:05 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.111 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31160.0 tgs: 261 data_time: 0.72s time: 119.24s eta: 4:26:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31996.0 tgs: 268 data_time: 0.96s time: 119.06s eta: 4:23:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30189.0 tgs: 254 data_time: 0.72s time: 118.85s eta: 4:21:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31631.0 tgs: 266 data_time: 0.84s time: 118.76s eta: 4:19:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.220 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31359.0 tgs: 262 data_time: 0.70s time: 119.42s eta: 4:18:44 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31656.0 tgs: 265 data_time: 0.63s time: 119.40s eta: 4:16:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31990.0 tgs: 268 data_time: 0.64s time: 119.34s eta: 4:14:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31922.0 tgs: 268 data_time: 0.69s time: 119.04s eta: 4:11:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 270 data_time: 0.80s time: 119.13s eta: 4:10:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32120.0 tgs: 269 data_time: 0.71s time: 119.14s eta: 4:08:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31279.0 tgs: 262 data_time: 0.80s time: 118.95s eta: 4:05:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30901.0 tgs: 259 data_time: 0.72s time: 119.15s eta: 4:04:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.114 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31748.0 tgs: 265 data_time: 0.87s time: 119.50s eta: 4:02:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.84s time: 119.24s eta: 4:00:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31431.0 tgs: 263 data_time: 0.66s time: 119.19s eta: 3:58:22 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.104 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28711.0 tgs: 240 data_time: 0.55s time: 119.36s eta: 3:56:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31444.0 tgs: 263 data_time: 0.89s time: 119.37s eta: 3:54:45 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 269 data_time: 0.61s time: 119.96s eta: 3:53:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31520.0 tgs: 264 data_time: 0.56s time: 119.31s eta: 3:50:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.105 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30647.0 tgs: 256 data_time: 0.69s time: 119.64s eta: 3:49:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31752.0 tgs: 267 data_time: 0.60s time: 118.85s eta: 3:45:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31262.0 tgs: 261 data_time: 0.69s time: 119.75s eta: 3:45:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31481.0 tgs: 263 data_time: 0.67s time: 119.32s eta: 3:42:44 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31925.0 tgs: 267 data_time: 0.68s time: 119.26s eta: 3:40:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31074.0 tgs: 261 data_time: 0.71s time: 118.99s eta: 3:38:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31522.0 tgs: 263 data_time: 0.75s time: 119.44s eta: 3:36:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 267 data_time: 0.53s time: 119.02s eta: 3:34:14 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30871.0 tgs: 259 data_time: 0.70s time: 119.00s eta: 3:32:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 261 data_time: 1.05s time: 119.21s eta: 3:30:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30661.0 tgs: 255 data_time: 0.90s time: 120.11s eta: 3:30:11 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31667.0 tgs: 265 data_time: 0.74s time: 119.13s eta: 3:26:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30884.0 tgs: 258 data_time: 0.62s time: 119.32s eta: 3:24:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31589.0 tgs: 265 data_time: 0.63s time: 119.06s eta: 3:22:24 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31608.0 tgs: 264 data_time: 0.60s time: 119.43s eta: 3:21:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 266 data_time: 0.76s time: 119.29s eta: 3:18:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32385.0 tgs: 271 data_time: 1.08s time: 119.22s eta: 3:16:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32058.0 tgs: 268 data_time: 0.77s time: 119.18s eta: 3:14:39 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.111 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32233.0 tgs: 270 data_time: 0.70s time: 119.02s eta: 3:12:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31515.0 tgs: 264 data_time: 0.89s time: 119.18s eta: 3:10:41 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 262 data_time: 0.68s time: 121.00s eta: 3:11:35 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31599.0 tgs: 264 data_time: 0.75s time: 119.50s eta: 3:07:12 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31312.0 tgs: 262 data_time: 0.81s time: 119.26s eta: 3:04:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31694.0 tgs: 266 data_time: 0.77s time: 118.99s eta: 3:02:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31602.0 tgs: 265 data_time: 1.00s time: 119.22s eta: 3:00:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.114 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 268 data_time: 0.67s time: 119.46s eta: 2:59:11 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31861.0 tgs: 267 data_time: 0.88s time: 118.98s eta: 2:56:28 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32357.0 tgs: 271 data_time: 0.83s time: 119.35s eta: 2:55:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.198 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31459.0 tgs: 263 data_time: 0.67s time: 119.42s eta: 2:53:09 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31680.0 tgs: 265 data_time: 0.94s time: 119.35s eta: 2:51:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31085.0 tgs: 255 data_time: 0.66s time: 121.45s eta: 2:52:03 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31704.0 tgs: 265 data_time: 0.62s time: 119.25s eta: 2:46:57 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 267 data_time: 0.91s time: 119.56s eta: 2:45:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 269 data_time: 0.86s time: 118.91s eta: 2:42:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31401.0 tgs: 263 data_time: 0.80s time: 119.18s eta: 2:40:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31741.0 tgs: 266 data_time: 0.83s time: 118.96s eta: 2:38:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31177.0 tgs: 261 data_time: 0.70s time: 119.23s eta: 2:36:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.94s time: 119.12s eta: 2:34:51 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.114 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31404.0 tgs: 262 data_time: 0.72s time: 119.42s eta: 2:33:15 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30851.0 tgs: 256 data_time: 0.91s time: 120.22s eta: 2:32:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 264 data_time: 0.92s time: 119.29s eta: 2:29:06 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 270 data_time: 1.01s time: 118.87s eta: 2:26:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.106 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29997.0 tgs: 250 data_time: 0.60s time: 119.52s eta: 2:25:24 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31188.0 tgs: 261 data_time: 0.58s time: 119.07s eta: 2:22:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30698.0 tgs: 257 data_time: 0.86s time: 119.13s eta: 2:20:58 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31945.0 tgs: 268 data_time: 0.71s time: 119.05s eta: 2:18:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31510.0 tgs: 262 data_time: 0.59s time: 119.84s eta: 2:17:48 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 269 data_time: 0.76s time: 119.36s eta: 2:15:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31723.0 tgs: 265 data_time: 0.64s time: 119.39s eta: 2:13:19 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 262 data_time: 0.73s time: 119.67s eta: 2:11:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32311.0 tgs: 270 data_time: 0.54s time: 119.49s eta: 2:09:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 264 data_time: 0.84s time: 118.97s eta: 2:06:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.90s time: 119.36s eta: 2:05:19 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31314.0 tgs: 262 data_time: 0.66s time: 119.29s eta: 2:03:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32188.0 tgs: 270 data_time: 0.85s time: 119.21s eta: 2:01:11 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 264 data_time: 0.57s time: 119.51s eta: 1:59:30 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31628.0 tgs: 264 data_time: 0.74s time: 119.42s eta: 1:57:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32191.0 tgs: 269 data_time: 0.99s time: 119.53s eta: 1:55:32 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32189.0 tgs: 269 data_time: 0.64s time: 119.31s eta: 1:53:20 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31456.0 tgs: 263 data_time: 1.13s time: 119.25s eta: 1:51:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31457.0 tgs: 264 data_time: 0.73s time: 118.92s eta: 1:49:00 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31595.0 tgs: 264 data_time: 0.57s time: 119.64s eta: 1:47:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.099 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31715.0 tgs: 264 data_time: 0.87s time: 119.72s eta: 1:45:44 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31227.0 tgs: 261 data_time: 0.70s time: 119.35s eta: 1:43:26 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31517.0 tgs: 264 data_time: 0.58s time: 119.31s eta: 1:41:24 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30828.0 tgs: 258 data_time: 1.12s time: 119.27s eta: 1:39:23 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 266 data_time: 0.84s time: 118.92s eta: 1:37:07 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 0.75s time: 119.15s eta: 1:35:19 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32169.0 tgs: 270 data_time: 0.90s time: 119.07s eta: 1:33:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32048.0 tgs: 268 data_time: 0.74s time: 119.25s eta: 1:31:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.263 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 264 data_time: 0.85s time: 119.73s eta: 1:29:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 270 data_time: 0.93s time: 119.02s eta: 1:27:16 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 268 data_time: 0.60s time: 118.86s eta: 1:25:10 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31771.0 tgs: 265 data_time: 0.62s time: 119.48s eta: 1:23:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31409.0 tgs: 262 data_time: 0.66s time: 119.48s eta: 1:21:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32078.0 tgs: 269 data_time: 0.97s time: 118.94s eta: 1:19:17 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30835.0 tgs: 259 data_time: 0.53s time: 118.94s eta: 1:17:18 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 259 data_time: 0.64s time: 122.15s eta: 1:17:21 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29971.0 tgs: 251 data_time: 0.68s time: 119.37s eta: 1:13:36 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.205 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32196.0 tgs: 270 data_time: 0.61s time: 119.05s eta: 1:11:25 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31560.0 tgs: 265 data_time: 0.70s time: 118.92s eta: 1:09:22 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32256.0 tgs: 270 data_time: 0.69s time: 119.10s eta: 1:07:29 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31973.0 tgs: 267 data_time: 0.64s time: 119.41s eta: 1:05:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30809.0 tgs: 258 data_time: 0.76s time: 119.34s eta: 1:03:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32157.0 tgs: 269 data_time: 0.74s time: 119.18s eta: 1:01:34 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.54s time: 119.27s eta: 0:59:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31766.0 tgs: 265 data_time: 0.91s time: 119.45s eta: 0:57:44 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32283.0 tgs: 270 data_time: 0.83s time: 119.39s eta: 0:55:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.204 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 269 data_time: 0.73s time: 119.49s eta: 0:53:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31809.0 tgs: 266 data_time: 0.61s time: 119.36s eta: 0:51:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.201 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31292.0 tgs: 261 data_time: 0.92s time: 119.47s eta: 0:49:46 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.223 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31407.0 tgs: 264 data_time: 0.77s time: 118.90s eta: 0:47:33 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.83s time: 119.28s eta: 0:45:43 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32159.0 tgs: 269 data_time: 0.87s time: 119.22s eta: 0:43:42 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31686.0 tgs: 266 data_time: 0.75s time: 119.00s eta: 0:41:38 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31867.0 tgs: 267 data_time: 0.84s time: 119.03s eta: 0:39:40 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32209.0 tgs: 269 data_time: 0.78s time: 119.66s eta: 0:37:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31218.0 tgs: 261 data_time: 0.51s time: 119.28s eta: 0:35:47 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31653.0 tgs: 264 data_time: 0.72s time: 119.74s eta: 0:33:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31428.0 tgs: 259 data_time: 0.70s time: 121.23s eta: 0:32:19 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.58s time: 119.37s eta: 0:29:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32149.0 tgs: 269 data_time: 0.97s time: 119.32s eta: 0:27:50 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31200.0 tgs: 260 data_time: 0.69s time: 119.64s eta: 0:25:55 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31894.0 tgs: 267 data_time: 0.65s time: 119.09s eta: 0:23:49 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31019.0 tgs: 259 data_time: 0.89s time: 119.39s eta: 0:21:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31951.0 tgs: 267 data_time: 0.80s time: 119.25s eta: 0:19:52 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.66s time: 119.25s eta: 0:17:53 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 263 data_time: 0.60s time: 119.90s eta: 0:15:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32221.0 tgs: 270 data_time: 0.85s time: 119.22s eta: 0:13:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30866.0 tgs: 259 data_time: 0.75s time: 119.09s eta: 0:11:54 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.067 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31751.0 tgs: 266 data_time: 0.52s time: 119.32s eta: 0:09:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30864.0 tgs: 259 data_time: 1.01s time: 119.16s eta: 0:07:56 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32549.0 tgs: 271 data_time: 0.53s time: 119.69s eta: 0:05:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31389.0 tgs: 258 data_time: 0.60s time: 121.38s eta: 0:04:02 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31510.0 tgs: 264 data_time: 0.64s time: 119.20s eta: 0:01:59 +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 28][DP 7][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank29.log b/20250121165312/rank29.log new file mode 100644 index 0000000000000000000000000000000000000000..d1e84dd2b08c03c3564c841199324d71f6bb8bb8 --- /dev/null +++ b/20250121165312/rank29.log @@ -0,0 +1,851 @@ +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.06s +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.42 seconds, peak gpu memory 13.4G +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.289 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 32271.0 tgs: 210 data_time: 1.78s time: 153.55s eta: 1 day, 1:17:34 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.208 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32224.0 tgs: 269 data_time: 0.96s time: 119.57s eta: 19:39:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.225 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31186.0 tgs: 260 data_time: 0.98s time: 119.68s eta: 19:38:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.213 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32092.0 tgs: 268 data_time: 0.86s time: 119.65s eta: 19:36:31 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.224 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 265 data_time: 0.84s time: 119.30s eta: 19:31:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.254 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31263.0 tgs: 261 data_time: 0.76s time: 119.53s eta: 19:31:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30097.0 tgs: 244 data_time: 0.83s time: 122.96s eta: 20:02:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.205 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 270 data_time: 0.91s time: 119.89s eta: 19:30:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.194 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30582.0 tgs: 253 data_time: 0.85s time: 120.65s eta: 19:36:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.248 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32372.0 tgs: 269 data_time: 0.82s time: 119.92s eta: 19:27:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 267 data_time: 0.71s time: 120.46s eta: 19:30:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31025.0 tgs: 259 data_time: 0.65s time: 119.34s eta: 19:17:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.320 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31804.0 tgs: 265 data_time: 0.92s time: 119.69s eta: 19:18:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.230 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32389.0 tgs: 271 data_time: 0.90s time: 119.17s eta: 19:11:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30946.0 tgs: 259 data_time: 0.91s time: 119.22s eta: 19:10:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31432.0 tgs: 264 data_time: 0.94s time: 118.92s eta: 19:05:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32194.0 tgs: 270 data_time: 0.76s time: 119.14s eta: 19:05:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31342.0 tgs: 262 data_time: 1.09s time: 119.18s eta: 19:04:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32054.0 tgs: 268 data_time: 0.70s time: 119.40s eta: 19:04:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.75s time: 119.33s eta: 19:01:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31497.0 tgs: 261 data_time: 0.63s time: 120.31s eta: 19:08:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32299.0 tgs: 262 data_time: 0.82s time: 122.92s eta: 19:31:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 268 data_time: 0.66s time: 119.48s eta: 18:57:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31131.0 tgs: 261 data_time: 0.65s time: 119.16s eta: 18:52:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.78s time: 119.24s eta: 18:50:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32152.0 tgs: 269 data_time: 0.69s time: 119.39s eta: 18:50:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31336.0 tgs: 261 data_time: 0.99s time: 119.66s eta: 18:50:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31412.0 tgs: 263 data_time: 0.77s time: 119.38s eta: 18:46:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 268 data_time: 0.72s time: 119.79s eta: 18:48:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 245 data_time: 0.86s time: 129.33s eta: 20:15:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31765.0 tgs: 265 data_time: 0.88s time: 119.56s eta: 18:41:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31949.0 tgs: 267 data_time: 0.96s time: 119.39s eta: 18:38:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.247 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 259 data_time: 0.62s time: 120.66s eta: 18:48:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31518.0 tgs: 263 data_time: 0.97s time: 119.39s eta: 18:34:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 268 data_time: 0.94s time: 119.16s eta: 18:30:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.275 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32039.0 tgs: 269 data_time: 1.23s time: 119.07s eta: 18:27:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.78s time: 119.13s eta: 18:25:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32352.0 tgs: 268 data_time: 0.79s time: 120.39s eta: 18:35:34 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31385.0 tgs: 259 data_time: 1.05s time: 121.17s eta: 18:40:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 261 data_time: 0.73s time: 120.66s eta: 18:34:06 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31223.0 tgs: 260 data_time: 0.78s time: 119.73s eta: 18:23:31 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.230 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32577.0 tgs: 262 data_time: 1.11s time: 123.88s eta: 18:59:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31503.0 tgs: 263 data_time: 0.79s time: 119.56s eta: 18:17:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 263 data_time: 0.75s time: 119.35s eta: 18:14:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31080.0 tgs: 260 data_time: 0.85s time: 119.22s eta: 18:10:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31762.0 tgs: 265 data_time: 0.79s time: 119.60s eta: 18:12:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.134 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31859.0 tgs: 267 data_time: 0.87s time: 119.26s eta: 18:07:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32052.0 tgs: 269 data_time: 0.64s time: 119.01s eta: 18:03:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31795.0 tgs: 266 data_time: 0.80s time: 119.10s eta: 18:01:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31424.0 tgs: 263 data_time: 0.78s time: 119.27s eta: 18:01:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.73s time: 119.33s eta: 17:59:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32221.0 tgs: 268 data_time: 0.93s time: 120.03s eta: 18:04:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31787.0 tgs: 266 data_time: 0.89s time: 119.09s eta: 17:53:48 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31727.0 tgs: 265 data_time: 0.70s time: 119.41s eta: 17:54:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31831.0 tgs: 267 data_time: 0.93s time: 119.00s eta: 17:49:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 269 data_time: 0.96s time: 119.01s eta: 17:47:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31492.0 tgs: 251 data_time: 0.83s time: 125.36s eta: 18:41:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31059.0 tgs: 258 data_time: 0.71s time: 120.10s eta: 17:52:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30232.0 tgs: 253 data_time: 0.80s time: 119.40s eta: 17:44:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31767.0 tgs: 266 data_time: 0.84s time: 119.34s eta: 17:42:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.61s time: 119.52s eta: 17:41:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 255 data_time: 0.67s time: 126.04s eta: 18:37:31 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31753.0 tgs: 258 data_time: 0.45s time: 122.63s eta: 18:05:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31116.0 tgs: 256 data_time: 0.71s time: 121.15s eta: 17:50:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31233.0 tgs: 262 data_time: 0.70s time: 119.03s eta: 17:29:28 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32074.0 tgs: 269 data_time: 0.74s time: 118.85s eta: 17:25:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 256 data_time: 0.82s time: 124.00s eta: 18:09:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32008.0 tgs: 256 data_time: 0.74s time: 124.56s eta: 18:11:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32207.0 tgs: 269 data_time: 0.79s time: 119.39s eta: 17:24:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31813.0 tgs: 265 data_time: 0.70s time: 119.70s eta: 17:25:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.247 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31610.0 tgs: 264 data_time: 0.66s time: 119.52s eta: 17:21:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 262 data_time: 0.72s time: 118.86s eta: 17:14:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.226 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 265 data_time: 0.99s time: 119.85s eta: 17:20:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.229 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31473.0 tgs: 263 data_time: 0.92s time: 119.56s eta: 17:16:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32047.0 tgs: 262 data_time: 0.76s time: 121.87s eta: 17:34:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29030.0 tgs: 235 data_time: 0.58s time: 123.51s eta: 17:46:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.74s time: 119.01s eta: 17:05:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.83s time: 119.24s eta: 17:05:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32013.0 tgs: 268 data_time: 0.58s time: 119.13s eta: 17:02:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30187.0 tgs: 253 data_time: 0.76s time: 119.26s eta: 17:01:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31309.0 tgs: 263 data_time: 0.80s time: 119.03s eta: 16:57:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 266 data_time: 0.78s time: 119.28s eta: 16:57:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31701.0 tgs: 266 data_time: 0.60s time: 119.16s eta: 16:54:48 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 269 data_time: 0.68s time: 119.58s eta: 16:56:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31201.0 tgs: 261 data_time: 0.72s time: 119.13s eta: 16:50:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31038.0 tgs: 261 data_time: 0.93s time: 118.84s eta: 16:46:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31025.0 tgs: 260 data_time: 0.85s time: 119.01s eta: 16:45:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30099.0 tgs: 251 data_time: 0.87s time: 119.85s eta: 16:50:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.205 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.95s time: 119.38s eta: 16:44:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31723.0 tgs: 266 data_time: 0.83s time: 119.01s eta: 16:39:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31759.0 tgs: 262 data_time: 0.89s time: 120.92s eta: 16:53:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30751.0 tgs: 258 data_time: 0.60s time: 119.00s eta: 16:35:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31819.0 tgs: 266 data_time: 0.65s time: 119.28s eta: 16:36:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31779.0 tgs: 266 data_time: 0.72s time: 119.23s eta: 16:33:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31279.0 tgs: 262 data_time: 0.78s time: 119.36s eta: 16:32:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31568.0 tgs: 263 data_time: 0.84s time: 119.70s eta: 16:33:31 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31989.0 tgs: 268 data_time: 0.84s time: 119.20s eta: 16:27:22 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31654.0 tgs: 264 data_time: 0.67s time: 119.76s eta: 16:30:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31764.0 tgs: 264 data_time: 0.77s time: 120.28s eta: 16:32:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31774.0 tgs: 267 data_time: 0.71s time: 118.97s eta: 16:19:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31673.0 tgs: 248 data_time: 0.72s time: 127.32s eta: 17:26:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.87s time: 119.22s eta: 16:17:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:28:55][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.97s time: 119.14s eta: 16:14:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 259 data_time: 0.53s time: 119.25s eta: 16:13:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.101 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31441.0 tgs: 263 data_time: 0.79s time: 119.19s eta: 16:11:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.219 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 270 data_time: 0.91s time: 118.95s eta: 16:07:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.119 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32186.0 tgs: 270 data_time: 0.91s time: 118.86s eta: 16:04:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31747.0 tgs: 266 data_time: 0.76s time: 118.96s eta: 16:03:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31339.0 tgs: 257 data_time: 0.72s time: 121.84s eta: 16:24:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.64s time: 119.00s eta: 15:59:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 269 data_time: 0.87s time: 119.22s eta: 15:59:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31503.0 tgs: 263 data_time: 1.05s time: 119.61s eta: 16:00:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.124 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31520.0 tgs: 264 data_time: 0.75s time: 119.04s eta: 15:54:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31714.0 tgs: 266 data_time: 0.79s time: 118.98s eta: 15:51:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31219.0 tgs: 261 data_time: 0.98s time: 119.16s eta: 15:51:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31548.0 tgs: 263 data_time: 0.73s time: 119.64s eta: 15:53:08 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 269 data_time: 0.92s time: 118.93s eta: 15:45:29 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31262.0 tgs: 262 data_time: 0.91s time: 119.28s eta: 15:46:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31993.0 tgs: 268 data_time: 0.60s time: 119.13s eta: 15:43:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.228 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31357.0 tgs: 263 data_time: 0.88s time: 119.00s eta: 15:40:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31058.0 tgs: 259 data_time: 0.87s time: 119.66s eta: 15:43:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32204.0 tgs: 265 data_time: 0.71s time: 121.50s eta: 15:55:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.209 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32363.0 tgs: 271 data_time: 0.67s time: 119.41s eta: 15:37:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 242 data_time: 0.77s time: 129.96s eta: 16:58:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31435.0 tgs: 263 data_time: 0.91s time: 119.47s eta: 15:33:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 270 data_time: 0.81s time: 119.60s eta: 15:32:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 266 data_time: 0.85s time: 119.53s eta: 15:30:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 259 data_time: 0.71s time: 122.14s eta: 15:48:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30593.0 tgs: 253 data_time: 0.86s time: 120.75s eta: 15:35:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31383.0 tgs: 262 data_time: 0.67s time: 119.41s eta: 15:23:28 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31513.0 tgs: 263 data_time: 0.58s time: 119.44s eta: 15:21:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32587.0 tgs: 247 data_time: 0.92s time: 131.63s eta: 16:53:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 264 data_time: 0.64s time: 119.29s eta: 15:16:33 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30622.0 tgs: 256 data_time: 0.74s time: 119.57s eta: 15:16:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30873.0 tgs: 252 data_time: 1.15s time: 122.26s eta: 15:35:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31673.0 tgs: 264 data_time: 0.61s time: 119.71s eta: 15:13:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.125 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30890.0 tgs: 258 data_time: 0.89s time: 119.35s eta: 15:09:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31272.0 tgs: 263 data_time: 0.78s time: 118.90s eta: 15:03:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32168.0 tgs: 270 data_time: 0.97s time: 119.00s eta: 15:02:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.214 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32521.0 tgs: 272 data_time: 0.82s time: 119.30s eta: 15:02:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.86s time: 118.92s eta: 14:57:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 263 data_time: 0.72s time: 120.52s eta: 15:07:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31767.0 tgs: 266 data_time: 0.65s time: 119.31s eta: 14:56:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.254 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31769.0 tgs: 259 data_time: 1.08s time: 122.66s eta: 15:19:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32206.0 tgs: 270 data_time: 0.81s time: 119.12s eta: 14:51:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30760.0 tgs: 256 data_time: 0.94s time: 119.76s eta: 14:54:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32005.0 tgs: 267 data_time: 0.57s time: 119.46s eta: 14:49:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.112 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31508.0 tgs: 264 data_time: 0.71s time: 118.99s eta: 14:44:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31893.0 tgs: 267 data_time: 0.70s time: 119.33s eta: 14:45:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.118 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31235.0 tgs: 262 data_time: 0.53s time: 118.94s eta: 14:40:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31995.0 tgs: 268 data_time: 0.71s time: 119.05s eta: 14:38:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32011.0 tgs: 268 data_time: 0.79s time: 119.06s eta: 14:37:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30863.0 tgs: 259 data_time: 0.68s time: 119.12s eta: 14:35:33 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30213.0 tgs: 253 data_time: 0.69s time: 119.00s eta: 14:32:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31367.0 tgs: 263 data_time: 0.87s time: 119.08s eta: 14:31:14 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31269.0 tgs: 262 data_time: 0.71s time: 119.15s eta: 14:29:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31668.0 tgs: 264 data_time: 0.66s time: 119.65s eta: 14:31:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.120 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31958.0 tgs: 268 data_time: 0.72s time: 119.18s eta: 14:26:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.208 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 269 data_time: 0.77s time: 119.20s eta: 14:24:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.82s time: 119.39s eta: 14:23:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31959.0 tgs: 264 data_time: 0.85s time: 120.73s eta: 14:31:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.116 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32334.0 tgs: 272 data_time: 0.74s time: 118.85s eta: 14:15:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31346.0 tgs: 263 data_time: 0.85s time: 119.10s eta: 14:15:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31361.0 tgs: 263 data_time: 0.68s time: 119.15s eta: 14:13:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31668.0 tgs: 263 data_time: 1.00s time: 119.97s eta: 14:17:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31098.0 tgs: 259 data_time: 0.82s time: 119.86s eta: 14:15:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30925.0 tgs: 258 data_time: 0.55s time: 119.58s eta: 14:11:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32117.0 tgs: 269 data_time: 0.65s time: 119.23s eta: 14:06:29 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31798.0 tgs: 266 data_time: 0.77s time: 119.26s eta: 14:04:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.220 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.77s time: 119.27s eta: 14:02:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32114.0 tgs: 269 data_time: 0.91s time: 119.14s eta: 13:59:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31630.0 tgs: 265 data_time: 0.61s time: 119.12s eta: 13:57:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32387.0 tgs: 271 data_time: 0.96s time: 119.17s eta: 13:56:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 266 data_time: 0.68s time: 119.35s eta: 13:55:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31302.0 tgs: 260 data_time: 0.76s time: 120.11s eta: 13:58:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31527.0 tgs: 257 data_time: 0.63s time: 122.56s eta: 14:13:50 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:56:46][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32428.0 tgs: 272 data_time: 0.75s time: 119.10s eta: 13:47:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31703.0 tgs: 265 data_time: 0.63s time: 119.39s eta: 13:47:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.131 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.96s time: 119.34s eta: 13:45:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31817.0 tgs: 266 data_time: 0.73s time: 119.35s eta: 13:43:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31491.0 tgs: 264 data_time: 0.56s time: 119.11s eta: 13:39:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31439.0 tgs: 263 data_time: 0.92s time: 119.25s eta: 13:38:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.214 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 1.05s time: 119.13s eta: 13:36:03 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.234 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31671.0 tgs: 265 data_time: 0.92s time: 119.22s eta: 13:34:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30523.0 tgs: 255 data_time: 0.91s time: 119.45s eta: 13:34:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31916.0 tgs: 268 data_time: 0.64s time: 119.08s eta: 13:29:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31049.0 tgs: 261 data_time: 0.43s time: 118.92s eta: 13:26:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31921.0 tgs: 266 data_time: 0.52s time: 119.81s eta: 13:30:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 268 data_time: 0.85s time: 119.25s eta: 13:24:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.128 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.91s time: 119.16s eta: 13:22:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32195.0 tgs: 269 data_time: 0.84s time: 119.34s eta: 13:21:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31241.0 tgs: 261 data_time: 0.90s time: 119.26s eta: 13:19:03 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 270 data_time: 0.76s time: 119.04s eta: 13:15:33 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.194 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.81s time: 119.37s eta: 13:15:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30896.0 tgs: 259 data_time: 0.82s time: 119.18s eta: 13:12:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 269 data_time: 0.82s time: 119.45s eta: 13:12:22 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30668.0 tgs: 257 data_time: 0.85s time: 119.29s eta: 13:09:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 264 data_time: 0.70s time: 119.15s eta: 13:06:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 265 data_time: 0.71s time: 120.38s eta: 13:12:28 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 265 data_time: 0.81s time: 120.16s eta: 13:09:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 269 data_time: 0.74s time: 119.50s eta: 13:02:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27721.0 tgs: 231 data_time: 0.92s time: 119.51s eta: 13:00:48 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31795.0 tgs: 265 data_time: 0.86s time: 119.57s eta: 12:59:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31254.0 tgs: 260 data_time: 0.64s time: 119.79s eta: 12:58:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31574.0 tgs: 265 data_time: 0.77s time: 119.04s eta: 12:51:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31155.0 tgs: 261 data_time: 0.83s time: 119.06s eta: 12:49:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31056.0 tgs: 260 data_time: 0.61s time: 119.11s eta: 12:48:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-21 23:58:28][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32326.0 tgs: 270 data_time: 0.85s time: 119.39s eta: 12:48:06 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 264 data_time: 0.74s time: 119.20s eta: 12:44:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32123.0 tgs: 268 data_time: 0.94s time: 119.44s eta: 12:44:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31749.0 tgs: 266 data_time: 0.67s time: 119.10s eta: 12:40:14 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30997.0 tgs: 260 data_time: 0.83s time: 119.07s eta: 12:38:03 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32055.0 tgs: 268 data_time: 0.79s time: 119.60s eta: 12:39:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 273 data_time: 0.77s time: 118.78s eta: 12:32:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.187 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 270 data_time: 0.76s time: 118.88s eta: 12:30:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31814.0 tgs: 266 data_time: 0.80s time: 119.17s eta: 12:30:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31922.0 tgs: 267 data_time: 0.83s time: 119.27s eta: 12:29:23 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.195 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32157.0 tgs: 269 data_time: 0.59s time: 119.34s eta: 12:27:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.198 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31834.0 tgs: 267 data_time: 0.72s time: 119.11s eta: 12:24:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.129 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31074.0 tgs: 260 data_time: 0.65s time: 119.34s eta: 12:23:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.209 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32137.0 tgs: 269 data_time: 0.67s time: 119.26s eta: 12:21:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31326.0 tgs: 263 data_time: 0.86s time: 119.02s eta: 12:17:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31646.0 tgs: 261 data_time: 0.64s time: 121.13s eta: 12:28:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31682.0 tgs: 264 data_time: 0.84s time: 119.99s eta: 12:19:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31959.0 tgs: 269 data_time: 0.63s time: 118.72s eta: 12:10:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.118 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30787.0 tgs: 257 data_time: 0.84s time: 119.63s eta: 12:13:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31823.0 tgs: 265 data_time: 0.76s time: 119.83s eta: 12:12:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31642.0 tgs: 265 data_time: 0.59s time: 119.14s eta: 12:06:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32138.0 tgs: 269 data_time: 0.67s time: 119.17s eta: 12:04:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.197 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32136.0 tgs: 269 data_time: 0.58s time: 119.11s eta: 12:02:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.72s time: 119.19s eta: 12:01:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 267 data_time: 0.72s time: 118.95s eta: 11:57:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 266 data_time: 0.64s time: 119.70s eta: 12:00:13 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.251 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 267 data_time: 0.85s time: 119.27s eta: 11:55:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.174 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 266 data_time: 0.91s time: 119.28s eta: 11:53:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31851.0 tgs: 266 data_time: 0.61s time: 119.70s eta: 11:54:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31892.0 tgs: 268 data_time: 0.90s time: 118.78s eta: 11:46:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.175 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31879.0 tgs: 267 data_time: 0.78s time: 119.15s eta: 11:46:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31023.0 tgs: 259 data_time: 0.75s time: 119.61s eta: 11:47:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32492.0 tgs: 273 data_time: 0.73s time: 118.87s eta: 11:41:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30868.0 tgs: 260 data_time: 0.64s time: 118.68s eta: 11:38:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31405.0 tgs: 263 data_time: 0.96s time: 119.21s eta: 11:39:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31589.0 tgs: 265 data_time: 0.96s time: 119.14s eta: 11:36:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 269 data_time: 0.70s time: 119.07s eta: 11:34:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31521.0 tgs: 264 data_time: 0.95s time: 118.95s eta: 11:31:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.206 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 269 data_time: 0.83s time: 118.91s eta: 11:29:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.201 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32398.0 tgs: 272 data_time: 0.81s time: 118.97s eta: 11:28:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.122 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31334.0 tgs: 262 data_time: 0.62s time: 119.29s eta: 11:27:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.175 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32549.0 tgs: 272 data_time: 0.73s time: 119.34s eta: 11:26:13 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.227 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32111.0 tgs: 270 data_time: 0.82s time: 118.89s eta: 11:21:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.75s time: 119.20s eta: 11:21:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.180 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 28070.0 tgs: 234 data_time: 0.71s time: 119.71s eta: 11:22:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 0.66s time: 119.12s eta: 11:16:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30675.0 tgs: 257 data_time: 0.63s time: 119.12s eta: 11:15:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32604.0 tgs: 273 data_time: 0.87s time: 118.99s eta: 11:12:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.181 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31034.0 tgs: 260 data_time: 0.71s time: 119.08s eta: 11:10:50 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 262 data_time: 0.64s time: 119.16s eta: 11:09:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.184 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 267 data_time: 0.79s time: 118.96s eta: 11:06:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32310.0 tgs: 270 data_time: 0.87s time: 119.47s eta: 11:07:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31737.0 tgs: 265 data_time: 0.99s time: 119.64s eta: 11:06:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 269 data_time: 0.53s time: 119.30s eta: 11:02:06 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.237 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 265 data_time: 1.01s time: 119.32s eta: 11:00:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31822.0 tgs: 267 data_time: 0.77s time: 118.81s eta: 10:55:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28741.0 tgs: 240 data_time: 0.74s time: 119.47s eta: 10:57:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31552.0 tgs: 264 data_time: 0.83s time: 119.37s eta: 10:54:31 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31005.0 tgs: 259 data_time: 0.56s time: 119.28s eta: 10:52:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31325.0 tgs: 262 data_time: 0.61s time: 119.26s eta: 10:49:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.197 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 270 data_time: 0.63s time: 119.10s eta: 10:47:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.212 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32367.0 tgs: 270 data_time: 0.84s time: 119.53s eta: 10:47:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31439.0 tgs: 263 data_time: 0.62s time: 119.33s eta: 10:44:23 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31650.0 tgs: 264 data_time: 1.05s time: 119.49s eta: 10:43:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32244.0 tgs: 269 data_time: 0.89s time: 119.51s eta: 10:41:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.196 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31044.0 tgs: 260 data_time: 1.05s time: 118.99s eta: 10:36:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.182 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.87s time: 119.22s eta: 10:35:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32149.0 tgs: 270 data_time: 0.75s time: 119.06s eta: 10:32:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31111.0 tgs: 261 data_time: 0.83s time: 118.78s eta: 10:29:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30613.0 tgs: 255 data_time: 0.82s time: 119.86s eta: 10:33:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.117 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31475.0 tgs: 263 data_time: 0.84s time: 119.38s eta: 10:28:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31903.0 tgs: 268 data_time: 0.91s time: 118.83s eta: 10:23:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.124 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31353.0 tgs: 263 data_time: 0.49s time: 118.98s eta: 10:22:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.146 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32109.0 tgs: 269 data_time: 0.64s time: 119.28s eta: 10:22:14 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31351.0 tgs: 263 data_time: 0.78s time: 119.20s eta: 10:19:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31577.0 tgs: 265 data_time: 0.99s time: 119.07s eta: 10:17:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32081.0 tgs: 268 data_time: 0.81s time: 119.48s eta: 10:17:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31521.0 tgs: 263 data_time: 0.89s time: 119.41s eta: 10:14:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32025.0 tgs: 268 data_time: 0.82s time: 119.44s eta: 10:13:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31915.0 tgs: 267 data_time: 0.78s time: 119.29s eta: 10:10:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.198 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31465.0 tgs: 264 data_time: 0.66s time: 119.11s eta: 10:07:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.195 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30904.0 tgs: 258 data_time: 0.71s time: 119.70s eta: 10:08:29 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.272 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32158.0 tgs: 268 data_time: 1.02s time: 119.83s eta: 10:07:08 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 263 data_time: 1.11s time: 121.55s eta: 10:13:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31631.0 tgs: 265 data_time: 0.72s time: 119.25s eta: 10:00:14 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31174.0 tgs: 262 data_time: 0.84s time: 118.80s eta: 9:55:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.118 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32127.0 tgs: 268 data_time: 0.86s time: 119.56s eta: 9:57:48 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.211 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31380.0 tgs: 262 data_time: 0.85s time: 119.70s eta: 9:56:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.102 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 262 data_time: 0.58s time: 119.26s eta: 9:52:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.203 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31612.0 tgs: 264 data_time: 0.77s time: 119.33s eta: 9:50:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32502.0 tgs: 272 data_time: 0.80s time: 119.22s eta: 9:48:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31153.0 tgs: 257 data_time: 0.69s time: 120.88s eta: 9:54:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.151 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31774.0 tgs: 266 data_time: 0.64s time: 119.27s eta: 9:44:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31461.0 tgs: 263 data_time: 0.65s time: 119.46s eta: 9:43:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31674.0 tgs: 264 data_time: 0.58s time: 119.54s eta: 9:41:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30423.0 tgs: 255 data_time: 0.71s time: 119.20s eta: 9:38:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 269 data_time: 0.64s time: 119.48s eta: 9:37:27 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31530.0 tgs: 264 data_time: 0.57s time: 119.05s eta: 9:33:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31115.0 tgs: 260 data_time: 0.70s time: 119.24s eta: 9:32:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 267 data_time: 0.82s time: 119.16s eta: 9:29:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32559.0 tgs: 273 data_time: 0.71s time: 119.22s eta: 9:28:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 266 data_time: 0.82s time: 119.46s eta: 9:27:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32349.0 tgs: 271 data_time: 0.74s time: 118.99s eta: 9:23:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.193 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 264 data_time: 0.72s time: 119.03s eta: 9:21:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31970.0 tgs: 267 data_time: 0.87s time: 119.53s eta: 9:21:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31238.0 tgs: 261 data_time: 0.73s time: 119.59s eta: 9:20:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30878.0 tgs: 259 data_time: 0.46s time: 119.13s eta: 9:15:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32095.0 tgs: 269 data_time: 1.02s time: 119.05s eta: 9:13:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30782.0 tgs: 258 data_time: 0.86s time: 118.88s eta: 9:10:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.99s time: 119.32s eta: 9:10:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31507.0 tgs: 264 data_time: 1.11s time: 119.09s eta: 9:07:48 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32289.0 tgs: 271 data_time: 0.77s time: 118.86s eta: 9:04:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.197 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 265 data_time: 0.75s time: 119.67s eta: 9:06:28 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 265 data_time: 0.87s time: 119.44s eta: 9:03:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.120 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31804.0 tgs: 266 data_time: 0.84s time: 119.43s eta: 9:01:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31150.0 tgs: 262 data_time: 0.81s time: 118.74s eta: 8:56:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.109 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31503.0 tgs: 265 data_time: 0.63s time: 118.87s eta: 8:54:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30717.0 tgs: 257 data_time: 0.74s time: 119.33s eta: 8:55:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 271 data_time: 0.95s time: 119.14s eta: 8:52:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.200 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 270 data_time: 0.79s time: 119.27s eta: 8:50:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 267 data_time: 0.58s time: 120.47s eta: 8:54:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32001.0 tgs: 268 data_time: 0.60s time: 119.24s eta: 8:46:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30960.0 tgs: 259 data_time: 0.73s time: 119.09s eta: 8:44:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31557.0 tgs: 261 data_time: 0.74s time: 120.82s eta: 8:49:36 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31578.0 tgs: 262 data_time: 0.93s time: 120.20s eta: 8:44:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.119 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 272 data_time: 0.79s time: 118.82s eta: 8:36:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.199 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 264 data_time: 0.91s time: 119.35s eta: 8:37:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.117 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31091.0 tgs: 260 data_time: 0.88s time: 119.33s eta: 8:35:06 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31802.0 tgs: 267 data_time: 0.79s time: 119.05s eta: 8:31:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29331.0 tgs: 246 data_time: 0.89s time: 119.16s eta: 8:30:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.182 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 267 data_time: 0.64s time: 119.04s eta: 8:27:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.90s time: 119.12s eta: 8:26:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.211 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31622.0 tgs: 265 data_time: 0.80s time: 119.24s eta: 8:24:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32132.0 tgs: 269 data_time: 0.72s time: 119.22s eta: 8:22:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30444.0 tgs: 255 data_time: 0.80s time: 119.38s eta: 8:21:23 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29681.0 tgs: 247 data_time: 0.71s time: 119.77s eta: 8:21:03 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.193 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31112.0 tgs: 261 data_time: 0.63s time: 119.20s eta: 8:16:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.206 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31417.0 tgs: 263 data_time: 0.97s time: 119.02s eta: 8:13:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 265 data_time: 0.49s time: 119.89s eta: 8:15:33 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31857.0 tgs: 266 data_time: 0.97s time: 119.55s eta: 8:12:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31694.0 tgs: 265 data_time: 0.83s time: 119.60s eta: 8:10:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31688.0 tgs: 266 data_time: 0.83s time: 118.99s eta: 8:05:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 268 data_time: 1.18s time: 119.84s eta: 8:07:22 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.212 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.87s time: 119.70s eta: 8:04:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31108.0 tgs: 260 data_time: 0.60s time: 119.40s eta: 8:01:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.203 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30617.0 tgs: 256 data_time: 0.71s time: 119.28s eta: 7:59:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32412.0 tgs: 271 data_time: 0.79s time: 119.19s eta: 7:56:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.276 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31155.0 tgs: 261 data_time: 0.82s time: 119.19s eta: 7:54:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31249.0 tgs: 262 data_time: 0.73s time: 118.94s eta: 7:51:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31767.0 tgs: 267 data_time: 0.60s time: 118.97s eta: 7:49:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32200.0 tgs: 269 data_time: 0.81s time: 119.42s eta: 7:49:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31028.0 tgs: 260 data_time: 0.74s time: 119.03s eta: 7:46:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30943.0 tgs: 259 data_time: 0.77s time: 119.10s eta: 7:44:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.187 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27601.0 tgs: 229 data_time: 0.98s time: 120.02s eta: 7:46:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.195 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 269 data_time: 0.81s time: 119.83s eta: 7:43:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.198 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31357.0 tgs: 263 data_time: 0.86s time: 119.16s eta: 7:38:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32042.0 tgs: 269 data_time: 0.66s time: 118.96s eta: 7:36:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30972.0 tgs: 259 data_time: 0.81s time: 119.32s eta: 7:35:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 264 data_time: 0.84s time: 119.38s eta: 7:33:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 268 data_time: 0.67s time: 119.06s eta: 7:30:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32006.0 tgs: 268 data_time: 0.72s time: 118.99s eta: 7:28:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31985.0 tgs: 268 data_time: 0.74s time: 119.23s eta: 7:27:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32083.0 tgs: 269 data_time: 0.52s time: 118.99s eta: 7:24:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 270 data_time: 0.68s time: 118.93s eta: 7:22:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.177 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 270 data_time: 0.75s time: 119.08s eta: 7:20:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31264.0 tgs: 262 data_time: 0.81s time: 119.01s eta: 7:18:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30557.0 tgs: 256 data_time: 0.87s time: 119.01s eta: 7:16:23 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32378.0 tgs: 271 data_time: 0.69s time: 119.25s eta: 7:15:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.185 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31049.0 tgs: 260 data_time: 0.69s time: 119.36s eta: 7:13:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31609.0 tgs: 265 data_time: 0.99s time: 118.94s eta: 7:10:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 269 data_time: 0.83s time: 119.20s eta: 7:09:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 268 data_time: 0.63s time: 119.58s eta: 7:08:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.113 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31888.0 tgs: 266 data_time: 0.66s time: 119.66s eta: 7:06:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.211 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31603.0 tgs: 265 data_time: 0.84s time: 118.95s eta: 7:02:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.232 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31844.0 tgs: 266 data_time: 1.27s time: 119.30s eta: 7:01:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31175.0 tgs: 260 data_time: 0.71s time: 119.52s eta: 7:00:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32251.0 tgs: 270 data_time: 0.62s time: 119.42s eta: 6:57:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32177.0 tgs: 270 data_time: 0.74s time: 118.91s eta: 6:54:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30656.0 tgs: 256 data_time: 0.76s time: 119.34s eta: 6:53:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31257.0 tgs: 262 data_time: 0.63s time: 118.93s eta: 6:50:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.189 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31953.0 tgs: 267 data_time: 0.68s time: 119.26s eta: 6:49:28 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32291.0 tgs: 271 data_time: 0.78s time: 119.02s eta: 6:46:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.113 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31443.0 tgs: 263 data_time: 0.66s time: 119.23s eta: 6:45:22 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31661.0 tgs: 265 data_time: 0.71s time: 119.42s eta: 6:44:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.83s time: 119.01s eta: 6:40:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 267 data_time: 0.92s time: 119.01s eta: 6:38:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32275.0 tgs: 270 data_time: 0.77s time: 119.45s eta: 6:38:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31728.0 tgs: 267 data_time: 0.88s time: 118.80s eta: 6:34:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30487.0 tgs: 255 data_time: 0.85s time: 119.43s eta: 6:34:08 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32181.0 tgs: 269 data_time: 0.57s time: 119.57s eta: 6:32:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31879.0 tgs: 267 data_time: 0.89s time: 119.38s eta: 6:29:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31770.0 tgs: 265 data_time: 0.69s time: 119.56s eta: 6:28:33 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31713.0 tgs: 265 data_time: 0.89s time: 119.37s eta: 6:25:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31751.0 tgs: 267 data_time: 0.81s time: 118.82s eta: 6:22:13 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31094.0 tgs: 261 data_time: 0.76s time: 119.01s eta: 6:20:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30046.0 tgs: 251 data_time: 0.75s time: 119.40s eta: 6:20:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30496.0 tgs: 255 data_time: 0.92s time: 119.34s eta: 6:17:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.90s time: 119.05s eta: 6:15:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32366.0 tgs: 271 data_time: 0.80s time: 119.43s eta: 6:14:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.71s time: 119.39s eta: 6:12:06 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.188 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32038.0 tgs: 268 data_time: 0.75s time: 119.28s eta: 6:09:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.100 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32204.0 tgs: 270 data_time: 0.68s time: 119.10s eta: 6:07:13 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 264 data_time: 0.85s time: 121.96s eta: 6:13:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31634.0 tgs: 264 data_time: 0.87s time: 119.39s eta: 6:04:07 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 267 data_time: 0.83s time: 119.57s eta: 6:02:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31460.0 tgs: 263 data_time: 0.44s time: 119.25s eta: 5:59:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.88s time: 119.26s eta: 5:57:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.095 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31519.0 tgs: 265 data_time: 0.92s time: 118.91s eta: 5:54:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.195 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31931.0 tgs: 268 data_time: 0.87s time: 119.02s eta: 5:53:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31804.0 tgs: 267 data_time: 0.92s time: 119.07s eta: 5:51:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31811.0 tgs: 266 data_time: 0.72s time: 119.29s eta: 5:49:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32211.0 tgs: 270 data_time: 0.91s time: 118.97s eta: 5:46:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31563.0 tgs: 264 data_time: 0.98s time: 119.27s eta: 5:45:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31523.0 tgs: 264 data_time: 0.84s time: 119.36s eta: 5:44:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.113 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31092.0 tgs: 261 data_time: 0.76s time: 119.06s eta: 5:41:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.198 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31645.0 tgs: 264 data_time: 0.63s time: 119.49s eta: 5:40:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 269 data_time: 0.96s time: 119.38s eta: 5:38:14 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32149.0 tgs: 268 data_time: 0.71s time: 119.57s eta: 5:36:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32055.0 tgs: 269 data_time: 0.64s time: 118.98s eta: 5:33:08 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31677.0 tgs: 266 data_time: 0.70s time: 118.95s eta: 5:31:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 266 data_time: 0.89s time: 119.23s eta: 5:29:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31308.0 tgs: 262 data_time: 0.60s time: 119.40s eta: 5:28:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31346.0 tgs: 262 data_time: 0.96s time: 119.46s eta: 5:26:31 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 266 data_time: 0.75s time: 118.95s eta: 5:23:08 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 267 data_time: 0.75s time: 119.63s eta: 5:23:00 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31885.0 tgs: 267 data_time: 0.57s time: 119.35s eta: 5:20:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31464.0 tgs: 264 data_time: 0.86s time: 119.11s eta: 5:17:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.104 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30545.0 tgs: 255 data_time: 0.53s time: 119.42s eta: 5:16:28 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30973.0 tgs: 259 data_time: 1.00s time: 119.58s eta: 5:14:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32125.0 tgs: 269 data_time: 0.94s time: 119.00s eta: 5:11:23 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32641.0 tgs: 274 data_time: 0.61s time: 119.00s eta: 5:09:23 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31741.0 tgs: 266 data_time: 0.70s time: 119.23s eta: 5:07:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.221 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 263 data_time: 0.60s time: 119.68s eta: 5:07:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31365.0 tgs: 263 data_time: 0.82s time: 119.11s eta: 5:03:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 269 data_time: 0.81s time: 119.07s eta: 5:01:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.111 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31410.0 tgs: 258 data_time: 0.90s time: 121.52s eta: 5:05:48 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31644.0 tgs: 264 data_time: 0.86s time: 119.57s eta: 4:58:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30798.0 tgs: 257 data_time: 0.65s time: 119.50s eta: 4:56:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31825.0 tgs: 265 data_time: 0.72s time: 119.73s eta: 4:55:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31122.0 tgs: 260 data_time: 0.87s time: 119.52s eta: 4:52:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 269 data_time: 0.96s time: 119.41s eta: 4:50:34 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32125.0 tgs: 269 data_time: 0.60s time: 119.20s eta: 4:48:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.199 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32413.0 tgs: 272 data_time: 0.62s time: 118.88s eta: 4:45:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31172.0 tgs: 261 data_time: 0.78s time: 119.12s eta: 4:43:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 265 data_time: 1.15s time: 119.98s eta: 4:43:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31651.0 tgs: 265 data_time: 0.81s time: 119.15s eta: 4:39:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32070.0 tgs: 269 data_time: 0.91s time: 119.01s eta: 4:37:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31895.0 tgs: 267 data_time: 0.81s time: 119.06s eta: 4:35:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31039.0 tgs: 260 data_time: 0.80s time: 119.37s eta: 4:34:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 263 data_time: 0.61s time: 119.41s eta: 4:32:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31654.0 tgs: 266 data_time: 0.60s time: 118.93s eta: 4:29:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 262 data_time: 0.84s time: 119.60s eta: 4:29:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31160.0 tgs: 261 data_time: 0.71s time: 119.24s eta: 4:26:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.200 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31996.0 tgs: 268 data_time: 0.96s time: 119.06s eta: 4:23:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.194 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30189.0 tgs: 253 data_time: 0.71s time: 118.87s eta: 4:21:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31631.0 tgs: 266 data_time: 0.85s time: 118.74s eta: 4:19:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.250 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31359.0 tgs: 262 data_time: 0.70s time: 119.43s eta: 4:18:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31656.0 tgs: 265 data_time: 0.66s time: 119.40s eta: 4:16:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31990.0 tgs: 268 data_time: 0.63s time: 119.34s eta: 4:14:35 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31922.0 tgs: 268 data_time: 0.69s time: 119.05s eta: 4:11:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 270 data_time: 0.79s time: 119.12s eta: 4:10:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32120.0 tgs: 269 data_time: 0.70s time: 119.14s eta: 4:08:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31279.0 tgs: 262 data_time: 0.81s time: 118.96s eta: 4:05:50 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30901.0 tgs: 259 data_time: 0.71s time: 119.15s eta: 4:04:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31748.0 tgs: 265 data_time: 0.87s time: 119.49s eta: 4:02:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.84s time: 119.24s eta: 4:00:28 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31431.0 tgs: 263 data_time: 0.65s time: 119.19s eta: 3:58:23 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28711.0 tgs: 240 data_time: 0.55s time: 119.36s eta: 3:56:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.106 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31444.0 tgs: 263 data_time: 0.89s time: 119.37s eta: 3:54:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 269 data_time: 0.61s time: 119.95s eta: 3:53:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31520.0 tgs: 264 data_time: 0.56s time: 119.32s eta: 3:50:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30647.0 tgs: 256 data_time: 0.69s time: 119.63s eta: 3:49:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31752.0 tgs: 267 data_time: 0.61s time: 118.85s eta: 3:45:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31262.0 tgs: 261 data_time: 0.68s time: 119.75s eta: 3:45:31 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31481.0 tgs: 263 data_time: 0.66s time: 119.32s eta: 3:42:44 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31925.0 tgs: 267 data_time: 0.69s time: 119.26s eta: 3:40:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31074.0 tgs: 261 data_time: 0.70s time: 119.00s eta: 3:38:09 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31522.0 tgs: 263 data_time: 0.74s time: 119.43s eta: 3:36:57 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 267 data_time: 0.53s time: 119.03s eta: 3:34:15 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30871.0 tgs: 259 data_time: 0.70s time: 118.99s eta: 3:32:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 261 data_time: 1.04s time: 119.22s eta: 3:30:36 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30661.0 tgs: 255 data_time: 0.91s time: 120.11s eta: 3:30:11 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31667.0 tgs: 265 data_time: 0.73s time: 119.13s eta: 3:26:29 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30884.0 tgs: 258 data_time: 0.62s time: 119.31s eta: 3:24:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31589.0 tgs: 265 data_time: 0.64s time: 119.08s eta: 3:22:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31608.0 tgs: 264 data_time: 0.60s time: 119.43s eta: 3:21:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 266 data_time: 0.76s time: 119.29s eta: 3:18:48 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32385.0 tgs: 271 data_time: 1.09s time: 119.22s eta: 3:16:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32058.0 tgs: 268 data_time: 0.77s time: 119.18s eta: 3:14:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32233.0 tgs: 270 data_time: 0.70s time: 119.02s eta: 3:12:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31515.0 tgs: 264 data_time: 0.89s time: 119.18s eta: 3:10:41 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 262 data_time: 0.69s time: 121.05s eta: 3:11:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31599.0 tgs: 264 data_time: 0.75s time: 119.50s eta: 3:07:13 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31312.0 tgs: 262 data_time: 0.81s time: 119.26s eta: 3:04:51 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31694.0 tgs: 266 data_time: 0.76s time: 118.98s eta: 3:02:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31602.0 tgs: 265 data_time: 0.98s time: 119.22s eta: 3:00:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 269 data_time: 0.67s time: 119.45s eta: 2:59:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31861.0 tgs: 267 data_time: 0.88s time: 118.98s eta: 2:56:29 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.204 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32357.0 tgs: 271 data_time: 0.82s time: 119.35s eta: 2:55:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31459.0 tgs: 263 data_time: 0.67s time: 119.41s eta: 2:53:08 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31680.0 tgs: 265 data_time: 0.96s time: 119.35s eta: 2:51:04 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31085.0 tgs: 255 data_time: 0.66s time: 121.46s eta: 2:52:03 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31704.0 tgs: 265 data_time: 0.63s time: 119.24s eta: 2:46:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 267 data_time: 0.91s time: 119.57s eta: 2:45:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 269 data_time: 0.86s time: 118.90s eta: 2:42:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31401.0 tgs: 263 data_time: 0.80s time: 119.19s eta: 2:40:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31741.0 tgs: 266 data_time: 0.84s time: 118.96s eta: 2:38:36 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31177.0 tgs: 261 data_time: 0.69s time: 119.23s eta: 2:36:58 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.94s time: 119.11s eta: 2:34:50 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31404.0 tgs: 262 data_time: 0.73s time: 119.43s eta: 2:33:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30851.0 tgs: 256 data_time: 0.90s time: 120.23s eta: 2:32:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 264 data_time: 0.92s time: 119.28s eta: 2:29:05 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 270 data_time: 1.01s time: 118.88s eta: 2:26:37 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29997.0 tgs: 250 data_time: 0.57s time: 119.51s eta: 2:25:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31188.0 tgs: 261 data_time: 0.56s time: 119.07s eta: 2:22:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.220 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30698.0 tgs: 257 data_time: 0.88s time: 119.14s eta: 2:20:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31945.0 tgs: 268 data_time: 0.72s time: 119.04s eta: 2:18:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31510.0 tgs: 262 data_time: 0.58s time: 119.84s eta: 2:17:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.266 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 270 data_time: 0.74s time: 119.36s eta: 2:15:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31723.0 tgs: 265 data_time: 0.63s time: 119.39s eta: 2:13:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 262 data_time: 0.73s time: 119.68s eta: 2:11:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32311.0 tgs: 270 data_time: 0.54s time: 119.49s eta: 2:09:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 264 data_time: 0.82s time: 118.95s eta: 2:06:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.91s time: 119.38s eta: 2:05:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31314.0 tgs: 262 data_time: 0.66s time: 119.27s eta: 2:03:14 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32188.0 tgs: 269 data_time: 0.81s time: 119.22s eta: 2:01:12 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 264 data_time: 0.56s time: 119.50s eta: 1:59:30 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31628.0 tgs: 264 data_time: 0.73s time: 119.42s eta: 1:57:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32191.0 tgs: 269 data_time: 0.99s time: 119.53s eta: 1:55:32 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32189.0 tgs: 269 data_time: 0.65s time: 119.30s eta: 1:53:20 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31456.0 tgs: 263 data_time: 1.12s time: 119.25s eta: 1:51:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31457.0 tgs: 264 data_time: 0.73s time: 118.93s eta: 1:49:01 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31595.0 tgs: 264 data_time: 0.58s time: 119.64s eta: 1:47:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31715.0 tgs: 264 data_time: 0.86s time: 119.72s eta: 1:45:45 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.266 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31227.0 tgs: 261 data_time: 0.70s time: 119.35s eta: 1:43:26 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31517.0 tgs: 264 data_time: 0.58s time: 119.30s eta: 1:41:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.217 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30828.0 tgs: 258 data_time: 1.07s time: 119.28s eta: 1:39:24 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 266 data_time: 0.84s time: 118.90s eta: 1:37:06 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 0.76s time: 119.16s eta: 1:35:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32169.0 tgs: 270 data_time: 0.90s time: 119.07s eta: 1:33:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32048.0 tgs: 268 data_time: 0.76s time: 119.25s eta: 1:31:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 264 data_time: 0.83s time: 119.72s eta: 1:29:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 270 data_time: 0.92s time: 119.02s eta: 1:27:16 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 268 data_time: 0.59s time: 118.86s eta: 1:25:10 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31771.0 tgs: 265 data_time: 0.61s time: 119.48s eta: 1:23:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31409.0 tgs: 262 data_time: 0.66s time: 119.48s eta: 1:21:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32078.0 tgs: 269 data_time: 0.97s time: 118.94s eta: 1:19:17 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30835.0 tgs: 259 data_time: 0.53s time: 118.93s eta: 1:17:18 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 259 data_time: 0.64s time: 122.16s eta: 1:17:21 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29971.0 tgs: 251 data_time: 0.68s time: 119.37s eta: 1:13:36 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32196.0 tgs: 270 data_time: 0.61s time: 119.05s eta: 1:11:25 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31560.0 tgs: 265 data_time: 0.70s time: 118.91s eta: 1:09:22 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32256.0 tgs: 270 data_time: 0.69s time: 119.10s eta: 1:07:29 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31973.0 tgs: 267 data_time: 0.64s time: 119.41s eta: 1:05:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30809.0 tgs: 258 data_time: 0.76s time: 119.34s eta: 1:03:39 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32157.0 tgs: 269 data_time: 0.74s time: 119.18s eta: 1:01:34 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.193 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.54s time: 119.28s eta: 0:59:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31766.0 tgs: 265 data_time: 0.92s time: 119.44s eta: 0:57:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32283.0 tgs: 270 data_time: 0.82s time: 119.40s eta: 0:55:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 269 data_time: 0.72s time: 119.48s eta: 0:53:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31809.0 tgs: 266 data_time: 0.61s time: 119.36s eta: 0:51:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31292.0 tgs: 261 data_time: 0.91s time: 119.47s eta: 0:49:46 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31407.0 tgs: 264 data_time: 0.77s time: 118.89s eta: 0:47:33 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.84s time: 119.28s eta: 0:45:43 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32159.0 tgs: 269 data_time: 0.88s time: 119.21s eta: 0:43:42 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31686.0 tgs: 266 data_time: 0.74s time: 118.99s eta: 0:41:38 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31867.0 tgs: 267 data_time: 0.88s time: 119.04s eta: 0:39:40 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32209.0 tgs: 269 data_time: 0.78s time: 119.66s eta: 0:37:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31218.0 tgs: 261 data_time: 0.51s time: 119.28s eta: 0:35:47 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31653.0 tgs: 264 data_time: 0.72s time: 119.74s eta: 0:33:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31428.0 tgs: 259 data_time: 0.70s time: 121.23s eta: 0:32:19 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.59s time: 119.37s eta: 0:29:50 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.205 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32149.0 tgs: 269 data_time: 1.00s time: 119.32s eta: 0:27:50 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31200.0 tgs: 260 data_time: 0.69s time: 119.64s eta: 0:25:55 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31894.0 tgs: 267 data_time: 0.65s time: 119.09s eta: 0:23:49 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31019.0 tgs: 259 data_time: 0.89s time: 119.38s eta: 0:21:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31951.0 tgs: 267 data_time: 0.79s time: 119.26s eta: 0:19:52 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.66s time: 119.24s eta: 0:17:53 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 263 data_time: 0.61s time: 119.91s eta: 0:15:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32221.0 tgs: 270 data_time: 0.85s time: 119.21s eta: 0:13:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.209 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30866.0 tgs: 259 data_time: 0.75s time: 119.09s eta: 0:11:54 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31751.0 tgs: 266 data_time: 0.52s time: 119.32s eta: 0:09:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30864.0 tgs: 259 data_time: 1.03s time: 119.17s eta: 0:07:56 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32549.0 tgs: 271 data_time: 0.52s time: 119.69s eta: 0:05:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31389.0 tgs: 258 data_time: 0.61s time: 121.44s eta: 0:04:02 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31510.0 tgs: 264 data_time: 0.64s time: 119.22s eta: 0:01:59 +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.26 seconds, peak gpu memory 17.4G +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 29][DP 7][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank31.log b/20250121165312/rank31.log new file mode 100644 index 0000000000000000000000000000000000000000..fc5b039856948c5055977a56103809f88796ef49 --- /dev/null +++ b/20250121165312/rank31.log @@ -0,0 +1,851 @@ +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.03s +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.67 seconds, peak gpu memory 13.4G +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.255 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 32271.0 tgs: 210 data_time: 1.75s time: 153.37s eta: 1 day, 1:15:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.267 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32224.0 tgs: 269 data_time: 0.93s time: 119.78s eta: 19:41:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.243 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31186.0 tgs: 260 data_time: 0.95s time: 119.68s eta: 19:38:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.200 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32092.0 tgs: 268 data_time: 0.82s time: 119.65s eta: 19:36:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.191 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 265 data_time: 0.77s time: 119.30s eta: 19:31:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.230 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31263.0 tgs: 261 data_time: 0.73s time: 119.53s eta: 19:31:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.296 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30097.0 tgs: 244 data_time: 0.80s time: 122.96s eta: 20:02:58 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.268 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 270 data_time: 0.88s time: 119.89s eta: 19:30:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.201 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30582.0 tgs: 253 data_time: 0.80s time: 120.65s eta: 19:36:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.231 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32372.0 tgs: 269 data_time: 0.79s time: 119.92s eta: 19:27:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.225 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 267 data_time: 0.67s time: 120.46s eta: 19:30:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31025.0 tgs: 259 data_time: 0.62s time: 119.34s eta: 19:17:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31804.0 tgs: 265 data_time: 0.88s time: 119.67s eta: 19:18:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32389.0 tgs: 271 data_time: 0.81s time: 119.18s eta: 19:12:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.258 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30946.0 tgs: 259 data_time: 0.89s time: 119.22s eta: 19:10:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31432.0 tgs: 264 data_time: 0.91s time: 118.93s eta: 19:05:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.250 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32194.0 tgs: 270 data_time: 0.74s time: 119.14s eta: 19:05:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31342.0 tgs: 262 data_time: 1.07s time: 119.18s eta: 19:04:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32054.0 tgs: 268 data_time: 0.68s time: 119.39s eta: 19:04:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.233 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.74s time: 119.33s eta: 19:01:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31497.0 tgs: 261 data_time: 0.61s time: 120.31s eta: 19:08:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32299.0 tgs: 262 data_time: 0.79s time: 122.91s eta: 19:31:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 268 data_time: 0.64s time: 119.48s eta: 18:57:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31131.0 tgs: 261 data_time: 0.62s time: 119.16s eta: 18:52:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.75s time: 119.24s eta: 18:50:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32152.0 tgs: 269 data_time: 0.66s time: 119.39s eta: 18:50:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31336.0 tgs: 261 data_time: 0.98s time: 119.66s eta: 18:50:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31412.0 tgs: 263 data_time: 0.71s time: 119.38s eta: 18:46:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 268 data_time: 0.70s time: 119.79s eta: 18:48:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 245 data_time: 0.84s time: 129.33s eta: 20:15:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31765.0 tgs: 265 data_time: 0.86s time: 119.57s eta: 18:41:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31949.0 tgs: 267 data_time: 0.94s time: 119.39s eta: 18:38:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 259 data_time: 0.61s time: 120.66s eta: 18:48:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31518.0 tgs: 263 data_time: 0.95s time: 119.40s eta: 18:34:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.131 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 268 data_time: 0.92s time: 119.16s eta: 18:30:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32039.0 tgs: 269 data_time: 1.22s time: 119.07s eta: 18:27:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.77s time: 119.13s eta: 18:25:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32352.0 tgs: 268 data_time: 0.73s time: 120.39s eta: 18:35:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.233 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31385.0 tgs: 259 data_time: 1.03s time: 121.17s eta: 18:40:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 261 data_time: 0.70s time: 120.66s eta: 18:34:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31223.0 tgs: 260 data_time: 0.75s time: 119.74s eta: 18:23:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32577.0 tgs: 262 data_time: 1.04s time: 123.87s eta: 18:59:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31503.0 tgs: 263 data_time: 0.76s time: 119.55s eta: 18:17:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 263 data_time: 0.73s time: 119.35s eta: 18:14:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31080.0 tgs: 260 data_time: 0.83s time: 119.22s eta: 18:10:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31762.0 tgs: 265 data_time: 0.73s time: 119.60s eta: 18:12:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31859.0 tgs: 267 data_time: 0.85s time: 119.27s eta: 18:07:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32052.0 tgs: 269 data_time: 0.60s time: 119.01s eta: 18:03:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31795.0 tgs: 266 data_time: 0.78s time: 119.09s eta: 18:01:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31424.0 tgs: 263 data_time: 0.77s time: 119.26s eta: 18:01:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.71s time: 119.33s eta: 17:59:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32221.0 tgs: 268 data_time: 0.90s time: 120.02s eta: 18:04:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31787.0 tgs: 266 data_time: 0.86s time: 119.10s eta: 17:53:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31727.0 tgs: 265 data_time: 0.66s time: 119.41s eta: 17:54:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31831.0 tgs: 267 data_time: 0.90s time: 118.99s eta: 17:48:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32054.0 tgs: 269 data_time: 0.93s time: 119.03s eta: 17:47:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31492.0 tgs: 251 data_time: 0.82s time: 125.36s eta: 18:41:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31059.0 tgs: 258 data_time: 0.69s time: 120.10s eta: 17:52:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30232.0 tgs: 253 data_time: 0.78s time: 119.39s eta: 17:44:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31767.0 tgs: 266 data_time: 0.83s time: 119.34s eta: 17:42:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.57s time: 119.51s eta: 17:41:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 255 data_time: 0.65s time: 126.05s eta: 18:37:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31753.0 tgs: 258 data_time: 0.43s time: 122.63s eta: 18:05:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31116.0 tgs: 256 data_time: 0.69s time: 121.15s eta: 17:50:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31233.0 tgs: 262 data_time: 0.68s time: 119.03s eta: 17:29:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32074.0 tgs: 269 data_time: 0.72s time: 118.86s eta: 17:25:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 256 data_time: 0.80s time: 123.99s eta: 18:09:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32008.0 tgs: 256 data_time: 0.71s time: 124.57s eta: 18:12:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32207.0 tgs: 269 data_time: 0.76s time: 119.39s eta: 17:24:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.270 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31813.0 tgs: 265 data_time: 0.66s time: 119.69s eta: 17:25:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31610.0 tgs: 264 data_time: 0.64s time: 119.53s eta: 17:21:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.232 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 262 data_time: 0.71s time: 118.86s eta: 17:14:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 265 data_time: 0.98s time: 119.85s eta: 17:20:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31473.0 tgs: 263 data_time: 0.87s time: 119.56s eta: 17:16:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32047.0 tgs: 262 data_time: 0.75s time: 121.87s eta: 17:34:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.128 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29030.0 tgs: 235 data_time: 0.61s time: 123.51s eta: 17:46:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.72s time: 119.01s eta: 17:05:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.82s time: 119.23s eta: 17:05:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32013.0 tgs: 268 data_time: 0.56s time: 119.14s eta: 17:02:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30187.0 tgs: 253 data_time: 0.74s time: 119.25s eta: 17:01:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31309.0 tgs: 263 data_time: 0.78s time: 119.04s eta: 16:57:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 266 data_time: 0.76s time: 119.27s eta: 16:57:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31701.0 tgs: 266 data_time: 0.58s time: 119.16s eta: 16:54:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 269 data_time: 0.66s time: 119.58s eta: 16:56:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31201.0 tgs: 261 data_time: 0.69s time: 119.13s eta: 16:50:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31038.0 tgs: 261 data_time: 0.91s time: 118.84s eta: 16:46:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31025.0 tgs: 260 data_time: 0.81s time: 119.01s eta: 16:45:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30099.0 tgs: 251 data_time: 0.85s time: 119.85s eta: 16:50:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.92s time: 119.38s eta: 16:44:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31723.0 tgs: 266 data_time: 0.79s time: 119.01s eta: 16:39:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31759.0 tgs: 262 data_time: 0.87s time: 120.92s eta: 16:53:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30751.0 tgs: 258 data_time: 0.58s time: 119.00s eta: 16:35:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31819.0 tgs: 266 data_time: 0.62s time: 119.28s eta: 16:36:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31779.0 tgs: 266 data_time: 0.69s time: 119.23s eta: 16:33:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31279.0 tgs: 262 data_time: 0.75s time: 119.36s eta: 16:32:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31568.0 tgs: 263 data_time: 0.82s time: 119.70s eta: 16:33:32 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31989.0 tgs: 268 data_time: 0.86s time: 119.20s eta: 16:27:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31654.0 tgs: 264 data_time: 0.65s time: 119.76s eta: 16:30:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31764.0 tgs: 263 data_time: 0.76s time: 120.35s eta: 16:32:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31774.0 tgs: 267 data_time: 0.69s time: 118.97s eta: 16:19:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31673.0 tgs: 248 data_time: 0.71s time: 127.32s eta: 17:26:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.236 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.86s time: 119.22s eta: 16:17:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.94s time: 119.15s eta: 16:15:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.124 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 259 data_time: 0.52s time: 119.24s eta: 16:13:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31441.0 tgs: 263 data_time: 0.77s time: 119.19s eta: 16:11:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 270 data_time: 0.90s time: 118.95s eta: 16:07:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32186.0 tgs: 270 data_time: 0.93s time: 118.86s eta: 16:04:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.221 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31747.0 tgs: 266 data_time: 0.74s time: 118.97s eta: 16:03:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31339.0 tgs: 257 data_time: 0.70s time: 121.83s eta: 16:24:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.61s time: 118.99s eta: 15:59:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 269 data_time: 0.85s time: 119.22s eta: 15:59:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31503.0 tgs: 263 data_time: 1.00s time: 119.61s eta: 16:00:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31520.0 tgs: 264 data_time: 0.74s time: 119.04s eta: 15:54:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.201 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31714.0 tgs: 266 data_time: 0.77s time: 118.98s eta: 15:51:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.233 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31219.0 tgs: 261 data_time: 0.95s time: 119.17s eta: 15:51:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.249 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31548.0 tgs: 263 data_time: 0.71s time: 119.63s eta: 15:53:05 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 269 data_time: 0.88s time: 118.94s eta: 15:45:32 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31262.0 tgs: 262 data_time: 0.88s time: 119.29s eta: 15:46:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31993.0 tgs: 268 data_time: 0.58s time: 119.12s eta: 15:43:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31357.0 tgs: 263 data_time: 0.87s time: 119.00s eta: 15:40:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31058.0 tgs: 259 data_time: 0.85s time: 119.66s eta: 15:43:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32204.0 tgs: 265 data_time: 0.69s time: 121.50s eta: 15:55:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32363.0 tgs: 271 data_time: 0.65s time: 119.41s eta: 15:37:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.205 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 242 data_time: 0.76s time: 129.96s eta: 16:58:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31435.0 tgs: 263 data_time: 0.88s time: 119.47s eta: 15:33:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.208 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 270 data_time: 0.79s time: 119.60s eta: 15:32:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 266 data_time: 0.83s time: 119.53s eta: 15:30:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 259 data_time: 0.68s time: 122.14s eta: 15:48:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.234 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30593.0 tgs: 253 data_time: 0.85s time: 120.75s eta: 15:35:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31383.0 tgs: 262 data_time: 0.65s time: 119.41s eta: 15:23:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31513.0 tgs: 263 data_time: 0.57s time: 119.44s eta: 15:21:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32587.0 tgs: 247 data_time: 0.90s time: 131.63s eta: 16:53:32 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 264 data_time: 0.62s time: 119.30s eta: 15:16:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.114 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30622.0 tgs: 256 data_time: 0.71s time: 119.56s eta: 15:16:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30873.0 tgs: 252 data_time: 1.11s time: 122.26s eta: 15:35:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31673.0 tgs: 264 data_time: 0.59s time: 119.71s eta: 15:13:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30890.0 tgs: 258 data_time: 1.00s time: 119.35s eta: 15:09:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31272.0 tgs: 263 data_time: 0.77s time: 118.90s eta: 15:03:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32168.0 tgs: 270 data_time: 0.94s time: 119.00s eta: 15:02:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32521.0 tgs: 272 data_time: 0.79s time: 119.30s eta: 15:02:42 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.84s time: 118.92s eta: 14:57:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 263 data_time: 0.68s time: 120.52s eta: 15:07:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31767.0 tgs: 266 data_time: 0.64s time: 119.31s eta: 14:56:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.238 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31769.0 tgs: 259 data_time: 1.14s time: 122.65s eta: 15:19:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32206.0 tgs: 270 data_time: 0.80s time: 119.12s eta: 14:51:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30760.0 tgs: 256 data_time: 0.92s time: 119.75s eta: 14:54:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32005.0 tgs: 267 data_time: 0.54s time: 119.46s eta: 14:50:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.212 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31508.0 tgs: 264 data_time: 0.69s time: 118.99s eta: 14:44:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31893.0 tgs: 267 data_time: 0.66s time: 119.33s eta: 14:45:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31235.0 tgs: 262 data_time: 0.52s time: 118.93s eta: 14:40:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31995.0 tgs: 268 data_time: 0.69s time: 119.05s eta: 14:38:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32011.0 tgs: 268 data_time: 0.76s time: 119.06s eta: 14:37:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30863.0 tgs: 259 data_time: 0.65s time: 119.13s eta: 14:35:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30213.0 tgs: 253 data_time: 0.68s time: 119.00s eta: 14:32:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31367.0 tgs: 263 data_time: 0.85s time: 119.07s eta: 14:31:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.227 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31269.0 tgs: 262 data_time: 0.69s time: 119.15s eta: 14:29:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31668.0 tgs: 264 data_time: 0.61s time: 119.65s eta: 14:31:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31958.0 tgs: 268 data_time: 0.70s time: 119.18s eta: 14:26:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32164.0 tgs: 269 data_time: 0.75s time: 119.20s eta: 14:24:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.79s time: 119.39s eta: 14:23:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31959.0 tgs: 264 data_time: 0.83s time: 120.73s eta: 14:31:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.199 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32334.0 tgs: 272 data_time: 0.72s time: 118.85s eta: 14:15:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.184 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31346.0 tgs: 263 data_time: 0.85s time: 119.10s eta: 14:15:32 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31361.0 tgs: 263 data_time: 0.66s time: 119.15s eta: 14:13:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.339 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31668.0 tgs: 263 data_time: 0.96s time: 119.96s eta: 14:17:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31098.0 tgs: 259 data_time: 0.80s time: 119.87s eta: 14:15:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30925.0 tgs: 258 data_time: 0.53s time: 119.58s eta: 14:11:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32117.0 tgs: 269 data_time: 0.62s time: 119.22s eta: 14:06:28 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31798.0 tgs: 266 data_time: 0.74s time: 119.27s eta: 14:04:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.74s time: 119.26s eta: 14:02:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32114.0 tgs: 269 data_time: 0.89s time: 119.14s eta: 13:59:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31630.0 tgs: 265 data_time: 0.60s time: 119.12s eta: 13:57:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32387.0 tgs: 271 data_time: 0.95s time: 119.17s eta: 13:56:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 266 data_time: 0.66s time: 119.35s eta: 13:55:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31302.0 tgs: 260 data_time: 0.75s time: 120.11s eta: 13:58:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31527.0 tgs: 257 data_time: 0.62s time: 122.55s eta: 14:13:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:56:46][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32428.0 tgs: 272 data_time: 0.73s time: 119.11s eta: 13:47:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.207 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31703.0 tgs: 265 data_time: 0.60s time: 119.37s eta: 13:47:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.94s time: 119.35s eta: 13:45:28 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31817.0 tgs: 266 data_time: 0.71s time: 119.35s eta: 13:43:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.119 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31491.0 tgs: 264 data_time: 0.54s time: 119.11s eta: 13:39:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31439.0 tgs: 263 data_time: 0.88s time: 119.25s eta: 13:38:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 1.01s time: 119.13s eta: 13:36:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.228 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31671.0 tgs: 265 data_time: 0.87s time: 119.22s eta: 13:34:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30523.0 tgs: 255 data_time: 0.87s time: 119.45s eta: 13:34:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31916.0 tgs: 268 data_time: 0.62s time: 119.08s eta: 13:29:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31049.0 tgs: 261 data_time: 0.41s time: 118.92s eta: 13:26:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31921.0 tgs: 266 data_time: 0.50s time: 119.81s eta: 13:30:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.188 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 268 data_time: 0.82s time: 119.25s eta: 13:24:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.89s time: 119.16s eta: 13:22:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.203 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32195.0 tgs: 269 data_time: 0.81s time: 119.35s eta: 13:21:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31241.0 tgs: 261 data_time: 0.86s time: 119.26s eta: 13:19:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 270 data_time: 0.74s time: 119.04s eta: 13:15:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.79s time: 119.37s eta: 13:15:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30896.0 tgs: 259 data_time: 0.81s time: 119.18s eta: 13:12:32 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 269 data_time: 0.81s time: 119.46s eta: 13:12:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30668.0 tgs: 257 data_time: 0.84s time: 119.29s eta: 13:09:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 264 data_time: 0.68s time: 119.15s eta: 13:06:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 265 data_time: 0.68s time: 120.46s eta: 13:13:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 265 data_time: 0.78s time: 120.16s eta: 13:09:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32202.0 tgs: 269 data_time: 0.73s time: 119.50s eta: 13:02:42 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27721.0 tgs: 231 data_time: 0.91s time: 119.51s eta: 13:00:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.201 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31795.0 tgs: 265 data_time: 0.84s time: 119.57s eta: 12:59:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31254.0 tgs: 260 data_time: 0.63s time: 119.79s eta: 12:58:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31574.0 tgs: 265 data_time: 0.74s time: 119.04s eta: 12:51:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31155.0 tgs: 261 data_time: 0.82s time: 119.05s eta: 12:49:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31056.0 tgs: 260 data_time: 0.59s time: 119.13s eta: 12:48:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32326.0 tgs: 270 data_time: 0.83s time: 119.40s eta: 12:48:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 264 data_time: 0.70s time: 119.19s eta: 12:44:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.231 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32123.0 tgs: 268 data_time: 0.93s time: 119.44s eta: 12:44:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31749.0 tgs: 266 data_time: 0.65s time: 119.09s eta: 12:40:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.193 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30997.0 tgs: 260 data_time: 0.81s time: 119.07s eta: 12:38:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.194 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32055.0 tgs: 268 data_time: 0.75s time: 119.59s eta: 12:39:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 273 data_time: 0.75s time: 118.78s eta: 12:32:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 270 data_time: 0.74s time: 118.88s eta: 12:30:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31814.0 tgs: 266 data_time: 0.77s time: 119.17s eta: 12:30:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31922.0 tgs: 267 data_time: 0.81s time: 119.26s eta: 12:29:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.113 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32157.0 tgs: 269 data_time: 0.58s time: 119.35s eta: 12:27:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31834.0 tgs: 267 data_time: 0.70s time: 119.12s eta: 12:24:28 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31074.0 tgs: 260 data_time: 0.63s time: 119.34s eta: 12:23:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.265 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32137.0 tgs: 269 data_time: 0.64s time: 119.27s eta: 12:21:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31326.0 tgs: 263 data_time: 0.83s time: 119.02s eta: 12:17:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31646.0 tgs: 261 data_time: 0.62s time: 121.13s eta: 12:28:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.201 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31682.0 tgs: 264 data_time: 0.82s time: 119.99s eta: 12:19:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31959.0 tgs: 269 data_time: 0.61s time: 118.72s eta: 12:10:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30787.0 tgs: 257 data_time: 0.80s time: 119.63s eta: 12:13:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31823.0 tgs: 265 data_time: 0.74s time: 119.83s eta: 12:12:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.215 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31642.0 tgs: 265 data_time: 0.58s time: 119.14s eta: 12:06:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32138.0 tgs: 269 data_time: 0.65s time: 119.17s eta: 12:04:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32136.0 tgs: 269 data_time: 0.57s time: 119.11s eta: 12:02:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.71s time: 119.18s eta: 12:01:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 267 data_time: 0.70s time: 118.95s eta: 11:57:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 266 data_time: 0.62s time: 119.71s eta: 12:00:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 267 data_time: 0.82s time: 119.27s eta: 11:55:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 266 data_time: 0.89s time: 119.27s eta: 11:53:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31851.0 tgs: 266 data_time: 0.59s time: 119.71s eta: 11:54:14 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31892.0 tgs: 268 data_time: 0.84s time: 118.78s eta: 11:46:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31879.0 tgs: 267 data_time: 0.76s time: 119.16s eta: 11:47:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.108 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31023.0 tgs: 259 data_time: 0.73s time: 119.61s eta: 11:47:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32492.0 tgs: 273 data_time: 0.70s time: 118.88s eta: 11:41:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.188 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30868.0 tgs: 260 data_time: 0.62s time: 118.67s eta: 11:38:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.217 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31405.0 tgs: 263 data_time: 0.93s time: 119.21s eta: 11:39:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31589.0 tgs: 265 data_time: 0.91s time: 119.14s eta: 11:36:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 269 data_time: 0.68s time: 119.07s eta: 11:34:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31521.0 tgs: 265 data_time: 0.92s time: 118.95s eta: 11:31:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 269 data_time: 0.81s time: 118.91s eta: 11:29:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.189 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32398.0 tgs: 272 data_time: 0.79s time: 118.97s eta: 11:28:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31334.0 tgs: 262 data_time: 0.60s time: 119.28s eta: 11:27:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32549.0 tgs: 272 data_time: 0.69s time: 119.35s eta: 11:26:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32111.0 tgs: 270 data_time: 0.77s time: 118.89s eta: 11:21:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.74s time: 119.20s eta: 11:21:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 28070.0 tgs: 234 data_time: 0.67s time: 119.72s eta: 11:22:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 0.64s time: 119.12s eta: 11:16:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30675.0 tgs: 257 data_time: 0.61s time: 119.12s eta: 11:14:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32604.0 tgs: 273 data_time: 0.83s time: 119.00s eta: 11:12:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31034.0 tgs: 260 data_time: 0.69s time: 119.08s eta: 11:10:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.175 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31335.0 tgs: 262 data_time: 0.61s time: 119.16s eta: 11:09:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 267 data_time: 0.76s time: 118.97s eta: 11:06:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32310.0 tgs: 270 data_time: 0.85s time: 119.47s eta: 11:07:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.193 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31737.0 tgs: 265 data_time: 0.95s time: 119.64s eta: 11:05:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 269 data_time: 0.52s time: 119.30s eta: 11:02:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.197 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 265 data_time: 0.98s time: 119.31s eta: 11:00:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31822.0 tgs: 267 data_time: 0.75s time: 118.82s eta: 10:55:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28741.0 tgs: 240 data_time: 0.71s time: 119.47s eta: 10:57:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.131 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31552.0 tgs: 264 data_time: 0.81s time: 119.37s eta: 10:54:31 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31005.0 tgs: 259 data_time: 0.55s time: 119.28s eta: 10:52:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31325.0 tgs: 262 data_time: 0.62s time: 119.26s eta: 10:49:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 270 data_time: 0.62s time: 119.10s eta: 10:47:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32367.0 tgs: 270 data_time: 0.81s time: 119.53s eta: 10:47:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31439.0 tgs: 263 data_time: 0.60s time: 119.33s eta: 10:44:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31650.0 tgs: 264 data_time: 0.99s time: 119.49s eta: 10:43:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32244.0 tgs: 269 data_time: 0.84s time: 119.51s eta: 10:41:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31044.0 tgs: 260 data_time: 1.03s time: 118.99s eta: 10:36:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.86s time: 119.22s eta: 10:35:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.230 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32149.0 tgs: 270 data_time: 0.74s time: 119.06s eta: 10:33:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.110 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31111.0 tgs: 261 data_time: 0.82s time: 118.78s eta: 10:29:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30613.0 tgs: 255 data_time: 0.79s time: 119.86s eta: 10:33:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.184 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31475.0 tgs: 263 data_time: 0.80s time: 119.38s eta: 10:28:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31903.0 tgs: 268 data_time: 0.89s time: 118.83s eta: 10:23:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.119 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31353.0 tgs: 263 data_time: 0.47s time: 118.98s eta: 10:22:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32109.0 tgs: 269 data_time: 0.61s time: 119.28s eta: 10:22:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31351.0 tgs: 263 data_time: 0.76s time: 119.20s eta: 10:19:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31577.0 tgs: 265 data_time: 0.98s time: 119.06s eta: 10:17:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32081.0 tgs: 268 data_time: 0.81s time: 119.48s eta: 10:17:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.198 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31521.0 tgs: 263 data_time: 0.86s time: 119.41s eta: 10:14:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32025.0 tgs: 268 data_time: 0.80s time: 119.45s eta: 10:13:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31915.0 tgs: 267 data_time: 0.75s time: 119.29s eta: 10:10:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31465.0 tgs: 264 data_time: 0.64s time: 119.11s eta: 10:07:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30904.0 tgs: 258 data_time: 0.69s time: 119.70s eta: 10:08:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32158.0 tgs: 268 data_time: 0.99s time: 119.83s eta: 10:07:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 263 data_time: 1.07s time: 121.55s eta: 10:13:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31631.0 tgs: 265 data_time: 0.71s time: 119.25s eta: 10:00:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.205 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31174.0 tgs: 262 data_time: 0.81s time: 118.80s eta: 9:56:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32127.0 tgs: 268 data_time: 0.84s time: 119.56s eta: 9:57:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.191 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31380.0 tgs: 262 data_time: 0.84s time: 119.70s eta: 9:56:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 262 data_time: 0.57s time: 119.26s eta: 9:52:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31612.0 tgs: 264 data_time: 0.75s time: 119.32s eta: 9:50:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.194 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32502.0 tgs: 272 data_time: 0.77s time: 119.22s eta: 9:48:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31153.0 tgs: 259 data_time: 0.66s time: 119.97s eta: 9:49:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.185 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31774.0 tgs: 266 data_time: 0.61s time: 119.27s eta: 9:44:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31461.0 tgs: 263 data_time: 0.60s time: 119.46s eta: 9:43:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31674.0 tgs: 264 data_time: 0.55s time: 119.54s eta: 9:41:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30423.0 tgs: 255 data_time: 0.68s time: 119.20s eta: 9:38:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 269 data_time: 0.60s time: 119.48s eta: 9:37:28 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31530.0 tgs: 264 data_time: 0.54s time: 119.05s eta: 9:33:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31115.0 tgs: 260 data_time: 0.67s time: 119.24s eta: 9:32:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 267 data_time: 0.77s time: 119.16s eta: 9:29:58 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.172 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32559.0 tgs: 273 data_time: 0.70s time: 119.22s eta: 9:28:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 266 data_time: 0.80s time: 119.46s eta: 9:27:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32349.0 tgs: 271 data_time: 0.71s time: 118.98s eta: 9:23:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 264 data_time: 0.69s time: 119.03s eta: 9:21:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.170 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31970.0 tgs: 267 data_time: 0.84s time: 119.54s eta: 9:21:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.242 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31238.0 tgs: 261 data_time: 0.71s time: 119.59s eta: 9:20:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.110 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30878.0 tgs: 259 data_time: 0.44s time: 119.13s eta: 9:15:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32095.0 tgs: 269 data_time: 0.97s time: 119.05s eta: 9:13:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30782.0 tgs: 258 data_time: 0.84s time: 118.88s eta: 9:10:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.191 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.95s time: 119.32s eta: 9:10:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.219 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31507.0 tgs: 264 data_time: 1.17s time: 119.09s eta: 9:07:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32289.0 tgs: 271 data_time: 0.76s time: 118.86s eta: 9:04:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.246 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 265 data_time: 0.74s time: 119.67s eta: 9:06:28 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 265 data_time: 0.85s time: 119.43s eta: 9:03:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31804.0 tgs: 266 data_time: 0.82s time: 119.43s eta: 9:01:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31150.0 tgs: 262 data_time: 0.80s time: 118.73s eta: 8:56:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31503.0 tgs: 265 data_time: 0.60s time: 118.86s eta: 8:54:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30717.0 tgs: 257 data_time: 0.75s time: 119.34s eta: 8:55:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 271 data_time: 0.94s time: 119.14s eta: 8:52:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32209.0 tgs: 270 data_time: 0.79s time: 119.27s eta: 8:50:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 267 data_time: 0.57s time: 120.46s eta: 8:54:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32001.0 tgs: 268 data_time: 0.58s time: 119.23s eta: 8:46:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30960.0 tgs: 259 data_time: 0.72s time: 119.09s eta: 8:44:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31557.0 tgs: 261 data_time: 0.74s time: 120.83s eta: 8:49:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31578.0 tgs: 262 data_time: 0.91s time: 120.19s eta: 8:44:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 272 data_time: 0.74s time: 118.83s eta: 8:36:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.223 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 264 data_time: 0.88s time: 119.35s eta: 8:37:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31091.0 tgs: 260 data_time: 0.82s time: 119.33s eta: 8:35:05 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31802.0 tgs: 267 data_time: 0.76s time: 119.05s eta: 8:31:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.209 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29331.0 tgs: 246 data_time: 0.84s time: 119.16s eta: 8:30:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 267 data_time: 0.61s time: 119.04s eta: 8:27:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.177 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.87s time: 119.12s eta: 8:26:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31622.0 tgs: 265 data_time: 0.78s time: 119.24s eta: 8:24:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32132.0 tgs: 269 data_time: 0.70s time: 119.22s eta: 8:22:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30444.0 tgs: 255 data_time: 0.79s time: 119.38s eta: 8:21:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29681.0 tgs: 247 data_time: 0.69s time: 119.77s eta: 8:21:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31112.0 tgs: 261 data_time: 0.61s time: 119.20s eta: 8:16:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31417.0 tgs: 263 data_time: 0.94s time: 119.02s eta: 8:13:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 265 data_time: 0.47s time: 119.90s eta: 8:15:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31857.0 tgs: 266 data_time: 0.93s time: 119.55s eta: 8:12:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31694.0 tgs: 265 data_time: 0.82s time: 119.59s eta: 8:10:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31688.0 tgs: 266 data_time: 0.80s time: 119.00s eta: 8:05:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 268 data_time: 1.16s time: 119.84s eta: 8:07:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.244 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 0.85s time: 119.70s eta: 8:04:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31108.0 tgs: 260 data_time: 0.58s time: 119.40s eta: 8:01:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30617.0 tgs: 256 data_time: 0.67s time: 119.29s eta: 7:59:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.107 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32412.0 tgs: 271 data_time: 0.74s time: 119.19s eta: 7:56:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.165 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31155.0 tgs: 261 data_time: 0.79s time: 119.20s eta: 7:54:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.112 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31249.0 tgs: 262 data_time: 0.72s time: 118.93s eta: 7:51:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31767.0 tgs: 267 data_time: 0.58s time: 118.97s eta: 7:49:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.248 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32200.0 tgs: 269 data_time: 0.79s time: 119.41s eta: 7:49:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.189 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31028.0 tgs: 260 data_time: 0.72s time: 119.03s eta: 7:46:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.217 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30943.0 tgs: 259 data_time: 0.75s time: 119.10s eta: 7:44:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27601.0 tgs: 229 data_time: 0.96s time: 120.02s eta: 7:46:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.207 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 269 data_time: 0.79s time: 119.82s eta: 7:43:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31357.0 tgs: 263 data_time: 0.86s time: 119.17s eta: 7:38:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32042.0 tgs: 269 data_time: 0.63s time: 118.96s eta: 7:36:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30972.0 tgs: 259 data_time: 0.80s time: 119.32s eta: 7:35:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 264 data_time: 0.81s time: 119.38s eta: 7:33:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 268 data_time: 0.65s time: 119.06s eta: 7:30:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32006.0 tgs: 268 data_time: 0.71s time: 118.99s eta: 7:28:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31985.0 tgs: 268 data_time: 0.71s time: 119.24s eta: 7:27:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.150 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32083.0 tgs: 269 data_time: 0.50s time: 118.98s eta: 7:24:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 270 data_time: 0.66s time: 118.93s eta: 7:22:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 270 data_time: 0.74s time: 119.08s eta: 7:20:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31264.0 tgs: 262 data_time: 0.79s time: 119.00s eta: 7:18:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30557.0 tgs: 256 data_time: 0.84s time: 119.02s eta: 7:16:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.208 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32378.0 tgs: 271 data_time: 0.68s time: 119.26s eta: 7:15:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31049.0 tgs: 260 data_time: 0.67s time: 119.36s eta: 7:13:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31609.0 tgs: 265 data_time: 0.96s time: 118.94s eta: 7:10:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.192 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 269 data_time: 0.80s time: 119.20s eta: 7:09:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 268 data_time: 0.61s time: 119.59s eta: 7:08:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31888.0 tgs: 266 data_time: 0.64s time: 119.66s eta: 7:06:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31603.0 tgs: 265 data_time: 0.81s time: 118.95s eta: 7:02:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31844.0 tgs: 266 data_time: 1.25s time: 119.30s eta: 7:01:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31175.0 tgs: 260 data_time: 0.70s time: 119.52s eta: 7:00:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32251.0 tgs: 270 data_time: 0.60s time: 119.42s eta: 6:57:58 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32177.0 tgs: 270 data_time: 0.72s time: 118.91s eta: 6:54:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30656.0 tgs: 256 data_time: 0.74s time: 119.34s eta: 6:53:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31257.0 tgs: 262 data_time: 0.61s time: 118.93s eta: 6:50:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31953.0 tgs: 267 data_time: 0.67s time: 119.26s eta: 6:49:28 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.193 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32291.0 tgs: 271 data_time: 0.76s time: 119.02s eta: 6:46:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31443.0 tgs: 263 data_time: 0.63s time: 119.23s eta: 6:45:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31661.0 tgs: 265 data_time: 0.66s time: 119.41s eta: 6:44:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.81s time: 119.01s eta: 6:40:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 267 data_time: 0.87s time: 119.01s eta: 6:38:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32275.0 tgs: 270 data_time: 0.73s time: 119.45s eta: 6:38:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31728.0 tgs: 267 data_time: 0.86s time: 118.80s eta: 6:34:01 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30487.0 tgs: 255 data_time: 0.82s time: 119.44s eta: 6:34:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32181.0 tgs: 269 data_time: 0.55s time: 119.57s eta: 6:32:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31879.0 tgs: 267 data_time: 0.85s time: 119.38s eta: 6:29:58 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31770.0 tgs: 266 data_time: 0.67s time: 119.16s eta: 6:27:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31713.0 tgs: 265 data_time: 0.85s time: 119.36s eta: 6:25:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31751.0 tgs: 267 data_time: 0.80s time: 118.82s eta: 6:22:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31094.0 tgs: 261 data_time: 0.73s time: 119.02s eta: 6:20:51 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30046.0 tgs: 251 data_time: 0.73s time: 119.40s eta: 6:20:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30496.0 tgs: 255 data_time: 0.87s time: 119.34s eta: 6:17:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.88s time: 119.05s eta: 6:15:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32366.0 tgs: 271 data_time: 0.77s time: 119.43s eta: 6:14:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32228.0 tgs: 269 data_time: 0.68s time: 119.39s eta: 6:12:05 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32038.0 tgs: 268 data_time: 0.73s time: 119.28s eta: 6:09:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.219 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32204.0 tgs: 270 data_time: 0.66s time: 119.10s eta: 6:07:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.201 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 264 data_time: 0.83s time: 121.96s eta: 6:14:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31634.0 tgs: 264 data_time: 0.85s time: 119.38s eta: 6:04:07 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 267 data_time: 0.80s time: 119.57s eta: 6:02:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31460.0 tgs: 263 data_time: 0.42s time: 119.25s eta: 5:59:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.114 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.85s time: 119.26s eta: 5:57:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31519.0 tgs: 265 data_time: 0.89s time: 118.91s eta: 5:54:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31931.0 tgs: 268 data_time: 0.85s time: 119.02s eta: 5:53:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31804.0 tgs: 267 data_time: 0.90s time: 119.07s eta: 5:51:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31811.0 tgs: 266 data_time: 0.70s time: 119.29s eta: 5:49:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32211.0 tgs: 270 data_time: 0.89s time: 118.97s eta: 5:46:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.242 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31563.0 tgs: 264 data_time: 0.91s time: 119.27s eta: 5:45:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31523.0 tgs: 264 data_time: 0.80s time: 119.36s eta: 5:44:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31092.0 tgs: 261 data_time: 0.74s time: 119.05s eta: 5:41:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31645.0 tgs: 264 data_time: 0.62s time: 119.49s eta: 5:40:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 269 data_time: 0.93s time: 119.38s eta: 5:38:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32149.0 tgs: 268 data_time: 0.68s time: 119.57s eta: 5:36:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32055.0 tgs: 269 data_time: 0.59s time: 118.98s eta: 5:33:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31677.0 tgs: 266 data_time: 0.67s time: 118.95s eta: 5:31:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 266 data_time: 0.88s time: 119.24s eta: 5:29:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31308.0 tgs: 262 data_time: 0.58s time: 119.40s eta: 5:28:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31346.0 tgs: 262 data_time: 0.95s time: 119.46s eta: 5:26:31 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 266 data_time: 0.71s time: 118.95s eta: 5:23:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 267 data_time: 0.73s time: 119.63s eta: 5:22:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31885.0 tgs: 267 data_time: 0.56s time: 119.35s eta: 5:20:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.199 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31464.0 tgs: 264 data_time: 0.83s time: 119.11s eta: 5:17:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.209 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30545.0 tgs: 255 data_time: 0.48s time: 119.42s eta: 5:16:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.103 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30973.0 tgs: 259 data_time: 0.96s time: 119.58s eta: 5:14:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32125.0 tgs: 269 data_time: 0.88s time: 119.00s eta: 5:11:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32641.0 tgs: 274 data_time: 0.58s time: 119.00s eta: 5:09:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31741.0 tgs: 266 data_time: 0.68s time: 119.23s eta: 5:07:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 263 data_time: 0.59s time: 119.68s eta: 5:07:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31365.0 tgs: 263 data_time: 0.79s time: 119.11s eta: 5:03:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.184 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 269 data_time: 0.79s time: 119.07s eta: 5:01:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31410.0 tgs: 258 data_time: 0.85s time: 121.52s eta: 5:05:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.219 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31644.0 tgs: 264 data_time: 0.83s time: 119.58s eta: 4:58:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30798.0 tgs: 257 data_time: 0.62s time: 119.49s eta: 4:56:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31825.0 tgs: 265 data_time: 0.71s time: 119.73s eta: 4:55:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31122.0 tgs: 260 data_time: 0.84s time: 119.52s eta: 4:52:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 269 data_time: 0.92s time: 119.41s eta: 4:50:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32125.0 tgs: 269 data_time: 0.58s time: 119.20s eta: 4:48:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32413.0 tgs: 272 data_time: 0.61s time: 118.88s eta: 4:45:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31172.0 tgs: 261 data_time: 0.74s time: 119.12s eta: 4:43:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.209 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 265 data_time: 1.11s time: 119.98s eta: 4:43:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31651.0 tgs: 265 data_time: 0.76s time: 119.14s eta: 4:39:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32070.0 tgs: 269 data_time: 0.89s time: 119.01s eta: 4:37:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.175 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31895.0 tgs: 267 data_time: 0.79s time: 119.06s eta: 4:35:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31039.0 tgs: 260 data_time: 0.79s time: 119.37s eta: 4:34:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 263 data_time: 0.59s time: 119.40s eta: 4:32:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31654.0 tgs: 266 data_time: 0.59s time: 118.94s eta: 4:29:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 262 data_time: 0.83s time: 119.60s eta: 4:29:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31160.0 tgs: 261 data_time: 0.69s time: 119.24s eta: 4:26:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31996.0 tgs: 268 data_time: 0.93s time: 119.06s eta: 4:23:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30189.0 tgs: 254 data_time: 0.68s time: 118.85s eta: 4:21:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31631.0 tgs: 266 data_time: 0.81s time: 118.76s eta: 4:19:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.192 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31359.0 tgs: 262 data_time: 0.67s time: 119.42s eta: 4:18:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31656.0 tgs: 265 data_time: 0.63s time: 119.40s eta: 4:16:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31990.0 tgs: 268 data_time: 0.61s time: 119.34s eta: 4:14:35 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31922.0 tgs: 268 data_time: 0.67s time: 119.05s eta: 4:11:58 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 270 data_time: 0.78s time: 119.12s eta: 4:10:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.207 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32120.0 tgs: 269 data_time: 0.67s time: 119.13s eta: 4:08:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31279.0 tgs: 262 data_time: 0.76s time: 118.96s eta: 4:05:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30901.0 tgs: 259 data_time: 0.69s time: 119.15s eta: 4:04:15 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31748.0 tgs: 265 data_time: 0.85s time: 119.50s eta: 4:02:58 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.81s time: 119.24s eta: 4:00:28 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31431.0 tgs: 263 data_time: 0.62s time: 119.19s eta: 3:58:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28711.0 tgs: 240 data_time: 0.54s time: 119.36s eta: 3:56:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31444.0 tgs: 263 data_time: 0.86s time: 119.37s eta: 3:54:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 269 data_time: 0.60s time: 119.95s eta: 3:53:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31520.0 tgs: 264 data_time: 0.55s time: 119.31s eta: 3:50:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30647.0 tgs: 256 data_time: 0.67s time: 119.64s eta: 3:49:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31752.0 tgs: 267 data_time: 0.59s time: 118.85s eta: 3:45:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31262.0 tgs: 261 data_time: 0.66s time: 119.75s eta: 3:45:31 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31481.0 tgs: 263 data_time: 0.64s time: 119.33s eta: 3:42:44 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31925.0 tgs: 267 data_time: 0.68s time: 119.26s eta: 3:40:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31074.0 tgs: 261 data_time: 0.69s time: 119.00s eta: 3:38:09 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.105 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31522.0 tgs: 263 data_time: 0.72s time: 119.43s eta: 3:36:57 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31831.0 tgs: 267 data_time: 0.51s time: 119.04s eta: 3:34:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30871.0 tgs: 259 data_time: 0.71s time: 118.98s eta: 3:32:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 261 data_time: 1.01s time: 119.22s eta: 3:30:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30661.0 tgs: 255 data_time: 0.86s time: 120.11s eta: 3:30:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.220 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31667.0 tgs: 265 data_time: 0.70s time: 119.13s eta: 3:26:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30884.0 tgs: 258 data_time: 0.60s time: 119.31s eta: 3:24:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31589.0 tgs: 265 data_time: 0.62s time: 119.07s eta: 3:22:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31608.0 tgs: 264 data_time: 0.58s time: 119.43s eta: 3:21:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 266 data_time: 0.73s time: 119.29s eta: 3:18:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32385.0 tgs: 271 data_time: 1.07s time: 119.21s eta: 3:16:42 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32058.0 tgs: 268 data_time: 0.76s time: 119.18s eta: 3:14:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32233.0 tgs: 270 data_time: 0.68s time: 119.02s eta: 3:12:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31515.0 tgs: 264 data_time: 0.86s time: 119.18s eta: 3:10:41 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 262 data_time: 0.66s time: 121.05s eta: 3:11:39 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31599.0 tgs: 264 data_time: 0.73s time: 119.50s eta: 3:07:13 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.218 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31312.0 tgs: 262 data_time: 0.80s time: 119.26s eta: 3:04:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:43:13][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31694.0 tgs: 266 data_time: 0.75s time: 118.99s eta: 3:02:27 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31602.0 tgs: 265 data_time: 0.96s time: 119.22s eta: 3:00:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.202 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 268 data_time: 0.65s time: 119.45s eta: 2:59:10 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.206 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31861.0 tgs: 267 data_time: 0.84s time: 118.98s eta: 2:56:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32357.0 tgs: 271 data_time: 0.80s time: 119.35s eta: 2:55:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31459.0 tgs: 263 data_time: 0.65s time: 119.41s eta: 2:53:08 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31680.0 tgs: 265 data_time: 0.90s time: 119.35s eta: 2:51:04 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31085.0 tgs: 255 data_time: 0.64s time: 121.46s eta: 2:52:03 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31704.0 tgs: 265 data_time: 0.60s time: 119.24s eta: 2:46:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 267 data_time: 0.89s time: 119.57s eta: 2:45:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 269 data_time: 0.84s time: 118.91s eta: 2:42:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31401.0 tgs: 263 data_time: 0.77s time: 119.18s eta: 2:40:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31741.0 tgs: 266 data_time: 0.81s time: 118.96s eta: 2:38:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31177.0 tgs: 261 data_time: 0.66s time: 119.23s eta: 2:36:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.192 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.92s time: 119.10s eta: 2:34:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31404.0 tgs: 262 data_time: 0.69s time: 119.43s eta: 2:33:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30851.0 tgs: 256 data_time: 0.88s time: 120.22s eta: 2:32:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31540.0 tgs: 264 data_time: 0.88s time: 119.28s eta: 2:29:05 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32184.0 tgs: 270 data_time: 0.98s time: 118.89s eta: 2:26:37 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29997.0 tgs: 251 data_time: 0.55s time: 119.51s eta: 2:25:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31188.0 tgs: 261 data_time: 0.54s time: 119.07s eta: 2:22:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30698.0 tgs: 257 data_time: 0.83s time: 119.14s eta: 2:20:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.232 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31945.0 tgs: 268 data_time: 0.69s time: 119.05s eta: 2:18:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31510.0 tgs: 262 data_time: 0.57s time: 119.84s eta: 2:17:48 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 270 data_time: 0.73s time: 119.36s eta: 2:15:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31723.0 tgs: 265 data_time: 0.62s time: 119.39s eta: 2:13:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.204 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 262 data_time: 0.71s time: 119.68s eta: 2:11:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32311.0 tgs: 270 data_time: 0.52s time: 119.49s eta: 2:09:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 264 data_time: 0.82s time: 118.96s eta: 2:06:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.193 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.88s time: 119.37s eta: 2:05:20 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31314.0 tgs: 262 data_time: 0.65s time: 119.27s eta: 2:03:14 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32188.0 tgs: 269 data_time: 0.80s time: 119.22s eta: 2:01:12 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 264 data_time: 0.55s time: 119.50s eta: 1:59:30 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31628.0 tgs: 264 data_time: 0.70s time: 119.42s eta: 1:57:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.259 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32191.0 tgs: 269 data_time: 0.96s time: 119.54s eta: 1:55:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32189.0 tgs: 269 data_time: 0.63s time: 119.30s eta: 1:53:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31456.0 tgs: 263 data_time: 1.07s time: 119.25s eta: 1:51:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31457.0 tgs: 264 data_time: 0.70s time: 118.93s eta: 1:49:00 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31595.0 tgs: 264 data_time: 0.55s time: 119.64s eta: 1:47:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31715.0 tgs: 264 data_time: 0.83s time: 119.72s eta: 1:45:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31227.0 tgs: 261 data_time: 0.68s time: 119.34s eta: 1:43:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31517.0 tgs: 264 data_time: 0.57s time: 119.31s eta: 1:41:24 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30828.0 tgs: 258 data_time: 1.05s time: 119.28s eta: 1:39:23 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 266 data_time: 0.83s time: 118.90s eta: 1:37:06 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 0.73s time: 119.16s eta: 1:35:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32169.0 tgs: 270 data_time: 0.88s time: 119.07s eta: 1:33:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.279 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32048.0 tgs: 268 data_time: 0.72s time: 119.25s eta: 1:31:25 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 264 data_time: 0.81s time: 119.72s eta: 1:29:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 270 data_time: 0.88s time: 119.01s eta: 1:27:16 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 268 data_time: 0.58s time: 118.87s eta: 1:25:11 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31771.0 tgs: 265 data_time: 0.59s time: 119.48s eta: 1:23:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31409.0 tgs: 262 data_time: 0.64s time: 119.48s eta: 1:21:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32078.0 tgs: 269 data_time: 0.94s time: 118.94s eta: 1:19:17 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30835.0 tgs: 259 data_time: 0.50s time: 118.93s eta: 1:17:18 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 259 data_time: 0.61s time: 122.16s eta: 1:17:21 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29971.0 tgs: 251 data_time: 0.65s time: 119.37s eta: 1:13:36 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32196.0 tgs: 270 data_time: 0.59s time: 119.06s eta: 1:11:26 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31560.0 tgs: 265 data_time: 0.67s time: 118.92s eta: 1:09:22 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32256.0 tgs: 270 data_time: 0.67s time: 119.10s eta: 1:07:29 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31973.0 tgs: 267 data_time: 0.61s time: 119.41s eta: 1:05:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30809.0 tgs: 258 data_time: 0.72s time: 119.34s eta: 1:03:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32157.0 tgs: 269 data_time: 0.72s time: 119.18s eta: 1:01:34 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.53s time: 119.27s eta: 0:59:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31766.0 tgs: 265 data_time: 0.88s time: 119.44s eta: 0:57:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32283.0 tgs: 270 data_time: 0.81s time: 119.39s eta: 0:55:42 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.210 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 269 data_time: 0.69s time: 119.48s eta: 0:53:45 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31809.0 tgs: 266 data_time: 0.59s time: 119.36s eta: 0:51:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.223 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31292.0 tgs: 261 data_time: 0.88s time: 119.47s eta: 0:49:46 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31407.0 tgs: 264 data_time: 0.75s time: 118.90s eta: 0:47:33 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.79s time: 119.27s eta: 0:45:43 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32159.0 tgs: 269 data_time: 0.86s time: 119.22s eta: 0:43:42 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31686.0 tgs: 266 data_time: 0.73s time: 118.99s eta: 0:41:38 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31867.0 tgs: 267 data_time: 0.81s time: 119.04s eta: 0:39:40 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32209.0 tgs: 269 data_time: 0.78s time: 119.66s eta: 0:37:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31218.0 tgs: 261 data_time: 0.49s time: 119.28s eta: 0:35:47 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31653.0 tgs: 264 data_time: 0.71s time: 119.74s eta: 0:33:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31428.0 tgs: 259 data_time: 0.68s time: 121.23s eta: 0:32:19 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.56s time: 119.37s eta: 0:29:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32149.0 tgs: 269 data_time: 0.95s time: 119.32s eta: 0:27:50 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31200.0 tgs: 260 data_time: 0.67s time: 119.63s eta: 0:25:55 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31894.0 tgs: 267 data_time: 0.63s time: 119.09s eta: 0:23:49 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31019.0 tgs: 259 data_time: 0.86s time: 119.38s eta: 0:21:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.231 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31951.0 tgs: 267 data_time: 0.78s time: 119.26s eta: 0:19:52 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.64s time: 119.24s eta: 0:17:53 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 263 data_time: 0.58s time: 119.91s eta: 0:15:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32221.0 tgs: 270 data_time: 0.83s time: 119.21s eta: 0:13:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30866.0 tgs: 259 data_time: 0.73s time: 119.09s eta: 0:11:54 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.093 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31751.0 tgs: 266 data_time: 0.51s time: 119.32s eta: 0:09:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30864.0 tgs: 258 data_time: 0.99s time: 119.17s eta: 0:07:56 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32549.0 tgs: 271 data_time: 0.51s time: 119.69s eta: 0:05:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31389.0 tgs: 258 data_time: 0.56s time: 121.42s eta: 0:04:02 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31510.0 tgs: 264 data_time: 0.62s time: 119.21s eta: 0:01:59 +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.26 seconds, peak gpu memory 17.4G +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 31][DP 7][SP 3][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank34.log b/20250121165312/rank34.log new file mode 100644 index 0000000000000000000000000000000000000000..16e463d0abdfa707c40732c3b26c2411130d89e4 --- /dev/null +++ b/20250121165312/rank34.log @@ -0,0 +1,851 @@ +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:35][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.83 seconds, peak gpu memory 13.4G +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.222 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31323.0 tgs: 205 data_time: 1.98s time: 152.16s eta: 1 day, 1:03:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.276 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31388.0 tgs: 261 data_time: 0.81s time: 120.16s eta: 19:45:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.236 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31981.0 tgs: 267 data_time: 1.01s time: 119.69s eta: 19:38:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31503.0 tgs: 263 data_time: 1.28s time: 119.65s eta: 19:36:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31606.0 tgs: 264 data_time: 0.67s time: 119.31s eta: 19:31:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.221 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31195.0 tgs: 260 data_time: 0.75s time: 119.54s eta: 19:31:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.187 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31593.0 tgs: 256 data_time: 0.74s time: 122.96s eta: 20:02:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.204 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31176.0 tgs: 260 data_time: 0.66s time: 119.90s eta: 19:30:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.215 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32152.0 tgs: 266 data_time: 0.67s time: 120.66s eta: 19:36:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29940.0 tgs: 249 data_time: 0.66s time: 119.92s eta: 19:27:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.171 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31618.0 tgs: 262 data_time: 0.89s time: 120.46s eta: 19:30:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31523.0 tgs: 264 data_time: 0.64s time: 119.34s eta: 19:17:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.212 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31105.0 tgs: 259 data_time: 0.83s time: 119.68s eta: 19:18:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31143.0 tgs: 261 data_time: 0.52s time: 119.18s eta: 19:12:05 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31851.0 tgs: 267 data_time: 0.58s time: 119.19s eta: 19:10:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31775.0 tgs: 267 data_time: 0.86s time: 118.92s eta: 19:05:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31815.0 tgs: 267 data_time: 0.88s time: 119.15s eta: 19:05:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31366.0 tgs: 263 data_time: 0.80s time: 119.18s eta: 19:04:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31364.0 tgs: 262 data_time: 0.83s time: 119.40s eta: 19:04:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32004.0 tgs: 268 data_time: 0.95s time: 119.33s eta: 19:01:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.310 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 262 data_time: 0.84s time: 120.31s eta: 19:08:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31087.0 tgs: 252 data_time: 0.69s time: 122.92s eta: 19:31:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31662.0 tgs: 264 data_time: 0.61s time: 119.48s eta: 18:57:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31237.0 tgs: 262 data_time: 0.81s time: 119.16s eta: 18:51:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31187.0 tgs: 261 data_time: 0.76s time: 119.25s eta: 18:50:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.236 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31088.0 tgs: 260 data_time: 0.71s time: 119.39s eta: 18:50:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31662.0 tgs: 264 data_time: 0.79s time: 119.66s eta: 18:50:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32039.0 tgs: 268 data_time: 0.98s time: 119.38s eta: 18:46:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32252.0 tgs: 269 data_time: 0.61s time: 119.77s eta: 18:47:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 246 data_time: 1.02s time: 129.31s eta: 20:15:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 263 data_time: 0.66s time: 119.57s eta: 18:41:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 29894.0 tgs: 250 data_time: 0.84s time: 119.39s eta: 18:38:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31299.0 tgs: 259 data_time: 0.88s time: 120.66s eta: 18:48:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31641.0 tgs: 264 data_time: 0.86s time: 119.40s eta: 18:34:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.117 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31053.0 tgs: 260 data_time: 0.57s time: 119.16s eta: 18:30:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31452.0 tgs: 264 data_time: 0.83s time: 119.07s eta: 18:27:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31537.0 tgs: 264 data_time: 0.82s time: 119.14s eta: 18:25:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32333.0 tgs: 268 data_time: 0.89s time: 120.39s eta: 18:35:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32264.0 tgs: 266 data_time: 0.78s time: 121.17s eta: 18:40:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31571.0 tgs: 261 data_time: 0.64s time: 120.66s eta: 18:34:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32023.0 tgs: 267 data_time: 0.54s time: 119.74s eta: 18:23:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32305.0 tgs: 260 data_time: 0.79s time: 123.87s eta: 18:59:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32197.0 tgs: 269 data_time: 1.05s time: 119.57s eta: 18:18:00 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31127.0 tgs: 260 data_time: 0.61s time: 119.32s eta: 18:13:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31389.0 tgs: 263 data_time: 0.84s time: 119.23s eta: 18:10:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31307.0 tgs: 261 data_time: 0.77s time: 119.60s eta: 18:12:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 266 data_time: 1.21s time: 119.27s eta: 18:07:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31146.0 tgs: 261 data_time: 0.98s time: 119.01s eta: 18:02:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.245 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30841.0 tgs: 258 data_time: 0.84s time: 119.11s eta: 18:01:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30859.0 tgs: 258 data_time: 1.19s time: 119.26s eta: 18:01:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31436.0 tgs: 263 data_time: 0.87s time: 119.33s eta: 17:59:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32042.0 tgs: 266 data_time: 0.91s time: 120.02s eta: 18:04:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31299.0 tgs: 262 data_time: 0.65s time: 119.10s eta: 17:53:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.252 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31820.0 tgs: 266 data_time: 0.56s time: 119.41s eta: 17:54:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31916.0 tgs: 268 data_time: 0.68s time: 118.99s eta: 17:48:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30595.0 tgs: 257 data_time: 0.80s time: 119.03s eta: 17:47:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32038.0 tgs: 255 data_time: 0.73s time: 125.36s eta: 18:42:00 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31912.0 tgs: 265 data_time: 0.81s time: 120.11s eta: 17:52:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 267 data_time: 0.95s time: 119.36s eta: 17:44:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 269 data_time: 0.70s time: 119.32s eta: 17:41:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32223.0 tgs: 269 data_time: 0.83s time: 119.51s eta: 17:41:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32203.0 tgs: 255 data_time: 0.66s time: 126.05s eta: 18:37:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30992.0 tgs: 252 data_time: 0.72s time: 122.63s eta: 18:05:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.257 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32317.0 tgs: 266 data_time: 1.02s time: 121.16s eta: 17:50:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31408.0 tgs: 263 data_time: 0.75s time: 119.04s eta: 17:29:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31491.0 tgs: 264 data_time: 0.80s time: 118.86s eta: 17:25:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30711.0 tgs: 247 data_time: 0.87s time: 124.00s eta: 18:09:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30847.0 tgs: 247 data_time: 0.80s time: 124.57s eta: 18:12:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32021.0 tgs: 268 data_time: 0.93s time: 119.39s eta: 17:24:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.277 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32505.0 tgs: 271 data_time: 1.00s time: 119.70s eta: 17:25:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31080.0 tgs: 260 data_time: 0.92s time: 119.53s eta: 17:21:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31246.0 tgs: 262 data_time: 0.69s time: 118.86s eta: 17:14:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32177.0 tgs: 268 data_time: 0.94s time: 119.85s eta: 17:20:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.134 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31382.0 tgs: 262 data_time: 0.75s time: 119.52s eta: 17:15:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32159.0 tgs: 263 data_time: 0.84s time: 121.86s eta: 17:34:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31877.0 tgs: 258 data_time: 0.66s time: 123.51s eta: 17:46:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31207.0 tgs: 262 data_time: 0.66s time: 119.02s eta: 17:05:31 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30898.0 tgs: 259 data_time: 0.84s time: 119.23s eta: 17:05:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30827.0 tgs: 258 data_time: 0.82s time: 119.14s eta: 17:02:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31777.0 tgs: 266 data_time: 0.55s time: 119.26s eta: 17:01:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31137.0 tgs: 261 data_time: 0.89s time: 119.04s eta: 16:57:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31356.0 tgs: 262 data_time: 0.80s time: 119.28s eta: 16:57:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 270 data_time: 0.84s time: 119.16s eta: 16:54:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.263 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32076.0 tgs: 268 data_time: 1.07s time: 119.59s eta: 16:56:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31126.0 tgs: 261 data_time: 0.72s time: 119.13s eta: 16:50:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 266 data_time: 0.92s time: 118.84s eta: 16:46:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31389.0 tgs: 263 data_time: 0.72s time: 119.02s eta: 16:45:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31471.0 tgs: 262 data_time: 0.79s time: 119.85s eta: 16:50:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30848.0 tgs: 258 data_time: 0.85s time: 119.36s eta: 16:44:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 268 data_time: 0.85s time: 119.01s eta: 16:39:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32558.0 tgs: 269 data_time: 0.76s time: 120.93s eta: 16:53:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32033.0 tgs: 269 data_time: 0.75s time: 119.00s eta: 16:35:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.222 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32243.0 tgs: 270 data_time: 0.69s time: 119.28s eta: 16:36:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32206.0 tgs: 270 data_time: 0.90s time: 119.23s eta: 16:33:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32438.0 tgs: 271 data_time: 0.66s time: 119.37s eta: 16:32:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32226.0 tgs: 269 data_time: 0.93s time: 119.70s eta: 16:33:32 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.69s time: 119.21s eta: 16:27:27 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 264 data_time: 0.58s time: 119.76s eta: 16:30:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 267 data_time: 0.89s time: 120.37s eta: 16:33:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 267 data_time: 0.68s time: 118.97s eta: 16:19:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31282.0 tgs: 245 data_time: 0.81s time: 127.33s eta: 17:26:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31733.0 tgs: 266 data_time: 0.74s time: 119.23s eta: 16:17:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 260 data_time: 0.98s time: 119.14s eta: 16:14:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31006.0 tgs: 260 data_time: 0.79s time: 119.21s eta: 16:13:32 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31562.0 tgs: 264 data_time: 0.60s time: 119.17s eta: 16:11:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31002.0 tgs: 260 data_time: 0.75s time: 118.95s eta: 16:07:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 263 data_time: 0.97s time: 118.86s eta: 16:04:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32454.0 tgs: 272 data_time: 0.80s time: 118.96s eta: 16:03:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 265 data_time: 0.63s time: 121.84s eta: 16:24:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30862.0 tgs: 259 data_time: 0.61s time: 119.00s eta: 15:59:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31650.0 tgs: 265 data_time: 0.97s time: 119.23s eta: 15:59:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31524.0 tgs: 263 data_time: 0.76s time: 119.61s eta: 16:00:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30848.0 tgs: 259 data_time: 0.77s time: 119.04s eta: 15:54:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 267 data_time: 0.65s time: 118.98s eta: 15:51:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 1.05s time: 119.17s eta: 15:51:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29869.0 tgs: 249 data_time: 0.75s time: 119.64s eta: 15:53:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32048.0 tgs: 269 data_time: 0.75s time: 118.94s eta: 15:45:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 266 data_time: 0.76s time: 119.28s eta: 15:46:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31414.0 tgs: 263 data_time: 0.91s time: 119.11s eta: 15:42:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32163.0 tgs: 270 data_time: 0.61s time: 118.99s eta: 15:40:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30553.0 tgs: 255 data_time: 0.77s time: 119.67s eta: 15:43:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31786.0 tgs: 261 data_time: 0.74s time: 121.50s eta: 15:55:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32336.0 tgs: 270 data_time: 1.12s time: 119.41s eta: 15:37:23 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30519.0 tgs: 234 data_time: 0.67s time: 129.96s eta: 16:58:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32337.0 tgs: 270 data_time: 0.86s time: 119.47s eta: 15:33:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.114 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31199.0 tgs: 260 data_time: 0.85s time: 119.61s eta: 15:32:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 269 data_time: 0.80s time: 119.53s eta: 15:30:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.230 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32322.0 tgs: 264 data_time: 0.96s time: 122.14s eta: 15:48:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30720.0 tgs: 254 data_time: 0.69s time: 120.75s eta: 15:35:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30939.0 tgs: 259 data_time: 1.09s time: 119.42s eta: 15:23:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31698.0 tgs: 265 data_time: 0.74s time: 119.44s eta: 15:21:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31820.0 tgs: 241 data_time: 0.89s time: 131.63s eta: 16:53:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31395.0 tgs: 263 data_time: 0.60s time: 119.29s eta: 15:16:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31617.0 tgs: 264 data_time: 0.66s time: 119.52s eta: 15:16:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.125 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32474.0 tgs: 265 data_time: 0.60s time: 122.25s eta: 15:35:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31858.0 tgs: 266 data_time: 0.85s time: 119.71s eta: 15:13:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31342.0 tgs: 262 data_time: 0.92s time: 119.35s eta: 15:09:05 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31870.0 tgs: 268 data_time: 0.66s time: 118.90s eta: 15:03:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.227 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31221.0 tgs: 262 data_time: 0.87s time: 119.00s eta: 15:02:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 267 data_time: 0.70s time: 119.31s eta: 15:02:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31524.0 tgs: 265 data_time: 1.08s time: 118.92s eta: 14:57:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32396.0 tgs: 268 data_time: 0.74s time: 120.53s eta: 15:07:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31891.0 tgs: 267 data_time: 0.91s time: 119.31s eta: 14:56:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.234 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32178.0 tgs: 262 data_time: 1.15s time: 122.65s eta: 15:19:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31452.0 tgs: 264 data_time: 0.83s time: 119.13s eta: 14:51:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.292 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31484.0 tgs: 262 data_time: 1.05s time: 119.76s eta: 14:54:10 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31734.0 tgs: 265 data_time: 0.68s time: 119.47s eta: 14:50:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31121.0 tgs: 261 data_time: 0.90s time: 118.99s eta: 14:44:31 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32300.0 tgs: 270 data_time: 0.67s time: 119.31s eta: 14:44:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31009.0 tgs: 260 data_time: 0.89s time: 118.94s eta: 14:40:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.197 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 270 data_time: 0.92s time: 119.05s eta: 14:39:00 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 267 data_time: 0.87s time: 119.06s eta: 14:37:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.113 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 265 data_time: 0.55s time: 119.13s eta: 14:35:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32011.0 tgs: 268 data_time: 0.72s time: 119.00s eta: 14:32:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31413.0 tgs: 263 data_time: 0.61s time: 119.08s eta: 14:31:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31819.0 tgs: 267 data_time: 0.76s time: 119.15s eta: 14:29:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30523.0 tgs: 255 data_time: 0.77s time: 119.65s eta: 14:31:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 265 data_time: 0.71s time: 119.18s eta: 14:26:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32117.0 tgs: 269 data_time: 0.98s time: 119.20s eta: 14:24:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30801.0 tgs: 257 data_time: 0.75s time: 119.39s eta: 14:23:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31854.0 tgs: 263 data_time: 0.74s time: 120.73s eta: 14:31:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.210 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 266 data_time: 0.96s time: 118.85s eta: 14:15:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.120 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32012.0 tgs: 268 data_time: 0.78s time: 119.10s eta: 14:15:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32216.0 tgs: 270 data_time: 0.91s time: 119.11s eta: 14:13:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.195 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32399.0 tgs: 270 data_time: 0.89s time: 119.95s eta: 14:17:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30625.0 tgs: 255 data_time: 0.45s time: 119.87s eta: 14:15:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32129.0 tgs: 268 data_time: 0.65s time: 119.59s eta: 14:11:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.201 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 268 data_time: 0.89s time: 119.23s eta: 14:06:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31621.0 tgs: 265 data_time: 0.64s time: 119.26s eta: 14:04:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32198.0 tgs: 269 data_time: 0.86s time: 119.27s eta: 14:02:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32209.0 tgs: 270 data_time: 0.80s time: 119.14s eta: 13:59:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31991.0 tgs: 268 data_time: 0.83s time: 119.12s eta: 13:57:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 269 data_time: 0.95s time: 119.17s eta: 13:56:10 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31588.0 tgs: 264 data_time: 0.76s time: 119.35s eta: 13:55:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31117.0 tgs: 259 data_time: 0.57s time: 120.11s eta: 13:58:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31650.0 tgs: 258 data_time: 1.04s time: 122.56s eta: 14:13:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.90s time: 119.11s eta: 13:47:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.253 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30848.0 tgs: 258 data_time: 0.68s time: 119.37s eta: 13:47:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31447.0 tgs: 263 data_time: 0.88s time: 119.33s eta: 13:45:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31659.0 tgs: 265 data_time: 0.72s time: 119.33s eta: 13:43:23 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.202 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32096.0 tgs: 269 data_time: 0.71s time: 119.12s eta: 13:39:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31683.0 tgs: 265 data_time: 0.99s time: 119.26s eta: 13:38:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32452.0 tgs: 272 data_time: 0.86s time: 119.14s eta: 13:36:05 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31475.0 tgs: 263 data_time: 0.68s time: 119.22s eta: 13:34:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.182 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30320.0 tgs: 253 data_time: 0.84s time: 119.46s eta: 13:34:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32453.0 tgs: 272 data_time: 0.80s time: 119.09s eta: 13:29:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31923.0 tgs: 268 data_time: 0.72s time: 118.92s eta: 13:26:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31882.0 tgs: 266 data_time: 0.62s time: 119.81s eta: 13:30:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.191 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30850.0 tgs: 258 data_time: 0.88s time: 119.27s eta: 13:25:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.117 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31955.0 tgs: 268 data_time: 0.68s time: 119.15s eta: 13:22:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.221 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32059.0 tgs: 268 data_time: 0.83s time: 119.35s eta: 13:21:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32227.0 tgs: 270 data_time: 0.70s time: 119.26s eta: 13:19:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31072.0 tgs: 261 data_time: 0.90s time: 119.04s eta: 13:15:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31866.0 tgs: 266 data_time: 0.87s time: 119.35s eta: 13:15:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31014.0 tgs: 260 data_time: 0.73s time: 119.17s eta: 13:12:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31575.0 tgs: 264 data_time: 0.95s time: 119.46s eta: 13:12:23 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.191 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31526.0 tgs: 264 data_time: 0.67s time: 119.29s eta: 13:09:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31466.0 tgs: 264 data_time: 0.65s time: 119.15s eta: 13:06:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 265 data_time: 0.74s time: 120.42s eta: 13:12:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32275.0 tgs: 268 data_time: 0.59s time: 120.16s eta: 13:09:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32338.0 tgs: 270 data_time: 0.87s time: 119.50s eta: 13:02:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 266 data_time: 0.84s time: 119.51s eta: 13:00:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31084.0 tgs: 259 data_time: 0.71s time: 119.57s eta: 12:59:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.194 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 268 data_time: 1.08s time: 119.80s eta: 12:58:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31602.0 tgs: 265 data_time: 0.65s time: 119.04s eta: 12:51:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31746.0 tgs: 266 data_time: 0.78s time: 119.05s eta: 12:49:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32700.0 tgs: 274 data_time: 0.91s time: 119.12s eta: 12:48:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 266 data_time: 0.78s time: 119.40s eta: 12:48:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.107 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30903.0 tgs: 259 data_time: 0.57s time: 119.18s eta: 12:44:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31993.0 tgs: 267 data_time: 0.88s time: 119.41s eta: 12:44:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.112 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31147.0 tgs: 261 data_time: 0.69s time: 119.09s eta: 12:40:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32199.0 tgs: 270 data_time: 0.87s time: 119.07s eta: 12:38:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 268 data_time: 0.72s time: 119.59s eta: 12:39:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31334.0 tgs: 263 data_time: 0.55s time: 118.78s eta: 12:32:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32316.0 tgs: 271 data_time: 0.75s time: 118.89s eta: 12:30:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.201 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31277.0 tgs: 262 data_time: 0.85s time: 119.17s eta: 12:30:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30556.0 tgs: 256 data_time: 0.83s time: 119.26s eta: 12:29:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31378.0 tgs: 262 data_time: 0.85s time: 119.35s eta: 12:27:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31623.0 tgs: 265 data_time: 0.70s time: 119.12s eta: 12:24:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32413.0 tgs: 271 data_time: 0.77s time: 119.35s eta: 12:23:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.210 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32407.0 tgs: 271 data_time: 0.81s time: 119.27s eta: 12:21:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31952.0 tgs: 268 data_time: 0.76s time: 119.02s eta: 12:17:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30898.0 tgs: 255 data_time: 0.53s time: 121.13s eta: 12:29:00 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31053.0 tgs: 258 data_time: 0.86s time: 119.98s eta: 12:19:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 271 data_time: 0.76s time: 118.68s eta: 12:09:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.112 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32136.0 tgs: 268 data_time: 0.69s time: 119.64s eta: 12:13:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30984.0 tgs: 258 data_time: 0.90s time: 119.83s eta: 12:12:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31629.0 tgs: 265 data_time: 0.70s time: 119.14s eta: 12:06:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.099 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31407.0 tgs: 263 data_time: 0.73s time: 119.17s eta: 12:04:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31371.0 tgs: 263 data_time: 0.87s time: 119.12s eta: 12:02:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 264 data_time: 0.70s time: 119.19s eta: 12:01:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.280 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.91s time: 118.96s eta: 11:57:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 261 data_time: 0.83s time: 119.71s eta: 12:00:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.219 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31595.0 tgs: 264 data_time: 1.13s time: 119.28s eta: 11:55:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32038.0 tgs: 268 data_time: 1.02s time: 119.28s eta: 11:53:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.203 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 266 data_time: 1.00s time: 119.70s eta: 11:54:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32036.0 tgs: 269 data_time: 0.77s time: 118.78s eta: 11:46:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.184 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32466.0 tgs: 272 data_time: 0.87s time: 119.16s eta: 11:47:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32269.0 tgs: 269 data_time: 0.75s time: 119.61s eta: 11:47:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31643.0 tgs: 266 data_time: 0.58s time: 118.82s eta: 11:41:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32243.0 tgs: 271 data_time: 0.88s time: 118.68s eta: 11:38:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31173.0 tgs: 261 data_time: 0.86s time: 119.22s eta: 11:39:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31340.0 tgs: 263 data_time: 0.93s time: 119.14s eta: 11:36:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31830.0 tgs: 267 data_time: 0.56s time: 119.08s eta: 11:34:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30844.0 tgs: 259 data_time: 0.82s time: 118.95s eta: 11:31:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31527.0 tgs: 265 data_time: 0.92s time: 118.92s eta: 11:29:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32520.0 tgs: 273 data_time: 0.84s time: 118.97s eta: 11:28:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31546.0 tgs: 264 data_time: 0.66s time: 119.29s eta: 11:27:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 267 data_time: 0.76s time: 119.36s eta: 11:26:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.202 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31561.0 tgs: 265 data_time: 0.88s time: 118.88s eta: 11:21:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31786.0 tgs: 266 data_time: 0.73s time: 119.21s eta: 11:21:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31311.0 tgs: 261 data_time: 0.72s time: 119.72s eta: 11:22:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.200 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 268 data_time: 0.87s time: 119.12s eta: 11:17:00 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.111 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32434.0 tgs: 272 data_time: 0.74s time: 119.12s eta: 11:15:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30761.0 tgs: 258 data_time: 1.07s time: 118.94s eta: 11:11:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32182.0 tgs: 270 data_time: 0.84s time: 119.09s eta: 11:10:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31067.0 tgs: 260 data_time: 0.85s time: 119.16s eta: 11:09:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30765.0 tgs: 258 data_time: 0.65s time: 118.97s eta: 11:06:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.173 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31217.0 tgs: 261 data_time: 0.93s time: 119.47s eta: 11:07:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31905.0 tgs: 266 data_time: 1.06s time: 119.65s eta: 11:06:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.178 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32070.0 tgs: 268 data_time: 0.59s time: 119.30s eta: 11:02:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31471.0 tgs: 263 data_time: 0.52s time: 119.32s eta: 11:00:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32171.0 tgs: 270 data_time: 0.97s time: 118.82s eta: 10:55:27 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30309.0 tgs: 253 data_time: 0.65s time: 119.48s eta: 10:57:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31658.0 tgs: 265 data_time: 0.68s time: 119.36s eta: 10:54:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31451.0 tgs: 263 data_time: 1.01s time: 119.29s eta: 10:52:05 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30783.0 tgs: 258 data_time: 1.01s time: 119.25s eta: 10:49:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.123 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32225.0 tgs: 270 data_time: 0.74s time: 119.10s eta: 10:47:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 270 data_time: 0.87s time: 119.54s eta: 10:47:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.79s time: 119.31s eta: 10:44:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30201.0 tgs: 252 data_time: 0.69s time: 119.49s eta: 10:43:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.213 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31419.0 tgs: 262 data_time: 0.65s time: 119.52s eta: 10:41:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31770.0 tgs: 266 data_time: 0.71s time: 118.99s eta: 10:36:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31674.0 tgs: 265 data_time: 0.55s time: 119.22s eta: 10:35:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31133.0 tgs: 261 data_time: 0.67s time: 119.06s eta: 10:32:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.227 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 270 data_time: 0.92s time: 118.78s eta: 10:29:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32098.0 tgs: 267 data_time: 0.99s time: 119.86s eta: 10:33:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 268 data_time: 0.78s time: 119.39s eta: 10:28:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 267 data_time: 0.89s time: 118.83s eta: 10:23:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.206 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31451.0 tgs: 264 data_time: 0.81s time: 118.99s eta: 10:22:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32341.0 tgs: 271 data_time: 0.96s time: 119.28s eta: 10:22:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31759.0 tgs: 266 data_time: 0.80s time: 119.21s eta: 10:19:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.190 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31711.0 tgs: 266 data_time: 0.75s time: 119.06s eta: 10:17:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 267 data_time: 1.06s time: 119.48s eta: 10:17:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.109 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31261.0 tgs: 261 data_time: 0.80s time: 119.37s eta: 10:14:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.126 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31206.0 tgs: 261 data_time: 0.54s time: 119.45s eta: 10:13:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31861.0 tgs: 267 data_time: 0.58s time: 119.29s eta: 10:10:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31804.0 tgs: 267 data_time: 0.80s time: 119.11s eta: 10:07:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31898.0 tgs: 266 data_time: 0.58s time: 119.70s eta: 10:08:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.213 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32301.0 tgs: 269 data_time: 0.89s time: 119.83s eta: 10:07:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.189 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30369.0 tgs: 249 data_time: 1.30s time: 121.56s eta: 10:13:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32201.0 tgs: 270 data_time: 0.78s time: 119.26s eta: 10:00:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32151.0 tgs: 270 data_time: 0.60s time: 118.81s eta: 9:56:00 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31583.0 tgs: 264 data_time: 0.69s time: 119.57s eta: 9:57:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.125 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27870.0 tgs: 232 data_time: 0.72s time: 119.70s eta: 9:56:31 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.79s time: 119.26s eta: 9:52:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32124.0 tgs: 269 data_time: 0.68s time: 119.33s eta: 9:50:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31530.0 tgs: 264 data_time: 0.89s time: 119.24s eta: 9:48:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32408.0 tgs: 268 data_time: 1.04s time: 120.83s eta: 9:54:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31634.0 tgs: 265 data_time: 0.82s time: 119.22s eta: 9:44:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.201 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 267 data_time: 0.75s time: 119.46s eta: 9:43:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.242 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31751.0 tgs: 265 data_time: 0.84s time: 119.55s eta: 9:41:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31843.0 tgs: 267 data_time: 0.73s time: 119.20s eta: 9:38:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 267 data_time: 0.74s time: 119.48s eta: 9:37:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31199.0 tgs: 262 data_time: 0.73s time: 119.05s eta: 9:33:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31135.0 tgs: 261 data_time: 0.75s time: 119.24s eta: 9:32:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.238 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31798.0 tgs: 266 data_time: 1.00s time: 119.16s eta: 9:29:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.214 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.83s time: 119.23s eta: 9:28:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32526.0 tgs: 272 data_time: 0.84s time: 119.46s eta: 9:27:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.218 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 269 data_time: 0.81s time: 118.99s eta: 9:23:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31837.0 tgs: 267 data_time: 0.64s time: 119.04s eta: 9:21:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31530.0 tgs: 263 data_time: 0.89s time: 119.53s eta: 9:21:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.195 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31694.0 tgs: 265 data_time: 0.61s time: 119.59s eta: 9:20:05 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.202 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31826.0 tgs: 267 data_time: 1.00s time: 119.13s eta: 9:15:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32408.0 tgs: 272 data_time: 0.73s time: 119.02s eta: 9:13:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32612.0 tgs: 274 data_time: 0.93s time: 118.89s eta: 9:10:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31444.0 tgs: 263 data_time: 1.05s time: 119.33s eta: 9:10:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31313.0 tgs: 262 data_time: 0.68s time: 119.09s eta: 9:07:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32168.0 tgs: 270 data_time: 0.72s time: 118.86s eta: 9:04:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 269 data_time: 0.78s time: 119.67s eta: 9:06:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32506.0 tgs: 272 data_time: 0.74s time: 119.43s eta: 9:03:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.145 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31034.0 tgs: 259 data_time: 0.94s time: 119.43s eta: 9:01:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31068.0 tgs: 261 data_time: 0.72s time: 118.74s eta: 8:56:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31474.0 tgs: 264 data_time: 0.72s time: 118.86s eta: 8:54:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31891.0 tgs: 267 data_time: 0.79s time: 119.34s eta: 8:55:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31624.0 tgs: 265 data_time: 0.76s time: 119.15s eta: 8:52:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31497.0 tgs: 264 data_time: 1.03s time: 119.27s eta: 8:50:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 264 data_time: 0.85s time: 120.48s eta: 8:54:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 267 data_time: 0.87s time: 119.24s eta: 8:46:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 269 data_time: 0.85s time: 119.04s eta: 8:43:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30869.0 tgs: 255 data_time: 0.78s time: 120.83s eta: 8:49:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32365.0 tgs: 269 data_time: 0.87s time: 120.20s eta: 8:44:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30526.0 tgs: 256 data_time: 0.85s time: 118.83s eta: 8:36:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.230 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 267 data_time: 1.08s time: 119.35s eta: 8:37:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31160.0 tgs: 261 data_time: 0.86s time: 119.33s eta: 8:35:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.117 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31729.0 tgs: 266 data_time: 0.91s time: 119.06s eta: 8:31:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31042.0 tgs: 260 data_time: 0.83s time: 119.17s eta: 8:30:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30842.0 tgs: 259 data_time: 0.78s time: 119.04s eta: 8:27:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.109 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31401.0 tgs: 263 data_time: 0.73s time: 119.12s eta: 8:26:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.191 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32270.0 tgs: 270 data_time: 1.03s time: 119.25s eta: 8:24:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30885.0 tgs: 259 data_time: 1.25s time: 119.22s eta: 8:22:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32113.0 tgs: 268 data_time: 0.92s time: 119.38s eta: 8:21:23 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30617.0 tgs: 255 data_time: 0.77s time: 119.78s eta: 8:21:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.179 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30966.0 tgs: 259 data_time: 0.99s time: 119.20s eta: 8:16:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31021.0 tgs: 260 data_time: 0.71s time: 118.98s eta: 8:13:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.177 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31584.0 tgs: 263 data_time: 0.58s time: 119.90s eta: 8:15:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.70s time: 119.56s eta: 8:12:10 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 265 data_time: 0.67s time: 119.59s eta: 8:10:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31358.0 tgs: 263 data_time: 0.71s time: 119.00s eta: 8:05:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 263 data_time: 0.80s time: 119.84s eta: 8:07:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 266 data_time: 1.15s time: 119.71s eta: 8:04:48 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32581.0 tgs: 272 data_time: 0.77s time: 119.40s eta: 8:01:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32202.0 tgs: 269 data_time: 0.48s time: 119.29s eta: 7:59:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32132.0 tgs: 269 data_time: 0.76s time: 119.19s eta: 7:56:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30752.0 tgs: 257 data_time: 0.74s time: 119.20s eta: 7:54:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30695.0 tgs: 258 data_time: 0.71s time: 118.94s eta: 7:51:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.128 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31639.0 tgs: 265 data_time: 0.73s time: 118.97s eta: 7:49:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.130 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 268 data_time: 0.65s time: 119.42s eta: 7:49:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32232.0 tgs: 270 data_time: 0.90s time: 119.04s eta: 7:46:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.192 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30972.0 tgs: 260 data_time: 0.94s time: 119.07s eta: 7:44:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.109 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31635.0 tgs: 263 data_time: 1.13s time: 120.01s eta: 7:46:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32102.0 tgs: 267 data_time: 0.72s time: 119.82s eta: 7:43:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31869.0 tgs: 267 data_time: 0.86s time: 119.16s eta: 7:38:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31579.0 tgs: 265 data_time: 0.70s time: 118.97s eta: 7:36:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31285.0 tgs: 262 data_time: 0.66s time: 119.32s eta: 7:35:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30900.0 tgs: 258 data_time: 0.74s time: 119.39s eta: 7:33:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32129.0 tgs: 269 data_time: 0.57s time: 119.06s eta: 7:30:27 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30797.0 tgs: 258 data_time: 0.59s time: 118.99s eta: 7:28:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31609.0 tgs: 265 data_time: 0.74s time: 119.24s eta: 7:27:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.249 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32408.0 tgs: 272 data_time: 0.71s time: 118.99s eta: 7:24:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.172 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31012.0 tgs: 260 data_time: 0.42s time: 118.93s eta: 7:22:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31290.0 tgs: 262 data_time: 0.67s time: 119.08s eta: 7:20:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31489.0 tgs: 264 data_time: 0.97s time: 119.00s eta: 7:18:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30842.0 tgs: 259 data_time: 0.58s time: 119.03s eta: 7:16:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.129 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31723.0 tgs: 266 data_time: 0.66s time: 119.25s eta: 7:15:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31898.0 tgs: 267 data_time: 0.69s time: 119.35s eta: 7:13:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31408.0 tgs: 264 data_time: 0.71s time: 118.93s eta: 7:10:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.225 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30956.0 tgs: 259 data_time: 1.08s time: 119.21s eta: 7:09:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31446.0 tgs: 262 data_time: 0.71s time: 119.58s eta: 7:08:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32057.0 tgs: 267 data_time: 0.71s time: 119.66s eta: 7:06:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30995.0 tgs: 260 data_time: 0.91s time: 118.96s eta: 7:02:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 0.80s time: 119.30s eta: 7:01:31 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:46:30][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31067.0 tgs: 259 data_time: 0.80s time: 119.52s eta: 7:00:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31134.0 tgs: 260 data_time: 0.60s time: 119.43s eta: 6:57:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.191 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31956.0 tgs: 268 data_time: 0.64s time: 118.91s eta: 6:54:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30944.0 tgs: 259 data_time: 0.75s time: 119.35s eta: 6:53:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31741.0 tgs: 266 data_time: 0.78s time: 118.94s eta: 6:50:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32494.0 tgs: 272 data_time: 0.78s time: 119.26s eta: 6:49:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.189 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 267 data_time: 0.59s time: 119.02s eta: 6:46:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31928.0 tgs: 267 data_time: 0.91s time: 119.22s eta: 6:45:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.201 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 265 data_time: 0.98s time: 119.37s eta: 6:43:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31380.0 tgs: 263 data_time: 0.82s time: 119.02s eta: 6:40:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.114 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30555.0 tgs: 256 data_time: 0.71s time: 119.01s eta: 6:38:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.89s time: 119.46s eta: 6:38:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32514.0 tgs: 273 data_time: 0.75s time: 118.81s eta: 6:34:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32429.0 tgs: 271 data_time: 0.83s time: 119.44s eta: 6:34:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.182 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32640.0 tgs: 272 data_time: 1.20s time: 119.57s eta: 6:32:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.180 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32408.0 tgs: 271 data_time: 0.82s time: 119.39s eta: 6:29:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 269 data_time: 0.95s time: 119.66s eta: 6:28:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31777.0 tgs: 266 data_time: 0.67s time: 119.36s eta: 6:25:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.188 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 266 data_time: 0.83s time: 118.83s eta: 6:22:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31981.0 tgs: 268 data_time: 0.79s time: 119.01s eta: 6:20:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31265.0 tgs: 261 data_time: 0.71s time: 119.41s eta: 6:20:07 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29951.0 tgs: 250 data_time: 0.75s time: 119.33s eta: 6:17:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31228.0 tgs: 262 data_time: 0.64s time: 119.05s eta: 6:14:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31526.0 tgs: 264 data_time: 0.52s time: 119.38s eta: 6:14:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 269 data_time: 0.67s time: 119.39s eta: 6:12:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31862.0 tgs: 267 data_time: 0.72s time: 119.29s eta: 6:09:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32242.0 tgs: 270 data_time: 0.71s time: 119.11s eta: 6:07:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31870.0 tgs: 261 data_time: 0.95s time: 121.96s eta: 6:14:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32375.0 tgs: 271 data_time: 0.53s time: 119.39s eta: 6:04:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31437.0 tgs: 262 data_time: 0.62s time: 119.58s eta: 6:02:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31866.0 tgs: 267 data_time: 0.64s time: 119.26s eta: 5:59:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31602.0 tgs: 264 data_time: 0.88s time: 119.26s eta: 5:57:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.196 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31467.0 tgs: 264 data_time: 0.91s time: 118.92s eta: 5:54:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 269 data_time: 0.73s time: 119.03s eta: 5:53:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31332.0 tgs: 263 data_time: 0.86s time: 119.06s eta: 5:51:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31507.0 tgs: 264 data_time: 0.98s time: 119.29s eta: 5:49:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31229.0 tgs: 262 data_time: 0.76s time: 118.97s eta: 5:47:00 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30876.0 tgs: 258 data_time: 0.94s time: 119.27s eta: 5:45:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.77s time: 119.32s eta: 5:44:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31958.0 tgs: 268 data_time: 0.89s time: 119.05s eta: 5:41:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.74s time: 119.50s eta: 5:40:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31766.0 tgs: 266 data_time: 0.79s time: 119.38s eta: 5:38:14 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32039.0 tgs: 267 data_time: 0.94s time: 119.57s eta: 5:36:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.87s time: 118.99s eta: 5:33:10 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 267 data_time: 0.85s time: 118.95s eta: 5:31:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 270 data_time: 0.78s time: 119.24s eta: 5:29:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 265 data_time: 0.82s time: 119.40s eta: 5:28:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31958.0 tgs: 267 data_time: 0.96s time: 119.47s eta: 5:26:32 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31897.0 tgs: 268 data_time: 0.57s time: 118.95s eta: 5:23:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32191.0 tgs: 269 data_time: 0.78s time: 119.63s eta: 5:22:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.220 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 26257.0 tgs: 220 data_time: 0.50s time: 119.35s eta: 5:20:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31210.0 tgs: 261 data_time: 0.86s time: 119.12s eta: 5:17:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31275.0 tgs: 261 data_time: 1.04s time: 119.44s eta: 5:16:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.221 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32370.0 tgs: 270 data_time: 0.88s time: 119.56s eta: 5:14:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.197 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31607.0 tgs: 265 data_time: 0.68s time: 119.01s eta: 5:11:23 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31756.0 tgs: 266 data_time: 0.80s time: 119.00s eta: 5:09:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.189 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 270 data_time: 0.89s time: 119.24s eta: 5:08:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30943.0 tgs: 258 data_time: 0.90s time: 119.67s eta: 5:07:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.184 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.86s time: 119.12s eta: 5:03:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30938.0 tgs: 259 data_time: 0.71s time: 119.06s eta: 5:01:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32052.0 tgs: 263 data_time: 0.68s time: 121.52s eta: 5:05:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30998.0 tgs: 259 data_time: 0.92s time: 119.58s eta: 4:58:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31855.0 tgs: 266 data_time: 0.72s time: 119.50s eta: 4:56:45 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31404.0 tgs: 262 data_time: 0.70s time: 119.74s eta: 4:55:21 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30901.0 tgs: 258 data_time: 0.65s time: 119.52s eta: 4:52:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.236 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31560.0 tgs: 264 data_time: 1.15s time: 119.41s eta: 4:50:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.82s time: 119.21s eta: 4:48:05 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.233 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 269 data_time: 0.68s time: 118.89s eta: 4:45:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30964.0 tgs: 260 data_time: 0.60s time: 119.03s eta: 4:43:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29944.0 tgs: 249 data_time: 0.69s time: 119.98s eta: 4:43:57 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.68s time: 119.15s eta: 4:39:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30152.0 tgs: 253 data_time: 0.79s time: 119.01s eta: 4:37:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.194 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31939.0 tgs: 268 data_time: 0.92s time: 119.06s eta: 4:35:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 267 data_time: 1.06s time: 119.37s eta: 4:34:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31461.0 tgs: 263 data_time: 0.70s time: 119.41s eta: 4:32:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30782.0 tgs: 258 data_time: 0.93s time: 118.94s eta: 4:29:36 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.184 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31712.0 tgs: 265 data_time: 0.78s time: 119.60s eta: 4:29:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31675.0 tgs: 265 data_time: 0.81s time: 119.24s eta: 4:26:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.213 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 0.54s time: 119.06s eta: 4:23:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31746.0 tgs: 267 data_time: 0.67s time: 118.85s eta: 4:21:28 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31114.0 tgs: 261 data_time: 0.90s time: 118.76s eta: 4:19:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.114 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 266 data_time: 0.70s time: 119.43s eta: 4:18:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31607.0 tgs: 264 data_time: 0.82s time: 119.40s eta: 4:16:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31454.0 tgs: 263 data_time: 0.72s time: 119.32s eta: 4:14:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32235.0 tgs: 270 data_time: 0.65s time: 119.05s eta: 4:11:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32047.0 tgs: 269 data_time: 1.08s time: 119.13s eta: 4:10:09 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31972.0 tgs: 268 data_time: 0.63s time: 119.13s eta: 4:08:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 270 data_time: 0.77s time: 118.96s eta: 4:05:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32059.0 tgs: 269 data_time: 0.70s time: 119.15s eta: 4:04:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32173.0 tgs: 269 data_time: 0.64s time: 119.50s eta: 4:02:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31061.0 tgs: 260 data_time: 0.77s time: 119.25s eta: 4:00:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31603.0 tgs: 265 data_time: 0.77s time: 119.19s eta: 3:58:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 270 data_time: 0.70s time: 119.36s eta: 3:56:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30680.0 tgs: 257 data_time: 0.91s time: 119.37s eta: 3:54:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32032.0 tgs: 267 data_time: 0.76s time: 119.95s eta: 3:53:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32258.0 tgs: 270 data_time: 0.56s time: 119.32s eta: 3:50:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31597.0 tgs: 264 data_time: 0.71s time: 119.64s eta: 3:49:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31840.0 tgs: 267 data_time: 0.95s time: 118.85s eta: 3:45:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32295.0 tgs: 269 data_time: 1.14s time: 119.71s eta: 3:45:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30898.0 tgs: 258 data_time: 0.71s time: 119.33s eta: 3:42:44 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32059.0 tgs: 268 data_time: 0.80s time: 119.26s eta: 3:40:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.81s time: 119.00s eta: 3:38:10 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31865.0 tgs: 266 data_time: 0.61s time: 119.43s eta: 3:36:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31605.0 tgs: 265 data_time: 0.71s time: 119.04s eta: 3:34:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31589.0 tgs: 265 data_time: 0.85s time: 118.98s eta: 3:32:10 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30677.0 tgs: 257 data_time: 0.93s time: 119.23s eta: 3:30:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.109 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31910.0 tgs: 265 data_time: 0.72s time: 120.11s eta: 3:30:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31878.0 tgs: 267 data_time: 0.90s time: 119.13s eta: 3:26:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30743.0 tgs: 257 data_time: 0.66s time: 119.32s eta: 3:24:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.72s time: 119.08s eta: 3:22:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31569.0 tgs: 264 data_time: 1.04s time: 119.44s eta: 3:21:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31958.0 tgs: 267 data_time: 0.72s time: 119.28s eta: 3:18:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30267.0 tgs: 253 data_time: 0.86s time: 119.22s eta: 3:16:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31449.0 tgs: 263 data_time: 0.98s time: 119.13s eta: 3:14:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31466.0 tgs: 264 data_time: 0.79s time: 119.02s eta: 3:12:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.211 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31925.0 tgs: 267 data_time: 0.72s time: 119.19s eta: 3:10:41 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31461.0 tgs: 259 data_time: 0.74s time: 121.05s eta: 3:11:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31837.0 tgs: 266 data_time: 0.81s time: 119.51s eta: 3:07:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31360.0 tgs: 262 data_time: 0.72s time: 119.26s eta: 3:04:51 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.110 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30682.0 tgs: 257 data_time: 0.68s time: 118.99s eta: 3:02:27 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31461.0 tgs: 263 data_time: 0.86s time: 119.22s eta: 3:00:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32106.0 tgs: 268 data_time: 0.54s time: 119.46s eta: 2:59:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.107 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30882.0 tgs: 259 data_time: 0.62s time: 118.98s eta: 2:56:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32623.0 tgs: 273 data_time: 1.01s time: 119.35s eta: 2:55:03 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.110 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 265 data_time: 0.69s time: 119.41s eta: 2:53:08 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31877.0 tgs: 267 data_time: 0.97s time: 119.35s eta: 2:51:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.237 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31673.0 tgs: 260 data_time: 0.63s time: 121.46s eta: 2:52:04 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30964.0 tgs: 259 data_time: 0.52s time: 119.24s eta: 2:46:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32388.0 tgs: 270 data_time: 0.88s time: 119.54s eta: 2:45:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 266 data_time: 0.71s time: 118.91s eta: 2:42:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31024.0 tgs: 260 data_time: 0.87s time: 119.18s eta: 2:40:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.187 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 269 data_time: 0.76s time: 118.97s eta: 2:38:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30283.0 tgs: 253 data_time: 1.02s time: 119.23s eta: 2:36:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31587.0 tgs: 265 data_time: 0.95s time: 119.11s eta: 2:34:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30966.0 tgs: 259 data_time: 0.74s time: 119.44s eta: 2:33:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31592.0 tgs: 262 data_time: 0.86s time: 120.22s eta: 2:32:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31907.0 tgs: 267 data_time: 0.75s time: 119.28s eta: 2:29:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 269 data_time: 0.65s time: 118.89s eta: 2:26:37 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31199.0 tgs: 261 data_time: 0.54s time: 119.52s eta: 2:25:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 267 data_time: 0.67s time: 119.08s eta: 2:22:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31412.0 tgs: 263 data_time: 0.62s time: 119.13s eta: 2:20:58 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30660.0 tgs: 257 data_time: 0.74s time: 119.05s eta: 2:18:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31816.0 tgs: 265 data_time: 0.74s time: 119.84s eta: 2:17:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32007.0 tgs: 268 data_time: 0.83s time: 119.31s eta: 2:15:13 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30289.0 tgs: 253 data_time: 0.78s time: 119.39s eta: 2:13:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 267 data_time: 0.77s time: 119.68s eta: 2:11:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31019.0 tgs: 259 data_time: 0.80s time: 119.49s eta: 2:09:27 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31730.0 tgs: 266 data_time: 0.86s time: 118.96s eta: 2:06:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 265 data_time: 0.90s time: 119.37s eta: 2:05:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30443.0 tgs: 255 data_time: 0.91s time: 119.28s eta: 2:03:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30905.0 tgs: 259 data_time: 0.66s time: 119.22s eta: 2:01:12 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31709.0 tgs: 265 data_time: 0.75s time: 119.51s eta: 1:59:30 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 264 data_time: 0.68s time: 119.43s eta: 1:57:26 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 265 data_time: 0.68s time: 119.53s eta: 1:55:32 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31060.0 tgs: 260 data_time: 0.62s time: 119.31s eta: 1:53:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 267 data_time: 0.67s time: 119.25s eta: 1:51:18 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 265 data_time: 0.87s time: 118.93s eta: 1:49:01 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30865.0 tgs: 257 data_time: 0.65s time: 119.64s eta: 1:47:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 267 data_time: 0.78s time: 119.69s eta: 1:45:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31319.0 tgs: 262 data_time: 0.62s time: 119.34s eta: 1:43:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31900.0 tgs: 267 data_time: 0.71s time: 119.31s eta: 1:41:24 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.81s time: 119.28s eta: 1:39:23 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31471.0 tgs: 264 data_time: 0.91s time: 118.91s eta: 1:37:06 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30797.0 tgs: 258 data_time: 0.82s time: 119.16s eta: 1:35:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31962.0 tgs: 268 data_time: 0.99s time: 119.06s eta: 1:33:15 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32074.0 tgs: 268 data_time: 0.70s time: 119.26s eta: 1:31:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 268 data_time: 0.61s time: 119.72s eta: 1:29:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31450.0 tgs: 264 data_time: 0.79s time: 119.02s eta: 1:27:16 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31320.0 tgs: 263 data_time: 0.55s time: 118.87s eta: 1:25:11 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31268.0 tgs: 261 data_time: 0.63s time: 119.48s eta: 1:23:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.089 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 269 data_time: 0.74s time: 119.49s eta: 1:21:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32419.0 tgs: 272 data_time: 0.70s time: 118.94s eta: 1:19:17 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31308.0 tgs: 263 data_time: 0.84s time: 118.95s eta: 1:17:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30939.0 tgs: 253 data_time: 0.78s time: 122.12s eta: 1:17:20 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31111.0 tgs: 260 data_time: 0.78s time: 119.35s eta: 1:13:35 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.61s time: 119.06s eta: 1:11:25 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31555.0 tgs: 265 data_time: 0.58s time: 118.92s eta: 1:09:22 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31167.0 tgs: 261 data_time: 0.83s time: 119.11s eta: 1:07:29 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31922.0 tgs: 267 data_time: 0.89s time: 119.42s eta: 1:05:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32251.0 tgs: 270 data_time: 0.64s time: 119.35s eta: 1:03:39 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31505.0 tgs: 264 data_time: 0.62s time: 119.18s eta: 1:01:34 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30900.0 tgs: 259 data_time: 0.55s time: 119.27s eta: 0:59:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 264 data_time: 1.10s time: 119.44s eta: 0:57:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31319.0 tgs: 262 data_time: 0.68s time: 119.39s eta: 0:55:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.241 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32585.0 tgs: 272 data_time: 0.71s time: 119.49s eta: 0:53:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31416.0 tgs: 263 data_time: 0.98s time: 119.37s eta: 0:51:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30506.0 tgs: 255 data_time: 0.69s time: 119.47s eta: 0:49:46 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31605.0 tgs: 265 data_time: 0.64s time: 118.90s eta: 0:47:33 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30844.0 tgs: 258 data_time: 0.89s time: 119.27s eta: 0:45:43 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 270 data_time: 0.72s time: 119.19s eta: 0:43:42 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.71s time: 119.00s eta: 0:41:38 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31081.0 tgs: 261 data_time: 1.00s time: 119.04s eta: 0:39:40 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32146.0 tgs: 268 data_time: 0.70s time: 119.66s eta: 0:37:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 266 data_time: 0.80s time: 119.29s eta: 0:35:47 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.237 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31969.0 tgs: 266 data_time: 0.83s time: 119.75s eta: 0:33:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31841.0 tgs: 262 data_time: 0.84s time: 121.23s eta: 0:32:19 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31373.0 tgs: 262 data_time: 0.60s time: 119.38s eta: 0:29:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 29852.0 tgs: 250 data_time: 0.77s time: 119.32s eta: 0:27:50 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32191.0 tgs: 269 data_time: 0.60s time: 119.64s eta: 0:25:55 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31841.0 tgs: 267 data_time: 0.93s time: 119.10s eta: 0:23:49 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32422.0 tgs: 271 data_time: 0.67s time: 119.39s eta: 0:21:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31197.0 tgs: 261 data_time: 0.88s time: 119.26s eta: 0:19:52 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30716.0 tgs: 257 data_time: 1.13s time: 119.24s eta: 0:17:53 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31330.0 tgs: 261 data_time: 0.59s time: 119.90s eta: 0:15:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32200.0 tgs: 270 data_time: 0.92s time: 119.18s eta: 0:13:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31804.0 tgs: 267 data_time: 0.91s time: 119.09s eta: 0:11:54 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 269 data_time: 0.76s time: 119.32s eta: 0:09:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.228 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31296.0 tgs: 262 data_time: 0.46s time: 119.17s eta: 0:07:56 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 269 data_time: 0.68s time: 119.69s eta: 0:05:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31209.0 tgs: 257 data_time: 0.80s time: 121.35s eta: 0:04:02 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30751.0 tgs: 257 data_time: 0.77s time: 119.21s eta: 0:01:59 +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.29 seconds, peak gpu memory 17.4G +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 34][DP 8][SP 2][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank35.log b/20250121165312/rank35.log new file mode 100644 index 0000000000000000000000000000000000000000..b7cea40a1501788e238be7f71710cd7528bbda43 --- /dev/null +++ b/20250121165312/rank35.log @@ -0,0 +1,851 @@ +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:35][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.05s +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.72 seconds, peak gpu memory 13.4G +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.259 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31323.0 tgs: 205 data_time: 1.97s time: 152.27s eta: 1 day, 1:04:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.310 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31388.0 tgs: 261 data_time: 0.82s time: 120.07s eta: 19:44:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.218 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31981.0 tgs: 267 data_time: 1.00s time: 119.69s eta: 19:38:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.236 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31503.0 tgs: 263 data_time: 1.33s time: 119.65s eta: 19:36:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31606.0 tgs: 264 data_time: 0.67s time: 119.31s eta: 19:31:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.225 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31195.0 tgs: 260 data_time: 0.76s time: 119.54s eta: 19:31:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.221 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31593.0 tgs: 256 data_time: 0.75s time: 122.96s eta: 20:02:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.242 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31176.0 tgs: 260 data_time: 0.67s time: 119.90s eta: 19:30:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.175 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32152.0 tgs: 266 data_time: 0.69s time: 120.65s eta: 19:36:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.198 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29940.0 tgs: 249 data_time: 0.65s time: 119.92s eta: 19:27:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.198 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31618.0 tgs: 262 data_time: 0.89s time: 120.46s eta: 19:30:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31523.0 tgs: 264 data_time: 0.65s time: 119.35s eta: 19:17:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.219 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31105.0 tgs: 259 data_time: 0.83s time: 119.67s eta: 19:18:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31143.0 tgs: 261 data_time: 0.52s time: 119.18s eta: 19:12:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31851.0 tgs: 267 data_time: 0.57s time: 119.19s eta: 19:10:09 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31775.0 tgs: 267 data_time: 0.87s time: 118.93s eta: 19:05:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31815.0 tgs: 267 data_time: 0.90s time: 119.15s eta: 19:05:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31366.0 tgs: 263 data_time: 0.88s time: 119.18s eta: 19:04:07 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31364.0 tgs: 262 data_time: 0.84s time: 119.40s eta: 19:04:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.241 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32004.0 tgs: 268 data_time: 0.96s time: 119.33s eta: 19:01:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 262 data_time: 0.85s time: 120.32s eta: 19:09:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31087.0 tgs: 252 data_time: 0.69s time: 122.92s eta: 19:31:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31662.0 tgs: 265 data_time: 0.58s time: 119.48s eta: 18:57:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31237.0 tgs: 262 data_time: 0.82s time: 119.16s eta: 18:51:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31187.0 tgs: 261 data_time: 0.76s time: 119.25s eta: 18:50:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31088.0 tgs: 260 data_time: 0.66s time: 119.39s eta: 18:50:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31662.0 tgs: 264 data_time: 0.78s time: 119.66s eta: 18:50:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32039.0 tgs: 268 data_time: 0.96s time: 119.39s eta: 18:46:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32252.0 tgs: 269 data_time: 0.56s time: 119.77s eta: 18:47:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 246 data_time: 0.96s time: 129.31s eta: 20:15:30 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 263 data_time: 0.58s time: 119.56s eta: 18:41:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 29894.0 tgs: 250 data_time: 0.79s time: 119.39s eta: 18:38:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.233 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31299.0 tgs: 259 data_time: 0.84s time: 120.66s eta: 18:48:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31641.0 tgs: 264 data_time: 0.83s time: 119.40s eta: 18:34:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31053.0 tgs: 260 data_time: 0.58s time: 119.16s eta: 18:30:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31452.0 tgs: 264 data_time: 0.81s time: 119.07s eta: 18:27:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31537.0 tgs: 264 data_time: 0.80s time: 119.14s eta: 18:26:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32333.0 tgs: 268 data_time: 0.86s time: 120.39s eta: 18:35:35 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32264.0 tgs: 266 data_time: 0.76s time: 121.17s eta: 18:40:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.122 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31571.0 tgs: 261 data_time: 0.62s time: 120.66s eta: 18:34:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.120 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32023.0 tgs: 267 data_time: 0.53s time: 119.74s eta: 18:23:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.287 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32305.0 tgs: 260 data_time: 0.76s time: 123.88s eta: 18:59:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32197.0 tgs: 269 data_time: 1.03s time: 119.56s eta: 18:17:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31127.0 tgs: 260 data_time: 0.59s time: 119.32s eta: 18:13:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31389.0 tgs: 263 data_time: 0.82s time: 119.22s eta: 18:10:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.119 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31307.0 tgs: 261 data_time: 0.76s time: 119.60s eta: 18:12:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.242 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 266 data_time: 1.19s time: 119.27s eta: 18:07:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31146.0 tgs: 261 data_time: 0.94s time: 119.02s eta: 18:03:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30841.0 tgs: 258 data_time: 0.81s time: 119.10s eta: 18:01:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30859.0 tgs: 258 data_time: 1.14s time: 119.26s eta: 18:01:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31436.0 tgs: 263 data_time: 0.84s time: 119.33s eta: 17:59:57 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32042.0 tgs: 266 data_time: 0.90s time: 120.02s eta: 18:04:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31299.0 tgs: 262 data_time: 0.63s time: 119.11s eta: 17:53:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31820.0 tgs: 266 data_time: 0.54s time: 119.41s eta: 17:54:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.121 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31916.0 tgs: 268 data_time: 0.67s time: 119.00s eta: 17:48:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30595.0 tgs: 257 data_time: 0.74s time: 119.03s eta: 17:47:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.230 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32038.0 tgs: 255 data_time: 0.71s time: 125.37s eta: 18:42:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31912.0 tgs: 265 data_time: 0.78s time: 120.11s eta: 17:52:57 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 267 data_time: 0.91s time: 119.37s eta: 17:44:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 269 data_time: 0.67s time: 119.31s eta: 17:41:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32223.0 tgs: 269 data_time: 0.78s time: 119.51s eta: 17:41:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32203.0 tgs: 255 data_time: 0.62s time: 126.05s eta: 18:37:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30992.0 tgs: 252 data_time: 0.69s time: 122.63s eta: 18:05:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.278 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32317.0 tgs: 266 data_time: 0.98s time: 121.16s eta: 17:50:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31408.0 tgs: 263 data_time: 0.68s time: 119.04s eta: 17:29:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31491.0 tgs: 264 data_time: 0.77s time: 118.86s eta: 17:25:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30711.0 tgs: 247 data_time: 0.82s time: 124.00s eta: 18:09:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30847.0 tgs: 247 data_time: 0.75s time: 124.57s eta: 18:12:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32021.0 tgs: 268 data_time: 0.89s time: 119.39s eta: 17:24:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32505.0 tgs: 271 data_time: 0.97s time: 119.70s eta: 17:25:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31080.0 tgs: 260 data_time: 0.91s time: 119.53s eta: 17:21:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31246.0 tgs: 262 data_time: 0.68s time: 118.86s eta: 17:14:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32177.0 tgs: 268 data_time: 0.91s time: 119.85s eta: 17:20:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31382.0 tgs: 262 data_time: 0.73s time: 119.52s eta: 17:15:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.213 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32159.0 tgs: 263 data_time: 0.83s time: 121.86s eta: 17:34:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31877.0 tgs: 258 data_time: 0.64s time: 123.52s eta: 17:46:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31207.0 tgs: 262 data_time: 0.62s time: 119.00s eta: 17:05:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30898.0 tgs: 259 data_time: 0.77s time: 119.23s eta: 17:05:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30827.0 tgs: 258 data_time: 0.81s time: 119.14s eta: 17:02:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31777.0 tgs: 266 data_time: 0.53s time: 119.26s eta: 17:01:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31137.0 tgs: 261 data_time: 0.86s time: 119.04s eta: 16:57:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31356.0 tgs: 262 data_time: 0.78s time: 119.28s eta: 16:57:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 270 data_time: 0.83s time: 119.16s eta: 16:54:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32076.0 tgs: 268 data_time: 1.04s time: 119.59s eta: 16:56:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31126.0 tgs: 261 data_time: 0.69s time: 119.13s eta: 16:50:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 266 data_time: 0.90s time: 118.84s eta: 16:46:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31389.0 tgs: 263 data_time: 0.69s time: 119.02s eta: 16:45:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.232 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31471.0 tgs: 262 data_time: 0.78s time: 119.85s eta: 16:50:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30848.0 tgs: 258 data_time: 0.83s time: 119.36s eta: 16:44:35 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 268 data_time: 0.82s time: 119.01s eta: 16:39:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32558.0 tgs: 269 data_time: 0.72s time: 120.93s eta: 16:53:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32033.0 tgs: 269 data_time: 0.72s time: 119.00s eta: 16:35:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32243.0 tgs: 270 data_time: 0.68s time: 119.29s eta: 16:36:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.236 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32206.0 tgs: 270 data_time: 0.87s time: 119.22s eta: 16:33:32 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32438.0 tgs: 271 data_time: 0.64s time: 119.37s eta: 16:32:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32226.0 tgs: 269 data_time: 0.90s time: 119.70s eta: 16:33:32 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.231 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.67s time: 119.21s eta: 16:27:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 264 data_time: 0.56s time: 119.76s eta: 16:30:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 267 data_time: 0.86s time: 120.27s eta: 16:32:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.235 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 267 data_time: 0.66s time: 118.98s eta: 16:19:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.236 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31282.0 tgs: 245 data_time: 0.78s time: 127.33s eta: 17:26:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31733.0 tgs: 266 data_time: 0.72s time: 119.23s eta: 16:17:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 260 data_time: 0.94s time: 119.14s eta: 16:15:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31006.0 tgs: 260 data_time: 0.76s time: 119.20s eta: 16:13:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31562.0 tgs: 264 data_time: 0.58s time: 119.17s eta: 16:11:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31002.0 tgs: 260 data_time: 0.72s time: 118.96s eta: 16:07:30 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 263 data_time: 0.95s time: 118.86s eta: 16:04:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32454.0 tgs: 272 data_time: 0.77s time: 118.96s eta: 16:03:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.120 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32314.0 tgs: 265 data_time: 0.61s time: 121.84s eta: 16:24:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30862.0 tgs: 259 data_time: 0.59s time: 119.00s eta: 15:59:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31650.0 tgs: 265 data_time: 0.95s time: 119.23s eta: 15:59:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31524.0 tgs: 263 data_time: 0.74s time: 119.62s eta: 16:00:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30848.0 tgs: 259 data_time: 0.76s time: 119.04s eta: 15:54:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 267 data_time: 0.63s time: 118.98s eta: 15:51:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 1.01s time: 119.17s eta: 15:51:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29869.0 tgs: 249 data_time: 0.73s time: 119.64s eta: 15:53:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32048.0 tgs: 269 data_time: 0.71s time: 118.93s eta: 15:45:31 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 266 data_time: 0.75s time: 119.29s eta: 15:46:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31414.0 tgs: 263 data_time: 0.89s time: 119.11s eta: 15:42:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32163.0 tgs: 270 data_time: 0.59s time: 118.99s eta: 15:40:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30553.0 tgs: 255 data_time: 0.77s time: 119.66s eta: 15:43:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31786.0 tgs: 261 data_time: 0.72s time: 121.51s eta: 15:55:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.220 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32336.0 tgs: 270 data_time: 1.10s time: 119.41s eta: 15:37:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30519.0 tgs: 234 data_time: 0.65s time: 129.97s eta: 16:58:05 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.223 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32337.0 tgs: 270 data_time: 0.82s time: 119.47s eta: 15:33:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31199.0 tgs: 260 data_time: 0.84s time: 119.61s eta: 15:32:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 269 data_time: 0.77s time: 119.54s eta: 15:30:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32322.0 tgs: 264 data_time: 0.89s time: 122.13s eta: 15:48:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30720.0 tgs: 254 data_time: 0.67s time: 120.76s eta: 15:35:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30939.0 tgs: 259 data_time: 1.07s time: 119.41s eta: 15:23:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31698.0 tgs: 265 data_time: 0.71s time: 119.45s eta: 15:21:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31820.0 tgs: 241 data_time: 0.86s time: 131.63s eta: 16:53:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31395.0 tgs: 263 data_time: 0.57s time: 119.29s eta: 15:16:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31617.0 tgs: 264 data_time: 0.63s time: 119.52s eta: 15:16:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32474.0 tgs: 265 data_time: 0.58s time: 122.25s eta: 15:35:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31858.0 tgs: 266 data_time: 0.82s time: 119.71s eta: 15:13:48 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31342.0 tgs: 262 data_time: 0.89s time: 119.35s eta: 15:09:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31870.0 tgs: 268 data_time: 0.64s time: 118.90s eta: 15:03:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31221.0 tgs: 262 data_time: 0.84s time: 119.00s eta: 15:02:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 267 data_time: 0.67s time: 119.31s eta: 15:02:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31524.0 tgs: 265 data_time: 1.04s time: 118.92s eta: 14:57:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32396.0 tgs: 268 data_time: 0.71s time: 120.52s eta: 15:07:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31891.0 tgs: 267 data_time: 0.88s time: 119.31s eta: 14:56:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32178.0 tgs: 262 data_time: 1.13s time: 122.66s eta: 15:19:57 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31452.0 tgs: 264 data_time: 0.81s time: 119.12s eta: 14:51:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31484.0 tgs: 262 data_time: 1.00s time: 119.76s eta: 14:54:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.123 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31734.0 tgs: 265 data_time: 0.67s time: 119.46s eta: 14:50:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31121.0 tgs: 261 data_time: 0.86s time: 118.99s eta: 14:44:31 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.307 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32300.0 tgs: 270 data_time: 0.65s time: 119.30s eta: 14:44:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31009.0 tgs: 260 data_time: 0.85s time: 118.94s eta: 14:40:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 270 data_time: 0.89s time: 119.05s eta: 14:39:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.118 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 267 data_time: 0.80s time: 119.07s eta: 14:37:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 265 data_time: 0.52s time: 119.13s eta: 14:35:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32011.0 tgs: 269 data_time: 0.70s time: 119.00s eta: 14:32:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31413.0 tgs: 263 data_time: 0.58s time: 119.08s eta: 14:31:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31819.0 tgs: 267 data_time: 0.73s time: 119.15s eta: 14:29:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30523.0 tgs: 255 data_time: 0.74s time: 119.65s eta: 14:31:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 265 data_time: 0.68s time: 119.18s eta: 14:26:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.205 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32117.0 tgs: 269 data_time: 0.94s time: 119.20s eta: 14:24:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30801.0 tgs: 257 data_time: 0.73s time: 119.39s eta: 14:23:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31854.0 tgs: 263 data_time: 0.72s time: 120.74s eta: 14:31:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.222 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 266 data_time: 0.93s time: 118.85s eta: 14:15:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32012.0 tgs: 268 data_time: 0.77s time: 119.10s eta: 14:15:32 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32216.0 tgs: 270 data_time: 0.87s time: 119.12s eta: 14:13:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32399.0 tgs: 270 data_time: 0.85s time: 119.95s eta: 14:17:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.127 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30625.0 tgs: 255 data_time: 0.43s time: 119.87s eta: 14:15:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.211 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32129.0 tgs: 268 data_time: 0.62s time: 119.59s eta: 14:11:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 268 data_time: 0.86s time: 119.24s eta: 14:06:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.111 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31621.0 tgs: 265 data_time: 0.62s time: 119.25s eta: 14:04:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.197 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32198.0 tgs: 269 data_time: 0.80s time: 119.27s eta: 14:02:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32209.0 tgs: 270 data_time: 0.77s time: 119.14s eta: 13:59:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.230 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31991.0 tgs: 268 data_time: 0.80s time: 119.12s eta: 13:57:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.210 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32164.0 tgs: 269 data_time: 0.91s time: 119.17s eta: 13:56:10 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31588.0 tgs: 264 data_time: 0.72s time: 119.35s eta: 13:55:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31117.0 tgs: 259 data_time: 0.56s time: 120.12s eta: 13:58:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31650.0 tgs: 258 data_time: 1.01s time: 122.55s eta: 14:13:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.343 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 268 data_time: 0.86s time: 119.11s eta: 13:47:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30848.0 tgs: 258 data_time: 0.66s time: 119.37s eta: 13:47:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31447.0 tgs: 263 data_time: 0.87s time: 119.33s eta: 13:45:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31659.0 tgs: 265 data_time: 0.69s time: 119.34s eta: 13:43:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.220 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32096.0 tgs: 269 data_time: 0.68s time: 119.11s eta: 13:39:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.120 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31683.0 tgs: 265 data_time: 0.96s time: 119.26s eta: 13:38:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.216 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32452.0 tgs: 272 data_time: 0.82s time: 119.14s eta: 13:36:07 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31475.0 tgs: 264 data_time: 0.65s time: 119.22s eta: 13:34:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30320.0 tgs: 253 data_time: 0.81s time: 119.46s eta: 13:34:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32453.0 tgs: 272 data_time: 0.77s time: 119.08s eta: 13:29:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31923.0 tgs: 268 data_time: 0.74s time: 118.92s eta: 13:26:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.192 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31882.0 tgs: 266 data_time: 0.61s time: 119.81s eta: 13:30:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.119 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30850.0 tgs: 258 data_time: 0.88s time: 119.26s eta: 13:25:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.210 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31955.0 tgs: 268 data_time: 0.64s time: 119.16s eta: 13:22:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32059.0 tgs: 268 data_time: 0.80s time: 119.35s eta: 13:21:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32227.0 tgs: 270 data_time: 0.68s time: 119.26s eta: 13:19:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31072.0 tgs: 261 data_time: 0.87s time: 119.05s eta: 13:15:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31866.0 tgs: 267 data_time: 0.84s time: 119.35s eta: 13:15:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31014.0 tgs: 260 data_time: 0.70s time: 119.17s eta: 13:12:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31575.0 tgs: 264 data_time: 0.92s time: 119.46s eta: 13:12:23 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31526.0 tgs: 264 data_time: 0.64s time: 119.30s eta: 13:09:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.096 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31466.0 tgs: 264 data_time: 0.62s time: 119.14s eta: 13:06:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 265 data_time: 0.72s time: 120.41s eta: 13:12:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32275.0 tgs: 268 data_time: 0.57s time: 120.16s eta: 13:09:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32338.0 tgs: 270 data_time: 0.83s time: 119.50s eta: 13:02:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.251 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 266 data_time: 0.80s time: 119.51s eta: 13:00:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31084.0 tgs: 259 data_time: 0.69s time: 119.57s eta: 12:59:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 268 data_time: 1.05s time: 119.80s eta: 12:58:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31602.0 tgs: 265 data_time: 0.63s time: 119.04s eta: 12:51:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.203 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31746.0 tgs: 266 data_time: 0.72s time: 119.05s eta: 12:49:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32700.0 tgs: 274 data_time: 0.86s time: 119.13s eta: 12:48:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 266 data_time: 0.74s time: 119.40s eta: 12:48:07 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30903.0 tgs: 259 data_time: 0.55s time: 119.19s eta: 12:44:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31993.0 tgs: 267 data_time: 0.85s time: 119.41s eta: 12:44:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.141 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31147.0 tgs: 261 data_time: 0.67s time: 119.10s eta: 12:40:14 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32199.0 tgs: 270 data_time: 0.81s time: 119.07s eta: 12:38:05 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 268 data_time: 0.71s time: 119.60s eta: 12:39:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31334.0 tgs: 263 data_time: 0.54s time: 118.78s eta: 12:32:14 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32316.0 tgs: 271 data_time: 0.74s time: 118.89s eta: 12:30:57 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31277.0 tgs: 262 data_time: 0.83s time: 119.17s eta: 12:30:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30556.0 tgs: 256 data_time: 0.81s time: 119.27s eta: 12:29:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.214 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31378.0 tgs: 262 data_time: 0.84s time: 119.35s eta: 12:27:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.198 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31623.0 tgs: 265 data_time: 0.68s time: 119.12s eta: 12:24:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32413.0 tgs: 271 data_time: 0.73s time: 119.35s eta: 12:23:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.201 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32407.0 tgs: 271 data_time: 0.78s time: 119.26s eta: 12:21:23 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.122 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31952.0 tgs: 268 data_time: 0.75s time: 119.03s eta: 12:17:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30898.0 tgs: 255 data_time: 0.49s time: 121.13s eta: 12:29:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31053.0 tgs: 258 data_time: 0.84s time: 119.99s eta: 12:19:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 271 data_time: 0.74s time: 118.68s eta: 12:09:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32136.0 tgs: 268 data_time: 0.68s time: 119.64s eta: 12:13:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30984.0 tgs: 258 data_time: 0.84s time: 119.83s eta: 12:12:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31629.0 tgs: 265 data_time: 0.69s time: 119.14s eta: 12:06:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31407.0 tgs: 263 data_time: 0.71s time: 119.17s eta: 12:04:57 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.110 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31371.0 tgs: 263 data_time: 0.83s time: 119.12s eta: 12:02:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.196 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 264 data_time: 0.68s time: 119.19s eta: 12:01:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.87s time: 118.95s eta: 11:57:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31328.0 tgs: 261 data_time: 0.80s time: 119.71s eta: 12:00:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.194 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31595.0 tgs: 264 data_time: 1.09s time: 119.28s eta: 11:55:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.193 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32038.0 tgs: 268 data_time: 0.97s time: 119.27s eta: 11:53:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.115 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 266 data_time: 0.96s time: 119.71s eta: 11:54:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32036.0 tgs: 269 data_time: 0.74s time: 118.78s eta: 11:46:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32466.0 tgs: 272 data_time: 0.84s time: 119.16s eta: 11:47:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32269.0 tgs: 269 data_time: 0.73s time: 119.61s eta: 11:47:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31643.0 tgs: 266 data_time: 0.55s time: 118.82s eta: 11:41:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32243.0 tgs: 271 data_time: 0.84s time: 118.69s eta: 11:38:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.119 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31173.0 tgs: 261 data_time: 0.84s time: 119.21s eta: 11:39:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31340.0 tgs: 263 data_time: 0.91s time: 119.14s eta: 11:36:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.188 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31830.0 tgs: 267 data_time: 0.54s time: 119.08s eta: 11:34:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.116 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30844.0 tgs: 259 data_time: 0.79s time: 118.95s eta: 11:31:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.197 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31527.0 tgs: 265 data_time: 0.92s time: 118.92s eta: 11:29:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.237 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32520.0 tgs: 273 data_time: 0.83s time: 118.97s eta: 11:28:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.114 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31546.0 tgs: 264 data_time: 0.64s time: 119.29s eta: 11:27:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.217 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 267 data_time: 0.71s time: 119.36s eta: 11:26:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31561.0 tgs: 265 data_time: 0.89s time: 118.89s eta: 11:21:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31786.0 tgs: 266 data_time: 0.71s time: 119.20s eta: 11:21:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31311.0 tgs: 261 data_time: 0.69s time: 119.72s eta: 11:22:23 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 268 data_time: 0.85s time: 119.12s eta: 11:16:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32434.0 tgs: 272 data_time: 0.72s time: 119.12s eta: 11:15:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.206 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30761.0 tgs: 258 data_time: 1.07s time: 118.94s eta: 11:11:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32182.0 tgs: 270 data_time: 0.82s time: 119.10s eta: 11:10:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.221 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31067.0 tgs: 260 data_time: 0.81s time: 119.16s eta: 11:09:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30765.0 tgs: 258 data_time: 0.63s time: 118.97s eta: 11:06:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.205 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31217.0 tgs: 261 data_time: 0.90s time: 119.47s eta: 11:07:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31905.0 tgs: 266 data_time: 1.04s time: 119.64s eta: 11:06:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.117 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32070.0 tgs: 268 data_time: 0.56s time: 119.30s eta: 11:02:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.112 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31471.0 tgs: 263 data_time: 0.51s time: 119.31s eta: 11:00:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.226 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32171.0 tgs: 270 data_time: 0.95s time: 118.82s eta: 10:55:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30309.0 tgs: 253 data_time: 0.62s time: 119.47s eta: 10:57:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31658.0 tgs: 265 data_time: 0.67s time: 119.37s eta: 10:54:32 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.199 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31451.0 tgs: 263 data_time: 0.99s time: 119.28s eta: 10:52:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30783.0 tgs: 258 data_time: 0.98s time: 119.26s eta: 10:49:57 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32225.0 tgs: 270 data_time: 0.72s time: 119.10s eta: 10:47:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.167 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 270 data_time: 0.84s time: 119.53s eta: 10:47:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.78s time: 119.31s eta: 10:44:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.088 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30201.0 tgs: 252 data_time: 0.66s time: 119.49s eta: 10:43:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.165 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31419.0 tgs: 262 data_time: 0.62s time: 119.52s eta: 10:41:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31770.0 tgs: 267 data_time: 0.69s time: 118.98s eta: 10:36:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.114 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31674.0 tgs: 265 data_time: 0.53s time: 119.22s eta: 10:35:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31133.0 tgs: 261 data_time: 0.66s time: 119.06s eta: 10:32:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 270 data_time: 0.90s time: 118.79s eta: 10:29:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32098.0 tgs: 267 data_time: 0.98s time: 119.86s eta: 10:33:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 268 data_time: 0.76s time: 119.38s eta: 10:28:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.130 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 267 data_time: 0.84s time: 118.84s eta: 10:23:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31451.0 tgs: 264 data_time: 0.78s time: 118.99s eta: 10:22:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.122 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32341.0 tgs: 271 data_time: 0.95s time: 119.29s eta: 10:22:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.124 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31759.0 tgs: 266 data_time: 0.78s time: 119.21s eta: 10:19:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.191 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31711.0 tgs: 266 data_time: 0.69s time: 119.06s eta: 10:17:07 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 267 data_time: 1.03s time: 119.48s eta: 10:17:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31261.0 tgs: 261 data_time: 0.73s time: 119.37s eta: 10:14:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.111 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31206.0 tgs: 261 data_time: 0.50s time: 119.45s eta: 10:13:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.202 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31861.0 tgs: 267 data_time: 0.56s time: 119.29s eta: 10:10:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31804.0 tgs: 267 data_time: 0.77s time: 119.11s eta: 10:07:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31898.0 tgs: 266 data_time: 0.56s time: 119.70s eta: 10:08:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.177 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32301.0 tgs: 269 data_time: 0.87s time: 119.83s eta: 10:07:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30369.0 tgs: 249 data_time: 1.28s time: 121.56s eta: 10:13:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32201.0 tgs: 270 data_time: 0.75s time: 119.26s eta: 10:00:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32151.0 tgs: 270 data_time: 0.58s time: 118.81s eta: 9:56:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31583.0 tgs: 264 data_time: 0.67s time: 119.58s eta: 9:57:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27870.0 tgs: 232 data_time: 0.69s time: 119.69s eta: 9:56:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.79s time: 119.26s eta: 9:52:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32124.0 tgs: 269 data_time: 0.66s time: 119.33s eta: 9:50:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31530.0 tgs: 264 data_time: 0.85s time: 119.23s eta: 9:48:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.234 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32408.0 tgs: 267 data_time: 1.01s time: 120.97s eta: 9:54:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.210 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31634.0 tgs: 265 data_time: 0.79s time: 119.22s eta: 9:44:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.212 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 267 data_time: 0.75s time: 119.46s eta: 9:43:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31751.0 tgs: 265 data_time: 0.82s time: 119.54s eta: 9:41:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31843.0 tgs: 267 data_time: 0.71s time: 119.21s eta: 9:38:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 267 data_time: 0.70s time: 119.48s eta: 9:37:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31199.0 tgs: 262 data_time: 0.70s time: 119.06s eta: 9:33:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31135.0 tgs: 261 data_time: 0.73s time: 119.23s eta: 9:32:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.190 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31798.0 tgs: 266 data_time: 0.98s time: 119.16s eta: 9:29:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.80s time: 119.23s eta: 9:28:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32526.0 tgs: 272 data_time: 0.79s time: 119.46s eta: 9:27:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.109 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32013.0 tgs: 269 data_time: 0.79s time: 118.99s eta: 9:23:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.202 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31837.0 tgs: 267 data_time: 0.61s time: 119.04s eta: 9:21:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31530.0 tgs: 263 data_time: 0.86s time: 119.54s eta: 9:21:48 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.195 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31694.0 tgs: 265 data_time: 0.58s time: 119.59s eta: 9:20:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31826.0 tgs: 267 data_time: 0.97s time: 119.13s eta: 9:15:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.235 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32408.0 tgs: 272 data_time: 0.70s time: 119.02s eta: 9:13:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.221 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32612.0 tgs: 274 data_time: 0.90s time: 118.88s eta: 9:10:48 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31444.0 tgs: 263 data_time: 1.02s time: 119.33s eta: 9:10:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31313.0 tgs: 262 data_time: 0.65s time: 119.08s eta: 9:07:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32168.0 tgs: 270 data_time: 0.70s time: 118.86s eta: 9:04:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.223 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 269 data_time: 0.69s time: 119.67s eta: 9:06:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.112 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32506.0 tgs: 272 data_time: 0.71s time: 119.44s eta: 9:03:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31034.0 tgs: 259 data_time: 0.89s time: 119.44s eta: 9:01:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31068.0 tgs: 261 data_time: 0.69s time: 118.73s eta: 8:56:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.177 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31474.0 tgs: 264 data_time: 0.72s time: 118.87s eta: 8:54:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31891.0 tgs: 267 data_time: 0.74s time: 119.34s eta: 8:55:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31624.0 tgs: 265 data_time: 0.76s time: 119.15s eta: 8:52:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.298 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31497.0 tgs: 264 data_time: 1.02s time: 119.27s eta: 8:50:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.183 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31830.0 tgs: 264 data_time: 0.82s time: 120.47s eta: 8:54:05 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 267 data_time: 0.84s time: 119.24s eta: 8:46:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 269 data_time: 0.84s time: 119.05s eta: 8:43:48 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30869.0 tgs: 255 data_time: 0.74s time: 120.82s eta: 8:49:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32365.0 tgs: 269 data_time: 0.86s time: 120.20s eta: 8:44:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30526.0 tgs: 256 data_time: 0.81s time: 118.83s eta: 8:36:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.201 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 267 data_time: 1.01s time: 119.35s eta: 8:37:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31160.0 tgs: 261 data_time: 0.84s time: 119.34s eta: 8:35:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31729.0 tgs: 266 data_time: 0.90s time: 119.05s eta: 8:31:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.192 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31042.0 tgs: 260 data_time: 0.80s time: 119.16s eta: 8:30:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.186 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30842.0 tgs: 259 data_time: 0.73s time: 119.04s eta: 8:27:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31401.0 tgs: 263 data_time: 0.70s time: 119.12s eta: 8:26:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32270.0 tgs: 270 data_time: 1.00s time: 119.24s eta: 8:24:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.232 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30885.0 tgs: 259 data_time: 1.21s time: 119.22s eta: 8:22:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32113.0 tgs: 268 data_time: 0.89s time: 119.38s eta: 8:21:23 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.183 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30617.0 tgs: 255 data_time: 0.75s time: 119.78s eta: 8:21:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30966.0 tgs: 259 data_time: 0.99s time: 119.20s eta: 8:16:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31021.0 tgs: 260 data_time: 0.69s time: 118.98s eta: 8:13:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.122 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31584.0 tgs: 263 data_time: 0.56s time: 119.90s eta: 8:15:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.67s time: 119.56s eta: 8:12:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 265 data_time: 0.64s time: 119.59s eta: 8:10:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31358.0 tgs: 263 data_time: 0.69s time: 119.00s eta: 8:05:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 263 data_time: 0.79s time: 119.84s eta: 8:07:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.203 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31947.0 tgs: 266 data_time: 1.18s time: 119.71s eta: 8:04:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32581.0 tgs: 272 data_time: 0.76s time: 119.40s eta: 8:01:35 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32202.0 tgs: 269 data_time: 0.45s time: 119.29s eta: 7:59:07 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.217 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32132.0 tgs: 269 data_time: 0.72s time: 119.19s eta: 7:56:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.194 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30752.0 tgs: 257 data_time: 0.73s time: 119.20s eta: 7:54:48 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30695.0 tgs: 258 data_time: 0.70s time: 118.94s eta: 7:51:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31639.0 tgs: 265 data_time: 0.71s time: 118.97s eta: 7:49:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 268 data_time: 0.62s time: 119.42s eta: 7:49:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32232.0 tgs: 270 data_time: 0.85s time: 119.04s eta: 7:46:14 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.238 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30972.0 tgs: 260 data_time: 0.93s time: 119.07s eta: 7:44:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.202 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31635.0 tgs: 263 data_time: 1.12s time: 120.00s eta: 7:45:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32102.0 tgs: 267 data_time: 0.69s time: 119.83s eta: 7:43:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31869.0 tgs: 267 data_time: 0.85s time: 119.16s eta: 7:38:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31579.0 tgs: 265 data_time: 0.64s time: 118.97s eta: 7:36:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31285.0 tgs: 262 data_time: 0.64s time: 119.33s eta: 7:35:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30900.0 tgs: 258 data_time: 0.71s time: 119.38s eta: 7:33:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32129.0 tgs: 269 data_time: 0.53s time: 119.06s eta: 7:30:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.161 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30797.0 tgs: 258 data_time: 0.57s time: 118.99s eta: 7:28:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31609.0 tgs: 265 data_time: 0.72s time: 119.24s eta: 7:27:09 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.106 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32408.0 tgs: 272 data_time: 0.70s time: 118.99s eta: 7:24:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31012.0 tgs: 260 data_time: 0.40s time: 118.93s eta: 7:22:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31290.0 tgs: 262 data_time: 0.64s time: 119.08s eta: 7:20:35 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31489.0 tgs: 264 data_time: 0.93s time: 118.99s eta: 7:18:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30842.0 tgs: 259 data_time: 0.56s time: 119.03s eta: 7:16:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31723.0 tgs: 266 data_time: 0.67s time: 119.25s eta: 7:15:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31898.0 tgs: 267 data_time: 0.69s time: 119.35s eta: 7:13:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31408.0 tgs: 264 data_time: 0.68s time: 118.94s eta: 7:10:10 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.107 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30956.0 tgs: 259 data_time: 1.02s time: 119.21s eta: 7:09:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31446.0 tgs: 262 data_time: 0.67s time: 119.59s eta: 7:08:30 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32057.0 tgs: 267 data_time: 0.70s time: 119.66s eta: 7:06:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30995.0 tgs: 260 data_time: 0.88s time: 118.96s eta: 7:02:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 0.76s time: 119.30s eta: 7:01:31 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.287 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31067.0 tgs: 259 data_time: 0.77s time: 119.52s eta: 7:00:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31134.0 tgs: 260 data_time: 0.56s time: 119.44s eta: 6:58:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31956.0 tgs: 268 data_time: 0.61s time: 118.90s eta: 6:54:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30944.0 tgs: 259 data_time: 0.72s time: 119.34s eta: 6:53:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31741.0 tgs: 266 data_time: 0.76s time: 118.94s eta: 6:50:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32494.0 tgs: 272 data_time: 0.75s time: 119.27s eta: 6:49:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 267 data_time: 0.56s time: 119.02s eta: 6:46:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.192 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31928.0 tgs: 267 data_time: 0.88s time: 119.23s eta: 6:45:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 265 data_time: 0.97s time: 119.37s eta: 6:43:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31380.0 tgs: 263 data_time: 0.78s time: 119.02s eta: 6:40:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30555.0 tgs: 256 data_time: 0.69s time: 119.01s eta: 6:38:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.109 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.86s time: 119.46s eta: 6:38:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32514.0 tgs: 273 data_time: 0.72s time: 118.80s eta: 6:34:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32429.0 tgs: 271 data_time: 0.82s time: 119.44s eta: 6:34:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.255 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32640.0 tgs: 272 data_time: 1.19s time: 119.57s eta: 6:32:35 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32408.0 tgs: 271 data_time: 0.79s time: 119.38s eta: 6:29:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32250.0 tgs: 270 data_time: 0.92s time: 119.07s eta: 6:26:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31777.0 tgs: 266 data_time: 0.66s time: 119.36s eta: 6:25:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 266 data_time: 0.80s time: 118.83s eta: 6:22:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31981.0 tgs: 268 data_time: 0.76s time: 119.01s eta: 6:20:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31265.0 tgs: 261 data_time: 0.68s time: 119.40s eta: 6:20:05 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29951.0 tgs: 250 data_time: 0.71s time: 119.35s eta: 6:17:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31228.0 tgs: 262 data_time: 0.61s time: 119.04s eta: 6:14:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31526.0 tgs: 264 data_time: 0.50s time: 119.38s eta: 6:14:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.106 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 269 data_time: 0.64s time: 119.39s eta: 6:12:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31862.0 tgs: 267 data_time: 0.69s time: 119.28s eta: 6:09:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32242.0 tgs: 270 data_time: 0.68s time: 119.10s eta: 6:07:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31870.0 tgs: 261 data_time: 0.93s time: 121.96s eta: 6:14:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32375.0 tgs: 271 data_time: 0.53s time: 119.39s eta: 6:04:07 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31437.0 tgs: 262 data_time: 0.61s time: 119.57s eta: 6:02:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.179 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31866.0 tgs: 267 data_time: 0.63s time: 119.26s eta: 5:59:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31602.0 tgs: 264 data_time: 0.86s time: 119.27s eta: 5:57:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31467.0 tgs: 264 data_time: 0.89s time: 118.91s eta: 5:54:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.186 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 269 data_time: 0.71s time: 119.02s eta: 5:53:05 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31332.0 tgs: 263 data_time: 0.85s time: 119.07s eta: 5:51:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31507.0 tgs: 264 data_time: 0.96s time: 119.29s eta: 5:49:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31229.0 tgs: 262 data_time: 0.75s time: 118.97s eta: 5:46:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.269 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30876.0 tgs: 258 data_time: 0.93s time: 119.27s eta: 5:45:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.73s time: 119.32s eta: 5:44:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31958.0 tgs: 268 data_time: 0.85s time: 119.05s eta: 5:41:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.179 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.72s time: 119.50s eta: 5:40:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31766.0 tgs: 266 data_time: 0.73s time: 119.38s eta: 5:38:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.186 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32039.0 tgs: 267 data_time: 0.92s time: 119.57s eta: 5:36:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.79s time: 118.99s eta: 5:33:10 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 267 data_time: 0.83s time: 118.94s eta: 5:31:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 270 data_time: 0.76s time: 119.24s eta: 5:29:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 265 data_time: 0.79s time: 119.40s eta: 5:28:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.187 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31958.0 tgs: 267 data_time: 0.93s time: 119.46s eta: 5:26:31 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.183 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31897.0 tgs: 268 data_time: 0.54s time: 118.95s eta: 5:23:09 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32191.0 tgs: 269 data_time: 0.81s time: 119.63s eta: 5:22:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.107 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 26257.0 tgs: 219 data_time: 0.49s time: 119.35s eta: 5:20:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31210.0 tgs: 262 data_time: 0.84s time: 119.12s eta: 5:17:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31275.0 tgs: 261 data_time: 1.02s time: 119.43s eta: 5:16:28 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32370.0 tgs: 270 data_time: 0.87s time: 119.57s eta: 5:14:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.110 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31607.0 tgs: 265 data_time: 0.67s time: 119.01s eta: 5:11:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.213 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31756.0 tgs: 266 data_time: 0.77s time: 119.00s eta: 5:09:23 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 270 data_time: 0.84s time: 119.23s eta: 5:08:00 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30943.0 tgs: 258 data_time: 0.86s time: 119.69s eta: 5:07:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.84s time: 119.12s eta: 5:03:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30938.0 tgs: 259 data_time: 0.69s time: 119.07s eta: 5:01:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32052.0 tgs: 263 data_time: 0.66s time: 121.52s eta: 5:05:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30998.0 tgs: 259 data_time: 0.90s time: 119.58s eta: 4:58:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31855.0 tgs: 266 data_time: 0.69s time: 119.51s eta: 4:56:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31404.0 tgs: 262 data_time: 0.68s time: 119.73s eta: 4:55:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30901.0 tgs: 258 data_time: 0.64s time: 119.52s eta: 4:52:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31560.0 tgs: 264 data_time: 1.09s time: 119.42s eta: 4:50:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.78s time: 119.20s eta: 4:48:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 269 data_time: 0.67s time: 118.89s eta: 4:45:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30964.0 tgs: 260 data_time: 0.57s time: 119.04s eta: 4:43:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29944.0 tgs: 249 data_time: 0.68s time: 119.98s eta: 4:43:57 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.65s time: 119.15s eta: 4:39:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30152.0 tgs: 253 data_time: 0.75s time: 119.01s eta: 4:37:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31939.0 tgs: 268 data_time: 0.88s time: 119.06s eta: 4:35:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.222 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 267 data_time: 1.02s time: 119.37s eta: 4:34:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.194 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31461.0 tgs: 263 data_time: 0.68s time: 119.41s eta: 4:32:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.112 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30782.0 tgs: 258 data_time: 0.90s time: 118.94s eta: 4:29:36 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31712.0 tgs: 265 data_time: 0.76s time: 119.60s eta: 4:29:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31675.0 tgs: 265 data_time: 0.79s time: 119.24s eta: 4:26:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 0.51s time: 119.07s eta: 4:23:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31746.0 tgs: 267 data_time: 0.65s time: 118.85s eta: 4:21:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31114.0 tgs: 261 data_time: 0.87s time: 118.76s eta: 4:19:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31885.0 tgs: 266 data_time: 0.68s time: 119.43s eta: 4:18:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31607.0 tgs: 264 data_time: 0.80s time: 119.40s eta: 4:16:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31454.0 tgs: 263 data_time: 0.69s time: 119.32s eta: 4:14:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32235.0 tgs: 270 data_time: 0.62s time: 119.05s eta: 4:11:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32047.0 tgs: 269 data_time: 1.06s time: 119.13s eta: 4:10:09 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31972.0 tgs: 268 data_time: 0.62s time: 119.13s eta: 4:08:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 270 data_time: 0.74s time: 118.96s eta: 4:05:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32059.0 tgs: 269 data_time: 0.71s time: 119.16s eta: 4:04:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32173.0 tgs: 269 data_time: 0.62s time: 119.50s eta: 4:02:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31061.0 tgs: 260 data_time: 0.80s time: 119.25s eta: 4:00:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31603.0 tgs: 265 data_time: 0.76s time: 119.19s eta: 3:58:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 270 data_time: 0.68s time: 119.36s eta: 3:56:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30680.0 tgs: 257 data_time: 0.88s time: 119.37s eta: 3:54:45 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32032.0 tgs: 267 data_time: 0.73s time: 119.95s eta: 3:53:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32258.0 tgs: 270 data_time: 0.54s time: 119.32s eta: 3:50:41 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31597.0 tgs: 264 data_time: 0.68s time: 119.63s eta: 3:49:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31840.0 tgs: 267 data_time: 0.93s time: 118.86s eta: 3:45:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32295.0 tgs: 269 data_time: 1.11s time: 119.70s eta: 3:45:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30898.0 tgs: 258 data_time: 0.69s time: 119.33s eta: 3:42:44 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32059.0 tgs: 268 data_time: 0.78s time: 119.26s eta: 3:40:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.79s time: 119.00s eta: 3:38:09 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31865.0 tgs: 266 data_time: 0.59s time: 119.44s eta: 3:36:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31605.0 tgs: 265 data_time: 0.68s time: 119.03s eta: 3:34:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.202 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31589.0 tgs: 265 data_time: 0.84s time: 118.99s eta: 3:32:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30677.0 tgs: 257 data_time: 0.88s time: 119.22s eta: 3:30:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31910.0 tgs: 265 data_time: 0.71s time: 120.11s eta: 3:30:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31878.0 tgs: 267 data_time: 0.89s time: 119.13s eta: 3:26:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30743.0 tgs: 257 data_time: 0.65s time: 119.32s eta: 3:24:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.69s time: 119.08s eta: 3:22:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.191 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31569.0 tgs: 264 data_time: 1.01s time: 119.43s eta: 3:21:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.207 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31958.0 tgs: 267 data_time: 0.69s time: 119.29s eta: 3:18:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30267.0 tgs: 253 data_time: 0.85s time: 119.22s eta: 3:16:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31449.0 tgs: 263 data_time: 0.94s time: 119.13s eta: 3:14:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31466.0 tgs: 264 data_time: 0.73s time: 119.03s eta: 3:12:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.205 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31925.0 tgs: 267 data_time: 0.70s time: 119.19s eta: 3:10:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31461.0 tgs: 260 data_time: 0.67s time: 120.92s eta: 3:11:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31837.0 tgs: 266 data_time: 0.75s time: 119.49s eta: 3:07:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31360.0 tgs: 262 data_time: 0.69s time: 119.26s eta: 3:04:51 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30682.0 tgs: 257 data_time: 0.66s time: 118.99s eta: 3:02:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31461.0 tgs: 263 data_time: 0.78s time: 119.22s eta: 3:00:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32106.0 tgs: 268 data_time: 0.52s time: 119.46s eta: 2:59:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30882.0 tgs: 259 data_time: 0.58s time: 118.98s eta: 2:56:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32623.0 tgs: 273 data_time: 0.98s time: 119.36s eta: 2:55:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 265 data_time: 0.67s time: 119.41s eta: 2:53:08 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31877.0 tgs: 267 data_time: 0.95s time: 119.35s eta: 2:51:04 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31673.0 tgs: 260 data_time: 0.62s time: 121.46s eta: 2:52:03 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30964.0 tgs: 259 data_time: 0.49s time: 119.24s eta: 2:46:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32388.0 tgs: 270 data_time: 0.85s time: 119.54s eta: 2:45:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 266 data_time: 0.68s time: 118.91s eta: 2:42:30 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31024.0 tgs: 260 data_time: 0.81s time: 119.18s eta: 2:40:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 269 data_time: 0.74s time: 118.97s eta: 2:38:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30283.0 tgs: 253 data_time: 1.00s time: 119.23s eta: 2:36:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31587.0 tgs: 265 data_time: 0.91s time: 119.11s eta: 2:34:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30966.0 tgs: 259 data_time: 0.71s time: 119.44s eta: 2:33:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31592.0 tgs: 262 data_time: 0.83s time: 120.22s eta: 2:32:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31907.0 tgs: 267 data_time: 0.71s time: 119.28s eta: 2:29:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 269 data_time: 0.62s time: 118.88s eta: 2:26:37 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31199.0 tgs: 261 data_time: 0.53s time: 119.52s eta: 2:25:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 267 data_time: 0.64s time: 119.07s eta: 2:22:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31412.0 tgs: 263 data_time: 0.61s time: 119.14s eta: 2:20:58 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30660.0 tgs: 257 data_time: 0.72s time: 119.05s eta: 2:18:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31816.0 tgs: 265 data_time: 0.70s time: 119.84s eta: 2:17:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32007.0 tgs: 268 data_time: 0.80s time: 119.31s eta: 2:15:13 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30289.0 tgs: 253 data_time: 0.75s time: 119.39s eta: 2:13:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 267 data_time: 0.74s time: 119.68s eta: 2:11:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31019.0 tgs: 259 data_time: 0.77s time: 119.50s eta: 2:09:27 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31730.0 tgs: 266 data_time: 0.79s time: 118.96s eta: 2:06:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 265 data_time: 0.85s time: 119.37s eta: 2:05:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.260 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30443.0 tgs: 255 data_time: 0.86s time: 119.28s eta: 2:03:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30905.0 tgs: 259 data_time: 0.63s time: 119.23s eta: 2:01:12 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31709.0 tgs: 265 data_time: 0.72s time: 119.50s eta: 1:59:30 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 264 data_time: 0.67s time: 119.42s eta: 1:57:26 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 265 data_time: 0.66s time: 119.53s eta: 1:55:32 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31060.0 tgs: 260 data_time: 0.58s time: 119.31s eta: 1:53:20 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 267 data_time: 0.65s time: 119.25s eta: 1:51:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 265 data_time: 0.84s time: 118.94s eta: 1:49:01 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30865.0 tgs: 257 data_time: 0.63s time: 119.63s eta: 1:47:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 267 data_time: 0.75s time: 119.70s eta: 1:45:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31319.0 tgs: 262 data_time: 0.60s time: 119.33s eta: 1:43:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31900.0 tgs: 267 data_time: 0.68s time: 119.31s eta: 1:41:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.75s time: 119.29s eta: 1:39:24 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31471.0 tgs: 264 data_time: 0.86s time: 118.91s eta: 1:37:06 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30797.0 tgs: 258 data_time: 0.79s time: 119.16s eta: 1:35:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31962.0 tgs: 268 data_time: 0.95s time: 119.06s eta: 1:33:15 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32074.0 tgs: 268 data_time: 0.67s time: 119.26s eta: 1:31:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 268 data_time: 0.58s time: 119.72s eta: 1:29:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.099 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31450.0 tgs: 264 data_time: 0.75s time: 119.02s eta: 1:27:16 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31320.0 tgs: 263 data_time: 0.52s time: 118.87s eta: 1:25:11 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31268.0 tgs: 261 data_time: 0.61s time: 119.48s eta: 1:23:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.204 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 269 data_time: 0.72s time: 119.49s eta: 1:21:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32419.0 tgs: 272 data_time: 0.67s time: 118.94s eta: 1:19:17 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31308.0 tgs: 263 data_time: 0.81s time: 118.93s eta: 1:17:18 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30939.0 tgs: 253 data_time: 0.76s time: 122.14s eta: 1:17:21 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31111.0 tgs: 260 data_time: 0.75s time: 119.34s eta: 1:13:35 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.215 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.58s time: 119.05s eta: 1:11:25 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31555.0 tgs: 265 data_time: 0.56s time: 118.92s eta: 1:09:22 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31167.0 tgs: 261 data_time: 0.80s time: 119.10s eta: 1:07:29 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31922.0 tgs: 267 data_time: 0.88s time: 119.42s eta: 1:05:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32251.0 tgs: 270 data_time: 0.61s time: 119.35s eta: 1:03:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31505.0 tgs: 264 data_time: 0.60s time: 119.18s eta: 1:01:34 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30900.0 tgs: 259 data_time: 0.53s time: 119.27s eta: 0:59:38 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 264 data_time: 1.04s time: 119.45s eta: 0:57:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31319.0 tgs: 262 data_time: 0.66s time: 119.40s eta: 0:55:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32585.0 tgs: 272 data_time: 0.67s time: 119.48s eta: 0:53:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31416.0 tgs: 263 data_time: 0.94s time: 119.37s eta: 0:51:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30506.0 tgs: 255 data_time: 0.68s time: 119.47s eta: 0:49:46 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31605.0 tgs: 265 data_time: 0.59s time: 118.90s eta: 0:47:33 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30844.0 tgs: 258 data_time: 0.85s time: 119.26s eta: 0:45:43 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32227.0 tgs: 270 data_time: 0.71s time: 119.19s eta: 0:43:42 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.196 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.67s time: 119.00s eta: 0:41:39 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31081.0 tgs: 261 data_time: 0.95s time: 119.04s eta: 0:39:40 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32146.0 tgs: 268 data_time: 0.67s time: 119.66s eta: 0:37:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.220 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 266 data_time: 0.78s time: 119.29s eta: 0:35:47 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31969.0 tgs: 266 data_time: 0.81s time: 119.75s eta: 0:33:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31841.0 tgs: 262 data_time: 0.82s time: 121.22s eta: 0:32:19 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31373.0 tgs: 262 data_time: 0.58s time: 119.38s eta: 0:29:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 29852.0 tgs: 250 data_time: 0.75s time: 119.32s eta: 0:27:50 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32191.0 tgs: 269 data_time: 0.58s time: 119.64s eta: 0:25:55 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31841.0 tgs: 267 data_time: 0.91s time: 119.09s eta: 0:23:49 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32422.0 tgs: 271 data_time: 0.64s time: 119.39s eta: 0:21:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31197.0 tgs: 261 data_time: 0.85s time: 119.26s eta: 0:19:52 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30716.0 tgs: 257 data_time: 1.12s time: 119.24s eta: 0:17:53 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31330.0 tgs: 261 data_time: 0.59s time: 119.90s eta: 0:15:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32200.0 tgs: 270 data_time: 0.89s time: 119.18s eta: 0:13:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31804.0 tgs: 267 data_time: 0.91s time: 119.09s eta: 0:11:54 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 269 data_time: 0.74s time: 119.32s eta: 0:09:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31296.0 tgs: 262 data_time: 0.45s time: 119.17s eta: 0:07:56 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 269 data_time: 0.66s time: 119.69s eta: 0:05:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31209.0 tgs: 256 data_time: 0.79s time: 121.45s eta: 0:04:02 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30751.0 tgs: 257 data_time: 0.76s time: 119.21s eta: 0:01:59 +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.29 seconds, peak gpu memory 17.4G +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 35][DP 8][SP 3][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank36.log b/20250121165312/rank36.log new file mode 100644 index 0000000000000000000000000000000000000000..411d684cb98e7ffde3d6845524cfe18db776be43 --- /dev/null +++ b/20250121165312/rank36.log @@ -0,0 +1,851 @@ +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:35][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.02s +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.78 seconds, peak gpu memory 13.4G +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.339 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 32281.0 tgs: 212 data_time: 2.27s time: 152.13s eta: 1 day, 1:03:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.237 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32206.0 tgs: 267 data_time: 0.74s time: 120.20s eta: 19:45:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.217 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32213.0 tgs: 269 data_time: 0.92s time: 119.68s eta: 19:38:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29331.0 tgs: 245 data_time: 0.82s time: 119.65s eta: 19:36:36 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.222 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31870.0 tgs: 267 data_time: 0.86s time: 119.31s eta: 19:31:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.193 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29472.0 tgs: 246 data_time: 0.68s time: 119.54s eta: 19:31:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 261 data_time: 0.69s time: 122.96s eta: 20:02:58 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.350 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30854.0 tgs: 257 data_time: 0.63s time: 119.89s eta: 19:30:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31900.0 tgs: 264 data_time: 0.71s time: 120.65s eta: 19:36:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.216 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32109.0 tgs: 267 data_time: 0.74s time: 119.92s eta: 19:27:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.220 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31161.0 tgs: 258 data_time: 0.57s time: 120.46s eta: 19:30:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.185 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31536.0 tgs: 264 data_time: 0.77s time: 119.35s eta: 19:17:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 267 data_time: 0.75s time: 119.67s eta: 19:18:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32359.0 tgs: 271 data_time: 0.73s time: 119.18s eta: 19:12:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.251 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32408.0 tgs: 271 data_time: 0.94s time: 119.18s eta: 19:10:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31413.0 tgs: 264 data_time: 0.92s time: 118.93s eta: 19:05:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31857.0 tgs: 267 data_time: 0.75s time: 119.14s eta: 19:05:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31570.0 tgs: 264 data_time: 0.71s time: 119.18s eta: 19:04:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31566.0 tgs: 264 data_time: 0.65s time: 119.40s eta: 19:04:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32203.0 tgs: 269 data_time: 0.77s time: 119.33s eta: 19:01:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30573.0 tgs: 254 data_time: 0.63s time: 120.31s eta: 19:08:58 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30164.0 tgs: 245 data_time: 0.68s time: 122.92s eta: 19:31:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31721.0 tgs: 265 data_time: 0.75s time: 119.48s eta: 18:57:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31235.0 tgs: 262 data_time: 0.74s time: 119.16s eta: 18:52:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 269 data_time: 0.69s time: 119.24s eta: 18:50:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31111.0 tgs: 260 data_time: 0.88s time: 119.39s eta: 18:50:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31722.0 tgs: 265 data_time: 0.81s time: 119.66s eta: 18:50:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30903.0 tgs: 258 data_time: 0.77s time: 119.39s eta: 18:46:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31566.0 tgs: 263 data_time: 0.98s time: 119.77s eta: 18:47:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 28740.0 tgs: 222 data_time: 0.80s time: 129.31s eta: 20:15:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31943.0 tgs: 267 data_time: 0.60s time: 119.58s eta: 18:42:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.108 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31226.0 tgs: 261 data_time: 0.82s time: 119.38s eta: 18:38:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31930.0 tgs: 264 data_time: 0.64s time: 120.66s eta: 18:48:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31763.0 tgs: 266 data_time: 0.87s time: 119.40s eta: 18:34:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32155.0 tgs: 269 data_time: 0.81s time: 119.16s eta: 18:30:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31819.0 tgs: 267 data_time: 0.87s time: 119.07s eta: 18:27:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32555.0 tgs: 273 data_time: 0.62s time: 119.14s eta: 18:25:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31507.0 tgs: 261 data_time: 0.73s time: 120.39s eta: 18:35:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32485.0 tgs: 268 data_time: 0.85s time: 121.17s eta: 18:40:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32286.0 tgs: 267 data_time: 0.89s time: 120.67s eta: 18:34:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32270.0 tgs: 269 data_time: 0.74s time: 119.73s eta: 18:23:32 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.355 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 257 data_time: 0.92s time: 123.88s eta: 18:59:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31139.0 tgs: 260 data_time: 0.73s time: 119.55s eta: 18:17:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 266 data_time: 0.66s time: 119.32s eta: 18:13:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32014.0 tgs: 268 data_time: 0.84s time: 119.23s eta: 18:10:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31950.0 tgs: 267 data_time: 0.97s time: 119.60s eta: 18:12:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31159.0 tgs: 261 data_time: 0.59s time: 119.27s eta: 18:07:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31876.0 tgs: 267 data_time: 0.91s time: 119.01s eta: 18:02:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31681.0 tgs: 265 data_time: 0.87s time: 119.10s eta: 18:01:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31594.0 tgs: 264 data_time: 1.14s time: 119.26s eta: 18:01:18 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31032.0 tgs: 260 data_time: 1.00s time: 119.33s eta: 17:59:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31014.0 tgs: 258 data_time: 0.92s time: 120.02s eta: 18:04:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32168.0 tgs: 270 data_time: 0.94s time: 119.10s eta: 17:53:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.71s time: 119.42s eta: 17:54:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 272 data_time: 0.91s time: 119.00s eta: 17:49:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31439.0 tgs: 264 data_time: 0.67s time: 119.03s eta: 17:47:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31610.0 tgs: 252 data_time: 0.56s time: 125.37s eta: 18:42:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31493.0 tgs: 262 data_time: 0.92s time: 120.11s eta: 17:52:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31578.0 tgs: 264 data_time: 0.76s time: 119.36s eta: 17:44:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.260 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32478.0 tgs: 272 data_time: 1.00s time: 119.31s eta: 17:41:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29943.0 tgs: 250 data_time: 0.83s time: 119.51s eta: 17:41:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31038.0 tgs: 246 data_time: 0.54s time: 126.05s eta: 18:37:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31043.0 tgs: 253 data_time: 0.80s time: 122.64s eta: 18:05:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32383.0 tgs: 267 data_time: 0.86s time: 121.16s eta: 17:50:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31991.0 tgs: 268 data_time: 0.63s time: 119.04s eta: 17:29:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30412.0 tgs: 255 data_time: 0.83s time: 118.86s eta: 17:25:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31377.0 tgs: 253 data_time: 1.01s time: 124.00s eta: 18:09:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 254 data_time: 0.74s time: 124.57s eta: 18:12:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32481.0 tgs: 272 data_time: 0.95s time: 119.40s eta: 17:24:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31743.0 tgs: 265 data_time: 0.97s time: 119.69s eta: 17:25:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32084.0 tgs: 268 data_time: 0.87s time: 119.53s eta: 17:21:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32042.0 tgs: 269 data_time: 0.77s time: 118.86s eta: 17:14:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 1.00s time: 119.85s eta: 17:20:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29019.0 tgs: 242 data_time: 0.78s time: 119.52s eta: 17:15:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.260 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31875.0 tgs: 261 data_time: 0.77s time: 121.86s eta: 17:34:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32461.0 tgs: 262 data_time: 0.59s time: 123.52s eta: 17:46:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.115 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31810.0 tgs: 267 data_time: 0.65s time: 119.01s eta: 17:05:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 267 data_time: 0.86s time: 119.23s eta: 17:05:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.239 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31702.0 tgs: 266 data_time: 0.77s time: 119.14s eta: 17:02:36 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.119 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 28580.0 tgs: 239 data_time: 0.87s time: 119.26s eta: 17:01:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31749.0 tgs: 266 data_time: 0.78s time: 119.04s eta: 16:57:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31455.0 tgs: 263 data_time: 0.84s time: 119.28s eta: 16:57:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30811.0 tgs: 258 data_time: 0.73s time: 119.16s eta: 16:54:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31311.0 tgs: 261 data_time: 0.65s time: 119.59s eta: 16:56:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32026.0 tgs: 268 data_time: 0.73s time: 119.13s eta: 16:50:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31318.0 tgs: 263 data_time: 0.64s time: 118.84s eta: 16:46:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.121 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31265.0 tgs: 262 data_time: 0.91s time: 119.02s eta: 16:45:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31698.0 tgs: 264 data_time: 0.71s time: 119.85s eta: 16:50:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32116.0 tgs: 269 data_time: 0.72s time: 119.36s eta: 16:44:36 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31820.0 tgs: 267 data_time: 0.56s time: 119.01s eta: 16:39:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.128 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32398.0 tgs: 267 data_time: 0.71s time: 120.93s eta: 16:53:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32077.0 tgs: 269 data_time: 0.78s time: 119.00s eta: 16:35:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 268 data_time: 0.83s time: 119.28s eta: 16:36:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 265 data_time: 0.62s time: 119.23s eta: 16:33:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31125.0 tgs: 260 data_time: 0.72s time: 119.37s eta: 16:32:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31589.0 tgs: 263 data_time: 0.85s time: 119.71s eta: 16:33:34 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.212 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 264 data_time: 0.68s time: 119.20s eta: 16:27:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30913.0 tgs: 258 data_time: 0.71s time: 119.77s eta: 16:30:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 260 data_time: 0.82s time: 120.28s eta: 16:32:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.245 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32113.0 tgs: 269 data_time: 1.05s time: 118.97s eta: 16:19:32 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31989.0 tgs: 251 data_time: 0.92s time: 127.33s eta: 17:26:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.230 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31887.0 tgs: 267 data_time: 0.71s time: 119.23s eta: 16:17:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 267 data_time: 0.55s time: 119.14s eta: 16:14:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.120 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31366.0 tgs: 263 data_time: 0.70s time: 119.21s eta: 16:13:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.125 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31395.0 tgs: 263 data_time: 0.83s time: 119.17s eta: 16:11:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32048.0 tgs: 269 data_time: 1.04s time: 118.96s eta: 16:07:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.251 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31406.0 tgs: 264 data_time: 0.68s time: 118.86s eta: 16:04:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31434.0 tgs: 264 data_time: 0.55s time: 118.96s eta: 16:03:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 258 data_time: 0.74s time: 121.85s eta: 16:24:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31767.0 tgs: 266 data_time: 0.73s time: 119.00s eta: 15:59:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31044.0 tgs: 260 data_time: 0.81s time: 119.22s eta: 15:59:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 267 data_time: 0.77s time: 119.62s eta: 16:00:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 271 data_time: 0.69s time: 119.05s eta: 15:54:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32439.0 tgs: 272 data_time: 0.64s time: 118.98s eta: 15:51:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.71s time: 119.17s eta: 15:51:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31837.0 tgs: 266 data_time: 0.56s time: 119.64s eta: 15:53:06 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 268 data_time: 0.74s time: 118.94s eta: 15:45:34 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32245.0 tgs: 270 data_time: 0.95s time: 119.29s eta: 15:46:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31475.0 tgs: 264 data_time: 0.97s time: 119.10s eta: 15:42:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31148.0 tgs: 261 data_time: 0.80s time: 118.99s eta: 15:40:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31417.0 tgs: 262 data_time: 0.82s time: 119.66s eta: 15:43:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 259 data_time: 0.75s time: 121.50s eta: 15:55:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31756.0 tgs: 265 data_time: 0.75s time: 119.41s eta: 15:37:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.249 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 247 data_time: 0.97s time: 129.96s eta: 16:58:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31805.0 tgs: 266 data_time: 1.01s time: 119.47s eta: 15:33:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31292.0 tgs: 261 data_time: 0.87s time: 119.61s eta: 15:32:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30814.0 tgs: 257 data_time: 0.83s time: 119.53s eta: 15:30:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31413.0 tgs: 257 data_time: 0.68s time: 122.14s eta: 15:48:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 265 data_time: 0.94s time: 120.76s eta: 15:35:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 265 data_time: 0.81s time: 119.41s eta: 15:23:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.68s time: 119.45s eta: 15:21:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31520.0 tgs: 239 data_time: 0.83s time: 131.64s eta: 16:53:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31932.0 tgs: 267 data_time: 0.59s time: 119.30s eta: 15:16:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30315.0 tgs: 253 data_time: 0.68s time: 119.52s eta: 15:16:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31784.0 tgs: 259 data_time: 0.78s time: 122.25s eta: 15:35:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.108 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 26939.0 tgs: 225 data_time: 0.42s time: 119.71s eta: 15:13:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31378.0 tgs: 262 data_time: 1.04s time: 119.36s eta: 15:09:08 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31717.0 tgs: 266 data_time: 0.79s time: 118.89s eta: 15:03:34 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 270 data_time: 0.82s time: 119.00s eta: 15:02:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.335 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32422.0 tgs: 271 data_time: 0.85s time: 119.31s eta: 15:02:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30691.0 tgs: 258 data_time: 0.74s time: 118.93s eta: 14:57:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31614.0 tgs: 262 data_time: 0.81s time: 120.52s eta: 15:07:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31621.0 tgs: 265 data_time: 0.54s time: 119.31s eta: 14:56:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.188 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29421.0 tgs: 239 data_time: 0.73s time: 122.66s eta: 15:19:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32141.0 tgs: 269 data_time: 0.76s time: 119.13s eta: 14:51:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31387.0 tgs: 262 data_time: 0.73s time: 119.76s eta: 14:54:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 263 data_time: 0.84s time: 119.47s eta: 14:50:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 260 data_time: 0.82s time: 118.99s eta: 14:44:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.105 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30723.0 tgs: 257 data_time: 0.51s time: 119.30s eta: 14:44:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.219 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31767.0 tgs: 267 data_time: 0.84s time: 118.94s eta: 14:40:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30699.0 tgs: 257 data_time: 0.82s time: 119.06s eta: 14:39:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.216 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.63s time: 119.06s eta: 14:37:06 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32007.0 tgs: 268 data_time: 0.52s time: 119.13s eta: 14:35:34 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.210 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32346.0 tgs: 271 data_time: 0.97s time: 119.00s eta: 14:32:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 269 data_time: 0.79s time: 119.08s eta: 14:31:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.111 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31216.0 tgs: 261 data_time: 0.79s time: 119.15s eta: 14:29:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 261 data_time: 0.94s time: 119.65s eta: 14:31:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32106.0 tgs: 269 data_time: 0.72s time: 119.18s eta: 14:26:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31674.0 tgs: 265 data_time: 0.93s time: 119.20s eta: 14:24:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31484.0 tgs: 263 data_time: 0.87s time: 119.39s eta: 14:23:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31704.0 tgs: 262 data_time: 0.60s time: 120.74s eta: 14:31:18 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31765.0 tgs: 267 data_time: 0.85s time: 118.85s eta: 14:15:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31034.0 tgs: 260 data_time: 0.89s time: 119.10s eta: 14:15:31 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.226 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32110.0 tgs: 269 data_time: 1.08s time: 119.11s eta: 14:13:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 261 data_time: 0.94s time: 119.94s eta: 14:17:36 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28445.0 tgs: 237 data_time: 0.67s time: 119.87s eta: 14:15:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 267 data_time: 0.76s time: 119.59s eta: 14:11:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31811.0 tgs: 266 data_time: 0.80s time: 119.23s eta: 14:06:31 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31451.0 tgs: 263 data_time: 0.72s time: 119.26s eta: 14:04:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.124 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31249.0 tgs: 262 data_time: 0.61s time: 119.27s eta: 14:02:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32506.0 tgs: 272 data_time: 0.79s time: 119.15s eta: 14:00:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31517.0 tgs: 264 data_time: 0.52s time: 119.12s eta: 13:57:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 268 data_time: 0.68s time: 119.17s eta: 13:56:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.126 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31591.0 tgs: 264 data_time: 0.73s time: 119.35s eta: 13:55:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.127 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31486.0 tgs: 262 data_time: 0.58s time: 120.12s eta: 13:58:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31404.0 tgs: 256 data_time: 0.99s time: 122.56s eta: 14:13:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.121 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30847.0 tgs: 258 data_time: 0.53s time: 119.11s eta: 13:47:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31370.0 tgs: 262 data_time: 0.64s time: 119.37s eta: 13:47:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.187 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32348.0 tgs: 271 data_time: 0.92s time: 119.33s eta: 13:45:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31441.0 tgs: 263 data_time: 0.75s time: 119.33s eta: 13:43:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32536.0 tgs: 273 data_time: 0.74s time: 119.12s eta: 13:39:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.211 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32140.0 tgs: 269 data_time: 0.85s time: 119.26s eta: 13:38:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.79s time: 119.13s eta: 13:36:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32511.0 tgs: 272 data_time: 0.66s time: 119.22s eta: 13:34:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32096.0 tgs: 268 data_time: 0.92s time: 119.46s eta: 13:34:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31630.0 tgs: 265 data_time: 0.77s time: 119.08s eta: 13:29:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.74s time: 118.92s eta: 13:26:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 0.55s time: 119.81s eta: 13:30:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 267 data_time: 0.82s time: 119.25s eta: 13:24:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30171.0 tgs: 253 data_time: 0.81s time: 119.16s eta: 13:22:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31004.0 tgs: 259 data_time: 0.83s time: 119.35s eta: 13:21:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31610.0 tgs: 265 data_time: 0.64s time: 119.26s eta: 13:19:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31674.0 tgs: 266 data_time: 0.72s time: 119.04s eta: 13:15:36 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.188 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32220.0 tgs: 269 data_time: 1.01s time: 119.35s eta: 13:15:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32555.0 tgs: 273 data_time: 0.72s time: 119.17s eta: 13:12:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32403.0 tgs: 271 data_time: 0.84s time: 119.46s eta: 13:12:24 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31960.0 tgs: 267 data_time: 0.91s time: 119.29s eta: 13:09:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30806.0 tgs: 258 data_time: 0.53s time: 119.15s eta: 13:06:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31620.0 tgs: 262 data_time: 0.55s time: 120.41s eta: 13:12:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32095.0 tgs: 267 data_time: 0.84s time: 120.16s eta: 13:09:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.115 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30004.0 tgs: 251 data_time: 0.73s time: 119.51s eta: 13:02:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31533.0 tgs: 263 data_time: 0.67s time: 119.51s eta: 13:00:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 265 data_time: 0.71s time: 119.57s eta: 12:59:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31699.0 tgs: 264 data_time: 0.54s time: 119.80s eta: 12:58:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.74s time: 119.04s eta: 12:51:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.197 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30943.0 tgs: 259 data_time: 0.54s time: 119.05s eta: 12:49:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 0.73s time: 119.12s eta: 12:48:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31768.0 tgs: 266 data_time: 0.59s time: 119.40s eta: 12:48:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31973.0 tgs: 268 data_time: 0.82s time: 119.18s eta: 12:44:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31047.0 tgs: 260 data_time: 0.77s time: 119.41s eta: 12:44:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.191 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.90s time: 119.10s eta: 12:40:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31914.0 tgs: 268 data_time: 0.86s time: 119.07s eta: 12:38:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31686.0 tgs: 264 data_time: 1.15s time: 119.60s eta: 12:39:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31041.0 tgs: 261 data_time: 0.84s time: 118.78s eta: 12:32:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31048.0 tgs: 261 data_time: 0.83s time: 118.89s eta: 12:30:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.202 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32065.0 tgs: 269 data_time: 0.61s time: 119.17s eta: 12:30:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 262 data_time: 0.73s time: 119.27s eta: 12:29:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 265 data_time: 0.74s time: 119.36s eta: 12:27:58 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.121 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32142.0 tgs: 269 data_time: 0.63s time: 119.11s eta: 12:24:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31871.0 tgs: 267 data_time: 0.76s time: 119.36s eta: 12:23:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31790.0 tgs: 266 data_time: 0.88s time: 119.26s eta: 12:21:24 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31075.0 tgs: 261 data_time: 0.68s time: 119.04s eta: 12:18:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.117 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 262 data_time: 0.55s time: 121.12s eta: 12:28:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.194 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30518.0 tgs: 254 data_time: 0.72s time: 119.98s eta: 12:19:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.198 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32133.0 tgs: 270 data_time: 0.67s time: 118.69s eta: 12:09:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.257 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30697.0 tgs: 256 data_time: 0.93s time: 119.63s eta: 12:13:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31613.0 tgs: 263 data_time: 0.83s time: 119.83s eta: 12:12:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.179 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31313.0 tgs: 262 data_time: 0.75s time: 119.14s eta: 12:06:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.217 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31978.0 tgs: 268 data_time: 1.01s time: 119.17s eta: 12:04:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.80s time: 119.12s eta: 12:02:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31458.0 tgs: 263 data_time: 0.72s time: 119.19s eta: 12:01:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.121 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31714.0 tgs: 266 data_time: 0.83s time: 118.96s eta: 11:57:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.131 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31425.0 tgs: 262 data_time: 0.51s time: 119.70s eta: 12:00:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32243.0 tgs: 270 data_time: 0.81s time: 119.28s eta: 11:55:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31927.0 tgs: 267 data_time: 0.64s time: 119.27s eta: 11:53:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31260.0 tgs: 261 data_time: 0.78s time: 119.71s eta: 11:54:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 271 data_time: 0.79s time: 118.78s eta: 11:46:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31547.0 tgs: 264 data_time: 0.56s time: 119.16s eta: 11:47:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31192.0 tgs: 260 data_time: 0.72s time: 119.61s eta: 11:47:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.189 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31341.0 tgs: 263 data_time: 0.96s time: 118.82s eta: 11:41:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.208 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 269 data_time: 0.52s time: 118.68s eta: 11:38:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30626.0 tgs: 256 data_time: 0.48s time: 119.22s eta: 11:39:24 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 266 data_time: 0.87s time: 119.14s eta: 11:36:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.204 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 268 data_time: 0.98s time: 119.08s eta: 11:34:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31960.0 tgs: 268 data_time: 0.82s time: 118.94s eta: 11:31:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31888.0 tgs: 268 data_time: 0.80s time: 118.91s eta: 11:29:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.185 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31588.0 tgs: 265 data_time: 0.78s time: 118.97s eta: 11:28:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.204 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.97s time: 119.29s eta: 11:27:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.223 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 268 data_time: 0.93s time: 119.35s eta: 11:26:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.115 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32187.0 tgs: 270 data_time: 0.89s time: 118.89s eta: 11:21:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31718.0 tgs: 266 data_time: 0.84s time: 119.20s eta: 11:21:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.115 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30398.0 tgs: 253 data_time: 0.57s time: 119.72s eta: 11:22:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.182 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30968.0 tgs: 259 data_time: 1.01s time: 119.12s eta: 11:16:58 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 268 data_time: 0.69s time: 119.13s eta: 11:15:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30670.0 tgs: 257 data_time: 0.64s time: 118.93s eta: 11:11:58 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31500.0 tgs: 264 data_time: 0.74s time: 119.10s eta: 11:10:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.199 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31599.0 tgs: 265 data_time: 0.69s time: 119.16s eta: 11:09:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31970.0 tgs: 268 data_time: 0.93s time: 118.98s eta: 11:06:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.232 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32253.0 tgs: 269 data_time: 0.93s time: 119.47s eta: 11:07:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31310.0 tgs: 261 data_time: 0.79s time: 119.65s eta: 11:06:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32190.0 tgs: 269 data_time: 0.73s time: 119.30s eta: 11:02:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31668.0 tgs: 265 data_time: 0.84s time: 119.31s eta: 11:00:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.191 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31349.0 tgs: 263 data_time: 1.04s time: 118.82s eta: 10:55:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 268 data_time: 0.87s time: 119.48s eta: 10:57:08 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.106 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 267 data_time: 0.55s time: 119.37s eta: 10:54:32 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.126 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31849.0 tgs: 267 data_time: 0.68s time: 119.28s eta: 10:52:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31869.0 tgs: 267 data_time: 0.74s time: 119.26s eta: 10:49:58 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31631.0 tgs: 265 data_time: 0.83s time: 119.10s eta: 10:47:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30525.0 tgs: 255 data_time: 0.58s time: 119.54s eta: 10:47:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.111 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.75s time: 119.31s eta: 10:44:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.203 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 267 data_time: 0.55s time: 119.49s eta: 10:43:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.112 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31464.0 tgs: 263 data_time: 0.75s time: 119.51s eta: 10:41:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31730.0 tgs: 266 data_time: 0.59s time: 118.99s eta: 10:36:36 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30533.0 tgs: 256 data_time: 0.72s time: 119.22s eta: 10:35:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31617.0 tgs: 265 data_time: 0.60s time: 119.06s eta: 10:33:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 272 data_time: 0.87s time: 118.78s eta: 10:29:32 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.180 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 263 data_time: 1.15s time: 119.86s eta: 10:33:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32281.0 tgs: 270 data_time: 0.76s time: 119.38s eta: 10:28:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31813.0 tgs: 267 data_time: 0.76s time: 118.84s eta: 10:23:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32442.0 tgs: 272 data_time: 0.77s time: 118.99s eta: 10:22:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31092.0 tgs: 260 data_time: 0.63s time: 119.28s eta: 10:22:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.119 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30707.0 tgs: 257 data_time: 1.02s time: 119.21s eta: 10:19:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31109.0 tgs: 261 data_time: 0.75s time: 119.06s eta: 10:17:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.265 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 267 data_time: 0.71s time: 119.48s eta: 10:17:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32353.0 tgs: 271 data_time: 0.81s time: 119.37s eta: 10:14:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.58s time: 119.45s eta: 10:13:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.182 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31262.0 tgs: 262 data_time: 0.75s time: 119.29s eta: 10:10:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 269 data_time: 0.83s time: 119.11s eta: 10:07:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30333.0 tgs: 253 data_time: 0.65s time: 119.70s eta: 10:08:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.74s time: 119.83s eta: 10:07:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 259 data_time: 1.01s time: 121.56s eta: 10:13:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31917.0 tgs: 267 data_time: 0.81s time: 119.25s eta: 10:00:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32009.0 tgs: 269 data_time: 0.83s time: 118.81s eta: 9:56:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.192 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30963.0 tgs: 258 data_time: 0.91s time: 119.57s eta: 9:57:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31840.0 tgs: 265 data_time: 0.88s time: 119.70s eta: 9:56:31 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.87s time: 119.26s eta: 9:52:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.199 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31173.0 tgs: 261 data_time: 0.82s time: 119.33s eta: 9:50:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 264 data_time: 0.76s time: 119.23s eta: 9:48:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 261 data_time: 0.73s time: 120.63s eta: 9:53:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.222 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32114.0 tgs: 269 data_time: 0.83s time: 119.22s eta: 9:44:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.193 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30399.0 tgs: 254 data_time: 0.64s time: 119.46s eta: 9:43:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.189 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32432.0 tgs: 271 data_time: 0.70s time: 119.55s eta: 9:41:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 270 data_time: 0.74s time: 119.20s eta: 9:38:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28686.0 tgs: 240 data_time: 0.70s time: 119.48s eta: 9:37:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31700.0 tgs: 266 data_time: 0.76s time: 119.05s eta: 9:33:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31095.0 tgs: 260 data_time: 0.89s time: 119.24s eta: 9:32:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31440.0 tgs: 263 data_time: 0.65s time: 119.16s eta: 9:30:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31957.0 tgs: 268 data_time: 0.69s time: 119.22s eta: 9:28:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.198 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31655.0 tgs: 264 data_time: 0.71s time: 119.46s eta: 9:27:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.185 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30538.0 tgs: 256 data_time: 0.89s time: 118.98s eta: 9:23:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31787.0 tgs: 267 data_time: 0.83s time: 119.04s eta: 9:21:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30794.0 tgs: 257 data_time: 0.79s time: 119.54s eta: 9:21:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31343.0 tgs: 262 data_time: 0.72s time: 119.59s eta: 9:20:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31575.0 tgs: 265 data_time: 0.84s time: 119.13s eta: 9:15:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.175 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32184.0 tgs: 270 data_time: 0.97s time: 119.02s eta: 9:13:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31968.0 tgs: 268 data_time: 0.84s time: 118.88s eta: 9:10:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30803.0 tgs: 258 data_time: 0.67s time: 119.32s eta: 9:10:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31664.0 tgs: 265 data_time: 1.00s time: 119.10s eta: 9:07:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31162.0 tgs: 262 data_time: 0.69s time: 118.86s eta: 9:04:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.181 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 265 data_time: 0.82s time: 119.67s eta: 9:06:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.188 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31125.0 tgs: 260 data_time: 0.58s time: 119.44s eta: 9:03:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 267 data_time: 0.67s time: 119.43s eta: 9:01:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.182 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 270 data_time: 0.87s time: 118.74s eta: 8:56:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.213 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 269 data_time: 0.87s time: 118.86s eta: 8:54:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30719.0 tgs: 257 data_time: 0.89s time: 119.34s eta: 8:55:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32401.0 tgs: 271 data_time: 0.75s time: 119.15s eta: 8:52:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.103 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32313.0 tgs: 270 data_time: 0.61s time: 119.27s eta: 8:50:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32434.0 tgs: 269 data_time: 0.75s time: 120.47s eta: 8:54:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31492.0 tgs: 264 data_time: 0.77s time: 119.24s eta: 8:46:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 266 data_time: 0.61s time: 119.04s eta: 8:43:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 261 data_time: 0.87s time: 120.83s eta: 8:49:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30313.0 tgs: 252 data_time: 0.72s time: 120.20s eta: 8:44:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 270 data_time: 0.78s time: 118.83s eta: 8:36:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32352.0 tgs: 271 data_time: 0.80s time: 119.35s eta: 8:37:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32118.0 tgs: 269 data_time: 0.77s time: 119.33s eta: 8:35:06 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.120 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31800.0 tgs: 267 data_time: 0.70s time: 119.05s eta: 8:31:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32278.0 tgs: 270 data_time: 0.65s time: 119.16s eta: 8:30:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32212.0 tgs: 270 data_time: 0.78s time: 119.05s eta: 8:27:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31620.0 tgs: 265 data_time: 0.88s time: 119.12s eta: 8:26:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.185 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31643.0 tgs: 265 data_time: 0.91s time: 119.25s eta: 8:24:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.280 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31306.0 tgs: 262 data_time: 0.97s time: 119.22s eta: 8:22:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.180 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 268 data_time: 1.14s time: 119.38s eta: 8:21:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31115.0 tgs: 259 data_time: 0.76s time: 119.78s eta: 8:21:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31672.0 tgs: 265 data_time: 0.86s time: 119.20s eta: 8:16:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31812.0 tgs: 267 data_time: 0.75s time: 118.98s eta: 8:13:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 257 data_time: 0.83s time: 119.90s eta: 8:15:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30483.0 tgs: 254 data_time: 0.61s time: 119.55s eta: 8:12:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30317.0 tgs: 253 data_time: 0.83s time: 119.59s eta: 8:10:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31943.0 tgs: 268 data_time: 0.63s time: 119.00s eta: 8:05:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31116.0 tgs: 259 data_time: 0.72s time: 119.85s eta: 8:07:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27629.0 tgs: 230 data_time: 0.74s time: 119.71s eta: 8:04:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.195 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32097.0 tgs: 268 data_time: 0.79s time: 119.40s eta: 8:01:34 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31895.0 tgs: 267 data_time: 0.55s time: 119.29s eta: 7:59:08 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31210.0 tgs: 261 data_time: 0.71s time: 119.19s eta: 7:56:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32432.0 tgs: 272 data_time: 0.70s time: 119.20s eta: 7:54:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32019.0 tgs: 269 data_time: 0.78s time: 118.94s eta: 7:51:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31680.0 tgs: 266 data_time: 0.83s time: 118.97s eta: 7:49:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 271 data_time: 1.21s time: 119.42s eta: 7:49:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.245 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31530.0 tgs: 264 data_time: 0.59s time: 119.04s eta: 7:46:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.231 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32008.0 tgs: 268 data_time: 0.72s time: 119.06s eta: 7:44:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 269 data_time: 0.67s time: 120.00s eta: 7:46:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30695.0 tgs: 256 data_time: 0.73s time: 119.84s eta: 7:43:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32301.0 tgs: 271 data_time: 0.54s time: 119.16s eta: 7:38:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31351.0 tgs: 263 data_time: 0.78s time: 118.97s eta: 7:36:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31466.0 tgs: 263 data_time: 0.61s time: 119.33s eta: 7:35:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.123 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32340.0 tgs: 270 data_time: 0.79s time: 119.38s eta: 7:33:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.256 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31327.0 tgs: 263 data_time: 0.98s time: 119.06s eta: 7:30:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31370.0 tgs: 263 data_time: 0.83s time: 118.99s eta: 7:28:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31790.0 tgs: 266 data_time: 0.71s time: 119.24s eta: 7:27:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32322.0 tgs: 271 data_time: 0.84s time: 118.99s eta: 7:24:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 268 data_time: 0.82s time: 118.93s eta: 7:22:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 270 data_time: 0.78s time: 119.08s eta: 7:20:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.219 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31253.0 tgs: 262 data_time: 0.96s time: 118.99s eta: 7:18:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31151.0 tgs: 261 data_time: 0.69s time: 119.03s eta: 7:16:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.229 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 268 data_time: 0.79s time: 119.25s eta: 7:15:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.116 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30135.0 tgs: 252 data_time: 0.69s time: 119.36s eta: 7:13:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 267 data_time: 0.77s time: 118.94s eta: 7:10:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.105 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31232.0 tgs: 262 data_time: 0.63s time: 119.21s eta: 7:09:08 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31162.0 tgs: 260 data_time: 0.76s time: 119.58s eta: 7:08:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.177 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 266 data_time: 0.83s time: 119.67s eta: 7:06:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.219 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31657.0 tgs: 266 data_time: 0.74s time: 118.95s eta: 7:02:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31308.0 tgs: 262 data_time: 0.68s time: 119.31s eta: 7:01:32 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:46:30][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 268 data_time: 0.72s time: 119.51s eta: 7:00:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31385.0 tgs: 262 data_time: 0.67s time: 119.43s eta: 6:58:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30623.0 tgs: 257 data_time: 0.79s time: 118.90s eta: 6:54:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31017.0 tgs: 259 data_time: 0.69s time: 119.35s eta: 6:53:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31151.0 tgs: 261 data_time: 0.85s time: 118.94s eta: 6:50:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32069.0 tgs: 268 data_time: 0.76s time: 119.27s eta: 6:49:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30882.0 tgs: 259 data_time: 0.62s time: 119.02s eta: 6:46:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31174.0 tgs: 261 data_time: 0.64s time: 119.22s eta: 6:45:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.121 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31824.0 tgs: 266 data_time: 0.59s time: 119.37s eta: 6:43:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 260 data_time: 0.71s time: 119.01s eta: 6:40:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.185 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32005.0 tgs: 268 data_time: 1.00s time: 119.01s eta: 6:38:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 269 data_time: 0.70s time: 119.45s eta: 6:38:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.176 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31684.0 tgs: 266 data_time: 0.99s time: 118.82s eta: 6:34:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31911.0 tgs: 267 data_time: 0.71s time: 119.43s eta: 6:34:06 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31685.0 tgs: 264 data_time: 0.83s time: 119.57s eta: 6:32:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.89s time: 119.38s eta: 6:29:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31698.0 tgs: 264 data_time: 0.46s time: 119.74s eta: 6:29:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.112 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30622.0 tgs: 256 data_time: 0.62s time: 119.36s eta: 6:25:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31439.0 tgs: 264 data_time: 0.77s time: 118.83s eta: 6:22:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.221 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 269 data_time: 0.88s time: 119.02s eta: 6:20:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31740.0 tgs: 265 data_time: 0.79s time: 119.40s eta: 6:20:06 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31709.0 tgs: 265 data_time: 0.71s time: 119.34s eta: 6:17:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31568.0 tgs: 265 data_time: 0.79s time: 119.04s eta: 6:14:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 267 data_time: 0.56s time: 119.38s eta: 6:14:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31134.0 tgs: 260 data_time: 0.67s time: 119.39s eta: 6:12:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.183 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31572.0 tgs: 264 data_time: 0.89s time: 119.29s eta: 6:09:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31923.0 tgs: 268 data_time: 0.65s time: 119.10s eta: 6:07:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30930.0 tgs: 253 data_time: 0.88s time: 121.96s eta: 6:14:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31469.0 tgs: 263 data_time: 0.81s time: 119.39s eta: 6:04:07 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.331 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 263 data_time: 0.72s time: 119.57s eta: 6:02:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 267 data_time: 0.81s time: 119.27s eta: 5:59:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30842.0 tgs: 258 data_time: 0.83s time: 119.27s eta: 5:57:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30791.0 tgs: 258 data_time: 0.86s time: 118.90s eta: 5:54:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32084.0 tgs: 269 data_time: 0.57s time: 119.02s eta: 5:53:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31337.0 tgs: 263 data_time: 0.69s time: 119.07s eta: 5:51:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30721.0 tgs: 257 data_time: 0.59s time: 119.30s eta: 5:49:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31601.0 tgs: 265 data_time: 0.57s time: 118.97s eta: 5:46:58 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31244.0 tgs: 261 data_time: 0.70s time: 119.27s eta: 5:45:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31356.0 tgs: 262 data_time: 0.59s time: 119.32s eta: 5:44:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31797.0 tgs: 267 data_time: 0.61s time: 119.05s eta: 5:41:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32197.0 tgs: 269 data_time: 0.79s time: 119.51s eta: 5:40:36 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.78s time: 119.37s eta: 5:38:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31323.0 tgs: 261 data_time: 0.68s time: 119.57s eta: 5:36:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 268 data_time: 0.68s time: 118.99s eta: 5:33:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.178 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31386.0 tgs: 263 data_time: 0.92s time: 118.94s eta: 5:31:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31859.0 tgs: 267 data_time: 0.77s time: 119.25s eta: 5:29:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.210 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31634.0 tgs: 264 data_time: 0.77s time: 119.40s eta: 5:28:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 264 data_time: 0.61s time: 119.46s eta: 5:26:31 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.191 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31642.0 tgs: 265 data_time: 0.63s time: 118.96s eta: 5:23:09 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31724.0 tgs: 265 data_time: 0.84s time: 119.63s eta: 5:23:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31170.0 tgs: 261 data_time: 0.69s time: 119.35s eta: 5:20:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.192 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31657.0 tgs: 265 data_time: 0.65s time: 119.12s eta: 5:17:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31409.0 tgs: 263 data_time: 0.64s time: 119.42s eta: 5:16:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.209 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31619.0 tgs: 264 data_time: 0.65s time: 119.57s eta: 5:14:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31695.0 tgs: 266 data_time: 0.66s time: 119.00s eta: 5:11:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31184.0 tgs: 262 data_time: 0.79s time: 119.00s eta: 5:09:23 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31733.0 tgs: 266 data_time: 0.95s time: 119.23s eta: 5:08:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 259 data_time: 0.66s time: 119.68s eta: 5:07:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.195 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.87s time: 119.11s eta: 5:03:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.107 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29736.0 tgs: 249 data_time: 0.50s time: 119.08s eta: 5:01:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30589.0 tgs: 251 data_time: 0.72s time: 121.52s eta: 5:05:48 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31327.0 tgs: 261 data_time: 0.58s time: 119.58s eta: 4:58:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31651.0 tgs: 264 data_time: 0.68s time: 119.50s eta: 4:56:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30981.0 tgs: 258 data_time: 0.88s time: 119.73s eta: 4:55:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 269 data_time: 0.61s time: 119.53s eta: 4:52:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31235.0 tgs: 261 data_time: 0.63s time: 119.42s eta: 4:50:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 266 data_time: 0.83s time: 119.20s eta: 4:48:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32147.0 tgs: 270 data_time: 0.67s time: 118.89s eta: 4:45:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31157.0 tgs: 261 data_time: 0.69s time: 119.04s eta: 4:43:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32556.0 tgs: 271 data_time: 0.75s time: 119.98s eta: 4:43:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 266 data_time: 0.77s time: 119.15s eta: 4:40:00 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30815.0 tgs: 258 data_time: 0.55s time: 119.00s eta: 4:37:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.221 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30824.0 tgs: 258 data_time: 0.74s time: 119.06s eta: 4:35:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31422.0 tgs: 263 data_time: 0.69s time: 119.38s eta: 4:34:33 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31278.0 tgs: 261 data_time: 0.60s time: 119.41s eta: 4:32:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32246.0 tgs: 271 data_time: 0.83s time: 118.94s eta: 4:29:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32500.0 tgs: 271 data_time: 0.82s time: 119.61s eta: 4:29:06 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31312.0 tgs: 262 data_time: 0.71s time: 119.24s eta: 4:26:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32090.0 tgs: 269 data_time: 0.97s time: 119.06s eta: 4:23:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32063.0 tgs: 269 data_time: 0.78s time: 118.85s eta: 4:21:28 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30082.0 tgs: 253 data_time: 0.85s time: 118.76s eta: 4:19:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31216.0 tgs: 261 data_time: 0.70s time: 119.43s eta: 4:18:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31636.0 tgs: 264 data_time: 0.51s time: 119.40s eta: 4:16:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31837.0 tgs: 266 data_time: 0.66s time: 119.32s eta: 4:14:33 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32423.0 tgs: 272 data_time: 0.68s time: 119.05s eta: 4:11:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32034.0 tgs: 268 data_time: 0.69s time: 119.13s eta: 4:10:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31699.0 tgs: 266 data_time: 0.88s time: 119.13s eta: 4:08:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31731.0 tgs: 266 data_time: 0.75s time: 118.96s eta: 4:05:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31173.0 tgs: 261 data_time: 0.70s time: 119.17s eta: 4:04:18 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30948.0 tgs: 259 data_time: 0.65s time: 119.48s eta: 4:02:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31499.0 tgs: 264 data_time: 0.83s time: 119.25s eta: 4:00:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31743.0 tgs: 266 data_time: 0.51s time: 119.19s eta: 3:58:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.187 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32189.0 tgs: 269 data_time: 0.59s time: 119.36s eta: 3:56:44 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31398.0 tgs: 263 data_time: 0.71s time: 119.37s eta: 3:54:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30454.0 tgs: 253 data_time: 0.53s time: 119.95s eta: 3:53:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31934.0 tgs: 267 data_time: 0.50s time: 119.32s eta: 3:50:41 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 262 data_time: 0.60s time: 119.64s eta: 3:49:18 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32449.0 tgs: 273 data_time: 0.67s time: 118.86s eta: 3:45:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32082.0 tgs: 268 data_time: 0.72s time: 119.70s eta: 3:45:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 268 data_time: 0.85s time: 119.33s eta: 3:42:45 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.097 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31845.0 tgs: 267 data_time: 0.43s time: 119.27s eta: 3:40:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31823.0 tgs: 267 data_time: 0.77s time: 119.00s eta: 3:38:10 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.262 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31240.0 tgs: 261 data_time: 0.72s time: 119.43s eta: 3:36:57 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31848.0 tgs: 267 data_time: 0.67s time: 119.03s eta: 3:34:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31764.0 tgs: 266 data_time: 0.79s time: 118.99s eta: 3:32:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.255 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31847.0 tgs: 267 data_time: 0.74s time: 119.22s eta: 3:30:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32068.0 tgs: 266 data_time: 0.56s time: 120.11s eta: 3:30:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.218 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 269 data_time: 0.88s time: 119.13s eta: 3:26:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.187 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 268 data_time: 0.63s time: 119.32s eta: 3:24:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31657.0 tgs: 265 data_time: 0.68s time: 119.08s eta: 3:22:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31659.0 tgs: 265 data_time: 0.98s time: 119.43s eta: 3:21:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31541.0 tgs: 264 data_time: 0.60s time: 119.29s eta: 3:18:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30513.0 tgs: 255 data_time: 0.74s time: 119.22s eta: 3:16:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31295.0 tgs: 262 data_time: 0.86s time: 119.13s eta: 3:14:34 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 267 data_time: 0.43s time: 119.02s eta: 3:12:24 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32171.0 tgs: 269 data_time: 0.77s time: 119.19s eta: 3:10:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32370.0 tgs: 267 data_time: 0.74s time: 120.97s eta: 3:11:32 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30920.0 tgs: 258 data_time: 0.64s time: 119.51s eta: 3:07:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32211.0 tgs: 270 data_time: 0.95s time: 119.27s eta: 3:04:51 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31530.0 tgs: 264 data_time: 0.50s time: 118.99s eta: 3:02:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 267 data_time: 0.76s time: 119.22s eta: 3:00:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 265 data_time: 0.69s time: 119.46s eta: 2:59:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32031.0 tgs: 269 data_time: 0.94s time: 118.98s eta: 2:56:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 265 data_time: 0.78s time: 119.35s eta: 2:55:03 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31643.0 tgs: 264 data_time: 0.84s time: 119.41s eta: 2:53:08 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31911.0 tgs: 267 data_time: 0.85s time: 119.35s eta: 2:51:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32067.0 tgs: 264 data_time: 0.82s time: 121.46s eta: 2:52:04 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 264 data_time: 0.66s time: 119.24s eta: 2:46:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31845.0 tgs: 266 data_time: 0.67s time: 119.54s eta: 2:45:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31181.0 tgs: 262 data_time: 0.70s time: 118.91s eta: 2:42:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31734.0 tgs: 266 data_time: 0.67s time: 119.18s eta: 2:40:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31980.0 tgs: 268 data_time: 0.55s time: 118.97s eta: 2:38:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 270 data_time: 0.78s time: 119.23s eta: 2:36:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31151.0 tgs: 261 data_time: 0.66s time: 119.11s eta: 2:34:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31354.0 tgs: 262 data_time: 0.76s time: 119.44s eta: 2:33:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30460.0 tgs: 253 data_time: 0.72s time: 120.23s eta: 2:32:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 262 data_time: 0.78s time: 119.28s eta: 2:29:05 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31902.0 tgs: 268 data_time: 0.77s time: 118.89s eta: 2:26:37 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 264 data_time: 0.58s time: 119.52s eta: 2:25:24 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31751.0 tgs: 266 data_time: 0.63s time: 119.07s eta: 2:22:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32194.0 tgs: 270 data_time: 0.65s time: 119.15s eta: 2:20:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31143.0 tgs: 261 data_time: 0.62s time: 119.04s eta: 2:18:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31451.0 tgs: 262 data_time: 1.11s time: 119.85s eta: 2:17:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32236.0 tgs: 270 data_time: 0.92s time: 119.31s eta: 2:15:13 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.099 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 265 data_time: 0.50s time: 119.38s eta: 2:13:18 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32271.0 tgs: 269 data_time: 0.67s time: 119.68s eta: 2:11:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.204 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31812.0 tgs: 266 data_time: 0.70s time: 119.50s eta: 2:09:27 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31345.0 tgs: 263 data_time: 0.61s time: 118.95s eta: 2:06:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.226 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32208.0 tgs: 269 data_time: 0.73s time: 119.38s eta: 2:05:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32034.0 tgs: 268 data_time: 0.74s time: 119.27s eta: 2:03:14 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31919.0 tgs: 267 data_time: 0.86s time: 119.22s eta: 2:01:12 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31755.0 tgs: 265 data_time: 0.61s time: 119.51s eta: 1:59:30 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 268 data_time: 0.87s time: 119.42s eta: 1:57:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30763.0 tgs: 257 data_time: 0.72s time: 119.53s eta: 1:55:33 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30975.0 tgs: 259 data_time: 0.72s time: 119.30s eta: 1:53:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31884.0 tgs: 267 data_time: 0.76s time: 119.26s eta: 1:51:18 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32717.0 tgs: 275 data_time: 0.75s time: 118.93s eta: 1:49:01 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31060.0 tgs: 259 data_time: 0.58s time: 119.64s eta: 1:47:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31628.0 tgs: 264 data_time: 1.06s time: 119.69s eta: 1:45:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31790.0 tgs: 266 data_time: 0.75s time: 119.34s eta: 1:43:25 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31093.0 tgs: 260 data_time: 0.64s time: 119.31s eta: 1:41:24 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31460.0 tgs: 263 data_time: 0.60s time: 119.29s eta: 1:39:24 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31548.0 tgs: 265 data_time: 0.70s time: 118.91s eta: 1:37:06 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31262.0 tgs: 262 data_time: 0.74s time: 119.17s eta: 1:35:20 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31884.0 tgs: 267 data_time: 0.68s time: 119.06s eta: 1:33:15 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32320.0 tgs: 270 data_time: 0.91s time: 119.27s eta: 1:31:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31061.0 tgs: 259 data_time: 0.74s time: 119.72s eta: 1:29:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31479.0 tgs: 264 data_time: 0.65s time: 119.01s eta: 1:27:16 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31756.0 tgs: 267 data_time: 0.55s time: 118.87s eta: 1:25:11 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.223 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31951.0 tgs: 267 data_time: 0.98s time: 119.48s eta: 1:23:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32022.0 tgs: 267 data_time: 0.84s time: 119.49s eta: 1:21:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32116.0 tgs: 270 data_time: 0.82s time: 118.94s eta: 1:19:17 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 268 data_time: 0.77s time: 118.94s eta: 1:17:18 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32443.0 tgs: 265 data_time: 0.66s time: 122.13s eta: 1:17:21 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31437.0 tgs: 263 data_time: 0.66s time: 119.34s eta: 1:13:35 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32125.0 tgs: 269 data_time: 0.76s time: 119.06s eta: 1:11:26 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31143.0 tgs: 261 data_time: 0.79s time: 118.92s eta: 1:09:22 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31354.0 tgs: 263 data_time: 0.73s time: 119.11s eta: 1:07:29 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32238.0 tgs: 269 data_time: 0.68s time: 119.41s eta: 1:05:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.69s time: 119.36s eta: 1:03:39 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31718.0 tgs: 266 data_time: 0.62s time: 119.17s eta: 1:01:34 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 268 data_time: 0.85s time: 119.28s eta: 0:59:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32284.0 tgs: 270 data_time: 0.75s time: 119.44s eta: 0:57:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 264 data_time: 0.89s time: 119.39s eta: 0:55:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 264 data_time: 0.68s time: 119.49s eta: 0:53:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31932.0 tgs: 267 data_time: 0.81s time: 119.37s eta: 0:51:43 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 268 data_time: 0.88s time: 119.47s eta: 0:49:46 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31612.0 tgs: 265 data_time: 0.83s time: 118.90s eta: 0:47:33 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31583.0 tgs: 264 data_time: 0.56s time: 119.26s eta: 0:45:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31801.0 tgs: 266 data_time: 0.74s time: 119.19s eta: 0:43:42 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32207.0 tgs: 270 data_time: 0.46s time: 118.99s eta: 0:41:38 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31280.0 tgs: 262 data_time: 0.94s time: 119.04s eta: 0:39:40 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 265 data_time: 0.69s time: 119.66s eta: 0:37:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 266 data_time: 0.60s time: 119.29s eta: 0:35:47 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31941.0 tgs: 266 data_time: 0.79s time: 119.75s eta: 0:33:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 263 data_time: 0.80s time: 121.23s eta: 0:32:19 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29970.0 tgs: 251 data_time: 0.39s time: 119.38s eta: 0:29:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31575.0 tgs: 264 data_time: 0.75s time: 119.32s eta: 0:27:50 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30964.0 tgs: 258 data_time: 0.56s time: 119.64s eta: 0:25:55 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30870.0 tgs: 259 data_time: 0.57s time: 119.10s eta: 0:23:49 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31970.0 tgs: 267 data_time: 0.64s time: 119.39s eta: 0:21:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31745.0 tgs: 266 data_time: 0.76s time: 119.26s eta: 0:19:52 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 268 data_time: 0.64s time: 119.24s eta: 0:17:53 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 265 data_time: 0.56s time: 119.90s eta: 0:15:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31585.0 tgs: 264 data_time: 0.66s time: 119.19s eta: 0:13:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31779.0 tgs: 266 data_time: 0.77s time: 119.09s eta: 0:11:54 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.69s time: 119.32s eta: 0:09:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.196 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31664.0 tgs: 265 data_time: 0.72s time: 119.17s eta: 0:07:56 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30712.0 tgs: 256 data_time: 0.64s time: 119.69s eta: 0:05:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31311.0 tgs: 258 data_time: 0.72s time: 121.28s eta: 0:04:02 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31199.0 tgs: 261 data_time: 0.70s time: 119.21s eta: 0:01:59 +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.29 seconds, peak gpu memory 17.4G +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 36][DP 9][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank37.log b/20250121165312/rank37.log new file mode 100644 index 0000000000000000000000000000000000000000..296df20c9bababd97ec2f43a08536738ad18a791 --- /dev/null +++ b/20250121165312/rank37.log @@ -0,0 +1,851 @@ +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:35][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.08s +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.83 seconds, peak gpu memory 13.4G +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.297 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 32281.0 tgs: 212 data_time: 2.37s time: 152.17s eta: 1 day, 1:03:57 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.255 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32206.0 tgs: 268 data_time: 0.77s time: 120.15s eta: 19:45:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.244 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32213.0 tgs: 269 data_time: 0.92s time: 119.68s eta: 19:38:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29331.0 tgs: 245 data_time: 0.83s time: 119.66s eta: 19:36:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.210 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31870.0 tgs: 267 data_time: 0.87s time: 119.30s eta: 19:31:09 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.201 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29472.0 tgs: 246 data_time: 0.68s time: 119.54s eta: 19:31:27 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 261 data_time: 0.71s time: 122.97s eta: 20:03:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.275 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30854.0 tgs: 257 data_time: 0.65s time: 119.90s eta: 19:31:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.234 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31900.0 tgs: 264 data_time: 0.73s time: 120.65s eta: 19:36:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.212 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32109.0 tgs: 267 data_time: 0.76s time: 119.92s eta: 19:27:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31161.0 tgs: 258 data_time: 0.58s time: 120.46s eta: 19:30:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31536.0 tgs: 264 data_time: 0.79s time: 119.35s eta: 19:17:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.236 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 267 data_time: 0.76s time: 119.68s eta: 19:18:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32359.0 tgs: 271 data_time: 0.75s time: 119.18s eta: 19:12:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32408.0 tgs: 271 data_time: 0.94s time: 119.19s eta: 19:10:10 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31413.0 tgs: 264 data_time: 0.96s time: 118.93s eta: 19:05:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.213 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31857.0 tgs: 267 data_time: 0.78s time: 119.14s eta: 19:05:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31570.0 tgs: 264 data_time: 0.73s time: 119.18s eta: 19:04:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31566.0 tgs: 264 data_time: 0.65s time: 119.40s eta: 19:04:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32203.0 tgs: 269 data_time: 0.79s time: 119.33s eta: 19:01:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30573.0 tgs: 254 data_time: 0.65s time: 120.31s eta: 19:09:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30164.0 tgs: 245 data_time: 0.68s time: 122.91s eta: 19:31:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31721.0 tgs: 265 data_time: 0.76s time: 119.48s eta: 18:57:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31235.0 tgs: 262 data_time: 0.76s time: 119.16s eta: 18:51:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 269 data_time: 0.69s time: 119.25s eta: 18:50:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.326 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31111.0 tgs: 260 data_time: 0.88s time: 119.39s eta: 18:50:14 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31722.0 tgs: 265 data_time: 0.82s time: 119.67s eta: 18:50:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30903.0 tgs: 258 data_time: 0.78s time: 119.39s eta: 18:46:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31566.0 tgs: 263 data_time: 1.02s time: 119.77s eta: 18:47:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 28740.0 tgs: 222 data_time: 0.79s time: 129.31s eta: 20:15:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.256 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31943.0 tgs: 267 data_time: 0.61s time: 119.57s eta: 18:41:57 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31226.0 tgs: 261 data_time: 0.82s time: 119.39s eta: 18:38:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31930.0 tgs: 264 data_time: 0.64s time: 120.66s eta: 18:48:09 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31763.0 tgs: 266 data_time: 0.91s time: 119.40s eta: 18:34:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32155.0 tgs: 269 data_time: 0.81s time: 119.16s eta: 18:30:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.233 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31819.0 tgs: 267 data_time: 0.87s time: 119.07s eta: 18:27:22 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32555.0 tgs: 273 data_time: 0.62s time: 119.14s eta: 18:25:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31507.0 tgs: 261 data_time: 0.74s time: 120.39s eta: 18:35:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32485.0 tgs: 268 data_time: 0.86s time: 121.18s eta: 18:40:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32286.0 tgs: 267 data_time: 0.89s time: 120.66s eta: 18:34:05 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32270.0 tgs: 269 data_time: 0.77s time: 119.73s eta: 18:23:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.515 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 257 data_time: 0.93s time: 123.88s eta: 18:59:41 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31139.0 tgs: 260 data_time: 0.73s time: 119.55s eta: 18:17:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 266 data_time: 0.66s time: 119.33s eta: 18:13:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32014.0 tgs: 268 data_time: 0.86s time: 119.22s eta: 18:10:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31950.0 tgs: 267 data_time: 0.98s time: 119.60s eta: 18:12:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31159.0 tgs: 261 data_time: 0.60s time: 119.27s eta: 18:07:22 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31876.0 tgs: 267 data_time: 0.92s time: 119.01s eta: 18:02:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31681.0 tgs: 266 data_time: 0.87s time: 119.10s eta: 18:01:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31594.0 tgs: 264 data_time: 1.18s time: 119.26s eta: 18:01:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31032.0 tgs: 260 data_time: 1.03s time: 119.34s eta: 17:59:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31014.0 tgs: 258 data_time: 0.93s time: 120.02s eta: 18:04:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32168.0 tgs: 270 data_time: 0.94s time: 119.10s eta: 17:53:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.72s time: 119.42s eta: 17:54:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 272 data_time: 0.91s time: 119.00s eta: 17:48:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31439.0 tgs: 264 data_time: 0.66s time: 119.02s eta: 17:47:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31610.0 tgs: 252 data_time: 0.57s time: 125.37s eta: 18:42:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31493.0 tgs: 262 data_time: 0.86s time: 120.10s eta: 17:52:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31578.0 tgs: 264 data_time: 0.79s time: 119.37s eta: 17:44:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32478.0 tgs: 272 data_time: 1.01s time: 119.32s eta: 17:41:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29943.0 tgs: 250 data_time: 0.82s time: 119.51s eta: 17:41:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31038.0 tgs: 246 data_time: 0.55s time: 126.05s eta: 18:37:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31043.0 tgs: 253 data_time: 0.80s time: 122.63s eta: 18:05:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32383.0 tgs: 267 data_time: 0.85s time: 121.16s eta: 17:50:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31991.0 tgs: 268 data_time: 0.63s time: 119.04s eta: 17:29:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30412.0 tgs: 255 data_time: 0.84s time: 118.86s eta: 17:25:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31377.0 tgs: 253 data_time: 1.01s time: 124.01s eta: 18:09:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 254 data_time: 0.74s time: 124.56s eta: 18:11:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32481.0 tgs: 272 data_time: 0.95s time: 119.39s eta: 17:24:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31743.0 tgs: 265 data_time: 0.98s time: 119.70s eta: 17:25:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32084.0 tgs: 268 data_time: 0.86s time: 119.53s eta: 17:21:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32042.0 tgs: 269 data_time: 0.78s time: 118.86s eta: 17:14:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 0.99s time: 119.85s eta: 17:20:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29019.0 tgs: 242 data_time: 0.78s time: 119.52s eta: 17:15:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.282 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31875.0 tgs: 261 data_time: 0.76s time: 121.86s eta: 17:34:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32461.0 tgs: 262 data_time: 0.59s time: 123.52s eta: 17:46:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.239 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31810.0 tgs: 267 data_time: 0.67s time: 119.02s eta: 17:05:31 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 267 data_time: 0.86s time: 119.23s eta: 17:05:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31702.0 tgs: 266 data_time: 0.76s time: 119.15s eta: 17:02:41 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 28580.0 tgs: 239 data_time: 0.83s time: 119.25s eta: 17:01:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31749.0 tgs: 266 data_time: 0.77s time: 119.04s eta: 16:57:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31455.0 tgs: 263 data_time: 0.82s time: 119.28s eta: 16:57:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30811.0 tgs: 258 data_time: 0.72s time: 119.16s eta: 16:54:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31311.0 tgs: 261 data_time: 0.64s time: 119.59s eta: 16:56:30 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32026.0 tgs: 268 data_time: 0.72s time: 119.13s eta: 16:50:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31318.0 tgs: 263 data_time: 0.63s time: 118.84s eta: 16:46:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31265.0 tgs: 262 data_time: 0.90s time: 119.02s eta: 16:45:41 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31698.0 tgs: 264 data_time: 0.71s time: 119.85s eta: 16:50:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32116.0 tgs: 269 data_time: 0.71s time: 119.36s eta: 16:44:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31820.0 tgs: 267 data_time: 0.55s time: 119.01s eta: 16:39:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32398.0 tgs: 267 data_time: 0.73s time: 120.93s eta: 16:53:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32077.0 tgs: 269 data_time: 0.78s time: 118.99s eta: 16:35:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 268 data_time: 0.82s time: 119.29s eta: 16:36:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 265 data_time: 0.61s time: 119.23s eta: 16:33:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31125.0 tgs: 260 data_time: 0.71s time: 119.37s eta: 16:32:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31589.0 tgs: 263 data_time: 0.85s time: 119.70s eta: 16:33:30 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.121 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 264 data_time: 0.68s time: 119.21s eta: 16:27:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.206 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30913.0 tgs: 258 data_time: 0.71s time: 119.76s eta: 16:30:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 260 data_time: 0.85s time: 120.33s eta: 16:32:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32113.0 tgs: 269 data_time: 1.05s time: 118.98s eta: 16:19:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31989.0 tgs: 251 data_time: 0.96s time: 127.33s eta: 17:26:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31887.0 tgs: 267 data_time: 0.72s time: 119.22s eta: 16:17:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 267 data_time: 0.54s time: 119.14s eta: 16:14:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.208 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31366.0 tgs: 263 data_time: 0.69s time: 119.20s eta: 16:13:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31395.0 tgs: 263 data_time: 0.75s time: 119.17s eta: 16:11:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32048.0 tgs: 269 data_time: 1.03s time: 118.95s eta: 16:07:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.247 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31406.0 tgs: 264 data_time: 0.68s time: 118.86s eta: 16:04:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31434.0 tgs: 264 data_time: 0.52s time: 118.96s eta: 16:03:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 258 data_time: 0.73s time: 121.84s eta: 16:24:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31767.0 tgs: 266 data_time: 0.71s time: 119.00s eta: 15:59:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31044.0 tgs: 260 data_time: 0.83s time: 119.22s eta: 15:59:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 267 data_time: 0.76s time: 119.61s eta: 16:00:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 271 data_time: 0.69s time: 119.05s eta: 15:54:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32439.0 tgs: 272 data_time: 0.64s time: 118.98s eta: 15:51:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.71s time: 119.18s eta: 15:51:25 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31837.0 tgs: 266 data_time: 0.54s time: 119.64s eta: 15:53:06 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 268 data_time: 0.73s time: 118.94s eta: 15:45:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32245.0 tgs: 270 data_time: 0.94s time: 119.29s eta: 15:46:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31475.0 tgs: 264 data_time: 0.97s time: 119.10s eta: 15:42:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31148.0 tgs: 261 data_time: 0.76s time: 118.99s eta: 15:40:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31417.0 tgs: 262 data_time: 0.81s time: 119.66s eta: 15:43:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 259 data_time: 0.74s time: 121.50s eta: 15:55:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31756.0 tgs: 265 data_time: 0.76s time: 119.41s eta: 15:37:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 247 data_time: 0.95s time: 129.97s eta: 16:58:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.233 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31805.0 tgs: 266 data_time: 1.02s time: 119.47s eta: 15:33:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31292.0 tgs: 261 data_time: 0.83s time: 119.60s eta: 15:32:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.116 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30814.0 tgs: 257 data_time: 0.83s time: 119.53s eta: 15:30:22 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31413.0 tgs: 257 data_time: 0.68s time: 122.14s eta: 15:48:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 265 data_time: 0.93s time: 120.76s eta: 15:35:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.226 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 265 data_time: 0.81s time: 119.41s eta: 15:23:27 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.68s time: 119.45s eta: 15:21:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31520.0 tgs: 239 data_time: 0.84s time: 131.64s eta: 16:53:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31932.0 tgs: 267 data_time: 0.60s time: 119.30s eta: 15:16:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30315.0 tgs: 253 data_time: 0.68s time: 119.52s eta: 15:16:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31784.0 tgs: 259 data_time: 0.78s time: 122.25s eta: 15:35:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 26939.0 tgs: 225 data_time: 0.42s time: 119.71s eta: 15:13:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31378.0 tgs: 262 data_time: 0.94s time: 119.35s eta: 15:09:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31717.0 tgs: 266 data_time: 0.80s time: 118.90s eta: 15:03:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.121 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 270 data_time: 0.82s time: 119.00s eta: 15:02:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32422.0 tgs: 271 data_time: 0.86s time: 119.30s eta: 15:02:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30691.0 tgs: 258 data_time: 0.73s time: 118.93s eta: 14:57:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.204 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31614.0 tgs: 262 data_time: 0.82s time: 120.52s eta: 15:07:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.118 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31621.0 tgs: 265 data_time: 0.54s time: 119.31s eta: 14:56:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29421.0 tgs: 239 data_time: 0.73s time: 122.66s eta: 15:19:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.251 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32141.0 tgs: 269 data_time: 0.76s time: 119.13s eta: 14:51:31 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31387.0 tgs: 262 data_time: 0.73s time: 119.75s eta: 14:54:10 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 263 data_time: 0.84s time: 119.47s eta: 14:50:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 260 data_time: 0.82s time: 118.99s eta: 14:44:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.114 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30723.0 tgs: 257 data_time: 0.52s time: 119.30s eta: 14:44:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31767.0 tgs: 267 data_time: 0.84s time: 118.95s eta: 14:40:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.229 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30699.0 tgs: 257 data_time: 0.85s time: 119.05s eta: 14:39:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.64s time: 119.06s eta: 14:37:06 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32007.0 tgs: 268 data_time: 0.53s time: 119.13s eta: 14:35:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32346.0 tgs: 271 data_time: 0.97s time: 119.01s eta: 14:32:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.174 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 269 data_time: 0.79s time: 119.07s eta: 14:31:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.127 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31216.0 tgs: 261 data_time: 0.79s time: 119.15s eta: 14:29:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 261 data_time: 0.93s time: 119.65s eta: 14:31:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32106.0 tgs: 269 data_time: 0.75s time: 119.17s eta: 14:25:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31674.0 tgs: 265 data_time: 0.90s time: 119.20s eta: 14:24:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31484.0 tgs: 263 data_time: 0.85s time: 119.39s eta: 14:23:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31704.0 tgs: 262 data_time: 0.60s time: 120.74s eta: 14:31:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31765.0 tgs: 267 data_time: 0.83s time: 118.85s eta: 14:15:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.247 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31034.0 tgs: 260 data_time: 0.89s time: 119.10s eta: 14:15:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32110.0 tgs: 269 data_time: 1.07s time: 119.11s eta: 14:13:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.218 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 261 data_time: 0.94s time: 119.95s eta: 14:17:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28445.0 tgs: 237 data_time: 0.62s time: 119.87s eta: 14:15:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 267 data_time: 0.74s time: 119.59s eta: 14:11:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31811.0 tgs: 266 data_time: 0.79s time: 119.22s eta: 14:06:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.122 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31451.0 tgs: 263 data_time: 0.71s time: 119.27s eta: 14:04:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31249.0 tgs: 262 data_time: 0.61s time: 119.27s eta: 14:02:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32506.0 tgs: 272 data_time: 0.80s time: 119.15s eta: 13:59:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31517.0 tgs: 264 data_time: 0.52s time: 119.12s eta: 13:57:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 268 data_time: 0.67s time: 119.17s eta: 13:56:09 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.201 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31591.0 tgs: 264 data_time: 0.73s time: 119.35s eta: 13:55:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31486.0 tgs: 262 data_time: 0.59s time: 120.12s eta: 13:58:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31404.0 tgs: 256 data_time: 0.99s time: 122.55s eta: 14:13:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.194 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30847.0 tgs: 258 data_time: 0.53s time: 119.11s eta: 13:47:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31370.0 tgs: 262 data_time: 0.64s time: 119.37s eta: 13:47:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.207 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32348.0 tgs: 271 data_time: 0.92s time: 119.32s eta: 13:45:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31441.0 tgs: 263 data_time: 0.75s time: 119.33s eta: 13:43:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32536.0 tgs: 273 data_time: 0.75s time: 119.12s eta: 13:39:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32140.0 tgs: 269 data_time: 0.85s time: 119.26s eta: 13:38:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.185 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.77s time: 119.14s eta: 13:36:05 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.286 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32511.0 tgs: 272 data_time: 0.64s time: 119.23s eta: 13:34:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.116 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32096.0 tgs: 268 data_time: 0.86s time: 119.46s eta: 13:34:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31630.0 tgs: 265 data_time: 0.75s time: 119.08s eta: 13:29:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.74s time: 118.92s eta: 13:26:41 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.128 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 0.55s time: 119.81s eta: 13:30:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 267 data_time: 0.81s time: 119.26s eta: 13:25:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30171.0 tgs: 253 data_time: 0.80s time: 119.16s eta: 13:22:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31004.0 tgs: 259 data_time: 0.83s time: 119.35s eta: 13:21:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.253 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31610.0 tgs: 265 data_time: 0.64s time: 119.26s eta: 13:19:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31674.0 tgs: 266 data_time: 0.71s time: 119.04s eta: 13:15:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32220.0 tgs: 269 data_time: 0.99s time: 119.35s eta: 13:15:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32555.0 tgs: 273 data_time: 0.72s time: 119.17s eta: 13:12:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32403.0 tgs: 271 data_time: 0.84s time: 119.46s eta: 13:12:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31960.0 tgs: 267 data_time: 0.92s time: 119.29s eta: 13:09:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.122 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30806.0 tgs: 258 data_time: 0.53s time: 119.15s eta: 13:06:22 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31620.0 tgs: 262 data_time: 0.56s time: 120.42s eta: 13:12:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.109 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32095.0 tgs: 267 data_time: 0.82s time: 120.16s eta: 13:09:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30004.0 tgs: 251 data_time: 0.73s time: 119.50s eta: 13:02:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.109 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31533.0 tgs: 263 data_time: 0.67s time: 119.51s eta: 13:00:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 265 data_time: 0.70s time: 119.58s eta: 12:59:14 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31699.0 tgs: 264 data_time: 0.56s time: 119.80s eta: 12:58:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.128 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.69s time: 119.04s eta: 12:51:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30943.0 tgs: 259 data_time: 0.54s time: 119.05s eta: 12:49:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.188 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 0.73s time: 119.13s eta: 12:48:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31768.0 tgs: 266 data_time: 0.59s time: 119.40s eta: 12:48:07 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31973.0 tgs: 268 data_time: 0.84s time: 119.18s eta: 12:44:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31047.0 tgs: 260 data_time: 0.75s time: 119.41s eta: 12:44:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.90s time: 119.10s eta: 12:40:14 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31914.0 tgs: 268 data_time: 0.86s time: 119.07s eta: 12:38:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.232 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31686.0 tgs: 264 data_time: 1.14s time: 119.61s eta: 12:39:30 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31041.0 tgs: 261 data_time: 0.78s time: 118.77s eta: 12:32:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31048.0 tgs: 261 data_time: 0.84s time: 118.89s eta: 12:30:57 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32065.0 tgs: 269 data_time: 0.61s time: 119.17s eta: 12:30:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 262 data_time: 0.71s time: 119.27s eta: 12:29:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.193 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 265 data_time: 0.73s time: 119.35s eta: 12:27:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32142.0 tgs: 269 data_time: 0.64s time: 119.12s eta: 12:24:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31871.0 tgs: 267 data_time: 0.76s time: 119.35s eta: 12:23:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31790.0 tgs: 266 data_time: 0.88s time: 119.26s eta: 12:21:25 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31075.0 tgs: 261 data_time: 0.68s time: 119.03s eta: 12:17:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 262 data_time: 0.52s time: 121.13s eta: 12:29:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30518.0 tgs: 254 data_time: 0.71s time: 119.98s eta: 12:19:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32133.0 tgs: 270 data_time: 0.67s time: 118.69s eta: 12:09:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30697.0 tgs: 256 data_time: 0.93s time: 119.63s eta: 12:13:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31613.0 tgs: 263 data_time: 0.83s time: 119.83s eta: 12:12:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31313.0 tgs: 262 data_time: 0.73s time: 119.14s eta: 12:06:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.206 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31978.0 tgs: 268 data_time: 1.01s time: 119.17s eta: 12:04:57 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.77s time: 119.12s eta: 12:02:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31458.0 tgs: 263 data_time: 0.72s time: 119.19s eta: 12:01:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31714.0 tgs: 266 data_time: 0.82s time: 118.96s eta: 11:57:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.131 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31425.0 tgs: 262 data_time: 0.51s time: 119.70s eta: 12:00:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32243.0 tgs: 270 data_time: 0.80s time: 119.28s eta: 11:55:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31927.0 tgs: 267 data_time: 0.58s time: 119.27s eta: 11:53:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.118 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31260.0 tgs: 261 data_time: 0.75s time: 119.71s eta: 11:54:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.113 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 271 data_time: 0.77s time: 118.79s eta: 11:46:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.179 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31547.0 tgs: 264 data_time: 0.54s time: 119.16s eta: 11:47:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31192.0 tgs: 260 data_time: 0.70s time: 119.61s eta: 11:47:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31341.0 tgs: 263 data_time: 0.95s time: 118.82s eta: 11:41:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.224 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 269 data_time: 0.51s time: 118.69s eta: 11:38:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.130 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30626.0 tgs: 256 data_time: 0.48s time: 119.22s eta: 11:39:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.119 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 266 data_time: 0.87s time: 119.14s eta: 11:36:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.259 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 268 data_time: 0.98s time: 119.07s eta: 11:34:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31960.0 tgs: 268 data_time: 0.82s time: 118.95s eta: 11:31:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31888.0 tgs: 268 data_time: 0.80s time: 118.92s eta: 11:29:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31588.0 tgs: 265 data_time: 0.77s time: 118.97s eta: 11:28:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.120 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.93s time: 119.29s eta: 11:27:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 268 data_time: 0.92s time: 119.36s eta: 11:26:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.195 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32187.0 tgs: 270 data_time: 0.88s time: 118.88s eta: 11:21:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31718.0 tgs: 266 data_time: 0.82s time: 119.21s eta: 11:21:27 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30398.0 tgs: 253 data_time: 0.57s time: 119.72s eta: 11:22:22 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.271 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30968.0 tgs: 259 data_time: 1.02s time: 119.12s eta: 11:16:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 268 data_time: 0.69s time: 119.13s eta: 11:15:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30670.0 tgs: 257 data_time: 0.64s time: 118.93s eta: 11:11:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31500.0 tgs: 264 data_time: 0.76s time: 119.09s eta: 11:10:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.200 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31599.0 tgs: 265 data_time: 0.60s time: 119.16s eta: 11:09:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.222 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31970.0 tgs: 268 data_time: 0.97s time: 118.97s eta: 11:06:14 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32253.0 tgs: 269 data_time: 0.90s time: 119.47s eta: 11:07:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.115 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31310.0 tgs: 261 data_time: 0.79s time: 119.64s eta: 11:06:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32190.0 tgs: 269 data_time: 0.80s time: 119.30s eta: 11:02:07 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31668.0 tgs: 265 data_time: 0.85s time: 119.31s eta: 11:00:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.170 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31349.0 tgs: 263 data_time: 1.03s time: 118.82s eta: 10:55:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 268 data_time: 0.87s time: 119.48s eta: 10:57:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.099 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 267 data_time: 0.55s time: 119.36s eta: 10:54:30 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31849.0 tgs: 266 data_time: 0.68s time: 119.29s eta: 10:52:06 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31869.0 tgs: 267 data_time: 0.75s time: 119.25s eta: 10:49:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31631.0 tgs: 265 data_time: 0.84s time: 119.10s eta: 10:47:05 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30525.0 tgs: 255 data_time: 0.57s time: 119.54s eta: 10:47:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.77s time: 119.30s eta: 10:44:14 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 267 data_time: 0.55s time: 119.49s eta: 10:43:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31464.0 tgs: 263 data_time: 0.73s time: 119.52s eta: 10:41:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31730.0 tgs: 266 data_time: 0.58s time: 118.99s eta: 10:36:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30533.0 tgs: 256 data_time: 0.76s time: 119.22s eta: 10:35:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.113 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31617.0 tgs: 265 data_time: 0.60s time: 119.06s eta: 10:33:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 272 data_time: 0.88s time: 118.78s eta: 10:29:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 263 data_time: 1.17s time: 119.86s eta: 10:33:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32281.0 tgs: 270 data_time: 0.77s time: 119.38s eta: 10:28:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31813.0 tgs: 267 data_time: 0.77s time: 118.84s eta: 10:23:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32442.0 tgs: 272 data_time: 0.76s time: 118.99s eta: 10:22:41 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31092.0 tgs: 260 data_time: 0.64s time: 119.29s eta: 10:22:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30707.0 tgs: 257 data_time: 1.02s time: 119.20s eta: 10:19:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31109.0 tgs: 261 data_time: 0.74s time: 119.06s eta: 10:17:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.124 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 267 data_time: 0.71s time: 119.48s eta: 10:17:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.205 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32353.0 tgs: 271 data_time: 0.84s time: 119.37s eta: 10:14:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.218 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.58s time: 119.45s eta: 10:13:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31262.0 tgs: 262 data_time: 0.74s time: 119.29s eta: 10:10:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 269 data_time: 0.83s time: 119.11s eta: 10:07:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30333.0 tgs: 253 data_time: 0.63s time: 119.70s eta: 10:08:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.74s time: 119.83s eta: 10:07:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 259 data_time: 1.02s time: 121.55s eta: 10:13:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.191 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31917.0 tgs: 267 data_time: 0.81s time: 119.26s eta: 10:00:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.112 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32009.0 tgs: 269 data_time: 0.83s time: 118.81s eta: 9:56:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.227 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30963.0 tgs: 258 data_time: 0.91s time: 119.57s eta: 9:57:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31840.0 tgs: 265 data_time: 0.88s time: 119.71s eta: 9:56:32 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.99s time: 119.26s eta: 9:52:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31173.0 tgs: 261 data_time: 0.80s time: 119.34s eta: 9:50:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 264 data_time: 0.77s time: 119.22s eta: 9:48:09 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 260 data_time: 0.73s time: 120.86s eta: 9:54:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.225 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32114.0 tgs: 269 data_time: 0.78s time: 119.23s eta: 9:44:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30399.0 tgs: 254 data_time: 0.65s time: 119.45s eta: 9:43:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32432.0 tgs: 271 data_time: 0.72s time: 119.55s eta: 9:41:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 270 data_time: 0.74s time: 119.20s eta: 9:38:07 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28686.0 tgs: 240 data_time: 0.69s time: 119.48s eta: 9:37:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31700.0 tgs: 266 data_time: 0.76s time: 119.05s eta: 9:33:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31095.0 tgs: 260 data_time: 0.91s time: 119.24s eta: 9:32:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31440.0 tgs: 263 data_time: 0.65s time: 119.16s eta: 9:30:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31957.0 tgs: 268 data_time: 0.70s time: 119.22s eta: 9:28:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31655.0 tgs: 264 data_time: 0.71s time: 119.46s eta: 9:27:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.191 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30538.0 tgs: 256 data_time: 0.89s time: 118.99s eta: 9:23:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31787.0 tgs: 267 data_time: 0.83s time: 119.04s eta: 9:21:27 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.233 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30794.0 tgs: 257 data_time: 0.81s time: 119.53s eta: 9:21:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.237 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31343.0 tgs: 262 data_time: 0.72s time: 119.60s eta: 9:20:06 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31575.0 tgs: 265 data_time: 0.81s time: 119.12s eta: 9:15:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32184.0 tgs: 270 data_time: 0.96s time: 119.02s eta: 9:13:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31968.0 tgs: 268 data_time: 0.82s time: 118.89s eta: 9:10:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30803.0 tgs: 258 data_time: 0.67s time: 119.33s eta: 9:10:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31664.0 tgs: 265 data_time: 1.01s time: 119.09s eta: 9:07:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31162.0 tgs: 262 data_time: 0.70s time: 118.86s eta: 9:04:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 265 data_time: 0.82s time: 119.67s eta: 9:06:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31125.0 tgs: 260 data_time: 0.60s time: 119.44s eta: 9:03:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 267 data_time: 0.67s time: 119.43s eta: 9:01:25 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 270 data_time: 0.86s time: 118.74s eta: 8:56:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 269 data_time: 0.88s time: 118.86s eta: 8:54:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30719.0 tgs: 257 data_time: 0.90s time: 119.34s eta: 8:55:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32401.0 tgs: 271 data_time: 0.78s time: 119.15s eta: 8:52:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32313.0 tgs: 270 data_time: 0.63s time: 119.28s eta: 8:50:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32434.0 tgs: 269 data_time: 0.77s time: 120.47s eta: 8:54:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31492.0 tgs: 264 data_time: 0.80s time: 119.24s eta: 8:46:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 266 data_time: 0.61s time: 119.04s eta: 8:43:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.163 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 261 data_time: 0.88s time: 120.83s eta: 8:49:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30313.0 tgs: 252 data_time: 0.72s time: 120.20s eta: 8:44:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 270 data_time: 0.78s time: 118.83s eta: 8:36:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.232 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32352.0 tgs: 271 data_time: 0.80s time: 119.35s eta: 8:37:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32118.0 tgs: 269 data_time: 0.78s time: 119.33s eta: 8:35:07 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31800.0 tgs: 267 data_time: 0.70s time: 119.05s eta: 8:31:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32278.0 tgs: 270 data_time: 0.65s time: 119.16s eta: 8:30:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.232 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32212.0 tgs: 270 data_time: 0.79s time: 119.04s eta: 8:27:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.145 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31620.0 tgs: 265 data_time: 0.88s time: 119.12s eta: 8:26:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31643.0 tgs: 265 data_time: 0.92s time: 119.25s eta: 8:24:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31306.0 tgs: 262 data_time: 0.98s time: 119.22s eta: 8:22:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.122 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 268 data_time: 1.16s time: 119.38s eta: 8:21:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31115.0 tgs: 259 data_time: 0.76s time: 119.78s eta: 8:21:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31672.0 tgs: 265 data_time: 0.87s time: 119.20s eta: 8:16:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31812.0 tgs: 267 data_time: 0.76s time: 118.98s eta: 8:13:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 257 data_time: 0.80s time: 119.90s eta: 8:15:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30483.0 tgs: 254 data_time: 0.61s time: 119.56s eta: 8:12:10 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30317.0 tgs: 253 data_time: 0.82s time: 119.60s eta: 8:10:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31943.0 tgs: 268 data_time: 0.62s time: 119.00s eta: 8:05:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31116.0 tgs: 259 data_time: 0.72s time: 119.84s eta: 8:07:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.120 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27629.0 tgs: 230 data_time: 0.74s time: 119.71s eta: 8:04:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32097.0 tgs: 268 data_time: 0.79s time: 119.40s eta: 8:01:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31895.0 tgs: 267 data_time: 0.54s time: 119.29s eta: 7:59:09 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31210.0 tgs: 261 data_time: 0.73s time: 119.19s eta: 7:56:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32432.0 tgs: 272 data_time: 0.70s time: 119.20s eta: 7:54:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.191 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32019.0 tgs: 269 data_time: 0.79s time: 118.94s eta: 7:51:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31680.0 tgs: 266 data_time: 0.83s time: 118.97s eta: 7:49:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.196 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 271 data_time: 1.21s time: 119.42s eta: 7:49:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31530.0 tgs: 264 data_time: 0.59s time: 119.03s eta: 7:46:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32008.0 tgs: 268 data_time: 0.72s time: 119.07s eta: 7:44:22 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.112 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 269 data_time: 0.66s time: 120.00s eta: 7:45:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30695.0 tgs: 256 data_time: 0.73s time: 119.83s eta: 7:43:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32301.0 tgs: 271 data_time: 0.55s time: 119.16s eta: 7:38:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31351.0 tgs: 263 data_time: 0.78s time: 118.97s eta: 7:36:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.118 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31466.0 tgs: 263 data_time: 0.61s time: 119.33s eta: 7:35:25 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32340.0 tgs: 270 data_time: 0.78s time: 119.39s eta: 7:33:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31327.0 tgs: 263 data_time: 0.97s time: 119.07s eta: 7:30:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31370.0 tgs: 263 data_time: 0.82s time: 118.99s eta: 7:28:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31790.0 tgs: 266 data_time: 0.72s time: 119.24s eta: 7:27:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32322.0 tgs: 271 data_time: 0.84s time: 118.99s eta: 7:24:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 268 data_time: 0.82s time: 118.93s eta: 7:22:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 270 data_time: 0.78s time: 119.09s eta: 7:20:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31253.0 tgs: 262 data_time: 0.96s time: 118.99s eta: 7:18:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.112 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31151.0 tgs: 261 data_time: 0.73s time: 119.03s eta: 7:16:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 268 data_time: 0.80s time: 119.25s eta: 7:15:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30135.0 tgs: 252 data_time: 0.70s time: 119.36s eta: 7:13:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 267 data_time: 0.74s time: 118.94s eta: 7:10:09 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.227 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31232.0 tgs: 261 data_time: 0.63s time: 119.21s eta: 7:09:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31162.0 tgs: 260 data_time: 0.77s time: 119.58s eta: 7:08:30 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.118 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 266 data_time: 0.82s time: 119.67s eta: 7:06:48 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.180 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31657.0 tgs: 266 data_time: 0.74s time: 118.95s eta: 7:02:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.112 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31308.0 tgs: 262 data_time: 0.68s time: 119.30s eta: 7:01:31 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:46:30][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 268 data_time: 0.71s time: 119.52s eta: 7:00:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31385.0 tgs: 262 data_time: 0.65s time: 119.43s eta: 6:58:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30623.0 tgs: 257 data_time: 0.78s time: 118.91s eta: 6:54:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31017.0 tgs: 259 data_time: 0.70s time: 119.35s eta: 6:53:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.273 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31151.0 tgs: 261 data_time: 0.82s time: 118.94s eta: 6:50:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32069.0 tgs: 268 data_time: 0.75s time: 119.27s eta: 6:49:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30882.0 tgs: 259 data_time: 0.62s time: 119.02s eta: 6:46:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31174.0 tgs: 261 data_time: 0.65s time: 119.22s eta: 6:45:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31824.0 tgs: 266 data_time: 0.62s time: 119.37s eta: 6:43:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 260 data_time: 0.73s time: 119.02s eta: 6:40:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.187 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32005.0 tgs: 268 data_time: 0.98s time: 119.01s eta: 6:38:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 269 data_time: 0.66s time: 119.46s eta: 6:38:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31684.0 tgs: 266 data_time: 0.98s time: 118.80s eta: 6:34:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31911.0 tgs: 267 data_time: 0.71s time: 119.45s eta: 6:34:10 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31685.0 tgs: 265 data_time: 0.85s time: 119.56s eta: 6:32:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.90s time: 119.38s eta: 6:29:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31698.0 tgs: 264 data_time: 0.46s time: 119.63s eta: 6:28:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30622.0 tgs: 256 data_time: 0.62s time: 119.36s eta: 6:25:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31439.0 tgs: 264 data_time: 0.77s time: 118.82s eta: 6:22:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 269 data_time: 0.88s time: 119.02s eta: 6:20:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31740.0 tgs: 265 data_time: 0.78s time: 119.40s eta: 6:20:05 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31709.0 tgs: 265 data_time: 0.72s time: 119.34s eta: 6:17:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31568.0 tgs: 265 data_time: 0.79s time: 119.05s eta: 6:15:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 267 data_time: 0.56s time: 119.37s eta: 6:14:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31134.0 tgs: 260 data_time: 0.67s time: 119.40s eta: 6:12:07 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31572.0 tgs: 264 data_time: 0.89s time: 119.28s eta: 6:09:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31923.0 tgs: 268 data_time: 0.65s time: 119.11s eta: 6:07:14 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30930.0 tgs: 253 data_time: 0.87s time: 121.96s eta: 6:14:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.111 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31469.0 tgs: 263 data_time: 0.80s time: 119.39s eta: 6:04:07 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 263 data_time: 0.72s time: 119.58s eta: 6:02:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 268 data_time: 0.81s time: 119.26s eta: 5:59:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30842.0 tgs: 258 data_time: 0.82s time: 119.26s eta: 5:57:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.208 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30791.0 tgs: 258 data_time: 0.86s time: 118.92s eta: 5:54:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32084.0 tgs: 269 data_time: 0.58s time: 119.02s eta: 5:53:05 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31337.0 tgs: 263 data_time: 0.70s time: 119.08s eta: 5:51:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30721.0 tgs: 257 data_time: 0.60s time: 119.30s eta: 5:49:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31601.0 tgs: 265 data_time: 0.58s time: 118.95s eta: 5:46:57 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.118 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31244.0 tgs: 261 data_time: 0.71s time: 119.27s eta: 5:45:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31356.0 tgs: 262 data_time: 0.60s time: 119.32s eta: 5:44:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31797.0 tgs: 267 data_time: 0.62s time: 119.05s eta: 5:41:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32197.0 tgs: 269 data_time: 0.78s time: 119.50s eta: 5:40:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.77s time: 119.39s eta: 5:38:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31323.0 tgs: 261 data_time: 0.69s time: 119.57s eta: 5:36:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 268 data_time: 0.69s time: 118.98s eta: 5:33:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.174 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31386.0 tgs: 263 data_time: 0.91s time: 118.95s eta: 5:31:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31859.0 tgs: 267 data_time: 0.77s time: 119.25s eta: 5:29:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31634.0 tgs: 264 data_time: 0.76s time: 119.40s eta: 5:28:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 264 data_time: 0.61s time: 119.46s eta: 5:26:31 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31642.0 tgs: 266 data_time: 0.63s time: 118.95s eta: 5:23:09 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31724.0 tgs: 265 data_time: 0.85s time: 119.63s eta: 5:23:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31170.0 tgs: 261 data_time: 0.68s time: 119.35s eta: 5:20:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31657.0 tgs: 265 data_time: 0.66s time: 119.12s eta: 5:17:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31409.0 tgs: 263 data_time: 0.64s time: 119.43s eta: 5:16:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31619.0 tgs: 264 data_time: 0.65s time: 119.57s eta: 5:14:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31695.0 tgs: 266 data_time: 0.66s time: 119.01s eta: 5:11:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31184.0 tgs: 262 data_time: 0.79s time: 119.00s eta: 5:09:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.268 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31733.0 tgs: 266 data_time: 0.96s time: 119.23s eta: 5:08:00 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.097 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 259 data_time: 0.66s time: 119.68s eta: 5:07:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.175 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.86s time: 119.11s eta: 5:03:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29736.0 tgs: 249 data_time: 0.51s time: 119.07s eta: 5:01:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30589.0 tgs: 251 data_time: 0.71s time: 121.52s eta: 5:05:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31327.0 tgs: 261 data_time: 0.57s time: 119.58s eta: 4:58:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31651.0 tgs: 264 data_time: 0.67s time: 119.51s eta: 4:56:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30981.0 tgs: 258 data_time: 0.89s time: 119.73s eta: 4:55:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 269 data_time: 0.63s time: 119.52s eta: 4:52:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31235.0 tgs: 261 data_time: 0.63s time: 119.42s eta: 4:50:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 266 data_time: 0.82s time: 119.20s eta: 4:48:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32147.0 tgs: 270 data_time: 0.66s time: 118.89s eta: 4:45:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.114 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31157.0 tgs: 261 data_time: 0.69s time: 119.04s eta: 4:43:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.178 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32556.0 tgs: 271 data_time: 0.74s time: 119.98s eta: 4:43:57 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 266 data_time: 0.77s time: 119.16s eta: 4:40:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.101 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30815.0 tgs: 258 data_time: 0.55s time: 119.00s eta: 4:37:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30824.0 tgs: 258 data_time: 0.74s time: 119.06s eta: 4:35:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31422.0 tgs: 263 data_time: 0.69s time: 119.38s eta: 4:34:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.228 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31278.0 tgs: 261 data_time: 0.60s time: 119.41s eta: 4:32:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.185 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32246.0 tgs: 271 data_time: 0.83s time: 118.94s eta: 4:29:35 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32500.0 tgs: 271 data_time: 0.84s time: 119.60s eta: 4:29:06 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31312.0 tgs: 262 data_time: 0.74s time: 119.24s eta: 4:26:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.191 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32090.0 tgs: 269 data_time: 1.00s time: 119.06s eta: 4:23:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32063.0 tgs: 269 data_time: 0.79s time: 118.85s eta: 4:21:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.112 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30082.0 tgs: 253 data_time: 0.86s time: 118.76s eta: 4:19:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31216.0 tgs: 261 data_time: 0.70s time: 119.43s eta: 4:18:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31636.0 tgs: 264 data_time: 0.52s time: 119.40s eta: 4:16:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.186 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31837.0 tgs: 266 data_time: 0.66s time: 119.32s eta: 4:14:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32423.0 tgs: 272 data_time: 0.69s time: 119.05s eta: 4:11:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32034.0 tgs: 268 data_time: 0.69s time: 119.13s eta: 4:10:10 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31699.0 tgs: 266 data_time: 0.88s time: 119.13s eta: 4:08:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.181 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31731.0 tgs: 266 data_time: 0.81s time: 118.96s eta: 4:05:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.205 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31173.0 tgs: 261 data_time: 0.71s time: 119.16s eta: 4:04:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.192 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30948.0 tgs: 258 data_time: 0.69s time: 119.49s eta: 4:02:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31499.0 tgs: 264 data_time: 0.86s time: 119.25s eta: 4:00:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31743.0 tgs: 266 data_time: 0.51s time: 119.19s eta: 3:58:23 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32189.0 tgs: 269 data_time: 0.60s time: 119.36s eta: 3:56:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31398.0 tgs: 263 data_time: 0.72s time: 119.38s eta: 3:54:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30454.0 tgs: 253 data_time: 0.52s time: 119.95s eta: 3:53:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31934.0 tgs: 267 data_time: 0.50s time: 119.32s eta: 3:50:41 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 262 data_time: 0.62s time: 119.64s eta: 3:49:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32449.0 tgs: 273 data_time: 0.68s time: 118.85s eta: 3:45:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32082.0 tgs: 268 data_time: 0.72s time: 119.70s eta: 3:45:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 268 data_time: 0.86s time: 119.33s eta: 3:42:44 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31845.0 tgs: 267 data_time: 0.43s time: 119.26s eta: 3:40:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31823.0 tgs: 267 data_time: 0.79s time: 119.00s eta: 3:38:10 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31240.0 tgs: 261 data_time: 0.73s time: 119.43s eta: 3:36:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31848.0 tgs: 267 data_time: 0.66s time: 119.03s eta: 3:34:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31764.0 tgs: 266 data_time: 0.82s time: 118.99s eta: 3:32:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31847.0 tgs: 267 data_time: 0.75s time: 119.22s eta: 3:30:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32068.0 tgs: 266 data_time: 0.56s time: 120.11s eta: 3:30:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 269 data_time: 0.89s time: 119.13s eta: 3:26:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 268 data_time: 0.64s time: 119.31s eta: 3:24:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31657.0 tgs: 265 data_time: 0.68s time: 119.08s eta: 3:22:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31659.0 tgs: 265 data_time: 0.98s time: 119.43s eta: 3:21:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31541.0 tgs: 264 data_time: 0.60s time: 119.29s eta: 3:18:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30513.0 tgs: 255 data_time: 0.73s time: 119.22s eta: 3:16:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31295.0 tgs: 262 data_time: 0.86s time: 119.13s eta: 3:14:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 267 data_time: 0.45s time: 119.02s eta: 3:12:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32171.0 tgs: 269 data_time: 0.80s time: 119.19s eta: 3:10:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32370.0 tgs: 267 data_time: 0.75s time: 120.94s eta: 3:11:28 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30920.0 tgs: 258 data_time: 0.64s time: 119.51s eta: 3:07:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32211.0 tgs: 270 data_time: 0.96s time: 119.27s eta: 3:04:51 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.185 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31530.0 tgs: 264 data_time: 0.51s time: 118.99s eta: 3:02:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 267 data_time: 0.77s time: 119.23s eta: 3:00:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 265 data_time: 0.73s time: 119.46s eta: 2:59:10 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32031.0 tgs: 269 data_time: 0.94s time: 118.99s eta: 2:56:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 265 data_time: 0.78s time: 119.35s eta: 2:55:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.207 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31643.0 tgs: 264 data_time: 0.84s time: 119.41s eta: 2:53:08 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31911.0 tgs: 267 data_time: 0.88s time: 119.36s eta: 2:51:04 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32067.0 tgs: 264 data_time: 0.80s time: 121.46s eta: 2:52:03 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 264 data_time: 0.67s time: 119.24s eta: 2:46:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31845.0 tgs: 266 data_time: 0.67s time: 119.54s eta: 2:45:21 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.200 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31181.0 tgs: 262 data_time: 0.71s time: 118.91s eta: 2:42:30 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31734.0 tgs: 266 data_time: 0.67s time: 119.18s eta: 2:40:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31980.0 tgs: 268 data_time: 0.55s time: 118.97s eta: 2:38:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 270 data_time: 0.80s time: 119.23s eta: 2:36:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31151.0 tgs: 261 data_time: 0.66s time: 119.11s eta: 2:34:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31354.0 tgs: 262 data_time: 0.76s time: 119.43s eta: 2:33:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30460.0 tgs: 253 data_time: 0.72s time: 120.23s eta: 2:32:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 262 data_time: 0.78s time: 119.29s eta: 2:29:06 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31902.0 tgs: 268 data_time: 0.78s time: 118.88s eta: 2:26:37 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 264 data_time: 0.58s time: 119.52s eta: 2:25:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31751.0 tgs: 266 data_time: 0.62s time: 119.07s eta: 2:22:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32194.0 tgs: 270 data_time: 0.65s time: 119.14s eta: 2:20:58 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31143.0 tgs: 261 data_time: 0.62s time: 119.05s eta: 2:18:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31451.0 tgs: 262 data_time: 1.10s time: 119.85s eta: 2:17:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32236.0 tgs: 270 data_time: 0.92s time: 119.31s eta: 2:15:13 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 265 data_time: 0.49s time: 119.38s eta: 2:13:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32271.0 tgs: 269 data_time: 0.66s time: 119.68s eta: 2:11:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31812.0 tgs: 266 data_time: 0.70s time: 119.49s eta: 2:09:27 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31345.0 tgs: 263 data_time: 0.62s time: 118.96s eta: 2:06:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32208.0 tgs: 269 data_time: 0.74s time: 119.37s eta: 2:05:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32034.0 tgs: 268 data_time: 0.75s time: 119.28s eta: 2:03:15 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31919.0 tgs: 267 data_time: 0.87s time: 119.22s eta: 2:01:12 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31755.0 tgs: 265 data_time: 0.61s time: 119.51s eta: 1:59:30 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 268 data_time: 0.87s time: 119.42s eta: 1:57:25 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30763.0 tgs: 257 data_time: 0.71s time: 119.54s eta: 1:55:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30975.0 tgs: 259 data_time: 0.72s time: 119.30s eta: 1:53:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31884.0 tgs: 267 data_time: 0.76s time: 119.26s eta: 1:51:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32717.0 tgs: 275 data_time: 0.76s time: 118.93s eta: 1:49:01 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31060.0 tgs: 259 data_time: 0.60s time: 119.64s eta: 1:47:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.200 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31628.0 tgs: 264 data_time: 1.05s time: 119.69s eta: 1:45:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31790.0 tgs: 266 data_time: 0.78s time: 119.34s eta: 1:43:25 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31093.0 tgs: 260 data_time: 0.65s time: 119.31s eta: 1:41:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31460.0 tgs: 263 data_time: 0.61s time: 119.28s eta: 1:39:24 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31548.0 tgs: 265 data_time: 0.70s time: 118.91s eta: 1:37:06 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31262.0 tgs: 262 data_time: 0.75s time: 119.16s eta: 1:35:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31884.0 tgs: 267 data_time: 0.71s time: 119.07s eta: 1:33:16 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32320.0 tgs: 271 data_time: 0.90s time: 119.26s eta: 1:31:26 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31061.0 tgs: 259 data_time: 0.74s time: 119.72s eta: 1:29:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31479.0 tgs: 264 data_time: 0.68s time: 119.02s eta: 1:27:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31756.0 tgs: 267 data_time: 0.55s time: 118.87s eta: 1:25:11 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31951.0 tgs: 267 data_time: 0.97s time: 119.48s eta: 1:23:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32022.0 tgs: 268 data_time: 0.84s time: 119.48s eta: 1:21:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32116.0 tgs: 270 data_time: 0.87s time: 118.94s eta: 1:19:17 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 268 data_time: 0.79s time: 118.93s eta: 1:17:18 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32443.0 tgs: 265 data_time: 0.69s time: 122.13s eta: 1:17:20 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31437.0 tgs: 263 data_time: 0.68s time: 119.35s eta: 1:13:36 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.223 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32125.0 tgs: 269 data_time: 0.76s time: 119.05s eta: 1:11:25 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31143.0 tgs: 261 data_time: 0.80s time: 118.92s eta: 1:09:22 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31354.0 tgs: 263 data_time: 0.73s time: 119.11s eta: 1:07:29 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.186 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32238.0 tgs: 269 data_time: 0.68s time: 119.42s eta: 1:05:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.70s time: 119.35s eta: 1:03:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31718.0 tgs: 266 data_time: 0.63s time: 119.18s eta: 1:01:34 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 268 data_time: 0.87s time: 119.27s eta: 0:59:38 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32284.0 tgs: 270 data_time: 0.76s time: 119.45s eta: 0:57:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.242 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 264 data_time: 0.90s time: 119.40s eta: 0:55:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 264 data_time: 0.69s time: 119.48s eta: 0:53:45 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.225 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31932.0 tgs: 267 data_time: 0.81s time: 119.36s eta: 0:51:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 268 data_time: 0.88s time: 119.47s eta: 0:49:46 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31612.0 tgs: 265 data_time: 0.83s time: 118.89s eta: 0:47:33 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31583.0 tgs: 264 data_time: 0.57s time: 119.27s eta: 0:45:43 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31801.0 tgs: 266 data_time: 0.74s time: 119.19s eta: 0:43:42 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32207.0 tgs: 270 data_time: 0.43s time: 119.01s eta: 0:41:39 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31280.0 tgs: 262 data_time: 0.94s time: 119.03s eta: 0:39:40 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 265 data_time: 0.70s time: 119.66s eta: 0:37:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 266 data_time: 0.60s time: 119.29s eta: 0:35:47 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31941.0 tgs: 266 data_time: 0.80s time: 119.75s eta: 0:33:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 263 data_time: 0.80s time: 121.23s eta: 0:32:19 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.090 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29970.0 tgs: 251 data_time: 0.39s time: 119.38s eta: 0:29:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31575.0 tgs: 264 data_time: 0.75s time: 119.32s eta: 0:27:50 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30964.0 tgs: 258 data_time: 0.56s time: 119.64s eta: 0:25:55 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30870.0 tgs: 259 data_time: 0.58s time: 119.10s eta: 0:23:49 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31970.0 tgs: 267 data_time: 0.65s time: 119.39s eta: 0:21:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31745.0 tgs: 266 data_time: 0.77s time: 119.26s eta: 0:19:52 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 268 data_time: 0.64s time: 119.25s eta: 0:17:53 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 265 data_time: 0.56s time: 119.90s eta: 0:15:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31585.0 tgs: 265 data_time: 0.66s time: 119.18s eta: 0:13:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31779.0 tgs: 266 data_time: 0.77s time: 119.10s eta: 0:11:54 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.70s time: 119.32s eta: 0:09:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31664.0 tgs: 265 data_time: 0.72s time: 119.17s eta: 0:07:56 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.186 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30712.0 tgs: 256 data_time: 0.64s time: 119.69s eta: 0:05:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31311.0 tgs: 257 data_time: 0.72s time: 121.43s eta: 0:04:02 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31199.0 tgs: 261 data_time: 0.71s time: 119.21s eta: 0:01:59 +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.29 seconds, peak gpu memory 17.4G +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 37][DP 9][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank39.log b/20250121165312/rank39.log new file mode 100644 index 0000000000000000000000000000000000000000..d433b7aa4727156e7de83addb6a195ce3ab7033c --- /dev/null +++ b/20250121165312/rank39.log @@ -0,0 +1,851 @@ +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:35][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.05s +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.40 seconds, peak gpu memory 13.4G +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.300 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 32281.0 tgs: 212 data_time: 2.37s time: 152.26s eta: 1 day, 1:04:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.233 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32206.0 tgs: 268 data_time: 0.77s time: 120.06s eta: 19:44:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.283 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32213.0 tgs: 269 data_time: 0.94s time: 119.68s eta: 19:38:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29331.0 tgs: 245 data_time: 0.87s time: 119.66s eta: 19:36:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.198 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31870.0 tgs: 267 data_time: 0.91s time: 119.31s eta: 19:31:10 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.213 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29472.0 tgs: 246 data_time: 0.70s time: 119.54s eta: 19:31:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.185 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 261 data_time: 0.73s time: 122.96s eta: 20:02:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30854.0 tgs: 257 data_time: 0.69s time: 119.89s eta: 19:30:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.209 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31900.0 tgs: 264 data_time: 0.75s time: 120.66s eta: 19:36:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.180 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32109.0 tgs: 267 data_time: 0.79s time: 119.92s eta: 19:27:12 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31161.0 tgs: 258 data_time: 0.60s time: 120.46s eta: 19:30:27 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.204 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31536.0 tgs: 264 data_time: 0.81s time: 119.35s eta: 19:17:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 267 data_time: 0.77s time: 119.67s eta: 19:18:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32359.0 tgs: 271 data_time: 0.77s time: 119.18s eta: 19:12:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.242 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32408.0 tgs: 271 data_time: 0.97s time: 119.18s eta: 19:10:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31413.0 tgs: 264 data_time: 1.00s time: 118.93s eta: 19:05:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31857.0 tgs: 267 data_time: 0.78s time: 119.15s eta: 19:05:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31570.0 tgs: 264 data_time: 0.74s time: 119.18s eta: 19:04:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31566.0 tgs: 264 data_time: 0.68s time: 119.40s eta: 19:04:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32203.0 tgs: 269 data_time: 0.80s time: 119.33s eta: 19:01:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30573.0 tgs: 254 data_time: 0.66s time: 120.31s eta: 19:08:58 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30164.0 tgs: 245 data_time: 0.69s time: 122.92s eta: 19:31:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31721.0 tgs: 265 data_time: 0.77s time: 119.48s eta: 18:57:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31235.0 tgs: 262 data_time: 0.77s time: 119.16s eta: 18:51:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 269 data_time: 0.69s time: 119.25s eta: 18:50:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.243 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31111.0 tgs: 260 data_time: 0.90s time: 119.39s eta: 18:50:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31722.0 tgs: 265 data_time: 0.82s time: 119.66s eta: 18:50:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30903.0 tgs: 258 data_time: 0.81s time: 119.39s eta: 18:46:12 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31566.0 tgs: 263 data_time: 1.00s time: 119.76s eta: 18:47:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 28740.0 tgs: 222 data_time: 0.80s time: 129.31s eta: 20:15:31 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31943.0 tgs: 267 data_time: 0.62s time: 119.57s eta: 18:41:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31226.0 tgs: 261 data_time: 0.84s time: 119.39s eta: 18:38:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31930.0 tgs: 264 data_time: 0.66s time: 120.66s eta: 18:48:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31763.0 tgs: 266 data_time: 0.89s time: 119.40s eta: 18:34:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32155.0 tgs: 269 data_time: 0.83s time: 119.16s eta: 18:30:09 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31819.0 tgs: 267 data_time: 0.89s time: 119.07s eta: 18:27:22 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32555.0 tgs: 273 data_time: 0.64s time: 119.14s eta: 18:25:58 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31507.0 tgs: 261 data_time: 0.75s time: 120.39s eta: 18:35:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32485.0 tgs: 268 data_time: 0.87s time: 121.17s eta: 18:40:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32286.0 tgs: 267 data_time: 0.92s time: 120.66s eta: 18:34:06 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32270.0 tgs: 269 data_time: 0.75s time: 119.74s eta: 18:23:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 257 data_time: 0.93s time: 123.87s eta: 18:59:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31139.0 tgs: 260 data_time: 0.74s time: 119.56s eta: 18:17:58 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 266 data_time: 0.68s time: 119.32s eta: 18:13:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32014.0 tgs: 268 data_time: 0.85s time: 119.22s eta: 18:10:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31950.0 tgs: 267 data_time: 0.99s time: 119.60s eta: 18:12:21 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31159.0 tgs: 261 data_time: 0.61s time: 119.27s eta: 18:07:21 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31876.0 tgs: 267 data_time: 0.95s time: 119.01s eta: 18:03:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.234 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31681.0 tgs: 266 data_time: 0.89s time: 119.10s eta: 18:01:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.123 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31594.0 tgs: 264 data_time: 1.18s time: 119.27s eta: 18:01:22 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31032.0 tgs: 260 data_time: 1.04s time: 119.33s eta: 17:59:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31014.0 tgs: 258 data_time: 0.95s time: 120.02s eta: 18:04:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.115 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32168.0 tgs: 270 data_time: 0.97s time: 119.11s eta: 17:53:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.73s time: 119.41s eta: 17:54:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 272 data_time: 0.93s time: 119.00s eta: 17:48:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31439.0 tgs: 264 data_time: 0.70s time: 119.03s eta: 17:47:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31610.0 tgs: 252 data_time: 0.58s time: 125.36s eta: 18:42:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31493.0 tgs: 262 data_time: 0.88s time: 120.10s eta: 17:52:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31578.0 tgs: 264 data_time: 0.76s time: 119.36s eta: 17:44:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32478.0 tgs: 272 data_time: 1.04s time: 119.32s eta: 17:41:58 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29943.0 tgs: 250 data_time: 0.84s time: 119.51s eta: 17:41:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31038.0 tgs: 246 data_time: 0.56s time: 126.05s eta: 18:37:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.134 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31043.0 tgs: 253 data_time: 0.81s time: 122.63s eta: 18:05:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32383.0 tgs: 267 data_time: 0.87s time: 121.16s eta: 17:50:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31991.0 tgs: 268 data_time: 0.66s time: 119.04s eta: 17:29:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30412.0 tgs: 255 data_time: 0.84s time: 118.86s eta: 17:25:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31377.0 tgs: 253 data_time: 1.03s time: 124.00s eta: 18:09:07 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 254 data_time: 0.76s time: 124.57s eta: 18:12:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32481.0 tgs: 272 data_time: 0.98s time: 119.40s eta: 17:24:44 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31743.0 tgs: 265 data_time: 0.99s time: 119.71s eta: 17:25:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.250 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32084.0 tgs: 268 data_time: 0.88s time: 119.52s eta: 17:21:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32042.0 tgs: 269 data_time: 0.81s time: 118.86s eta: 17:14:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31645.0 tgs: 264 data_time: 1.01s time: 119.85s eta: 17:20:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29019.0 tgs: 242 data_time: 0.79s time: 119.53s eta: 17:15:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31875.0 tgs: 261 data_time: 0.78s time: 121.86s eta: 17:34:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32461.0 tgs: 262 data_time: 0.60s time: 123.51s eta: 17:46:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31810.0 tgs: 267 data_time: 0.65s time: 119.01s eta: 17:05:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.114 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 267 data_time: 0.88s time: 119.23s eta: 17:05:22 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31702.0 tgs: 266 data_time: 0.78s time: 119.14s eta: 17:02:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 28580.0 tgs: 239 data_time: 0.86s time: 119.26s eta: 17:01:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31749.0 tgs: 266 data_time: 0.81s time: 119.04s eta: 16:57:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.108 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31455.0 tgs: 263 data_time: 0.84s time: 119.28s eta: 16:57:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30811.0 tgs: 258 data_time: 0.73s time: 119.16s eta: 16:54:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31311.0 tgs: 261 data_time: 0.66s time: 119.59s eta: 16:56:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32026.0 tgs: 268 data_time: 0.79s time: 119.14s eta: 16:50:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31318.0 tgs: 263 data_time: 0.65s time: 118.84s eta: 16:46:09 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31265.0 tgs: 262 data_time: 0.94s time: 119.02s eta: 16:45:44 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31698.0 tgs: 264 data_time: 0.72s time: 119.84s eta: 16:50:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32116.0 tgs: 269 data_time: 0.73s time: 119.36s eta: 16:44:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31820.0 tgs: 267 data_time: 0.57s time: 119.01s eta: 16:39:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32398.0 tgs: 267 data_time: 0.74s time: 120.93s eta: 16:53:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32077.0 tgs: 269 data_time: 0.79s time: 118.99s eta: 16:35:35 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 268 data_time: 0.84s time: 119.30s eta: 16:36:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 265 data_time: 0.63s time: 119.22s eta: 16:33:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.210 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31125.0 tgs: 260 data_time: 0.73s time: 119.37s eta: 16:32:44 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31589.0 tgs: 263 data_time: 0.86s time: 119.70s eta: 16:33:31 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 264 data_time: 0.70s time: 119.21s eta: 16:27:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30913.0 tgs: 258 data_time: 0.73s time: 119.76s eta: 16:30:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 260 data_time: 0.86s time: 120.36s eta: 16:32:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32113.0 tgs: 269 data_time: 1.06s time: 118.97s eta: 16:19:32 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31989.0 tgs: 251 data_time: 0.94s time: 127.33s eta: 17:26:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31887.0 tgs: 267 data_time: 0.73s time: 119.22s eta: 16:17:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 267 data_time: 0.56s time: 119.14s eta: 16:14:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31366.0 tgs: 263 data_time: 0.72s time: 119.21s eta: 16:13:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31395.0 tgs: 263 data_time: 0.80s time: 119.17s eta: 16:11:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32048.0 tgs: 269 data_time: 1.04s time: 118.95s eta: 16:07:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31406.0 tgs: 264 data_time: 0.69s time: 118.86s eta: 16:04:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31434.0 tgs: 264 data_time: 0.53s time: 118.96s eta: 16:03:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 258 data_time: 0.74s time: 121.84s eta: 16:24:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31767.0 tgs: 266 data_time: 0.73s time: 119.00s eta: 15:59:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.229 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31044.0 tgs: 260 data_time: 0.82s time: 119.23s eta: 15:59:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.232 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 267 data_time: 0.77s time: 119.61s eta: 16:00:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 271 data_time: 0.70s time: 119.04s eta: 15:54:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:50:49][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32439.0 tgs: 272 data_time: 0.66s time: 118.99s eta: 15:51:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.75s time: 119.17s eta: 15:51:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31837.0 tgs: 266 data_time: 0.56s time: 119.64s eta: 15:53:06 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.201 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 268 data_time: 0.75s time: 118.94s eta: 15:45:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32245.0 tgs: 270 data_time: 0.96s time: 119.28s eta: 15:46:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31475.0 tgs: 264 data_time: 1.00s time: 119.10s eta: 15:42:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31148.0 tgs: 261 data_time: 0.78s time: 119.00s eta: 15:40:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31417.0 tgs: 262 data_time: 0.83s time: 119.66s eta: 15:43:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 259 data_time: 0.77s time: 121.50s eta: 15:55:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31756.0 tgs: 265 data_time: 0.77s time: 119.41s eta: 15:37:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 247 data_time: 1.00s time: 129.96s eta: 16:58:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31805.0 tgs: 266 data_time: 1.02s time: 119.47s eta: 15:33:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31292.0 tgs: 261 data_time: 0.84s time: 119.60s eta: 15:32:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30814.0 tgs: 257 data_time: 0.84s time: 119.53s eta: 15:30:21 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31413.0 tgs: 257 data_time: 0.69s time: 122.14s eta: 15:48:34 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.122 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 265 data_time: 0.94s time: 120.75s eta: 15:35:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 265 data_time: 0.82s time: 119.41s eta: 15:23:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.70s time: 119.45s eta: 15:21:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31520.0 tgs: 239 data_time: 0.86s time: 131.64s eta: 16:53:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31932.0 tgs: 267 data_time: 0.62s time: 119.29s eta: 15:16:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.120 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30315.0 tgs: 253 data_time: 0.69s time: 119.52s eta: 15:16:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31784.0 tgs: 259 data_time: 0.80s time: 122.25s eta: 15:35:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 26939.0 tgs: 225 data_time: 0.44s time: 119.72s eta: 15:13:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31378.0 tgs: 262 data_time: 0.96s time: 119.36s eta: 15:09:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31717.0 tgs: 266 data_time: 0.82s time: 118.90s eta: 15:03:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.181 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 270 data_time: 0.91s time: 118.99s eta: 15:02:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32422.0 tgs: 271 data_time: 0.87s time: 119.31s eta: 15:02:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30691.0 tgs: 258 data_time: 0.76s time: 118.92s eta: 14:57:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31614.0 tgs: 262 data_time: 0.85s time: 120.52s eta: 15:07:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31621.0 tgs: 265 data_time: 0.55s time: 119.32s eta: 14:56:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29421.0 tgs: 239 data_time: 0.74s time: 122.65s eta: 15:19:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.214 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32141.0 tgs: 269 data_time: 0.78s time: 119.13s eta: 14:51:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31387.0 tgs: 262 data_time: 0.74s time: 119.76s eta: 14:54:10 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 263 data_time: 0.85s time: 119.47s eta: 14:50:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 260 data_time: 0.87s time: 118.99s eta: 14:44:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30723.0 tgs: 257 data_time: 0.52s time: 119.32s eta: 14:44:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31767.0 tgs: 267 data_time: 0.89s time: 118.93s eta: 14:40:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30699.0 tgs: 257 data_time: 0.84s time: 119.05s eta: 14:39:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.65s time: 119.06s eta: 14:37:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32007.0 tgs: 268 data_time: 0.54s time: 119.13s eta: 14:35:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32346.0 tgs: 271 data_time: 0.98s time: 119.00s eta: 14:32:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 269 data_time: 0.83s time: 119.08s eta: 14:31:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.229 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31216.0 tgs: 261 data_time: 0.83s time: 119.15s eta: 14:29:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 261 data_time: 0.98s time: 119.65s eta: 14:31:27 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32106.0 tgs: 269 data_time: 0.76s time: 119.18s eta: 14:26:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31674.0 tgs: 265 data_time: 0.93s time: 119.20s eta: 14:24:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31484.0 tgs: 263 data_time: 0.88s time: 119.39s eta: 14:23:34 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31704.0 tgs: 262 data_time: 0.62s time: 120.74s eta: 14:31:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31765.0 tgs: 267 data_time: 0.88s time: 118.86s eta: 14:15:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31034.0 tgs: 260 data_time: 0.91s time: 119.10s eta: 14:15:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.218 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32110.0 tgs: 269 data_time: 1.10s time: 119.12s eta: 14:13:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.209 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 261 data_time: 0.97s time: 119.95s eta: 14:17:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28445.0 tgs: 237 data_time: 0.64s time: 119.86s eta: 14:15:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 267 data_time: 0.77s time: 119.59s eta: 14:11:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31811.0 tgs: 266 data_time: 0.82s time: 119.22s eta: 14:06:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31451.0 tgs: 263 data_time: 0.74s time: 119.27s eta: 14:04:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.174 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31249.0 tgs: 262 data_time: 0.63s time: 119.27s eta: 14:02:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32506.0 tgs: 272 data_time: 0.81s time: 119.15s eta: 13:59:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31517.0 tgs: 264 data_time: 0.53s time: 119.12s eta: 13:57:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 268 data_time: 0.68s time: 119.17s eta: 13:56:09 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31591.0 tgs: 264 data_time: 0.76s time: 119.35s eta: 13:55:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31486.0 tgs: 262 data_time: 0.60s time: 120.12s eta: 13:58:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31404.0 tgs: 256 data_time: 1.00s time: 122.56s eta: 14:13:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30847.0 tgs: 258 data_time: 0.55s time: 119.10s eta: 13:47:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31370.0 tgs: 262 data_time: 0.66s time: 119.37s eta: 13:47:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32348.0 tgs: 271 data_time: 0.94s time: 119.33s eta: 13:45:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.107 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31441.0 tgs: 263 data_time: 0.77s time: 119.33s eta: 13:43:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32536.0 tgs: 273 data_time: 0.77s time: 119.12s eta: 13:39:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32140.0 tgs: 269 data_time: 0.87s time: 119.26s eta: 13:38:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.184 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.81s time: 119.14s eta: 13:36:04 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32511.0 tgs: 272 data_time: 0.67s time: 119.22s eta: 13:34:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32096.0 tgs: 268 data_time: 0.89s time: 119.46s eta: 13:34:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31630.0 tgs: 265 data_time: 0.77s time: 119.09s eta: 13:29:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.76s time: 118.92s eta: 13:26:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 0.56s time: 119.81s eta: 13:30:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 267 data_time: 0.85s time: 119.26s eta: 13:24:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30171.0 tgs: 253 data_time: 0.82s time: 119.16s eta: 13:22:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.216 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31004.0 tgs: 259 data_time: 0.86s time: 119.35s eta: 13:21:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31610.0 tgs: 265 data_time: 0.65s time: 119.26s eta: 13:19:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31674.0 tgs: 266 data_time: 0.72s time: 119.04s eta: 13:15:35 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32220.0 tgs: 269 data_time: 1.02s time: 119.35s eta: 13:15:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32555.0 tgs: 273 data_time: 0.73s time: 119.17s eta: 13:12:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.188 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32403.0 tgs: 271 data_time: 0.86s time: 119.46s eta: 13:12:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.203 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31960.0 tgs: 267 data_time: 0.96s time: 119.29s eta: 13:09:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30806.0 tgs: 258 data_time: 0.54s time: 119.15s eta: 13:06:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.197 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31620.0 tgs: 262 data_time: 0.55s time: 120.36s eta: 13:12:22 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32095.0 tgs: 267 data_time: 0.84s time: 120.16s eta: 13:09:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30004.0 tgs: 251 data_time: 0.75s time: 119.50s eta: 13:02:44 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.249 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31533.0 tgs: 263 data_time: 0.70s time: 119.51s eta: 13:00:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.211 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 265 data_time: 0.72s time: 119.58s eta: 12:59:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31699.0 tgs: 264 data_time: 0.54s time: 119.79s eta: 12:58:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.70s time: 119.04s eta: 12:51:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30943.0 tgs: 259 data_time: 0.54s time: 119.05s eta: 12:49:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.304 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 0.74s time: 119.12s eta: 12:48:21 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31768.0 tgs: 266 data_time: 0.60s time: 119.40s eta: 12:48:07 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31973.0 tgs: 268 data_time: 0.86s time: 119.18s eta: 12:44:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31047.0 tgs: 259 data_time: 0.78s time: 119.41s eta: 12:44:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.91s time: 119.10s eta: 12:40:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.117 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31914.0 tgs: 268 data_time: 0.87s time: 119.07s eta: 12:38:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.179 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31686.0 tgs: 264 data_time: 1.15s time: 119.60s eta: 12:39:27 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31041.0 tgs: 261 data_time: 0.80s time: 118.78s eta: 12:32:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31048.0 tgs: 261 data_time: 0.84s time: 118.89s eta: 12:30:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32065.0 tgs: 269 data_time: 0.62s time: 119.17s eta: 12:30:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 262 data_time: 0.72s time: 119.27s eta: 12:29:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 265 data_time: 0.74s time: 119.35s eta: 12:27:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32142.0 tgs: 269 data_time: 0.65s time: 119.12s eta: 12:24:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.200 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31871.0 tgs: 267 data_time: 0.78s time: 119.36s eta: 12:24:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31790.0 tgs: 266 data_time: 0.90s time: 119.25s eta: 12:21:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31075.0 tgs: 261 data_time: 0.69s time: 119.03s eta: 12:17:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.196 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 262 data_time: 0.54s time: 121.13s eta: 12:28:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30518.0 tgs: 254 data_time: 0.73s time: 119.99s eta: 12:19:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32133.0 tgs: 270 data_time: 0.69s time: 118.69s eta: 12:09:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.185 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30697.0 tgs: 256 data_time: 0.90s time: 119.64s eta: 12:13:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31613.0 tgs: 263 data_time: 0.85s time: 119.83s eta: 12:12:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31313.0 tgs: 262 data_time: 0.74s time: 119.14s eta: 12:06:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31978.0 tgs: 268 data_time: 1.03s time: 119.17s eta: 12:04:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.80s time: 119.12s eta: 12:02:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31458.0 tgs: 263 data_time: 0.73s time: 119.19s eta: 12:01:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31714.0 tgs: 266 data_time: 0.84s time: 118.96s eta: 11:57:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31425.0 tgs: 262 data_time: 0.53s time: 119.70s eta: 12:00:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32243.0 tgs: 270 data_time: 0.82s time: 119.28s eta: 11:55:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31927.0 tgs: 267 data_time: 0.60s time: 119.28s eta: 11:53:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31260.0 tgs: 261 data_time: 0.77s time: 119.71s eta: 11:54:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 271 data_time: 0.79s time: 118.79s eta: 11:46:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.186 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31547.0 tgs: 264 data_time: 0.56s time: 119.16s eta: 11:47:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31192.0 tgs: 260 data_time: 0.72s time: 119.61s eta: 11:47:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31341.0 tgs: 263 data_time: 0.98s time: 118.82s eta: 11:41:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 269 data_time: 0.53s time: 118.68s eta: 11:38:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30626.0 tgs: 256 data_time: 0.51s time: 119.22s eta: 11:39:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.211 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 266 data_time: 0.90s time: 119.14s eta: 11:36:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 268 data_time: 1.01s time: 119.08s eta: 11:34:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31960.0 tgs: 268 data_time: 0.84s time: 118.94s eta: 11:31:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.204 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31888.0 tgs: 268 data_time: 0.83s time: 118.92s eta: 11:29:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31588.0 tgs: 265 data_time: 0.79s time: 118.97s eta: 11:28:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.96s time: 119.29s eta: 11:27:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 268 data_time: 0.95s time: 119.36s eta: 11:26:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.222 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32187.0 tgs: 270 data_time: 0.91s time: 118.88s eta: 11:21:35 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31718.0 tgs: 266 data_time: 0.86s time: 119.21s eta: 11:21:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30398.0 tgs: 253 data_time: 0.59s time: 119.72s eta: 11:22:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.214 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30968.0 tgs: 259 data_time: 1.03s time: 119.12s eta: 11:16:58 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.170 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 268 data_time: 0.70s time: 119.13s eta: 11:15:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30670.0 tgs: 257 data_time: 0.67s time: 118.94s eta: 11:11:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31500.0 tgs: 264 data_time: 0.77s time: 119.09s eta: 11:10:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.190 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31599.0 tgs: 265 data_time: 0.65s time: 119.16s eta: 11:09:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31970.0 tgs: 268 data_time: 0.97s time: 118.98s eta: 11:06:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.180 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32253.0 tgs: 269 data_time: 0.91s time: 119.46s eta: 11:07:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31310.0 tgs: 261 data_time: 0.80s time: 119.64s eta: 11:06:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32190.0 tgs: 269 data_time: 0.76s time: 119.30s eta: 11:02:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.197 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31668.0 tgs: 265 data_time: 0.85s time: 119.31s eta: 11:00:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31349.0 tgs: 263 data_time: 1.05s time: 118.82s eta: 10:55:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 268 data_time: 0.90s time: 119.47s eta: 10:57:06 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 267 data_time: 0.56s time: 119.37s eta: 10:54:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31849.0 tgs: 267 data_time: 0.69s time: 119.28s eta: 10:52:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31869.0 tgs: 267 data_time: 0.77s time: 119.26s eta: 10:49:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31631.0 tgs: 265 data_time: 0.85s time: 119.10s eta: 10:47:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30525.0 tgs: 255 data_time: 0.58s time: 119.54s eta: 10:47:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.79s time: 119.30s eta: 10:44:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 267 data_time: 0.57s time: 119.49s eta: 10:43:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31464.0 tgs: 263 data_time: 0.75s time: 119.52s eta: 10:41:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31730.0 tgs: 266 data_time: 0.59s time: 118.99s eta: 10:36:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30533.0 tgs: 256 data_time: 0.73s time: 119.22s eta: 10:35:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.101 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31617.0 tgs: 265 data_time: 0.61s time: 119.06s eta: 10:33:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 272 data_time: 0.89s time: 118.78s eta: 10:29:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 263 data_time: 1.18s time: 119.86s eta: 10:33:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32281.0 tgs: 270 data_time: 0.79s time: 119.39s eta: 10:28:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31813.0 tgs: 267 data_time: 0.78s time: 118.83s eta: 10:23:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.124 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32442.0 tgs: 272 data_time: 0.79s time: 118.99s eta: 10:22:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31092.0 tgs: 260 data_time: 0.65s time: 119.29s eta: 10:22:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30707.0 tgs: 257 data_time: 1.06s time: 119.20s eta: 10:19:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.121 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31109.0 tgs: 261 data_time: 0.76s time: 119.06s eta: 10:17:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31987.0 tgs: 267 data_time: 0.72s time: 119.48s eta: 10:17:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32353.0 tgs: 271 data_time: 0.81s time: 119.37s eta: 10:14:44 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.131 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.60s time: 119.45s eta: 10:13:10 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31262.0 tgs: 262 data_time: 0.76s time: 119.29s eta: 10:10:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 269 data_time: 0.84s time: 119.11s eta: 10:07:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30333.0 tgs: 253 data_time: 0.65s time: 119.70s eta: 10:08:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.76s time: 119.83s eta: 10:07:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 259 data_time: 1.01s time: 121.56s eta: 10:13:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31917.0 tgs: 267 data_time: 0.84s time: 119.26s eta: 10:00:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32009.0 tgs: 269 data_time: 0.87s time: 118.81s eta: 9:56:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30963.0 tgs: 258 data_time: 0.94s time: 119.56s eta: 9:57:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.125 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31840.0 tgs: 265 data_time: 0.90s time: 119.71s eta: 9:56:32 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.91s time: 119.26s eta: 9:52:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31173.0 tgs: 261 data_time: 0.83s time: 119.33s eta: 9:50:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 264 data_time: 0.76s time: 119.22s eta: 9:48:10 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31499.0 tgs: 260 data_time: 0.74s time: 120.88s eta: 9:54:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32114.0 tgs: 269 data_time: 0.78s time: 119.23s eta: 9:44:12 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.209 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30399.0 tgs: 254 data_time: 0.69s time: 119.45s eta: 9:43:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32432.0 tgs: 271 data_time: 0.70s time: 119.55s eta: 9:41:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 270 data_time: 0.77s time: 119.20s eta: 9:38:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28686.0 tgs: 240 data_time: 0.71s time: 119.48s eta: 9:37:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31700.0 tgs: 266 data_time: 0.80s time: 119.05s eta: 9:33:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31095.0 tgs: 260 data_time: 0.95s time: 119.24s eta: 9:32:21 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31440.0 tgs: 263 data_time: 0.68s time: 119.16s eta: 9:29:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.183 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31957.0 tgs: 268 data_time: 0.74s time: 119.22s eta: 9:28:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31655.0 tgs: 264 data_time: 0.73s time: 119.47s eta: 9:27:27 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.222 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30538.0 tgs: 256 data_time: 0.91s time: 118.98s eta: 9:23:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31787.0 tgs: 267 data_time: 0.85s time: 119.04s eta: 9:21:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30794.0 tgs: 257 data_time: 0.81s time: 119.53s eta: 9:21:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31343.0 tgs: 262 data_time: 0.73s time: 119.59s eta: 9:20:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31575.0 tgs: 265 data_time: 0.86s time: 119.13s eta: 9:15:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32184.0 tgs: 270 data_time: 0.98s time: 119.02s eta: 9:13:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31968.0 tgs: 268 data_time: 0.86s time: 118.89s eta: 9:10:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30803.0 tgs: 258 data_time: 0.68s time: 119.33s eta: 9:10:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31664.0 tgs: 265 data_time: 1.03s time: 119.09s eta: 9:07:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31162.0 tgs: 262 data_time: 0.71s time: 118.86s eta: 9:04:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 265 data_time: 0.83s time: 119.67s eta: 9:06:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31125.0 tgs: 260 data_time: 0.61s time: 119.44s eta: 9:03:27 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 267 data_time: 0.67s time: 119.43s eta: 9:01:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 270 data_time: 0.90s time: 118.75s eta: 8:56:21 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.204 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 269 data_time: 0.92s time: 118.86s eta: 8:54:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30719.0 tgs: 257 data_time: 0.92s time: 119.34s eta: 8:55:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32401.0 tgs: 271 data_time: 0.77s time: 119.15s eta: 8:52:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32313.0 tgs: 270 data_time: 0.62s time: 119.26s eta: 8:50:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.256 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32434.0 tgs: 269 data_time: 0.77s time: 120.47s eta: 8:54:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31492.0 tgs: 264 data_time: 0.81s time: 119.24s eta: 8:46:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31743.0 tgs: 266 data_time: 0.61s time: 119.04s eta: 8:43:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 261 data_time: 0.92s time: 120.83s eta: 8:49:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30313.0 tgs: 252 data_time: 0.74s time: 120.20s eta: 8:44:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 270 data_time: 0.79s time: 118.82s eta: 8:36:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32352.0 tgs: 271 data_time: 0.84s time: 119.35s eta: 8:37:12 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32118.0 tgs: 269 data_time: 0.79s time: 119.33s eta: 8:35:07 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.188 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31800.0 tgs: 267 data_time: 0.70s time: 119.05s eta: 8:31:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32278.0 tgs: 270 data_time: 0.68s time: 119.16s eta: 8:30:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32212.0 tgs: 270 data_time: 0.80s time: 119.04s eta: 8:27:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31620.0 tgs: 265 data_time: 0.89s time: 119.12s eta: 8:26:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31643.0 tgs: 265 data_time: 0.94s time: 119.25s eta: 8:24:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31306.0 tgs: 262 data_time: 0.99s time: 119.22s eta: 8:22:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.199 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 268 data_time: 1.17s time: 119.38s eta: 8:21:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31115.0 tgs: 259 data_time: 0.82s time: 119.78s eta: 8:21:04 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.107 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31672.0 tgs: 265 data_time: 0.87s time: 119.20s eta: 8:16:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31812.0 tgs: 267 data_time: 0.79s time: 118.98s eta: 8:13:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 257 data_time: 0.82s time: 119.91s eta: 8:15:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30483.0 tgs: 254 data_time: 0.62s time: 119.55s eta: 8:12:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30317.0 tgs: 253 data_time: 0.90s time: 119.59s eta: 8:10:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31943.0 tgs: 268 data_time: 0.64s time: 119.00s eta: 8:05:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31116.0 tgs: 259 data_time: 0.73s time: 119.84s eta: 8:07:21 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.214 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27629.0 tgs: 230 data_time: 0.76s time: 119.71s eta: 8:04:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32097.0 tgs: 268 data_time: 0.82s time: 119.41s eta: 8:01:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31895.0 tgs: 267 data_time: 0.57s time: 119.29s eta: 7:59:07 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31210.0 tgs: 261 data_time: 0.76s time: 119.19s eta: 7:56:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32432.0 tgs: 272 data_time: 0.71s time: 119.20s eta: 7:54:48 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32019.0 tgs: 269 data_time: 0.81s time: 118.94s eta: 7:51:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.172 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31680.0 tgs: 266 data_time: 0.86s time: 118.97s eta: 7:49:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.216 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 271 data_time: 1.23s time: 119.42s eta: 7:49:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31530.0 tgs: 264 data_time: 0.59s time: 119.04s eta: 7:46:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.113 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32008.0 tgs: 268 data_time: 0.74s time: 119.07s eta: 7:44:22 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32364.0 tgs: 269 data_time: 0.68s time: 120.00s eta: 7:46:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30695.0 tgs: 256 data_time: 0.74s time: 119.82s eta: 7:43:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.187 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32301.0 tgs: 271 data_time: 0.57s time: 119.17s eta: 7:38:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31351.0 tgs: 263 data_time: 0.80s time: 118.97s eta: 7:36:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31466.0 tgs: 263 data_time: 0.63s time: 119.33s eta: 7:35:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.172 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32340.0 tgs: 270 data_time: 0.83s time: 119.39s eta: 7:33:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31327.0 tgs: 263 data_time: 1.00s time: 119.07s eta: 7:30:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31370.0 tgs: 263 data_time: 0.86s time: 119.00s eta: 7:28:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31790.0 tgs: 266 data_time: 0.74s time: 119.23s eta: 7:27:07 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32322.0 tgs: 271 data_time: 0.86s time: 118.99s eta: 7:24:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 268 data_time: 0.86s time: 118.93s eta: 7:22:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.233 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32194.0 tgs: 270 data_time: 0.82s time: 119.08s eta: 7:20:35 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.169 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31253.0 tgs: 262 data_time: 0.99s time: 118.99s eta: 7:18:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31151.0 tgs: 261 data_time: 0.73s time: 119.03s eta: 7:16:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 268 data_time: 0.83s time: 119.25s eta: 7:15:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30135.0 tgs: 252 data_time: 0.71s time: 119.35s eta: 7:13:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 267 data_time: 0.78s time: 118.94s eta: 7:10:10 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31232.0 tgs: 261 data_time: 0.65s time: 119.21s eta: 7:09:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.174 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31162.0 tgs: 260 data_time: 0.79s time: 119.58s eta: 7:08:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 266 data_time: 0.86s time: 119.66s eta: 7:06:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31657.0 tgs: 266 data_time: 0.77s time: 118.96s eta: 7:02:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31308.0 tgs: 262 data_time: 0.73s time: 119.30s eta: 7:01:31 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:46:30][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 268 data_time: 0.76s time: 119.52s eta: 7:00:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.240 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31385.0 tgs: 262 data_time: 0.66s time: 119.43s eta: 6:58:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30623.0 tgs: 257 data_time: 0.82s time: 118.91s eta: 6:54:12 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31017.0 tgs: 259 data_time: 0.71s time: 119.34s eta: 6:53:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31151.0 tgs: 261 data_time: 0.85s time: 118.93s eta: 6:50:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.203 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32069.0 tgs: 268 data_time: 0.78s time: 119.27s eta: 6:49:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30882.0 tgs: 259 data_time: 0.65s time: 119.03s eta: 6:46:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31174.0 tgs: 261 data_time: 0.66s time: 119.21s eta: 6:45:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.105 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31824.0 tgs: 266 data_time: 0.64s time: 119.37s eta: 6:43:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30978.0 tgs: 260 data_time: 0.74s time: 119.01s eta: 6:40:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.213 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32005.0 tgs: 268 data_time: 1.01s time: 119.01s eta: 6:38:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 269 data_time: 0.68s time: 119.46s eta: 6:38:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.106 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31684.0 tgs: 266 data_time: 1.02s time: 118.81s eta: 6:34:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31911.0 tgs: 267 data_time: 0.74s time: 119.43s eta: 6:34:07 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31685.0 tgs: 264 data_time: 0.88s time: 119.58s eta: 6:32:36 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.210 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.92s time: 119.39s eta: 6:30:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31698.0 tgs: 265 data_time: 0.48s time: 119.56s eta: 6:28:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30622.0 tgs: 256 data_time: 0.65s time: 119.36s eta: 6:25:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31439.0 tgs: 264 data_time: 0.79s time: 118.83s eta: 6:22:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 269 data_time: 0.92s time: 119.02s eta: 6:20:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.188 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31740.0 tgs: 265 data_time: 0.80s time: 119.41s eta: 6:20:06 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31709.0 tgs: 265 data_time: 0.73s time: 119.34s eta: 6:17:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31568.0 tgs: 265 data_time: 0.81s time: 119.04s eta: 6:14:58 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31930.0 tgs: 267 data_time: 0.58s time: 119.38s eta: 6:14:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31134.0 tgs: 260 data_time: 0.71s time: 119.39s eta: 6:12:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31572.0 tgs: 264 data_time: 0.92s time: 119.28s eta: 6:09:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31923.0 tgs: 268 data_time: 0.67s time: 119.10s eta: 6:07:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.216 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30930.0 tgs: 253 data_time: 0.90s time: 121.96s eta: 6:14:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.114 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31469.0 tgs: 263 data_time: 0.87s time: 119.38s eta: 6:04:07 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 263 data_time: 0.74s time: 119.57s eta: 6:02:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 268 data_time: 0.85s time: 119.26s eta: 5:59:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.210 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30842.0 tgs: 258 data_time: 0.86s time: 119.26s eta: 5:57:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.203 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30791.0 tgs: 258 data_time: 0.87s time: 118.92s eta: 5:54:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32084.0 tgs: 269 data_time: 0.63s time: 119.01s eta: 5:53:04 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31337.0 tgs: 263 data_time: 0.71s time: 119.07s eta: 5:51:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30721.0 tgs: 257 data_time: 0.62s time: 119.29s eta: 5:49:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31601.0 tgs: 265 data_time: 0.59s time: 118.97s eta: 5:46:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.221 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31244.0 tgs: 261 data_time: 0.72s time: 119.27s eta: 5:45:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.236 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31356.0 tgs: 262 data_time: 0.61s time: 119.32s eta: 5:44:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31797.0 tgs: 267 data_time: 0.64s time: 119.05s eta: 5:41:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.099 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32197.0 tgs: 269 data_time: 0.82s time: 119.50s eta: 5:40:35 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32378.0 tgs: 271 data_time: 0.79s time: 119.38s eta: 5:38:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31323.0 tgs: 261 data_time: 0.70s time: 119.57s eta: 5:36:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 268 data_time: 0.72s time: 118.98s eta: 5:33:09 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.178 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31386.0 tgs: 263 data_time: 0.95s time: 118.95s eta: 5:31:04 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31859.0 tgs: 267 data_time: 0.79s time: 119.25s eta: 5:29:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31634.0 tgs: 264 data_time: 0.78s time: 119.40s eta: 5:28:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 264 data_time: 0.63s time: 119.46s eta: 5:26:31 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31642.0 tgs: 265 data_time: 0.67s time: 118.96s eta: 5:23:09 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31724.0 tgs: 265 data_time: 0.87s time: 119.63s eta: 5:23:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31170.0 tgs: 261 data_time: 0.73s time: 119.35s eta: 5:20:14 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31657.0 tgs: 265 data_time: 0.70s time: 119.12s eta: 5:17:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31409.0 tgs: 263 data_time: 0.70s time: 119.42s eta: 5:16:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31619.0 tgs: 264 data_time: 0.68s time: 119.57s eta: 5:14:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31695.0 tgs: 266 data_time: 0.68s time: 119.01s eta: 5:11:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.094 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31184.0 tgs: 262 data_time: 0.81s time: 119.00s eta: 5:09:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31733.0 tgs: 266 data_time: 1.00s time: 119.23s eta: 5:08:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.191 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 259 data_time: 0.68s time: 119.68s eta: 5:07:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.90s time: 119.11s eta: 5:03:44 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.179 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29736.0 tgs: 249 data_time: 0.52s time: 119.07s eta: 5:01:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30589.0 tgs: 251 data_time: 0.73s time: 121.52s eta: 5:05:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31327.0 tgs: 261 data_time: 0.59s time: 119.58s eta: 4:58:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31651.0 tgs: 264 data_time: 0.69s time: 119.50s eta: 4:56:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30981.0 tgs: 258 data_time: 0.91s time: 119.73s eta: 4:55:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.188 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32259.0 tgs: 269 data_time: 0.62s time: 119.52s eta: 4:52:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31235.0 tgs: 261 data_time: 0.65s time: 119.42s eta: 4:50:35 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.176 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31823.0 tgs: 266 data_time: 0.86s time: 119.20s eta: 4:48:04 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.110 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32147.0 tgs: 270 data_time: 0.68s time: 118.89s eta: 4:45:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31157.0 tgs: 261 data_time: 0.70s time: 119.04s eta: 4:43:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32556.0 tgs: 271 data_time: 0.77s time: 119.98s eta: 4:43:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.207 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 266 data_time: 0.81s time: 119.15s eta: 4:40:00 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30815.0 tgs: 258 data_time: 0.57s time: 119.01s eta: 4:37:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30824.0 tgs: 258 data_time: 0.77s time: 119.06s eta: 4:35:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.175 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31422.0 tgs: 263 data_time: 0.72s time: 119.38s eta: 4:34:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31278.0 tgs: 261 data_time: 0.63s time: 119.42s eta: 4:32:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32246.0 tgs: 271 data_time: 0.86s time: 118.93s eta: 4:29:34 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32500.0 tgs: 271 data_time: 0.88s time: 119.60s eta: 4:29:05 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31312.0 tgs: 262 data_time: 0.74s time: 119.24s eta: 4:26:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.272 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32090.0 tgs: 269 data_time: 1.00s time: 119.06s eta: 4:23:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32063.0 tgs: 269 data_time: 0.80s time: 118.85s eta: 4:21:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.126 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30082.0 tgs: 253 data_time: 0.87s time: 118.76s eta: 4:19:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31216.0 tgs: 261 data_time: 0.71s time: 119.43s eta: 4:18:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31636.0 tgs: 264 data_time: 0.54s time: 119.40s eta: 4:16:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31837.0 tgs: 266 data_time: 0.69s time: 119.32s eta: 4:14:32 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32423.0 tgs: 272 data_time: 0.71s time: 119.05s eta: 4:11:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32034.0 tgs: 268 data_time: 0.70s time: 119.13s eta: 4:10:10 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.189 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31699.0 tgs: 266 data_time: 0.90s time: 119.13s eta: 4:08:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.114 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31731.0 tgs: 266 data_time: 0.78s time: 118.96s eta: 4:05:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31173.0 tgs: 261 data_time: 0.72s time: 119.15s eta: 4:04:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30948.0 tgs: 258 data_time: 0.67s time: 119.50s eta: 4:02:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31499.0 tgs: 264 data_time: 0.86s time: 119.25s eta: 4:00:28 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31743.0 tgs: 266 data_time: 0.53s time: 119.19s eta: 3:58:23 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32189.0 tgs: 269 data_time: 0.62s time: 119.36s eta: 3:56:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31398.0 tgs: 263 data_time: 0.75s time: 119.37s eta: 3:54:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.189 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30454.0 tgs: 253 data_time: 0.55s time: 119.95s eta: 3:53:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31934.0 tgs: 267 data_time: 0.52s time: 119.32s eta: 3:50:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 262 data_time: 0.63s time: 119.64s eta: 3:49:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32449.0 tgs: 273 data_time: 0.71s time: 118.86s eta: 3:45:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32082.0 tgs: 268 data_time: 0.74s time: 119.70s eta: 3:45:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 268 data_time: 0.89s time: 119.33s eta: 3:42:45 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31845.0 tgs: 267 data_time: 0.44s time: 119.27s eta: 3:40:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31823.0 tgs: 267 data_time: 0.80s time: 119.00s eta: 3:38:10 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.106 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31240.0 tgs: 261 data_time: 0.74s time: 119.43s eta: 3:36:57 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31848.0 tgs: 267 data_time: 0.70s time: 119.03s eta: 3:34:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31764.0 tgs: 266 data_time: 0.84s time: 118.99s eta: 3:32:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31847.0 tgs: 267 data_time: 0.78s time: 119.22s eta: 3:30:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32068.0 tgs: 266 data_time: 0.59s time: 120.11s eta: 3:30:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 269 data_time: 0.92s time: 119.13s eta: 3:26:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.204 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 268 data_time: 0.65s time: 119.32s eta: 3:24:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31657.0 tgs: 265 data_time: 0.70s time: 119.08s eta: 3:22:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.204 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31659.0 tgs: 265 data_time: 1.00s time: 119.43s eta: 3:21:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31541.0 tgs: 264 data_time: 0.62s time: 119.29s eta: 3:18:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30513.0 tgs: 255 data_time: 0.78s time: 119.22s eta: 3:16:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31295.0 tgs: 262 data_time: 0.90s time: 119.13s eta: 3:14:34 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 267 data_time: 0.45s time: 119.02s eta: 3:12:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.111 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32171.0 tgs: 269 data_time: 0.81s time: 119.19s eta: 3:10:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32370.0 tgs: 267 data_time: 0.76s time: 121.07s eta: 3:11:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30920.0 tgs: 258 data_time: 0.65s time: 119.50s eta: 3:07:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32211.0 tgs: 270 data_time: 0.96s time: 119.27s eta: 3:04:51 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31530.0 tgs: 264 data_time: 0.52s time: 118.98s eta: 3:02:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.114 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 267 data_time: 0.77s time: 119.22s eta: 3:00:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31744.0 tgs: 265 data_time: 0.73s time: 119.46s eta: 2:59:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.229 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32031.0 tgs: 269 data_time: 1.03s time: 118.98s eta: 2:56:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 265 data_time: 0.79s time: 119.35s eta: 2:55:03 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31643.0 tgs: 264 data_time: 0.87s time: 119.41s eta: 2:53:08 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.200 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31911.0 tgs: 267 data_time: 0.90s time: 119.36s eta: 2:51:04 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32067.0 tgs: 264 data_time: 0.82s time: 121.46s eta: 2:52:04 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 264 data_time: 0.69s time: 119.24s eta: 2:46:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31845.0 tgs: 266 data_time: 0.70s time: 119.54s eta: 2:45:22 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31181.0 tgs: 262 data_time: 0.74s time: 118.91s eta: 2:42:30 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31734.0 tgs: 266 data_time: 0.70s time: 119.18s eta: 2:40:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31980.0 tgs: 268 data_time: 0.57s time: 118.97s eta: 2:38:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 270 data_time: 0.84s time: 119.23s eta: 2:36:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31151.0 tgs: 261 data_time: 0.69s time: 119.11s eta: 2:34:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31354.0 tgs: 262 data_time: 0.81s time: 119.43s eta: 2:33:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30460.0 tgs: 253 data_time: 0.75s time: 120.23s eta: 2:32:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 262 data_time: 0.81s time: 119.28s eta: 2:29:06 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31902.0 tgs: 268 data_time: 0.82s time: 118.88s eta: 2:26:37 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 264 data_time: 0.60s time: 119.52s eta: 2:25:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.110 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31751.0 tgs: 266 data_time: 0.65s time: 119.07s eta: 2:22:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32194.0 tgs: 270 data_time: 0.68s time: 119.14s eta: 2:20:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31143.0 tgs: 261 data_time: 0.64s time: 119.05s eta: 2:18:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31451.0 tgs: 262 data_time: 1.14s time: 119.85s eta: 2:17:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32236.0 tgs: 270 data_time: 0.94s time: 119.31s eta: 2:15:12 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31733.0 tgs: 265 data_time: 0.52s time: 119.39s eta: 2:13:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32271.0 tgs: 269 data_time: 0.68s time: 119.68s eta: 2:11:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.235 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31812.0 tgs: 266 data_time: 0.73s time: 119.50s eta: 2:09:27 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31345.0 tgs: 263 data_time: 0.64s time: 118.96s eta: 2:06:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32208.0 tgs: 269 data_time: 0.77s time: 119.38s eta: 2:05:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32034.0 tgs: 268 data_time: 0.77s time: 119.28s eta: 2:03:15 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31919.0 tgs: 267 data_time: 0.89s time: 119.24s eta: 2:01:13 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31755.0 tgs: 265 data_time: 0.63s time: 119.50s eta: 1:59:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 268 data_time: 0.91s time: 119.42s eta: 1:57:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30763.0 tgs: 257 data_time: 0.72s time: 119.54s eta: 1:55:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30975.0 tgs: 259 data_time: 0.75s time: 119.30s eta: 1:53:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31884.0 tgs: 267 data_time: 0.80s time: 119.25s eta: 1:51:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32717.0 tgs: 275 data_time: 0.76s time: 118.93s eta: 1:49:01 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31060.0 tgs: 259 data_time: 0.62s time: 119.65s eta: 1:47:41 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31628.0 tgs: 264 data_time: 1.08s time: 119.68s eta: 1:45:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31790.0 tgs: 266 data_time: 0.78s time: 119.34s eta: 1:43:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31093.0 tgs: 260 data_time: 0.67s time: 119.31s eta: 1:41:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31460.0 tgs: 263 data_time: 0.63s time: 119.28s eta: 1:39:24 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31548.0 tgs: 265 data_time: 0.72s time: 118.91s eta: 1:37:06 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31262.0 tgs: 262 data_time: 0.76s time: 119.16s eta: 1:35:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31884.0 tgs: 267 data_time: 0.72s time: 119.07s eta: 1:33:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32320.0 tgs: 271 data_time: 0.92s time: 119.26s eta: 1:31:25 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31061.0 tgs: 259 data_time: 0.76s time: 119.72s eta: 1:29:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31479.0 tgs: 264 data_time: 0.70s time: 119.02s eta: 1:27:16 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31756.0 tgs: 267 data_time: 0.56s time: 118.87s eta: 1:25:11 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31951.0 tgs: 267 data_time: 1.00s time: 119.48s eta: 1:23:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32022.0 tgs: 267 data_time: 0.85s time: 119.49s eta: 1:21:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32116.0 tgs: 270 data_time: 0.85s time: 118.94s eta: 1:19:17 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 268 data_time: 0.80s time: 118.93s eta: 1:17:18 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32443.0 tgs: 265 data_time: 0.68s time: 122.13s eta: 1:17:20 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.256 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31437.0 tgs: 263 data_time: 0.69s time: 119.35s eta: 1:13:35 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32125.0 tgs: 269 data_time: 0.77s time: 119.06s eta: 1:11:26 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31143.0 tgs: 261 data_time: 0.82s time: 118.92s eta: 1:09:22 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.196 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31354.0 tgs: 263 data_time: 0.74s time: 119.11s eta: 1:07:29 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32238.0 tgs: 269 data_time: 0.71s time: 119.42s eta: 1:05:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.72s time: 119.35s eta: 1:03:39 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31718.0 tgs: 266 data_time: 0.66s time: 119.18s eta: 1:01:34 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32010.0 tgs: 268 data_time: 0.88s time: 119.28s eta: 0:59:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32284.0 tgs: 270 data_time: 0.77s time: 119.44s eta: 0:57:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31535.0 tgs: 264 data_time: 0.92s time: 119.39s eta: 0:55:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 264 data_time: 0.70s time: 119.49s eta: 0:53:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31932.0 tgs: 267 data_time: 0.83s time: 119.37s eta: 0:51:43 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 268 data_time: 0.91s time: 119.47s eta: 0:49:46 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31612.0 tgs: 265 data_time: 0.86s time: 118.90s eta: 0:47:33 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31583.0 tgs: 264 data_time: 0.58s time: 119.26s eta: 0:45:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31801.0 tgs: 266 data_time: 0.77s time: 119.19s eta: 0:43:42 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32207.0 tgs: 270 data_time: 0.45s time: 118.99s eta: 0:41:38 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31280.0 tgs: 262 data_time: 0.98s time: 119.04s eta: 0:39:40 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 265 data_time: 0.71s time: 119.66s eta: 0:37:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 266 data_time: 0.62s time: 119.29s eta: 0:35:47 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.211 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31941.0 tgs: 266 data_time: 0.82s time: 119.75s eta: 0:33:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 263 data_time: 0.83s time: 121.23s eta: 0:32:19 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29970.0 tgs: 251 data_time: 0.41s time: 119.37s eta: 0:29:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31575.0 tgs: 264 data_time: 0.77s time: 119.32s eta: 0:27:50 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30964.0 tgs: 258 data_time: 0.59s time: 119.64s eta: 0:25:55 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30870.0 tgs: 259 data_time: 0.58s time: 119.10s eta: 0:23:49 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31970.0 tgs: 267 data_time: 0.67s time: 119.39s eta: 0:21:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31745.0 tgs: 266 data_time: 0.79s time: 119.26s eta: 0:19:52 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 268 data_time: 0.66s time: 119.25s eta: 0:17:53 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 265 data_time: 0.57s time: 119.90s eta: 0:15:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31585.0 tgs: 265 data_time: 0.67s time: 119.18s eta: 0:13:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31779.0 tgs: 266 data_time: 0.78s time: 119.09s eta: 0:11:54 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.72s time: 119.32s eta: 0:09:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31664.0 tgs: 265 data_time: 0.74s time: 119.17s eta: 0:07:56 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30712.0 tgs: 256 data_time: 0.65s time: 119.69s eta: 0:05:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31311.0 tgs: 257 data_time: 0.75s time: 121.45s eta: 0:04:02 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31199.0 tgs: 261 data_time: 0.72s time: 119.21s eta: 0:01:59 +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.29 seconds, peak gpu memory 17.4G +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 39][DP 9][SP 3][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank4.log b/20250121165312/rank4.log new file mode 100644 index 0000000000000000000000000000000000000000..b7fb0b46a10b8393127e5f6da6c8f94230630de6 --- /dev/null +++ b/20250121165312/rank4.log @@ -0,0 +1,851 @@ +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.05s +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:00:44][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 69.91 seconds, peak gpu memory 13.4G +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.248 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.6GB text_tokens: 31121.0 tgs: 202 data_time: 2.19s time: 153.42s eta: 1 day, 1:16:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.340 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31200.0 tgs: 259 data_time: 0.84s time: 120.13s eta: 19:45:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.221 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30861.0 tgs: 257 data_time: 0.94s time: 119.68s eta: 19:38:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.267 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31960.0 tgs: 267 data_time: 0.94s time: 119.65s eta: 19:36:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.228 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31620.0 tgs: 265 data_time: 1.16s time: 119.30s eta: 19:31:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.208 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31899.0 tgs: 266 data_time: 0.74s time: 119.54s eta: 19:31:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32173.0 tgs: 261 data_time: 0.98s time: 122.96s eta: 20:02:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.222 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32073.0 tgs: 267 data_time: 0.74s time: 119.89s eta: 19:30:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.207 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32230.0 tgs: 267 data_time: 0.81s time: 120.65s eta: 19:36:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32211.0 tgs: 268 data_time: 0.70s time: 119.92s eta: 19:27:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31612.0 tgs: 262 data_time: 0.86s time: 120.46s eta: 19:30:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.207 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31664.0 tgs: 265 data_time: 0.66s time: 119.34s eta: 19:17:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31279.0 tgs: 261 data_time: 0.67s time: 119.67s eta: 19:18:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.236 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31574.0 tgs: 264 data_time: 0.87s time: 119.21s eta: 19:12:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31608.0 tgs: 265 data_time: 0.72s time: 119.22s eta: 19:10:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32396.0 tgs: 272 data_time: 0.85s time: 118.93s eta: 19:05:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32445.0 tgs: 272 data_time: 0.75s time: 119.14s eta: 19:05:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32483.0 tgs: 272 data_time: 0.85s time: 119.18s eta: 19:04:06 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30947.0 tgs: 259 data_time: 0.87s time: 119.39s eta: 19:04:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.336 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32390.0 tgs: 271 data_time: 0.75s time: 119.33s eta: 19:01:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31866.0 tgs: 264 data_time: 0.85s time: 120.31s eta: 19:09:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31128.0 tgs: 253 data_time: 0.80s time: 122.91s eta: 19:31:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32297.0 tgs: 270 data_time: 0.67s time: 119.48s eta: 18:57:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.272 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32495.0 tgs: 272 data_time: 1.01s time: 119.16s eta: 18:51:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31533.0 tgs: 264 data_time: 0.70s time: 119.24s eta: 18:50:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31074.0 tgs: 260 data_time: 0.73s time: 119.39s eta: 18:50:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32081.0 tgs: 268 data_time: 0.84s time: 119.66s eta: 18:50:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31457.0 tgs: 263 data_time: 0.78s time: 119.38s eta: 18:46:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31099.0 tgs: 259 data_time: 0.70s time: 119.79s eta: 18:48:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32202.0 tgs: 249 data_time: 0.74s time: 129.32s eta: 20:15:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30996.0 tgs: 259 data_time: 0.70s time: 119.57s eta: 18:41:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31401.0 tgs: 263 data_time: 0.91s time: 119.38s eta: 18:38:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.311 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31506.0 tgs: 261 data_time: 0.82s time: 120.66s eta: 18:48:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31849.0 tgs: 266 data_time: 0.79s time: 119.39s eta: 18:34:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31503.0 tgs: 264 data_time: 0.63s time: 119.16s eta: 18:30:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 270 data_time: 0.75s time: 119.07s eta: 18:27:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 264 data_time: 0.88s time: 119.13s eta: 18:25:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29886.0 tgs: 248 data_time: 0.94s time: 120.39s eta: 18:35:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 266 data_time: 0.76s time: 121.17s eta: 18:40:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31388.0 tgs: 260 data_time: 0.85s time: 120.66s eta: 18:34:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.246 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31748.0 tgs: 265 data_time: 0.79s time: 119.73s eta: 18:23:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31342.0 tgs: 253 data_time: 0.67s time: 123.87s eta: 18:59:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31840.0 tgs: 266 data_time: 0.85s time: 119.56s eta: 18:17:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31789.0 tgs: 266 data_time: 0.63s time: 119.36s eta: 18:14:06 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31941.0 tgs: 267 data_time: 0.78s time: 119.22s eta: 18:10:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30569.0 tgs: 255 data_time: 0.72s time: 119.60s eta: 18:12:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31307.0 tgs: 262 data_time: 0.82s time: 119.27s eta: 18:07:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31568.0 tgs: 265 data_time: 0.75s time: 119.01s eta: 18:02:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31707.0 tgs: 266 data_time: 1.05s time: 119.10s eta: 18:01:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31866.0 tgs: 267 data_time: 0.87s time: 119.26s eta: 18:01:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.258 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 264 data_time: 0.89s time: 119.33s eta: 17:59:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.288 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 265 data_time: 0.92s time: 120.02s eta: 18:04:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30551.0 tgs: 256 data_time: 0.58s time: 119.10s eta: 17:53:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31961.0 tgs: 267 data_time: 0.80s time: 119.41s eta: 17:54:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.264 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31763.0 tgs: 266 data_time: 0.98s time: 118.99s eta: 17:48:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31917.0 tgs: 268 data_time: 1.01s time: 119.03s eta: 17:47:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31804.0 tgs: 253 data_time: 0.69s time: 125.36s eta: 18:41:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 268 data_time: 0.97s time: 120.10s eta: 17:52:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 267 data_time: 0.79s time: 119.39s eta: 17:44:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 262 data_time: 0.96s time: 119.34s eta: 17:42:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32214.0 tgs: 269 data_time: 0.81s time: 119.51s eta: 17:41:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31816.0 tgs: 252 data_time: 0.90s time: 126.03s eta: 18:37:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31958.0 tgs: 260 data_time: 0.60s time: 122.65s eta: 18:05:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31545.0 tgs: 260 data_time: 0.97s time: 121.16s eta: 17:50:12 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31565.0 tgs: 265 data_time: 0.75s time: 119.03s eta: 17:29:27 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31753.0 tgs: 267 data_time: 0.70s time: 118.86s eta: 17:25:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30827.0 tgs: 248 data_time: 0.72s time: 123.99s eta: 18:09:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31476.0 tgs: 252 data_time: 1.06s time: 124.57s eta: 18:12:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31389.0 tgs: 262 data_time: 0.43s time: 119.39s eta: 17:24:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.121 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30909.0 tgs: 258 data_time: 0.89s time: 119.70s eta: 17:25:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32110.0 tgs: 268 data_time: 0.80s time: 119.53s eta: 17:21:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31386.0 tgs: 264 data_time: 0.98s time: 118.86s eta: 17:14:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31265.0 tgs: 260 data_time: 0.76s time: 119.85s eta: 17:20:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31690.0 tgs: 265 data_time: 0.85s time: 119.56s eta: 17:16:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31628.0 tgs: 259 data_time: 0.74s time: 121.86s eta: 17:34:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.278 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30827.0 tgs: 249 data_time: 0.87s time: 123.51s eta: 17:46:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 267 data_time: 0.77s time: 119.01s eta: 17:05:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31929.0 tgs: 267 data_time: 0.55s time: 119.23s eta: 17:05:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31847.0 tgs: 267 data_time: 0.63s time: 119.14s eta: 17:02:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30582.0 tgs: 256 data_time: 0.44s time: 119.25s eta: 17:01:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 268 data_time: 1.00s time: 119.04s eta: 16:57:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.131 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 266 data_time: 0.89s time: 119.28s eta: 16:57:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31980.0 tgs: 268 data_time: 0.61s time: 119.16s eta: 16:54:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31403.0 tgs: 262 data_time: 0.66s time: 119.58s eta: 16:56:27 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31893.0 tgs: 267 data_time: 0.73s time: 119.13s eta: 16:50:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.208 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31919.0 tgs: 268 data_time: 0.73s time: 118.84s eta: 16:46:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31438.0 tgs: 264 data_time: 0.62s time: 119.01s eta: 16:45:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32392.0 tgs: 270 data_time: 0.63s time: 119.85s eta: 16:50:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31675.0 tgs: 265 data_time: 0.77s time: 119.40s eta: 16:44:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31027.0 tgs: 260 data_time: 0.93s time: 119.01s eta: 16:39:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31959.0 tgs: 264 data_time: 0.69s time: 120.92s eta: 16:53:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31398.0 tgs: 263 data_time: 0.68s time: 118.99s eta: 16:35:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30730.0 tgs: 257 data_time: 0.56s time: 119.28s eta: 16:36:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 267 data_time: 0.84s time: 119.23s eta: 16:33:34 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30756.0 tgs: 257 data_time: 0.71s time: 119.37s eta: 16:32:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31124.0 tgs: 260 data_time: 0.86s time: 119.70s eta: 16:33:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31870.0 tgs: 267 data_time: 0.82s time: 119.20s eta: 16:27:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 267 data_time: 0.66s time: 119.76s eta: 16:30:02 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32040.0 tgs: 266 data_time: 0.85s time: 120.23s eta: 16:31:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31523.0 tgs: 264 data_time: 0.70s time: 118.97s eta: 16:19:31 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.207 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31958.0 tgs: 250 data_time: 0.80s time: 127.33s eta: 17:26:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31178.0 tgs: 261 data_time: 0.84s time: 119.22s eta: 16:17:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31326.0 tgs: 262 data_time: 0.63s time: 119.14s eta: 16:14:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.232 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31894.0 tgs: 267 data_time: 0.66s time: 119.24s eta: 16:13:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32095.0 tgs: 269 data_time: 0.70s time: 119.17s eta: 16:11:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32095.0 tgs: 269 data_time: 0.71s time: 118.95s eta: 16:07:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 260 data_time: 0.72s time: 118.86s eta: 16:04:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32098.0 tgs: 269 data_time: 0.70s time: 118.96s eta: 16:03:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32395.0 tgs: 265 data_time: 0.71s time: 121.84s eta: 16:24:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.80s time: 118.99s eta: 15:59:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31865.0 tgs: 267 data_time: 0.89s time: 119.22s eta: 15:59:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32108.0 tgs: 268 data_time: 0.58s time: 119.62s eta: 16:00:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:48:49][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.242 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31679.0 tgs: 266 data_time: 0.86s time: 119.04s eta: 15:54:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31865.0 tgs: 267 data_time: 0.81s time: 118.98s eta: 15:51:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 271 data_time: 0.52s time: 119.17s eta: 15:51:22 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32210.0 tgs: 269 data_time: 0.76s time: 119.64s eta: 15:53:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 267 data_time: 1.00s time: 118.93s eta: 15:45:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31383.0 tgs: 263 data_time: 0.90s time: 119.28s eta: 15:46:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.128 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31380.0 tgs: 263 data_time: 0.63s time: 119.14s eta: 15:43:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.240 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30738.0 tgs: 258 data_time: 0.81s time: 119.01s eta: 15:40:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30780.0 tgs: 257 data_time: 0.69s time: 119.65s eta: 15:43:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31099.0 tgs: 255 data_time: 0.75s time: 121.50s eta: 15:55:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30077.0 tgs: 251 data_time: 0.90s time: 119.41s eta: 15:37:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.124 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 244 data_time: 0.54s time: 129.97s eta: 16:58:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32419.0 tgs: 271 data_time: 1.02s time: 119.47s eta: 15:33:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30440.0 tgs: 254 data_time: 0.60s time: 119.60s eta: 15:32:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32007.0 tgs: 267 data_time: 0.74s time: 119.54s eta: 15:30:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31413.0 tgs: 257 data_time: 0.77s time: 122.12s eta: 15:48:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30763.0 tgs: 254 data_time: 0.96s time: 120.75s eta: 15:35:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31478.0 tgs: 263 data_time: 1.06s time: 119.41s eta: 15:23:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31029.0 tgs: 259 data_time: 0.89s time: 119.44s eta: 15:21:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31717.0 tgs: 240 data_time: 1.00s time: 131.63s eta: 16:53:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31425.0 tgs: 263 data_time: 0.82s time: 119.29s eta: 15:16:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.213 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31450.0 tgs: 263 data_time: 0.59s time: 119.56s eta: 15:16:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30888.0 tgs: 252 data_time: 0.62s time: 122.25s eta: 15:35:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31977.0 tgs: 267 data_time: 0.71s time: 119.71s eta: 15:13:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.098 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31730.0 tgs: 265 data_time: 0.65s time: 119.35s eta: 15:09:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32646.0 tgs: 274 data_time: 0.83s time: 118.90s eta: 15:03:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30572.0 tgs: 256 data_time: 0.69s time: 118.99s eta: 15:02:22 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 267 data_time: 0.64s time: 119.31s eta: 15:02:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 267 data_time: 0.77s time: 118.92s eta: 14:57:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.238 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32139.0 tgs: 266 data_time: 0.63s time: 120.53s eta: 15:07:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.122 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31865.0 tgs: 267 data_time: 0.52s time: 119.30s eta: 14:56:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.239 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31565.0 tgs: 257 data_time: 0.93s time: 122.65s eta: 15:19:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32240.0 tgs: 270 data_time: 0.93s time: 119.12s eta: 14:51:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31375.0 tgs: 261 data_time: 0.54s time: 119.75s eta: 14:54:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31901.0 tgs: 267 data_time: 1.07s time: 119.46s eta: 14:50:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.235 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31560.0 tgs: 265 data_time: 1.06s time: 118.99s eta: 14:44:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 267 data_time: 0.82s time: 119.35s eta: 14:45:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.103 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31745.0 tgs: 266 data_time: 0.55s time: 118.94s eta: 14:40:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.209 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32466.0 tgs: 272 data_time: 0.78s time: 119.05s eta: 14:38:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.296 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32075.0 tgs: 269 data_time: 0.89s time: 119.06s eta: 14:37:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31204.0 tgs: 261 data_time: 0.64s time: 119.12s eta: 14:35:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.115 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32130.0 tgs: 270 data_time: 0.78s time: 119.00s eta: 14:32:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.117 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31219.0 tgs: 262 data_time: 0.68s time: 119.07s eta: 14:31:12 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31694.0 tgs: 266 data_time: 0.67s time: 119.15s eta: 14:29:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.113 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29250.0 tgs: 244 data_time: 0.74s time: 119.65s eta: 14:31:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32365.0 tgs: 271 data_time: 0.78s time: 119.18s eta: 14:26:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.66s time: 119.20s eta: 14:24:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31236.0 tgs: 261 data_time: 0.68s time: 119.39s eta: 14:23:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 267 data_time: 0.79s time: 120.73s eta: 14:31:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32092.0 tgs: 270 data_time: 0.79s time: 118.85s eta: 14:15:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31177.0 tgs: 261 data_time: 0.93s time: 119.10s eta: 14:15:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29651.0 tgs: 248 data_time: 0.81s time: 119.15s eta: 14:13:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.309 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31710.0 tgs: 264 data_time: 0.85s time: 119.95s eta: 14:17:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31200.0 tgs: 260 data_time: 0.84s time: 119.87s eta: 14:15:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 264 data_time: 0.43s time: 119.59s eta: 14:11:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31725.0 tgs: 266 data_time: 0.56s time: 119.21s eta: 14:06:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31653.0 tgs: 265 data_time: 0.78s time: 119.26s eta: 14:04:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 28771.0 tgs: 241 data_time: 0.46s time: 119.27s eta: 14:02:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30512.0 tgs: 256 data_time: 0.71s time: 119.14s eta: 13:59:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31814.0 tgs: 267 data_time: 0.70s time: 119.12s eta: 13:57:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 268 data_time: 0.64s time: 119.17s eta: 13:56:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 0.80s time: 119.34s eta: 13:55:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.216 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31806.0 tgs: 264 data_time: 0.54s time: 120.11s eta: 13:58:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.180 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30861.0 tgs: 251 data_time: 0.54s time: 122.55s eta: 14:13:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:56:46][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31496.0 tgs: 264 data_time: 0.69s time: 119.10s eta: 13:47:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 268 data_time: 0.85s time: 119.37s eta: 13:47:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32122.0 tgs: 269 data_time: 1.11s time: 119.35s eta: 13:45:31 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.117 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31538.0 tgs: 264 data_time: 0.73s time: 119.34s eta: 13:43:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.262 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32418.0 tgs: 272 data_time: 1.06s time: 119.11s eta: 13:39:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.257 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31784.0 tgs: 266 data_time: 0.93s time: 119.25s eta: 13:38:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.235 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31232.0 tgs: 262 data_time: 0.64s time: 119.13s eta: 13:36:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.172 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30502.0 tgs: 255 data_time: 0.72s time: 119.22s eta: 13:34:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31234.0 tgs: 261 data_time: 0.79s time: 119.46s eta: 13:34:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 269 data_time: 0.96s time: 119.08s eta: 13:29:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31706.0 tgs: 266 data_time: 0.74s time: 118.92s eta: 13:26:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31733.0 tgs: 264 data_time: 0.88s time: 119.81s eta: 13:30:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.208 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31375.0 tgs: 263 data_time: 0.97s time: 119.25s eta: 13:24:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31316.0 tgs: 262 data_time: 0.83s time: 119.16s eta: 13:22:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31743.0 tgs: 265 data_time: 0.73s time: 119.35s eta: 13:21:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32176.0 tgs: 269 data_time: 0.89s time: 119.26s eta: 13:19:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31482.0 tgs: 264 data_time: 0.90s time: 119.04s eta: 13:15:34 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.239 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31244.0 tgs: 261 data_time: 0.70s time: 119.37s eta: 13:15:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31400.0 tgs: 263 data_time: 0.90s time: 119.19s eta: 13:12:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32223.0 tgs: 269 data_time: 0.84s time: 119.46s eta: 13:12:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.193 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32167.0 tgs: 269 data_time: 0.84s time: 119.28s eta: 13:09:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32252.0 tgs: 270 data_time: 0.82s time: 119.15s eta: 13:06:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31860.0 tgs: 264 data_time: 0.93s time: 120.43s eta: 13:12:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32295.0 tgs: 268 data_time: 0.89s time: 120.16s eta: 13:09:02 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 269 data_time: 1.06s time: 119.50s eta: 13:02:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.65s time: 119.51s eta: 13:00:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 267 data_time: 0.63s time: 119.57s eta: 12:59:12 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.121 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32469.0 tgs: 271 data_time: 0.90s time: 119.79s eta: 12:58:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31212.0 tgs: 262 data_time: 0.73s time: 119.04s eta: 12:51:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 270 data_time: 0.95s time: 119.05s eta: 12:49:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31458.0 tgs: 264 data_time: 0.78s time: 119.12s eta: 12:48:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31682.0 tgs: 265 data_time: 0.49s time: 119.39s eta: 12:48:06 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30946.0 tgs: 259 data_time: 0.72s time: 119.20s eta: 12:44:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.202 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30503.0 tgs: 255 data_time: 0.89s time: 119.44s eta: 12:44:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31664.0 tgs: 265 data_time: 0.75s time: 119.09s eta: 12:40:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32465.0 tgs: 272 data_time: 0.68s time: 119.07s eta: 12:38:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30539.0 tgs: 255 data_time: 0.64s time: 119.59s eta: 12:39:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 266 data_time: 0.50s time: 118.78s eta: 12:32:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 263 data_time: 0.86s time: 118.88s eta: 12:30:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31891.0 tgs: 267 data_time: 0.62s time: 119.17s eta: 12:30:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31597.0 tgs: 264 data_time: 0.97s time: 119.27s eta: 12:29:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31596.0 tgs: 264 data_time: 0.62s time: 119.35s eta: 12:27:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.194 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 265 data_time: 0.54s time: 119.11s eta: 12:24:27 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31693.0 tgs: 265 data_time: 1.05s time: 119.34s eta: 12:23:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31275.0 tgs: 262 data_time: 0.81s time: 119.26s eta: 12:21:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31414.0 tgs: 263 data_time: 0.71s time: 119.03s eta: 12:17:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31884.0 tgs: 263 data_time: 0.70s time: 121.13s eta: 12:28:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31301.0 tgs: 260 data_time: 0.76s time: 120.00s eta: 12:19:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32116.0 tgs: 270 data_time: 0.68s time: 118.72s eta: 12:10:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 268 data_time: 0.68s time: 119.63s eta: 12:13:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 268 data_time: 0.70s time: 119.82s eta: 12:12:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.121 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31232.0 tgs: 262 data_time: 0.60s time: 119.14s eta: 12:06:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.124 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31364.0 tgs: 263 data_time: 0.96s time: 119.17s eta: 12:04:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30624.0 tgs: 257 data_time: 0.99s time: 119.11s eta: 12:02:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31739.0 tgs: 266 data_time: 0.95s time: 119.18s eta: 12:01:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 269 data_time: 0.86s time: 118.95s eta: 11:57:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.107 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.69s time: 119.72s eta: 12:00:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32679.0 tgs: 274 data_time: 0.79s time: 119.26s eta: 11:55:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31804.0 tgs: 266 data_time: 0.68s time: 119.27s eta: 11:53:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.187 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32111.0 tgs: 268 data_time: 1.07s time: 119.72s eta: 11:54:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31402.0 tgs: 264 data_time: 0.88s time: 118.77s eta: 11:46:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 267 data_time: 0.64s time: 119.16s eta: 11:46:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32371.0 tgs: 270 data_time: 0.76s time: 119.61s eta: 11:47:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32279.0 tgs: 271 data_time: 0.72s time: 118.87s eta: 11:41:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31716.0 tgs: 267 data_time: 0.87s time: 118.68s eta: 11:38:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.114 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31294.0 tgs: 262 data_time: 0.83s time: 119.22s eta: 11:39:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31867.0 tgs: 267 data_time: 0.68s time: 119.14s eta: 11:36:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32098.0 tgs: 269 data_time: 0.99s time: 119.08s eta: 11:34:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.182 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31651.0 tgs: 266 data_time: 0.78s time: 118.94s eta: 11:31:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.194 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31370.0 tgs: 263 data_time: 0.49s time: 118.91s eta: 11:29:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31139.0 tgs: 261 data_time: 0.92s time: 118.98s eta: 11:28:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.186 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31723.0 tgs: 265 data_time: 0.68s time: 119.28s eta: 11:27:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.191 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31960.0 tgs: 267 data_time: 0.74s time: 119.36s eta: 11:26:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31081.0 tgs: 261 data_time: 0.75s time: 118.88s eta: 11:21:34 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.245 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.67s time: 119.20s eta: 11:21:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.206 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31357.0 tgs: 261 data_time: 0.64s time: 119.72s eta: 11:22:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30252.0 tgs: 253 data_time: 0.57s time: 119.11s eta: 11:16:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32160.0 tgs: 269 data_time: 0.64s time: 119.12s eta: 11:15:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32206.0 tgs: 270 data_time: 0.68s time: 118.99s eta: 11:12:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 267 data_time: 0.81s time: 119.09s eta: 11:10:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 270 data_time: 0.76s time: 119.15s eta: 11:09:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32335.0 tgs: 271 data_time: 0.83s time: 118.97s eta: 11:06:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.121 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31996.0 tgs: 267 data_time: 0.68s time: 119.47s eta: 11:07:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 266 data_time: 0.96s time: 119.64s eta: 11:06:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32246.0 tgs: 270 data_time: 0.60s time: 119.30s eta: 11:02:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31885.0 tgs: 267 data_time: 0.71s time: 119.31s eta: 11:00:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31053.0 tgs: 261 data_time: 0.77s time: 118.82s eta: 10:55:29 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.127 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 266 data_time: 0.58s time: 119.47s eta: 10:57:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.205 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32110.0 tgs: 268 data_time: 0.59s time: 119.37s eta: 10:54:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31650.0 tgs: 265 data_time: 0.56s time: 119.28s eta: 10:52:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.102 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31124.0 tgs: 260 data_time: 0.69s time: 119.25s eta: 10:49:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32085.0 tgs: 269 data_time: 0.69s time: 119.09s eta: 10:47:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31688.0 tgs: 265 data_time: 0.65s time: 119.54s eta: 10:47:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30428.0 tgs: 254 data_time: 0.64s time: 119.34s eta: 10:44:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.179 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 264 data_time: 0.88s time: 119.49s eta: 10:43:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31922.0 tgs: 267 data_time: 0.75s time: 119.51s eta: 10:41:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30350.0 tgs: 255 data_time: 0.74s time: 118.99s eta: 10:36:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31440.0 tgs: 263 data_time: 0.61s time: 119.21s eta: 10:35:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31722.0 tgs: 266 data_time: 0.56s time: 119.06s eta: 10:32:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31123.0 tgs: 262 data_time: 0.73s time: 118.78s eta: 10:29:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32160.0 tgs: 268 data_time: 0.68s time: 119.86s eta: 10:33:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31121.0 tgs: 260 data_time: 0.81s time: 119.38s eta: 10:28:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31568.0 tgs: 265 data_time: 0.84s time: 118.83s eta: 10:23:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32076.0 tgs: 269 data_time: 1.01s time: 118.98s eta: 10:22:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 270 data_time: 0.84s time: 119.29s eta: 10:22:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31533.0 tgs: 264 data_time: 0.74s time: 119.20s eta: 10:19:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.208 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30617.0 tgs: 257 data_time: 0.77s time: 119.06s eta: 10:17:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31023.0 tgs: 259 data_time: 0.63s time: 119.47s eta: 10:17:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.273 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31981.0 tgs: 267 data_time: 0.82s time: 119.42s eta: 10:15:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.200 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 269 data_time: 0.82s time: 119.46s eta: 10:13:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31454.0 tgs: 263 data_time: 0.73s time: 119.29s eta: 10:10:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31844.0 tgs: 267 data_time: 0.64s time: 119.11s eta: 10:07:27 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31256.0 tgs: 261 data_time: 0.85s time: 119.70s eta: 10:08:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31006.0 tgs: 258 data_time: 0.85s time: 119.83s eta: 10:07:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32260.0 tgs: 265 data_time: 0.39s time: 121.55s eta: 10:13:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.106 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 266 data_time: 0.82s time: 119.25s eta: 10:00:14 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.110 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 263 data_time: 0.83s time: 118.80s eta: 9:55:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32025.0 tgs: 267 data_time: 0.80s time: 119.56s eta: 9:57:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32432.0 tgs: 270 data_time: 0.87s time: 119.70s eta: 9:56:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.197 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31904.0 tgs: 267 data_time: 0.61s time: 119.26s eta: 9:52:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31232.0 tgs: 261 data_time: 0.55s time: 119.33s eta: 9:50:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32155.0 tgs: 269 data_time: 0.72s time: 119.22s eta: 9:48:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.176 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32199.0 tgs: 266 data_time: 0.97s time: 120.80s eta: 9:53:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.154 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31943.0 tgs: 267 data_time: 0.75s time: 119.25s eta: 9:44:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32072.0 tgs: 268 data_time: 0.66s time: 119.46s eta: 9:43:22 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.206 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31197.0 tgs: 260 data_time: 0.81s time: 119.54s eta: 9:41:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31536.0 tgs: 264 data_time: 0.87s time: 119.20s eta: 9:38:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31646.0 tgs: 264 data_time: 0.84s time: 119.48s eta: 9:37:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31473.0 tgs: 264 data_time: 0.68s time: 119.05s eta: 9:33:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 270 data_time: 0.52s time: 119.24s eta: 9:32:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.113 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30515.0 tgs: 256 data_time: 0.50s time: 119.16s eta: 9:29:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 269 data_time: 0.63s time: 119.22s eta: 9:28:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.103 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31478.0 tgs: 263 data_time: 0.55s time: 119.46s eta: 9:27:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32223.0 tgs: 270 data_time: 0.87s time: 118.98s eta: 9:23:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30954.0 tgs: 260 data_time: 0.83s time: 119.03s eta: 9:21:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 269 data_time: 0.64s time: 119.53s eta: 9:21:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 268 data_time: 0.65s time: 119.59s eta: 9:20:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.183 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31640.0 tgs: 265 data_time: 0.55s time: 119.12s eta: 9:15:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.105 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31717.0 tgs: 266 data_time: 0.62s time: 119.07s eta: 9:13:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31846.0 tgs: 267 data_time: 0.98s time: 118.90s eta: 9:10:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.151 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31510.0 tgs: 264 data_time: 0.76s time: 119.32s eta: 9:10:51 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 268 data_time: 0.75s time: 119.09s eta: 9:07:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.202 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31288.0 tgs: 263 data_time: 0.65s time: 118.86s eta: 9:04:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31775.0 tgs: 265 data_time: 0.63s time: 119.67s eta: 9:06:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31506.0 tgs: 263 data_time: 0.69s time: 119.43s eta: 9:03:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32383.0 tgs: 271 data_time: 0.77s time: 119.43s eta: 9:01:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 268 data_time: 0.67s time: 118.74s eta: 8:56:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31809.0 tgs: 267 data_time: 0.80s time: 118.86s eta: 8:54:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.201 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31804.0 tgs: 266 data_time: 0.74s time: 119.33s eta: 8:55:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.234 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31819.0 tgs: 267 data_time: 0.90s time: 119.15s eta: 8:52:12 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31395.0 tgs: 263 data_time: 0.95s time: 119.26s eta: 8:50:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30913.0 tgs: 256 data_time: 0.64s time: 120.47s eta: 8:54:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.131 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30896.0 tgs: 259 data_time: 0.47s time: 119.24s eta: 8:46:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31426.0 tgs: 263 data_time: 0.62s time: 119.06s eta: 8:43:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30607.0 tgs: 253 data_time: 0.88s time: 120.83s eta: 8:49:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31666.0 tgs: 263 data_time: 1.06s time: 120.19s eta: 8:44:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31422.0 tgs: 264 data_time: 0.77s time: 118.82s eta: 8:36:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32065.0 tgs: 268 data_time: 0.75s time: 119.35s eta: 8:37:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29051.0 tgs: 243 data_time: 0.71s time: 119.33s eta: 8:35:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.198 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31478.0 tgs: 264 data_time: 0.86s time: 119.05s eta: 8:31:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.70s time: 119.16s eta: 8:30:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 266 data_time: 0.93s time: 119.04s eta: 8:27:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.191 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31903.0 tgs: 267 data_time: 0.88s time: 119.12s eta: 8:26:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.179 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31462.0 tgs: 263 data_time: 0.51s time: 119.24s eta: 8:24:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31230.0 tgs: 261 data_time: 0.89s time: 119.22s eta: 8:22:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31975.0 tgs: 267 data_time: 1.03s time: 119.38s eta: 8:21:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31603.0 tgs: 263 data_time: 0.79s time: 119.78s eta: 8:21:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.180 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31330.0 tgs: 262 data_time: 0.86s time: 119.20s eta: 8:16:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32314.0 tgs: 271 data_time: 0.57s time: 119.02s eta: 8:13:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30649.0 tgs: 255 data_time: 0.69s time: 119.90s eta: 8:15:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.115 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31730.0 tgs: 265 data_time: 0.76s time: 119.55s eta: 8:12:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32136.0 tgs: 268 data_time: 0.71s time: 119.59s eta: 8:10:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32233.0 tgs: 270 data_time: 0.70s time: 119.00s eta: 8:05:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32420.0 tgs: 270 data_time: 0.88s time: 119.84s eta: 8:07:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 267 data_time: 0.71s time: 119.71s eta: 8:04:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30701.0 tgs: 257 data_time: 0.62s time: 119.40s eta: 8:01:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30999.0 tgs: 259 data_time: 0.56s time: 119.28s eta: 7:59:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.098 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30910.0 tgs: 259 data_time: 0.75s time: 119.20s eta: 7:56:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31626.0 tgs: 265 data_time: 0.84s time: 119.19s eta: 7:54:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.109 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31205.0 tgs: 262 data_time: 0.69s time: 118.93s eta: 7:51:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31301.0 tgs: 263 data_time: 0.50s time: 118.97s eta: 7:49:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.115 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30878.0 tgs: 258 data_time: 0.76s time: 119.41s eta: 7:49:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32167.0 tgs: 270 data_time: 0.66s time: 119.03s eta: 7:46:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.120 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30569.0 tgs: 256 data_time: 0.79s time: 119.10s eta: 7:44:29 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31276.0 tgs: 260 data_time: 0.81s time: 120.02s eta: 7:46:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31148.0 tgs: 259 data_time: 0.70s time: 119.82s eta: 7:43:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.122 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 264 data_time: 0.61s time: 119.16s eta: 7:38:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32303.0 tgs: 271 data_time: 0.79s time: 118.97s eta: 7:36:02 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 264 data_time: 0.58s time: 119.32s eta: 7:35:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.221 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31801.0 tgs: 266 data_time: 0.94s time: 119.38s eta: 7:33:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.210 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31606.0 tgs: 265 data_time: 0.72s time: 119.06s eta: 7:30:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.237 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 270 data_time: 0.54s time: 118.99s eta: 7:28:12 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.177 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32225.0 tgs: 270 data_time: 0.77s time: 119.24s eta: 7:27:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31317.0 tgs: 263 data_time: 0.76s time: 118.99s eta: 7:24:12 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.096 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32487.0 tgs: 273 data_time: 0.93s time: 118.93s eta: 7:22:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31473.0 tgs: 264 data_time: 0.58s time: 119.08s eta: 7:20:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 267 data_time: 1.08s time: 118.99s eta: 7:18:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 0.82s time: 119.03s eta: 7:16:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31564.0 tgs: 264 data_time: 0.90s time: 119.26s eta: 7:15:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31494.0 tgs: 263 data_time: 1.02s time: 119.36s eta: 7:13:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32085.0 tgs: 269 data_time: 0.87s time: 118.94s eta: 7:10:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31085.0 tgs: 260 data_time: 0.75s time: 119.20s eta: 7:09:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30739.0 tgs: 257 data_time: 0.49s time: 119.59s eta: 7:08:31 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30876.0 tgs: 258 data_time: 0.90s time: 119.65s eta: 7:06:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32187.0 tgs: 270 data_time: 0.75s time: 118.95s eta: 7:02:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.182 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31679.0 tgs: 265 data_time: 0.59s time: 119.30s eta: 7:01:31 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31599.0 tgs: 264 data_time: 0.67s time: 119.51s eta: 7:00:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30788.0 tgs: 257 data_time: 0.68s time: 119.42s eta: 6:57:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31658.0 tgs: 266 data_time: 0.78s time: 118.91s eta: 6:54:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.207 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31288.0 tgs: 262 data_time: 0.90s time: 119.34s eta: 6:53:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 265 data_time: 0.66s time: 118.93s eta: 6:50:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.113 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29411.0 tgs: 246 data_time: 0.76s time: 119.26s eta: 6:49:28 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31947.0 tgs: 268 data_time: 0.81s time: 119.02s eta: 6:46:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.193 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32296.0 tgs: 270 data_time: 0.59s time: 119.23s eta: 6:45:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32317.0 tgs: 270 data_time: 0.91s time: 119.41s eta: 6:44:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31893.0 tgs: 267 data_time: 0.60s time: 119.01s eta: 6:40:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.229 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31933.0 tgs: 268 data_time: 0.72s time: 119.01s eta: 6:38:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32279.0 tgs: 270 data_time: 0.63s time: 119.45s eta: 6:38:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32098.0 tgs: 270 data_time: 0.79s time: 118.80s eta: 6:34:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.82s time: 119.44s eta: 6:34:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.093 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31963.0 tgs: 267 data_time: 0.78s time: 119.56s eta: 6:32:34 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32047.0 tgs: 268 data_time: 0.59s time: 119.38s eta: 6:29:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32344.0 tgs: 270 data_time: 0.73s time: 119.73s eta: 6:29:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30861.0 tgs: 258 data_time: 0.62s time: 119.36s eta: 6:25:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30730.0 tgs: 258 data_time: 0.75s time: 118.81s eta: 6:22:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31433.0 tgs: 264 data_time: 0.73s time: 119.02s eta: 6:20:51 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.161 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 264 data_time: 0.95s time: 119.40s eta: 6:20:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.214 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30937.0 tgs: 259 data_time: 1.05s time: 119.33s eta: 6:17:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31473.0 tgs: 264 data_time: 0.80s time: 119.05s eta: 6:14:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30918.0 tgs: 258 data_time: 0.67s time: 119.41s eta: 6:14:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.111 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31563.0 tgs: 264 data_time: 0.78s time: 119.39s eta: 6:12:05 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32363.0 tgs: 271 data_time: 0.76s time: 119.28s eta: 6:09:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30922.0 tgs: 259 data_time: 0.67s time: 119.10s eta: 6:07:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31702.0 tgs: 259 data_time: 1.02s time: 121.96s eta: 6:14:00 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29878.0 tgs: 250 data_time: 0.47s time: 119.38s eta: 6:04:07 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32043.0 tgs: 267 data_time: 0.52s time: 119.57s eta: 6:02:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31252.0 tgs: 262 data_time: 0.65s time: 119.26s eta: 5:59:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.85s time: 119.26s eta: 5:57:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31301.0 tgs: 263 data_time: 0.82s time: 118.92s eta: 5:54:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.186 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 270 data_time: 0.52s time: 119.01s eta: 5:53:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30499.0 tgs: 256 data_time: 0.74s time: 119.07s eta: 5:51:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.62s time: 119.29s eta: 5:49:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.232 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31346.0 tgs: 263 data_time: 0.71s time: 118.95s eta: 5:46:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:00:05][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31986.0 tgs: 268 data_time: 0.61s time: 119.28s eta: 5:45:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31759.0 tgs: 266 data_time: 0.84s time: 119.38s eta: 5:44:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32118.0 tgs: 269 data_time: 0.76s time: 119.05s eta: 5:41:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32177.0 tgs: 269 data_time: 0.55s time: 119.49s eta: 5:40:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31619.0 tgs: 264 data_time: 0.55s time: 119.38s eta: 5:38:14 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.120 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31770.0 tgs: 265 data_time: 0.89s time: 119.57s eta: 5:36:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.226 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 271 data_time: 0.82s time: 118.98s eta: 5:33:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 270 data_time: 0.56s time: 118.95s eta: 5:31:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31909.0 tgs: 267 data_time: 0.66s time: 119.23s eta: 5:29:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.258 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32070.0 tgs: 268 data_time: 1.01s time: 119.40s eta: 5:28:21 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30860.0 tgs: 258 data_time: 0.88s time: 119.46s eta: 5:26:31 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31447.0 tgs: 264 data_time: 0.80s time: 118.95s eta: 5:23:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.208 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32029.0 tgs: 267 data_time: 0.56s time: 119.63s eta: 5:22:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32272.0 tgs: 270 data_time: 0.82s time: 119.35s eta: 5:20:14 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31746.0 tgs: 266 data_time: 0.79s time: 119.12s eta: 5:17:39 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.113 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31223.0 tgs: 261 data_time: 0.53s time: 119.42s eta: 5:16:27 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30443.0 tgs: 254 data_time: 0.67s time: 119.58s eta: 5:14:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32008.0 tgs: 268 data_time: 0.77s time: 119.00s eta: 5:11:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31898.0 tgs: 268 data_time: 0.83s time: 119.00s eta: 5:09:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30983.0 tgs: 259 data_time: 0.64s time: 119.22s eta: 5:07:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.103 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32114.0 tgs: 268 data_time: 0.58s time: 119.68s eta: 5:07:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31267.0 tgs: 262 data_time: 0.93s time: 119.11s eta: 5:03:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32116.0 tgs: 269 data_time: 0.55s time: 119.07s eta: 5:01:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31611.0 tgs: 260 data_time: 0.78s time: 121.52s eta: 5:05:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32120.0 tgs: 268 data_time: 0.84s time: 119.57s eta: 4:58:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30857.0 tgs: 258 data_time: 0.60s time: 119.50s eta: 4:56:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31122.0 tgs: 259 data_time: 0.65s time: 119.73s eta: 4:55:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 267 data_time: 1.00s time: 119.52s eta: 4:52:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.188 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31299.0 tgs: 262 data_time: 0.81s time: 119.41s eta: 4:50:34 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31067.0 tgs: 260 data_time: 0.77s time: 119.20s eta: 4:48:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31110.0 tgs: 261 data_time: 0.76s time: 118.88s eta: 4:45:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30959.0 tgs: 259 data_time: 0.80s time: 119.12s eta: 4:43:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.252 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32155.0 tgs: 268 data_time: 0.80s time: 119.98s eta: 4:43:57 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30478.0 tgs: 255 data_time: 0.63s time: 119.14s eta: 4:39:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31802.0 tgs: 267 data_time: 0.85s time: 119.01s eta: 4:37:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31279.0 tgs: 262 data_time: 0.72s time: 119.05s eta: 4:35:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30979.0 tgs: 259 data_time: 0.75s time: 119.37s eta: 4:34:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31410.0 tgs: 263 data_time: 0.68s time: 119.41s eta: 4:32:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.119 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30876.0 tgs: 259 data_time: 0.67s time: 118.94s eta: 4:29:35 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31330.0 tgs: 261 data_time: 0.81s time: 119.60s eta: 4:29:06 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31223.0 tgs: 261 data_time: 0.75s time: 119.24s eta: 4:26:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31372.0 tgs: 263 data_time: 0.71s time: 119.06s eta: 4:23:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.112 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32058.0 tgs: 269 data_time: 0.54s time: 118.85s eta: 4:21:27 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.105 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31508.0 tgs: 265 data_time: 0.75s time: 118.76s eta: 4:19:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 265 data_time: 0.70s time: 119.43s eta: 4:18:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 267 data_time: 0.70s time: 119.40s eta: 4:16:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.113 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 266 data_time: 0.81s time: 119.35s eta: 4:14:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.198 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 270 data_time: 0.75s time: 119.05s eta: 4:11:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31414.0 tgs: 263 data_time: 0.96s time: 119.12s eta: 4:10:09 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 266 data_time: 0.63s time: 119.13s eta: 4:08:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.207 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31335.0 tgs: 263 data_time: 0.76s time: 118.96s eta: 4:05:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31978.0 tgs: 268 data_time: 0.59s time: 119.15s eta: 4:04:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32231.0 tgs: 269 data_time: 0.66s time: 119.50s eta: 4:02:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31339.0 tgs: 262 data_time: 0.85s time: 119.25s eta: 4:00:29 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.228 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32120.0 tgs: 269 data_time: 0.72s time: 119.19s eta: 3:58:22 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30076.0 tgs: 251 data_time: 0.48s time: 119.36s eta: 3:56:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.230 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31943.0 tgs: 267 data_time: 0.53s time: 119.37s eta: 3:54:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31315.0 tgs: 261 data_time: 0.75s time: 119.95s eta: 3:53:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31944.0 tgs: 267 data_time: 0.53s time: 119.31s eta: 3:50:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32224.0 tgs: 269 data_time: 0.92s time: 119.63s eta: 3:49:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31645.0 tgs: 266 data_time: 0.63s time: 118.86s eta: 3:45:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31541.0 tgs: 263 data_time: 0.79s time: 119.75s eta: 3:45:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 265 data_time: 0.73s time: 119.33s eta: 3:42:44 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32351.0 tgs: 271 data_time: 0.67s time: 119.26s eta: 3:40:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.256 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31994.0 tgs: 268 data_time: 0.58s time: 118.99s eta: 3:38:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31662.0 tgs: 265 data_time: 0.83s time: 119.44s eta: 3:36:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32004.0 tgs: 268 data_time: 0.79s time: 119.03s eta: 3:34:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 267 data_time: 0.78s time: 118.99s eta: 3:32:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32211.0 tgs: 270 data_time: 0.80s time: 119.22s eta: 3:30:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 265 data_time: 0.64s time: 120.11s eta: 3:30:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31343.0 tgs: 263 data_time: 1.03s time: 119.13s eta: 3:26:29 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31637.0 tgs: 265 data_time: 0.64s time: 119.31s eta: 3:24:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.107 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30575.0 tgs: 256 data_time: 0.61s time: 119.08s eta: 3:22:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.204 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32117.0 tgs: 268 data_time: 0.71s time: 119.43s eta: 3:21:02 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32390.0 tgs: 271 data_time: 0.90s time: 119.29s eta: 3:18:48 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31448.0 tgs: 263 data_time: 0.73s time: 119.22s eta: 3:16:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30578.0 tgs: 256 data_time: 0.84s time: 119.16s eta: 3:14:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30317.0 tgs: 254 data_time: 0.59s time: 119.02s eta: 3:12:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32080.0 tgs: 269 data_time: 0.87s time: 119.18s eta: 3:10:41 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 267 data_time: 0.66s time: 121.08s eta: 3:11:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31456.0 tgs: 263 data_time: 1.02s time: 119.50s eta: 3:07:13 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32208.0 tgs: 270 data_time: 0.77s time: 119.26s eta: 3:04:51 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30833.0 tgs: 259 data_time: 0.51s time: 118.98s eta: 3:02:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32482.0 tgs: 272 data_time: 0.92s time: 119.22s eta: 3:00:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31234.0 tgs: 261 data_time: 0.61s time: 119.45s eta: 2:59:10 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30873.0 tgs: 259 data_time: 0.95s time: 118.98s eta: 2:56:29 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31321.0 tgs: 262 data_time: 0.71s time: 119.36s eta: 2:55:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30164.0 tgs: 252 data_time: 0.78s time: 119.40s eta: 2:53:08 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32131.0 tgs: 269 data_time: 0.50s time: 119.35s eta: 2:51:04 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 262 data_time: 0.61s time: 121.46s eta: 2:52:03 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31358.0 tgs: 262 data_time: 0.78s time: 119.24s eta: 2:46:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31428.0 tgs: 262 data_time: 0.77s time: 119.60s eta: 2:45:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31506.0 tgs: 264 data_time: 0.65s time: 118.92s eta: 2:42:31 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31983.0 tgs: 268 data_time: 0.78s time: 119.18s eta: 2:40:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32377.0 tgs: 272 data_time: 0.59s time: 118.97s eta: 2:38:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32134.0 tgs: 269 data_time: 0.84s time: 119.24s eta: 2:36:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.81s time: 119.10s eta: 2:34:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 265 data_time: 0.62s time: 119.43s eta: 2:33:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32478.0 tgs: 270 data_time: 0.72s time: 120.22s eta: 2:32:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32539.0 tgs: 272 data_time: 0.99s time: 119.28s eta: 2:29:06 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31644.0 tgs: 266 data_time: 0.73s time: 118.88s eta: 2:26:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 267 data_time: 1.10s time: 119.51s eta: 2:25:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30803.0 tgs: 258 data_time: 0.79s time: 119.07s eta: 2:22:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31056.0 tgs: 260 data_time: 0.72s time: 119.14s eta: 2:20:58 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32306.0 tgs: 271 data_time: 1.02s time: 119.05s eta: 2:18:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 265 data_time: 1.01s time: 119.84s eta: 2:17:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31162.0 tgs: 261 data_time: 0.50s time: 119.34s eta: 2:15:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 268 data_time: 0.55s time: 119.38s eta: 2:13:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31794.0 tgs: 265 data_time: 0.61s time: 119.68s eta: 2:11:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31899.0 tgs: 266 data_time: 0.69s time: 119.49s eta: 2:09:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31755.0 tgs: 266 data_time: 0.64s time: 118.96s eta: 2:06:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31617.0 tgs: 264 data_time: 0.67s time: 119.37s eta: 2:05:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31404.0 tgs: 263 data_time: 0.71s time: 119.27s eta: 2:03:14 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31039.0 tgs: 260 data_time: 0.63s time: 119.22s eta: 2:01:12 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.315 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 266 data_time: 0.74s time: 119.51s eta: 1:59:30 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 265 data_time: 0.79s time: 119.42s eta: 1:57:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31706.0 tgs: 265 data_time: 0.84s time: 119.53s eta: 1:55:32 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32145.0 tgs: 269 data_time: 0.77s time: 119.30s eta: 1:53:20 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31837.0 tgs: 266 data_time: 0.99s time: 119.25s eta: 1:51:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31735.0 tgs: 266 data_time: 0.82s time: 118.93s eta: 1:49:01 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31939.0 tgs: 266 data_time: 0.48s time: 119.64s eta: 1:47:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31480.0 tgs: 262 data_time: 0.89s time: 119.72s eta: 1:45:45 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32280.0 tgs: 270 data_time: 0.66s time: 119.35s eta: 1:43:26 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31796.0 tgs: 266 data_time: 0.64s time: 119.30s eta: 1:41:24 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31968.0 tgs: 268 data_time: 0.85s time: 119.28s eta: 1:39:23 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32116.0 tgs: 270 data_time: 0.85s time: 118.91s eta: 1:37:06 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31277.0 tgs: 262 data_time: 0.81s time: 119.16s eta: 1:35:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.65s time: 119.06s eta: 1:33:15 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31830.0 tgs: 266 data_time: 0.59s time: 119.25s eta: 1:31:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 269 data_time: 0.63s time: 119.72s eta: 1:29:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 263 data_time: 0.78s time: 119.01s eta: 1:27:16 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32167.0 tgs: 270 data_time: 0.81s time: 118.87s eta: 1:25:11 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29522.0 tgs: 247 data_time: 0.55s time: 119.48s eta: 1:23:37 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30894.0 tgs: 258 data_time: 0.58s time: 119.48s eta: 1:21:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31530.0 tgs: 265 data_time: 0.62s time: 118.94s eta: 1:19:17 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31873.0 tgs: 267 data_time: 0.80s time: 118.93s eta: 1:17:18 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31926.0 tgs: 261 data_time: 0.83s time: 122.16s eta: 1:17:22 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32056.0 tgs: 268 data_time: 0.57s time: 119.36s eta: 1:13:36 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31302.0 tgs: 262 data_time: 0.60s time: 119.05s eta: 1:11:25 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31700.0 tgs: 266 data_time: 0.77s time: 118.92s eta: 1:09:22 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31840.0 tgs: 267 data_time: 0.77s time: 119.10s eta: 1:07:29 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.213 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 264 data_time: 0.78s time: 119.42s eta: 1:05:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30968.0 tgs: 259 data_time: 0.85s time: 119.33s eta: 1:03:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31765.0 tgs: 266 data_time: 0.91s time: 119.18s eta: 1:01:34 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31224.0 tgs: 261 data_time: 0.64s time: 119.27s eta: 0:59:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31042.0 tgs: 259 data_time: 0.58s time: 119.44s eta: 0:57:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30602.0 tgs: 256 data_time: 0.88s time: 119.39s eta: 0:55:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32154.0 tgs: 269 data_time: 0.92s time: 119.48s eta: 0:53:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31928.0 tgs: 267 data_time: 0.75s time: 119.36s eta: 0:51:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32167.0 tgs: 269 data_time: 0.91s time: 119.47s eta: 0:49:46 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32088.0 tgs: 269 data_time: 0.78s time: 118.89s eta: 0:47:33 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31677.0 tgs: 265 data_time: 0.54s time: 119.28s eta: 0:45:43 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32432.0 tgs: 272 data_time: 0.72s time: 119.22s eta: 0:43:42 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32086.0 tgs: 269 data_time: 0.67s time: 118.99s eta: 0:41:38 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.63s time: 119.04s eta: 0:39:40 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31604.0 tgs: 264 data_time: 0.82s time: 119.66s eta: 0:37:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 267 data_time: 0.51s time: 119.28s eta: 0:35:47 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32555.0 tgs: 271 data_time: 0.76s time: 119.74s eta: 0:33:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31713.0 tgs: 261 data_time: 0.88s time: 121.22s eta: 0:32:19 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 265 data_time: 0.69s time: 119.38s eta: 0:29:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31575.0 tgs: 264 data_time: 0.76s time: 119.32s eta: 0:27:50 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31044.0 tgs: 259 data_time: 0.74s time: 119.63s eta: 0:25:55 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32100.0 tgs: 269 data_time: 1.02s time: 119.09s eta: 0:23:49 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31500.0 tgs: 263 data_time: 0.87s time: 119.38s eta: 0:21:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31519.0 tgs: 264 data_time: 0.61s time: 119.26s eta: 0:19:52 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 268 data_time: 1.15s time: 119.24s eta: 0:17:53 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32063.0 tgs: 267 data_time: 0.71s time: 119.91s eta: 0:15:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32312.0 tgs: 271 data_time: 0.66s time: 119.22s eta: 0:13:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31645.0 tgs: 265 data_time: 0.77s time: 119.09s eta: 0:11:54 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30933.0 tgs: 259 data_time: 0.79s time: 119.32s eta: 0:09:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31953.0 tgs: 268 data_time: 0.72s time: 119.17s eta: 0:07:56 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32212.0 tgs: 269 data_time: 0.81s time: 119.69s eta: 0:05:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30966.0 tgs: 254 data_time: 0.55s time: 121.45s eta: 0:04:02 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31936.0 tgs: 267 data_time: 0.97s time: 119.21s eta: 0:01:59 +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 4][DP 1][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank41.log b/20250121165312/rank41.log new file mode 100644 index 0000000000000000000000000000000000000000..61ea093d5a6d17ecfa9435a9d4ddc3955df136d3 --- /dev/null +++ b/20250121165312/rank41.log @@ -0,0 +1,851 @@ +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.49 seconds, peak gpu memory 13.4G +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.253 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 32069.0 tgs: 211 data_time: 1.97s time: 151.64s eta: 1 day, 0:58:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.258 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32159.0 tgs: 267 data_time: 0.89s time: 120.12s eta: 19:45:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.275 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31619.0 tgs: 264 data_time: 1.04s time: 119.69s eta: 19:38:57 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.205 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32056.0 tgs: 267 data_time: 0.91s time: 119.65s eta: 19:36:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.196 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31273.0 tgs: 262 data_time: 1.10s time: 119.31s eta: 19:31:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.201 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31327.0 tgs: 262 data_time: 0.98s time: 119.54s eta: 19:31:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.212 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32396.0 tgs: 263 data_time: 0.91s time: 122.97s eta: 20:03:00 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.207 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31929.0 tgs: 266 data_time: 0.85s time: 119.88s eta: 19:30:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.201 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31933.0 tgs: 264 data_time: 0.85s time: 120.66s eta: 19:36:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31433.0 tgs: 262 data_time: 0.84s time: 119.92s eta: 19:27:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.200 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32287.0 tgs: 268 data_time: 0.99s time: 120.46s eta: 19:30:27 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.249 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31446.0 tgs: 263 data_time: 0.77s time: 119.34s eta: 19:17:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.255 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 266 data_time: 0.78s time: 119.68s eta: 19:18:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31057.0 tgs: 260 data_time: 0.74s time: 119.17s eta: 19:11:58 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32202.0 tgs: 270 data_time: 0.94s time: 119.20s eta: 19:10:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32492.0 tgs: 273 data_time: 0.65s time: 118.93s eta: 19:05:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31095.0 tgs: 260 data_time: 0.64s time: 119.15s eta: 19:05:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32615.0 tgs: 273 data_time: 0.71s time: 119.18s eta: 19:04:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32172.0 tgs: 269 data_time: 0.67s time: 119.40s eta: 19:04:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32340.0 tgs: 271 data_time: 0.68s time: 119.33s eta: 19:01:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31148.0 tgs: 258 data_time: 0.80s time: 120.31s eta: 19:08:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30975.0 tgs: 252 data_time: 0.64s time: 122.91s eta: 19:31:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32237.0 tgs: 269 data_time: 0.95s time: 119.48s eta: 18:57:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31482.0 tgs: 264 data_time: 0.90s time: 119.16s eta: 18:51:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31615.0 tgs: 265 data_time: 0.72s time: 119.25s eta: 18:50:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32372.0 tgs: 271 data_time: 1.07s time: 119.39s eta: 18:50:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 27656.0 tgs: 231 data_time: 0.71s time: 119.67s eta: 18:50:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31754.0 tgs: 265 data_time: 0.58s time: 119.38s eta: 18:46:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32225.0 tgs: 269 data_time: 1.03s time: 119.78s eta: 18:47:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32178.0 tgs: 248 data_time: 0.92s time: 129.31s eta: 20:15:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32356.0 tgs: 270 data_time: 0.99s time: 119.57s eta: 18:41:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32084.0 tgs: 268 data_time: 0.61s time: 119.39s eta: 18:38:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.260 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31489.0 tgs: 260 data_time: 0.83s time: 120.66s eta: 18:48:08 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31760.0 tgs: 266 data_time: 0.80s time: 119.40s eta: 18:34:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31719.0 tgs: 266 data_time: 0.95s time: 119.17s eta: 18:30:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31694.0 tgs: 266 data_time: 0.91s time: 119.07s eta: 18:27:22 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31260.0 tgs: 262 data_time: 0.61s time: 119.13s eta: 18:25:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32424.0 tgs: 269 data_time: 1.23s time: 120.39s eta: 18:35:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31747.0 tgs: 262 data_time: 0.72s time: 121.17s eta: 18:40:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31806.0 tgs: 263 data_time: 0.91s time: 120.67s eta: 18:34:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31095.0 tgs: 259 data_time: 0.83s time: 119.74s eta: 18:23:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31720.0 tgs: 256 data_time: 0.98s time: 123.87s eta: 18:59:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31981.0 tgs: 267 data_time: 1.17s time: 119.56s eta: 18:17:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31115.0 tgs: 260 data_time: 0.87s time: 119.32s eta: 18:13:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32109.0 tgs: 269 data_time: 0.86s time: 119.23s eta: 18:10:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31354.0 tgs: 262 data_time: 0.73s time: 119.60s eta: 18:12:20 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.258 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31295.0 tgs: 262 data_time: 0.75s time: 119.27s eta: 18:07:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32080.0 tgs: 269 data_time: 0.81s time: 119.02s eta: 18:03:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32230.0 tgs: 270 data_time: 0.63s time: 119.09s eta: 18:01:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32091.0 tgs: 269 data_time: 0.90s time: 119.26s eta: 18:01:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31199.0 tgs: 261 data_time: 0.99s time: 119.32s eta: 17:59:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31340.0 tgs: 261 data_time: 0.79s time: 120.02s eta: 18:04:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31916.0 tgs: 267 data_time: 0.79s time: 119.10s eta: 17:53:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 267 data_time: 0.64s time: 119.42s eta: 17:54:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31153.0 tgs: 261 data_time: 0.83s time: 118.99s eta: 17:48:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31392.0 tgs: 263 data_time: 0.93s time: 119.02s eta: 17:47:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31449.0 tgs: 250 data_time: 0.88s time: 125.37s eta: 18:42:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32236.0 tgs: 268 data_time: 0.79s time: 120.11s eta: 17:52:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31580.0 tgs: 264 data_time: 0.65s time: 119.36s eta: 17:44:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31801.0 tgs: 266 data_time: 0.65s time: 119.34s eta: 17:42:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31747.0 tgs: 265 data_time: 0.82s time: 119.51s eta: 17:41:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.341 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 251 data_time: 1.02s time: 126.04s eta: 18:37:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31543.0 tgs: 257 data_time: 0.70s time: 122.63s eta: 18:05:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32204.0 tgs: 265 data_time: 0.70s time: 121.16s eta: 17:50:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30805.0 tgs: 258 data_time: 0.76s time: 119.04s eta: 17:29:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31921.0 tgs: 268 data_time: 0.97s time: 118.85s eta: 17:25:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32335.0 tgs: 260 data_time: 0.84s time: 124.00s eta: 18:09:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32297.0 tgs: 259 data_time: 0.87s time: 124.57s eta: 18:12:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31866.0 tgs: 266 data_time: 0.76s time: 119.40s eta: 17:24:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32561.0 tgs: 272 data_time: 0.62s time: 119.69s eta: 17:25:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 267 data_time: 0.73s time: 119.53s eta: 17:21:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31696.0 tgs: 266 data_time: 0.87s time: 118.86s eta: 17:14:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30731.0 tgs: 256 data_time: 0.85s time: 119.85s eta: 17:20:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.233 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32264.0 tgs: 269 data_time: 0.95s time: 119.53s eta: 17:15:57 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.277 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31133.0 tgs: 255 data_time: 0.59s time: 121.86s eta: 17:34:03 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32236.0 tgs: 260 data_time: 0.76s time: 123.51s eta: 17:46:20 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.122 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30927.0 tgs: 259 data_time: 0.72s time: 119.01s eta: 17:05:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32197.0 tgs: 270 data_time: 0.79s time: 119.23s eta: 17:05:21 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31126.0 tgs: 261 data_time: 0.89s time: 119.14s eta: 17:02:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30513.0 tgs: 255 data_time: 0.81s time: 119.25s eta: 17:01:35 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31357.0 tgs: 263 data_time: 0.72s time: 119.04s eta: 16:57:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32154.0 tgs: 269 data_time: 0.69s time: 119.29s eta: 16:57:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31416.0 tgs: 263 data_time: 0.64s time: 119.16s eta: 16:54:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31901.0 tgs: 266 data_time: 0.70s time: 119.58s eta: 16:56:27 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31596.0 tgs: 265 data_time: 0.67s time: 119.13s eta: 16:50:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31902.0 tgs: 268 data_time: 0.68s time: 118.85s eta: 16:46:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.274 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32051.0 tgs: 269 data_time: 0.76s time: 119.01s eta: 16:45:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.227 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31807.0 tgs: 265 data_time: 0.75s time: 119.86s eta: 16:50:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31978.0 tgs: 267 data_time: 0.88s time: 119.36s eta: 16:44:35 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32055.0 tgs: 269 data_time: 0.72s time: 119.03s eta: 16:39:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31508.0 tgs: 260 data_time: 0.72s time: 120.91s eta: 16:53:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.112 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31549.0 tgs: 265 data_time: 0.70s time: 119.00s eta: 16:35:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32149.0 tgs: 269 data_time: 0.57s time: 119.29s eta: 16:36:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.201 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32167.0 tgs: 269 data_time: 0.89s time: 119.23s eta: 16:33:33 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.213 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31587.0 tgs: 264 data_time: 0.79s time: 119.37s eta: 16:32:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31707.0 tgs: 264 data_time: 1.01s time: 119.70s eta: 16:33:32 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31488.0 tgs: 264 data_time: 0.81s time: 119.21s eta: 16:27:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31928.0 tgs: 266 data_time: 0.64s time: 119.76s eta: 16:30:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31519.0 tgs: 261 data_time: 0.89s time: 120.33s eta: 16:32:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32320.0 tgs: 271 data_time: 0.97s time: 118.97s eta: 16:19:32 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30987.0 tgs: 243 data_time: 0.72s time: 127.33s eta: 17:26:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.245 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 269 data_time: 0.65s time: 119.22s eta: 16:17:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31996.0 tgs: 268 data_time: 0.50s time: 119.14s eta: 16:14:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 266 data_time: 0.83s time: 119.22s eta: 16:13:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32282.0 tgs: 270 data_time: 0.77s time: 119.17s eta: 16:11:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32337.0 tgs: 271 data_time: 0.62s time: 118.95s eta: 16:07:27 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.271 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30162.0 tgs: 253 data_time: 0.81s time: 118.86s eta: 16:04:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32170.0 tgs: 270 data_time: 0.72s time: 118.96s eta: 16:03:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.094 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28719.0 tgs: 235 data_time: 0.52s time: 121.84s eta: 16:24:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31313.0 tgs: 263 data_time: 0.66s time: 119.00s eta: 15:59:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32026.0 tgs: 268 data_time: 0.62s time: 119.22s eta: 15:59:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30116.0 tgs: 251 data_time: 0.60s time: 119.62s eta: 16:00:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31059.0 tgs: 260 data_time: 0.91s time: 119.04s eta: 15:54:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31929.0 tgs: 268 data_time: 0.67s time: 118.98s eta: 15:51:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.212 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31991.0 tgs: 268 data_time: 0.66s time: 119.17s eta: 15:51:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 263 data_time: 0.70s time: 119.64s eta: 15:53:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31053.0 tgs: 261 data_time: 0.76s time: 118.94s eta: 15:45:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31692.0 tgs: 265 data_time: 0.63s time: 119.28s eta: 15:46:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32347.0 tgs: 271 data_time: 0.94s time: 119.11s eta: 15:42:57 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31725.0 tgs: 266 data_time: 0.60s time: 118.99s eta: 15:40:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31531.0 tgs: 263 data_time: 0.79s time: 119.66s eta: 15:43:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32398.0 tgs: 266 data_time: 0.73s time: 121.50s eta: 15:55:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.120 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32317.0 tgs: 270 data_time: 0.71s time: 119.42s eta: 15:37:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.232 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31324.0 tgs: 241 data_time: 0.72s time: 129.97s eta: 16:58:03 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31625.0 tgs: 264 data_time: 0.78s time: 119.46s eta: 15:33:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31866.0 tgs: 266 data_time: 0.81s time: 119.61s eta: 15:32:57 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32349.0 tgs: 270 data_time: 1.02s time: 119.53s eta: 15:30:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 260 data_time: 0.60s time: 122.14s eta: 15:48:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32369.0 tgs: 268 data_time: 0.73s time: 120.76s eta: 15:35:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31523.0 tgs: 263 data_time: 0.75s time: 119.41s eta: 15:23:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 268 data_time: 1.18s time: 119.44s eta: 15:21:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31827.0 tgs: 241 data_time: 0.86s time: 131.63s eta: 16:53:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31520.0 tgs: 264 data_time: 0.67s time: 119.29s eta: 15:16:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32328.0 tgs: 270 data_time: 0.69s time: 119.55s eta: 15:16:35 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31312.0 tgs: 256 data_time: 1.00s time: 122.25s eta: 15:35:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30570.0 tgs: 255 data_time: 0.59s time: 119.71s eta: 15:13:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31656.0 tgs: 265 data_time: 0.83s time: 119.35s eta: 15:09:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30609.0 tgs: 257 data_time: 0.96s time: 118.90s eta: 15:03:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30979.0 tgs: 260 data_time: 0.79s time: 119.00s eta: 15:02:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.70s time: 119.31s eta: 15:02:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31276.0 tgs: 263 data_time: 1.14s time: 118.92s eta: 14:57:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 268 data_time: 0.70s time: 120.52s eta: 15:07:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31767.0 tgs: 266 data_time: 1.02s time: 119.31s eta: 14:56:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31979.0 tgs: 260 data_time: 0.76s time: 122.66s eta: 15:19:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 266 data_time: 1.05s time: 119.12s eta: 14:51:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31832.0 tgs: 265 data_time: 1.24s time: 119.76s eta: 14:54:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31292.0 tgs: 261 data_time: 0.79s time: 119.47s eta: 14:50:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31876.0 tgs: 267 data_time: 0.86s time: 118.99s eta: 14:44:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31124.0 tgs: 260 data_time: 1.00s time: 119.30s eta: 14:44:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31881.0 tgs: 268 data_time: 0.80s time: 118.93s eta: 14:40:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32174.0 tgs: 270 data_time: 0.83s time: 119.06s eta: 14:39:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 270 data_time: 0.77s time: 119.06s eta: 14:37:03 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31116.0 tgs: 261 data_time: 0.79s time: 119.13s eta: 14:35:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 270 data_time: 0.78s time: 119.00s eta: 14:32:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32510.0 tgs: 273 data_time: 0.88s time: 119.08s eta: 14:31:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31713.0 tgs: 266 data_time: 0.77s time: 119.15s eta: 14:29:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31085.0 tgs: 259 data_time: 0.68s time: 119.66s eta: 14:31:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31637.0 tgs: 265 data_time: 0.86s time: 119.18s eta: 14:26:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31800.0 tgs: 266 data_time: 0.85s time: 119.20s eta: 14:24:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31902.0 tgs: 267 data_time: 0.58s time: 119.39s eta: 14:23:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31062.0 tgs: 257 data_time: 0.90s time: 120.74s eta: 14:31:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31051.0 tgs: 261 data_time: 0.73s time: 118.85s eta: 14:15:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.251 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31772.0 tgs: 266 data_time: 1.03s time: 119.10s eta: 14:15:31 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30836.0 tgs: 258 data_time: 0.85s time: 119.13s eta: 14:13:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32108.0 tgs: 267 data_time: 0.78s time: 119.96s eta: 14:17:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30850.0 tgs: 257 data_time: 0.78s time: 119.87s eta: 14:15:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31463.0 tgs: 263 data_time: 0.53s time: 119.59s eta: 14:11:04 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31448.0 tgs: 263 data_time: 0.64s time: 119.22s eta: 14:06:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31334.0 tgs: 262 data_time: 0.85s time: 119.26s eta: 14:04:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31864.0 tgs: 267 data_time: 0.86s time: 119.27s eta: 14:02:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31647.0 tgs: 265 data_time: 0.80s time: 119.14s eta: 13:59:57 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31563.0 tgs: 264 data_time: 0.74s time: 119.12s eta: 13:57:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31902.0 tgs: 267 data_time: 1.05s time: 119.17s eta: 13:56:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 29759.0 tgs: 249 data_time: 0.77s time: 119.34s eta: 13:55:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.243 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 265 data_time: 1.01s time: 120.11s eta: 13:58:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31373.0 tgs: 255 data_time: 0.69s time: 122.56s eta: 14:13:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31224.0 tgs: 262 data_time: 0.66s time: 119.10s eta: 13:47:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.186 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30358.0 tgs: 254 data_time: 1.10s time: 119.37s eta: 13:47:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31375.0 tgs: 262 data_time: 0.98s time: 119.32s eta: 13:45:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32365.0 tgs: 271 data_time: 0.74s time: 119.33s eta: 13:43:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 268 data_time: 0.66s time: 119.11s eta: 13:39:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31269.0 tgs: 262 data_time: 0.70s time: 119.26s eta: 13:38:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31408.0 tgs: 263 data_time: 0.69s time: 119.13s eta: 13:36:03 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.224 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32101.0 tgs: 269 data_time: 0.99s time: 119.23s eta: 13:34:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31850.0 tgs: 266 data_time: 0.72s time: 119.46s eta: 13:34:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30710.0 tgs: 257 data_time: 0.90s time: 119.08s eta: 13:29:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31779.0 tgs: 267 data_time: 1.07s time: 118.93s eta: 13:26:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32041.0 tgs: 267 data_time: 0.46s time: 119.80s eta: 13:30:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.125 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30280.0 tgs: 253 data_time: 0.65s time: 119.26s eta: 13:24:58 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.192 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31354.0 tgs: 263 data_time: 1.03s time: 119.17s eta: 13:22:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29293.0 tgs: 245 data_time: 0.74s time: 119.34s eta: 13:21:32 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31774.0 tgs: 266 data_time: 0.67s time: 119.27s eta: 13:19:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.208 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 268 data_time: 0.71s time: 119.04s eta: 13:15:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32536.0 tgs: 272 data_time: 0.60s time: 119.36s eta: 13:15:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31876.0 tgs: 267 data_time: 0.94s time: 119.17s eta: 13:12:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31549.0 tgs: 264 data_time: 1.02s time: 119.46s eta: 13:12:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32604.0 tgs: 273 data_time: 0.88s time: 119.28s eta: 13:09:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31952.0 tgs: 268 data_time: 0.80s time: 119.15s eta: 13:06:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.113 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30917.0 tgs: 256 data_time: 0.60s time: 120.38s eta: 13:12:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32412.0 tgs: 269 data_time: 0.85s time: 120.16s eta: 13:09:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.108 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 264 data_time: 0.84s time: 119.50s eta: 13:02:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32242.0 tgs: 269 data_time: 0.72s time: 119.51s eta: 13:00:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 268 data_time: 0.71s time: 119.57s eta: 12:59:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.227 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30519.0 tgs: 254 data_time: 0.67s time: 119.80s eta: 12:58:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32014.0 tgs: 268 data_time: 0.63s time: 119.04s eta: 12:51:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32306.0 tgs: 271 data_time: 0.77s time: 119.05s eta: 12:49:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30696.0 tgs: 257 data_time: 0.96s time: 119.13s eta: 12:48:22 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31929.0 tgs: 267 data_time: 0.89s time: 119.39s eta: 12:48:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.213 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31661.0 tgs: 265 data_time: 1.03s time: 119.18s eta: 12:44:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30836.0 tgs: 258 data_time: 0.74s time: 119.44s eta: 12:44:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30787.0 tgs: 258 data_time: 0.80s time: 119.10s eta: 12:40:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31337.0 tgs: 263 data_time: 0.62s time: 119.08s eta: 12:38:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.207 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31863.0 tgs: 266 data_time: 1.10s time: 119.60s eta: 12:39:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.117 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29739.0 tgs: 250 data_time: 0.49s time: 118.78s eta: 12:32:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30929.0 tgs: 260 data_time: 1.05s time: 118.88s eta: 12:30:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31806.0 tgs: 266 data_time: 0.58s time: 119.18s eta: 12:30:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31801.0 tgs: 266 data_time: 0.88s time: 119.26s eta: 12:29:22 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31832.0 tgs: 266 data_time: 0.56s time: 119.34s eta: 12:27:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31829.0 tgs: 267 data_time: 0.64s time: 119.12s eta: 12:24:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32446.0 tgs: 271 data_time: 0.89s time: 119.34s eta: 12:23:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.108 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 264 data_time: 0.76s time: 119.26s eta: 12:21:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.214 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32031.0 tgs: 269 data_time: 0.65s time: 119.03s eta: 12:17:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30524.0 tgs: 251 data_time: 1.04s time: 121.13s eta: 12:28:58 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31527.0 tgs: 262 data_time: 0.94s time: 119.98s eta: 12:19:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32383.0 tgs: 272 data_time: 0.89s time: 118.70s eta: 12:09:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31619.0 tgs: 264 data_time: 0.70s time: 119.64s eta: 12:13:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31708.0 tgs: 264 data_time: 0.74s time: 119.83s eta: 12:12:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32314.0 tgs: 271 data_time: 0.87s time: 119.14s eta: 12:06:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32414.0 tgs: 271 data_time: 0.84s time: 119.17s eta: 12:04:58 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30908.0 tgs: 259 data_time: 0.79s time: 119.12s eta: 12:02:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31538.0 tgs: 264 data_time: 0.66s time: 119.18s eta: 12:01:03 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.202 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31544.0 tgs: 265 data_time: 0.71s time: 118.96s eta: 11:57:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30810.0 tgs: 257 data_time: 0.62s time: 119.71s eta: 12:00:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31610.0 tgs: 265 data_time: 0.69s time: 119.28s eta: 11:55:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.122 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31609.0 tgs: 265 data_time: 0.79s time: 119.27s eta: 11:53:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31766.0 tgs: 265 data_time: 0.85s time: 119.71s eta: 11:54:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31691.0 tgs: 266 data_time: 0.75s time: 118.78s eta: 11:46:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 266 data_time: 0.97s time: 119.16s eta: 11:46:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31304.0 tgs: 261 data_time: 0.57s time: 119.61s eta: 11:47:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30964.0 tgs: 260 data_time: 0.79s time: 118.83s eta: 11:41:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30226.0 tgs: 254 data_time: 0.72s time: 118.68s eta: 11:38:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31175.0 tgs: 261 data_time: 0.93s time: 119.22s eta: 11:39:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31823.0 tgs: 267 data_time: 0.89s time: 119.14s eta: 11:36:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 0.75s time: 119.07s eta: 11:34:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31877.0 tgs: 267 data_time: 1.00s time: 118.95s eta: 11:31:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.215 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31777.0 tgs: 267 data_time: 0.63s time: 118.91s eta: 11:29:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.207 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31103.0 tgs: 261 data_time: 1.06s time: 118.97s eta: 11:28:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31645.0 tgs: 265 data_time: 0.63s time: 119.29s eta: 11:27:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30994.0 tgs: 259 data_time: 0.64s time: 119.35s eta: 11:26:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 269 data_time: 0.68s time: 118.89s eta: 11:21:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32190.0 tgs: 270 data_time: 0.92s time: 119.21s eta: 11:21:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 269 data_time: 0.74s time: 119.71s eta: 11:22:21 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32161.0 tgs: 269 data_time: 0.88s time: 119.12s eta: 11:16:58 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32455.0 tgs: 272 data_time: 0.61s time: 119.12s eta: 11:15:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30945.0 tgs: 260 data_time: 0.71s time: 118.96s eta: 11:12:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.218 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32229.0 tgs: 270 data_time: 0.55s time: 119.09s eta: 11:10:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.63s time: 119.16s eta: 11:09:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.125 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29619.0 tgs: 248 data_time: 0.54s time: 118.97s eta: 11:06:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 262 data_time: 0.61s time: 119.47s eta: 11:07:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.201 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30573.0 tgs: 255 data_time: 0.91s time: 119.64s eta: 11:06:00 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31948.0 tgs: 267 data_time: 0.92s time: 119.31s eta: 11:02:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31760.0 tgs: 266 data_time: 0.97s time: 119.31s eta: 11:00:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31705.0 tgs: 266 data_time: 0.82s time: 118.82s eta: 10:55:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31282.0 tgs: 261 data_time: 0.84s time: 119.47s eta: 10:57:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.120 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31668.0 tgs: 265 data_time: 0.61s time: 119.37s eta: 10:54:31 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31174.0 tgs: 261 data_time: 0.88s time: 119.28s eta: 10:52:03 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.130 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31438.0 tgs: 263 data_time: 0.88s time: 119.25s eta: 10:49:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32226.0 tgs: 270 data_time: 1.03s time: 119.10s eta: 10:47:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32215.0 tgs: 269 data_time: 0.68s time: 119.53s eta: 10:47:27 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31281.0 tgs: 262 data_time: 0.83s time: 119.31s eta: 10:44:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.186 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31956.0 tgs: 267 data_time: 0.81s time: 119.49s eta: 10:43:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30871.0 tgs: 258 data_time: 0.68s time: 119.51s eta: 10:41:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32095.0 tgs: 269 data_time: 0.99s time: 118.99s eta: 10:36:35 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31526.0 tgs: 264 data_time: 0.69s time: 119.22s eta: 10:35:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32049.0 tgs: 269 data_time: 0.60s time: 119.06s eta: 10:32:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31574.0 tgs: 265 data_time: 0.58s time: 118.78s eta: 10:29:33 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.186 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31674.0 tgs: 264 data_time: 0.80s time: 119.87s eta: 10:33:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31536.0 tgs: 264 data_time: 0.87s time: 119.38s eta: 10:28:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31959.0 tgs: 268 data_time: 0.85s time: 118.83s eta: 10:23:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.220 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31953.0 tgs: 268 data_time: 0.72s time: 118.99s eta: 10:22:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.183 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31534.0 tgs: 264 data_time: 0.58s time: 119.29s eta: 10:22:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31346.0 tgs: 262 data_time: 0.72s time: 119.21s eta: 10:19:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31427.0 tgs: 263 data_time: 0.56s time: 119.06s eta: 10:17:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31588.0 tgs: 264 data_time: 0.82s time: 119.48s eta: 10:17:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.180 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31644.0 tgs: 265 data_time: 0.72s time: 119.37s eta: 10:14:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31681.0 tgs: 265 data_time: 0.68s time: 119.45s eta: 10:13:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.211 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30269.0 tgs: 253 data_time: 0.91s time: 119.29s eta: 10:10:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.249 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31607.0 tgs: 265 data_time: 0.93s time: 119.11s eta: 10:07:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31978.0 tgs: 267 data_time: 1.04s time: 119.70s eta: 10:08:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.236 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 267 data_time: 0.69s time: 119.84s eta: 10:07:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.246 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32009.0 tgs: 263 data_time: 1.06s time: 121.55s eta: 10:13:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 265 data_time: 0.68s time: 119.26s eta: 10:00:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31657.0 tgs: 266 data_time: 0.81s time: 118.81s eta: 9:56:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.170 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31748.0 tgs: 265 data_time: 0.90s time: 119.56s eta: 9:57:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 266 data_time: 0.86s time: 119.71s eta: 9:56:32 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31472.0 tgs: 263 data_time: 0.74s time: 119.26s eta: 9:52:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.197 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30746.0 tgs: 257 data_time: 0.95s time: 119.33s eta: 9:50:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.216 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 264 data_time: 1.24s time: 119.23s eta: 9:48:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.170 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32046.0 tgs: 265 data_time: 0.77s time: 120.76s eta: 9:53:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31795.0 tgs: 266 data_time: 1.03s time: 119.27s eta: 9:44:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31162.0 tgs: 260 data_time: 0.65s time: 119.46s eta: 9:43:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 261 data_time: 0.67s time: 119.55s eta: 9:41:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30543.0 tgs: 256 data_time: 0.58s time: 119.20s eta: 9:38:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32342.0 tgs: 270 data_time: 0.70s time: 119.48s eta: 9:37:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.120 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31087.0 tgs: 261 data_time: 0.82s time: 119.05s eta: 9:33:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31590.0 tgs: 264 data_time: 0.62s time: 119.25s eta: 9:32:22 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31959.0 tgs: 268 data_time: 0.66s time: 119.16s eta: 9:29:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32311.0 tgs: 271 data_time: 0.80s time: 119.22s eta: 9:28:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32572.0 tgs: 272 data_time: 0.78s time: 119.47s eta: 9:27:27 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31915.0 tgs: 268 data_time: 0.70s time: 118.98s eta: 9:23:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32347.0 tgs: 271 data_time: 0.65s time: 119.03s eta: 9:21:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31010.0 tgs: 259 data_time: 0.71s time: 119.53s eta: 9:21:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.366 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 0.90s time: 119.60s eta: 9:20:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.119 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32379.0 tgs: 271 data_time: 0.84s time: 119.12s eta: 9:15:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.117 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31877.0 tgs: 267 data_time: 0.75s time: 119.00s eta: 9:13:21 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31387.0 tgs: 264 data_time: 0.97s time: 118.88s eta: 9:10:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31216.0 tgs: 261 data_time: 0.78s time: 119.32s eta: 9:10:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31204.0 tgs: 262 data_time: 0.79s time: 119.09s eta: 9:07:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32276.0 tgs: 271 data_time: 0.91s time: 118.86s eta: 9:04:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.192 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 265 data_time: 0.89s time: 119.67s eta: 9:06:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 268 data_time: 0.70s time: 119.44s eta: 9:03:27 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32374.0 tgs: 271 data_time: 0.73s time: 119.43s eta: 9:01:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.193 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31117.0 tgs: 262 data_time: 0.85s time: 118.74s eta: 8:56:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30842.0 tgs: 259 data_time: 0.95s time: 118.87s eta: 8:54:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.099 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30180.0 tgs: 252 data_time: 0.75s time: 119.33s eta: 8:55:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31017.0 tgs: 260 data_time: 0.65s time: 119.15s eta: 8:52:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.223 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31638.0 tgs: 265 data_time: 0.90s time: 119.26s eta: 8:50:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32513.0 tgs: 269 data_time: 0.89s time: 120.49s eta: 8:54:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31625.0 tgs: 265 data_time: 0.87s time: 119.22s eta: 8:46:33 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.189 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32193.0 tgs: 270 data_time: 0.98s time: 119.07s eta: 8:43:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31839.0 tgs: 263 data_time: 1.08s time: 120.83s eta: 8:49:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32062.0 tgs: 266 data_time: 0.91s time: 120.19s eta: 8:44:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31753.0 tgs: 267 data_time: 0.88s time: 118.83s eta: 8:36:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31790.0 tgs: 266 data_time: 0.77s time: 119.35s eta: 8:37:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31170.0 tgs: 261 data_time: 0.71s time: 119.34s eta: 8:35:08 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 266 data_time: 0.87s time: 119.05s eta: 8:31:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31159.0 tgs: 261 data_time: 0.98s time: 119.16s eta: 8:30:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32203.0 tgs: 270 data_time: 0.88s time: 119.04s eta: 8:27:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31528.0 tgs: 264 data_time: 0.59s time: 119.12s eta: 8:26:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31863.0 tgs: 267 data_time: 0.87s time: 119.24s eta: 8:24:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32267.0 tgs: 270 data_time: 0.98s time: 119.22s eta: 8:22:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.184 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31629.0 tgs: 264 data_time: 0.65s time: 119.38s eta: 8:21:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31190.0 tgs: 260 data_time: 0.76s time: 119.78s eta: 8:21:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32151.0 tgs: 269 data_time: 0.61s time: 119.19s eta: 8:16:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.112 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31971.0 tgs: 268 data_time: 0.60s time: 118.98s eta: 8:13:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.206 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31363.0 tgs: 261 data_time: 0.57s time: 119.90s eta: 8:15:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31813.0 tgs: 266 data_time: 0.88s time: 119.55s eta: 8:12:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.108 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30647.0 tgs: 256 data_time: 0.57s time: 119.60s eta: 8:10:20 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31515.0 tgs: 264 data_time: 0.83s time: 119.00s eta: 8:05:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31065.0 tgs: 259 data_time: 0.74s time: 119.85s eta: 8:07:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31305.0 tgs: 261 data_time: 0.65s time: 119.70s eta: 8:04:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32043.0 tgs: 268 data_time: 0.74s time: 119.41s eta: 8:01:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.190 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32033.0 tgs: 268 data_time: 0.49s time: 119.29s eta: 7:59:08 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31947.0 tgs: 268 data_time: 0.86s time: 119.19s eta: 7:56:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.104 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31553.0 tgs: 264 data_time: 0.67s time: 119.20s eta: 7:54:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31340.0 tgs: 263 data_time: 0.69s time: 118.93s eta: 7:51:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32067.0 tgs: 269 data_time: 0.63s time: 118.97s eta: 7:49:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31589.0 tgs: 264 data_time: 0.72s time: 119.42s eta: 7:49:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32091.0 tgs: 269 data_time: 0.71s time: 119.04s eta: 7:46:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31432.0 tgs: 263 data_time: 0.81s time: 119.07s eta: 7:44:22 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.161 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31040.0 tgs: 258 data_time: 0.51s time: 120.01s eta: 7:46:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31742.0 tgs: 264 data_time: 0.58s time: 119.83s eta: 7:43:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30989.0 tgs: 260 data_time: 0.68s time: 119.16s eta: 7:38:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31437.0 tgs: 264 data_time: 0.82s time: 118.97s eta: 7:36:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31398.0 tgs: 263 data_time: 0.75s time: 119.32s eta: 7:35:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 269 data_time: 0.70s time: 119.39s eta: 7:33:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30974.0 tgs: 260 data_time: 0.53s time: 119.07s eta: 7:30:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 267 data_time: 0.74s time: 118.99s eta: 7:28:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31009.0 tgs: 260 data_time: 0.63s time: 119.24s eta: 7:27:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32384.0 tgs: 272 data_time: 0.83s time: 119.00s eta: 7:24:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.197 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31920.0 tgs: 268 data_time: 0.83s time: 118.93s eta: 7:22:00 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31730.0 tgs: 266 data_time: 0.78s time: 119.08s eta: 7:20:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32118.0 tgs: 269 data_time: 0.71s time: 118.99s eta: 7:18:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31369.0 tgs: 263 data_time: 0.88s time: 119.03s eta: 7:16:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 268 data_time: 0.96s time: 119.25s eta: 7:15:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31593.0 tgs: 264 data_time: 0.65s time: 119.36s eta: 7:13:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 271 data_time: 0.88s time: 118.94s eta: 7:10:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32426.0 tgs: 272 data_time: 0.90s time: 119.20s eta: 7:09:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30817.0 tgs: 257 data_time: 0.67s time: 119.58s eta: 7:08:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.217 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31613.0 tgs: 264 data_time: 0.90s time: 119.67s eta: 7:06:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31541.0 tgs: 265 data_time: 0.75s time: 118.95s eta: 7:02:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.176 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31533.0 tgs: 264 data_time: 1.02s time: 119.30s eta: 7:01:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31066.0 tgs: 259 data_time: 0.79s time: 119.52s eta: 7:00:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 263 data_time: 0.67s time: 119.43s eta: 6:57:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30787.0 tgs: 258 data_time: 0.86s time: 118.91s eta: 6:54:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32228.0 tgs: 270 data_time: 1.06s time: 119.35s eta: 6:53:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31807.0 tgs: 267 data_time: 0.91s time: 118.94s eta: 6:50:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31413.0 tgs: 263 data_time: 0.65s time: 119.26s eta: 6:49:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31571.0 tgs: 265 data_time: 0.59s time: 119.02s eta: 6:46:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32209.0 tgs: 270 data_time: 0.87s time: 119.22s eta: 6:45:21 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.108 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31538.0 tgs: 264 data_time: 0.59s time: 119.38s eta: 6:43:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31429.0 tgs: 264 data_time: 0.69s time: 119.01s eta: 6:40:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.188 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30966.0 tgs: 260 data_time: 0.62s time: 119.01s eta: 6:38:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.202 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31853.0 tgs: 266 data_time: 0.68s time: 119.46s eta: 6:38:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 270 data_time: 0.99s time: 118.80s eta: 6:34:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31018.0 tgs: 259 data_time: 0.77s time: 119.44s eta: 6:34:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.197 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31845.0 tgs: 266 data_time: 0.89s time: 119.57s eta: 6:32:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 268 data_time: 0.90s time: 119.38s eta: 6:29:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31379.0 tgs: 262 data_time: 0.59s time: 119.71s eta: 6:29:04 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.65s time: 119.36s eta: 6:25:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.190 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30785.0 tgs: 259 data_time: 0.79s time: 118.82s eta: 6:22:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31894.0 tgs: 267 data_time: 1.00s time: 119.02s eta: 6:20:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.111 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31839.0 tgs: 266 data_time: 0.67s time: 119.40s eta: 6:20:04 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31476.0 tgs: 263 data_time: 0.72s time: 119.34s eta: 6:17:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31386.0 tgs: 263 data_time: 0.84s time: 119.05s eta: 6:14:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32280.0 tgs: 270 data_time: 0.80s time: 119.39s eta: 6:14:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31022.0 tgs: 259 data_time: 1.00s time: 119.39s eta: 6:12:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30370.0 tgs: 254 data_time: 0.91s time: 119.28s eta: 6:09:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.112 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29926.0 tgs: 251 data_time: 0.57s time: 119.10s eta: 6:07:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31224.0 tgs: 256 data_time: 0.69s time: 121.96s eta: 6:14:01 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31180.0 tgs: 261 data_time: 0.74s time: 119.39s eta: 6:04:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.190 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31966.0 tgs: 267 data_time: 0.66s time: 119.57s eta: 6:02:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31988.0 tgs: 268 data_time: 0.82s time: 119.26s eta: 5:59:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32020.0 tgs: 268 data_time: 0.79s time: 119.26s eta: 5:57:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31459.0 tgs: 264 data_time: 0.75s time: 118.92s eta: 5:54:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32038.0 tgs: 269 data_time: 0.51s time: 119.02s eta: 5:53:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30936.0 tgs: 259 data_time: 0.75s time: 119.08s eta: 5:51:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31889.0 tgs: 267 data_time: 0.80s time: 119.29s eta: 5:49:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30581.0 tgs: 257 data_time: 1.00s time: 118.97s eta: 5:46:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.113 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31326.0 tgs: 262 data_time: 0.77s time: 119.27s eta: 5:45:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30447.0 tgs: 255 data_time: 0.58s time: 119.33s eta: 5:44:04 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32490.0 tgs: 272 data_time: 0.82s time: 119.06s eta: 5:41:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32449.0 tgs: 271 data_time: 0.80s time: 119.49s eta: 5:40:32 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31621.0 tgs: 264 data_time: 0.74s time: 119.39s eta: 5:38:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31550.0 tgs: 263 data_time: 0.84s time: 119.57s eta: 5:36:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31934.0 tgs: 268 data_time: 0.79s time: 118.98s eta: 5:33:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31820.0 tgs: 267 data_time: 0.81s time: 118.95s eta: 5:31:04 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31599.0 tgs: 265 data_time: 0.82s time: 119.24s eta: 5:29:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32334.0 tgs: 270 data_time: 0.81s time: 119.40s eta: 5:28:21 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.101 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31177.0 tgs: 260 data_time: 0.59s time: 119.47s eta: 5:26:32 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31289.0 tgs: 263 data_time: 0.74s time: 118.94s eta: 5:23:07 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.208 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32117.0 tgs: 268 data_time: 0.47s time: 119.63s eta: 5:23:00 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.291 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31353.0 tgs: 262 data_time: 0.88s time: 119.35s eta: 5:20:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 268 data_time: 0.99s time: 119.12s eta: 5:17:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31818.0 tgs: 266 data_time: 0.58s time: 119.43s eta: 5:16:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.196 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32054.0 tgs: 268 data_time: 0.90s time: 119.57s eta: 5:14:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31928.0 tgs: 268 data_time: 0.59s time: 119.01s eta: 5:11:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31699.0 tgs: 266 data_time: 0.77s time: 119.00s eta: 5:09:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31672.0 tgs: 265 data_time: 0.85s time: 119.23s eta: 5:08:00 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.185 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31600.0 tgs: 264 data_time: 0.63s time: 119.68s eta: 5:07:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31804.0 tgs: 267 data_time: 0.56s time: 119.12s eta: 5:03:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32066.0 tgs: 269 data_time: 0.81s time: 119.07s eta: 5:01:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32074.0 tgs: 263 data_time: 0.90s time: 121.52s eta: 5:05:48 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31359.0 tgs: 262 data_time: 1.15s time: 119.57s eta: 4:58:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 268 data_time: 0.99s time: 119.50s eta: 4:56:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31262.0 tgs: 261 data_time: 0.76s time: 119.73s eta: 4:55:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.196 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32047.0 tgs: 268 data_time: 0.65s time: 119.53s eta: 4:52:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31771.0 tgs: 266 data_time: 0.66s time: 119.41s eta: 4:50:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31966.0 tgs: 268 data_time: 0.63s time: 119.21s eta: 4:48:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31697.0 tgs: 266 data_time: 0.64s time: 118.88s eta: 4:45:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.227 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31642.0 tgs: 265 data_time: 0.81s time: 119.04s eta: 4:43:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 257 data_time: 0.75s time: 119.98s eta: 4:43:57 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31475.0 tgs: 264 data_time: 0.82s time: 119.15s eta: 4:39:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31795.0 tgs: 267 data_time: 0.74s time: 119.01s eta: 4:37:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.126 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31221.0 tgs: 262 data_time: 0.52s time: 119.06s eta: 4:35:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.215 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31462.0 tgs: 263 data_time: 0.70s time: 119.38s eta: 4:34:33 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32505.0 tgs: 272 data_time: 0.62s time: 119.41s eta: 4:32:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.106 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31616.0 tgs: 265 data_time: 0.68s time: 118.94s eta: 4:29:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31983.0 tgs: 267 data_time: 0.79s time: 119.60s eta: 4:29:05 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32021.0 tgs: 268 data_time: 1.06s time: 119.25s eta: 4:26:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.126 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30121.0 tgs: 252 data_time: 0.49s time: 119.06s eta: 4:23:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31574.0 tgs: 265 data_time: 0.74s time: 118.85s eta: 4:21:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.111 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31181.0 tgs: 262 data_time: 0.79s time: 118.76s eta: 4:19:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31591.0 tgs: 264 data_time: 0.96s time: 119.42s eta: 4:18:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32086.0 tgs: 268 data_time: 0.65s time: 119.40s eta: 4:16:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31915.0 tgs: 267 data_time: 0.84s time: 119.34s eta: 4:14:35 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.119 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32347.0 tgs: 271 data_time: 0.63s time: 119.05s eta: 4:11:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32068.0 tgs: 269 data_time: 0.73s time: 119.13s eta: 4:10:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.210 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31296.0 tgs: 262 data_time: 0.66s time: 119.13s eta: 4:08:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31432.0 tgs: 264 data_time: 0.67s time: 118.96s eta: 4:05:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31809.0 tgs: 266 data_time: 0.69s time: 119.15s eta: 4:04:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.100 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31912.0 tgs: 267 data_time: 1.03s time: 119.50s eta: 4:02:58 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31972.0 tgs: 268 data_time: 0.84s time: 119.25s eta: 4:00:28 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32571.0 tgs: 273 data_time: 0.90s time: 119.19s eta: 3:58:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29691.0 tgs: 248 data_time: 0.76s time: 119.36s eta: 3:56:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31820.0 tgs: 266 data_time: 0.78s time: 119.37s eta: 3:54:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32297.0 tgs: 269 data_time: 0.62s time: 119.95s eta: 3:53:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32242.0 tgs: 270 data_time: 0.68s time: 119.33s eta: 3:50:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32043.0 tgs: 267 data_time: 0.69s time: 119.63s eta: 3:49:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 261 data_time: 0.66s time: 118.86s eta: 3:45:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31691.0 tgs: 264 data_time: 0.62s time: 119.70s eta: 3:45:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31852.0 tgs: 266 data_time: 0.61s time: 119.33s eta: 3:42:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.203 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32414.0 tgs: 271 data_time: 0.85s time: 119.26s eta: 3:40:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30859.0 tgs: 259 data_time: 0.83s time: 119.01s eta: 3:38:10 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31071.0 tgs: 260 data_time: 0.64s time: 119.43s eta: 3:36:57 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31843.0 tgs: 267 data_time: 0.81s time: 119.03s eta: 3:34:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31289.0 tgs: 262 data_time: 0.95s time: 118.99s eta: 3:32:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31659.0 tgs: 265 data_time: 0.73s time: 119.22s eta: 3:30:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31256.0 tgs: 260 data_time: 0.93s time: 120.11s eta: 3:30:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32127.0 tgs: 269 data_time: 0.62s time: 119.14s eta: 3:26:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32299.0 tgs: 270 data_time: 0.70s time: 119.31s eta: 3:24:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31474.0 tgs: 264 data_time: 0.75s time: 119.08s eta: 3:22:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.223 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31501.0 tgs: 263 data_time: 0.66s time: 119.43s eta: 3:21:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31576.0 tgs: 264 data_time: 0.61s time: 119.29s eta: 3:18:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31755.0 tgs: 266 data_time: 0.63s time: 119.22s eta: 3:16:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30981.0 tgs: 260 data_time: 0.90s time: 119.14s eta: 3:14:35 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 28869.0 tgs: 242 data_time: 0.60s time: 119.02s eta: 3:12:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 267 data_time: 0.64s time: 119.18s eta: 3:10:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 264 data_time: 0.92s time: 121.00s eta: 3:11:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 265 data_time: 0.66s time: 119.51s eta: 3:07:13 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32573.0 tgs: 273 data_time: 0.77s time: 119.26s eta: 3:04:51 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.3GB text_tokens: 30164.0 tgs: 253 data_time: 0.54s time: 118.99s eta: 3:02:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31251.0 tgs: 262 data_time: 0.94s time: 119.22s eta: 3:00:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31281.0 tgs: 261 data_time: 0.80s time: 119.46s eta: 2:59:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.186 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31304.0 tgs: 263 data_time: 0.61s time: 118.98s eta: 2:56:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31612.0 tgs: 264 data_time: 0.62s time: 119.35s eta: 2:55:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.105 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31385.0 tgs: 262 data_time: 0.71s time: 119.42s eta: 2:53:09 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.218 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32439.0 tgs: 271 data_time: 0.75s time: 119.35s eta: 2:51:04 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31777.0 tgs: 261 data_time: 0.65s time: 121.46s eta: 2:52:04 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32406.0 tgs: 271 data_time: 0.79s time: 119.24s eta: 2:46:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31636.0 tgs: 264 data_time: 0.87s time: 119.56s eta: 2:45:23 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32324.0 tgs: 271 data_time: 0.65s time: 118.91s eta: 2:42:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 268 data_time: 0.74s time: 119.18s eta: 2:40:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32160.0 tgs: 270 data_time: 0.89s time: 118.97s eta: 2:38:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30511.0 tgs: 255 data_time: 0.95s time: 119.23s eta: 2:36:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31776.0 tgs: 266 data_time: 0.79s time: 119.10s eta: 2:34:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30965.0 tgs: 259 data_time: 0.66s time: 119.43s eta: 2:33:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32050.0 tgs: 266 data_time: 0.91s time: 120.23s eta: 2:32:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30958.0 tgs: 259 data_time: 0.62s time: 119.29s eta: 2:29:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31489.0 tgs: 264 data_time: 0.93s time: 118.88s eta: 2:26:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 268 data_time: 0.82s time: 119.52s eta: 2:25:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 265 data_time: 1.01s time: 119.07s eta: 2:22:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32517.0 tgs: 272 data_time: 0.63s time: 119.14s eta: 2:20:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31450.0 tgs: 264 data_time: 0.63s time: 119.05s eta: 2:18:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.215 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32353.0 tgs: 269 data_time: 0.71s time: 119.84s eta: 2:17:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32250.0 tgs: 270 data_time: 0.88s time: 119.33s eta: 2:15:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.68s time: 119.37s eta: 2:13:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31558.0 tgs: 263 data_time: 0.83s time: 119.68s eta: 2:11:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32166.0 tgs: 269 data_time: 0.82s time: 119.49s eta: 2:09:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31209.0 tgs: 262 data_time: 0.72s time: 118.96s eta: 2:06:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32587.0 tgs: 272 data_time: 0.83s time: 119.38s eta: 2:05:20 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31579.0 tgs: 264 data_time: 0.88s time: 119.27s eta: 2:03:14 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32076.0 tgs: 269 data_time: 0.75s time: 119.23s eta: 2:01:12 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 266 data_time: 0.44s time: 119.51s eta: 1:59:30 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31143.0 tgs: 260 data_time: 0.76s time: 119.42s eta: 1:57:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32146.0 tgs: 268 data_time: 0.75s time: 119.54s eta: 1:55:33 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.64s time: 119.30s eta: 1:53:20 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31591.0 tgs: 264 data_time: 0.85s time: 119.26s eta: 1:51:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31906.0 tgs: 268 data_time: 0.58s time: 118.93s eta: 1:49:00 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32128.0 tgs: 268 data_time: 0.57s time: 119.65s eta: 1:47:41 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31159.0 tgs: 260 data_time: 0.97s time: 119.69s eta: 1:45:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31968.0 tgs: 267 data_time: 0.42s time: 119.35s eta: 1:43:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31429.0 tgs: 263 data_time: 0.72s time: 119.31s eta: 1:41:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31845.0 tgs: 266 data_time: 1.00s time: 119.28s eta: 1:39:24 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31843.0 tgs: 267 data_time: 0.68s time: 118.91s eta: 1:37:06 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32139.0 tgs: 269 data_time: 0.61s time: 119.16s eta: 1:35:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32245.0 tgs: 270 data_time: 0.62s time: 119.06s eta: 1:33:15 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 267 data_time: 0.72s time: 119.25s eta: 1:31:25 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31938.0 tgs: 266 data_time: 0.70s time: 119.73s eta: 1:29:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 268 data_time: 0.76s time: 119.01s eta: 1:27:16 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 270 data_time: 0.42s time: 118.87s eta: 1:25:11 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 268 data_time: 0.61s time: 119.48s eta: 1:23:37 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31439.0 tgs: 263 data_time: 0.72s time: 119.48s eta: 1:21:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31150.0 tgs: 261 data_time: 0.83s time: 118.94s eta: 1:19:17 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31051.0 tgs: 261 data_time: 0.57s time: 118.94s eta: 1:17:18 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32440.0 tgs: 265 data_time: 0.63s time: 122.13s eta: 1:17:20 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31572.0 tgs: 264 data_time: 0.72s time: 119.36s eta: 1:13:36 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31814.0 tgs: 267 data_time: 0.72s time: 119.06s eta: 1:11:26 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32165.0 tgs: 270 data_time: 0.70s time: 118.92s eta: 1:09:22 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32068.0 tgs: 269 data_time: 0.98s time: 119.11s eta: 1:07:29 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.53s time: 119.42s eta: 1:05:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31668.0 tgs: 265 data_time: 0.69s time: 119.35s eta: 1:03:39 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32098.0 tgs: 269 data_time: 0.74s time: 119.17s eta: 1:01:34 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31767.0 tgs: 266 data_time: 0.83s time: 119.27s eta: 0:59:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32539.0 tgs: 272 data_time: 0.95s time: 119.45s eta: 0:57:44 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30903.0 tgs: 258 data_time: 0.80s time: 119.40s eta: 0:55:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 266 data_time: 0.68s time: 119.48s eta: 0:53:45 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32307.0 tgs: 270 data_time: 0.71s time: 119.37s eta: 0:51:43 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32450.0 tgs: 271 data_time: 0.58s time: 119.47s eta: 0:49:46 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31319.0 tgs: 263 data_time: 0.62s time: 118.89s eta: 0:47:33 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 271 data_time: 0.58s time: 119.26s eta: 0:45:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31776.0 tgs: 266 data_time: 0.78s time: 119.20s eta: 0:43:42 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32478.0 tgs: 272 data_time: 0.71s time: 119.00s eta: 0:41:38 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31743.0 tgs: 266 data_time: 0.75s time: 119.03s eta: 0:39:40 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32052.0 tgs: 267 data_time: 0.73s time: 119.66s eta: 0:37:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31704.0 tgs: 265 data_time: 0.64s time: 119.28s eta: 0:35:47 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.201 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 268 data_time: 0.79s time: 119.75s eta: 0:33:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31468.0 tgs: 259 data_time: 0.58s time: 121.23s eta: 0:32:19 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32287.0 tgs: 270 data_time: 0.68s time: 119.38s eta: 0:29:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.59s time: 119.32s eta: 0:27:50 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31782.0 tgs: 265 data_time: 0.52s time: 119.64s eta: 0:25:55 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31282.0 tgs: 262 data_time: 0.83s time: 119.10s eta: 0:23:49 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31285.0 tgs: 262 data_time: 0.75s time: 119.38s eta: 0:21:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 263 data_time: 0.65s time: 119.26s eta: 0:19:52 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32422.0 tgs: 271 data_time: 0.83s time: 119.24s eta: 0:17:53 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.094 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31305.0 tgs: 261 data_time: 0.58s time: 119.90s eta: 0:15:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.079 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31974.0 tgs: 268 data_time: 0.60s time: 119.19s eta: 0:13:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30043.0 tgs: 252 data_time: 0.80s time: 119.09s eta: 0:11:54 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 264 data_time: 0.66s time: 119.32s eta: 0:09:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31639.0 tgs: 265 data_time: 0.65s time: 119.17s eta: 0:07:56 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31287.0 tgs: 261 data_time: 0.63s time: 119.69s eta: 0:05:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 259 data_time: 0.67s time: 121.42s eta: 0:04:02 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32169.0 tgs: 269 data_time: 0.63s time: 119.21s eta: 0:01:59 +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.28 seconds, peak gpu memory 17.4G +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 41][DP 10][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank44.log b/20250121165312/rank44.log new file mode 100644 index 0000000000000000000000000000000000000000..87be0f5eaae06862d5f16e2f94e22cecce3ce471 --- /dev/null +++ b/20250121165312/rank44.log @@ -0,0 +1,851 @@ +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:46][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.51 seconds, peak gpu memory 13.4G +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.359 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 31361.0 tgs: 205 data_time: 1.74s time: 152.67s eta: 1 day, 1:08:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.243 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32373.0 tgs: 269 data_time: 1.23s time: 120.10s eta: 19:44:57 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.159 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31271.0 tgs: 261 data_time: 1.18s time: 119.69s eta: 19:38:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.222 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31959.0 tgs: 267 data_time: 0.85s time: 119.66s eta: 19:36:36 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.239 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32119.0 tgs: 269 data_time: 0.63s time: 119.31s eta: 19:31:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.192 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31094.0 tgs: 260 data_time: 0.83s time: 119.53s eta: 19:31:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.212 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 259 data_time: 1.02s time: 122.96s eta: 20:02:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.324 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31574.0 tgs: 263 data_time: 0.79s time: 119.89s eta: 19:30:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31714.0 tgs: 262 data_time: 0.81s time: 120.65s eta: 19:36:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.224 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31891.0 tgs: 265 data_time: 0.88s time: 119.92s eta: 19:27:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.286 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31327.0 tgs: 260 data_time: 0.88s time: 120.46s eta: 19:30:27 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 0.84s time: 119.34s eta: 19:17:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.217 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31061.0 tgs: 259 data_time: 0.76s time: 119.68s eta: 19:18:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30891.0 tgs: 259 data_time: 0.76s time: 119.17s eta: 19:11:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31341.0 tgs: 262 data_time: 0.87s time: 119.20s eta: 19:10:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31438.0 tgs: 264 data_time: 0.91s time: 118.93s eta: 19:05:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 269 data_time: 0.67s time: 119.15s eta: 19:05:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 1.00s time: 119.18s eta: 19:04:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 262 data_time: 0.77s time: 119.40s eta: 19:04:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.78s time: 119.33s eta: 19:01:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 264 data_time: 0.70s time: 120.31s eta: 19:08:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 262 data_time: 0.79s time: 122.92s eta: 19:31:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31172.0 tgs: 260 data_time: 0.99s time: 119.48s eta: 18:57:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.230 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32324.0 tgs: 271 data_time: 1.08s time: 119.16s eta: 18:51:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 271 data_time: 0.93s time: 119.25s eta: 18:50:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31182.0 tgs: 261 data_time: 0.85s time: 119.39s eta: 18:50:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31901.0 tgs: 266 data_time: 0.77s time: 119.66s eta: 18:50:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30544.0 tgs: 255 data_time: 1.05s time: 119.39s eta: 18:46:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31830.0 tgs: 265 data_time: 0.77s time: 119.77s eta: 18:47:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.239 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31792.0 tgs: 245 data_time: 1.33s time: 129.31s eta: 20:15:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31968.0 tgs: 267 data_time: 0.91s time: 119.57s eta: 18:41:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31133.0 tgs: 260 data_time: 0.89s time: 119.39s eta: 18:38:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.223 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 267 data_time: 0.85s time: 120.66s eta: 18:48:09 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31570.0 tgs: 264 data_time: 1.02s time: 119.40s eta: 18:34:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32328.0 tgs: 271 data_time: 0.99s time: 119.16s eta: 18:30:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 265 data_time: 0.95s time: 119.07s eta: 18:27:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31631.0 tgs: 265 data_time: 1.01s time: 119.13s eta: 18:25:57 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31761.0 tgs: 263 data_time: 0.76s time: 120.39s eta: 18:35:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31545.0 tgs: 260 data_time: 0.86s time: 121.17s eta: 18:40:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30209.0 tgs: 250 data_time: 0.76s time: 120.66s eta: 18:34:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.249 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32351.0 tgs: 270 data_time: 0.63s time: 119.73s eta: 18:23:32 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31789.0 tgs: 256 data_time: 0.75s time: 123.88s eta: 18:59:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32049.0 tgs: 268 data_time: 0.76s time: 119.57s eta: 18:18:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32186.0 tgs: 269 data_time: 0.79s time: 119.31s eta: 18:13:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31935.0 tgs: 267 data_time: 0.85s time: 119.23s eta: 18:10:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 271 data_time: 0.78s time: 119.60s eta: 18:12:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32027.0 tgs: 268 data_time: 0.88s time: 119.27s eta: 18:07:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31765.0 tgs: 266 data_time: 0.59s time: 119.02s eta: 18:03:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31151.0 tgs: 261 data_time: 0.83s time: 119.10s eta: 18:01:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32470.0 tgs: 272 data_time: 0.84s time: 119.27s eta: 18:01:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.128 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.98s time: 119.33s eta: 17:59:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32319.0 tgs: 269 data_time: 0.90s time: 120.02s eta: 18:04:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30805.0 tgs: 258 data_time: 0.87s time: 119.11s eta: 17:53:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32262.0 tgs: 270 data_time: 0.79s time: 119.41s eta: 17:54:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32600.0 tgs: 273 data_time: 0.84s time: 119.00s eta: 17:49:01 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.81s time: 119.03s eta: 17:47:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32253.0 tgs: 257 data_time: 0.97s time: 125.36s eta: 18:41:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29875.0 tgs: 248 data_time: 0.79s time: 120.10s eta: 17:52:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31345.0 tgs: 262 data_time: 0.77s time: 119.36s eta: 17:44:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30572.0 tgs: 256 data_time: 0.95s time: 119.35s eta: 17:42:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31052.0 tgs: 259 data_time: 0.87s time: 119.50s eta: 17:41:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 249 data_time: 0.75s time: 126.05s eta: 18:37:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31868.0 tgs: 259 data_time: 0.85s time: 122.63s eta: 18:05:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31269.0 tgs: 258 data_time: 0.80s time: 121.16s eta: 17:50:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31590.0 tgs: 265 data_time: 0.70s time: 119.03s eta: 17:29:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32429.0 tgs: 272 data_time: 0.75s time: 118.86s eta: 17:25:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29122.0 tgs: 234 data_time: 0.87s time: 124.00s eta: 18:09:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31224.0 tgs: 250 data_time: 0.72s time: 124.57s eta: 18:12:01 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31914.0 tgs: 267 data_time: 1.08s time: 119.40s eta: 17:24:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.118 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31436.0 tgs: 262 data_time: 1.06s time: 119.70s eta: 17:25:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.69s time: 119.53s eta: 17:21:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32406.0 tgs: 272 data_time: 0.86s time: 118.86s eta: 17:14:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31471.0 tgs: 262 data_time: 0.78s time: 119.85s eta: 17:20:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32611.0 tgs: 272 data_time: 0.79s time: 119.53s eta: 17:15:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30756.0 tgs: 252 data_time: 0.84s time: 121.86s eta: 17:34:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31603.0 tgs: 255 data_time: 0.62s time: 123.51s eta: 17:46:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 266 data_time: 0.63s time: 119.01s eta: 17:05:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.117 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30407.0 tgs: 255 data_time: 0.83s time: 119.23s eta: 17:05:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31496.0 tgs: 264 data_time: 0.99s time: 119.14s eta: 17:02:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 267 data_time: 0.85s time: 119.25s eta: 17:01:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31305.0 tgs: 262 data_time: 0.89s time: 119.04s eta: 16:57:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31975.0 tgs: 268 data_time: 1.03s time: 119.28s eta: 16:57:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31974.0 tgs: 268 data_time: 0.69s time: 119.16s eta: 16:54:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.131 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31648.0 tgs: 264 data_time: 0.66s time: 119.58s eta: 16:56:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32343.0 tgs: 271 data_time: 1.02s time: 119.13s eta: 16:50:36 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31124.0 tgs: 261 data_time: 0.78s time: 118.84s eta: 16:46:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 269 data_time: 0.50s time: 119.02s eta: 16:45:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31889.0 tgs: 266 data_time: 0.66s time: 119.85s eta: 16:50:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32260.0 tgs: 270 data_time: 0.77s time: 119.36s eta: 16:44:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31721.0 tgs: 266 data_time: 0.50s time: 119.01s eta: 16:39:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32135.0 tgs: 265 data_time: 0.96s time: 120.93s eta: 16:53:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30808.0 tgs: 258 data_time: 0.96s time: 118.99s eta: 16:35:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.257 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31500.0 tgs: 264 data_time: 0.97s time: 119.29s eta: 16:36:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.274 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 265 data_time: 1.04s time: 119.23s eta: 16:33:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 268 data_time: 0.71s time: 119.37s eta: 16:32:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32157.0 tgs: 268 data_time: 0.80s time: 119.70s eta: 16:33:32 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32241.0 tgs: 270 data_time: 1.06s time: 119.21s eta: 16:27:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32080.0 tgs: 267 data_time: 0.83s time: 119.76s eta: 16:30:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 263 data_time: 0.83s time: 120.37s eta: 16:33:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31611.0 tgs: 265 data_time: 0.70s time: 118.98s eta: 16:19:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31344.0 tgs: 246 data_time: 0.68s time: 127.32s eta: 17:26:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31867.0 tgs: 267 data_time: 0.68s time: 119.22s eta: 16:17:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 268 data_time: 0.77s time: 119.14s eta: 16:14:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31208.0 tgs: 261 data_time: 0.61s time: 119.22s eta: 16:13:36 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32027.0 tgs: 268 data_time: 0.76s time: 119.18s eta: 16:11:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.230 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31555.0 tgs: 265 data_time: 0.88s time: 118.95s eta: 16:07:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.253 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30723.0 tgs: 258 data_time: 0.62s time: 118.86s eta: 16:04:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30348.0 tgs: 255 data_time: 0.70s time: 118.97s eta: 16:03:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 258 data_time: 0.67s time: 121.84s eta: 16:24:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31608.0 tgs: 265 data_time: 0.66s time: 118.99s eta: 15:59:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30383.0 tgs: 254 data_time: 1.12s time: 119.23s eta: 15:59:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.117 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32454.0 tgs: 271 data_time: 0.60s time: 119.61s eta: 16:00:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31170.0 tgs: 261 data_time: 0.89s time: 119.04s eta: 15:54:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31468.0 tgs: 264 data_time: 0.85s time: 118.98s eta: 15:51:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 258 data_time: 0.63s time: 119.17s eta: 15:51:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 269 data_time: 0.62s time: 119.64s eta: 15:53:09 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31922.0 tgs: 268 data_time: 0.80s time: 118.93s eta: 15:45:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 269 data_time: 0.78s time: 119.29s eta: 15:46:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32089.0 tgs: 269 data_time: 0.70s time: 119.10s eta: 15:42:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30705.0 tgs: 258 data_time: 0.97s time: 119.00s eta: 15:40:07 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31201.0 tgs: 260 data_time: 0.70s time: 119.65s eta: 15:43:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.123 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30621.0 tgs: 252 data_time: 0.64s time: 121.51s eta: 15:55:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.56s time: 119.41s eta: 15:37:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.278 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31983.0 tgs: 246 data_time: 0.69s time: 129.97s eta: 16:58:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32096.0 tgs: 268 data_time: 0.69s time: 119.46s eta: 15:33:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32218.0 tgs: 269 data_time: 0.83s time: 119.60s eta: 15:32:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31551.0 tgs: 263 data_time: 0.60s time: 119.53s eta: 15:30:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31482.0 tgs: 257 data_time: 0.69s time: 122.14s eta: 15:48:36 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30656.0 tgs: 253 data_time: 0.89s time: 120.76s eta: 15:35:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 268 data_time: 1.03s time: 119.41s eta: 15:23:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31441.0 tgs: 263 data_time: 0.92s time: 119.44s eta: 15:21:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31665.0 tgs: 240 data_time: 0.96s time: 131.63s eta: 16:53:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.107 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32148.0 tgs: 269 data_time: 0.76s time: 119.29s eta: 15:16:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 268 data_time: 0.59s time: 119.55s eta: 15:16:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.125 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31842.0 tgs: 260 data_time: 0.91s time: 122.25s eta: 15:35:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31745.0 tgs: 265 data_time: 1.02s time: 119.72s eta: 15:13:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30759.0 tgs: 257 data_time: 0.75s time: 119.34s eta: 15:09:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31030.0 tgs: 260 data_time: 1.00s time: 118.90s eta: 15:03:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31701.0 tgs: 266 data_time: 0.96s time: 119.00s eta: 15:02:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31700.0 tgs: 265 data_time: 0.87s time: 119.31s eta: 15:02:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31035.0 tgs: 260 data_time: 0.87s time: 118.92s eta: 14:57:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 258 data_time: 0.82s time: 120.52s eta: 15:07:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31653.0 tgs: 265 data_time: 0.74s time: 119.32s eta: 14:56:51 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.238 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32533.0 tgs: 265 data_time: 0.76s time: 122.66s eta: 15:19:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 266 data_time: 0.67s time: 119.13s eta: 14:51:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.211 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30713.0 tgs: 256 data_time: 1.12s time: 119.75s eta: 14:54:09 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.72s time: 119.47s eta: 14:50:01 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.245 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.80s time: 118.99s eta: 14:44:31 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32563.0 tgs: 272 data_time: 0.73s time: 119.30s eta: 14:44:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32619.0 tgs: 274 data_time: 0.89s time: 118.93s eta: 14:40:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31114.0 tgs: 261 data_time: 0.86s time: 119.05s eta: 14:39:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 27050.0 tgs: 227 data_time: 0.41s time: 119.06s eta: 14:37:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31203.0 tgs: 261 data_time: 0.64s time: 119.13s eta: 14:35:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.210 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 270 data_time: 0.98s time: 119.00s eta: 14:32:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.213 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31684.0 tgs: 266 data_time: 0.92s time: 119.07s eta: 14:31:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.117 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31253.0 tgs: 262 data_time: 0.64s time: 119.15s eta: 14:29:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.117 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31033.0 tgs: 259 data_time: 0.75s time: 119.65s eta: 14:31:27 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 268 data_time: 0.62s time: 119.18s eta: 14:26:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32325.0 tgs: 271 data_time: 0.90s time: 119.20s eta: 14:24:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31524.0 tgs: 264 data_time: 0.65s time: 119.39s eta: 14:23:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.196 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31265.0 tgs: 258 data_time: 0.99s time: 120.74s eta: 14:31:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31189.0 tgs: 262 data_time: 0.96s time: 118.86s eta: 14:15:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31533.0 tgs: 264 data_time: 0.93s time: 119.10s eta: 14:15:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.108 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.70s time: 119.13s eta: 14:13:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31560.0 tgs: 263 data_time: 0.83s time: 119.96s eta: 14:17:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31511.0 tgs: 262 data_time: 0.71s time: 119.87s eta: 14:15:03 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31916.0 tgs: 266 data_time: 0.63s time: 119.59s eta: 14:11:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31170.0 tgs: 261 data_time: 0.87s time: 119.22s eta: 14:06:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.115 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31944.0 tgs: 267 data_time: 0.73s time: 119.27s eta: 14:04:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.213 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32252.0 tgs: 270 data_time: 0.68s time: 119.27s eta: 14:02:51 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 269 data_time: 1.00s time: 119.14s eta: 13:59:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 270 data_time: 0.59s time: 119.13s eta: 13:57:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.87s time: 119.17s eta: 13:56:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.181 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31622.0 tgs: 264 data_time: 1.06s time: 119.35s eta: 13:55:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.196 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32179.0 tgs: 267 data_time: 1.39s time: 120.11s eta: 13:58:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32226.0 tgs: 262 data_time: 1.12s time: 122.56s eta: 14:13:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.125 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31889.0 tgs: 267 data_time: 0.95s time: 119.10s eta: 13:47:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31886.0 tgs: 267 data_time: 0.74s time: 119.37s eta: 13:47:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31640.0 tgs: 265 data_time: 0.75s time: 119.32s eta: 13:45:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 0.99s time: 119.33s eta: 13:43:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32138.0 tgs: 269 data_time: 0.90s time: 119.12s eta: 13:39:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31108.0 tgs: 260 data_time: 0.98s time: 119.25s eta: 13:38:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.211 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31461.0 tgs: 264 data_time: 0.68s time: 119.14s eta: 13:36:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.191 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 265 data_time: 0.93s time: 119.22s eta: 13:34:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32501.0 tgs: 272 data_time: 1.00s time: 119.46s eta: 13:34:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.213 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32122.0 tgs: 269 data_time: 0.80s time: 119.09s eta: 13:29:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.176 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31803.0 tgs: 267 data_time: 0.62s time: 118.92s eta: 13:26:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32254.0 tgs: 269 data_time: 0.81s time: 119.80s eta: 13:30:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 272 data_time: 1.02s time: 119.26s eta: 13:24:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31945.0 tgs: 268 data_time: 1.15s time: 119.17s eta: 13:22:24 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31334.0 tgs: 262 data_time: 1.02s time: 119.34s eta: 13:21:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.192 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 32048.0 tgs: 268 data_time: 0.94s time: 119.26s eta: 13:19:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32046.0 tgs: 269 data_time: 0.73s time: 119.04s eta: 13:15:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 264 data_time: 0.97s time: 119.35s eta: 13:15:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30547.0 tgs: 256 data_time: 0.65s time: 119.18s eta: 13:12:32 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 266 data_time: 0.98s time: 119.45s eta: 13:12:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 269 data_time: 1.12s time: 119.29s eta: 13:09:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31788.0 tgs: 266 data_time: 0.96s time: 119.15s eta: 13:06:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31874.0 tgs: 264 data_time: 0.98s time: 120.45s eta: 13:12:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32387.0 tgs: 269 data_time: 0.93s time: 120.16s eta: 13:09:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.224 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30335.0 tgs: 253 data_time: 0.53s time: 119.50s eta: 13:02:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30033.0 tgs: 251 data_time: 0.79s time: 119.51s eta: 13:00:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32025.0 tgs: 267 data_time: 0.82s time: 119.57s eta: 12:59:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.267 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31801.0 tgs: 265 data_time: 0.69s time: 119.80s eta: 12:58:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.223 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30774.0 tgs: 258 data_time: 0.88s time: 119.04s eta: 12:51:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.220 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31518.0 tgs: 264 data_time: 0.91s time: 119.05s eta: 12:49:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31386.0 tgs: 263 data_time: 0.65s time: 119.12s eta: 12:48:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.69s time: 119.40s eta: 12:48:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32166.0 tgs: 269 data_time: 0.85s time: 119.18s eta: 12:44:44 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32107.0 tgs: 268 data_time: 0.92s time: 119.44s eta: 12:44:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 1.05s time: 119.10s eta: 12:40:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.174 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31512.0 tgs: 264 data_time: 0.73s time: 119.07s eta: 12:38:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.184 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31533.0 tgs: 263 data_time: 0.86s time: 119.59s eta: 12:39:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31439.0 tgs: 264 data_time: 0.69s time: 118.78s eta: 12:32:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31130.0 tgs: 261 data_time: 0.67s time: 118.89s eta: 12:30:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32011.0 tgs: 268 data_time: 0.86s time: 119.17s eta: 12:30:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 25762.0 tgs: 216 data_time: 0.51s time: 119.26s eta: 12:29:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 270 data_time: 0.88s time: 119.35s eta: 12:27:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31986.0 tgs: 268 data_time: 0.90s time: 119.12s eta: 12:24:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.198 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 263 data_time: 0.99s time: 119.35s eta: 12:23:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.116 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32126.0 tgs: 269 data_time: 0.86s time: 119.26s eta: 12:21:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31207.0 tgs: 262 data_time: 0.98s time: 119.03s eta: 12:18:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.115 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 262 data_time: 0.60s time: 121.13s eta: 12:28:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31262.0 tgs: 260 data_time: 0.84s time: 119.99s eta: 12:19:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31381.0 tgs: 264 data_time: 0.87s time: 118.70s eta: 12:10:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.162 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31519.0 tgs: 263 data_time: 0.89s time: 119.63s eta: 12:13:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 268 data_time: 0.88s time: 119.84s eta: 12:12:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 264 data_time: 0.86s time: 119.15s eta: 12:06:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31152.0 tgs: 261 data_time: 0.79s time: 119.17s eta: 12:04:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.128 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.98s time: 119.12s eta: 12:02:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31252.0 tgs: 262 data_time: 0.98s time: 119.19s eta: 12:01:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.114 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30537.0 tgs: 256 data_time: 1.06s time: 118.96s eta: 11:57:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30712.0 tgs: 256 data_time: 0.55s time: 119.71s eta: 12:00:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31275.0 tgs: 262 data_time: 0.46s time: 119.28s eta: 11:55:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32041.0 tgs: 268 data_time: 0.71s time: 119.28s eta: 11:53:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.109 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31777.0 tgs: 265 data_time: 0.84s time: 119.71s eta: 11:54:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32515.0 tgs: 273 data_time: 0.71s time: 118.80s eta: 11:46:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.133 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.85s time: 119.14s eta: 11:46:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31652.0 tgs: 264 data_time: 0.92s time: 119.61s eta: 11:47:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31256.0 tgs: 263 data_time: 0.92s time: 118.83s eta: 11:41:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32109.0 tgs: 270 data_time: 0.98s time: 118.68s eta: 11:38:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.56s time: 119.22s eta: 11:39:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31108.0 tgs: 261 data_time: 0.92s time: 119.14s eta: 11:36:57 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 268 data_time: 0.88s time: 119.08s eta: 11:34:36 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 263 data_time: 0.68s time: 118.95s eta: 11:31:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31063.0 tgs: 261 data_time: 0.87s time: 118.91s eta: 11:29:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31995.0 tgs: 268 data_time: 0.74s time: 118.97s eta: 11:28:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32662.0 tgs: 273 data_time: 0.84s time: 119.29s eta: 11:27:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.118 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31996.0 tgs: 268 data_time: 0.66s time: 119.36s eta: 11:26:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31004.0 tgs: 260 data_time: 0.75s time: 118.88s eta: 11:21:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32430.0 tgs: 272 data_time: 0.79s time: 119.21s eta: 11:21:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30334.0 tgs: 253 data_time: 0.70s time: 119.72s eta: 11:22:24 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.114 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31948.0 tgs: 268 data_time: 0.96s time: 119.11s eta: 11:16:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31792.0 tgs: 266 data_time: 0.57s time: 119.12s eta: 11:15:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.189 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31595.0 tgs: 265 data_time: 0.92s time: 118.95s eta: 11:12:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30196.0 tgs: 253 data_time: 0.80s time: 119.10s eta: 11:10:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31338.0 tgs: 262 data_time: 0.63s time: 119.16s eta: 11:09:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31464.0 tgs: 264 data_time: 0.94s time: 118.97s eta: 11:06:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30987.0 tgs: 259 data_time: 0.80s time: 119.47s eta: 11:07:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31777.0 tgs: 265 data_time: 0.87s time: 119.64s eta: 11:06:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31707.0 tgs: 265 data_time: 0.81s time: 119.30s eta: 11:02:07 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.252 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30829.0 tgs: 258 data_time: 0.90s time: 119.31s eta: 11:00:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31273.0 tgs: 263 data_time: 0.86s time: 118.82s eta: 10:55:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32308.0 tgs: 270 data_time: 0.67s time: 119.47s eta: 10:57:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31749.0 tgs: 265 data_time: 0.57s time: 119.37s eta: 10:54:32 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31767.0 tgs: 266 data_time: 0.70s time: 119.28s eta: 10:52:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 267 data_time: 0.74s time: 119.26s eta: 10:49:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 267 data_time: 0.82s time: 119.10s eta: 10:47:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31298.0 tgs: 261 data_time: 0.76s time: 119.53s eta: 10:47:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31337.0 tgs: 262 data_time: 0.69s time: 119.31s eta: 10:44:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31711.0 tgs: 265 data_time: 0.82s time: 119.49s eta: 10:43:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31513.0 tgs: 263 data_time: 0.83s time: 119.51s eta: 10:41:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.123 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31466.0 tgs: 264 data_time: 0.93s time: 119.00s eta: 10:36:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.259 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31991.0 tgs: 268 data_time: 1.16s time: 119.22s eta: 10:35:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.191 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31635.0 tgs: 265 data_time: 0.74s time: 119.06s eta: 10:33:01 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31474.0 tgs: 264 data_time: 0.63s time: 118.78s eta: 10:29:31 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31774.0 tgs: 265 data_time: 0.78s time: 119.86s eta: 10:33:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31488.0 tgs: 263 data_time: 0.93s time: 119.38s eta: 10:28:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32535.0 tgs: 273 data_time: 0.82s time: 118.83s eta: 10:23:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 267 data_time: 0.88s time: 118.99s eta: 10:22:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.117 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31715.0 tgs: 265 data_time: 0.76s time: 119.28s eta: 10:22:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.114 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31553.0 tgs: 264 data_time: 0.63s time: 119.21s eta: 10:19:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 270 data_time: 0.76s time: 119.06s eta: 10:17:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.76s time: 119.48s eta: 10:17:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31907.0 tgs: 267 data_time: 0.60s time: 119.37s eta: 10:14:44 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32006.0 tgs: 267 data_time: 0.41s time: 119.45s eta: 10:13:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29892.0 tgs: 250 data_time: 0.70s time: 119.29s eta: 10:10:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30783.0 tgs: 258 data_time: 0.57s time: 119.11s eta: 10:07:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32288.0 tgs: 269 data_time: 0.83s time: 119.70s eta: 10:08:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 265 data_time: 0.87s time: 119.83s eta: 10:07:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32161.0 tgs: 264 data_time: 1.03s time: 121.55s eta: 10:13:51 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.85s time: 119.26s eta: 10:00:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31992.0 tgs: 269 data_time: 0.98s time: 118.81s eta: 9:56:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32072.0 tgs: 268 data_time: 0.68s time: 119.56s eta: 9:57:48 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.159 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31033.0 tgs: 259 data_time: 0.94s time: 119.71s eta: 9:56:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31291.0 tgs: 262 data_time: 0.85s time: 119.25s eta: 9:52:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.179 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 268 data_time: 1.04s time: 119.33s eta: 9:50:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30506.0 tgs: 255 data_time: 0.75s time: 119.22s eta: 9:48:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31095.0 tgs: 258 data_time: 0.88s time: 120.43s eta: 9:52:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32378.0 tgs: 271 data_time: 0.65s time: 119.26s eta: 9:44:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31737.0 tgs: 265 data_time: 0.84s time: 119.46s eta: 9:43:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31010.0 tgs: 259 data_time: 0.83s time: 119.55s eta: 9:41:48 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31878.0 tgs: 267 data_time: 0.49s time: 119.20s eta: 9:38:07 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31652.0 tgs: 264 data_time: 0.71s time: 119.48s eta: 9:37:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 265 data_time: 0.93s time: 119.05s eta: 9:33:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 262 data_time: 0.66s time: 119.24s eta: 9:32:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31283.0 tgs: 262 data_time: 0.79s time: 119.16s eta: 9:29:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.189 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32064.0 tgs: 268 data_time: 1.01s time: 119.22s eta: 9:28:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31438.0 tgs: 263 data_time: 0.81s time: 119.46s eta: 9:27:27 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.196 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31072.0 tgs: 261 data_time: 0.73s time: 118.98s eta: 9:23:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31779.0 tgs: 266 data_time: 0.71s time: 119.03s eta: 9:21:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.196 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31632.0 tgs: 264 data_time: 1.00s time: 119.53s eta: 9:21:48 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32297.0 tgs: 270 data_time: 0.67s time: 119.60s eta: 9:20:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.125 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31681.0 tgs: 265 data_time: 0.67s time: 119.12s eta: 9:15:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32415.0 tgs: 272 data_time: 0.69s time: 119.01s eta: 9:13:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31327.0 tgs: 263 data_time: 0.77s time: 118.88s eta: 9:10:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.96s time: 119.32s eta: 9:10:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.212 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 1.02s time: 119.09s eta: 9:07:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31427.0 tgs: 264 data_time: 0.77s time: 118.86s eta: 9:04:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 264 data_time: 0.68s time: 119.67s eta: 9:06:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31164.0 tgs: 260 data_time: 0.95s time: 119.44s eta: 9:03:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.187 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31258.0 tgs: 261 data_time: 0.73s time: 119.43s eta: 9:01:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.203 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 265 data_time: 0.81s time: 118.74s eta: 8:56:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31022.0 tgs: 260 data_time: 0.87s time: 118.87s eta: 8:54:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31953.0 tgs: 267 data_time: 0.91s time: 119.33s eta: 8:55:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30599.0 tgs: 256 data_time: 0.80s time: 119.15s eta: 8:52:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31640.0 tgs: 265 data_time: 0.52s time: 119.27s eta: 8:50:44 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31856.0 tgs: 264 data_time: 0.84s time: 120.48s eta: 8:54:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.108 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 267 data_time: 0.78s time: 119.23s eta: 8:46:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 266 data_time: 0.93s time: 119.07s eta: 8:43:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.197 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 263 data_time: 1.01s time: 120.84s eta: 8:49:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31839.0 tgs: 264 data_time: 0.81s time: 120.19s eta: 8:44:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31794.0 tgs: 267 data_time: 0.53s time: 118.82s eta: 8:36:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.207 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32118.0 tgs: 269 data_time: 0.81s time: 119.36s eta: 8:37:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.73s time: 119.33s eta: 8:35:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31845.0 tgs: 267 data_time: 0.74s time: 119.05s eta: 8:31:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 269 data_time: 0.79s time: 119.16s eta: 8:30:24 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.209 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.71s time: 119.04s eta: 8:27:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32051.0 tgs: 269 data_time: 0.71s time: 119.12s eta: 8:26:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.092 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.74s time: 119.25s eta: 8:24:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.78s time: 119.22s eta: 8:22:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31756.0 tgs: 266 data_time: 0.79s time: 119.38s eta: 8:21:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31863.0 tgs: 266 data_time: 0.88s time: 119.78s eta: 8:21:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31688.0 tgs: 265 data_time: 0.87s time: 119.20s eta: 8:16:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31434.0 tgs: 264 data_time: 0.63s time: 118.98s eta: 8:13:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.315 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31934.0 tgs: 266 data_time: 0.67s time: 119.90s eta: 8:15:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.118 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 268 data_time: 0.65s time: 119.55s eta: 8:12:09 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.119 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32172.0 tgs: 269 data_time: 0.86s time: 119.59s eta: 8:10:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31739.0 tgs: 266 data_time: 0.89s time: 119.00s eta: 8:05:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.120 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31506.0 tgs: 262 data_time: 0.81s time: 119.84s eta: 8:07:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 266 data_time: 0.74s time: 119.71s eta: 8:04:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31864.0 tgs: 266 data_time: 0.93s time: 119.40s eta: 8:01:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.165 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32174.0 tgs: 269 data_time: 0.79s time: 119.29s eta: 7:59:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31700.0 tgs: 265 data_time: 0.63s time: 119.19s eta: 7:56:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31403.0 tgs: 263 data_time: 0.66s time: 119.20s eta: 7:54:48 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31207.0 tgs: 262 data_time: 0.61s time: 118.94s eta: 7:51:48 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31558.0 tgs: 265 data_time: 0.99s time: 118.97s eta: 7:49:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 270 data_time: 0.69s time: 119.42s eta: 7:49:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 270 data_time: 0.60s time: 119.04s eta: 7:46:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32332.0 tgs: 271 data_time: 0.91s time: 119.07s eta: 7:44:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 268 data_time: 0.81s time: 120.01s eta: 7:46:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31324.0 tgs: 261 data_time: 0.70s time: 119.82s eta: 7:43:19 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32501.0 tgs: 272 data_time: 0.63s time: 119.17s eta: 7:38:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32108.0 tgs: 269 data_time: 0.86s time: 118.97s eta: 7:36:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31892.0 tgs: 267 data_time: 0.81s time: 119.33s eta: 7:35:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31726.0 tgs: 265 data_time: 0.83s time: 119.38s eta: 7:33:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31653.0 tgs: 265 data_time: 0.60s time: 119.06s eta: 7:30:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31597.0 tgs: 265 data_time: 0.81s time: 118.99s eta: 7:28:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30377.0 tgs: 254 data_time: 0.58s time: 119.24s eta: 7:27:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31018.0 tgs: 260 data_time: 0.46s time: 118.99s eta: 7:24:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30388.0 tgs: 255 data_time: 0.68s time: 118.93s eta: 7:22:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 262 data_time: 0.70s time: 119.09s eta: 7:20:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31932.0 tgs: 268 data_time: 0.88s time: 118.99s eta: 7:18:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31833.0 tgs: 267 data_time: 0.82s time: 119.03s eta: 7:16:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31649.0 tgs: 265 data_time: 0.70s time: 119.24s eta: 7:15:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31558.0 tgs: 264 data_time: 0.74s time: 119.36s eta: 7:13:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.184 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31390.0 tgs: 263 data_time: 0.65s time: 118.94s eta: 7:10:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.56s time: 119.21s eta: 7:09:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31845.0 tgs: 266 data_time: 0.55s time: 119.58s eta: 7:08:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.187 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29935.0 tgs: 250 data_time: 0.76s time: 119.66s eta: 7:06:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.109 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31746.0 tgs: 266 data_time: 0.71s time: 118.95s eta: 7:02:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31721.0 tgs: 265 data_time: 0.67s time: 119.31s eta: 7:01:32 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31320.0 tgs: 262 data_time: 0.98s time: 119.51s eta: 7:00:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32203.0 tgs: 269 data_time: 0.66s time: 119.43s eta: 6:57:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 272 data_time: 0.70s time: 118.91s eta: 6:54:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 270 data_time: 0.95s time: 119.34s eta: 6:53:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 263 data_time: 0.76s time: 118.93s eta: 6:50:19 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32205.0 tgs: 270 data_time: 0.73s time: 119.27s eta: 6:49:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.208 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30968.0 tgs: 260 data_time: 0.81s time: 119.02s eta: 6:46:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.114 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31854.0 tgs: 267 data_time: 0.59s time: 119.22s eta: 6:45:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.110 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 269 data_time: 0.83s time: 119.38s eta: 6:43:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32035.0 tgs: 269 data_time: 0.59s time: 119.01s eta: 6:40:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32042.0 tgs: 269 data_time: 0.95s time: 119.01s eta: 6:38:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 264 data_time: 0.71s time: 119.46s eta: 6:38:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.097 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31318.0 tgs: 263 data_time: 0.85s time: 118.80s eta: 6:34:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.70s time: 119.44s eta: 6:34:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.215 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31646.0 tgs: 264 data_time: 0.74s time: 119.57s eta: 6:32:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31349.0 tgs: 262 data_time: 0.70s time: 119.38s eta: 6:29:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.098 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30824.0 tgs: 257 data_time: 0.53s time: 119.60s eta: 6:28:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32354.0 tgs: 271 data_time: 0.64s time: 119.36s eta: 6:25:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31166.0 tgs: 262 data_time: 0.71s time: 118.82s eta: 6:22:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31988.0 tgs: 268 data_time: 0.96s time: 119.02s eta: 6:20:51 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31964.0 tgs: 267 data_time: 0.98s time: 119.40s eta: 6:20:05 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32179.0 tgs: 269 data_time: 0.99s time: 119.34s eta: 6:17:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31511.0 tgs: 264 data_time: 0.93s time: 119.05s eta: 6:14:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31463.0 tgs: 263 data_time: 0.72s time: 119.40s eta: 6:14:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31773.0 tgs: 266 data_time: 0.86s time: 119.38s eta: 6:12:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31893.0 tgs: 267 data_time: 0.71s time: 119.29s eta: 6:09:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.209 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.83s time: 119.10s eta: 6:07:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31700.0 tgs: 259 data_time: 0.80s time: 121.96s eta: 6:14:01 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31492.0 tgs: 263 data_time: 0.75s time: 119.39s eta: 6:04:07 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31332.0 tgs: 262 data_time: 0.71s time: 119.57s eta: 6:02:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31086.0 tgs: 260 data_time: 0.84s time: 119.26s eta: 5:59:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.108 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 27321.0 tgs: 229 data_time: 0.67s time: 119.26s eta: 5:57:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31862.0 tgs: 267 data_time: 0.97s time: 118.92s eta: 5:54:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 259 data_time: 0.94s time: 119.02s eta: 5:53:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31457.0 tgs: 264 data_time: 0.85s time: 119.07s eta: 5:51:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30236.0 tgs: 253 data_time: 0.90s time: 119.29s eta: 5:49:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31960.0 tgs: 268 data_time: 0.75s time: 118.97s eta: 5:46:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.205 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32273.0 tgs: 270 data_time: 0.72s time: 119.27s eta: 5:45:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32334.0 tgs: 270 data_time: 0.84s time: 119.33s eta: 5:44:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32077.0 tgs: 269 data_time: 0.71s time: 119.05s eta: 5:41:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.112 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31705.0 tgs: 265 data_time: 0.67s time: 119.50s eta: 5:40:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31560.0 tgs: 264 data_time: 0.81s time: 119.38s eta: 5:38:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30543.0 tgs: 255 data_time: 0.89s time: 119.57s eta: 5:36:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.113 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31756.0 tgs: 266 data_time: 0.77s time: 118.98s eta: 5:33:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.115 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31548.0 tgs: 265 data_time: 0.87s time: 118.95s eta: 5:31:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.60s time: 119.24s eta: 5:29:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30745.0 tgs: 257 data_time: 0.95s time: 119.40s eta: 5:28:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.216 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31753.0 tgs: 265 data_time: 0.93s time: 119.46s eta: 5:26:31 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31738.0 tgs: 266 data_time: 0.70s time: 118.95s eta: 5:23:09 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31359.0 tgs: 262 data_time: 0.59s time: 119.63s eta: 5:22:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31478.0 tgs: 263 data_time: 0.74s time: 119.35s eta: 5:20:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31081.0 tgs: 260 data_time: 0.70s time: 119.12s eta: 5:17:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31347.0 tgs: 262 data_time: 0.53s time: 119.42s eta: 5:16:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.209 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31565.0 tgs: 263 data_time: 0.93s time: 119.58s eta: 5:14:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.177 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 267 data_time: 0.83s time: 119.00s eta: 5:11:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 265 data_time: 0.71s time: 119.00s eta: 5:09:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31608.0 tgs: 265 data_time: 0.92s time: 119.23s eta: 5:08:00 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 268 data_time: 0.76s time: 119.69s eta: 5:07:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31961.0 tgs: 268 data_time: 0.72s time: 119.11s eta: 5:03:44 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32214.0 tgs: 270 data_time: 0.89s time: 119.07s eta: 5:01:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.210 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32104.0 tgs: 264 data_time: 0.91s time: 121.53s eta: 5:05:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.177 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 266 data_time: 0.81s time: 119.57s eta: 4:58:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30327.0 tgs: 253 data_time: 0.53s time: 119.50s eta: 4:56:44 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.100 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32344.0 tgs: 270 data_time: 1.09s time: 119.73s eta: 4:55:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.104 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31880.0 tgs: 266 data_time: 0.51s time: 119.52s eta: 4:52:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30378.0 tgs: 254 data_time: 0.87s time: 119.42s eta: 4:50:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31216.0 tgs: 261 data_time: 0.76s time: 119.20s eta: 4:48:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 265 data_time: 0.65s time: 118.89s eta: 4:45:19 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.186 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 265 data_time: 0.87s time: 119.05s eta: 4:43:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 28332.0 tgs: 236 data_time: 0.65s time: 119.98s eta: 4:43:57 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 270 data_time: 0.68s time: 119.15s eta: 4:39:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31138.0 tgs: 261 data_time: 0.72s time: 119.01s eta: 4:37:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31315.0 tgs: 263 data_time: 0.86s time: 119.07s eta: 4:35:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30717.0 tgs: 257 data_time: 0.84s time: 119.36s eta: 4:34:32 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31628.0 tgs: 264 data_time: 0.59s time: 119.41s eta: 4:32:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30401.0 tgs: 255 data_time: 0.58s time: 118.94s eta: 4:29:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32144.0 tgs: 268 data_time: 0.74s time: 119.60s eta: 4:29:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31502.0 tgs: 264 data_time: 0.94s time: 119.24s eta: 4:26:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.182 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.71s time: 119.07s eta: 4:23:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 269 data_time: 0.62s time: 118.85s eta: 4:21:28 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30956.0 tgs: 260 data_time: 0.70s time: 118.76s eta: 4:19:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29499.0 tgs: 246 data_time: 0.68s time: 119.43s eta: 4:18:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 266 data_time: 0.64s time: 119.40s eta: 4:16:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.101 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 28527.0 tgs: 239 data_time: 0.59s time: 119.34s eta: 4:14:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 261 data_time: 0.80s time: 119.05s eta: 4:11:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31333.0 tgs: 263 data_time: 0.81s time: 119.12s eta: 4:10:09 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31398.0 tgs: 263 data_time: 0.67s time: 119.13s eta: 4:08:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30799.0 tgs: 258 data_time: 0.65s time: 118.96s eta: 4:05:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31257.0 tgs: 262 data_time: 0.69s time: 119.15s eta: 4:04:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.203 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32315.0 tgs: 270 data_time: 0.75s time: 119.50s eta: 4:02:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.259 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32550.0 tgs: 272 data_time: 0.60s time: 119.25s eta: 4:00:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.103 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31342.0 tgs: 262 data_time: 0.72s time: 119.19s eta: 3:58:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31153.0 tgs: 261 data_time: 0.64s time: 119.36s eta: 3:56:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.110 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31063.0 tgs: 260 data_time: 0.73s time: 119.37s eta: 3:54:45 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 265 data_time: 0.85s time: 119.96s eta: 3:53:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 271 data_time: 0.69s time: 119.32s eta: 3:50:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30971.0 tgs: 258 data_time: 0.93s time: 119.63s eta: 3:49:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31112.0 tgs: 261 data_time: 0.68s time: 118.86s eta: 3:45:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 267 data_time: 0.98s time: 119.71s eta: 3:45:27 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 269 data_time: 0.84s time: 119.32s eta: 3:42:44 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31498.0 tgs: 264 data_time: 0.79s time: 119.26s eta: 3:40:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30580.0 tgs: 256 data_time: 0.92s time: 119.00s eta: 3:38:10 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 1.22s time: 119.43s eta: 3:36:57 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31384.0 tgs: 263 data_time: 0.88s time: 119.03s eta: 3:34:15 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31768.0 tgs: 266 data_time: 0.70s time: 118.99s eta: 3:32:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.72s time: 119.22s eta: 3:30:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30898.0 tgs: 257 data_time: 0.78s time: 120.11s eta: 3:30:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 268 data_time: 0.86s time: 119.13s eta: 3:26:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31751.0 tgs: 266 data_time: 0.82s time: 119.31s eta: 3:24:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 270 data_time: 0.49s time: 119.08s eta: 3:22:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32224.0 tgs: 269 data_time: 0.91s time: 119.43s eta: 3:21:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30820.0 tgs: 258 data_time: 0.78s time: 119.29s eta: 3:18:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31507.0 tgs: 264 data_time: 0.80s time: 119.22s eta: 3:16:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.95s time: 119.14s eta: 3:14:35 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30908.0 tgs: 259 data_time: 0.80s time: 119.02s eta: 3:12:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32441.0 tgs: 272 data_time: 0.64s time: 119.18s eta: 3:10:41 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30654.0 tgs: 253 data_time: 0.91s time: 121.05s eta: 3:11:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32062.0 tgs: 268 data_time: 0.70s time: 119.50s eta: 3:07:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32494.0 tgs: 272 data_time: 0.96s time: 119.26s eta: 3:04:51 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30956.0 tgs: 260 data_time: 0.73s time: 118.99s eta: 3:02:27 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31562.0 tgs: 264 data_time: 0.77s time: 119.22s eta: 3:00:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31709.0 tgs: 265 data_time: 0.79s time: 119.46s eta: 2:59:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 262 data_time: 0.73s time: 118.98s eta: 2:56:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32415.0 tgs: 271 data_time: 1.14s time: 119.35s eta: 2:55:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.109 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29085.0 tgs: 243 data_time: 0.81s time: 119.41s eta: 2:53:08 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31686.0 tgs: 265 data_time: 0.82s time: 119.35s eta: 2:51:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31625.0 tgs: 260 data_time: 0.70s time: 121.46s eta: 2:52:04 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.087 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31391.0 tgs: 263 data_time: 0.55s time: 119.24s eta: 2:46:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31410.0 tgs: 262 data_time: 0.98s time: 119.56s eta: 2:45:23 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 270 data_time: 0.69s time: 118.91s eta: 2:42:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 0.76s time: 119.19s eta: 2:40:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 267 data_time: 0.53s time: 118.97s eta: 2:38:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32061.0 tgs: 268 data_time: 0.89s time: 119.23s eta: 2:36:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31916.0 tgs: 267 data_time: 0.74s time: 119.11s eta: 2:34:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 266 data_time: 0.84s time: 119.43s eta: 2:33:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31433.0 tgs: 261 data_time: 1.14s time: 120.23s eta: 2:32:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 264 data_time: 0.92s time: 119.28s eta: 2:29:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32083.0 tgs: 269 data_time: 0.84s time: 118.88s eta: 2:26:37 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.111 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31396.0 tgs: 262 data_time: 0.69s time: 119.52s eta: 2:25:24 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31508.0 tgs: 264 data_time: 0.54s time: 119.07s eta: 2:22:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31513.0 tgs: 264 data_time: 0.60s time: 119.14s eta: 2:20:58 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31506.0 tgs: 264 data_time: 0.78s time: 119.05s eta: 2:18:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32410.0 tgs: 270 data_time: 0.71s time: 119.85s eta: 2:17:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 268 data_time: 1.02s time: 119.31s eta: 2:15:13 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31906.0 tgs: 267 data_time: 0.64s time: 119.38s eta: 2:13:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 267 data_time: 0.86s time: 119.68s eta: 2:11:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31768.0 tgs: 265 data_time: 0.88s time: 119.49s eta: 2:09:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 270 data_time: 0.67s time: 118.96s eta: 2:06:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31616.0 tgs: 264 data_time: 0.61s time: 119.38s eta: 2:05:21 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30679.0 tgs: 257 data_time: 0.75s time: 119.27s eta: 2:03:14 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30337.0 tgs: 254 data_time: 0.60s time: 119.22s eta: 2:01:12 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32212.0 tgs: 269 data_time: 0.79s time: 119.51s eta: 1:59:30 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.215 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.66s time: 119.43s eta: 1:57:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32505.0 tgs: 271 data_time: 0.85s time: 119.54s eta: 1:55:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31980.0 tgs: 268 data_time: 0.61s time: 119.30s eta: 1:53:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31335.0 tgs: 262 data_time: 0.88s time: 119.25s eta: 1:51:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32336.0 tgs: 271 data_time: 0.81s time: 118.93s eta: 1:49:01 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31302.0 tgs: 261 data_time: 0.57s time: 119.64s eta: 1:47:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.199 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31557.0 tgs: 263 data_time: 0.81s time: 119.69s eta: 1:45:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31294.0 tgs: 262 data_time: 0.75s time: 119.35s eta: 1:43:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31227.0 tgs: 261 data_time: 0.70s time: 119.30s eta: 1:41:24 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31983.0 tgs: 268 data_time: 1.00s time: 119.28s eta: 1:39:24 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32110.0 tgs: 270 data_time: 0.78s time: 118.91s eta: 1:37:06 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 263 data_time: 0.84s time: 119.16s eta: 1:35:19 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32043.0 tgs: 269 data_time: 1.02s time: 119.07s eta: 1:33:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32253.0 tgs: 270 data_time: 0.70s time: 119.25s eta: 1:31:25 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30673.0 tgs: 256 data_time: 0.84s time: 119.73s eta: 1:29:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31562.0 tgs: 265 data_time: 0.75s time: 119.02s eta: 1:27:16 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 263 data_time: 0.76s time: 118.87s eta: 1:25:11 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32139.0 tgs: 268 data_time: 0.71s time: 119.48s eta: 1:23:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30337.0 tgs: 253 data_time: 0.76s time: 119.48s eta: 1:21:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32527.0 tgs: 273 data_time: 0.83s time: 118.94s eta: 1:19:17 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31543.0 tgs: 265 data_time: 0.66s time: 118.93s eta: 1:17:18 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 263 data_time: 0.60s time: 122.13s eta: 1:17:20 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32360.0 tgs: 271 data_time: 0.55s time: 119.36s eta: 1:13:36 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31728.0 tgs: 266 data_time: 0.73s time: 119.06s eta: 1:11:26 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31053.0 tgs: 261 data_time: 0.70s time: 118.93s eta: 1:09:22 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31778.0 tgs: 266 data_time: 0.85s time: 119.10s eta: 1:07:29 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31475.0 tgs: 263 data_time: 0.70s time: 119.41s eta: 1:05:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29530.0 tgs: 247 data_time: 0.66s time: 119.35s eta: 1:03:39 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30940.0 tgs: 259 data_time: 0.62s time: 119.18s eta: 1:01:34 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31634.0 tgs: 265 data_time: 0.65s time: 119.27s eta: 0:59:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.106 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 269 data_time: 0.67s time: 119.44s eta: 0:57:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30613.0 tgs: 256 data_time: 0.84s time: 119.39s eta: 0:55:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31940.0 tgs: 267 data_time: 0.82s time: 119.49s eta: 0:53:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.63s time: 119.37s eta: 0:51:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31155.0 tgs: 260 data_time: 0.61s time: 119.47s eta: 0:49:46 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31618.0 tgs: 265 data_time: 0.86s time: 118.90s eta: 0:47:33 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32178.0 tgs: 269 data_time: 0.69s time: 119.27s eta: 0:45:43 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.66s time: 119.20s eta: 0:43:42 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32354.0 tgs: 271 data_time: 0.86s time: 118.99s eta: 0:41:38 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32505.0 tgs: 273 data_time: 0.79s time: 119.04s eta: 0:39:40 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32119.0 tgs: 268 data_time: 0.64s time: 119.66s eta: 0:37:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.106 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32210.0 tgs: 270 data_time: 0.59s time: 119.29s eta: 0:35:47 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32281.0 tgs: 269 data_time: 0.75s time: 119.75s eta: 0:33:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30657.0 tgs: 252 data_time: 0.65s time: 121.23s eta: 0:32:19 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31457.0 tgs: 263 data_time: 0.83s time: 119.38s eta: 0:29:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.201 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.78s time: 119.32s eta: 0:27:50 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31402.0 tgs: 262 data_time: 0.60s time: 119.64s eta: 0:25:55 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30715.0 tgs: 257 data_time: 0.60s time: 119.09s eta: 0:23:49 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31496.0 tgs: 263 data_time: 1.01s time: 119.39s eta: 0:21:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 264 data_time: 0.70s time: 119.26s eta: 0:19:52 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31642.0 tgs: 265 data_time: 0.73s time: 119.24s eta: 0:17:53 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 270 data_time: 0.65s time: 119.90s eta: 0:15:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31768.0 tgs: 266 data_time: 0.82s time: 119.19s eta: 0:13:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32268.0 tgs: 270 data_time: 0.89s time: 119.09s eta: 0:11:54 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32354.0 tgs: 271 data_time: 0.86s time: 119.32s eta: 0:09:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32019.0 tgs: 268 data_time: 0.71s time: 119.17s eta: 0:07:56 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.095 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31728.0 tgs: 265 data_time: 0.72s time: 119.68s eta: 0:05:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30597.0 tgs: 252 data_time: 1.08s time: 121.39s eta: 0:04:02 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32060.0 tgs: 268 data_time: 0.66s time: 119.21s eta: 0:01:59 +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.28 seconds, peak gpu memory 17.4G +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 44][DP 11][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank46.log b/20250121165312/rank46.log new file mode 100644 index 0000000000000000000000000000000000000000..c77e6b2ef27fd15a58e96d7e6d630312ba6a07b0 --- /dev/null +++ b/20250121165312/rank46.log @@ -0,0 +1,851 @@ +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 71.37 seconds, peak gpu memory 13.4G +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.315 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 31361.0 tgs: 205 data_time: 1.66s time: 152.61s eta: 1 day, 1:08:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.233 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32373.0 tgs: 269 data_time: 1.25s time: 120.24s eta: 19:46:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.225 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31271.0 tgs: 261 data_time: 1.12s time: 119.69s eta: 19:38:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31959.0 tgs: 267 data_time: 0.87s time: 119.66s eta: 19:36:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.246 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32119.0 tgs: 269 data_time: 0.68s time: 119.31s eta: 19:31:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31094.0 tgs: 260 data_time: 0.88s time: 119.53s eta: 19:31:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.199 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 259 data_time: 1.03s time: 122.96s eta: 20:02:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.286 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31574.0 tgs: 263 data_time: 0.78s time: 119.89s eta: 19:30:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.181 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31714.0 tgs: 262 data_time: 0.86s time: 120.66s eta: 19:36:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31891.0 tgs: 265 data_time: 0.85s time: 119.92s eta: 19:27:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.206 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31327.0 tgs: 260 data_time: 0.88s time: 120.46s eta: 19:30:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 0.88s time: 119.34s eta: 19:17:37 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.216 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31061.0 tgs: 259 data_time: 0.79s time: 119.68s eta: 19:18:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30891.0 tgs: 259 data_time: 0.79s time: 119.17s eta: 19:12:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31341.0 tgs: 262 data_time: 0.90s time: 119.20s eta: 19:10:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.232 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31438.0 tgs: 264 data_time: 0.93s time: 118.94s eta: 19:05:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 269 data_time: 0.66s time: 119.14s eta: 19:05:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 1.05s time: 119.18s eta: 19:04:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 262 data_time: 0.78s time: 119.40s eta: 19:04:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.81s time: 119.33s eta: 19:01:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 264 data_time: 0.69s time: 120.31s eta: 19:08:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 262 data_time: 0.75s time: 122.91s eta: 19:31:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31172.0 tgs: 260 data_time: 0.99s time: 119.48s eta: 18:57:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32324.0 tgs: 271 data_time: 0.98s time: 119.16s eta: 18:51:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 271 data_time: 0.91s time: 119.25s eta: 18:50:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.280 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31182.0 tgs: 261 data_time: 0.88s time: 119.39s eta: 18:50:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31901.0 tgs: 266 data_time: 0.79s time: 119.66s eta: 18:50:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30544.0 tgs: 255 data_time: 1.01s time: 119.39s eta: 18:46:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31830.0 tgs: 265 data_time: 0.77s time: 119.77s eta: 18:47:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31792.0 tgs: 245 data_time: 1.20s time: 129.31s eta: 20:15:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31968.0 tgs: 267 data_time: 0.87s time: 119.57s eta: 18:41:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31133.0 tgs: 260 data_time: 0.85s time: 119.39s eta: 18:38:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 267 data_time: 0.86s time: 120.65s eta: 18:48:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31570.0 tgs: 264 data_time: 0.98s time: 119.40s eta: 18:34:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32328.0 tgs: 271 data_time: 1.00s time: 119.16s eta: 18:30:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 265 data_time: 0.95s time: 119.07s eta: 18:27:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31631.0 tgs: 265 data_time: 1.03s time: 119.14s eta: 18:26:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31761.0 tgs: 263 data_time: 0.76s time: 120.39s eta: 18:35:34 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31545.0 tgs: 260 data_time: 0.86s time: 121.18s eta: 18:40:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30209.0 tgs: 250 data_time: 0.75s time: 120.67s eta: 18:34:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.241 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32351.0 tgs: 270 data_time: 0.65s time: 119.73s eta: 18:23:30 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.240 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31789.0 tgs: 256 data_time: 0.76s time: 123.88s eta: 18:59:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32049.0 tgs: 268 data_time: 0.76s time: 119.56s eta: 18:17:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32186.0 tgs: 269 data_time: 0.80s time: 119.33s eta: 18:13:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31935.0 tgs: 267 data_time: 0.86s time: 119.22s eta: 18:10:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 271 data_time: 0.77s time: 119.60s eta: 18:12:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32027.0 tgs: 268 data_time: 0.88s time: 119.26s eta: 18:07:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31765.0 tgs: 266 data_time: 0.61s time: 119.01s eta: 18:03:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31151.0 tgs: 261 data_time: 0.85s time: 119.10s eta: 18:01:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32470.0 tgs: 272 data_time: 0.87s time: 119.26s eta: 18:01:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 1.00s time: 119.33s eta: 17:59:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32319.0 tgs: 269 data_time: 0.91s time: 120.02s eta: 18:04:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30805.0 tgs: 258 data_time: 0.86s time: 119.11s eta: 17:53:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32262.0 tgs: 270 data_time: 0.76s time: 119.41s eta: 17:54:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32600.0 tgs: 273 data_time: 0.84s time: 119.00s eta: 17:49:01 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.80s time: 119.03s eta: 17:47:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32253.0 tgs: 257 data_time: 0.98s time: 125.36s eta: 18:42:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29875.0 tgs: 248 data_time: 0.79s time: 120.10s eta: 17:52:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31345.0 tgs: 262 data_time: 0.78s time: 119.36s eta: 17:44:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30572.0 tgs: 256 data_time: 0.94s time: 119.34s eta: 17:42:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31052.0 tgs: 259 data_time: 0.93s time: 119.51s eta: 17:41:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 249 data_time: 0.78s time: 126.05s eta: 18:37:37 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31868.0 tgs: 259 data_time: 0.83s time: 122.64s eta: 18:05:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31269.0 tgs: 258 data_time: 0.84s time: 121.16s eta: 17:50:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31590.0 tgs: 265 data_time: 0.71s time: 119.04s eta: 17:29:30 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32429.0 tgs: 272 data_time: 0.78s time: 118.85s eta: 17:25:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29122.0 tgs: 234 data_time: 0.87s time: 124.00s eta: 18:09:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31224.0 tgs: 250 data_time: 0.74s time: 124.56s eta: 18:12:01 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31914.0 tgs: 267 data_time: 1.08s time: 119.40s eta: 17:24:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31436.0 tgs: 262 data_time: 1.06s time: 119.70s eta: 17:25:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.72s time: 119.53s eta: 17:21:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32406.0 tgs: 272 data_time: 0.84s time: 118.86s eta: 17:14:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31471.0 tgs: 262 data_time: 0.82s time: 119.85s eta: 17:20:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32611.0 tgs: 272 data_time: 0.79s time: 119.54s eta: 17:16:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30756.0 tgs: 252 data_time: 0.83s time: 121.85s eta: 17:34:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31603.0 tgs: 255 data_time: 0.62s time: 123.51s eta: 17:46:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 266 data_time: 0.63s time: 119.01s eta: 17:05:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30407.0 tgs: 255 data_time: 0.83s time: 119.23s eta: 17:05:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31496.0 tgs: 264 data_time: 1.04s time: 119.14s eta: 17:02:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 267 data_time: 0.86s time: 119.26s eta: 17:01:37 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31305.0 tgs: 262 data_time: 0.89s time: 119.04s eta: 16:57:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31975.0 tgs: 268 data_time: 1.04s time: 119.28s eta: 16:57:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31974.0 tgs: 268 data_time: 0.70s time: 119.16s eta: 16:54:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31648.0 tgs: 264 data_time: 0.67s time: 119.58s eta: 16:56:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32343.0 tgs: 271 data_time: 1.01s time: 119.13s eta: 16:50:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31124.0 tgs: 261 data_time: 0.76s time: 118.85s eta: 16:46:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 269 data_time: 0.50s time: 119.01s eta: 16:45:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31889.0 tgs: 266 data_time: 0.64s time: 119.85s eta: 16:50:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32260.0 tgs: 270 data_time: 0.69s time: 119.37s eta: 16:44:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31721.0 tgs: 266 data_time: 0.49s time: 119.01s eta: 16:39:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32135.0 tgs: 265 data_time: 0.95s time: 120.92s eta: 16:53:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30808.0 tgs: 258 data_time: 0.98s time: 119.00s eta: 16:35:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31500.0 tgs: 264 data_time: 0.98s time: 119.29s eta: 16:36:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 265 data_time: 1.07s time: 119.23s eta: 16:33:33 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 268 data_time: 0.69s time: 119.37s eta: 16:32:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32157.0 tgs: 268 data_time: 0.79s time: 119.70s eta: 16:33:32 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32241.0 tgs: 270 data_time: 1.07s time: 119.20s eta: 16:27:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.213 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32080.0 tgs: 267 data_time: 0.83s time: 119.76s eta: 16:30:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 263 data_time: 0.83s time: 120.38s eta: 16:33:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31611.0 tgs: 265 data_time: 0.68s time: 118.97s eta: 16:19:32 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.244 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31344.0 tgs: 246 data_time: 0.67s time: 127.33s eta: 17:26:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31867.0 tgs: 267 data_time: 0.66s time: 119.22s eta: 16:17:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 268 data_time: 0.80s time: 119.14s eta: 16:14:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31208.0 tgs: 261 data_time: 0.63s time: 119.18s eta: 16:13:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32027.0 tgs: 268 data_time: 0.75s time: 119.17s eta: 16:11:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31555.0 tgs: 265 data_time: 0.91s time: 118.95s eta: 16:07:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30723.0 tgs: 258 data_time: 0.62s time: 118.86s eta: 16:04:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30348.0 tgs: 255 data_time: 0.70s time: 118.96s eta: 16:03:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 258 data_time: 0.67s time: 121.84s eta: 16:24:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31608.0 tgs: 265 data_time: 0.65s time: 118.99s eta: 15:59:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.244 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30383.0 tgs: 254 data_time: 1.09s time: 119.23s eta: 15:59:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32454.0 tgs: 271 data_time: 0.58s time: 119.62s eta: 16:00:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31170.0 tgs: 261 data_time: 0.89s time: 119.04s eta: 15:54:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31468.0 tgs: 264 data_time: 0.84s time: 118.98s eta: 15:51:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 258 data_time: 0.62s time: 119.18s eta: 15:51:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 269 data_time: 0.61s time: 119.63s eta: 15:53:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31922.0 tgs: 268 data_time: 0.84s time: 118.94s eta: 15:45:33 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 269 data_time: 0.78s time: 119.28s eta: 15:46:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32089.0 tgs: 269 data_time: 0.69s time: 119.10s eta: 15:42:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30705.0 tgs: 258 data_time: 0.97s time: 119.00s eta: 15:40:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31201.0 tgs: 260 data_time: 0.69s time: 119.66s eta: 15:43:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30621.0 tgs: 252 data_time: 0.61s time: 121.51s eta: 15:55:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.53s time: 119.41s eta: 15:37:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31983.0 tgs: 246 data_time: 0.74s time: 129.96s eta: 16:58:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32096.0 tgs: 268 data_time: 0.73s time: 119.47s eta: 15:33:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32218.0 tgs: 269 data_time: 0.84s time: 119.61s eta: 15:32:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31551.0 tgs: 263 data_time: 0.57s time: 119.53s eta: 15:30:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.212 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31482.0 tgs: 257 data_time: 0.70s time: 122.14s eta: 15:48:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30656.0 tgs: 253 data_time: 0.90s time: 120.75s eta: 15:35:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 268 data_time: 1.03s time: 119.42s eta: 15:23:32 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31441.0 tgs: 263 data_time: 1.10s time: 119.44s eta: 15:21:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31665.0 tgs: 240 data_time: 1.01s time: 131.63s eta: 16:53:33 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32148.0 tgs: 269 data_time: 0.85s time: 119.29s eta: 15:16:33 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 268 data_time: 0.63s time: 119.55s eta: 15:16:32 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.208 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31842.0 tgs: 260 data_time: 0.93s time: 122.25s eta: 15:35:12 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31745.0 tgs: 265 data_time: 1.03s time: 119.72s eta: 15:13:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30759.0 tgs: 257 data_time: 0.79s time: 119.34s eta: 15:08:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31030.0 tgs: 260 data_time: 1.02s time: 118.90s eta: 15:03:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31701.0 tgs: 266 data_time: 1.01s time: 119.00s eta: 15:02:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31700.0 tgs: 265 data_time: 0.95s time: 119.31s eta: 15:02:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.206 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31035.0 tgs: 260 data_time: 0.89s time: 118.92s eta: 14:57:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 258 data_time: 0.80s time: 120.52s eta: 15:07:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31653.0 tgs: 265 data_time: 0.73s time: 119.31s eta: 14:56:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32533.0 tgs: 265 data_time: 0.82s time: 122.66s eta: 15:19:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 266 data_time: 0.67s time: 119.13s eta: 14:51:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30713.0 tgs: 256 data_time: 1.07s time: 119.75s eta: 14:54:10 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.73s time: 119.47s eta: 14:50:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.83s time: 118.99s eta: 14:44:30 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32563.0 tgs: 272 data_time: 0.78s time: 119.30s eta: 14:44:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.218 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32619.0 tgs: 274 data_time: 0.89s time: 118.93s eta: 14:40:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31114.0 tgs: 261 data_time: 0.90s time: 119.05s eta: 14:39:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 27050.0 tgs: 227 data_time: 0.44s time: 119.06s eta: 14:37:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31203.0 tgs: 261 data_time: 0.73s time: 119.13s eta: 14:35:37 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.201 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 270 data_time: 1.02s time: 118.99s eta: 14:32:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31684.0 tgs: 266 data_time: 0.93s time: 119.07s eta: 14:31:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31253.0 tgs: 262 data_time: 0.67s time: 119.15s eta: 14:29:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.174 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31033.0 tgs: 259 data_time: 0.71s time: 119.65s eta: 14:31:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 268 data_time: 0.67s time: 119.18s eta: 14:26:01 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32325.0 tgs: 271 data_time: 0.91s time: 119.20s eta: 14:24:10 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31524.0 tgs: 264 data_time: 0.65s time: 119.39s eta: 14:23:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.174 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31265.0 tgs: 258 data_time: 1.07s time: 120.74s eta: 14:31:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31189.0 tgs: 262 data_time: 0.93s time: 118.86s eta: 14:15:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31533.0 tgs: 264 data_time: 0.93s time: 119.09s eta: 14:15:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.70s time: 119.13s eta: 14:13:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.253 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31560.0 tgs: 263 data_time: 0.87s time: 119.96s eta: 14:17:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31511.0 tgs: 262 data_time: 0.74s time: 119.87s eta: 14:15:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31916.0 tgs: 266 data_time: 0.66s time: 119.59s eta: 14:11:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31170.0 tgs: 261 data_time: 0.83s time: 119.22s eta: 14:06:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31944.0 tgs: 267 data_time: 0.75s time: 119.27s eta: 14:04:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32252.0 tgs: 270 data_time: 0.69s time: 119.26s eta: 14:02:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.202 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 269 data_time: 1.10s time: 119.14s eta: 13:59:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 270 data_time: 0.61s time: 119.12s eta: 13:57:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.84s time: 119.17s eta: 13:56:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.189 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31622.0 tgs: 264 data_time: 1.08s time: 119.35s eta: 13:55:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.256 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32179.0 tgs: 267 data_time: 1.40s time: 120.12s eta: 13:58:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32226.0 tgs: 262 data_time: 1.13s time: 122.56s eta: 14:13:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31889.0 tgs: 267 data_time: 1.00s time: 119.10s eta: 13:47:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.251 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31886.0 tgs: 267 data_time: 0.74s time: 119.37s eta: 13:47:37 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31640.0 tgs: 265 data_time: 0.76s time: 119.32s eta: 13:45:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 1.00s time: 119.33s eta: 13:43:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32138.0 tgs: 269 data_time: 0.89s time: 119.12s eta: 13:39:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31108.0 tgs: 260 data_time: 0.97s time: 119.25s eta: 13:38:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31461.0 tgs: 264 data_time: 0.71s time: 119.14s eta: 13:36:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 265 data_time: 0.89s time: 119.22s eta: 13:34:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32501.0 tgs: 272 data_time: 1.00s time: 119.46s eta: 13:34:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32122.0 tgs: 269 data_time: 0.81s time: 119.08s eta: 13:29:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31803.0 tgs: 267 data_time: 0.66s time: 118.92s eta: 13:26:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32254.0 tgs: 269 data_time: 0.77s time: 119.81s eta: 13:30:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 272 data_time: 1.04s time: 119.26s eta: 13:24:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.223 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31945.0 tgs: 268 data_time: 1.17s time: 119.17s eta: 13:22:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31334.0 tgs: 262 data_time: 1.04s time: 119.34s eta: 13:21:34 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 32048.0 tgs: 268 data_time: 0.92s time: 119.26s eta: 13:19:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32046.0 tgs: 269 data_time: 0.73s time: 119.04s eta: 13:15:34 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 264 data_time: 0.96s time: 119.35s eta: 13:15:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30547.0 tgs: 256 data_time: 0.64s time: 119.18s eta: 13:12:31 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 266 data_time: 0.94s time: 119.46s eta: 13:12:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.207 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 269 data_time: 1.11s time: 119.29s eta: 13:09:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.206 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31788.0 tgs: 266 data_time: 0.98s time: 119.15s eta: 13:06:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31874.0 tgs: 264 data_time: 0.98s time: 120.47s eta: 13:13:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32387.0 tgs: 269 data_time: 0.97s time: 120.16s eta: 13:09:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.120 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30335.0 tgs: 253 data_time: 0.55s time: 119.50s eta: 13:02:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30033.0 tgs: 251 data_time: 0.78s time: 119.51s eta: 13:00:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32025.0 tgs: 267 data_time: 0.82s time: 119.57s eta: 12:59:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31801.0 tgs: 265 data_time: 0.67s time: 119.80s eta: 12:58:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.115 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30774.0 tgs: 258 data_time: 0.86s time: 119.04s eta: 12:51:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31518.0 tgs: 264 data_time: 0.96s time: 119.05s eta: 12:49:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.120 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31386.0 tgs: 263 data_time: 0.70s time: 119.12s eta: 12:48:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.68s time: 119.40s eta: 12:48:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32166.0 tgs: 269 data_time: 0.88s time: 119.18s eta: 12:44:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32107.0 tgs: 268 data_time: 0.94s time: 119.44s eta: 12:44:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.183 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 1.05s time: 119.10s eta: 12:40:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31512.0 tgs: 264 data_time: 0.71s time: 119.07s eta: 12:38:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31533.0 tgs: 263 data_time: 0.88s time: 119.60s eta: 12:39:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31439.0 tgs: 264 data_time: 0.68s time: 118.79s eta: 12:32:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.174 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31130.0 tgs: 261 data_time: 0.67s time: 118.88s eta: 12:30:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32011.0 tgs: 268 data_time: 0.87s time: 119.17s eta: 12:30:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.111 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 25762.0 tgs: 215 data_time: 0.52s time: 119.27s eta: 12:29:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 270 data_time: 0.91s time: 119.35s eta: 12:27:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31986.0 tgs: 268 data_time: 0.86s time: 119.12s eta: 12:24:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 263 data_time: 1.00s time: 119.35s eta: 12:23:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32126.0 tgs: 269 data_time: 0.80s time: 119.26s eta: 12:21:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.200 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31207.0 tgs: 262 data_time: 0.95s time: 119.03s eta: 12:17:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.110 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 262 data_time: 0.60s time: 121.13s eta: 12:28:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31262.0 tgs: 260 data_time: 0.85s time: 119.98s eta: 12:19:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.208 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31381.0 tgs: 264 data_time: 0.83s time: 118.70s eta: 12:09:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31519.0 tgs: 263 data_time: 0.84s time: 119.64s eta: 12:13:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.196 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 268 data_time: 0.82s time: 119.83s eta: 12:12:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.144 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 264 data_time: 0.85s time: 119.14s eta: 12:06:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31152.0 tgs: 261 data_time: 0.78s time: 119.17s eta: 12:04:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.97s time: 119.12s eta: 12:02:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.215 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31252.0 tgs: 262 data_time: 0.94s time: 119.19s eta: 12:01:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30537.0 tgs: 256 data_time: 1.04s time: 118.96s eta: 11:57:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.193 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30712.0 tgs: 256 data_time: 0.55s time: 119.71s eta: 12:00:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31275.0 tgs: 262 data_time: 0.46s time: 119.27s eta: 11:55:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32041.0 tgs: 268 data_time: 0.74s time: 119.28s eta: 11:53:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31777.0 tgs: 265 data_time: 0.83s time: 119.71s eta: 11:54:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32515.0 tgs: 273 data_time: 0.74s time: 118.80s eta: 11:46:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.81s time: 119.15s eta: 11:46:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31652.0 tgs: 264 data_time: 0.97s time: 119.61s eta: 11:47:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31256.0 tgs: 263 data_time: 0.92s time: 118.83s eta: 11:41:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32109.0 tgs: 270 data_time: 0.99s time: 118.68s eta: 11:38:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.57s time: 119.22s eta: 11:39:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31108.0 tgs: 261 data_time: 0.93s time: 119.14s eta: 11:36:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.217 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 268 data_time: 0.90s time: 119.07s eta: 11:34:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.112 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 263 data_time: 0.73s time: 118.95s eta: 11:31:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31063.0 tgs: 261 data_time: 0.88s time: 118.91s eta: 11:29:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31995.0 tgs: 268 data_time: 0.76s time: 118.97s eta: 11:28:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.219 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32662.0 tgs: 273 data_time: 0.81s time: 119.28s eta: 11:27:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31996.0 tgs: 268 data_time: 0.63s time: 119.35s eta: 11:26:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31004.0 tgs: 260 data_time: 0.74s time: 118.88s eta: 11:21:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32430.0 tgs: 272 data_time: 0.78s time: 119.21s eta: 11:21:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30334.0 tgs: 253 data_time: 0.69s time: 119.72s eta: 11:22:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.180 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31948.0 tgs: 268 data_time: 1.00s time: 119.12s eta: 11:16:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31792.0 tgs: 266 data_time: 0.61s time: 119.12s eta: 11:15:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31595.0 tgs: 265 data_time: 0.94s time: 118.95s eta: 11:12:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30196.0 tgs: 253 data_time: 0.79s time: 119.10s eta: 11:10:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.194 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31338.0 tgs: 262 data_time: 0.64s time: 119.16s eta: 11:09:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31464.0 tgs: 264 data_time: 0.94s time: 118.97s eta: 11:06:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30987.0 tgs: 259 data_time: 0.74s time: 119.47s eta: 11:07:01 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31777.0 tgs: 265 data_time: 0.89s time: 119.65s eta: 11:06:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31707.0 tgs: 265 data_time: 0.82s time: 119.31s eta: 11:02:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.122 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30829.0 tgs: 258 data_time: 0.91s time: 119.30s eta: 11:00:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.202 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31273.0 tgs: 263 data_time: 0.89s time: 118.82s eta: 10:55:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32308.0 tgs: 270 data_time: 0.69s time: 119.47s eta: 10:57:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.126 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31749.0 tgs: 265 data_time: 0.57s time: 119.37s eta: 10:54:31 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31767.0 tgs: 266 data_time: 0.72s time: 119.28s eta: 10:52:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 267 data_time: 0.74s time: 119.26s eta: 10:49:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 267 data_time: 0.85s time: 119.10s eta: 10:47:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31298.0 tgs: 261 data_time: 0.77s time: 119.53s eta: 10:47:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.201 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31337.0 tgs: 262 data_time: 0.69s time: 119.31s eta: 10:44:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31711.0 tgs: 265 data_time: 0.78s time: 119.49s eta: 10:43:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31513.0 tgs: 263 data_time: 0.83s time: 119.52s eta: 10:41:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31466.0 tgs: 264 data_time: 0.98s time: 118.99s eta: 10:36:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31991.0 tgs: 268 data_time: 1.17s time: 119.22s eta: 10:35:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31635.0 tgs: 265 data_time: 0.75s time: 119.06s eta: 10:33:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31474.0 tgs: 264 data_time: 0.61s time: 118.78s eta: 10:29:32 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31774.0 tgs: 265 data_time: 0.75s time: 119.88s eta: 10:33:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31488.0 tgs: 263 data_time: 0.89s time: 119.37s eta: 10:28:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32535.0 tgs: 273 data_time: 0.80s time: 118.84s eta: 10:23:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 267 data_time: 0.94s time: 118.99s eta: 10:22:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31715.0 tgs: 265 data_time: 0.74s time: 119.28s eta: 10:22:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.216 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31553.0 tgs: 264 data_time: 0.63s time: 119.20s eta: 10:19:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 270 data_time: 0.78s time: 119.06s eta: 10:17:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.75s time: 119.48s eta: 10:17:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31907.0 tgs: 267 data_time: 0.61s time: 119.37s eta: 10:14:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.199 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32006.0 tgs: 267 data_time: 0.41s time: 119.45s eta: 10:13:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29892.0 tgs: 250 data_time: 0.69s time: 119.29s eta: 10:10:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.126 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30783.0 tgs: 258 data_time: 0.60s time: 119.11s eta: 10:07:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32288.0 tgs: 269 data_time: 0.80s time: 119.70s eta: 10:08:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.211 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 265 data_time: 0.88s time: 119.83s eta: 10:07:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.211 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32161.0 tgs: 264 data_time: 1.00s time: 121.56s eta: 10:13:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.85s time: 119.26s eta: 10:00:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31992.0 tgs: 269 data_time: 1.00s time: 118.80s eta: 9:56:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32072.0 tgs: 268 data_time: 0.68s time: 119.57s eta: 9:57:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31033.0 tgs: 259 data_time: 0.94s time: 119.70s eta: 9:56:31 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31291.0 tgs: 262 data_time: 0.88s time: 119.26s eta: 9:52:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 268 data_time: 1.01s time: 119.33s eta: 9:50:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30506.0 tgs: 255 data_time: 0.74s time: 119.22s eta: 9:48:10 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.117 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31095.0 tgs: 257 data_time: 0.89s time: 120.89s eta: 9:54:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.114 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32378.0 tgs: 271 data_time: 0.61s time: 119.27s eta: 9:44:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.238 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31737.0 tgs: 265 data_time: 0.84s time: 119.47s eta: 9:43:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31010.0 tgs: 259 data_time: 0.82s time: 119.55s eta: 9:41:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31878.0 tgs: 267 data_time: 0.49s time: 119.20s eta: 9:38:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.118 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31652.0 tgs: 264 data_time: 0.71s time: 119.48s eta: 9:37:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 265 data_time: 0.92s time: 119.05s eta: 9:33:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 262 data_time: 0.65s time: 119.24s eta: 9:32:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31283.0 tgs: 262 data_time: 0.82s time: 119.16s eta: 9:29:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32064.0 tgs: 268 data_time: 1.05s time: 119.23s eta: 9:28:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31438.0 tgs: 263 data_time: 0.83s time: 119.46s eta: 9:27:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31072.0 tgs: 261 data_time: 0.75s time: 118.99s eta: 9:23:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31779.0 tgs: 266 data_time: 0.71s time: 119.03s eta: 9:21:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31632.0 tgs: 264 data_time: 1.00s time: 119.54s eta: 9:21:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.113 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32297.0 tgs: 270 data_time: 0.70s time: 119.59s eta: 9:20:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.174 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31681.0 tgs: 265 data_time: 0.66s time: 119.12s eta: 9:15:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.166 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32415.0 tgs: 272 data_time: 0.68s time: 119.00s eta: 9:13:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31327.0 tgs: 263 data_time: 0.76s time: 118.88s eta: 9:10:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.172 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.97s time: 119.32s eta: 9:10:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 1.01s time: 119.09s eta: 9:07:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31427.0 tgs: 264 data_time: 0.78s time: 118.86s eta: 9:04:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.187 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 264 data_time: 0.70s time: 119.67s eta: 9:06:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31164.0 tgs: 260 data_time: 0.96s time: 119.44s eta: 9:03:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31258.0 tgs: 261 data_time: 0.74s time: 119.43s eta: 9:01:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 265 data_time: 0.76s time: 118.74s eta: 8:56:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31022.0 tgs: 260 data_time: 0.87s time: 118.87s eta: 8:54:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.199 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31953.0 tgs: 267 data_time: 0.91s time: 119.33s eta: 8:55:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.115 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30599.0 tgs: 256 data_time: 0.80s time: 119.15s eta: 8:52:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.100 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31640.0 tgs: 265 data_time: 0.52s time: 119.27s eta: 8:50:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31856.0 tgs: 264 data_time: 0.83s time: 120.48s eta: 8:54:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 267 data_time: 0.78s time: 119.24s eta: 8:46:37 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 266 data_time: 0.93s time: 119.07s eta: 8:43:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 263 data_time: 1.01s time: 120.83s eta: 8:49:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31839.0 tgs: 264 data_time: 0.78s time: 120.19s eta: 8:44:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31794.0 tgs: 267 data_time: 0.54s time: 118.83s eta: 8:36:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32118.0 tgs: 269 data_time: 0.82s time: 119.35s eta: 8:37:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.230 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.74s time: 119.33s eta: 8:35:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31845.0 tgs: 267 data_time: 0.74s time: 119.06s eta: 8:31:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.193 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 269 data_time: 0.78s time: 119.16s eta: 8:30:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.71s time: 119.04s eta: 8:27:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.185 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32051.0 tgs: 269 data_time: 0.70s time: 119.12s eta: 8:26:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.74s time: 119.25s eta: 8:24:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.192 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.77s time: 119.22s eta: 8:22:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31756.0 tgs: 266 data_time: 0.80s time: 119.38s eta: 8:21:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.177 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31863.0 tgs: 266 data_time: 0.88s time: 119.78s eta: 8:21:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31688.0 tgs: 265 data_time: 0.85s time: 119.20s eta: 8:16:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31434.0 tgs: 264 data_time: 0.63s time: 118.99s eta: 8:13:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31934.0 tgs: 266 data_time: 0.67s time: 119.89s eta: 8:15:33 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 268 data_time: 0.63s time: 119.55s eta: 8:12:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32172.0 tgs: 269 data_time: 0.86s time: 119.60s eta: 8:10:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31739.0 tgs: 266 data_time: 0.87s time: 119.00s eta: 8:05:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31506.0 tgs: 262 data_time: 0.84s time: 119.84s eta: 8:07:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.114 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 266 data_time: 0.76s time: 119.71s eta: 8:04:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31864.0 tgs: 266 data_time: 0.92s time: 119.41s eta: 8:01:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.104 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32174.0 tgs: 269 data_time: 0.81s time: 119.28s eta: 7:59:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31700.0 tgs: 265 data_time: 0.64s time: 119.19s eta: 7:56:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.111 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31403.0 tgs: 263 data_time: 0.67s time: 119.20s eta: 7:54:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31207.0 tgs: 262 data_time: 0.61s time: 118.93s eta: 7:51:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.224 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31558.0 tgs: 265 data_time: 1.02s time: 118.97s eta: 7:49:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 270 data_time: 0.66s time: 119.42s eta: 7:49:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.174 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 270 data_time: 0.61s time: 119.04s eta: 7:46:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.225 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32332.0 tgs: 271 data_time: 0.90s time: 119.07s eta: 7:44:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 268 data_time: 0.81s time: 120.01s eta: 7:46:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31324.0 tgs: 261 data_time: 0.69s time: 119.82s eta: 7:43:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32501.0 tgs: 272 data_time: 0.64s time: 119.17s eta: 7:38:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32108.0 tgs: 269 data_time: 0.86s time: 118.96s eta: 7:36:01 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.154 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31892.0 tgs: 267 data_time: 0.80s time: 119.33s eta: 7:35:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.216 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31726.0 tgs: 265 data_time: 0.84s time: 119.38s eta: 7:33:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.128 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31653.0 tgs: 265 data_time: 0.59s time: 119.06s eta: 7:30:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.103 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31597.0 tgs: 265 data_time: 0.79s time: 118.99s eta: 7:28:12 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30377.0 tgs: 254 data_time: 0.57s time: 119.24s eta: 7:27:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31018.0 tgs: 260 data_time: 0.46s time: 118.99s eta: 7:24:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.195 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30388.0 tgs: 255 data_time: 0.68s time: 118.93s eta: 7:22:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 262 data_time: 0.71s time: 119.08s eta: 7:20:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31932.0 tgs: 268 data_time: 0.83s time: 118.99s eta: 7:18:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31833.0 tgs: 267 data_time: 0.83s time: 119.03s eta: 7:16:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31649.0 tgs: 265 data_time: 0.71s time: 119.24s eta: 7:15:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31558.0 tgs: 264 data_time: 0.74s time: 119.36s eta: 7:13:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.100 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31390.0 tgs: 263 data_time: 0.65s time: 118.94s eta: 7:10:10 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.196 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.55s time: 119.21s eta: 7:09:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.190 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31845.0 tgs: 266 data_time: 0.56s time: 119.58s eta: 7:08:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29935.0 tgs: 250 data_time: 0.77s time: 119.66s eta: 7:06:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.192 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31746.0 tgs: 266 data_time: 0.72s time: 118.96s eta: 7:02:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31721.0 tgs: 265 data_time: 0.70s time: 119.30s eta: 7:01:31 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31320.0 tgs: 262 data_time: 1.02s time: 119.52s eta: 7:00:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32203.0 tgs: 269 data_time: 0.67s time: 119.43s eta: 6:57:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.194 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 272 data_time: 0.69s time: 118.91s eta: 6:54:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.239 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 270 data_time: 0.95s time: 119.34s eta: 6:53:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 263 data_time: 0.76s time: 118.93s eta: 6:50:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32205.0 tgs: 270 data_time: 0.73s time: 119.27s eta: 6:49:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30968.0 tgs: 260 data_time: 0.82s time: 119.02s eta: 6:46:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31854.0 tgs: 267 data_time: 0.60s time: 119.22s eta: 6:45:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 269 data_time: 0.83s time: 119.38s eta: 6:43:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32035.0 tgs: 269 data_time: 0.60s time: 119.01s eta: 6:40:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32042.0 tgs: 269 data_time: 0.97s time: 119.01s eta: 6:38:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.203 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 264 data_time: 0.73s time: 119.46s eta: 6:38:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31318.0 tgs: 263 data_time: 0.85s time: 118.80s eta: 6:34:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.70s time: 119.44s eta: 6:34:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31646.0 tgs: 264 data_time: 0.75s time: 119.57s eta: 6:32:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31349.0 tgs: 262 data_time: 0.71s time: 119.38s eta: 6:29:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.237 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30824.0 tgs: 257 data_time: 0.53s time: 119.71s eta: 6:29:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32354.0 tgs: 271 data_time: 0.65s time: 119.36s eta: 6:25:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31166.0 tgs: 262 data_time: 0.75s time: 118.82s eta: 6:22:12 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31988.0 tgs: 268 data_time: 0.99s time: 119.02s eta: 6:20:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31964.0 tgs: 267 data_time: 0.98s time: 119.40s eta: 6:20:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.198 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32179.0 tgs: 269 data_time: 0.98s time: 119.34s eta: 6:17:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31511.0 tgs: 264 data_time: 0.91s time: 119.05s eta: 6:15:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31463.0 tgs: 263 data_time: 0.72s time: 119.38s eta: 6:14:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31773.0 tgs: 266 data_time: 0.86s time: 119.39s eta: 6:12:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.202 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31893.0 tgs: 267 data_time: 0.72s time: 119.29s eta: 6:09:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.83s time: 119.10s eta: 6:07:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31700.0 tgs: 259 data_time: 0.81s time: 121.96s eta: 6:14:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31492.0 tgs: 263 data_time: 0.75s time: 119.39s eta: 6:04:07 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31332.0 tgs: 262 data_time: 0.70s time: 119.57s eta: 6:02:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31086.0 tgs: 260 data_time: 0.84s time: 119.26s eta: 5:59:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 27321.0 tgs: 229 data_time: 0.65s time: 119.27s eta: 5:57:48 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.205 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31862.0 tgs: 267 data_time: 0.98s time: 118.91s eta: 5:54:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 259 data_time: 0.92s time: 119.02s eta: 5:53:05 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31457.0 tgs: 264 data_time: 0.85s time: 119.07s eta: 5:51:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30236.0 tgs: 253 data_time: 0.90s time: 119.29s eta: 5:49:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31960.0 tgs: 268 data_time: 0.76s time: 118.97s eta: 5:47:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32273.0 tgs: 270 data_time: 0.73s time: 119.27s eta: 5:45:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.232 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32334.0 tgs: 270 data_time: 0.86s time: 119.33s eta: 5:44:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32077.0 tgs: 269 data_time: 0.73s time: 119.06s eta: 5:41:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31705.0 tgs: 265 data_time: 0.69s time: 119.50s eta: 5:40:34 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31560.0 tgs: 264 data_time: 0.82s time: 119.38s eta: 5:38:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.113 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30543.0 tgs: 255 data_time: 0.92s time: 119.57s eta: 5:36:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31756.0 tgs: 266 data_time: 0.76s time: 118.98s eta: 5:33:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31548.0 tgs: 265 data_time: 0.87s time: 118.95s eta: 5:31:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.111 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.61s time: 119.24s eta: 5:29:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30745.0 tgs: 257 data_time: 0.98s time: 119.40s eta: 5:28:21 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31753.0 tgs: 265 data_time: 0.93s time: 119.46s eta: 5:26:31 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31738.0 tgs: 266 data_time: 0.69s time: 118.95s eta: 5:23:08 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31359.0 tgs: 262 data_time: 0.59s time: 119.63s eta: 5:23:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31478.0 tgs: 263 data_time: 0.73s time: 119.35s eta: 5:20:14 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31081.0 tgs: 260 data_time: 0.69s time: 119.12s eta: 5:17:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.105 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31347.0 tgs: 262 data_time: 0.53s time: 119.42s eta: 5:16:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31565.0 tgs: 263 data_time: 0.98s time: 119.58s eta: 5:14:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.246 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 267 data_time: 0.84s time: 119.00s eta: 5:11:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 265 data_time: 0.71s time: 119.00s eta: 5:09:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31608.0 tgs: 265 data_time: 0.91s time: 119.23s eta: 5:08:00 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.259 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 268 data_time: 0.76s time: 119.69s eta: 5:07:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31961.0 tgs: 268 data_time: 0.71s time: 119.11s eta: 5:03:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32214.0 tgs: 270 data_time: 0.90s time: 119.07s eta: 5:01:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32104.0 tgs: 264 data_time: 0.91s time: 121.52s eta: 5:05:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.118 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 266 data_time: 0.82s time: 119.58s eta: 4:58:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.185 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30327.0 tgs: 253 data_time: 0.52s time: 119.50s eta: 4:56:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.199 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32344.0 tgs: 270 data_time: 1.12s time: 119.73s eta: 4:55:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.114 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31880.0 tgs: 266 data_time: 0.52s time: 119.52s eta: 4:52:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30378.0 tgs: 254 data_time: 0.86s time: 119.41s eta: 4:50:34 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31216.0 tgs: 261 data_time: 0.77s time: 119.21s eta: 4:48:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 265 data_time: 0.65s time: 118.88s eta: 4:45:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.179 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 265 data_time: 0.86s time: 119.05s eta: 4:43:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.110 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 28332.0 tgs: 236 data_time: 0.65s time: 119.98s eta: 4:43:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 270 data_time: 0.69s time: 119.15s eta: 4:39:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31138.0 tgs: 261 data_time: 0.72s time: 119.01s eta: 4:37:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31315.0 tgs: 263 data_time: 0.87s time: 119.06s eta: 4:35:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30717.0 tgs: 257 data_time: 0.86s time: 119.37s eta: 4:34:33 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31628.0 tgs: 264 data_time: 0.59s time: 119.41s eta: 4:32:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30401.0 tgs: 255 data_time: 0.59s time: 118.94s eta: 4:29:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32144.0 tgs: 268 data_time: 0.74s time: 119.60s eta: 4:29:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31502.0 tgs: 264 data_time: 0.94s time: 119.24s eta: 4:26:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.72s time: 119.06s eta: 4:23:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 269 data_time: 0.63s time: 118.85s eta: 4:21:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30956.0 tgs: 260 data_time: 0.73s time: 118.76s eta: 4:19:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29499.0 tgs: 247 data_time: 0.69s time: 119.43s eta: 4:18:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.119 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 266 data_time: 0.63s time: 119.41s eta: 4:16:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 28527.0 tgs: 239 data_time: 0.60s time: 119.34s eta: 4:14:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 261 data_time: 0.78s time: 119.05s eta: 4:11:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31333.0 tgs: 263 data_time: 0.85s time: 119.12s eta: 4:10:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31398.0 tgs: 263 data_time: 0.65s time: 119.14s eta: 4:08:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.112 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30799.0 tgs: 258 data_time: 0.64s time: 118.96s eta: 4:05:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31257.0 tgs: 262 data_time: 0.70s time: 119.15s eta: 4:04:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32315.0 tgs: 270 data_time: 0.75s time: 119.50s eta: 4:02:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32550.0 tgs: 272 data_time: 0.63s time: 119.25s eta: 4:00:28 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31342.0 tgs: 262 data_time: 0.72s time: 119.19s eta: 3:58:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31153.0 tgs: 261 data_time: 0.62s time: 119.36s eta: 3:56:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31063.0 tgs: 260 data_time: 0.73s time: 119.37s eta: 3:54:45 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 265 data_time: 0.85s time: 119.95s eta: 3:53:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 271 data_time: 0.69s time: 119.32s eta: 3:50:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30971.0 tgs: 258 data_time: 0.93s time: 119.64s eta: 3:49:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31112.0 tgs: 261 data_time: 0.68s time: 118.86s eta: 3:45:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 267 data_time: 0.99s time: 119.70s eta: 3:45:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 269 data_time: 0.83s time: 119.33s eta: 3:42:44 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31498.0 tgs: 264 data_time: 0.79s time: 119.26s eta: 3:40:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.238 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30580.0 tgs: 256 data_time: 0.91s time: 119.00s eta: 3:38:10 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 1.22s time: 119.43s eta: 3:36:57 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31384.0 tgs: 263 data_time: 0.89s time: 119.04s eta: 3:34:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31768.0 tgs: 266 data_time: 0.70s time: 118.99s eta: 3:32:12 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.74s time: 119.22s eta: 3:30:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30898.0 tgs: 257 data_time: 0.78s time: 120.12s eta: 3:30:12 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 268 data_time: 0.89s time: 119.13s eta: 3:26:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31751.0 tgs: 266 data_time: 0.83s time: 119.32s eta: 3:24:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 270 data_time: 0.50s time: 119.08s eta: 3:22:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32224.0 tgs: 269 data_time: 0.90s time: 119.43s eta: 3:21:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30820.0 tgs: 258 data_time: 0.78s time: 119.29s eta: 3:18:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31507.0 tgs: 264 data_time: 0.79s time: 119.22s eta: 3:16:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.95s time: 119.14s eta: 3:14:35 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30908.0 tgs: 259 data_time: 0.78s time: 119.02s eta: 3:12:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32441.0 tgs: 272 data_time: 0.63s time: 119.18s eta: 3:10:41 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.203 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30654.0 tgs: 253 data_time: 0.91s time: 121.06s eta: 3:11:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32062.0 tgs: 268 data_time: 0.70s time: 119.50s eta: 3:07:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32494.0 tgs: 272 data_time: 0.97s time: 119.26s eta: 3:04:51 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30956.0 tgs: 260 data_time: 0.74s time: 118.99s eta: 3:02:27 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31562.0 tgs: 264 data_time: 0.77s time: 119.22s eta: 3:00:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.185 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31709.0 tgs: 265 data_time: 0.79s time: 119.46s eta: 2:59:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 262 data_time: 0.73s time: 118.98s eta: 2:56:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.198 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32415.0 tgs: 271 data_time: 1.16s time: 119.35s eta: 2:55:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29085.0 tgs: 243 data_time: 0.84s time: 119.42s eta: 2:53:09 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31686.0 tgs: 265 data_time: 0.77s time: 119.35s eta: 2:51:03 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31625.0 tgs: 260 data_time: 0.70s time: 121.46s eta: 2:52:04 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31391.0 tgs: 263 data_time: 0.58s time: 119.24s eta: 2:46:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31410.0 tgs: 262 data_time: 1.01s time: 119.56s eta: 2:45:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 270 data_time: 0.71s time: 118.90s eta: 2:42:30 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 0.76s time: 119.19s eta: 2:40:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 267 data_time: 0.53s time: 118.96s eta: 2:38:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32061.0 tgs: 268 data_time: 0.83s time: 119.23s eta: 2:36:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.189 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31916.0 tgs: 267 data_time: 0.76s time: 119.11s eta: 2:34:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 266 data_time: 0.82s time: 119.43s eta: 2:33:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31433.0 tgs: 261 data_time: 1.13s time: 120.23s eta: 2:32:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 264 data_time: 0.92s time: 119.29s eta: 2:29:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32083.0 tgs: 269 data_time: 0.83s time: 118.88s eta: 2:26:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31396.0 tgs: 262 data_time: 0.70s time: 119.52s eta: 2:25:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31508.0 tgs: 264 data_time: 0.54s time: 119.07s eta: 2:22:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31513.0 tgs: 264 data_time: 0.60s time: 119.14s eta: 2:20:58 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31506.0 tgs: 264 data_time: 0.79s time: 119.05s eta: 2:18:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32410.0 tgs: 270 data_time: 0.71s time: 119.84s eta: 2:17:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 268 data_time: 1.03s time: 119.32s eta: 2:15:13 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31906.0 tgs: 267 data_time: 0.64s time: 119.38s eta: 2:13:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 267 data_time: 0.86s time: 119.68s eta: 2:11:39 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31768.0 tgs: 265 data_time: 0.88s time: 119.49s eta: 2:09:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 270 data_time: 0.66s time: 118.96s eta: 2:06:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31616.0 tgs: 264 data_time: 0.60s time: 119.38s eta: 2:05:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30679.0 tgs: 257 data_time: 0.74s time: 119.27s eta: 2:03:15 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30337.0 tgs: 254 data_time: 0.60s time: 119.22s eta: 2:01:12 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32212.0 tgs: 269 data_time: 0.79s time: 119.51s eta: 1:59:30 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.66s time: 119.42s eta: 1:57:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32505.0 tgs: 271 data_time: 0.85s time: 119.53s eta: 1:55:32 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31980.0 tgs: 268 data_time: 0.63s time: 119.30s eta: 1:53:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31335.0 tgs: 262 data_time: 0.88s time: 119.25s eta: 1:51:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32336.0 tgs: 271 data_time: 0.80s time: 118.93s eta: 1:49:01 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31302.0 tgs: 261 data_time: 0.57s time: 119.64s eta: 1:47:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31557.0 tgs: 263 data_time: 0.79s time: 119.69s eta: 1:45:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31294.0 tgs: 262 data_time: 0.75s time: 119.35s eta: 1:43:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.235 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31227.0 tgs: 261 data_time: 0.68s time: 119.31s eta: 1:41:24 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31983.0 tgs: 268 data_time: 1.00s time: 119.28s eta: 1:39:23 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32110.0 tgs: 270 data_time: 0.78s time: 118.92s eta: 1:37:06 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 263 data_time: 0.86s time: 119.15s eta: 1:35:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32043.0 tgs: 269 data_time: 1.02s time: 119.07s eta: 1:33:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32253.0 tgs: 270 data_time: 0.70s time: 119.25s eta: 1:31:25 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30673.0 tgs: 256 data_time: 0.85s time: 119.73s eta: 1:29:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31562.0 tgs: 265 data_time: 0.73s time: 119.02s eta: 1:27:16 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 263 data_time: 0.76s time: 118.87s eta: 1:25:11 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32139.0 tgs: 268 data_time: 0.74s time: 119.48s eta: 1:23:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.238 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30337.0 tgs: 253 data_time: 0.77s time: 119.48s eta: 1:21:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32527.0 tgs: 273 data_time: 0.83s time: 118.94s eta: 1:19:17 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31543.0 tgs: 265 data_time: 0.66s time: 118.94s eta: 1:17:18 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 263 data_time: 0.59s time: 122.13s eta: 1:17:20 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32360.0 tgs: 271 data_time: 0.57s time: 119.36s eta: 1:13:36 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31728.0 tgs: 266 data_time: 0.72s time: 119.06s eta: 1:11:26 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31053.0 tgs: 261 data_time: 0.69s time: 118.92s eta: 1:09:22 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31778.0 tgs: 266 data_time: 0.86s time: 119.11s eta: 1:07:29 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31475.0 tgs: 263 data_time: 0.71s time: 119.42s eta: 1:05:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29530.0 tgs: 247 data_time: 0.65s time: 119.34s eta: 1:03:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30940.0 tgs: 259 data_time: 0.64s time: 119.18s eta: 1:01:34 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.240 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31634.0 tgs: 265 data_time: 0.62s time: 119.27s eta: 0:59:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 269 data_time: 0.67s time: 119.44s eta: 0:57:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30613.0 tgs: 256 data_time: 0.85s time: 119.40s eta: 0:55:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31940.0 tgs: 267 data_time: 0.81s time: 119.48s eta: 0:53:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.64s time: 119.37s eta: 0:51:43 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31155.0 tgs: 260 data_time: 0.66s time: 119.47s eta: 0:49:46 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31618.0 tgs: 265 data_time: 0.82s time: 118.90s eta: 0:47:33 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32178.0 tgs: 269 data_time: 0.69s time: 119.26s eta: 0:45:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.66s time: 119.20s eta: 0:43:42 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32354.0 tgs: 271 data_time: 0.85s time: 119.00s eta: 0:41:38 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32505.0 tgs: 273 data_time: 0.77s time: 119.04s eta: 0:39:40 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.210 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32119.0 tgs: 268 data_time: 0.69s time: 119.66s eta: 0:37:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32210.0 tgs: 270 data_time: 0.58s time: 119.29s eta: 0:35:47 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32281.0 tgs: 269 data_time: 0.75s time: 119.75s eta: 0:33:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30657.0 tgs: 252 data_time: 0.65s time: 121.23s eta: 0:32:19 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31457.0 tgs: 263 data_time: 0.83s time: 119.39s eta: 0:29:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.77s time: 119.32s eta: 0:27:50 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31402.0 tgs: 262 data_time: 0.61s time: 119.64s eta: 0:25:55 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30715.0 tgs: 257 data_time: 0.58s time: 119.10s eta: 0:23:49 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.106 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31496.0 tgs: 263 data_time: 1.01s time: 119.38s eta: 0:21:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 264 data_time: 0.69s time: 119.26s eta: 0:19:52 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31642.0 tgs: 265 data_time: 0.73s time: 119.24s eta: 0:17:53 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 270 data_time: 0.65s time: 119.90s eta: 0:15:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.214 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31768.0 tgs: 266 data_time: 0.83s time: 119.19s eta: 0:13:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32268.0 tgs: 270 data_time: 0.90s time: 119.09s eta: 0:11:54 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32354.0 tgs: 271 data_time: 0.87s time: 119.32s eta: 0:09:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32019.0 tgs: 268 data_time: 0.71s time: 119.17s eta: 0:07:56 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31728.0 tgs: 265 data_time: 0.73s time: 119.69s eta: 0:05:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30597.0 tgs: 252 data_time: 1.09s time: 121.37s eta: 0:04:02 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32060.0 tgs: 268 data_time: 0.66s time: 119.21s eta: 0:01:59 +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.28 seconds, peak gpu memory 17.4G +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 46][DP 11][SP 2][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank47.log b/20250121165312/rank47.log new file mode 100644 index 0000000000000000000000000000000000000000..27815710ed3a081b01c3f5b047166cce6a674aa0 --- /dev/null +++ b/20250121165312/rank47.log @@ -0,0 +1,851 @@ +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:33][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.70 seconds, peak gpu memory 13.4G +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.238 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.0GB text_tokens: 31361.0 tgs: 204 data_time: 1.75s time: 153.30s eta: 1 day, 1:15:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.265 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32373.0 tgs: 269 data_time: 1.20s time: 120.17s eta: 19:45:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.226 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31271.0 tgs: 261 data_time: 1.16s time: 119.69s eta: 19:38:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31959.0 tgs: 267 data_time: 0.77s time: 119.65s eta: 19:36:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.210 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32119.0 tgs: 269 data_time: 0.59s time: 119.30s eta: 19:31:07 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.181 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31094.0 tgs: 260 data_time: 0.77s time: 119.54s eta: 19:31:32 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 259 data_time: 1.16s time: 122.96s eta: 20:02:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.239 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31574.0 tgs: 263 data_time: 0.73s time: 119.89s eta: 19:30:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.223 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31714.0 tgs: 262 data_time: 0.76s time: 120.66s eta: 19:36:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.208 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31891.0 tgs: 265 data_time: 0.80s time: 119.91s eta: 19:27:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31327.0 tgs: 260 data_time: 0.83s time: 120.46s eta: 19:30:28 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.261 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 0.82s time: 119.34s eta: 19:17:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31061.0 tgs: 259 data_time: 0.65s time: 119.68s eta: 19:18:52 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30891.0 tgs: 259 data_time: 0.72s time: 119.17s eta: 19:12:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31341.0 tgs: 262 data_time: 0.78s time: 119.20s eta: 19:10:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31438.0 tgs: 264 data_time: 0.86s time: 118.93s eta: 19:05:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 269 data_time: 0.64s time: 119.15s eta: 19:05:51 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 0.93s time: 119.18s eta: 19:04:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 262 data_time: 0.69s time: 119.39s eta: 19:04:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31896.0 tgs: 267 data_time: 0.76s time: 119.33s eta: 19:01:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 264 data_time: 0.64s time: 120.31s eta: 19:08:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.229 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 262 data_time: 0.74s time: 122.92s eta: 19:31:51 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31172.0 tgs: 260 data_time: 0.90s time: 119.48s eta: 18:57:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32324.0 tgs: 271 data_time: 0.92s time: 119.15s eta: 18:51:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 271 data_time: 0.87s time: 119.25s eta: 18:50:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31182.0 tgs: 261 data_time: 0.82s time: 119.39s eta: 18:50:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31901.0 tgs: 266 data_time: 0.68s time: 119.66s eta: 18:50:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30544.0 tgs: 255 data_time: 0.94s time: 119.38s eta: 18:46:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31830.0 tgs: 265 data_time: 0.77s time: 119.78s eta: 18:47:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31792.0 tgs: 245 data_time: 1.14s time: 129.31s eta: 20:15:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31968.0 tgs: 267 data_time: 0.79s time: 119.57s eta: 18:41:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31133.0 tgs: 260 data_time: 0.81s time: 119.39s eta: 18:38:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 267 data_time: 0.80s time: 120.66s eta: 18:48:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31570.0 tgs: 264 data_time: 0.91s time: 119.40s eta: 18:34:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32328.0 tgs: 271 data_time: 0.95s time: 119.16s eta: 18:30:07 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31564.0 tgs: 265 data_time: 0.93s time: 119.07s eta: 18:27:22 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31631.0 tgs: 265 data_time: 0.93s time: 119.14s eta: 18:25:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31761.0 tgs: 263 data_time: 0.72s time: 120.39s eta: 18:35:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31545.0 tgs: 260 data_time: 0.83s time: 121.17s eta: 18:40:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30209.0 tgs: 250 data_time: 0.73s time: 120.67s eta: 18:34:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32351.0 tgs: 270 data_time: 0.60s time: 119.74s eta: 18:23:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31789.0 tgs: 256 data_time: 0.72s time: 123.88s eta: 18:59:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32049.0 tgs: 268 data_time: 0.69s time: 119.56s eta: 18:17:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32186.0 tgs: 269 data_time: 0.76s time: 119.32s eta: 18:13:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31935.0 tgs: 267 data_time: 0.84s time: 119.23s eta: 18:10:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32419.0 tgs: 271 data_time: 0.74s time: 119.60s eta: 18:12:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32027.0 tgs: 268 data_time: 0.85s time: 119.27s eta: 18:07:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31765.0 tgs: 266 data_time: 0.55s time: 119.02s eta: 18:03:03 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31151.0 tgs: 261 data_time: 0.80s time: 119.10s eta: 18:01:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32470.0 tgs: 272 data_time: 0.82s time: 119.26s eta: 18:01:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31763.0 tgs: 266 data_time: 0.89s time: 119.33s eta: 17:59:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32319.0 tgs: 269 data_time: 0.89s time: 120.02s eta: 18:04:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30805.0 tgs: 258 data_time: 0.83s time: 119.11s eta: 17:53:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32262.0 tgs: 270 data_time: 0.73s time: 119.41s eta: 17:54:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32600.0 tgs: 273 data_time: 0.81s time: 119.01s eta: 17:49:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.77s time: 119.02s eta: 17:47:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32253.0 tgs: 257 data_time: 0.93s time: 125.37s eta: 18:42:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29875.0 tgs: 248 data_time: 0.77s time: 120.10s eta: 17:52:52 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31345.0 tgs: 262 data_time: 0.75s time: 119.36s eta: 17:44:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30572.0 tgs: 256 data_time: 0.93s time: 119.34s eta: 17:42:07 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31052.0 tgs: 259 data_time: 0.87s time: 119.51s eta: 17:41:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 249 data_time: 0.72s time: 126.05s eta: 18:37:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31868.0 tgs: 259 data_time: 0.81s time: 122.63s eta: 18:05:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31269.0 tgs: 258 data_time: 0.77s time: 121.16s eta: 17:50:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31590.0 tgs: 265 data_time: 0.69s time: 119.04s eta: 17:29:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32429.0 tgs: 272 data_time: 0.72s time: 118.85s eta: 17:25:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29122.0 tgs: 234 data_time: 0.87s time: 124.00s eta: 18:09:07 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31224.0 tgs: 250 data_time: 0.70s time: 124.57s eta: 18:12:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31914.0 tgs: 267 data_time: 1.07s time: 119.40s eta: 17:24:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31436.0 tgs: 262 data_time: 1.04s time: 119.70s eta: 17:25:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32074.0 tgs: 268 data_time: 0.65s time: 119.53s eta: 17:21:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32406.0 tgs: 272 data_time: 0.83s time: 118.86s eta: 17:14:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31471.0 tgs: 262 data_time: 0.76s time: 119.85s eta: 17:20:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32611.0 tgs: 272 data_time: 0.75s time: 119.53s eta: 17:15:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30756.0 tgs: 252 data_time: 0.82s time: 121.86s eta: 17:34:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31603.0 tgs: 255 data_time: 0.61s time: 123.51s eta: 17:46:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 266 data_time: 0.61s time: 119.01s eta: 17:05:30 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30407.0 tgs: 255 data_time: 0.81s time: 119.23s eta: 17:05:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31496.0 tgs: 264 data_time: 0.97s time: 119.15s eta: 17:02:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.268 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 267 data_time: 0.83s time: 119.25s eta: 17:01:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31305.0 tgs: 262 data_time: 0.87s time: 119.04s eta: 16:57:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31975.0 tgs: 268 data_time: 1.00s time: 119.27s eta: 16:57:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31974.0 tgs: 268 data_time: 0.68s time: 119.16s eta: 16:54:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31648.0 tgs: 264 data_time: 0.64s time: 119.58s eta: 16:56:28 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32343.0 tgs: 271 data_time: 0.99s time: 119.13s eta: 16:50:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.269 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31124.0 tgs: 261 data_time: 0.75s time: 118.85s eta: 16:46:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 269 data_time: 0.49s time: 119.02s eta: 16:45:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31889.0 tgs: 266 data_time: 0.65s time: 119.85s eta: 16:50:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.222 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32260.0 tgs: 270 data_time: 0.72s time: 119.37s eta: 16:44:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.122 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31721.0 tgs: 266 data_time: 0.49s time: 119.01s eta: 16:39:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32135.0 tgs: 265 data_time: 0.93s time: 120.94s eta: 16:53:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.249 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30808.0 tgs: 258 data_time: 0.99s time: 118.99s eta: 16:35:33 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31500.0 tgs: 264 data_time: 0.93s time: 119.29s eta: 16:36:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 265 data_time: 1.01s time: 119.23s eta: 16:33:35 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32046.0 tgs: 268 data_time: 0.69s time: 119.37s eta: 16:32:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32157.0 tgs: 268 data_time: 0.77s time: 119.70s eta: 16:33:32 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.221 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32241.0 tgs: 270 data_time: 1.02s time: 119.20s eta: 16:27:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.119 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32080.0 tgs: 267 data_time: 0.80s time: 119.76s eta: 16:30:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31745.0 tgs: 263 data_time: 0.81s time: 120.28s eta: 16:32:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.213 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31611.0 tgs: 265 data_time: 0.67s time: 118.98s eta: 16:19:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31344.0 tgs: 246 data_time: 0.65s time: 127.33s eta: 17:26:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31867.0 tgs: 267 data_time: 0.66s time: 119.22s eta: 16:17:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.288 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 268 data_time: 0.74s time: 119.15s eta: 16:15:00 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31208.0 tgs: 261 data_time: 0.59s time: 119.21s eta: 16:13:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.266 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32027.0 tgs: 268 data_time: 0.74s time: 119.17s eta: 16:11:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31555.0 tgs: 265 data_time: 0.88s time: 118.95s eta: 16:07:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30723.0 tgs: 258 data_time: 0.60s time: 118.86s eta: 16:04:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30348.0 tgs: 255 data_time: 0.68s time: 118.96s eta: 16:03:35 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31497.0 tgs: 258 data_time: 0.64s time: 121.84s eta: 16:24:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31608.0 tgs: 265 data_time: 0.65s time: 119.00s eta: 15:59:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30383.0 tgs: 254 data_time: 1.09s time: 119.23s eta: 15:59:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.234 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32454.0 tgs: 271 data_time: 0.58s time: 119.62s eta: 16:00:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.254 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31170.0 tgs: 261 data_time: 0.89s time: 119.05s eta: 15:54:24 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31468.0 tgs: 264 data_time: 0.81s time: 118.97s eta: 15:51:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 258 data_time: 0.62s time: 119.18s eta: 15:51:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32197.0 tgs: 269 data_time: 0.60s time: 119.64s eta: 15:53:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31922.0 tgs: 268 data_time: 0.78s time: 118.94s eta: 15:45:32 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32099.0 tgs: 269 data_time: 0.76s time: 119.29s eta: 15:46:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32089.0 tgs: 269 data_time: 0.67s time: 119.10s eta: 15:42:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30705.0 tgs: 258 data_time: 0.94s time: 119.01s eta: 15:40:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31201.0 tgs: 260 data_time: 0.67s time: 119.65s eta: 15:43:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30621.0 tgs: 252 data_time: 0.61s time: 121.50s eta: 15:55:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31948.0 tgs: 267 data_time: 0.54s time: 119.41s eta: 15:37:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31983.0 tgs: 246 data_time: 0.67s time: 129.96s eta: 16:58:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32096.0 tgs: 268 data_time: 0.67s time: 119.47s eta: 15:33:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32218.0 tgs: 269 data_time: 0.76s time: 119.60s eta: 15:32:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31551.0 tgs: 263 data_time: 0.58s time: 119.53s eta: 15:30:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31482.0 tgs: 257 data_time: 0.67s time: 122.14s eta: 15:48:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30656.0 tgs: 253 data_time: 0.81s time: 120.76s eta: 15:35:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32071.0 tgs: 268 data_time: 1.01s time: 119.41s eta: 15:23:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31441.0 tgs: 263 data_time: 0.87s time: 119.44s eta: 15:21:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.123 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31665.0 tgs: 240 data_time: 0.89s time: 131.64s eta: 16:53:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32148.0 tgs: 269 data_time: 0.71s time: 119.29s eta: 15:16:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 268 data_time: 0.55s time: 119.55s eta: 15:16:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31842.0 tgs: 260 data_time: 0.87s time: 122.25s eta: 15:35:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31745.0 tgs: 265 data_time: 0.94s time: 119.71s eta: 15:13:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.215 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30759.0 tgs: 257 data_time: 0.72s time: 119.35s eta: 15:09:03 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31030.0 tgs: 260 data_time: 0.98s time: 118.90s eta: 15:03:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31701.0 tgs: 266 data_time: 0.94s time: 119.00s eta: 15:02:24 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31700.0 tgs: 265 data_time: 0.81s time: 119.31s eta: 15:02:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31035.0 tgs: 260 data_time: 0.79s time: 118.92s eta: 14:57:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.264 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31192.0 tgs: 258 data_time: 0.77s time: 120.52s eta: 15:07:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31653.0 tgs: 265 data_time: 0.67s time: 119.33s eta: 14:56:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32533.0 tgs: 265 data_time: 0.69s time: 122.65s eta: 15:19:51 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 266 data_time: 0.64s time: 119.13s eta: 14:51:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30713.0 tgs: 256 data_time: 1.02s time: 119.75s eta: 14:54:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.214 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32082.0 tgs: 268 data_time: 0.68s time: 119.47s eta: 14:50:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.75s time: 118.99s eta: 14:44:30 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32563.0 tgs: 272 data_time: 0.72s time: 119.31s eta: 14:44:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32619.0 tgs: 274 data_time: 0.87s time: 118.93s eta: 14:40:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31114.0 tgs: 261 data_time: 0.80s time: 119.05s eta: 14:39:00 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.215 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 27050.0 tgs: 227 data_time: 0.40s time: 119.06s eta: 14:37:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31203.0 tgs: 261 data_time: 0.60s time: 119.13s eta: 14:35:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 270 data_time: 0.96s time: 119.01s eta: 14:32:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31684.0 tgs: 266 data_time: 0.85s time: 119.07s eta: 14:31:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31253.0 tgs: 262 data_time: 0.57s time: 119.15s eta: 14:29:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31033.0 tgs: 259 data_time: 0.65s time: 119.65s eta: 14:31:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.182 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32022.0 tgs: 268 data_time: 0.60s time: 119.18s eta: 14:26:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32325.0 tgs: 271 data_time: 0.82s time: 119.20s eta: 14:24:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.108 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31524.0 tgs: 264 data_time: 0.63s time: 119.39s eta: 14:23:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31265.0 tgs: 258 data_time: 0.91s time: 120.74s eta: 14:31:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31189.0 tgs: 262 data_time: 0.82s time: 118.85s eta: 14:15:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.204 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31533.0 tgs: 264 data_time: 0.84s time: 119.10s eta: 14:15:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31709.0 tgs: 266 data_time: 0.66s time: 119.13s eta: 14:13:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31560.0 tgs: 263 data_time: 0.75s time: 119.96s eta: 14:17:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.188 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31511.0 tgs: 262 data_time: 0.64s time: 119.87s eta: 14:15:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.129 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31916.0 tgs: 266 data_time: 0.58s time: 119.59s eta: 14:11:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31170.0 tgs: 261 data_time: 0.76s time: 119.22s eta: 14:06:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31944.0 tgs: 267 data_time: 0.68s time: 119.27s eta: 14:04:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.217 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32252.0 tgs: 270 data_time: 0.64s time: 119.27s eta: 14:02:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.114 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 269 data_time: 0.91s time: 119.14s eta: 13:59:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.120 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 270 data_time: 0.56s time: 119.12s eta: 13:57:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.78s time: 119.17s eta: 13:56:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31622.0 tgs: 264 data_time: 1.02s time: 119.35s eta: 13:55:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.208 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32179.0 tgs: 267 data_time: 1.28s time: 120.11s eta: 13:58:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.187 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32226.0 tgs: 262 data_time: 1.09s time: 122.56s eta: 14:13:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.131 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31889.0 tgs: 267 data_time: 0.89s time: 119.11s eta: 13:47:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31886.0 tgs: 267 data_time: 0.70s time: 119.37s eta: 13:47:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31640.0 tgs: 265 data_time: 0.67s time: 119.32s eta: 13:45:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 0.91s time: 119.33s eta: 13:43:22 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32138.0 tgs: 269 data_time: 0.87s time: 119.11s eta: 13:39:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31108.0 tgs: 260 data_time: 0.93s time: 119.26s eta: 13:38:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.122 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31461.0 tgs: 264 data_time: 0.65s time: 119.13s eta: 13:36:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 265 data_time: 0.83s time: 119.22s eta: 13:34:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.222 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32501.0 tgs: 272 data_time: 0.96s time: 119.46s eta: 13:34:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32122.0 tgs: 269 data_time: 0.72s time: 119.09s eta: 13:29:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31803.0 tgs: 267 data_time: 0.60s time: 118.92s eta: 13:26:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32254.0 tgs: 269 data_time: 0.67s time: 119.81s eta: 13:30:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32485.0 tgs: 272 data_time: 0.93s time: 119.26s eta: 13:24:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31945.0 tgs: 268 data_time: 1.09s time: 119.18s eta: 13:22:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31334.0 tgs: 262 data_time: 0.90s time: 119.33s eta: 13:21:30 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 32048.0 tgs: 268 data_time: 0.83s time: 119.26s eta: 13:19:03 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32046.0 tgs: 269 data_time: 0.68s time: 119.04s eta: 13:15:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 264 data_time: 0.91s time: 119.35s eta: 13:15:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.244 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30547.0 tgs: 256 data_time: 0.62s time: 119.18s eta: 13:12:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31787.0 tgs: 266 data_time: 0.88s time: 119.45s eta: 13:12:22 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 269 data_time: 1.05s time: 119.29s eta: 13:09:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31788.0 tgs: 266 data_time: 0.90s time: 119.15s eta: 13:06:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31874.0 tgs: 265 data_time: 0.93s time: 120.14s eta: 13:10:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.299 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32387.0 tgs: 269 data_time: 0.89s time: 120.16s eta: 13:09:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30335.0 tgs: 253 data_time: 0.49s time: 119.51s eta: 13:02:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30033.0 tgs: 251 data_time: 0.72s time: 119.51s eta: 13:00:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32025.0 tgs: 267 data_time: 0.73s time: 119.57s eta: 12:59:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31801.0 tgs: 265 data_time: 0.63s time: 119.80s eta: 12:58:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30774.0 tgs: 258 data_time: 0.78s time: 119.04s eta: 12:51:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.123 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31518.0 tgs: 264 data_time: 0.82s time: 119.05s eta: 12:49:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31386.0 tgs: 263 data_time: 0.62s time: 119.12s eta: 12:48:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.174 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.64s time: 119.40s eta: 12:48:07 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32166.0 tgs: 269 data_time: 0.78s time: 119.18s eta: 12:44:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32107.0 tgs: 268 data_time: 0.87s time: 119.44s eta: 12:44:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 0.97s time: 119.10s eta: 12:40:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31512.0 tgs: 264 data_time: 0.66s time: 119.07s eta: 12:38:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31533.0 tgs: 263 data_time: 0.84s time: 119.60s eta: 12:39:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31439.0 tgs: 264 data_time: 0.65s time: 118.78s eta: 12:32:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31130.0 tgs: 261 data_time: 0.62s time: 118.89s eta: 12:30:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32011.0 tgs: 268 data_time: 0.80s time: 119.17s eta: 12:30:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 25762.0 tgs: 216 data_time: 0.49s time: 119.27s eta: 12:29:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.171 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 270 data_time: 0.76s time: 119.35s eta: 12:27:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.213 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31986.0 tgs: 268 data_time: 0.83s time: 119.12s eta: 12:24:28 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.196 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 263 data_time: 0.96s time: 119.35s eta: 12:23:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32126.0 tgs: 269 data_time: 0.76s time: 119.26s eta: 12:21:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31207.0 tgs: 262 data_time: 0.91s time: 119.03s eta: 12:17:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.200 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 262 data_time: 0.57s time: 121.13s eta: 12:28:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.207 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31262.0 tgs: 260 data_time: 0.77s time: 119.99s eta: 12:19:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.201 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31381.0 tgs: 264 data_time: 0.80s time: 118.70s eta: 12:10:00 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.217 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31519.0 tgs: 263 data_time: 0.78s time: 119.64s eta: 12:13:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 268 data_time: 0.79s time: 119.83s eta: 12:12:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31516.0 tgs: 264 data_time: 0.77s time: 119.14s eta: 12:06:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31152.0 tgs: 261 data_time: 0.74s time: 119.17s eta: 12:04:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.88s time: 119.12s eta: 12:02:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.198 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31252.0 tgs: 262 data_time: 0.88s time: 119.19s eta: 12:01:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.205 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30537.0 tgs: 256 data_time: 0.95s time: 118.96s eta: 11:57:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.128 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30712.0 tgs: 256 data_time: 0.52s time: 119.70s eta: 12:00:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31275.0 tgs: 262 data_time: 0.44s time: 119.28s eta: 11:55:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.179 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32041.0 tgs: 268 data_time: 0.70s time: 119.27s eta: 11:53:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.217 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31777.0 tgs: 265 data_time: 0.77s time: 119.71s eta: 11:54:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32515.0 tgs: 273 data_time: 0.66s time: 118.80s eta: 11:46:51 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.73s time: 119.14s eta: 11:46:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31652.0 tgs: 264 data_time: 0.84s time: 119.61s eta: 11:47:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31256.0 tgs: 263 data_time: 0.82s time: 118.83s eta: 11:41:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.180 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32109.0 tgs: 270 data_time: 0.90s time: 118.68s eta: 11:38:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.52s time: 119.22s eta: 11:39:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31108.0 tgs: 261 data_time: 0.89s time: 119.14s eta: 11:36:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.138 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 268 data_time: 0.84s time: 119.07s eta: 11:34:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.121 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 263 data_time: 0.64s time: 118.95s eta: 11:31:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31063.0 tgs: 261 data_time: 0.79s time: 118.91s eta: 11:29:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31995.0 tgs: 268 data_time: 0.70s time: 118.97s eta: 11:28:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.206 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32662.0 tgs: 273 data_time: 0.75s time: 119.28s eta: 11:27:52 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.309 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31996.0 tgs: 268 data_time: 0.57s time: 119.36s eta: 11:26:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31004.0 tgs: 260 data_time: 0.71s time: 118.88s eta: 11:21:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.126 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32430.0 tgs: 272 data_time: 0.69s time: 119.21s eta: 11:21:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.168 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30334.0 tgs: 253 data_time: 0.63s time: 119.72s eta: 11:22:22 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31948.0 tgs: 268 data_time: 0.86s time: 119.12s eta: 11:16:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31792.0 tgs: 266 data_time: 0.52s time: 119.13s eta: 11:15:03 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31595.0 tgs: 265 data_time: 0.89s time: 118.95s eta: 11:12:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30196.0 tgs: 253 data_time: 0.72s time: 119.10s eta: 11:10:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31338.0 tgs: 263 data_time: 0.57s time: 119.15s eta: 11:09:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31464.0 tgs: 264 data_time: 0.86s time: 118.97s eta: 11:06:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30987.0 tgs: 259 data_time: 0.71s time: 119.47s eta: 11:07:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31777.0 tgs: 265 data_time: 0.81s time: 119.64s eta: 11:06:00 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31707.0 tgs: 265 data_time: 0.72s time: 119.31s eta: 11:02:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.110 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30829.0 tgs: 258 data_time: 0.82s time: 119.31s eta: 11:00:10 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31273.0 tgs: 263 data_time: 0.81s time: 118.82s eta: 10:55:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.107 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32308.0 tgs: 270 data_time: 0.63s time: 119.48s eta: 10:57:07 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31749.0 tgs: 265 data_time: 0.54s time: 119.36s eta: 10:54:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31767.0 tgs: 266 data_time: 0.67s time: 119.28s eta: 10:52:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 267 data_time: 0.70s time: 119.26s eta: 10:49:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31808.0 tgs: 267 data_time: 0.80s time: 119.10s eta: 10:47:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31298.0 tgs: 261 data_time: 0.69s time: 119.53s eta: 10:47:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31337.0 tgs: 262 data_time: 0.68s time: 119.31s eta: 10:44:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.197 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31711.0 tgs: 265 data_time: 0.75s time: 119.49s eta: 10:43:15 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.199 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31513.0 tgs: 263 data_time: 0.76s time: 119.51s eta: 10:41:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31466.0 tgs: 264 data_time: 0.88s time: 118.99s eta: 10:36:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31991.0 tgs: 268 data_time: 1.12s time: 119.22s eta: 10:35:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31635.0 tgs: 265 data_time: 0.68s time: 119.06s eta: 10:33:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31474.0 tgs: 264 data_time: 0.57s time: 118.78s eta: 10:29:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31774.0 tgs: 265 data_time: 0.70s time: 119.86s eta: 10:33:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.228 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31488.0 tgs: 263 data_time: 0.91s time: 119.38s eta: 10:28:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32535.0 tgs: 273 data_time: 0.74s time: 118.84s eta: 10:23:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 267 data_time: 0.85s time: 118.99s eta: 10:22:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31715.0 tgs: 265 data_time: 0.73s time: 119.28s eta: 10:22:15 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31553.0 tgs: 264 data_time: 0.60s time: 119.21s eta: 10:19:52 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 270 data_time: 0.72s time: 119.06s eta: 10:17:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 266 data_time: 0.71s time: 119.48s eta: 10:17:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.194 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31907.0 tgs: 267 data_time: 0.58s time: 119.37s eta: 10:14:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.226 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32006.0 tgs: 267 data_time: 0.39s time: 119.45s eta: 10:13:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29892.0 tgs: 250 data_time: 0.68s time: 119.29s eta: 10:10:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30783.0 tgs: 258 data_time: 0.54s time: 119.11s eta: 10:07:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.177 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32288.0 tgs: 269 data_time: 0.72s time: 119.70s eta: 10:08:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 265 data_time: 0.86s time: 119.83s eta: 10:07:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32161.0 tgs: 264 data_time: 0.98s time: 121.55s eta: 10:13:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.235 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.83s time: 119.25s eta: 10:00:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31992.0 tgs: 269 data_time: 0.96s time: 118.81s eta: 9:56:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.174 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32072.0 tgs: 268 data_time: 0.63s time: 119.56s eta: 9:57:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31033.0 tgs: 259 data_time: 0.93s time: 119.70s eta: 9:56:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31291.0 tgs: 262 data_time: 0.91s time: 119.26s eta: 9:52:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.175 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 267 data_time: 0.97s time: 119.33s eta: 9:50:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30506.0 tgs: 255 data_time: 0.72s time: 119.23s eta: 9:48:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.183 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31095.0 tgs: 257 data_time: 0.86s time: 120.65s eta: 9:53:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32378.0 tgs: 271 data_time: 0.56s time: 119.27s eta: 9:44:24 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.189 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31737.0 tgs: 265 data_time: 0.80s time: 119.46s eta: 9:43:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.125 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31010.0 tgs: 259 data_time: 0.81s time: 119.54s eta: 9:41:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31878.0 tgs: 267 data_time: 0.47s time: 119.21s eta: 9:38:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31652.0 tgs: 264 data_time: 0.70s time: 119.48s eta: 9:37:28 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 265 data_time: 0.91s time: 119.05s eta: 9:33:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.112 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 262 data_time: 0.65s time: 119.24s eta: 9:32:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31283.0 tgs: 262 data_time: 0.78s time: 119.16s eta: 9:29:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.206 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32064.0 tgs: 268 data_time: 0.97s time: 119.22s eta: 9:28:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31438.0 tgs: 263 data_time: 0.81s time: 119.46s eta: 9:27:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31072.0 tgs: 261 data_time: 0.72s time: 118.99s eta: 9:23:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31779.0 tgs: 266 data_time: 0.69s time: 119.03s eta: 9:21:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31632.0 tgs: 264 data_time: 0.98s time: 119.54s eta: 9:21:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32297.0 tgs: 270 data_time: 0.53s time: 119.59s eta: 9:20:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31681.0 tgs: 265 data_time: 0.64s time: 119.12s eta: 9:15:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32415.0 tgs: 272 data_time: 0.66s time: 119.00s eta: 9:13:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.251 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31327.0 tgs: 263 data_time: 0.67s time: 118.88s eta: 9:10:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.94s time: 119.33s eta: 9:10:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.217 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 0.98s time: 119.09s eta: 9:07:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31427.0 tgs: 264 data_time: 0.76s time: 118.86s eta: 9:04:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 264 data_time: 0.66s time: 119.67s eta: 9:06:28 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31164.0 tgs: 260 data_time: 0.94s time: 119.44s eta: 9:03:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31258.0 tgs: 261 data_time: 0.71s time: 119.44s eta: 9:01:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 265 data_time: 0.73s time: 118.74s eta: 8:56:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31022.0 tgs: 260 data_time: 0.86s time: 118.87s eta: 8:54:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31953.0 tgs: 267 data_time: 0.87s time: 119.33s eta: 8:55:00 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30599.0 tgs: 256 data_time: 0.77s time: 119.15s eta: 8:52:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.122 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31640.0 tgs: 265 data_time: 0.51s time: 119.27s eta: 8:50:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31856.0 tgs: 264 data_time: 0.82s time: 120.47s eta: 8:54:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.180 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 267 data_time: 0.76s time: 119.24s eta: 8:46:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 266 data_time: 0.90s time: 119.08s eta: 8:43:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 263 data_time: 0.99s time: 120.83s eta: 8:49:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31839.0 tgs: 264 data_time: 0.83s time: 120.20s eta: 8:44:52 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31794.0 tgs: 267 data_time: 0.46s time: 118.83s eta: 8:36:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32118.0 tgs: 269 data_time: 0.79s time: 119.35s eta: 8:37:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.70s time: 119.33s eta: 8:35:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31845.0 tgs: 267 data_time: 0.72s time: 119.06s eta: 8:31:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.211 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 269 data_time: 0.77s time: 119.16s eta: 8:30:24 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.73s time: 119.05s eta: 8:27:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32051.0 tgs: 269 data_time: 0.69s time: 119.12s eta: 8:26:15 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 268 data_time: 0.73s time: 119.24s eta: 8:24:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32153.0 tgs: 269 data_time: 0.75s time: 119.23s eta: 8:22:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31756.0 tgs: 266 data_time: 0.76s time: 119.38s eta: 8:21:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.204 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31863.0 tgs: 266 data_time: 0.82s time: 119.78s eta: 8:21:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31688.0 tgs: 265 data_time: 0.84s time: 119.20s eta: 8:16:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31434.0 tgs: 264 data_time: 0.61s time: 118.98s eta: 8:13:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31934.0 tgs: 266 data_time: 0.66s time: 119.90s eta: 8:15:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 268 data_time: 0.60s time: 119.55s eta: 8:12:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32172.0 tgs: 269 data_time: 0.82s time: 119.60s eta: 8:10:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31739.0 tgs: 266 data_time: 0.86s time: 119.00s eta: 8:05:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31506.0 tgs: 262 data_time: 0.78s time: 119.84s eta: 8:07:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 266 data_time: 0.74s time: 119.71s eta: 8:04:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31864.0 tgs: 266 data_time: 0.91s time: 119.40s eta: 8:01:35 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32174.0 tgs: 269 data_time: 0.76s time: 119.29s eta: 7:59:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31700.0 tgs: 265 data_time: 0.61s time: 119.19s eta: 7:56:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31403.0 tgs: 263 data_time: 0.61s time: 119.20s eta: 7:54:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.128 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31207.0 tgs: 262 data_time: 0.59s time: 118.94s eta: 7:51:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.162 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31558.0 tgs: 265 data_time: 0.99s time: 118.97s eta: 7:49:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.192 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 270 data_time: 0.66s time: 119.42s eta: 7:49:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32143.0 tgs: 270 data_time: 0.59s time: 119.04s eta: 7:46:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.178 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32332.0 tgs: 271 data_time: 0.90s time: 119.07s eta: 7:44:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.191 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 268 data_time: 0.78s time: 120.01s eta: 7:46:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31324.0 tgs: 261 data_time: 0.66s time: 119.83s eta: 7:43:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32501.0 tgs: 272 data_time: 0.61s time: 119.16s eta: 7:38:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32108.0 tgs: 269 data_time: 0.80s time: 118.97s eta: 7:36:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31892.0 tgs: 267 data_time: 0.78s time: 119.32s eta: 7:35:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31726.0 tgs: 265 data_time: 0.80s time: 119.38s eta: 7:33:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31653.0 tgs: 265 data_time: 0.56s time: 119.06s eta: 7:30:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31597.0 tgs: 265 data_time: 0.73s time: 118.99s eta: 7:28:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30377.0 tgs: 254 data_time: 0.55s time: 119.24s eta: 7:27:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.202 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31018.0 tgs: 260 data_time: 0.43s time: 118.99s eta: 7:24:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30388.0 tgs: 255 data_time: 0.64s time: 118.94s eta: 7:22:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.179 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 262 data_time: 0.69s time: 119.08s eta: 7:20:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31932.0 tgs: 268 data_time: 0.80s time: 118.99s eta: 7:18:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31833.0 tgs: 267 data_time: 0.78s time: 119.03s eta: 7:16:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31649.0 tgs: 265 data_time: 0.67s time: 119.25s eta: 7:15:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.230 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31558.0 tgs: 264 data_time: 0.69s time: 119.36s eta: 7:13:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31390.0 tgs: 263 data_time: 0.62s time: 118.94s eta: 7:10:10 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.50s time: 119.20s eta: 7:09:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.103 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31845.0 tgs: 266 data_time: 0.53s time: 119.59s eta: 7:08:30 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.112 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29935.0 tgs: 250 data_time: 0.73s time: 119.66s eta: 7:06:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31746.0 tgs: 266 data_time: 0.66s time: 118.95s eta: 7:02:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31721.0 tgs: 265 data_time: 0.63s time: 119.30s eta: 7:01:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31320.0 tgs: 262 data_time: 0.94s time: 119.52s eta: 7:00:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.207 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32203.0 tgs: 269 data_time: 0.62s time: 119.42s eta: 6:57:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 272 data_time: 0.66s time: 118.91s eta: 6:54:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 270 data_time: 0.91s time: 119.34s eta: 6:53:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 263 data_time: 0.74s time: 118.94s eta: 6:50:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32205.0 tgs: 270 data_time: 0.69s time: 119.27s eta: 6:49:28 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30968.0 tgs: 260 data_time: 0.78s time: 119.02s eta: 6:46:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31854.0 tgs: 267 data_time: 0.56s time: 119.22s eta: 6:45:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32232.0 tgs: 269 data_time: 0.79s time: 119.38s eta: 6:43:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.118 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32035.0 tgs: 269 data_time: 0.55s time: 119.01s eta: 6:40:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32042.0 tgs: 269 data_time: 0.93s time: 119.01s eta: 6:38:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 264 data_time: 0.69s time: 119.46s eta: 6:38:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31318.0 tgs: 263 data_time: 0.80s time: 118.80s eta: 6:34:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.205 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31780.0 tgs: 266 data_time: 0.67s time: 119.44s eta: 6:34:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.193 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31646.0 tgs: 264 data_time: 0.69s time: 119.57s eta: 6:32:35 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.188 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31349.0 tgs: 262 data_time: 0.68s time: 119.38s eta: 6:29:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.198 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30824.0 tgs: 257 data_time: 0.50s time: 119.60s eta: 6:28:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32354.0 tgs: 271 data_time: 0.61s time: 119.36s eta: 6:25:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31166.0 tgs: 262 data_time: 0.68s time: 118.82s eta: 6:22:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31988.0 tgs: 268 data_time: 0.92s time: 119.02s eta: 6:20:51 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31964.0 tgs: 267 data_time: 0.94s time: 119.40s eta: 6:20:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32179.0 tgs: 269 data_time: 0.93s time: 119.34s eta: 6:17:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.210 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31511.0 tgs: 264 data_time: 0.87s time: 119.05s eta: 6:14:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31463.0 tgs: 263 data_time: 0.69s time: 119.39s eta: 6:14:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.105 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31773.0 tgs: 266 data_time: 0.85s time: 119.39s eta: 6:12:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31893.0 tgs: 267 data_time: 0.70s time: 119.29s eta: 6:09:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31775.0 tgs: 266 data_time: 0.79s time: 119.10s eta: 6:07:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31700.0 tgs: 259 data_time: 0.77s time: 121.96s eta: 6:14:00 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31492.0 tgs: 263 data_time: 0.72s time: 119.39s eta: 6:04:07 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31332.0 tgs: 262 data_time: 0.68s time: 119.57s eta: 6:02:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31086.0 tgs: 260 data_time: 0.78s time: 119.25s eta: 5:59:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 27321.0 tgs: 229 data_time: 0.63s time: 119.27s eta: 5:57:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31862.0 tgs: 267 data_time: 0.92s time: 118.91s eta: 5:54:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30862.0 tgs: 259 data_time: 0.91s time: 119.02s eta: 5:53:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31457.0 tgs: 264 data_time: 0.82s time: 119.07s eta: 5:51:15 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30236.0 tgs: 253 data_time: 0.86s time: 119.29s eta: 5:49:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.116 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31960.0 tgs: 268 data_time: 0.77s time: 118.97s eta: 5:46:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32273.0 tgs: 270 data_time: 0.70s time: 119.27s eta: 5:45:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32334.0 tgs: 270 data_time: 0.80s time: 119.33s eta: 5:44:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32077.0 tgs: 269 data_time: 0.69s time: 119.06s eta: 5:41:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.120 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31705.0 tgs: 265 data_time: 0.59s time: 119.50s eta: 5:40:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.179 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31560.0 tgs: 264 data_time: 0.79s time: 119.38s eta: 5:38:14 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30543.0 tgs: 255 data_time: 0.87s time: 119.57s eta: 5:36:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31756.0 tgs: 266 data_time: 0.73s time: 118.98s eta: 5:33:08 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31548.0 tgs: 265 data_time: 0.83s time: 118.95s eta: 5:31:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.179 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31506.0 tgs: 264 data_time: 0.59s time: 119.24s eta: 5:29:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.095 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30745.0 tgs: 257 data_time: 0.95s time: 119.40s eta: 5:28:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31753.0 tgs: 265 data_time: 0.88s time: 119.46s eta: 5:26:31 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31738.0 tgs: 266 data_time: 0.66s time: 118.95s eta: 5:23:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31359.0 tgs: 262 data_time: 0.56s time: 119.63s eta: 5:22:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31478.0 tgs: 263 data_time: 0.71s time: 119.35s eta: 5:20:15 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31081.0 tgs: 260 data_time: 0.67s time: 119.14s eta: 5:17:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31347.0 tgs: 262 data_time: 0.51s time: 119.41s eta: 5:16:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.101 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31565.0 tgs: 263 data_time: 0.86s time: 119.58s eta: 5:14:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 267 data_time: 0.81s time: 119.00s eta: 5:11:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31614.0 tgs: 265 data_time: 0.69s time: 119.00s eta: 5:09:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.227 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31608.0 tgs: 265 data_time: 0.88s time: 119.23s eta: 5:08:00 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.189 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 268 data_time: 0.75s time: 119.69s eta: 5:07:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31961.0 tgs: 268 data_time: 0.69s time: 119.11s eta: 5:03:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32214.0 tgs: 270 data_time: 0.87s time: 119.07s eta: 5:01:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32104.0 tgs: 264 data_time: 0.87s time: 121.52s eta: 5:05:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 266 data_time: 0.70s time: 119.57s eta: 4:58:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30327.0 tgs: 253 data_time: 0.47s time: 119.50s eta: 4:56:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.223 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32344.0 tgs: 270 data_time: 1.04s time: 119.73s eta: 4:55:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31880.0 tgs: 266 data_time: 0.51s time: 119.52s eta: 4:52:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.213 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30378.0 tgs: 254 data_time: 0.82s time: 119.42s eta: 4:50:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31216.0 tgs: 261 data_time: 0.73s time: 119.20s eta: 4:48:04 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 265 data_time: 0.64s time: 118.89s eta: 4:45:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 265 data_time: 0.83s time: 119.05s eta: 4:43:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 28332.0 tgs: 236 data_time: 0.61s time: 119.98s eta: 4:43:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.098 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 270 data_time: 0.66s time: 119.15s eta: 4:39:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31138.0 tgs: 261 data_time: 0.70s time: 119.01s eta: 4:37:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31315.0 tgs: 263 data_time: 0.83s time: 119.06s eta: 4:35:48 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30717.0 tgs: 257 data_time: 0.83s time: 119.37s eta: 4:34:33 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31628.0 tgs: 264 data_time: 0.58s time: 119.41s eta: 4:32:39 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30401.0 tgs: 255 data_time: 0.56s time: 118.94s eta: 4:29:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32144.0 tgs: 268 data_time: 0.73s time: 119.60s eta: 4:29:05 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31502.0 tgs: 264 data_time: 0.89s time: 119.24s eta: 4:26:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.137 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.68s time: 119.07s eta: 4:23:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 269 data_time: 0.60s time: 118.85s eta: 4:21:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30956.0 tgs: 260 data_time: 0.68s time: 118.76s eta: 4:19:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.240 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29499.0 tgs: 247 data_time: 0.64s time: 119.43s eta: 4:18:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 266 data_time: 0.61s time: 119.41s eta: 4:16:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 28527.0 tgs: 239 data_time: 0.56s time: 119.33s eta: 4:14:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 261 data_time: 0.72s time: 119.05s eta: 4:11:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31333.0 tgs: 263 data_time: 0.78s time: 119.12s eta: 4:10:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31398.0 tgs: 263 data_time: 0.59s time: 119.14s eta: 4:08:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.108 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30799.0 tgs: 258 data_time: 0.63s time: 118.95s eta: 4:05:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31257.0 tgs: 262 data_time: 0.66s time: 119.16s eta: 4:04:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32315.0 tgs: 270 data_time: 0.73s time: 119.50s eta: 4:02:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32550.0 tgs: 272 data_time: 0.58s time: 119.25s eta: 4:00:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31342.0 tgs: 262 data_time: 0.69s time: 119.19s eta: 3:58:22 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31153.0 tgs: 260 data_time: 0.60s time: 119.36s eta: 3:56:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31063.0 tgs: 260 data_time: 0.71s time: 119.37s eta: 3:54:45 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 265 data_time: 0.81s time: 119.95s eta: 3:53:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 271 data_time: 0.67s time: 119.32s eta: 3:50:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30971.0 tgs: 258 data_time: 0.89s time: 119.63s eta: 3:49:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31112.0 tgs: 261 data_time: 0.66s time: 118.86s eta: 3:45:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 267 data_time: 0.96s time: 119.71s eta: 3:45:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 269 data_time: 0.79s time: 119.33s eta: 3:42:44 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31498.0 tgs: 264 data_time: 0.75s time: 119.26s eta: 3:40:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30580.0 tgs: 256 data_time: 0.88s time: 119.01s eta: 3:38:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31727.0 tgs: 265 data_time: 1.19s time: 119.42s eta: 3:36:57 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31384.0 tgs: 263 data_time: 0.83s time: 119.03s eta: 3:34:15 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31768.0 tgs: 266 data_time: 0.66s time: 118.99s eta: 3:32:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.70s time: 119.22s eta: 3:30:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30898.0 tgs: 257 data_time: 0.72s time: 120.11s eta: 3:30:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 268 data_time: 0.85s time: 119.13s eta: 3:26:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31751.0 tgs: 266 data_time: 0.79s time: 119.32s eta: 3:24:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.194 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 270 data_time: 0.47s time: 119.08s eta: 3:22:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32224.0 tgs: 269 data_time: 0.85s time: 119.43s eta: 3:21:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30820.0 tgs: 258 data_time: 0.75s time: 119.29s eta: 3:18:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31507.0 tgs: 264 data_time: 0.75s time: 119.22s eta: 3:16:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.191 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.81s time: 119.13s eta: 3:14:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30908.0 tgs: 259 data_time: 0.73s time: 119.02s eta: 3:12:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32441.0 tgs: 272 data_time: 0.60s time: 119.19s eta: 3:10:41 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30654.0 tgs: 253 data_time: 0.87s time: 121.03s eta: 3:11:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32062.0 tgs: 268 data_time: 0.67s time: 119.50s eta: 3:07:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32494.0 tgs: 272 data_time: 0.90s time: 119.26s eta: 3:04:51 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30956.0 tgs: 260 data_time: 0.70s time: 118.99s eta: 3:02:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31562.0 tgs: 264 data_time: 0.74s time: 119.23s eta: 3:00:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31709.0 tgs: 265 data_time: 0.76s time: 119.45s eta: 2:59:10 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.204 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 262 data_time: 0.68s time: 118.98s eta: 2:56:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32415.0 tgs: 271 data_time: 1.05s time: 119.35s eta: 2:55:02 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29085.0 tgs: 243 data_time: 0.77s time: 119.42s eta: 2:53:09 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31686.0 tgs: 265 data_time: 0.77s time: 119.35s eta: 2:51:03 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31625.0 tgs: 260 data_time: 0.66s time: 121.46s eta: 2:52:03 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31391.0 tgs: 263 data_time: 0.53s time: 119.24s eta: 2:46:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.172 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31410.0 tgs: 262 data_time: 0.95s time: 119.56s eta: 2:45:23 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32137.0 tgs: 270 data_time: 0.67s time: 118.91s eta: 2:42:30 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 0.73s time: 119.18s eta: 2:40:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 267 data_time: 0.51s time: 118.97s eta: 2:38:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.223 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32061.0 tgs: 268 data_time: 0.76s time: 119.23s eta: 2:36:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31916.0 tgs: 267 data_time: 0.68s time: 119.11s eta: 2:34:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 266 data_time: 0.81s time: 119.43s eta: 2:33:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.176 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31433.0 tgs: 261 data_time: 1.11s time: 120.23s eta: 2:32:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.106 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31575.0 tgs: 264 data_time: 0.90s time: 119.28s eta: 2:29:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32083.0 tgs: 269 data_time: 0.80s time: 118.88s eta: 2:26:37 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31396.0 tgs: 262 data_time: 0.66s time: 119.52s eta: 2:25:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31508.0 tgs: 264 data_time: 0.52s time: 119.06s eta: 2:22:52 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31513.0 tgs: 264 data_time: 0.57s time: 119.14s eta: 2:20:58 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31506.0 tgs: 264 data_time: 0.75s time: 119.05s eta: 2:18:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32410.0 tgs: 270 data_time: 0.66s time: 119.84s eta: 2:17:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31981.0 tgs: 268 data_time: 0.98s time: 119.32s eta: 2:15:13 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31906.0 tgs: 267 data_time: 0.61s time: 119.38s eta: 2:13:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.112 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32067.0 tgs: 267 data_time: 0.82s time: 119.68s eta: 2:11:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.201 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31768.0 tgs: 265 data_time: 0.86s time: 119.49s eta: 2:09:27 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 270 data_time: 0.65s time: 118.96s eta: 2:06:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31616.0 tgs: 264 data_time: 0.58s time: 119.38s eta: 2:05:21 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30679.0 tgs: 257 data_time: 0.70s time: 119.28s eta: 2:03:15 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30337.0 tgs: 254 data_time: 0.58s time: 119.22s eta: 2:01:12 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32212.0 tgs: 269 data_time: 0.75s time: 119.50s eta: 1:59:30 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.66s time: 119.42s eta: 1:57:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32505.0 tgs: 271 data_time: 0.80s time: 119.54s eta: 1:55:33 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31980.0 tgs: 268 data_time: 0.59s time: 119.30s eta: 1:53:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31335.0 tgs: 262 data_time: 0.85s time: 119.25s eta: 1:51:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32336.0 tgs: 271 data_time: 0.74s time: 118.94s eta: 1:49:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31302.0 tgs: 261 data_time: 0.54s time: 119.64s eta: 1:47:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31557.0 tgs: 263 data_time: 0.78s time: 119.69s eta: 1:45:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31294.0 tgs: 262 data_time: 0.72s time: 119.35s eta: 1:43:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.103 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31227.0 tgs: 261 data_time: 0.61s time: 119.31s eta: 1:41:24 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31983.0 tgs: 268 data_time: 0.98s time: 119.28s eta: 1:39:24 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32110.0 tgs: 270 data_time: 0.74s time: 118.90s eta: 1:37:06 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 263 data_time: 0.81s time: 119.16s eta: 1:35:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32043.0 tgs: 269 data_time: 0.98s time: 119.07s eta: 1:33:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32253.0 tgs: 270 data_time: 0.68s time: 119.25s eta: 1:31:25 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30673.0 tgs: 256 data_time: 0.81s time: 119.73s eta: 1:29:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31562.0 tgs: 265 data_time: 0.70s time: 119.02s eta: 1:27:16 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 263 data_time: 0.72s time: 118.87s eta: 1:25:11 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32139.0 tgs: 268 data_time: 0.68s time: 119.48s eta: 1:23:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30337.0 tgs: 253 data_time: 0.73s time: 119.48s eta: 1:21:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32527.0 tgs: 273 data_time: 0.80s time: 118.94s eta: 1:19:17 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31543.0 tgs: 265 data_time: 0.63s time: 118.93s eta: 1:17:18 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32142.0 tgs: 263 data_time: 0.58s time: 122.13s eta: 1:17:20 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32360.0 tgs: 271 data_time: 0.54s time: 119.36s eta: 1:13:36 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31728.0 tgs: 266 data_time: 0.71s time: 119.06s eta: 1:11:26 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31053.0 tgs: 261 data_time: 0.67s time: 118.92s eta: 1:09:22 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31778.0 tgs: 266 data_time: 0.83s time: 119.10s eta: 1:07:29 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31475.0 tgs: 263 data_time: 0.67s time: 119.42s eta: 1:05:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29530.0 tgs: 247 data_time: 0.63s time: 119.34s eta: 1:03:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30940.0 tgs: 259 data_time: 0.61s time: 119.18s eta: 1:01:34 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31634.0 tgs: 265 data_time: 0.60s time: 119.27s eta: 0:59:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 269 data_time: 0.64s time: 119.44s eta: 0:57:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30613.0 tgs: 256 data_time: 0.81s time: 119.39s eta: 0:55:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31940.0 tgs: 267 data_time: 0.74s time: 119.49s eta: 0:53:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 262 data_time: 0.61s time: 119.36s eta: 0:51:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31155.0 tgs: 260 data_time: 0.56s time: 119.47s eta: 0:49:46 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31618.0 tgs: 265 data_time: 0.79s time: 118.90s eta: 0:47:33 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32178.0 tgs: 269 data_time: 0.66s time: 119.26s eta: 0:45:43 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 268 data_time: 0.63s time: 119.20s eta: 0:43:42 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32354.0 tgs: 271 data_time: 0.83s time: 119.00s eta: 0:41:38 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32505.0 tgs: 273 data_time: 0.76s time: 119.04s eta: 0:39:40 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.193 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32119.0 tgs: 268 data_time: 0.65s time: 119.66s eta: 0:37:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32210.0 tgs: 270 data_time: 0.56s time: 119.28s eta: 0:35:47 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32281.0 tgs: 269 data_time: 0.72s time: 119.75s eta: 0:33:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30657.0 tgs: 252 data_time: 0.64s time: 121.23s eta: 0:32:19 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31457.0 tgs: 263 data_time: 0.80s time: 119.38s eta: 0:29:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.106 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32247.0 tgs: 270 data_time: 0.75s time: 119.32s eta: 0:27:50 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31402.0 tgs: 262 data_time: 0.53s time: 119.65s eta: 0:25:55 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.102 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30715.0 tgs: 257 data_time: 0.57s time: 119.09s eta: 0:23:49 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31496.0 tgs: 263 data_time: 0.98s time: 119.38s eta: 0:21:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 264 data_time: 0.65s time: 119.26s eta: 0:19:52 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31642.0 tgs: 265 data_time: 0.70s time: 119.24s eta: 0:17:53 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 270 data_time: 0.62s time: 119.90s eta: 0:15:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.207 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31768.0 tgs: 266 data_time: 0.81s time: 119.19s eta: 0:13:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.210 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32268.0 tgs: 270 data_time: 0.88s time: 119.09s eta: 0:11:54 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32354.0 tgs: 271 data_time: 0.81s time: 119.32s eta: 0:09:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32019.0 tgs: 268 data_time: 0.69s time: 119.17s eta: 0:07:56 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31728.0 tgs: 265 data_time: 0.70s time: 119.69s eta: 0:05:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30597.0 tgs: 253 data_time: 1.01s time: 120.64s eta: 0:04:01 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32060.0 tgs: 268 data_time: 0.64s time: 119.20s eta: 0:01:59 +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.30 seconds, peak gpu memory 17.4G +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 47][DP 11][SP 3][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank52.log b/20250121165312/rank52.log new file mode 100644 index 0000000000000000000000000000000000000000..3309d9f8a8c248e3a4b1246c9a6bcf91ea47aaa4 --- /dev/null +++ b/20250121165312/rank52.log @@ -0,0 +1,851 @@ +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.14s +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:00:47][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 67.13 seconds, peak gpu memory 13.4G +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.250 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31804.0 tgs: 210 data_time: 1.88s time: 150.96s eta: 1 day, 0:51:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.256 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31938.0 tgs: 265 data_time: 0.66s time: 120.17s eta: 19:45:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.332 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31420.0 tgs: 262 data_time: 0.72s time: 119.68s eta: 19:38:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32320.0 tgs: 270 data_time: 0.93s time: 119.65s eta: 19:36:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.235 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31755.0 tgs: 266 data_time: 0.80s time: 119.30s eta: 19:31:10 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.220 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31116.0 tgs: 260 data_time: 0.96s time: 119.54s eta: 19:31:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32297.0 tgs: 262 data_time: 0.89s time: 122.96s eta: 20:02:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32099.0 tgs: 267 data_time: 0.85s time: 119.89s eta: 19:30:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.202 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31943.0 tgs: 264 data_time: 0.90s time: 120.66s eta: 19:36:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32321.0 tgs: 269 data_time: 0.81s time: 119.91s eta: 19:27:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.235 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31685.0 tgs: 263 data_time: 1.11s time: 120.46s eta: 19:30:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.242 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31570.0 tgs: 264 data_time: 1.03s time: 119.34s eta: 19:17:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.266 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32271.0 tgs: 269 data_time: 0.97s time: 119.68s eta: 19:18:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31826.0 tgs: 267 data_time: 0.79s time: 119.18s eta: 19:12:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31152.0 tgs: 261 data_time: 0.69s time: 119.22s eta: 19:10:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31903.0 tgs: 268 data_time: 0.81s time: 118.92s eta: 19:05:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30941.0 tgs: 259 data_time: 0.63s time: 119.14s eta: 19:05:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 29085.0 tgs: 244 data_time: 0.64s time: 119.18s eta: 19:04:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32477.0 tgs: 272 data_time: 1.03s time: 119.40s eta: 19:04:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31064.0 tgs: 260 data_time: 0.67s time: 119.33s eta: 19:01:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31960.0 tgs: 265 data_time: 0.93s time: 120.32s eta: 19:09:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31106.0 tgs: 253 data_time: 0.91s time: 122.90s eta: 19:31:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31337.0 tgs: 262 data_time: 0.94s time: 119.48s eta: 18:57:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31611.0 tgs: 265 data_time: 0.93s time: 119.15s eta: 18:51:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31573.0 tgs: 264 data_time: 0.64s time: 119.25s eta: 18:50:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.218 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31897.0 tgs: 267 data_time: 0.97s time: 119.40s eta: 18:50:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32058.0 tgs: 267 data_time: 0.87s time: 119.65s eta: 18:50:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31119.0 tgs: 260 data_time: 0.80s time: 119.39s eta: 18:46:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31356.0 tgs: 261 data_time: 0.55s time: 119.79s eta: 18:47:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32225.0 tgs: 249 data_time: 1.02s time: 129.31s eta: 20:15:30 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31625.0 tgs: 264 data_time: 0.72s time: 119.57s eta: 18:41:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31942.0 tgs: 267 data_time: 0.82s time: 119.39s eta: 18:38:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32061.0 tgs: 265 data_time: 0.85s time: 120.66s eta: 18:48:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30275.0 tgs: 253 data_time: 0.90s time: 119.40s eta: 18:34:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31108.0 tgs: 261 data_time: 0.73s time: 119.15s eta: 18:30:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 268 data_time: 0.74s time: 119.07s eta: 18:27:22 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 29801.0 tgs: 250 data_time: 0.91s time: 119.13s eta: 18:25:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.244 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31102.0 tgs: 258 data_time: 0.81s time: 120.39s eta: 18:35:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.108 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32173.0 tgs: 265 data_time: 0.74s time: 121.17s eta: 18:40:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31492.0 tgs: 260 data_time: 0.84s time: 120.66s eta: 18:34:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31539.0 tgs: 263 data_time: 0.84s time: 119.74s eta: 18:23:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32436.0 tgs: 261 data_time: 1.12s time: 123.87s eta: 18:59:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32092.0 tgs: 268 data_time: 0.65s time: 119.56s eta: 18:17:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31585.0 tgs: 264 data_time: 0.82s time: 119.34s eta: 18:13:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30729.0 tgs: 257 data_time: 0.44s time: 119.22s eta: 18:10:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31914.0 tgs: 266 data_time: 0.71s time: 119.60s eta: 18:12:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.90s time: 119.26s eta: 18:07:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31767.0 tgs: 266 data_time: 0.88s time: 119.01s eta: 18:02:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31842.0 tgs: 267 data_time: 0.81s time: 119.10s eta: 18:01:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.253 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31119.0 tgs: 260 data_time: 0.84s time: 119.27s eta: 18:01:20 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31541.0 tgs: 264 data_time: 1.16s time: 119.33s eta: 17:59:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 263 data_time: 0.75s time: 120.02s eta: 18:04:09 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 270 data_time: 0.92s time: 119.11s eta: 17:53:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.109 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31104.0 tgs: 260 data_time: 0.66s time: 119.41s eta: 17:54:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32112.0 tgs: 269 data_time: 0.72s time: 118.99s eta: 17:48:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31195.0 tgs: 262 data_time: 0.71s time: 119.02s eta: 17:47:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31401.0 tgs: 250 data_time: 0.60s time: 125.37s eta: 18:42:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 260 data_time: 0.78s time: 120.10s eta: 17:52:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31585.0 tgs: 264 data_time: 0.77s time: 119.38s eta: 17:44:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31441.0 tgs: 263 data_time: 0.76s time: 119.34s eta: 17:42:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 269 data_time: 1.00s time: 119.51s eta: 17:41:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32203.0 tgs: 255 data_time: 0.79s time: 126.05s eta: 18:37:37 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31194.0 tgs: 254 data_time: 0.78s time: 122.64s eta: 18:05:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.253 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31045.0 tgs: 256 data_time: 0.72s time: 121.16s eta: 17:50:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.232 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32341.0 tgs: 271 data_time: 0.76s time: 119.04s eta: 17:29:30 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32008.0 tgs: 269 data_time: 0.75s time: 118.85s eta: 17:25:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31320.0 tgs: 252 data_time: 1.17s time: 124.00s eta: 18:09:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31671.0 tgs: 254 data_time: 0.39s time: 124.57s eta: 18:12:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31891.0 tgs: 267 data_time: 0.71s time: 119.39s eta: 17:24:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.290 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31509.0 tgs: 263 data_time: 0.92s time: 119.69s eta: 17:25:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.127 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31543.0 tgs: 263 data_time: 0.75s time: 119.53s eta: 17:21:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31409.0 tgs: 264 data_time: 0.70s time: 118.86s eta: 17:14:04 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31460.0 tgs: 262 data_time: 0.78s time: 119.85s eta: 17:20:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32375.0 tgs: 270 data_time: 0.76s time: 119.54s eta: 17:16:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 263 data_time: 0.90s time: 121.86s eta: 17:34:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31021.0 tgs: 251 data_time: 0.60s time: 123.52s eta: 17:46:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31677.0 tgs: 266 data_time: 0.50s time: 119.00s eta: 17:05:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31283.0 tgs: 262 data_time: 0.70s time: 119.23s eta: 17:05:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31695.0 tgs: 266 data_time: 0.65s time: 119.13s eta: 17:02:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31430.0 tgs: 263 data_time: 0.70s time: 119.26s eta: 17:01:37 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31860.0 tgs: 267 data_time: 0.64s time: 119.04s eta: 16:57:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31735.0 tgs: 266 data_time: 0.87s time: 119.28s eta: 16:57:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32162.0 tgs: 269 data_time: 0.60s time: 119.16s eta: 16:54:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.121 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32305.0 tgs: 270 data_time: 0.81s time: 119.58s eta: 16:56:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31604.0 tgs: 265 data_time: 0.63s time: 119.13s eta: 16:50:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31021.0 tgs: 261 data_time: 0.72s time: 118.84s eta: 16:46:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30926.0 tgs: 259 data_time: 0.79s time: 119.01s eta: 16:45:39 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.128 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31731.0 tgs: 264 data_time: 0.58s time: 119.85s eta: 16:50:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32114.0 tgs: 269 data_time: 0.65s time: 119.38s eta: 16:44:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32395.0 tgs: 272 data_time: 0.71s time: 119.01s eta: 16:39:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32192.0 tgs: 266 data_time: 1.00s time: 120.93s eta: 16:53:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32117.0 tgs: 269 data_time: 0.88s time: 118.99s eta: 16:35:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30997.0 tgs: 259 data_time: 0.56s time: 119.28s eta: 16:36:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31290.0 tgs: 262 data_time: 0.86s time: 119.23s eta: 16:33:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31546.0 tgs: 264 data_time: 0.68s time: 119.37s eta: 16:32:44 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30543.0 tgs: 255 data_time: 0.88s time: 119.70s eta: 16:33:30 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 267 data_time: 1.08s time: 119.21s eta: 16:27:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31304.0 tgs: 261 data_time: 0.61s time: 119.76s eta: 16:30:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29923.0 tgs: 248 data_time: 0.92s time: 120.35s eta: 16:32:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31723.0 tgs: 266 data_time: 0.60s time: 118.97s eta: 16:19:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31481.0 tgs: 247 data_time: 0.77s time: 127.33s eta: 17:26:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30947.0 tgs: 259 data_time: 0.51s time: 119.22s eta: 16:17:37 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31715.0 tgs: 266 data_time: 0.85s time: 119.14s eta: 16:14:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31002.0 tgs: 260 data_time: 0.80s time: 119.23s eta: 16:13:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 264 data_time: 0.69s time: 119.18s eta: 16:11:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31512.0 tgs: 264 data_time: 0.66s time: 118.95s eta: 16:07:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 266 data_time: 0.58s time: 118.86s eta: 16:04:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.116 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32180.0 tgs: 270 data_time: 0.82s time: 118.96s eta: 16:03:34 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 264 data_time: 0.83s time: 121.84s eta: 16:24:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31857.0 tgs: 267 data_time: 0.81s time: 119.00s eta: 15:59:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.83s time: 119.22s eta: 15:59:44 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30936.0 tgs: 258 data_time: 0.64s time: 119.61s eta: 16:00:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:48:49][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32371.0 tgs: 271 data_time: 0.70s time: 119.04s eta: 15:54:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31838.0 tgs: 267 data_time: 0.69s time: 118.98s eta: 15:51:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31824.0 tgs: 267 data_time: 0.99s time: 119.18s eta: 15:51:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31572.0 tgs: 263 data_time: 0.53s time: 119.63s eta: 15:53:04 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31084.0 tgs: 261 data_time: 0.60s time: 118.94s eta: 15:45:32 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.59s time: 119.28s eta: 15:46:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32223.0 tgs: 270 data_time: 0.78s time: 119.12s eta: 15:43:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32068.0 tgs: 269 data_time: 0.80s time: 119.00s eta: 15:40:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.203 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31424.0 tgs: 262 data_time: 0.64s time: 119.65s eta: 15:43:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32014.0 tgs: 263 data_time: 0.94s time: 121.50s eta: 15:55:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.217 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31957.0 tgs: 267 data_time: 0.92s time: 119.41s eta: 15:37:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32083.0 tgs: 246 data_time: 0.78s time: 129.97s eta: 16:58:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31479.0 tgs: 263 data_time: 0.98s time: 119.47s eta: 15:33:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.215 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30764.0 tgs: 257 data_time: 0.96s time: 119.60s eta: 15:32:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 269 data_time: 0.52s time: 119.53s eta: 15:30:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31377.0 tgs: 256 data_time: 0.78s time: 122.14s eta: 15:48:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32207.0 tgs: 266 data_time: 0.76s time: 120.75s eta: 15:35:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31143.0 tgs: 260 data_time: 0.71s time: 119.41s eta: 15:23:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32001.0 tgs: 267 data_time: 0.83s time: 119.44s eta: 15:21:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31599.0 tgs: 240 data_time: 1.10s time: 131.64s eta: 16:53:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32100.0 tgs: 269 data_time: 0.78s time: 119.29s eta: 15:16:34 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31853.0 tgs: 266 data_time: 0.57s time: 119.56s eta: 15:16:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31641.0 tgs: 258 data_time: 0.73s time: 122.25s eta: 15:35:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31967.0 tgs: 267 data_time: 0.80s time: 119.72s eta: 15:13:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.87s time: 119.35s eta: 15:09:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.232 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32339.0 tgs: 271 data_time: 0.81s time: 118.90s eta: 15:03:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32060.0 tgs: 269 data_time: 0.76s time: 119.00s eta: 15:02:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31794.0 tgs: 266 data_time: 0.95s time: 119.31s eta: 15:02:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.185 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30231.0 tgs: 254 data_time: 0.70s time: 118.92s eta: 14:57:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31939.0 tgs: 265 data_time: 0.76s time: 120.52s eta: 15:07:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31155.0 tgs: 261 data_time: 0.75s time: 119.32s eta: 14:56:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 260 data_time: 0.67s time: 122.66s eta: 15:19:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.184 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31398.0 tgs: 263 data_time: 0.51s time: 119.13s eta: 14:51:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.184 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.3GB text_tokens: 30027.0 tgs: 250 data_time: 0.73s time: 119.76s eta: 14:54:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31856.0 tgs: 266 data_time: 1.02s time: 119.46s eta: 14:49:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.213 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31765.0 tgs: 266 data_time: 0.59s time: 118.99s eta: 14:44:28 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 265 data_time: 0.94s time: 119.32s eta: 14:44:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31084.0 tgs: 261 data_time: 0.67s time: 118.94s eta: 14:40:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 268 data_time: 0.80s time: 119.05s eta: 14:39:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31449.0 tgs: 264 data_time: 0.69s time: 119.06s eta: 14:37:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.224 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.79s time: 119.13s eta: 14:35:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31945.0 tgs: 268 data_time: 1.01s time: 119.01s eta: 14:32:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30767.0 tgs: 258 data_time: 0.58s time: 119.06s eta: 14:31:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.86s time: 119.15s eta: 14:29:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31573.0 tgs: 263 data_time: 0.82s time: 119.65s eta: 14:31:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30682.0 tgs: 257 data_time: 0.98s time: 119.18s eta: 14:26:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 268 data_time: 0.93s time: 119.20s eta: 14:24:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31619.0 tgs: 264 data_time: 0.78s time: 119.39s eta: 14:23:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31409.0 tgs: 260 data_time: 0.65s time: 120.73s eta: 14:31:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31745.0 tgs: 267 data_time: 0.58s time: 118.85s eta: 14:15:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.126 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31139.0 tgs: 261 data_time: 0.70s time: 119.10s eta: 14:15:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31695.0 tgs: 266 data_time: 0.84s time: 119.14s eta: 14:13:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31698.0 tgs: 264 data_time: 0.73s time: 119.96s eta: 14:17:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31865.0 tgs: 265 data_time: 0.66s time: 119.87s eta: 14:15:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.211 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 261 data_time: 0.86s time: 119.59s eta: 14:11:04 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32104.0 tgs: 269 data_time: 0.91s time: 119.22s eta: 14:06:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.181 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32166.0 tgs: 269 data_time: 0.73s time: 119.26s eta: 14:04:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31283.0 tgs: 262 data_time: 0.71s time: 119.28s eta: 14:02:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31863.0 tgs: 267 data_time: 0.69s time: 119.13s eta: 13:59:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31634.0 tgs: 265 data_time: 0.63s time: 119.12s eta: 13:57:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32170.0 tgs: 269 data_time: 0.93s time: 119.17s eta: 13:56:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.338 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31578.0 tgs: 264 data_time: 0.83s time: 119.35s eta: 13:55:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30931.0 tgs: 257 data_time: 0.56s time: 120.11s eta: 13:58:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.196 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30410.0 tgs: 248 data_time: 0.83s time: 122.55s eta: 14:13:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 265 data_time: 0.78s time: 119.10s eta: 13:47:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.110 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 28123.0 tgs: 235 data_time: 0.63s time: 119.37s eta: 13:47:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32503.0 tgs: 272 data_time: 0.79s time: 119.34s eta: 13:45:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.125 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32544.0 tgs: 272 data_time: 0.73s time: 119.33s eta: 13:43:22 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31532.0 tgs: 264 data_time: 0.71s time: 119.11s eta: 13:39:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31600.0 tgs: 264 data_time: 0.76s time: 119.25s eta: 13:38:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31796.0 tgs: 266 data_time: 0.89s time: 119.13s eta: 13:36:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 264 data_time: 0.76s time: 119.22s eta: 13:34:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31789.0 tgs: 266 data_time: 0.97s time: 119.46s eta: 13:34:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31573.0 tgs: 265 data_time: 0.92s time: 119.08s eta: 13:29:44 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.70s time: 118.92s eta: 13:26:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31531.0 tgs: 263 data_time: 0.43s time: 119.81s eta: 13:30:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.209 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32533.0 tgs: 272 data_time: 0.91s time: 119.25s eta: 13:24:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30958.0 tgs: 259 data_time: 0.76s time: 119.16s eta: 13:22:20 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.094 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31592.0 tgs: 264 data_time: 0.68s time: 119.35s eta: 13:21:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32567.0 tgs: 273 data_time: 0.66s time: 119.26s eta: 13:19:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31043.0 tgs: 260 data_time: 0.64s time: 119.04s eta: 13:15:34 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31096.0 tgs: 260 data_time: 0.73s time: 119.36s eta: 13:15:44 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32269.0 tgs: 270 data_time: 0.58s time: 119.18s eta: 13:12:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31543.0 tgs: 264 data_time: 0.72s time: 119.46s eta: 13:12:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31111.0 tgs: 260 data_time: 0.97s time: 119.28s eta: 13:09:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31565.0 tgs: 264 data_time: 0.77s time: 119.15s eta: 13:06:22 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31842.0 tgs: 264 data_time: 0.79s time: 120.41s eta: 13:12:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31191.0 tgs: 259 data_time: 0.46s time: 120.16s eta: 13:09:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.164 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32299.0 tgs: 270 data_time: 0.80s time: 119.50s eta: 13:02:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32546.0 tgs: 272 data_time: 0.87s time: 119.51s eta: 13:00:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.172 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31772.0 tgs: 265 data_time: 0.63s time: 119.57s eta: 12:59:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.120 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30728.0 tgs: 256 data_time: 0.64s time: 119.79s eta: 12:58:39 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.188 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31607.0 tgs: 265 data_time: 0.92s time: 119.04s eta: 12:51:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.116 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31143.0 tgs: 261 data_time: 0.78s time: 119.05s eta: 12:49:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.188 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32090.0 tgs: 269 data_time: 1.08s time: 119.12s eta: 12:48:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31240.0 tgs: 261 data_time: 0.59s time: 119.40s eta: 12:48:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31952.0 tgs: 268 data_time: 0.89s time: 119.19s eta: 12:44:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.105 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31361.0 tgs: 262 data_time: 0.47s time: 119.44s eta: 12:44:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31360.0 tgs: 263 data_time: 0.63s time: 119.09s eta: 12:40:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.124 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 269 data_time: 0.64s time: 119.07s eta: 12:38:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31045.0 tgs: 259 data_time: 0.66s time: 119.59s eta: 12:39:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30512.0 tgs: 256 data_time: 0.53s time: 118.79s eta: 12:32:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31380.0 tgs: 263 data_time: 0.65s time: 118.88s eta: 12:30:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32513.0 tgs: 272 data_time: 0.72s time: 119.16s eta: 12:30:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31177.0 tgs: 261 data_time: 0.64s time: 119.27s eta: 12:29:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.127 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31722.0 tgs: 265 data_time: 0.61s time: 119.35s eta: 12:27:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31523.0 tgs: 264 data_time: 0.83s time: 119.12s eta: 12:24:28 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31972.0 tgs: 267 data_time: 1.03s time: 119.34s eta: 12:23:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.207 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31702.0 tgs: 265 data_time: 0.65s time: 119.26s eta: 12:21:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.79s time: 119.03s eta: 12:17:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.129 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 260 data_time: 0.60s time: 121.13s eta: 12:28:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31737.0 tgs: 264 data_time: 0.71s time: 119.99s eta: 12:19:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30726.0 tgs: 258 data_time: 0.56s time: 118.71s eta: 12:10:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31801.0 tgs: 265 data_time: 0.71s time: 119.63s eta: 12:13:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31699.0 tgs: 264 data_time: 0.68s time: 119.83s eta: 12:12:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31092.0 tgs: 260 data_time: 0.79s time: 119.14s eta: 12:06:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.194 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32141.0 tgs: 269 data_time: 0.67s time: 119.17s eta: 12:04:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.2GB text_tokens: 30077.0 tgs: 252 data_time: 0.57s time: 119.11s eta: 12:02:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31595.0 tgs: 265 data_time: 0.67s time: 119.18s eta: 12:01:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31495.0 tgs: 264 data_time: 0.76s time: 118.95s eta: 11:57:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31823.0 tgs: 265 data_time: 0.88s time: 119.70s eta: 12:00:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 271 data_time: 0.71s time: 119.27s eta: 11:55:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.168 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31428.0 tgs: 263 data_time: 0.89s time: 119.27s eta: 11:53:39 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 265 data_time: 0.83s time: 119.71s eta: 11:54:16 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.209 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32231.0 tgs: 271 data_time: 0.88s time: 118.78s eta: 11:46:44 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30750.0 tgs: 258 data_time: 0.66s time: 119.16s eta: 11:46:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 269 data_time: 0.71s time: 119.61s eta: 11:47:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 270 data_time: 0.80s time: 118.86s eta: 11:41:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31937.0 tgs: 269 data_time: 0.73s time: 118.68s eta: 11:38:14 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.198 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31383.0 tgs: 263 data_time: 0.79s time: 119.22s eta: 11:39:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.184 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 270 data_time: 0.81s time: 119.13s eta: 11:36:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31957.0 tgs: 268 data_time: 0.62s time: 119.07s eta: 11:34:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31613.0 tgs: 265 data_time: 0.87s time: 118.95s eta: 11:31:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31294.0 tgs: 263 data_time: 0.79s time: 118.91s eta: 11:29:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 266 data_time: 0.76s time: 118.97s eta: 11:28:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32010.0 tgs: 268 data_time: 0.69s time: 119.28s eta: 11:27:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31706.0 tgs: 265 data_time: 0.58s time: 119.35s eta: 11:26:16 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31687.0 tgs: 266 data_time: 0.64s time: 118.89s eta: 11:21:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31892.0 tgs: 267 data_time: 0.76s time: 119.20s eta: 11:21:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31923.0 tgs: 266 data_time: 0.53s time: 119.72s eta: 11:22:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30642.0 tgs: 257 data_time: 0.71s time: 119.12s eta: 11:16:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.183 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31304.0 tgs: 262 data_time: 0.84s time: 119.12s eta: 11:15:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30222.0 tgs: 254 data_time: 0.70s time: 118.95s eta: 11:12:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32183.0 tgs: 270 data_time: 0.83s time: 119.09s eta: 11:10:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 264 data_time: 0.79s time: 119.16s eta: 11:09:16 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30971.0 tgs: 260 data_time: 0.88s time: 118.97s eta: 11:06:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31802.0 tgs: 266 data_time: 0.60s time: 119.47s eta: 11:07:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 268 data_time: 0.75s time: 119.64s eta: 11:06:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.212 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31468.0 tgs: 263 data_time: 0.81s time: 119.32s eta: 11:02:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30564.0 tgs: 256 data_time: 0.68s time: 119.30s eta: 11:00:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32260.0 tgs: 271 data_time: 0.88s time: 118.82s eta: 10:55:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.116 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31636.0 tgs: 264 data_time: 1.02s time: 119.47s eta: 10:57:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31742.0 tgs: 265 data_time: 0.58s time: 119.37s eta: 10:54:31 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 267 data_time: 0.84s time: 119.28s eta: 10:52:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31543.0 tgs: 264 data_time: 0.84s time: 119.25s eta: 10:49:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31203.0 tgs: 261 data_time: 0.58s time: 119.10s eta: 10:47:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.60s time: 119.53s eta: 10:47:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30113.0 tgs: 252 data_time: 0.67s time: 119.34s eta: 10:44:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.225 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32047.0 tgs: 268 data_time: 0.87s time: 119.49s eta: 10:43:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 265 data_time: 0.71s time: 119.51s eta: 10:41:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.148 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31679.0 tgs: 266 data_time: 0.87s time: 118.99s eta: 10:36:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 0.66s time: 119.22s eta: 10:35:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32066.0 tgs: 269 data_time: 0.60s time: 119.06s eta: 10:33:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32237.0 tgs: 271 data_time: 0.76s time: 118.78s eta: 10:29:31 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32091.0 tgs: 267 data_time: 1.04s time: 119.86s eta: 10:33:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31105.0 tgs: 260 data_time: 0.70s time: 119.38s eta: 10:28:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31678.0 tgs: 266 data_time: 0.64s time: 118.84s eta: 10:23:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32063.0 tgs: 269 data_time: 0.80s time: 118.98s eta: 10:22:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31238.0 tgs: 261 data_time: 0.87s time: 119.28s eta: 10:22:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.96s time: 119.20s eta: 10:19:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31036.0 tgs: 260 data_time: 0.60s time: 119.06s eta: 10:17:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32673.0 tgs: 273 data_time: 0.68s time: 119.48s eta: 10:17:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.56s time: 119.38s eta: 10:14:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.313 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31813.0 tgs: 266 data_time: 0.78s time: 119.46s eta: 10:13:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31265.0 tgs: 262 data_time: 0.81s time: 119.28s eta: 10:10:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29916.0 tgs: 251 data_time: 0.61s time: 119.11s eta: 10:07:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31398.0 tgs: 262 data_time: 0.68s time: 119.70s eta: 10:08:28 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.117 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31790.0 tgs: 265 data_time: 0.59s time: 119.83s eta: 10:07:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30621.0 tgs: 251 data_time: 0.54s time: 121.55s eta: 10:13:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.214 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31482.0 tgs: 263 data_time: 0.80s time: 119.25s eta: 10:00:14 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.190 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30887.0 tgs: 259 data_time: 0.69s time: 118.80s eta: 9:56:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.197 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31016.0 tgs: 259 data_time: 0.87s time: 119.56s eta: 9:57:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.176 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31567.0 tgs: 263 data_time: 0.65s time: 119.71s eta: 9:56:32 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.102 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31464.0 tgs: 263 data_time: 0.56s time: 119.26s eta: 9:52:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31591.0 tgs: 264 data_time: 0.79s time: 119.33s eta: 9:50:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31766.0 tgs: 266 data_time: 0.64s time: 119.22s eta: 9:48:09 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31211.0 tgs: 258 data_time: 0.81s time: 120.88s eta: 9:54:20 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32200.0 tgs: 269 data_time: 0.62s time: 119.28s eta: 9:44:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31498.0 tgs: 263 data_time: 0.74s time: 119.46s eta: 9:43:21 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31207.0 tgs: 261 data_time: 0.70s time: 119.54s eta: 9:41:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30631.0 tgs: 256 data_time: 0.86s time: 119.20s eta: 9:38:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31489.0 tgs: 263 data_time: 0.69s time: 119.48s eta: 9:37:28 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32287.0 tgs: 271 data_time: 0.62s time: 119.05s eta: 9:33:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.198 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31601.0 tgs: 265 data_time: 0.64s time: 119.24s eta: 9:32:21 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.194 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31121.0 tgs: 261 data_time: 0.57s time: 119.16s eta: 9:29:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.141 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31063.0 tgs: 260 data_time: 0.78s time: 119.22s eta: 9:28:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30929.0 tgs: 258 data_time: 0.70s time: 119.46s eta: 9:27:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31038.0 tgs: 260 data_time: 0.58s time: 118.99s eta: 9:23:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31962.0 tgs: 268 data_time: 0.96s time: 119.04s eta: 9:21:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32040.0 tgs: 268 data_time: 0.66s time: 119.53s eta: 9:21:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32110.0 tgs: 268 data_time: 0.56s time: 119.59s eta: 9:20:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32125.0 tgs: 269 data_time: 0.52s time: 119.12s eta: 9:15:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.082 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30865.0 tgs: 259 data_time: 0.56s time: 119.02s eta: 9:13:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30444.0 tgs: 256 data_time: 0.65s time: 118.88s eta: 9:10:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.199 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 266 data_time: 0.68s time: 119.32s eta: 9:10:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.125 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32554.0 tgs: 273 data_time: 0.65s time: 119.09s eta: 9:07:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.118 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31838.0 tgs: 267 data_time: 0.55s time: 118.86s eta: 9:04:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.204 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31585.0 tgs: 263 data_time: 0.82s time: 119.67s eta: 9:06:28 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.098 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32065.0 tgs: 268 data_time: 0.57s time: 119.43s eta: 9:03:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32365.0 tgs: 270 data_time: 0.72s time: 119.43s eta: 9:01:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31530.0 tgs: 265 data_time: 0.60s time: 118.74s eta: 8:56:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31843.0 tgs: 267 data_time: 0.86s time: 118.87s eta: 8:54:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 266 data_time: 0.95s time: 119.34s eta: 8:55:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31883.0 tgs: 267 data_time: 0.83s time: 119.14s eta: 8:52:10 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31263.0 tgs: 262 data_time: 0.60s time: 119.27s eta: 8:50:44 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31502.0 tgs: 261 data_time: 0.62s time: 120.47s eta: 8:54:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 265 data_time: 0.60s time: 119.24s eta: 8:46:37 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.118 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32212.0 tgs: 270 data_time: 1.00s time: 119.09s eta: 8:44:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.194 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32265.0 tgs: 267 data_time: 0.80s time: 120.82s eta: 8:49:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31501.0 tgs: 262 data_time: 0.76s time: 120.19s eta: 8:44:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31740.0 tgs: 267 data_time: 0.72s time: 118.83s eta: 8:36:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32142.0 tgs: 269 data_time: 0.94s time: 119.35s eta: 8:37:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.119 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31537.0 tgs: 264 data_time: 0.78s time: 119.33s eta: 8:35:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 268 data_time: 0.59s time: 119.05s eta: 8:31:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31323.0 tgs: 262 data_time: 0.82s time: 119.17s eta: 8:30:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30331.0 tgs: 254 data_time: 0.59s time: 119.04s eta: 8:27:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 267 data_time: 0.55s time: 119.12s eta: 8:26:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 267 data_time: 0.62s time: 119.25s eta: 8:24:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31288.0 tgs: 262 data_time: 0.91s time: 119.22s eta: 8:22:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30457.0 tgs: 255 data_time: 0.82s time: 119.38s eta: 8:21:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 266 data_time: 1.11s time: 119.78s eta: 8:21:04 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.234 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31161.0 tgs: 261 data_time: 0.83s time: 119.19s eta: 8:16:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 269 data_time: 0.59s time: 119.00s eta: 8:13:51 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31143.0 tgs: 259 data_time: 0.79s time: 119.89s eta: 8:15:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.185 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31602.0 tgs: 264 data_time: 0.54s time: 119.56s eta: 8:12:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.212 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32321.0 tgs: 270 data_time: 0.70s time: 119.59s eta: 8:10:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32436.0 tgs: 272 data_time: 0.79s time: 119.00s eta: 8:05:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.172 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31579.0 tgs: 263 data_time: 0.87s time: 119.84s eta: 8:07:21 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.196 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32370.0 tgs: 270 data_time: 1.04s time: 119.70s eta: 8:04:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31525.0 tgs: 264 data_time: 0.69s time: 119.40s eta: 8:01:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31237.0 tgs: 261 data_time: 0.66s time: 119.28s eta: 7:59:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.249 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31094.0 tgs: 260 data_time: 0.56s time: 119.19s eta: 7:56:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.097 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30446.0 tgs: 255 data_time: 0.62s time: 119.20s eta: 7:54:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31584.0 tgs: 265 data_time: 0.69s time: 118.94s eta: 7:51:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32064.0 tgs: 269 data_time: 0.71s time: 118.97s eta: 7:49:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 271 data_time: 0.80s time: 119.42s eta: 7:49:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31385.0 tgs: 263 data_time: 0.43s time: 119.03s eta: 7:46:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.133 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31223.0 tgs: 262 data_time: 1.01s time: 119.10s eta: 7:44:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29144.0 tgs: 242 data_time: 0.40s time: 120.01s eta: 7:46:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31991.0 tgs: 266 data_time: 0.65s time: 119.83s eta: 7:43:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31446.0 tgs: 263 data_time: 0.80s time: 119.16s eta: 7:38:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 267 data_time: 0.74s time: 118.97s eta: 7:36:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32275.0 tgs: 270 data_time: 0.73s time: 119.32s eta: 7:35:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31450.0 tgs: 263 data_time: 0.70s time: 119.38s eta: 7:33:39 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30569.0 tgs: 256 data_time: 0.60s time: 119.06s eta: 7:30:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31563.0 tgs: 265 data_time: 0.61s time: 118.99s eta: 7:28:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 266 data_time: 0.65s time: 119.24s eta: 7:27:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30549.0 tgs: 256 data_time: 0.55s time: 118.99s eta: 7:24:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30501.0 tgs: 256 data_time: 0.93s time: 118.93s eta: 7:22:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31415.0 tgs: 263 data_time: 0.61s time: 119.08s eta: 7:20:34 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 270 data_time: 0.67s time: 118.99s eta: 7:18:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 268 data_time: 0.72s time: 119.03s eta: 7:16:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31394.0 tgs: 263 data_time: 0.72s time: 119.24s eta: 7:15:14 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32162.0 tgs: 269 data_time: 0.71s time: 119.36s eta: 7:13:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32229.0 tgs: 270 data_time: 0.95s time: 118.94s eta: 7:10:10 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31850.0 tgs: 267 data_time: 0.91s time: 119.20s eta: 7:09:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31664.0 tgs: 264 data_time: 0.64s time: 119.58s eta: 7:08:30 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.211 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 266 data_time: 1.12s time: 119.66s eta: 7:06:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32366.0 tgs: 272 data_time: 0.83s time: 118.96s eta: 7:02:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32393.0 tgs: 271 data_time: 0.82s time: 119.29s eta: 7:01:30 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.182 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32115.0 tgs: 268 data_time: 0.86s time: 119.52s eta: 7:00:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 265 data_time: 0.57s time: 119.43s eta: 6:57:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30729.0 tgs: 258 data_time: 0.91s time: 118.91s eta: 6:54:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31082.0 tgs: 260 data_time: 0.91s time: 119.34s eta: 6:53:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.188 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 267 data_time: 0.49s time: 118.93s eta: 6:50:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31354.0 tgs: 262 data_time: 0.82s time: 119.27s eta: 6:49:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32196.0 tgs: 270 data_time: 0.67s time: 119.02s eta: 6:46:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32244.0 tgs: 270 data_time: 0.91s time: 119.23s eta: 6:45:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.185 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31104.0 tgs: 260 data_time: 0.76s time: 119.40s eta: 6:43:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 0.85s time: 119.01s eta: 6:40:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.115 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.71s time: 119.01s eta: 6:38:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32059.0 tgs: 268 data_time: 0.62s time: 119.46s eta: 6:38:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.235 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31353.0 tgs: 263 data_time: 0.80s time: 118.80s eta: 6:34:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31972.0 tgs: 267 data_time: 0.67s time: 119.44s eta: 6:34:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32366.0 tgs: 270 data_time: 0.81s time: 119.57s eta: 6:32:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31755.0 tgs: 266 data_time: 0.85s time: 119.38s eta: 6:29:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 270 data_time: 0.78s time: 119.72s eta: 6:29:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31955.0 tgs: 267 data_time: 0.65s time: 119.36s eta: 6:25:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31614.0 tgs: 266 data_time: 0.73s time: 118.82s eta: 6:22:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30900.0 tgs: 259 data_time: 0.64s time: 119.02s eta: 6:20:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.112 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31254.0 tgs: 261 data_time: 0.68s time: 119.40s eta: 6:20:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 269 data_time: 0.60s time: 119.34s eta: 6:17:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31294.0 tgs: 262 data_time: 0.73s time: 119.05s eta: 6:14:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31826.0 tgs: 266 data_time: 0.94s time: 119.40s eta: 6:14:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31203.0 tgs: 261 data_time: 0.82s time: 119.39s eta: 6:12:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 270 data_time: 0.64s time: 119.29s eta: 6:09:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.73s time: 119.09s eta: 6:07:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32002.0 tgs: 262 data_time: 0.80s time: 121.96s eta: 6:14:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30972.0 tgs: 259 data_time: 0.61s time: 119.38s eta: 6:04:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32374.0 tgs: 270 data_time: 0.75s time: 119.57s eta: 6:02:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31817.0 tgs: 266 data_time: 0.79s time: 119.26s eta: 5:59:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32097.0 tgs: 269 data_time: 0.85s time: 119.26s eta: 5:57:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31825.0 tgs: 267 data_time: 0.65s time: 118.91s eta: 5:54:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32384.0 tgs: 272 data_time: 0.58s time: 119.02s eta: 5:53:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32222.0 tgs: 270 data_time: 0.67s time: 119.07s eta: 5:51:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31756.0 tgs: 266 data_time: 0.64s time: 119.29s eta: 5:49:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32129.0 tgs: 270 data_time: 0.72s time: 118.97s eta: 5:46:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 267 data_time: 0.71s time: 119.27s eta: 5:45:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.195 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32168.0 tgs: 269 data_time: 0.73s time: 119.35s eta: 5:44:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31865.0 tgs: 267 data_time: 0.57s time: 119.05s eta: 5:41:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.126 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30957.0 tgs: 259 data_time: 0.55s time: 119.50s eta: 5:40:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31644.0 tgs: 265 data_time: 0.75s time: 119.38s eta: 5:38:14 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 264 data_time: 0.78s time: 119.57s eta: 5:36:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31720.0 tgs: 266 data_time: 0.75s time: 118.98s eta: 5:33:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31219.0 tgs: 262 data_time: 0.81s time: 118.95s eta: 5:31:04 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.108 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 269 data_time: 0.64s time: 119.24s eta: 5:29:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31411.0 tgs: 263 data_time: 0.67s time: 119.40s eta: 5:28:20 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31174.0 tgs: 260 data_time: 0.78s time: 119.47s eta: 5:26:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.191 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31866.0 tgs: 267 data_time: 0.60s time: 118.94s eta: 5:23:07 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31577.0 tgs: 263 data_time: 0.70s time: 119.63s eta: 5:23:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31278.0 tgs: 262 data_time: 0.90s time: 119.35s eta: 5:20:14 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31963.0 tgs: 268 data_time: 0.53s time: 119.12s eta: 5:17:39 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31226.0 tgs: 261 data_time: 0.87s time: 119.42s eta: 5:16:28 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30283.0 tgs: 253 data_time: 0.46s time: 119.58s eta: 5:14:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.111 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31070.0 tgs: 261 data_time: 0.70s time: 119.00s eta: 5:11:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.173 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30568.0 tgs: 256 data_time: 0.66s time: 119.00s eta: 5:09:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.188 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31675.0 tgs: 265 data_time: 0.63s time: 119.22s eta: 5:07:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.100 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31086.0 tgs: 259 data_time: 0.77s time: 119.68s eta: 5:07:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31177.0 tgs: 261 data_time: 0.67s time: 119.11s eta: 5:03:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.273 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32290.0 tgs: 271 data_time: 0.66s time: 119.07s eta: 5:01:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31867.0 tgs: 262 data_time: 0.68s time: 121.52s eta: 5:05:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32000.0 tgs: 267 data_time: 0.67s time: 119.57s eta: 4:58:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31542.0 tgs: 263 data_time: 0.67s time: 119.50s eta: 4:56:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31591.0 tgs: 263 data_time: 0.72s time: 119.73s eta: 4:55:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.224 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30698.0 tgs: 256 data_time: 0.74s time: 119.53s eta: 4:52:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32464.0 tgs: 271 data_time: 0.58s time: 119.42s eta: 4:50:34 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.205 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31364.0 tgs: 263 data_time: 0.63s time: 119.20s eta: 4:48:03 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 269 data_time: 0.84s time: 118.89s eta: 4:45:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31698.0 tgs: 266 data_time: 0.58s time: 119.08s eta: 4:43:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31021.0 tgs: 258 data_time: 0.78s time: 119.98s eta: 4:43:57 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.232 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30787.0 tgs: 258 data_time: 0.90s time: 119.15s eta: 4:40:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 269 data_time: 0.62s time: 119.00s eta: 4:37:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32058.0 tgs: 269 data_time: 0.83s time: 119.06s eta: 4:35:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31562.0 tgs: 264 data_time: 0.78s time: 119.38s eta: 4:34:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.185 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31301.0 tgs: 262 data_time: 0.78s time: 119.40s eta: 4:32:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31994.0 tgs: 268 data_time: 0.90s time: 118.94s eta: 4:29:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31561.0 tgs: 263 data_time: 0.48s time: 119.60s eta: 4:29:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32218.0 tgs: 270 data_time: 0.80s time: 119.25s eta: 4:26:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31243.0 tgs: 262 data_time: 0.61s time: 119.05s eta: 4:23:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31943.0 tgs: 268 data_time: 0.66s time: 118.85s eta: 4:21:28 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31236.0 tgs: 263 data_time: 0.74s time: 118.76s eta: 4:19:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31514.0 tgs: 263 data_time: 0.58s time: 119.43s eta: 4:18:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32078.0 tgs: 268 data_time: 0.78s time: 119.40s eta: 4:16:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.176 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32077.0 tgs: 268 data_time: 1.01s time: 119.36s eta: 4:14:37 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30994.0 tgs: 260 data_time: 0.86s time: 119.04s eta: 4:11:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.108 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32248.0 tgs: 270 data_time: 0.71s time: 119.12s eta: 4:10:09 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31188.0 tgs: 261 data_time: 0.80s time: 119.15s eta: 4:08:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.194 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.88s time: 118.95s eta: 4:05:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31711.0 tgs: 266 data_time: 0.71s time: 119.15s eta: 4:04:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31054.0 tgs: 259 data_time: 0.99s time: 119.51s eta: 4:03:00 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31011.0 tgs: 260 data_time: 0.83s time: 119.23s eta: 4:00:27 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.203 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30871.0 tgs: 259 data_time: 0.80s time: 119.19s eta: 3:58:22 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 268 data_time: 0.59s time: 119.36s eta: 3:56:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32193.0 tgs: 269 data_time: 0.65s time: 119.38s eta: 3:54:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32213.0 tgs: 268 data_time: 0.67s time: 119.95s eta: 3:53:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31726.0 tgs: 265 data_time: 0.66s time: 119.31s eta: 3:50:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31869.0 tgs: 266 data_time: 0.68s time: 119.63s eta: 3:49:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.66s time: 118.85s eta: 3:45:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31024.0 tgs: 259 data_time: 0.79s time: 119.73s eta: 3:45:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 268 data_time: 0.64s time: 119.32s eta: 3:42:44 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31887.0 tgs: 267 data_time: 0.66s time: 119.27s eta: 3:40:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 265 data_time: 0.67s time: 118.99s eta: 3:38:09 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31690.0 tgs: 265 data_time: 0.70s time: 119.43s eta: 3:36:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32030.0 tgs: 269 data_time: 0.68s time: 119.03s eta: 3:34:15 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30510.0 tgs: 256 data_time: 0.66s time: 119.01s eta: 3:32:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32465.0 tgs: 272 data_time: 0.93s time: 119.20s eta: 3:30:35 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31518.0 tgs: 262 data_time: 0.87s time: 120.12s eta: 3:30:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32288.0 tgs: 271 data_time: 0.90s time: 119.13s eta: 3:26:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 262 data_time: 0.69s time: 119.31s eta: 3:24:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31498.0 tgs: 264 data_time: 0.71s time: 119.08s eta: 3:22:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 268 data_time: 0.96s time: 119.43s eta: 3:21:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31071.0 tgs: 260 data_time: 0.65s time: 119.29s eta: 3:18:48 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30698.0 tgs: 257 data_time: 0.83s time: 119.22s eta: 3:16:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.194 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 269 data_time: 1.19s time: 119.15s eta: 3:14:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.371 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31250.0 tgs: 262 data_time: 0.98s time: 119.02s eta: 3:12:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31309.0 tgs: 262 data_time: 0.71s time: 119.19s eta: 3:10:41 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31128.0 tgs: 257 data_time: 0.78s time: 121.04s eta: 3:11:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.224 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32181.0 tgs: 269 data_time: 0.74s time: 119.51s eta: 3:07:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30468.0 tgs: 255 data_time: 0.52s time: 119.26s eta: 3:04:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.206 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31901.0 tgs: 268 data_time: 0.67s time: 118.99s eta: 3:02:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31201.0 tgs: 261 data_time: 0.67s time: 119.22s eta: 3:00:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31439.0 tgs: 263 data_time: 0.67s time: 119.46s eta: 2:59:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31340.0 tgs: 263 data_time: 0.61s time: 118.98s eta: 2:56:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 262 data_time: 0.72s time: 119.35s eta: 2:55:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31670.0 tgs: 265 data_time: 0.75s time: 119.41s eta: 2:53:08 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31880.0 tgs: 267 data_time: 0.67s time: 119.35s eta: 2:51:04 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30258.0 tgs: 249 data_time: 0.84s time: 121.47s eta: 2:52:04 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 31186.0 tgs: 261 data_time: 0.77s time: 119.23s eta: 2:46:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31634.0 tgs: 264 data_time: 0.72s time: 119.60s eta: 2:45:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31792.0 tgs: 267 data_time: 0.79s time: 118.91s eta: 2:42:30 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 262 data_time: 1.05s time: 119.18s eta: 2:40:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29697.0 tgs: 249 data_time: 0.52s time: 118.96s eta: 2:38:37 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29970.0 tgs: 251 data_time: 0.86s time: 119.24s eta: 2:36:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31873.0 tgs: 267 data_time: 0.92s time: 119.10s eta: 2:34:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32327.0 tgs: 270 data_time: 0.96s time: 119.43s eta: 2:33:16 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 265 data_time: 0.72s time: 120.23s eta: 2:32:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31477.0 tgs: 263 data_time: 0.72s time: 119.28s eta: 2:29:05 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31438.0 tgs: 264 data_time: 0.86s time: 118.88s eta: 2:26:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30928.0 tgs: 258 data_time: 0.78s time: 119.51s eta: 2:25:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31563.0 tgs: 265 data_time: 0.83s time: 119.07s eta: 2:22:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 28841.0 tgs: 242 data_time: 0.49s time: 119.14s eta: 2:20:58 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31586.0 tgs: 265 data_time: 0.74s time: 119.05s eta: 2:18:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30966.0 tgs: 258 data_time: 0.74s time: 119.84s eta: 2:17:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 264 data_time: 0.86s time: 119.31s eta: 2:15:13 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30931.0 tgs: 259 data_time: 0.70s time: 119.38s eta: 2:13:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32160.0 tgs: 268 data_time: 0.66s time: 119.68s eta: 2:11:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31921.0 tgs: 267 data_time: 0.76s time: 119.49s eta: 2:09:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 29647.0 tgs: 249 data_time: 0.69s time: 118.96s eta: 2:06:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 268 data_time: 0.75s time: 119.37s eta: 2:05:20 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32209.0 tgs: 270 data_time: 0.61s time: 119.27s eta: 2:03:14 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 269 data_time: 0.73s time: 119.22s eta: 2:01:12 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.186 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32502.0 tgs: 271 data_time: 0.61s time: 119.50s eta: 1:59:30 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31039.0 tgs: 259 data_time: 0.54s time: 119.42s eta: 1:57:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32202.0 tgs: 269 data_time: 0.82s time: 119.53s eta: 1:55:32 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 261 data_time: 0.66s time: 119.30s eta: 1:53:20 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31901.0 tgs: 267 data_time: 0.94s time: 119.25s eta: 1:51:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.080 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 264 data_time: 0.67s time: 118.93s eta: 1:49:01 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 263 data_time: 0.74s time: 119.64s eta: 1:47:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 265 data_time: 0.84s time: 119.72s eta: 1:45:45 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31853.0 tgs: 266 data_time: 0.71s time: 119.35s eta: 1:43:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31822.0 tgs: 266 data_time: 0.55s time: 119.30s eta: 1:41:24 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 269 data_time: 0.75s time: 119.28s eta: 1:39:23 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31174.0 tgs: 262 data_time: 0.49s time: 118.91s eta: 1:37:06 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 266 data_time: 0.59s time: 119.16s eta: 1:35:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31093.0 tgs: 261 data_time: 0.72s time: 119.06s eta: 1:33:16 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31055.0 tgs: 260 data_time: 0.75s time: 119.26s eta: 1:31:25 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32220.0 tgs: 269 data_time: 0.60s time: 119.73s eta: 1:29:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30017.0 tgs: 252 data_time: 0.65s time: 119.02s eta: 1:27:16 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31706.0 tgs: 266 data_time: 0.47s time: 118.87s eta: 1:25:11 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30728.0 tgs: 257 data_time: 0.71s time: 119.48s eta: 1:23:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31784.0 tgs: 266 data_time: 0.71s time: 119.48s eta: 1:21:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32315.0 tgs: 271 data_time: 0.64s time: 118.94s eta: 1:19:17 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31231.0 tgs: 262 data_time: 0.77s time: 118.93s eta: 1:17:18 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31895.0 tgs: 261 data_time: 0.56s time: 122.14s eta: 1:17:21 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 267 data_time: 0.76s time: 119.36s eta: 1:13:36 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30755.0 tgs: 258 data_time: 0.67s time: 119.06s eta: 1:11:26 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31483.0 tgs: 264 data_time: 0.68s time: 118.92s eta: 1:09:22 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.82s time: 119.11s eta: 1:07:29 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31077.0 tgs: 260 data_time: 0.86s time: 119.41s eta: 1:05:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.244 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31785.0 tgs: 266 data_time: 0.85s time: 119.34s eta: 1:03:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.209 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31792.0 tgs: 266 data_time: 0.73s time: 119.18s eta: 1:01:34 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.275 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31622.0 tgs: 265 data_time: 0.65s time: 119.27s eta: 0:59:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 269 data_time: 0.97s time: 119.44s eta: 0:57:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31637.0 tgs: 264 data_time: 0.57s time: 119.39s eta: 0:55:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31072.0 tgs: 260 data_time: 0.56s time: 119.48s eta: 0:53:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32553.0 tgs: 272 data_time: 0.61s time: 119.37s eta: 0:51:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 263 data_time: 0.83s time: 119.47s eta: 0:49:46 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31327.0 tgs: 263 data_time: 0.81s time: 118.89s eta: 0:47:33 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32564.0 tgs: 272 data_time: 0.69s time: 119.28s eta: 0:45:43 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31478.0 tgs: 264 data_time: 0.62s time: 119.20s eta: 0:43:42 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31089.0 tgs: 261 data_time: 0.51s time: 119.00s eta: 0:41:38 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 267 data_time: 0.84s time: 119.03s eta: 0:39:40 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31638.0 tgs: 264 data_time: 0.70s time: 119.66s eta: 0:37:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31621.0 tgs: 265 data_time: 0.84s time: 119.28s eta: 0:35:47 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30405.0 tgs: 253 data_time: 0.61s time: 119.74s eta: 0:33:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32094.0 tgs: 264 data_time: 0.58s time: 121.23s eta: 0:32:19 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 270 data_time: 0.73s time: 119.38s eta: 0:29:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.233 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31990.0 tgs: 268 data_time: 0.95s time: 119.32s eta: 0:27:50 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31243.0 tgs: 261 data_time: 0.88s time: 119.64s eta: 0:25:55 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 270 data_time: 0.69s time: 119.10s eta: 0:23:49 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31470.0 tgs: 263 data_time: 0.80s time: 119.38s eta: 0:21:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.94s time: 119.26s eta: 0:19:52 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31770.0 tgs: 266 data_time: 0.55s time: 119.24s eta: 0:17:53 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31771.0 tgs: 264 data_time: 0.68s time: 119.90s eta: 0:15:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31510.0 tgs: 264 data_time: 0.60s time: 119.20s eta: 0:13:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31813.0 tgs: 267 data_time: 0.79s time: 119.09s eta: 0:11:54 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 269 data_time: 1.09s time: 119.33s eta: 0:09:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 272 data_time: 0.80s time: 119.16s eta: 0:07:56 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31826.0 tgs: 265 data_time: 0.68s time: 119.69s eta: 0:05:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31863.0 tgs: 262 data_time: 0.94s time: 121.45s eta: 0:04:02 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.84s time: 119.21s eta: 0:01:59 +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.28 seconds, peak gpu memory 17.4G +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 52][DP 13][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank53.log b/20250121165312/rank53.log new file mode 100644 index 0000000000000000000000000000000000000000..3f6b5a21d79ff841682effaee7119d643549e361 --- /dev/null +++ b/20250121165312/rank53.log @@ -0,0 +1,851 @@ +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.03s +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:00:41][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 72.93 seconds, peak gpu memory 13.4G +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.390 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31804.0 tgs: 210 data_time: 1.80s time: 150.94s eta: 1 day, 0:51:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.235 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31938.0 tgs: 265 data_time: 0.65s time: 120.19s eta: 19:45:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.220 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31420.0 tgs: 262 data_time: 0.73s time: 119.68s eta: 19:38:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32320.0 tgs: 270 data_time: 0.88s time: 119.65s eta: 19:36:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.203 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31755.0 tgs: 266 data_time: 0.79s time: 119.31s eta: 19:31:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.284 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31116.0 tgs: 260 data_time: 0.89s time: 119.54s eta: 19:31:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.246 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32297.0 tgs: 262 data_time: 0.78s time: 122.96s eta: 20:02:58 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.236 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32099.0 tgs: 267 data_time: 0.82s time: 119.89s eta: 19:30:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.254 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31943.0 tgs: 264 data_time: 0.88s time: 120.66s eta: 19:36:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.356 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32321.0 tgs: 269 data_time: 0.80s time: 119.91s eta: 19:27:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.204 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31685.0 tgs: 263 data_time: 1.09s time: 120.45s eta: 19:30:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.203 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31570.0 tgs: 264 data_time: 1.02s time: 119.34s eta: 19:17:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.228 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32271.0 tgs: 269 data_time: 0.95s time: 119.68s eta: 19:18:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31826.0 tgs: 267 data_time: 0.79s time: 119.17s eta: 19:12:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31152.0 tgs: 261 data_time: 0.70s time: 119.22s eta: 19:10:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31903.0 tgs: 268 data_time: 0.81s time: 118.92s eta: 19:05:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.244 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30941.0 tgs: 259 data_time: 0.63s time: 119.15s eta: 19:05:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 29085.0 tgs: 244 data_time: 0.62s time: 119.17s eta: 19:04:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32477.0 tgs: 272 data_time: 1.04s time: 119.40s eta: 19:04:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31064.0 tgs: 260 data_time: 0.69s time: 119.33s eta: 19:01:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31960.0 tgs: 265 data_time: 0.92s time: 120.31s eta: 19:08:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31106.0 tgs: 253 data_time: 0.90s time: 122.91s eta: 19:31:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.225 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31337.0 tgs: 262 data_time: 0.94s time: 119.48s eta: 18:57:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.285 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31611.0 tgs: 265 data_time: 0.92s time: 119.16s eta: 18:51:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31573.0 tgs: 264 data_time: 0.63s time: 119.25s eta: 18:50:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.202 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31897.0 tgs: 267 data_time: 0.98s time: 119.39s eta: 18:50:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32058.0 tgs: 267 data_time: 0.88s time: 119.66s eta: 18:50:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31119.0 tgs: 260 data_time: 0.76s time: 119.38s eta: 18:46:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31356.0 tgs: 261 data_time: 0.54s time: 119.79s eta: 18:48:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.243 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32225.0 tgs: 249 data_time: 1.00s time: 129.32s eta: 20:15:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31625.0 tgs: 264 data_time: 0.71s time: 119.56s eta: 18:41:51 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31942.0 tgs: 267 data_time: 0.82s time: 119.39s eta: 18:38:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32061.0 tgs: 265 data_time: 0.84s time: 120.66s eta: 18:48:10 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30275.0 tgs: 253 data_time: 0.89s time: 119.39s eta: 18:34:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.232 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31108.0 tgs: 261 data_time: 0.74s time: 119.16s eta: 18:30:09 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 268 data_time: 0.73s time: 119.07s eta: 18:27:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 29801.0 tgs: 250 data_time: 0.91s time: 119.14s eta: 18:26:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31102.0 tgs: 258 data_time: 0.77s time: 120.39s eta: 18:35:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32173.0 tgs: 265 data_time: 0.74s time: 121.18s eta: 18:40:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31492.0 tgs: 261 data_time: 0.84s time: 120.66s eta: 18:34:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31539.0 tgs: 263 data_time: 0.84s time: 119.74s eta: 18:23:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:26:51][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.226 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32436.0 tgs: 261 data_time: 1.12s time: 123.87s eta: 18:59:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32092.0 tgs: 268 data_time: 0.66s time: 119.56s eta: 18:17:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31585.0 tgs: 264 data_time: 0.82s time: 119.34s eta: 18:13:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30729.0 tgs: 257 data_time: 0.44s time: 119.22s eta: 18:10:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.126 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31914.0 tgs: 266 data_time: 0.70s time: 119.60s eta: 18:12:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.90s time: 119.28s eta: 18:07:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31767.0 tgs: 266 data_time: 0.89s time: 119.01s eta: 18:03:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31842.0 tgs: 267 data_time: 0.84s time: 119.09s eta: 18:01:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31119.0 tgs: 260 data_time: 0.83s time: 119.26s eta: 18:01:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.310 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31541.0 tgs: 264 data_time: 1.16s time: 119.33s eta: 17:59:56 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31567.0 tgs: 262 data_time: 0.75s time: 120.03s eta: 18:04:14 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32276.0 tgs: 270 data_time: 0.92s time: 119.10s eta: 17:53:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.131 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31104.0 tgs: 260 data_time: 0.66s time: 119.41s eta: 17:54:39 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32112.0 tgs: 269 data_time: 0.73s time: 119.00s eta: 17:48:58 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31195.0 tgs: 262 data_time: 0.71s time: 119.02s eta: 17:47:14 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31401.0 tgs: 250 data_time: 0.60s time: 125.37s eta: 18:42:02 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 260 data_time: 0.78s time: 120.10s eta: 17:52:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31585.0 tgs: 264 data_time: 0.79s time: 119.38s eta: 17:44:30 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31441.0 tgs: 263 data_time: 0.76s time: 119.34s eta: 17:42:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 269 data_time: 1.00s time: 119.51s eta: 17:41:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32203.0 tgs: 255 data_time: 0.79s time: 126.06s eta: 18:37:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31194.0 tgs: 254 data_time: 0.79s time: 122.63s eta: 18:05:15 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31045.0 tgs: 256 data_time: 0.72s time: 121.16s eta: 17:50:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.229 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32341.0 tgs: 271 data_time: 0.76s time: 119.04s eta: 17:29:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32008.0 tgs: 269 data_time: 0.73s time: 118.87s eta: 17:26:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31320.0 tgs: 252 data_time: 1.15s time: 123.98s eta: 18:08:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31671.0 tgs: 254 data_time: 0.39s time: 124.57s eta: 18:12:02 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31891.0 tgs: 267 data_time: 0.70s time: 119.39s eta: 17:24:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31509.0 tgs: 263 data_time: 0.91s time: 119.70s eta: 17:25:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.219 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31543.0 tgs: 263 data_time: 0.75s time: 119.53s eta: 17:21:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31409.0 tgs: 264 data_time: 0.70s time: 118.86s eta: 17:14:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31460.0 tgs: 262 data_time: 0.77s time: 119.85s eta: 17:20:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32375.0 tgs: 270 data_time: 0.76s time: 119.54s eta: 17:16:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 263 data_time: 0.90s time: 121.86s eta: 17:34:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31021.0 tgs: 251 data_time: 0.59s time: 123.52s eta: 17:46:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.263 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31677.0 tgs: 266 data_time: 0.51s time: 119.01s eta: 17:05:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31283.0 tgs: 262 data_time: 0.70s time: 119.23s eta: 17:05:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31695.0 tgs: 266 data_time: 0.64s time: 119.14s eta: 17:02:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.203 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31430.0 tgs: 263 data_time: 0.69s time: 119.26s eta: 17:01:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31860.0 tgs: 267 data_time: 0.63s time: 119.04s eta: 16:57:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.120 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31735.0 tgs: 266 data_time: 0.87s time: 119.28s eta: 16:57:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32162.0 tgs: 269 data_time: 0.60s time: 119.17s eta: 16:54:58 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32305.0 tgs: 270 data_time: 0.83s time: 119.57s eta: 16:56:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31604.0 tgs: 265 data_time: 0.63s time: 119.13s eta: 16:50:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31021.0 tgs: 261 data_time: 0.72s time: 118.85s eta: 16:46:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30926.0 tgs: 259 data_time: 0.80s time: 119.01s eta: 16:45:39 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31731.0 tgs: 264 data_time: 0.59s time: 119.85s eta: 16:50:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32114.0 tgs: 268 data_time: 0.67s time: 119.39s eta: 16:44:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32395.0 tgs: 272 data_time: 0.71s time: 119.01s eta: 16:39:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32192.0 tgs: 266 data_time: 0.99s time: 120.93s eta: 16:53:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32117.0 tgs: 269 data_time: 0.88s time: 118.99s eta: 16:35:30 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30997.0 tgs: 259 data_time: 0.56s time: 119.29s eta: 16:36:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.208 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31290.0 tgs: 262 data_time: 0.84s time: 119.23s eta: 16:33:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31546.0 tgs: 264 data_time: 0.67s time: 119.37s eta: 16:32:44 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30543.0 tgs: 255 data_time: 0.87s time: 119.70s eta: 16:33:31 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 267 data_time: 1.09s time: 119.20s eta: 16:27:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.123 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31304.0 tgs: 261 data_time: 0.61s time: 119.76s eta: 16:30:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29923.0 tgs: 248 data_time: 0.92s time: 120.31s eta: 16:32:32 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31723.0 tgs: 266 data_time: 0.61s time: 118.97s eta: 16:19:32 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31481.0 tgs: 247 data_time: 0.77s time: 127.33s eta: 17:26:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30947.0 tgs: 259 data_time: 0.50s time: 119.23s eta: 16:17:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31715.0 tgs: 266 data_time: 0.84s time: 119.14s eta: 16:14:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31002.0 tgs: 260 data_time: 0.83s time: 119.24s eta: 16:13:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.213 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 264 data_time: 0.69s time: 119.17s eta: 16:11:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31512.0 tgs: 264 data_time: 0.67s time: 118.95s eta: 16:07:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 266 data_time: 0.59s time: 118.86s eta: 16:04:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32180.0 tgs: 270 data_time: 0.82s time: 118.96s eta: 16:03:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 264 data_time: 0.83s time: 121.84s eta: 16:24:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.117 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31857.0 tgs: 267 data_time: 0.81s time: 118.99s eta: 15:59:51 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 269 data_time: 0.83s time: 119.23s eta: 15:59:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.125 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30936.0 tgs: 258 data_time: 0.64s time: 119.62s eta: 16:00:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32371.0 tgs: 271 data_time: 0.70s time: 119.04s eta: 15:54:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31838.0 tgs: 267 data_time: 0.69s time: 118.98s eta: 15:51:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31824.0 tgs: 267 data_time: 1.01s time: 119.18s eta: 15:51:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.210 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31572.0 tgs: 263 data_time: 0.54s time: 119.64s eta: 15:53:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31084.0 tgs: 261 data_time: 0.60s time: 118.93s eta: 15:45:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.128 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32123.0 tgs: 269 data_time: 0.59s time: 119.29s eta: 15:46:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.181 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32223.0 tgs: 270 data_time: 0.79s time: 119.12s eta: 15:43:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32068.0 tgs: 269 data_time: 0.83s time: 119.00s eta: 15:40:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31424.0 tgs: 262 data_time: 0.64s time: 119.65s eta: 15:43:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32014.0 tgs: 263 data_time: 0.94s time: 121.50s eta: 15:55:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31957.0 tgs: 267 data_time: 0.93s time: 119.41s eta: 15:37:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32083.0 tgs: 246 data_time: 0.78s time: 129.96s eta: 16:58:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31479.0 tgs: 263 data_time: 0.98s time: 119.47s eta: 15:33:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30764.0 tgs: 257 data_time: 0.95s time: 119.60s eta: 15:32:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.156 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 269 data_time: 0.52s time: 119.53s eta: 15:30:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31377.0 tgs: 256 data_time: 0.78s time: 122.14s eta: 15:48:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.154 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32207.0 tgs: 266 data_time: 0.77s time: 120.75s eta: 15:35:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31143.0 tgs: 260 data_time: 0.71s time: 119.42s eta: 15:23:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.201 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32001.0 tgs: 267 data_time: 0.83s time: 119.44s eta: 15:21:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.226 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31599.0 tgs: 240 data_time: 1.11s time: 131.64s eta: 16:53:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32100.0 tgs: 269 data_time: 0.78s time: 119.29s eta: 15:16:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31853.0 tgs: 266 data_time: 0.57s time: 119.56s eta: 15:16:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31641.0 tgs: 258 data_time: 0.73s time: 122.25s eta: 15:35:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31967.0 tgs: 267 data_time: 0.78s time: 119.71s eta: 15:13:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.86s time: 119.35s eta: 15:09:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32339.0 tgs: 271 data_time: 0.83s time: 118.90s eta: 15:03:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32060.0 tgs: 269 data_time: 0.77s time: 118.99s eta: 15:02:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31794.0 tgs: 266 data_time: 0.97s time: 119.31s eta: 15:02:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30231.0 tgs: 254 data_time: 0.69s time: 118.92s eta: 14:57:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31939.0 tgs: 265 data_time: 0.76s time: 120.52s eta: 15:07:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.188 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31155.0 tgs: 261 data_time: 0.75s time: 119.32s eta: 14:56:51 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 260 data_time: 0.67s time: 122.65s eta: 15:19:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31398.0 tgs: 263 data_time: 0.52s time: 119.13s eta: 14:51:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.3GB text_tokens: 30027.0 tgs: 250 data_time: 0.74s time: 119.76s eta: 14:54:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.213 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31856.0 tgs: 266 data_time: 1.02s time: 119.46s eta: 14:50:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31765.0 tgs: 266 data_time: 0.59s time: 118.99s eta: 14:44:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 265 data_time: 0.95s time: 119.32s eta: 14:44:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31084.0 tgs: 261 data_time: 0.68s time: 118.94s eta: 14:40:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 268 data_time: 0.80s time: 119.05s eta: 14:39:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.205 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31449.0 tgs: 264 data_time: 0.69s time: 119.06s eta: 14:37:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32103.0 tgs: 269 data_time: 0.79s time: 119.13s eta: 14:35:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31945.0 tgs: 268 data_time: 1.00s time: 119.00s eta: 14:32:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30767.0 tgs: 258 data_time: 0.59s time: 119.07s eta: 14:31:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 263 data_time: 0.87s time: 119.15s eta: 14:29:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31573.0 tgs: 263 data_time: 0.82s time: 119.65s eta: 14:31:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.309 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30682.0 tgs: 257 data_time: 0.96s time: 119.18s eta: 14:26:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 268 data_time: 0.93s time: 119.20s eta: 14:24:10 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31619.0 tgs: 264 data_time: 0.78s time: 119.39s eta: 14:23:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31409.0 tgs: 260 data_time: 0.65s time: 120.73s eta: 14:31:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31745.0 tgs: 267 data_time: 0.57s time: 118.85s eta: 14:15:44 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31139.0 tgs: 261 data_time: 0.68s time: 119.10s eta: 14:15:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31695.0 tgs: 266 data_time: 0.84s time: 119.14s eta: 14:13:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31698.0 tgs: 264 data_time: 0.73s time: 119.96s eta: 14:17:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31865.0 tgs: 265 data_time: 0.66s time: 119.87s eta: 14:15:02 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 261 data_time: 0.87s time: 119.59s eta: 14:11:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.194 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32104.0 tgs: 269 data_time: 0.93s time: 119.22s eta: 14:06:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32166.0 tgs: 269 data_time: 0.72s time: 119.26s eta: 14:04:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.204 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31283.0 tgs: 262 data_time: 0.70s time: 119.27s eta: 14:02:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31863.0 tgs: 267 data_time: 0.66s time: 119.14s eta: 13:59:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31634.0 tgs: 265 data_time: 0.62s time: 119.12s eta: 13:57:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32170.0 tgs: 269 data_time: 0.92s time: 119.17s eta: 13:56:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.173 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31578.0 tgs: 264 data_time: 0.83s time: 119.35s eta: 13:55:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30931.0 tgs: 257 data_time: 0.55s time: 120.11s eta: 13:58:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.112 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30410.0 tgs: 248 data_time: 0.82s time: 122.56s eta: 14:13:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 265 data_time: 0.77s time: 119.10s eta: 13:47:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 28123.0 tgs: 235 data_time: 0.63s time: 119.37s eta: 13:47:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32503.0 tgs: 272 data_time: 0.78s time: 119.34s eta: 13:45:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32544.0 tgs: 272 data_time: 0.73s time: 119.34s eta: 13:43:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31532.0 tgs: 264 data_time: 0.70s time: 119.11s eta: 13:39:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31600.0 tgs: 264 data_time: 0.76s time: 119.25s eta: 13:38:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31796.0 tgs: 266 data_time: 0.89s time: 119.14s eta: 13:36:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 264 data_time: 0.75s time: 119.22s eta: 13:34:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31789.0 tgs: 266 data_time: 0.96s time: 119.45s eta: 13:34:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31573.0 tgs: 265 data_time: 0.91s time: 119.08s eta: 13:29:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.230 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.71s time: 118.92s eta: 13:26:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31531.0 tgs: 263 data_time: 0.46s time: 119.81s eta: 13:30:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.216 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32533.0 tgs: 272 data_time: 0.90s time: 119.25s eta: 13:24:58 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30958.0 tgs: 259 data_time: 0.76s time: 119.16s eta: 13:22:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31592.0 tgs: 264 data_time: 0.68s time: 119.35s eta: 13:21:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32567.0 tgs: 273 data_time: 0.66s time: 119.26s eta: 13:19:02 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31043.0 tgs: 260 data_time: 0.63s time: 119.04s eta: 13:15:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.187 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31096.0 tgs: 260 data_time: 0.71s time: 119.36s eta: 13:15:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32269.0 tgs: 270 data_time: 0.58s time: 119.18s eta: 13:12:32 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31543.0 tgs: 264 data_time: 0.72s time: 119.45s eta: 13:12:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31111.0 tgs: 260 data_time: 0.98s time: 119.29s eta: 13:09:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31565.0 tgs: 264 data_time: 0.77s time: 119.15s eta: 13:06:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.226 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31842.0 tgs: 264 data_time: 0.78s time: 120.42s eta: 13:12:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.118 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31191.0 tgs: 259 data_time: 0.46s time: 120.16s eta: 13:09:02 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32299.0 tgs: 270 data_time: 0.79s time: 119.50s eta: 13:02:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32546.0 tgs: 272 data_time: 0.86s time: 119.51s eta: 13:00:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.182 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31772.0 tgs: 265 data_time: 0.63s time: 119.57s eta: 12:59:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.214 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30728.0 tgs: 256 data_time: 0.63s time: 119.79s eta: 12:58:39 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.204 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31607.0 tgs: 265 data_time: 0.92s time: 119.04s eta: 12:51:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31143.0 tgs: 261 data_time: 0.77s time: 119.05s eta: 12:49:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.195 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32090.0 tgs: 269 data_time: 1.08s time: 119.12s eta: 12:48:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31240.0 tgs: 261 data_time: 0.59s time: 119.40s eta: 12:48:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.168 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31952.0 tgs: 268 data_time: 0.87s time: 119.19s eta: 12:44:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.200 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31361.0 tgs: 262 data_time: 0.46s time: 119.44s eta: 12:44:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31360.0 tgs: 263 data_time: 0.63s time: 119.09s eta: 12:40:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.193 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 269 data_time: 0.62s time: 119.07s eta: 12:38:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31045.0 tgs: 259 data_time: 0.66s time: 119.60s eta: 12:39:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.118 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30512.0 tgs: 256 data_time: 0.52s time: 118.78s eta: 12:32:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31380.0 tgs: 263 data_time: 0.64s time: 118.88s eta: 12:30:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.132 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32513.0 tgs: 272 data_time: 0.67s time: 119.16s eta: 12:30:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31177.0 tgs: 261 data_time: 0.64s time: 119.27s eta: 12:29:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.290 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31722.0 tgs: 265 data_time: 0.60s time: 119.35s eta: 12:27:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.217 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31523.0 tgs: 264 data_time: 0.84s time: 119.11s eta: 12:24:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31972.0 tgs: 267 data_time: 1.03s time: 119.34s eta: 12:23:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31702.0 tgs: 265 data_time: 0.65s time: 119.26s eta: 12:21:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.79s time: 119.03s eta: 12:17:58 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.213 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31564.0 tgs: 260 data_time: 0.59s time: 121.13s eta: 12:29:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31737.0 tgs: 264 data_time: 0.70s time: 119.99s eta: 12:19:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30726.0 tgs: 258 data_time: 0.56s time: 118.70s eta: 12:10:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31801.0 tgs: 265 data_time: 0.71s time: 119.63s eta: 12:13:44 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.109 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31699.0 tgs: 264 data_time: 0.71s time: 119.83s eta: 12:12:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31092.0 tgs: 260 data_time: 0.80s time: 119.14s eta: 12:06:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32141.0 tgs: 269 data_time: 0.67s time: 119.18s eta: 12:04:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.2GB text_tokens: 30077.0 tgs: 252 data_time: 0.57s time: 119.11s eta: 12:02:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31595.0 tgs: 265 data_time: 0.67s time: 119.18s eta: 12:01:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31495.0 tgs: 264 data_time: 0.76s time: 118.95s eta: 11:57:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.212 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31823.0 tgs: 265 data_time: 0.87s time: 119.70s eta: 12:00:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 271 data_time: 0.71s time: 119.27s eta: 11:55:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31428.0 tgs: 263 data_time: 0.89s time: 119.27s eta: 11:53:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 265 data_time: 0.82s time: 119.71s eta: 11:54:15 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32231.0 tgs: 271 data_time: 0.87s time: 118.78s eta: 11:46:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30750.0 tgs: 258 data_time: 0.65s time: 119.15s eta: 11:46:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 269 data_time: 0.71s time: 119.61s eta: 11:47:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32140.0 tgs: 270 data_time: 0.79s time: 118.86s eta: 11:41:15 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31937.0 tgs: 269 data_time: 0.73s time: 118.69s eta: 11:38:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31383.0 tgs: 263 data_time: 0.79s time: 119.21s eta: 11:39:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.205 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 270 data_time: 0.79s time: 119.14s eta: 11:36:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31957.0 tgs: 268 data_time: 0.63s time: 119.07s eta: 11:34:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31613.0 tgs: 265 data_time: 0.89s time: 118.95s eta: 11:31:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31294.0 tgs: 263 data_time: 0.79s time: 118.91s eta: 11:29:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31714.0 tgs: 266 data_time: 0.76s time: 118.97s eta: 11:28:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32010.0 tgs: 268 data_time: 0.70s time: 119.28s eta: 11:27:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31706.0 tgs: 265 data_time: 0.58s time: 119.35s eta: 11:26:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31687.0 tgs: 266 data_time: 0.63s time: 118.89s eta: 11:21:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31892.0 tgs: 267 data_time: 0.75s time: 119.20s eta: 11:21:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31923.0 tgs: 266 data_time: 0.53s time: 119.71s eta: 11:22:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.121 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30642.0 tgs: 257 data_time: 0.71s time: 119.12s eta: 11:17:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31304.0 tgs: 262 data_time: 0.85s time: 119.12s eta: 11:15:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30222.0 tgs: 254 data_time: 0.69s time: 118.95s eta: 11:12:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32183.0 tgs: 270 data_time: 0.83s time: 119.09s eta: 11:10:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 264 data_time: 0.78s time: 119.16s eta: 11:09:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30971.0 tgs: 260 data_time: 0.88s time: 118.97s eta: 11:06:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31802.0 tgs: 266 data_time: 0.60s time: 119.47s eta: 11:07:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.215 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 268 data_time: 0.75s time: 119.64s eta: 11:06:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31468.0 tgs: 263 data_time: 0.82s time: 119.30s eta: 11:02:07 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30564.0 tgs: 256 data_time: 0.67s time: 119.31s eta: 11:00:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32260.0 tgs: 271 data_time: 0.86s time: 118.82s eta: 10:55:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31636.0 tgs: 264 data_time: 0.89s time: 119.47s eta: 10:57:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.114 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31742.0 tgs: 265 data_time: 0.57s time: 119.37s eta: 10:54:31 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 267 data_time: 0.83s time: 119.28s eta: 10:52:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31543.0 tgs: 264 data_time: 0.84s time: 119.25s eta: 10:49:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31203.0 tgs: 261 data_time: 0.57s time: 119.10s eta: 10:47:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32156.0 tgs: 269 data_time: 0.60s time: 119.53s eta: 10:47:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.121 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30113.0 tgs: 252 data_time: 0.67s time: 119.34s eta: 10:44:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32047.0 tgs: 268 data_time: 0.86s time: 119.49s eta: 10:43:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.176 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31770.0 tgs: 265 data_time: 0.71s time: 119.51s eta: 10:41:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31679.0 tgs: 266 data_time: 0.86s time: 118.99s eta: 10:36:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.177 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 0.66s time: 119.22s eta: 10:35:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32066.0 tgs: 269 data_time: 0.60s time: 119.06s eta: 10:33:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.194 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32237.0 tgs: 271 data_time: 0.77s time: 118.78s eta: 10:29:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.197 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32091.0 tgs: 267 data_time: 1.04s time: 119.86s eta: 10:33:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.204 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31105.0 tgs: 260 data_time: 0.70s time: 119.37s eta: 10:28:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31678.0 tgs: 266 data_time: 0.64s time: 118.84s eta: 10:23:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32063.0 tgs: 269 data_time: 0.80s time: 118.98s eta: 10:22:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31238.0 tgs: 261 data_time: 0.87s time: 119.28s eta: 10:22:15 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.96s time: 119.20s eta: 10:19:51 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31036.0 tgs: 260 data_time: 0.60s time: 119.07s eta: 10:17:09 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32673.0 tgs: 273 data_time: 0.68s time: 119.47s eta: 10:17:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 268 data_time: 0.57s time: 119.38s eta: 10:14:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.200 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31813.0 tgs: 266 data_time: 0.78s time: 119.45s eta: 10:13:09 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31265.0 tgs: 262 data_time: 0.80s time: 119.29s eta: 10:10:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.227 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29916.0 tgs: 251 data_time: 0.61s time: 119.11s eta: 10:07:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31398.0 tgs: 262 data_time: 0.67s time: 119.70s eta: 10:08:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31790.0 tgs: 265 data_time: 0.59s time: 119.83s eta: 10:07:07 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.132 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30621.0 tgs: 251 data_time: 0.53s time: 121.55s eta: 10:13:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.205 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31482.0 tgs: 263 data_time: 0.79s time: 119.25s eta: 10:00:14 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30887.0 tgs: 259 data_time: 0.69s time: 118.81s eta: 9:56:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.159 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31016.0 tgs: 259 data_time: 0.88s time: 119.56s eta: 9:57:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31567.0 tgs: 263 data_time: 0.65s time: 119.70s eta: 9:56:31 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.116 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31464.0 tgs: 263 data_time: 0.57s time: 119.26s eta: 9:52:19 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31591.0 tgs: 264 data_time: 0.79s time: 119.33s eta: 9:50:39 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.179 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31766.0 tgs: 266 data_time: 0.64s time: 119.22s eta: 9:48:10 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31211.0 tgs: 258 data_time: 0.80s time: 120.87s eta: 9:54:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32200.0 tgs: 269 data_time: 0.61s time: 119.28s eta: 9:44:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.173 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31498.0 tgs: 263 data_time: 0.75s time: 119.46s eta: 9:43:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31207.0 tgs: 261 data_time: 0.70s time: 119.55s eta: 9:41:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30631.0 tgs: 256 data_time: 0.85s time: 119.20s eta: 9:38:07 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31489.0 tgs: 263 data_time: 0.69s time: 119.48s eta: 9:37:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32287.0 tgs: 271 data_time: 0.62s time: 119.05s eta: 9:33:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31601.0 tgs: 265 data_time: 0.66s time: 119.24s eta: 9:32:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31121.0 tgs: 261 data_time: 0.57s time: 119.16s eta: 9:29:58 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.121 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31063.0 tgs: 260 data_time: 0.77s time: 119.22s eta: 9:28:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30929.0 tgs: 258 data_time: 0.69s time: 119.46s eta: 9:27:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31038.0 tgs: 260 data_time: 0.58s time: 118.98s eta: 9:23:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31962.0 tgs: 268 data_time: 0.96s time: 119.04s eta: 9:21:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32040.0 tgs: 268 data_time: 0.66s time: 119.53s eta: 9:21:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32110.0 tgs: 268 data_time: 0.56s time: 119.59s eta: 9:20:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32125.0 tgs: 269 data_time: 0.52s time: 119.13s eta: 9:15:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.117 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30865.0 tgs: 259 data_time: 0.56s time: 119.01s eta: 9:13:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30444.0 tgs: 256 data_time: 0.65s time: 118.88s eta: 9:10:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 266 data_time: 0.69s time: 119.33s eta: 9:10:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.201 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32554.0 tgs: 273 data_time: 0.65s time: 119.09s eta: 9:07:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31838.0 tgs: 267 data_time: 0.56s time: 118.86s eta: 9:04:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.189 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31585.0 tgs: 263 data_time: 0.82s time: 119.66s eta: 9:06:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32065.0 tgs: 268 data_time: 0.58s time: 119.43s eta: 9:03:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32365.0 tgs: 270 data_time: 0.73s time: 119.43s eta: 9:01:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31530.0 tgs: 265 data_time: 0.60s time: 118.74s eta: 8:56:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31843.0 tgs: 267 data_time: 0.86s time: 118.87s eta: 8:54:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 266 data_time: 0.96s time: 119.34s eta: 8:55:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31883.0 tgs: 267 data_time: 0.83s time: 119.14s eta: 8:52:09 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31263.0 tgs: 262 data_time: 0.60s time: 119.27s eta: 8:50:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31502.0 tgs: 261 data_time: 0.62s time: 120.47s eta: 8:54:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 265 data_time: 0.60s time: 119.24s eta: 8:46:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.215 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32212.0 tgs: 270 data_time: 1.00s time: 119.09s eta: 8:43:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32265.0 tgs: 267 data_time: 0.81s time: 120.83s eta: 8:49:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31501.0 tgs: 262 data_time: 0.76s time: 120.20s eta: 8:44:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31740.0 tgs: 267 data_time: 0.72s time: 118.82s eta: 8:36:51 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32142.0 tgs: 269 data_time: 0.95s time: 119.35s eta: 8:37:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.176 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31537.0 tgs: 264 data_time: 0.78s time: 119.33s eta: 8:35:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 268 data_time: 0.59s time: 119.05s eta: 8:31:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.200 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31323.0 tgs: 262 data_time: 0.82s time: 119.16s eta: 8:30:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30331.0 tgs: 254 data_time: 0.59s time: 119.04s eta: 8:27:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 267 data_time: 0.55s time: 119.12s eta: 8:26:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.179 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31917.0 tgs: 267 data_time: 0.63s time: 119.24s eta: 8:24:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31288.0 tgs: 262 data_time: 0.90s time: 119.22s eta: 8:22:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.125 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30457.0 tgs: 255 data_time: 0.83s time: 119.38s eta: 8:21:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.177 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 266 data_time: 1.11s time: 119.78s eta: 8:21:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31161.0 tgs: 261 data_time: 0.85s time: 119.21s eta: 8:16:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32130.0 tgs: 270 data_time: 0.59s time: 118.99s eta: 8:13:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31143.0 tgs: 259 data_time: 0.81s time: 119.90s eta: 8:15:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31602.0 tgs: 264 data_time: 0.54s time: 119.56s eta: 8:12:10 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32321.0 tgs: 270 data_time: 0.70s time: 119.59s eta: 8:10:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32436.0 tgs: 272 data_time: 0.81s time: 119.00s eta: 8:05:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31579.0 tgs: 263 data_time: 0.88s time: 119.84s eta: 8:07:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32370.0 tgs: 270 data_time: 1.04s time: 119.70s eta: 8:04:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31525.0 tgs: 264 data_time: 0.70s time: 119.41s eta: 8:01:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31237.0 tgs: 261 data_time: 0.66s time: 119.28s eta: 7:59:07 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.190 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31094.0 tgs: 260 data_time: 0.55s time: 119.19s eta: 7:56:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30446.0 tgs: 255 data_time: 0.62s time: 119.20s eta: 7:54:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.231 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31584.0 tgs: 265 data_time: 0.70s time: 118.94s eta: 7:51:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32064.0 tgs: 269 data_time: 0.71s time: 118.97s eta: 7:49:56 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.179 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 271 data_time: 0.80s time: 119.41s eta: 7:49:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31385.0 tgs: 263 data_time: 0.44s time: 119.04s eta: 7:46:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31223.0 tgs: 262 data_time: 1.01s time: 119.11s eta: 7:44:31 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.099 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29144.0 tgs: 242 data_time: 0.39s time: 120.00s eta: 7:46:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31991.0 tgs: 266 data_time: 0.66s time: 119.83s eta: 7:43:19 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.123 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31446.0 tgs: 263 data_time: 0.81s time: 119.16s eta: 7:38:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 267 data_time: 0.75s time: 118.97s eta: 7:36:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32275.0 tgs: 270 data_time: 0.73s time: 119.33s eta: 7:35:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31450.0 tgs: 263 data_time: 0.71s time: 119.38s eta: 7:33:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.116 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30569.0 tgs: 256 data_time: 0.60s time: 119.07s eta: 7:30:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31563.0 tgs: 265 data_time: 0.61s time: 118.99s eta: 7:28:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 266 data_time: 0.65s time: 119.24s eta: 7:27:09 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30549.0 tgs: 256 data_time: 0.55s time: 118.99s eta: 7:24:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30501.0 tgs: 256 data_time: 0.94s time: 118.93s eta: 7:22:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31415.0 tgs: 263 data_time: 0.61s time: 119.08s eta: 7:20:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 270 data_time: 0.67s time: 119.00s eta: 7:18:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31977.0 tgs: 268 data_time: 0.72s time: 119.02s eta: 7:16:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31394.0 tgs: 263 data_time: 0.72s time: 119.24s eta: 7:15:14 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.169 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32162.0 tgs: 269 data_time: 0.70s time: 119.36s eta: 7:13:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.201 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32229.0 tgs: 270 data_time: 0.95s time: 118.94s eta: 7:10:10 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.163 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31850.0 tgs: 267 data_time: 0.92s time: 119.20s eta: 7:09:07 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31664.0 tgs: 264 data_time: 0.65s time: 119.59s eta: 7:08:30 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.209 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31949.0 tgs: 266 data_time: 1.12s time: 119.66s eta: 7:06:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32366.0 tgs: 272 data_time: 0.83s time: 118.95s eta: 7:02:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32393.0 tgs: 271 data_time: 0.83s time: 119.30s eta: 7:01:31 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32115.0 tgs: 268 data_time: 0.86s time: 119.51s eta: 7:00:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 265 data_time: 0.58s time: 119.42s eta: 6:57:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30729.0 tgs: 258 data_time: 0.95s time: 118.91s eta: 6:54:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31082.0 tgs: 260 data_time: 0.92s time: 119.35s eta: 6:53:44 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 267 data_time: 0.49s time: 118.93s eta: 6:50:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31354.0 tgs: 262 data_time: 0.81s time: 119.26s eta: 6:49:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.118 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32196.0 tgs: 270 data_time: 0.67s time: 119.02s eta: 6:46:39 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32244.0 tgs: 270 data_time: 0.92s time: 119.23s eta: 6:45:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31104.0 tgs: 260 data_time: 0.78s time: 119.40s eta: 6:43:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.121 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31758.0 tgs: 266 data_time: 0.86s time: 119.01s eta: 6:40:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.202 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.71s time: 119.01s eta: 6:38:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.173 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32059.0 tgs: 268 data_time: 0.64s time: 119.45s eta: 6:38:10 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.199 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31353.0 tgs: 263 data_time: 0.81s time: 118.80s eta: 6:34:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.130 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31972.0 tgs: 267 data_time: 0.67s time: 119.44s eta: 6:34:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32366.0 tgs: 270 data_time: 0.81s time: 119.57s eta: 6:32:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.276 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31755.0 tgs: 266 data_time: 0.85s time: 119.38s eta: 6:29:58 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32404.0 tgs: 270 data_time: 0.80s time: 119.71s eta: 6:29:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.192 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31955.0 tgs: 267 data_time: 0.65s time: 119.36s eta: 6:25:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31614.0 tgs: 266 data_time: 0.73s time: 118.82s eta: 6:22:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30900.0 tgs: 259 data_time: 0.64s time: 119.02s eta: 6:20:51 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31254.0 tgs: 261 data_time: 0.69s time: 119.40s eta: 6:20:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.161 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 269 data_time: 0.62s time: 119.34s eta: 6:17:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.221 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31294.0 tgs: 262 data_time: 0.73s time: 119.05s eta: 6:15:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31826.0 tgs: 266 data_time: 0.95s time: 119.40s eta: 6:14:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31203.0 tgs: 261 data_time: 0.81s time: 119.39s eta: 6:12:05 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.193 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 270 data_time: 0.64s time: 119.29s eta: 6:09:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.74s time: 119.09s eta: 6:07:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32002.0 tgs: 262 data_time: 0.80s time: 121.96s eta: 6:14:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30972.0 tgs: 259 data_time: 0.59s time: 119.38s eta: 6:04:07 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32374.0 tgs: 270 data_time: 0.76s time: 119.57s eta: 6:02:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31817.0 tgs: 266 data_time: 0.79s time: 119.26s eta: 5:59:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32097.0 tgs: 269 data_time: 0.85s time: 119.26s eta: 5:57:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.190 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31825.0 tgs: 267 data_time: 0.65s time: 118.93s eta: 5:54:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.196 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32384.0 tgs: 272 data_time: 0.57s time: 119.01s eta: 5:53:03 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.189 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32222.0 tgs: 270 data_time: 0.67s time: 119.07s eta: 5:51:15 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31756.0 tgs: 266 data_time: 0.64s time: 119.29s eta: 5:49:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32129.0 tgs: 270 data_time: 0.73s time: 118.97s eta: 5:46:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.192 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31881.0 tgs: 267 data_time: 0.71s time: 119.27s eta: 5:45:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.183 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32168.0 tgs: 269 data_time: 0.74s time: 119.35s eta: 5:44:07 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31865.0 tgs: 267 data_time: 0.57s time: 119.05s eta: 5:41:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30957.0 tgs: 259 data_time: 0.54s time: 119.49s eta: 5:40:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.198 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31644.0 tgs: 265 data_time: 0.76s time: 119.39s eta: 5:38:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 264 data_time: 0.81s time: 119.56s eta: 5:36:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31720.0 tgs: 266 data_time: 0.75s time: 118.98s eta: 5:33:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31219.0 tgs: 262 data_time: 0.80s time: 118.95s eta: 5:31:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 269 data_time: 0.64s time: 119.24s eta: 5:29:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.203 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31411.0 tgs: 263 data_time: 0.67s time: 119.41s eta: 5:28:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.188 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31174.0 tgs: 260 data_time: 0.78s time: 119.46s eta: 5:26:30 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31866.0 tgs: 267 data_time: 0.60s time: 118.95s eta: 5:23:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31577.0 tgs: 263 data_time: 0.71s time: 119.63s eta: 5:22:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.161 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31278.0 tgs: 262 data_time: 0.89s time: 119.35s eta: 5:20:14 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31963.0 tgs: 268 data_time: 0.53s time: 119.12s eta: 5:17:39 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31226.0 tgs: 261 data_time: 0.86s time: 119.42s eta: 5:16:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.190 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30283.0 tgs: 253 data_time: 0.46s time: 119.58s eta: 5:14:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31070.0 tgs: 261 data_time: 0.71s time: 119.00s eta: 5:11:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.178 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30568.0 tgs: 256 data_time: 0.66s time: 119.00s eta: 5:09:23 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31675.0 tgs: 265 data_time: 0.63s time: 119.23s eta: 5:08:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31086.0 tgs: 259 data_time: 0.77s time: 119.68s eta: 5:07:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31177.0 tgs: 261 data_time: 0.66s time: 119.11s eta: 5:03:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.192 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32290.0 tgs: 271 data_time: 0.66s time: 119.07s eta: 5:01:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31867.0 tgs: 262 data_time: 0.69s time: 121.52s eta: 5:05:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.126 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32000.0 tgs: 267 data_time: 0.67s time: 119.58s eta: 4:58:56 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.114 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31542.0 tgs: 263 data_time: 0.67s time: 119.50s eta: 4:56:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31591.0 tgs: 263 data_time: 0.73s time: 119.73s eta: 4:55:19 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30698.0 tgs: 256 data_time: 0.74s time: 119.52s eta: 4:52:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.177 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32464.0 tgs: 271 data_time: 0.59s time: 119.41s eta: 4:50:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31364.0 tgs: 263 data_time: 0.63s time: 119.20s eta: 4:48:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 269 data_time: 0.84s time: 118.89s eta: 4:45:19 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31698.0 tgs: 266 data_time: 0.58s time: 119.08s eta: 4:43:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31021.0 tgs: 258 data_time: 0.74s time: 119.98s eta: 4:43:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30787.0 tgs: 258 data_time: 0.90s time: 119.15s eta: 4:39:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 269 data_time: 0.62s time: 119.01s eta: 4:37:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32058.0 tgs: 269 data_time: 0.83s time: 119.06s eta: 4:35:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31562.0 tgs: 264 data_time: 0.79s time: 119.37s eta: 4:34:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.112 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31301.0 tgs: 262 data_time: 0.78s time: 119.41s eta: 4:32:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31994.0 tgs: 268 data_time: 0.90s time: 118.94s eta: 4:29:35 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31561.0 tgs: 263 data_time: 0.49s time: 119.60s eta: 4:29:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32218.0 tgs: 270 data_time: 0.80s time: 119.24s eta: 4:26:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31243.0 tgs: 262 data_time: 0.60s time: 119.06s eta: 4:23:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31943.0 tgs: 268 data_time: 0.66s time: 118.85s eta: 4:21:27 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31236.0 tgs: 263 data_time: 0.74s time: 118.76s eta: 4:19:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.194 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31514.0 tgs: 263 data_time: 0.58s time: 119.43s eta: 4:18:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.194 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32078.0 tgs: 268 data_time: 0.79s time: 119.40s eta: 4:16:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32077.0 tgs: 268 data_time: 1.03s time: 119.36s eta: 4:14:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30994.0 tgs: 260 data_time: 0.86s time: 119.05s eta: 4:11:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32248.0 tgs: 270 data_time: 0.73s time: 119.12s eta: 4:10:09 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31188.0 tgs: 261 data_time: 0.80s time: 119.13s eta: 4:08:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 270 data_time: 0.86s time: 118.95s eta: 4:05:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31711.0 tgs: 266 data_time: 0.70s time: 119.15s eta: 4:04:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31054.0 tgs: 259 data_time: 0.99s time: 119.50s eta: 4:02:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31011.0 tgs: 260 data_time: 0.84s time: 119.24s eta: 4:00:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30871.0 tgs: 259 data_time: 0.80s time: 119.19s eta: 3:58:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32004.0 tgs: 268 data_time: 0.60s time: 119.36s eta: 3:56:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.111 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32193.0 tgs: 269 data_time: 0.65s time: 119.38s eta: 3:54:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32213.0 tgs: 268 data_time: 0.70s time: 119.95s eta: 3:53:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31726.0 tgs: 265 data_time: 0.68s time: 119.32s eta: 3:50:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31869.0 tgs: 266 data_time: 0.68s time: 119.63s eta: 3:49:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.210 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 271 data_time: 0.65s time: 118.86s eta: 3:45:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31024.0 tgs: 259 data_time: 0.80s time: 119.72s eta: 3:45:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 268 data_time: 0.64s time: 119.33s eta: 3:42:44 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31887.0 tgs: 267 data_time: 0.67s time: 119.27s eta: 3:40:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 265 data_time: 0.67s time: 119.00s eta: 3:38:09 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31690.0 tgs: 265 data_time: 0.70s time: 119.43s eta: 3:36:57 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32030.0 tgs: 269 data_time: 0.69s time: 119.03s eta: 3:34:15 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30510.0 tgs: 256 data_time: 0.66s time: 118.99s eta: 3:32:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32465.0 tgs: 272 data_time: 0.94s time: 119.22s eta: 3:30:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31518.0 tgs: 262 data_time: 0.87s time: 120.11s eta: 3:30:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32288.0 tgs: 271 data_time: 0.91s time: 119.13s eta: 3:26:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.112 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 262 data_time: 0.69s time: 119.31s eta: 3:24:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31498.0 tgs: 264 data_time: 0.71s time: 119.08s eta: 3:22:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.212 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32067.0 tgs: 268 data_time: 0.95s time: 119.43s eta: 3:21:02 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31071.0 tgs: 260 data_time: 0.66s time: 119.29s eta: 3:18:48 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30698.0 tgs: 257 data_time: 0.83s time: 119.22s eta: 3:16:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32057.0 tgs: 269 data_time: 1.19s time: 119.15s eta: 3:14:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31250.0 tgs: 262 data_time: 1.00s time: 119.02s eta: 3:12:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31309.0 tgs: 262 data_time: 0.71s time: 119.18s eta: 3:10:41 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31128.0 tgs: 257 data_time: 0.79s time: 121.08s eta: 3:11:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32181.0 tgs: 269 data_time: 0.75s time: 119.51s eta: 3:07:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30468.0 tgs: 255 data_time: 0.52s time: 119.26s eta: 3:04:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31901.0 tgs: 268 data_time: 0.67s time: 118.99s eta: 3:02:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31201.0 tgs: 261 data_time: 0.67s time: 119.22s eta: 3:00:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31439.0 tgs: 263 data_time: 0.67s time: 119.46s eta: 2:59:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31340.0 tgs: 263 data_time: 0.60s time: 118.98s eta: 2:56:28 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 262 data_time: 0.72s time: 119.35s eta: 2:55:02 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31670.0 tgs: 265 data_time: 0.78s time: 119.41s eta: 2:53:08 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31880.0 tgs: 267 data_time: 0.67s time: 119.35s eta: 2:51:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30258.0 tgs: 249 data_time: 0.85s time: 121.46s eta: 2:52:04 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.191 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 31186.0 tgs: 261 data_time: 0.78s time: 119.25s eta: 2:46:56 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31634.0 tgs: 264 data_time: 0.72s time: 119.59s eta: 2:45:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.102 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31792.0 tgs: 267 data_time: 0.79s time: 118.90s eta: 2:42:30 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 262 data_time: 0.93s time: 119.18s eta: 2:40:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29697.0 tgs: 249 data_time: 0.52s time: 118.96s eta: 2:38:37 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29970.0 tgs: 251 data_time: 0.86s time: 119.23s eta: 2:36:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31873.0 tgs: 267 data_time: 0.93s time: 119.11s eta: 2:34:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32327.0 tgs: 270 data_time: 0.96s time: 119.43s eta: 2:33:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.195 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 265 data_time: 0.73s time: 120.22s eta: 2:32:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31477.0 tgs: 263 data_time: 0.72s time: 119.28s eta: 2:29:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31438.0 tgs: 264 data_time: 0.84s time: 118.88s eta: 2:26:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30928.0 tgs: 258 data_time: 0.76s time: 119.52s eta: 2:25:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.097 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31563.0 tgs: 265 data_time: 0.82s time: 119.06s eta: 2:22:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 28841.0 tgs: 242 data_time: 0.49s time: 119.14s eta: 2:20:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31586.0 tgs: 265 data_time: 0.73s time: 119.04s eta: 2:18:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30966.0 tgs: 258 data_time: 0.74s time: 119.85s eta: 2:17:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 264 data_time: 0.85s time: 119.31s eta: 2:15:13 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.099 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30931.0 tgs: 259 data_time: 0.70s time: 119.38s eta: 2:13:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32160.0 tgs: 268 data_time: 0.66s time: 119.68s eta: 2:11:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31921.0 tgs: 267 data_time: 0.75s time: 119.49s eta: 2:09:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 29647.0 tgs: 249 data_time: 0.65s time: 118.96s eta: 2:06:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32103.0 tgs: 268 data_time: 0.75s time: 119.37s eta: 2:05:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32209.0 tgs: 270 data_time: 0.61s time: 119.27s eta: 2:03:14 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 269 data_time: 0.72s time: 119.22s eta: 2:01:12 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32502.0 tgs: 271 data_time: 0.61s time: 119.50s eta: 1:59:30 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31039.0 tgs: 259 data_time: 0.54s time: 119.42s eta: 1:57:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32202.0 tgs: 269 data_time: 0.81s time: 119.53s eta: 1:55:32 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 261 data_time: 0.66s time: 119.30s eta: 1:53:20 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31901.0 tgs: 267 data_time: 0.93s time: 119.26s eta: 1:51:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 264 data_time: 0.68s time: 118.92s eta: 1:49:00 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31570.0 tgs: 263 data_time: 0.74s time: 119.64s eta: 1:47:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 265 data_time: 0.85s time: 119.72s eta: 1:45:45 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31853.0 tgs: 266 data_time: 0.70s time: 119.35s eta: 1:43:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31822.0 tgs: 266 data_time: 0.54s time: 119.30s eta: 1:41:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32163.0 tgs: 269 data_time: 0.75s time: 119.28s eta: 1:39:24 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31174.0 tgs: 262 data_time: 0.49s time: 118.90s eta: 1:37:06 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31797.0 tgs: 266 data_time: 0.59s time: 119.16s eta: 1:35:19 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31093.0 tgs: 261 data_time: 0.72s time: 119.07s eta: 1:33:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31055.0 tgs: 260 data_time: 0.76s time: 119.25s eta: 1:31:25 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32220.0 tgs: 269 data_time: 0.60s time: 119.73s eta: 1:29:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30017.0 tgs: 252 data_time: 0.66s time: 119.01s eta: 1:27:16 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31706.0 tgs: 266 data_time: 0.47s time: 118.87s eta: 1:25:11 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30728.0 tgs: 257 data_time: 0.70s time: 119.48s eta: 1:23:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31784.0 tgs: 266 data_time: 0.71s time: 119.48s eta: 1:21:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32315.0 tgs: 271 data_time: 0.64s time: 118.94s eta: 1:19:17 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31231.0 tgs: 262 data_time: 0.78s time: 118.93s eta: 1:17:18 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31895.0 tgs: 261 data_time: 0.56s time: 122.14s eta: 1:17:21 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 267 data_time: 0.76s time: 119.36s eta: 1:13:36 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30755.0 tgs: 258 data_time: 0.67s time: 119.06s eta: 1:11:26 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31483.0 tgs: 264 data_time: 0.68s time: 118.92s eta: 1:09:22 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.82s time: 119.11s eta: 1:07:29 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31077.0 tgs: 260 data_time: 0.86s time: 119.41s eta: 1:05:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31785.0 tgs: 266 data_time: 0.84s time: 119.34s eta: 1:03:39 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.198 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31792.0 tgs: 266 data_time: 0.73s time: 119.18s eta: 1:01:34 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31622.0 tgs: 265 data_time: 0.65s time: 119.27s eta: 0:59:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.162 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 269 data_time: 0.97s time: 119.44s eta: 0:57:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31637.0 tgs: 264 data_time: 0.57s time: 119.39s eta: 0:55:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31072.0 tgs: 260 data_time: 0.56s time: 119.48s eta: 0:53:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32553.0 tgs: 272 data_time: 0.61s time: 119.36s eta: 0:51:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31526.0 tgs: 263 data_time: 0.83s time: 119.47s eta: 0:49:46 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31327.0 tgs: 263 data_time: 0.81s time: 118.89s eta: 0:47:33 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32564.0 tgs: 273 data_time: 0.68s time: 119.28s eta: 0:45:43 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31478.0 tgs: 264 data_time: 0.62s time: 119.21s eta: 0:43:42 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31089.0 tgs: 261 data_time: 0.51s time: 118.99s eta: 0:41:38 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 267 data_time: 0.84s time: 119.04s eta: 0:39:40 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31638.0 tgs: 264 data_time: 0.70s time: 119.66s eta: 0:37:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31621.0 tgs: 265 data_time: 0.83s time: 119.28s eta: 0:35:47 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30405.0 tgs: 253 data_time: 0.61s time: 119.74s eta: 0:33:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32094.0 tgs: 264 data_time: 0.57s time: 121.22s eta: 0:32:19 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32277.0 tgs: 270 data_time: 0.72s time: 119.38s eta: 0:29:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.238 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31990.0 tgs: 268 data_time: 0.95s time: 119.32s eta: 0:27:50 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31243.0 tgs: 261 data_time: 0.86s time: 119.64s eta: 0:25:55 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32264.0 tgs: 270 data_time: 0.67s time: 119.09s eta: 0:23:49 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31470.0 tgs: 263 data_time: 0.80s time: 119.39s eta: 0:21:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32215.0 tgs: 270 data_time: 0.93s time: 119.26s eta: 0:19:52 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31770.0 tgs: 266 data_time: 0.55s time: 119.24s eta: 0:17:53 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.103 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31771.0 tgs: 264 data_time: 0.68s time: 119.90s eta: 0:15:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31510.0 tgs: 264 data_time: 0.61s time: 119.20s eta: 0:13:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.225 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31813.0 tgs: 267 data_time: 0.80s time: 119.09s eta: 0:11:54 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.250 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32150.0 tgs: 269 data_time: 1.09s time: 119.33s eta: 0:09:56 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.220 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 272 data_time: 0.80s time: 119.17s eta: 0:07:56 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31826.0 tgs: 265 data_time: 0.68s time: 119.69s eta: 0:05:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31863.0 tgs: 263 data_time: 0.93s time: 120.97s eta: 0:04:01 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.83s time: 119.21s eta: 0:01:59 +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.28 seconds, peak gpu memory 17.4G +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 53][DP 13][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank57.log b/20250121165312/rank57.log new file mode 100644 index 0000000000000000000000000000000000000000..c36a4ed6a3052db3caeba7aae7e2edffc3415198 --- /dev/null +++ b/20250121165312/rank57.log @@ -0,0 +1,851 @@ +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:19][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.03s +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:00:48][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 65.82 seconds, peak gpu memory 13.4G +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.257 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 32438.0 tgs: 212 data_time: 2.47s time: 152.92s eta: 1 day, 1:11:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.255 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31049.0 tgs: 258 data_time: 0.92s time: 120.15s eta: 19:45:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.210 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31400.0 tgs: 262 data_time: 1.05s time: 119.68s eta: 19:38:52 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31528.0 tgs: 263 data_time: 0.74s time: 119.65s eta: 19:36:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.324 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31808.0 tgs: 266 data_time: 0.98s time: 119.31s eta: 19:31:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.218 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32312.0 tgs: 270 data_time: 0.81s time: 119.53s eta: 19:31:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.246 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31349.0 tgs: 254 data_time: 0.75s time: 122.96s eta: 20:02:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.199 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31993.0 tgs: 266 data_time: 0.71s time: 119.89s eta: 19:30:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.181 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32365.0 tgs: 268 data_time: 0.69s time: 120.66s eta: 19:36:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.211 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31497.0 tgs: 262 data_time: 0.86s time: 119.91s eta: 19:27:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.173 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31507.0 tgs: 261 data_time: 0.84s time: 120.46s eta: 19:30:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.230 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 29899.0 tgs: 250 data_time: 0.79s time: 119.34s eta: 19:17:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 268 data_time: 0.85s time: 119.67s eta: 19:18:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.325 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31562.0 tgs: 264 data_time: 0.88s time: 119.20s eta: 19:12:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30596.0 tgs: 256 data_time: 0.82s time: 119.22s eta: 19:10:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31446.0 tgs: 264 data_time: 0.92s time: 118.92s eta: 19:05:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30935.0 tgs: 259 data_time: 0.62s time: 119.15s eta: 19:05:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.238 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31582.0 tgs: 264 data_time: 0.72s time: 119.18s eta: 19:04:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32037.0 tgs: 268 data_time: 0.78s time: 119.39s eta: 19:04:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32131.0 tgs: 269 data_time: 0.71s time: 119.33s eta: 19:01:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32342.0 tgs: 268 data_time: 0.73s time: 120.31s eta: 19:08:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31716.0 tgs: 258 data_time: 0.66s time: 122.92s eta: 19:31:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31956.0 tgs: 267 data_time: 0.87s time: 119.48s eta: 18:57:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.232 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32071.0 tgs: 269 data_time: 1.01s time: 119.16s eta: 18:51:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31325.0 tgs: 262 data_time: 0.71s time: 119.24s eta: 18:50:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30772.0 tgs: 257 data_time: 0.93s time: 119.39s eta: 18:50:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31620.0 tgs: 264 data_time: 1.03s time: 119.66s eta: 18:50:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.236 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31970.0 tgs: 267 data_time: 1.02s time: 119.39s eta: 18:46:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32184.0 tgs: 268 data_time: 0.62s time: 119.79s eta: 18:48:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31024.0 tgs: 239 data_time: 0.91s time: 129.30s eta: 20:15:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30697.0 tgs: 256 data_time: 0.95s time: 119.57s eta: 18:41:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 270 data_time: 0.77s time: 119.39s eta: 18:38:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31376.0 tgs: 260 data_time: 0.72s time: 120.66s eta: 18:48:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31253.0 tgs: 261 data_time: 0.67s time: 119.40s eta: 18:34:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.1GB text_tokens: 27839.0 tgs: 233 data_time: 0.68s time: 119.16s eta: 18:30:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31971.0 tgs: 268 data_time: 0.84s time: 119.07s eta: 18:27:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32205.0 tgs: 270 data_time: 0.75s time: 119.13s eta: 18:25:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30720.0 tgs: 255 data_time: 0.77s time: 120.39s eta: 18:35:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32017.0 tgs: 264 data_time: 0.89s time: 121.17s eta: 18:40:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.217 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31169.0 tgs: 258 data_time: 0.63s time: 120.67s eta: 18:34:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32306.0 tgs: 269 data_time: 0.63s time: 119.73s eta: 18:23:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30552.0 tgs: 246 data_time: 0.75s time: 123.87s eta: 18:59:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.3GB text_tokens: 27829.0 tgs: 232 data_time: 0.65s time: 119.56s eta: 18:17:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32389.0 tgs: 271 data_time: 0.59s time: 119.36s eta: 18:14:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31067.0 tgs: 260 data_time: 0.67s time: 119.22s eta: 18:10:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32547.0 tgs: 272 data_time: 0.81s time: 119.60s eta: 18:12:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32221.0 tgs: 270 data_time: 0.72s time: 119.27s eta: 18:07:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31075.0 tgs: 261 data_time: 0.55s time: 119.01s eta: 18:03:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30517.0 tgs: 256 data_time: 0.74s time: 119.10s eta: 18:01:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31792.0 tgs: 266 data_time: 0.82s time: 119.27s eta: 18:01:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31413.0 tgs: 263 data_time: 1.04s time: 119.33s eta: 17:59:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 264 data_time: 0.81s time: 120.02s eta: 18:04:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32130.0 tgs: 269 data_time: 0.65s time: 119.10s eta: 17:53:52 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31532.0 tgs: 264 data_time: 0.60s time: 119.42s eta: 17:54:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31076.0 tgs: 261 data_time: 0.81s time: 118.98s eta: 17:48:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31660.0 tgs: 265 data_time: 0.70s time: 119.02s eta: 17:47:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32304.0 tgs: 257 data_time: 0.66s time: 125.36s eta: 18:41:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.120 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 265 data_time: 0.85s time: 120.10s eta: 17:52:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31620.0 tgs: 264 data_time: 0.61s time: 119.40s eta: 17:44:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32321.0 tgs: 270 data_time: 0.89s time: 119.34s eta: 17:42:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32465.0 tgs: 271 data_time: 0.73s time: 119.51s eta: 17:41:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31067.0 tgs: 246 data_time: 0.71s time: 126.05s eta: 18:37:36 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31472.0 tgs: 256 data_time: 0.80s time: 122.63s eta: 18:05:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30640.0 tgs: 252 data_time: 1.09s time: 121.15s eta: 17:50:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32323.0 tgs: 271 data_time: 0.77s time: 119.03s eta: 17:29:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32138.0 tgs: 270 data_time: 0.83s time: 118.85s eta: 17:25:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31246.0 tgs: 251 data_time: 0.73s time: 124.00s eta: 18:09:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30657.0 tgs: 246 data_time: 0.62s time: 124.56s eta: 18:12:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 265 data_time: 0.98s time: 119.39s eta: 17:24:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31495.0 tgs: 263 data_time: 0.73s time: 119.70s eta: 17:25:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31015.0 tgs: 259 data_time: 0.89s time: 119.52s eta: 17:21:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.220 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31302.0 tgs: 263 data_time: 0.82s time: 118.86s eta: 17:14:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31112.0 tgs: 259 data_time: 0.92s time: 119.85s eta: 17:20:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32124.0 tgs: 268 data_time: 0.81s time: 119.56s eta: 17:16:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31810.0 tgs: 261 data_time: 0.83s time: 121.86s eta: 17:34:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30978.0 tgs: 250 data_time: 0.80s time: 123.52s eta: 17:46:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32134.0 tgs: 270 data_time: 0.70s time: 119.01s eta: 17:05:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31835.0 tgs: 267 data_time: 0.64s time: 119.23s eta: 17:05:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31732.0 tgs: 266 data_time: 0.55s time: 119.14s eta: 17:02:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31897.0 tgs: 267 data_time: 0.51s time: 119.26s eta: 17:01:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32299.0 tgs: 271 data_time: 0.83s time: 119.04s eta: 16:57:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31880.0 tgs: 267 data_time: 0.80s time: 119.27s eta: 16:57:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32075.0 tgs: 269 data_time: 0.90s time: 119.16s eta: 16:54:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31715.0 tgs: 265 data_time: 0.80s time: 119.58s eta: 16:56:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31024.0 tgs: 260 data_time: 0.83s time: 119.13s eta: 16:50:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31288.0 tgs: 263 data_time: 0.57s time: 118.84s eta: 16:46:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31610.0 tgs: 265 data_time: 0.65s time: 119.02s eta: 16:45:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31777.0 tgs: 265 data_time: 0.77s time: 119.84s eta: 16:50:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 267 data_time: 0.75s time: 119.37s eta: 16:44:44 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30906.0 tgs: 259 data_time: 0.97s time: 119.01s eta: 16:39:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32048.0 tgs: 265 data_time: 0.88s time: 120.93s eta: 16:53:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.222 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31894.0 tgs: 268 data_time: 0.83s time: 118.99s eta: 16:35:32 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31418.0 tgs: 263 data_time: 0.81s time: 119.28s eta: 16:36:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31229.0 tgs: 261 data_time: 0.99s time: 119.23s eta: 16:33:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31371.0 tgs: 262 data_time: 0.72s time: 119.37s eta: 16:32:44 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32260.0 tgs: 269 data_time: 0.74s time: 119.70s eta: 16:33:31 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32060.0 tgs: 268 data_time: 0.91s time: 119.20s eta: 16:27:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32428.0 tgs: 270 data_time: 0.75s time: 119.76s eta: 16:30:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31708.0 tgs: 263 data_time: 1.09s time: 120.31s eta: 16:32:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31572.0 tgs: 265 data_time: 0.81s time: 118.97s eta: 16:19:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 251 data_time: 0.68s time: 127.32s eta: 17:26:10 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.220 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32407.0 tgs: 271 data_time: 1.00s time: 119.23s eta: 16:17:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32016.0 tgs: 268 data_time: 0.56s time: 119.14s eta: 16:14:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31913.0 tgs: 267 data_time: 0.75s time: 119.26s eta: 16:13:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32080.0 tgs: 269 data_time: 0.83s time: 119.16s eta: 16:11:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31182.0 tgs: 262 data_time: 0.58s time: 118.96s eta: 16:07:30 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.216 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32211.0 tgs: 271 data_time: 0.69s time: 118.86s eta: 16:04:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.91s time: 118.96s eta: 16:03:32 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31739.0 tgs: 260 data_time: 0.66s time: 121.85s eta: 16:24:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.221 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31493.0 tgs: 264 data_time: 0.68s time: 119.00s eta: 15:59:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32255.0 tgs: 270 data_time: 0.79s time: 119.22s eta: 15:59:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.115 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31835.0 tgs: 266 data_time: 0.56s time: 119.61s eta: 16:00:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31708.0 tgs: 266 data_time: 0.95s time: 119.04s eta: 15:54:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31931.0 tgs: 268 data_time: 0.64s time: 118.98s eta: 15:51:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31816.0 tgs: 266 data_time: 0.89s time: 119.17s eta: 15:51:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30743.0 tgs: 256 data_time: 0.62s time: 119.63s eta: 15:53:03 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30786.0 tgs: 258 data_time: 0.82s time: 118.94s eta: 15:45:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31842.0 tgs: 266 data_time: 0.81s time: 119.28s eta: 15:46:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32072.0 tgs: 269 data_time: 0.88s time: 119.13s eta: 15:43:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31539.0 tgs: 265 data_time: 0.61s time: 118.99s eta: 15:40:03 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.126 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31953.0 tgs: 267 data_time: 0.55s time: 119.66s eta: 15:43:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31622.0 tgs: 260 data_time: 0.60s time: 121.51s eta: 15:55:52 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32409.0 tgs: 271 data_time: 0.88s time: 119.41s eta: 15:37:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31679.0 tgs: 243 data_time: 0.71s time: 129.97s eta: 16:58:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31942.0 tgs: 267 data_time: 0.80s time: 119.46s eta: 15:33:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32181.0 tgs: 269 data_time: 0.73s time: 119.61s eta: 15:32:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31715.0 tgs: 265 data_time: 0.76s time: 119.53s eta: 15:30:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.223 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31654.0 tgs: 259 data_time: 0.88s time: 122.14s eta: 15:48:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31703.0 tgs: 262 data_time: 0.92s time: 120.75s eta: 15:35:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.236 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28323.0 tgs: 237 data_time: 0.99s time: 119.41s eta: 15:23:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31298.0 tgs: 262 data_time: 0.62s time: 119.45s eta: 15:21:44 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31641.0 tgs: 240 data_time: 0.77s time: 131.63s eta: 16:53:32 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31755.0 tgs: 266 data_time: 0.90s time: 119.29s eta: 15:16:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32105.0 tgs: 268 data_time: 0.81s time: 119.56s eta: 15:16:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 260 data_time: 1.06s time: 122.25s eta: 15:35:10 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.093 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 26382.0 tgs: 220 data_time: 0.92s time: 119.72s eta: 15:13:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32036.0 tgs: 268 data_time: 0.69s time: 119.35s eta: 15:09:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 270 data_time: 0.78s time: 118.89s eta: 15:03:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31153.0 tgs: 261 data_time: 0.54s time: 118.99s eta: 15:02:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32017.0 tgs: 268 data_time: 0.94s time: 119.31s eta: 15:02:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31544.0 tgs: 265 data_time: 0.79s time: 118.92s eta: 14:57:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31342.0 tgs: 260 data_time: 0.84s time: 120.52s eta: 15:07:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 268 data_time: 0.90s time: 119.32s eta: 14:56:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30092.0 tgs: 245 data_time: 0.61s time: 122.65s eta: 15:19:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31458.0 tgs: 264 data_time: 0.73s time: 119.13s eta: 14:51:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.200 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30869.0 tgs: 257 data_time: 0.71s time: 119.75s eta: 14:54:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31878.0 tgs: 266 data_time: 0.73s time: 119.47s eta: 14:50:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30925.0 tgs: 259 data_time: 0.94s time: 118.98s eta: 14:44:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.80s time: 119.33s eta: 14:44:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31349.0 tgs: 263 data_time: 0.82s time: 118.94s eta: 14:40:10 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31145.0 tgs: 261 data_time: 0.65s time: 119.05s eta: 14:38:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31408.0 tgs: 263 data_time: 0.59s time: 119.06s eta: 14:37:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31681.0 tgs: 265 data_time: 0.62s time: 119.12s eta: 14:35:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 269 data_time: 0.77s time: 119.00s eta: 14:32:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32178.0 tgs: 270 data_time: 0.71s time: 119.07s eta: 14:31:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.188 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31292.0 tgs: 262 data_time: 0.82s time: 119.15s eta: 14:29:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31692.0 tgs: 264 data_time: 0.61s time: 119.65s eta: 14:31:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31540.0 tgs: 264 data_time: 0.67s time: 119.18s eta: 14:26:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31519.0 tgs: 264 data_time: 0.88s time: 119.20s eta: 14:24:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.198 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31517.0 tgs: 263 data_time: 0.88s time: 119.39s eta: 14:23:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32035.0 tgs: 265 data_time: 0.63s time: 120.73s eta: 14:31:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31474.0 tgs: 264 data_time: 0.61s time: 118.85s eta: 14:15:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31341.0 tgs: 263 data_time: 0.68s time: 119.10s eta: 14:15:31 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.127 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32377.0 tgs: 271 data_time: 0.87s time: 119.15s eta: 14:13:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 260 data_time: 0.78s time: 119.96s eta: 14:17:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31961.0 tgs: 266 data_time: 0.92s time: 119.87s eta: 14:15:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.116 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31440.0 tgs: 262 data_time: 0.64s time: 119.59s eta: 14:11:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31244.0 tgs: 262 data_time: 0.84s time: 119.22s eta: 14:06:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 269 data_time: 0.55s time: 119.26s eta: 14:04:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31294.0 tgs: 262 data_time: 0.87s time: 119.27s eta: 14:02:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31862.0 tgs: 267 data_time: 0.66s time: 119.14s eta: 13:59:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31303.0 tgs: 262 data_time: 0.74s time: 119.12s eta: 13:57:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30612.0 tgs: 256 data_time: 0.67s time: 119.17s eta: 13:56:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 268 data_time: 0.91s time: 119.35s eta: 13:55:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30969.0 tgs: 257 data_time: 0.80s time: 120.12s eta: 13:58:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.131 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31239.0 tgs: 254 data_time: 0.69s time: 122.55s eta: 14:13:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31799.0 tgs: 266 data_time: 0.62s time: 119.11s eta: 13:47:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31643.0 tgs: 265 data_time: 0.51s time: 119.36s eta: 13:47:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31391.0 tgs: 263 data_time: 0.71s time: 119.35s eta: 13:45:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.124 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31178.0 tgs: 261 data_time: 0.81s time: 119.33s eta: 13:43:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.191 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31666.0 tgs: 265 data_time: 0.71s time: 119.12s eta: 13:39:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31285.0 tgs: 262 data_time: 0.83s time: 119.25s eta: 13:38:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.143 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31121.0 tgs: 261 data_time: 0.81s time: 119.13s eta: 13:36:03 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 267 data_time: 0.87s time: 119.22s eta: 13:34:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.097 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30527.0 tgs: 255 data_time: 0.91s time: 119.46s eta: 13:34:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.123 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31229.0 tgs: 262 data_time: 0.73s time: 119.08s eta: 13:29:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31442.0 tgs: 264 data_time: 0.76s time: 118.92s eta: 13:26:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.115 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31015.0 tgs: 258 data_time: 0.72s time: 119.81s eta: 13:30:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32331.0 tgs: 271 data_time: 0.82s time: 119.25s eta: 13:24:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.207 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31941.0 tgs: 268 data_time: 0.89s time: 119.16s eta: 13:22:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.119 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31398.0 tgs: 263 data_time: 0.69s time: 119.34s eta: 13:21:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30333.0 tgs: 254 data_time: 0.55s time: 119.27s eta: 13:19:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.201 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31897.0 tgs: 267 data_time: 0.75s time: 119.04s eta: 13:15:36 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31407.0 tgs: 263 data_time: 0.57s time: 119.37s eta: 13:15:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31795.0 tgs: 266 data_time: 0.75s time: 119.20s eta: 13:12:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31786.0 tgs: 266 data_time: 0.79s time: 119.45s eta: 13:12:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.124 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31676.0 tgs: 265 data_time: 0.79s time: 119.29s eta: 13:09:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31898.0 tgs: 267 data_time: 0.62s time: 119.15s eta: 13:06:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31205.0 tgs: 259 data_time: 0.80s time: 120.30s eta: 13:11:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31691.0 tgs: 263 data_time: 0.74s time: 120.16s eta: 13:09:03 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31599.0 tgs: 264 data_time: 0.52s time: 119.50s eta: 13:02:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31284.0 tgs: 261 data_time: 0.62s time: 119.51s eta: 13:00:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31522.0 tgs: 263 data_time: 0.82s time: 119.57s eta: 12:59:12 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.268 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31899.0 tgs: 266 data_time: 0.78s time: 119.80s eta: 12:58:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30720.0 tgs: 258 data_time: 0.72s time: 119.04s eta: 12:51:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31413.0 tgs: 263 data_time: 0.72s time: 119.05s eta: 12:49:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.206 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31164.0 tgs: 261 data_time: 0.80s time: 119.12s eta: 12:48:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.208 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32244.0 tgs: 270 data_time: 0.59s time: 119.40s eta: 12:48:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.86s time: 119.17s eta: 12:44:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30757.0 tgs: 257 data_time: 0.78s time: 119.43s eta: 12:44:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.123 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31423.0 tgs: 263 data_time: 0.60s time: 119.10s eta: 12:40:14 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.193 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31516.0 tgs: 264 data_time: 0.74s time: 119.07s eta: 12:38:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31901.0 tgs: 266 data_time: 0.80s time: 119.59s eta: 12:39:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.171 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32348.0 tgs: 272 data_time: 0.58s time: 118.78s eta: 12:32:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.195 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 270 data_time: 0.92s time: 118.88s eta: 12:30:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31583.0 tgs: 265 data_time: 0.79s time: 119.17s eta: 12:30:44 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31576.0 tgs: 264 data_time: 0.67s time: 119.27s eta: 12:29:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31703.0 tgs: 265 data_time: 0.58s time: 119.35s eta: 12:27:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31491.0 tgs: 264 data_time: 0.48s time: 119.12s eta: 12:24:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.174 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29242.0 tgs: 245 data_time: 0.59s time: 119.34s eta: 12:23:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.212 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31835.0 tgs: 266 data_time: 1.02s time: 119.26s eta: 12:21:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.110 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32031.0 tgs: 269 data_time: 0.68s time: 119.03s eta: 12:18:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.195 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31600.0 tgs: 260 data_time: 0.51s time: 121.13s eta: 12:28:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30542.0 tgs: 254 data_time: 0.56s time: 119.99s eta: 12:19:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32014.0 tgs: 269 data_time: 0.55s time: 118.73s eta: 12:10:10 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30957.0 tgs: 258 data_time: 0.68s time: 119.64s eta: 12:13:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31886.0 tgs: 266 data_time: 0.69s time: 119.83s eta: 12:12:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31160.0 tgs: 261 data_time: 0.58s time: 119.14s eta: 12:06:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31331.0 tgs: 262 data_time: 0.75s time: 119.17s eta: 12:04:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31705.0 tgs: 266 data_time: 0.68s time: 119.13s eta: 12:02:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.141 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30592.0 tgs: 256 data_time: 0.61s time: 119.17s eta: 12:01:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.211 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31575.0 tgs: 265 data_time: 0.73s time: 118.96s eta: 11:57:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31447.0 tgs: 262 data_time: 0.68s time: 119.71s eta: 12:00:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.239 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31788.0 tgs: 266 data_time: 0.81s time: 119.28s eta: 11:55:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31499.0 tgs: 264 data_time: 0.66s time: 119.27s eta: 11:53:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31807.0 tgs: 265 data_time: 0.52s time: 119.71s eta: 11:54:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31813.0 tgs: 267 data_time: 0.60s time: 118.79s eta: 11:46:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 266 data_time: 0.75s time: 119.15s eta: 11:46:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31325.0 tgs: 261 data_time: 0.79s time: 119.61s eta: 11:47:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31384.0 tgs: 264 data_time: 0.69s time: 118.85s eta: 11:41:14 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32162.0 tgs: 270 data_time: 0.72s time: 118.68s eta: 11:38:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32124.0 tgs: 269 data_time: 0.80s time: 119.17s eta: 11:39:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 270 data_time: 0.68s time: 119.14s eta: 11:36:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30361.0 tgs: 254 data_time: 0.61s time: 119.08s eta: 11:34:36 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.175 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31871.0 tgs: 267 data_time: 0.76s time: 118.95s eta: 11:31:52 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31505.0 tgs: 264 data_time: 0.99s time: 118.91s eta: 11:29:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.203 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31943.0 tgs: 268 data_time: 0.74s time: 118.97s eta: 11:28:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31304.0 tgs: 262 data_time: 0.48s time: 119.28s eta: 11:27:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31483.0 tgs: 263 data_time: 0.73s time: 119.37s eta: 11:26:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31666.0 tgs: 266 data_time: 0.77s time: 118.87s eta: 11:21:32 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.126 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 264 data_time: 0.88s time: 119.20s eta: 11:21:27 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31514.0 tgs: 263 data_time: 0.80s time: 119.72s eta: 11:22:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.197 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31982.0 tgs: 268 data_time: 0.87s time: 119.11s eta: 11:16:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32449.0 tgs: 272 data_time: 0.74s time: 119.12s eta: 11:15:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31812.0 tgs: 267 data_time: 0.68s time: 118.98s eta: 11:12:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.172 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 267 data_time: 0.68s time: 119.10s eta: 11:10:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.115 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31683.0 tgs: 265 data_time: 0.60s time: 119.16s eta: 11:09:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30762.0 tgs: 258 data_time: 0.67s time: 118.97s eta: 11:06:12 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.202 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31598.0 tgs: 264 data_time: 0.61s time: 119.47s eta: 11:07:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31831.0 tgs: 266 data_time: 1.05s time: 119.64s eta: 11:05:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.156 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32031.0 tgs: 268 data_time: 0.99s time: 119.30s eta: 11:02:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31190.0 tgs: 261 data_time: 0.60s time: 119.31s eta: 11:00:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31891.0 tgs: 268 data_time: 0.59s time: 118.82s eta: 10:55:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.177 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30188.0 tgs: 252 data_time: 0.52s time: 119.47s eta: 10:57:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30448.0 tgs: 255 data_time: 0.69s time: 119.37s eta: 10:54:32 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 266 data_time: 0.65s time: 119.27s eta: 10:52:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31413.0 tgs: 263 data_time: 0.69s time: 119.26s eta: 10:49:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31741.0 tgs: 266 data_time: 0.88s time: 119.10s eta: 10:47:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.195 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31154.0 tgs: 260 data_time: 0.61s time: 119.54s eta: 10:47:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31739.0 tgs: 265 data_time: 0.63s time: 119.33s eta: 10:44:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31724.0 tgs: 265 data_time: 1.00s time: 119.49s eta: 10:43:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31075.0 tgs: 260 data_time: 0.86s time: 119.51s eta: 10:41:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.143 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30901.0 tgs: 259 data_time: 0.87s time: 118.99s eta: 10:36:36 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.177 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 267 data_time: 0.94s time: 119.22s eta: 10:35:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32095.0 tgs: 269 data_time: 0.75s time: 119.06s eta: 10:33:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31227.0 tgs: 262 data_time: 0.75s time: 118.77s eta: 10:29:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32062.0 tgs: 267 data_time: 0.88s time: 119.86s eta: 10:33:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32041.0 tgs: 268 data_time: 0.84s time: 119.38s eta: 10:28:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32102.0 tgs: 270 data_time: 0.70s time: 118.83s eta: 10:23:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31813.0 tgs: 267 data_time: 0.62s time: 118.99s eta: 10:22:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32587.0 tgs: 273 data_time: 0.71s time: 119.29s eta: 10:22:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 268 data_time: 0.78s time: 119.20s eta: 10:19:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.130 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31099.0 tgs: 261 data_time: 0.78s time: 119.07s eta: 10:17:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31614.0 tgs: 264 data_time: 0.77s time: 119.47s eta: 10:17:14 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32070.0 tgs: 268 data_time: 0.71s time: 119.39s eta: 10:14:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31472.0 tgs: 263 data_time: 0.60s time: 119.45s eta: 10:13:10 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 267 data_time: 0.98s time: 119.29s eta: 10:10:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32012.0 tgs: 268 data_time: 0.57s time: 119.11s eta: 10:07:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30739.0 tgs: 256 data_time: 0.71s time: 119.70s eta: 10:08:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.125 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31454.0 tgs: 262 data_time: 0.66s time: 119.83s eta: 10:07:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.146 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30962.0 tgs: 254 data_time: 0.52s time: 121.56s eta: 10:13:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31616.0 tgs: 265 data_time: 0.78s time: 119.25s eta: 10:00:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32129.0 tgs: 270 data_time: 0.67s time: 118.81s eta: 9:56:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 267 data_time: 0.78s time: 119.56s eta: 9:57:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32008.0 tgs: 267 data_time: 0.93s time: 119.70s eta: 9:56:31 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.73s time: 119.26s eta: 9:52:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31567.0 tgs: 264 data_time: 1.04s time: 119.33s eta: 9:50:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31184.0 tgs: 261 data_time: 0.94s time: 119.23s eta: 9:48:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31932.0 tgs: 268 data_time: 1.07s time: 118.82s eta: 9:44:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.179 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30918.0 tgs: 259 data_time: 1.06s time: 119.28s eta: 9:44:27 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.195 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31846.0 tgs: 266 data_time: 0.73s time: 119.46s eta: 9:43:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.124 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31457.0 tgs: 263 data_time: 0.86s time: 119.55s eta: 9:41:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31481.0 tgs: 264 data_time: 1.02s time: 119.20s eta: 9:38:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.123 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31313.0 tgs: 262 data_time: 0.78s time: 119.48s eta: 9:37:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.192 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31821.0 tgs: 267 data_time: 0.47s time: 119.06s eta: 9:33:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.230 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31695.0 tgs: 265 data_time: 0.82s time: 119.24s eta: 9:32:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.157 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32073.0 tgs: 269 data_time: 0.80s time: 119.16s eta: 9:29:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30992.0 tgs: 259 data_time: 0.58s time: 119.23s eta: 9:28:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.166 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31918.0 tgs: 267 data_time: 0.61s time: 119.46s eta: 9:27:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 269 data_time: 0.79s time: 118.98s eta: 9:23:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.174 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31931.0 tgs: 268 data_time: 0.66s time: 119.04s eta: 9:21:27 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31455.0 tgs: 263 data_time: 0.81s time: 119.53s eta: 9:21:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31926.0 tgs: 266 data_time: 0.87s time: 119.60s eta: 9:20:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31454.0 tgs: 264 data_time: 0.87s time: 119.11s eta: 9:15:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31883.0 tgs: 267 data_time: 0.57s time: 119.06s eta: 9:13:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 269 data_time: 0.64s time: 118.88s eta: 9:10:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31653.0 tgs: 265 data_time: 0.77s time: 119.32s eta: 9:10:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32274.0 tgs: 271 data_time: 0.84s time: 119.09s eta: 9:07:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31286.0 tgs: 263 data_time: 0.67s time: 118.86s eta: 9:04:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32263.0 tgs: 269 data_time: 0.74s time: 119.67s eta: 9:06:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31971.0 tgs: 267 data_time: 0.96s time: 119.43s eta: 9:03:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32004.0 tgs: 267 data_time: 0.76s time: 119.43s eta: 9:01:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.222 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31946.0 tgs: 269 data_time: 0.76s time: 118.74s eta: 8:56:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31200.0 tgs: 262 data_time: 0.46s time: 118.86s eta: 8:54:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 267 data_time: 0.85s time: 119.34s eta: 8:55:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 261 data_time: 0.81s time: 119.15s eta: 8:52:10 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.115 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.65s time: 119.27s eta: 8:50:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 265 data_time: 0.82s time: 120.46s eta: 8:54:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32113.0 tgs: 269 data_time: 0.58s time: 119.24s eta: 8:46:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31893.0 tgs: 267 data_time: 0.74s time: 119.07s eta: 8:43:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32119.0 tgs: 265 data_time: 0.76s time: 120.83s eta: 8:49:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 263 data_time: 0.50s time: 120.20s eta: 8:44:52 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31453.0 tgs: 264 data_time: 0.57s time: 118.82s eta: 8:36:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.193 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31793.0 tgs: 266 data_time: 0.93s time: 119.35s eta: 8:37:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31942.0 tgs: 267 data_time: 0.75s time: 119.33s eta: 8:35:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32459.0 tgs: 272 data_time: 0.95s time: 119.06s eta: 8:31:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32107.0 tgs: 269 data_time: 0.82s time: 119.16s eta: 8:30:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31536.0 tgs: 264 data_time: 0.61s time: 119.05s eta: 8:27:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31293.0 tgs: 262 data_time: 0.81s time: 119.11s eta: 8:26:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.116 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31608.0 tgs: 265 data_time: 0.63s time: 119.24s eta: 8:24:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 265 data_time: 0.95s time: 119.22s eta: 8:22:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.108 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30454.0 tgs: 255 data_time: 0.64s time: 119.38s eta: 8:21:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32005.0 tgs: 267 data_time: 0.99s time: 119.78s eta: 8:21:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.150 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32016.0 tgs: 268 data_time: 0.75s time: 119.20s eta: 8:16:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.262 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32147.0 tgs: 270 data_time: 0.74s time: 119.02s eta: 8:13:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31854.0 tgs: 265 data_time: 0.69s time: 119.89s eta: 8:15:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32245.0 tgs: 269 data_time: 1.15s time: 119.55s eta: 8:12:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31493.0 tgs: 263 data_time: 0.77s time: 119.59s eta: 8:10:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32026.0 tgs: 269 data_time: 0.86s time: 119.00s eta: 8:05:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.176 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30549.0 tgs: 254 data_time: 0.92s time: 119.84s eta: 8:07:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.249 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32343.0 tgs: 270 data_time: 0.73s time: 119.71s eta: 8:04:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31206.0 tgs: 261 data_time: 0.61s time: 119.40s eta: 8:01:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31594.0 tgs: 264 data_time: 0.77s time: 119.28s eta: 7:59:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30887.0 tgs: 259 data_time: 0.75s time: 119.19s eta: 7:56:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31503.0 tgs: 264 data_time: 0.49s time: 119.20s eta: 7:54:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.166 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32381.0 tgs: 272 data_time: 0.56s time: 118.93s eta: 7:51:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32309.0 tgs: 271 data_time: 0.83s time: 118.97s eta: 7:49:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.120 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31660.0 tgs: 265 data_time: 0.67s time: 119.42s eta: 7:49:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31337.0 tgs: 263 data_time: 0.81s time: 119.03s eta: 7:46:12 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 25182.0 tgs: 211 data_time: 0.66s time: 119.10s eta: 7:44:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32028.0 tgs: 266 data_time: 0.73s time: 120.00s eta: 7:46:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31767.0 tgs: 265 data_time: 0.56s time: 119.83s eta: 7:43:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31799.0 tgs: 266 data_time: 0.81s time: 119.16s eta: 7:38:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.187 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31759.0 tgs: 266 data_time: 1.00s time: 118.97s eta: 7:36:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31630.0 tgs: 265 data_time: 0.70s time: 119.32s eta: 7:35:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30443.0 tgs: 255 data_time: 0.79s time: 119.38s eta: 7:33:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31228.0 tgs: 262 data_time: 0.61s time: 119.07s eta: 7:30:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.195 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 268 data_time: 1.03s time: 119.00s eta: 7:28:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32484.0 tgs: 272 data_time: 0.72s time: 119.23s eta: 7:27:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.109 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31202.0 tgs: 262 data_time: 0.66s time: 118.99s eta: 7:24:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30642.0 tgs: 257 data_time: 0.77s time: 118.93s eta: 7:22:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.171 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31756.0 tgs: 266 data_time: 0.59s time: 119.08s eta: 7:20:35 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.198 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 268 data_time: 0.84s time: 118.99s eta: 7:18:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32516.0 tgs: 273 data_time: 0.63s time: 119.02s eta: 7:16:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.229 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31528.0 tgs: 264 data_time: 1.10s time: 119.27s eta: 7:15:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31875.0 tgs: 267 data_time: 0.42s time: 119.37s eta: 7:13:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30635.0 tgs: 257 data_time: 0.76s time: 118.94s eta: 7:10:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32122.0 tgs: 269 data_time: 0.68s time: 119.20s eta: 7:09:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32345.0 tgs: 270 data_time: 0.98s time: 119.58s eta: 7:08:30 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31069.0 tgs: 259 data_time: 0.81s time: 119.66s eta: 7:06:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31326.0 tgs: 263 data_time: 0.88s time: 118.95s eta: 7:02:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31306.0 tgs: 262 data_time: 0.53s time: 119.30s eta: 7:01:30 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31082.0 tgs: 260 data_time: 0.87s time: 119.52s eta: 7:00:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31948.0 tgs: 267 data_time: 0.80s time: 119.42s eta: 6:57:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.179 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31260.0 tgs: 262 data_time: 0.68s time: 118.92s eta: 6:54:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31907.0 tgs: 267 data_time: 1.07s time: 119.33s eta: 6:53:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30801.0 tgs: 258 data_time: 0.82s time: 118.93s eta: 6:50:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31625.0 tgs: 265 data_time: 1.01s time: 119.27s eta: 6:49:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31799.0 tgs: 267 data_time: 0.68s time: 119.03s eta: 6:46:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.091 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30983.0 tgs: 259 data_time: 0.66s time: 119.24s eta: 6:45:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30953.0 tgs: 259 data_time: 0.57s time: 119.38s eta: 6:43:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32121.0 tgs: 269 data_time: 0.92s time: 119.01s eta: 6:40:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32112.0 tgs: 269 data_time: 1.05s time: 119.01s eta: 6:38:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32111.0 tgs: 268 data_time: 0.70s time: 119.45s eta: 6:38:10 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31064.0 tgs: 261 data_time: 0.83s time: 118.80s eta: 6:34:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31257.0 tgs: 261 data_time: 0.78s time: 119.44s eta: 6:34:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31319.0 tgs: 261 data_time: 0.82s time: 119.57s eta: 6:32:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.113 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31573.0 tgs: 264 data_time: 0.79s time: 119.38s eta: 6:29:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30942.0 tgs: 259 data_time: 0.69s time: 119.41s eta: 6:28:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30872.0 tgs: 258 data_time: 0.70s time: 119.37s eta: 6:25:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31643.0 tgs: 266 data_time: 0.66s time: 118.82s eta: 6:22:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31431.0 tgs: 264 data_time: 0.99s time: 119.01s eta: 6:20:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.113 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31563.0 tgs: 264 data_time: 0.54s time: 119.40s eta: 6:20:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31098.0 tgs: 260 data_time: 0.55s time: 119.35s eta: 6:17:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31016.0 tgs: 260 data_time: 0.85s time: 119.05s eta: 6:15:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.198 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29927.0 tgs: 250 data_time: 0.60s time: 119.40s eta: 6:14:07 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32021.0 tgs: 268 data_time: 0.68s time: 119.40s eta: 6:12:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30210.0 tgs: 253 data_time: 0.90s time: 119.28s eta: 6:09:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31420.0 tgs: 263 data_time: 0.87s time: 119.10s eta: 6:07:14 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31113.0 tgs: 255 data_time: 1.06s time: 121.96s eta: 6:13:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.214 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31980.0 tgs: 267 data_time: 0.64s time: 119.39s eta: 6:04:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31011.0 tgs: 259 data_time: 0.86s time: 119.56s eta: 6:02:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31348.0 tgs: 262 data_time: 0.54s time: 119.28s eta: 5:59:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30900.0 tgs: 259 data_time: 0.90s time: 119.24s eta: 5:57:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32156.0 tgs: 270 data_time: 0.61s time: 118.91s eta: 5:54:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30116.0 tgs: 253 data_time: 0.67s time: 119.02s eta: 5:53:05 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.85s time: 119.07s eta: 5:51:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.203 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31046.0 tgs: 260 data_time: 0.84s time: 119.29s eta: 5:49:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31475.0 tgs: 264 data_time: 0.87s time: 118.97s eta: 5:46:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32044.0 tgs: 268 data_time: 0.73s time: 119.27s eta: 5:45:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31385.0 tgs: 262 data_time: 0.67s time: 119.35s eta: 5:44:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31786.0 tgs: 266 data_time: 0.91s time: 119.05s eta: 5:41:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.221 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31949.0 tgs: 267 data_time: 0.72s time: 119.50s eta: 5:40:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31691.0 tgs: 265 data_time: 0.60s time: 119.38s eta: 5:38:14 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31537.0 tgs: 263 data_time: 0.90s time: 119.57s eta: 5:36:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31791.0 tgs: 267 data_time: 0.60s time: 118.98s eta: 5:33:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31380.0 tgs: 263 data_time: 0.98s time: 118.95s eta: 5:31:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31003.0 tgs: 260 data_time: 0.74s time: 119.24s eta: 5:29:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31478.0 tgs: 263 data_time: 0.79s time: 119.40s eta: 5:28:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31004.0 tgs: 259 data_time: 0.87s time: 119.46s eta: 5:26:31 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32147.0 tgs: 270 data_time: 0.64s time: 118.95s eta: 5:23:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31005.0 tgs: 259 data_time: 0.57s time: 119.63s eta: 5:22:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.192 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31620.0 tgs: 264 data_time: 0.80s time: 119.36s eta: 5:20:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.181 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31568.0 tgs: 265 data_time: 0.93s time: 119.11s eta: 5:17:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.71s time: 119.42s eta: 5:16:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31935.0 tgs: 267 data_time: 0.55s time: 119.60s eta: 5:14:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31286.0 tgs: 262 data_time: 0.89s time: 119.00s eta: 5:11:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32052.0 tgs: 269 data_time: 0.65s time: 119.00s eta: 5:09:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31161.0 tgs: 261 data_time: 0.75s time: 119.22s eta: 5:07:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31157.0 tgs: 260 data_time: 0.62s time: 119.69s eta: 5:07:12 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.213 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31563.0 tgs: 264 data_time: 0.85s time: 119.11s eta: 5:03:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30869.0 tgs: 259 data_time: 0.57s time: 119.07s eta: 5:01:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.240 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31814.0 tgs: 261 data_time: 0.84s time: 121.52s eta: 5:05:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 266 data_time: 0.94s time: 119.57s eta: 4:58:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32437.0 tgs: 271 data_time: 0.76s time: 119.50s eta: 4:56:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 26307.0 tgs: 219 data_time: 0.71s time: 119.73s eta: 4:55:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.205 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31684.0 tgs: 265 data_time: 0.73s time: 119.52s eta: 4:52:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32287.0 tgs: 270 data_time: 0.61s time: 119.41s eta: 4:50:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.172 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31623.0 tgs: 265 data_time: 0.79s time: 119.20s eta: 4:48:04 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32163.0 tgs: 270 data_time: 0.66s time: 118.88s eta: 4:45:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31675.0 tgs: 265 data_time: 0.70s time: 119.09s eta: 4:43:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32028.0 tgs: 266 data_time: 0.67s time: 119.97s eta: 4:43:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31647.0 tgs: 265 data_time: 0.92s time: 119.15s eta: 4:39:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.186 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.4GB text_tokens: 30528.0 tgs: 256 data_time: 0.80s time: 119.01s eta: 4:37:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32233.0 tgs: 270 data_time: 0.81s time: 119.05s eta: 4:35:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31490.0 tgs: 263 data_time: 0.57s time: 119.37s eta: 4:34:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.163 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31678.0 tgs: 265 data_time: 0.59s time: 119.41s eta: 4:32:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32165.0 tgs: 270 data_time: 0.85s time: 118.93s eta: 4:29:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32539.0 tgs: 272 data_time: 0.66s time: 119.60s eta: 4:29:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31606.0 tgs: 265 data_time: 0.80s time: 119.24s eta: 4:26:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31685.0 tgs: 266 data_time: 0.62s time: 119.06s eta: 4:23:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31526.0 tgs: 265 data_time: 1.02s time: 118.86s eta: 4:21:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31911.0 tgs: 268 data_time: 0.86s time: 118.75s eta: 4:19:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32220.0 tgs: 269 data_time: 0.79s time: 119.43s eta: 4:18:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31145.0 tgs: 260 data_time: 0.71s time: 119.40s eta: 4:16:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.192 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31708.0 tgs: 265 data_time: 0.96s time: 119.35s eta: 4:14:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 269 data_time: 0.79s time: 119.05s eta: 4:11:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.198 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31027.0 tgs: 260 data_time: 0.84s time: 119.12s eta: 4:10:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31633.0 tgs: 265 data_time: 0.53s time: 119.14s eta: 4:08:12 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.111 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32239.0 tgs: 271 data_time: 0.47s time: 118.95s eta: 4:05:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.245 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31226.0 tgs: 262 data_time: 0.75s time: 119.15s eta: 4:04:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31694.0 tgs: 265 data_time: 0.75s time: 119.50s eta: 4:02:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.284 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31483.0 tgs: 264 data_time: 0.97s time: 119.25s eta: 4:00:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29467.0 tgs: 247 data_time: 0.55s time: 119.20s eta: 3:58:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31517.0 tgs: 264 data_time: 0.70s time: 119.35s eta: 3:56:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31651.0 tgs: 265 data_time: 0.67s time: 119.37s eta: 3:54:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31897.0 tgs: 265 data_time: 0.75s time: 119.95s eta: 3:53:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32458.0 tgs: 272 data_time: 0.67s time: 119.32s eta: 3:50:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31760.0 tgs: 265 data_time: 0.63s time: 119.64s eta: 3:49:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30861.0 tgs: 259 data_time: 0.65s time: 118.85s eta: 3:45:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31861.0 tgs: 266 data_time: 1.03s time: 119.75s eta: 3:45:31 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.105 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28188.0 tgs: 236 data_time: 0.77s time: 119.30s eta: 3:42:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30289.0 tgs: 253 data_time: 0.74s time: 119.27s eta: 3:40:39 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31869.0 tgs: 267 data_time: 0.70s time: 118.99s eta: 3:38:09 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.192 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32273.0 tgs: 270 data_time: 0.75s time: 119.42s eta: 3:36:57 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31189.0 tgs: 262 data_time: 0.93s time: 119.03s eta: 3:34:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32195.0 tgs: 270 data_time: 0.84s time: 118.99s eta: 3:32:12 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31331.0 tgs: 262 data_time: 0.91s time: 119.22s eta: 3:30:36 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32435.0 tgs: 270 data_time: 0.86s time: 120.11s eta: 3:30:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.197 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31302.0 tgs: 262 data_time: 0.67s time: 119.14s eta: 3:26:30 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31486.0 tgs: 263 data_time: 0.49s time: 119.32s eta: 3:24:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31801.0 tgs: 267 data_time: 0.55s time: 119.07s eta: 3:22:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31581.0 tgs: 264 data_time: 0.65s time: 119.43s eta: 3:21:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.244 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31654.0 tgs: 265 data_time: 0.49s time: 119.29s eta: 3:18:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32087.0 tgs: 269 data_time: 1.14s time: 119.22s eta: 3:16:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31710.0 tgs: 266 data_time: 0.70s time: 119.18s eta: 3:14:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32326.0 tgs: 271 data_time: 0.72s time: 119.02s eta: 3:12:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.187 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31997.0 tgs: 268 data_time: 0.72s time: 119.19s eta: 3:10:41 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32110.0 tgs: 265 data_time: 0.75s time: 120.85s eta: 3:11:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31236.0 tgs: 261 data_time: 0.65s time: 119.50s eta: 3:07:13 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32178.0 tgs: 269 data_time: 0.84s time: 119.26s eta: 3:04:51 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.191 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30900.0 tgs: 259 data_time: 0.71s time: 118.98s eta: 3:02:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32149.0 tgs: 269 data_time: 0.82s time: 119.22s eta: 3:00:48 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31652.0 tgs: 264 data_time: 0.90s time: 119.46s eta: 2:59:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31170.0 tgs: 261 data_time: 0.76s time: 118.98s eta: 2:56:28 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.220 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31338.0 tgs: 262 data_time: 0.72s time: 119.35s eta: 2:55:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31676.0 tgs: 265 data_time: 0.93s time: 119.41s eta: 2:53:08 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32005.0 tgs: 268 data_time: 0.84s time: 119.35s eta: 2:51:03 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.287 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31895.0 tgs: 262 data_time: 0.79s time: 121.46s eta: 2:52:03 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31041.0 tgs: 260 data_time: 0.78s time: 119.24s eta: 2:46:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30787.0 tgs: 257 data_time: 0.79s time: 119.57s eta: 2:45:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.191 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31572.0 tgs: 265 data_time: 0.81s time: 118.90s eta: 2:42:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30071.0 tgs: 252 data_time: 1.04s time: 119.18s eta: 2:40:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.175 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32160.0 tgs: 270 data_time: 0.56s time: 118.96s eta: 2:38:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30535.0 tgs: 256 data_time: 0.63s time: 119.24s eta: 2:37:00 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32279.0 tgs: 271 data_time: 0.83s time: 119.10s eta: 2:34:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.130 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31510.0 tgs: 263 data_time: 0.84s time: 119.43s eta: 2:33:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 264 data_time: 0.92s time: 120.22s eta: 2:32:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31828.0 tgs: 266 data_time: 0.76s time: 119.29s eta: 2:29:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31998.0 tgs: 269 data_time: 0.71s time: 118.88s eta: 2:26:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31473.0 tgs: 263 data_time: 0.84s time: 119.51s eta: 2:25:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31979.0 tgs: 268 data_time: 0.81s time: 119.07s eta: 2:22:52 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32169.0 tgs: 270 data_time: 0.70s time: 119.14s eta: 2:20:58 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30734.0 tgs: 258 data_time: 0.83s time: 119.05s eta: 2:18:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32016.0 tgs: 267 data_time: 0.60s time: 119.85s eta: 2:17:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31679.0 tgs: 265 data_time: 0.95s time: 119.35s eta: 2:15:15 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31179.0 tgs: 261 data_time: 0.88s time: 119.39s eta: 2:13:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32285.0 tgs: 269 data_time: 0.85s time: 119.68s eta: 2:11:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31867.0 tgs: 266 data_time: 0.72s time: 119.49s eta: 2:09:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.204 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32192.0 tgs: 270 data_time: 0.61s time: 118.96s eta: 2:06:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.114 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 267 data_time: 0.55s time: 119.38s eta: 2:05:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31268.0 tgs: 262 data_time: 0.60s time: 119.27s eta: 2:03:14 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.193 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31972.0 tgs: 268 data_time: 0.64s time: 119.22s eta: 2:01:12 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32105.0 tgs: 268 data_time: 0.69s time: 119.50s eta: 1:59:30 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.110 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31854.0 tgs: 266 data_time: 0.66s time: 119.42s eta: 1:57:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.3GB text_tokens: 29773.0 tgs: 249 data_time: 0.61s time: 119.55s eta: 1:55:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.196 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31566.0 tgs: 264 data_time: 0.77s time: 119.30s eta: 1:53:20 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32023.0 tgs: 268 data_time: 0.88s time: 119.24s eta: 1:51:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31523.0 tgs: 265 data_time: 1.04s time: 118.93s eta: 1:49:01 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31316.0 tgs: 261 data_time: 0.48s time: 119.63s eta: 1:47:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32316.0 tgs: 269 data_time: 0.72s time: 119.72s eta: 1:45:45 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31506.0 tgs: 263 data_time: 0.68s time: 119.35s eta: 1:43:26 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.246 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31696.0 tgs: 265 data_time: 0.83s time: 119.30s eta: 1:41:24 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.77s time: 119.28s eta: 1:39:23 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31821.0 tgs: 267 data_time: 0.57s time: 118.90s eta: 1:37:06 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31955.0 tgs: 268 data_time: 0.55s time: 119.16s eta: 1:35:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31859.0 tgs: 267 data_time: 0.86s time: 119.07s eta: 1:33:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32108.0 tgs: 269 data_time: 0.68s time: 119.26s eta: 1:31:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31293.0 tgs: 261 data_time: 0.67s time: 119.72s eta: 1:29:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32056.0 tgs: 269 data_time: 0.78s time: 119.02s eta: 1:27:16 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 268 data_time: 0.68s time: 118.86s eta: 1:25:11 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30951.0 tgs: 259 data_time: 0.73s time: 119.47s eta: 1:23:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.212 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32311.0 tgs: 270 data_time: 1.10s time: 119.48s eta: 1:21:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31327.0 tgs: 263 data_time: 0.57s time: 118.94s eta: 1:19:17 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 272 data_time: 0.61s time: 118.94s eta: 1:17:18 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32327.0 tgs: 264 data_time: 0.68s time: 122.15s eta: 1:17:21 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31631.0 tgs: 264 data_time: 0.80s time: 119.37s eta: 1:13:36 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31944.0 tgs: 268 data_time: 0.60s time: 119.05s eta: 1:11:25 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.195 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32267.0 tgs: 271 data_time: 0.81s time: 118.92s eta: 1:09:22 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32058.0 tgs: 269 data_time: 0.86s time: 119.11s eta: 1:07:29 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32293.0 tgs: 270 data_time: 0.73s time: 119.42s eta: 1:05:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32021.0 tgs: 268 data_time: 0.63s time: 119.34s eta: 1:03:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31485.0 tgs: 264 data_time: 0.93s time: 119.19s eta: 1:01:34 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30686.0 tgs: 257 data_time: 0.79s time: 119.26s eta: 0:59:37 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31480.0 tgs: 263 data_time: 0.76s time: 119.44s eta: 0:57:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.186 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32253.0 tgs: 270 data_time: 0.80s time: 119.39s eta: 0:55:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31871.0 tgs: 266 data_time: 0.97s time: 119.48s eta: 0:53:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30874.0 tgs: 258 data_time: 0.64s time: 119.37s eta: 0:51:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31851.0 tgs: 266 data_time: 0.63s time: 119.46s eta: 0:49:46 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32302.0 tgs: 271 data_time: 0.84s time: 118.89s eta: 0:47:33 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31295.0 tgs: 262 data_time: 0.84s time: 119.29s eta: 0:45:43 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31799.0 tgs: 266 data_time: 0.61s time: 119.21s eta: 0:43:42 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31303.0 tgs: 263 data_time: 0.84s time: 118.99s eta: 0:41:38 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.75s time: 119.04s eta: 0:39:40 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32044.0 tgs: 267 data_time: 0.38s time: 119.66s eta: 0:37:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31888.0 tgs: 267 data_time: 0.76s time: 119.29s eta: 0:35:47 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.195 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31660.0 tgs: 264 data_time: 0.93s time: 119.74s eta: 0:33:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30811.0 tgs: 254 data_time: 0.69s time: 121.23s eta: 0:32:19 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31739.0 tgs: 265 data_time: 0.84s time: 119.38s eta: 0:29:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32381.0 tgs: 271 data_time: 0.72s time: 119.32s eta: 0:27:50 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31276.0 tgs: 261 data_time: 0.56s time: 119.64s eta: 0:25:55 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31583.0 tgs: 265 data_time: 0.88s time: 119.10s eta: 0:23:49 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31300.0 tgs: 262 data_time: 0.92s time: 119.39s eta: 0:21:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31720.0 tgs: 265 data_time: 0.66s time: 119.25s eta: 0:19:52 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31606.0 tgs: 265 data_time: 0.86s time: 119.24s eta: 0:17:53 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 265 data_time: 0.76s time: 119.91s eta: 0:15:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32213.0 tgs: 270 data_time: 0.75s time: 119.21s eta: 0:13:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31027.0 tgs: 260 data_time: 0.78s time: 119.09s eta: 0:11:54 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30893.0 tgs: 258 data_time: 0.86s time: 119.32s eta: 0:09:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30638.0 tgs: 257 data_time: 0.69s time: 119.17s eta: 0:07:56 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 265 data_time: 0.97s time: 119.69s eta: 0:05:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 266 data_time: 0.55s time: 121.23s eta: 0:04:02 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31447.0 tgs: 263 data_time: 0.73s time: 119.22s eta: 0:01:59 +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 57][DP 14][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank60.log b/20250121165312/rank60.log new file mode 100644 index 0000000000000000000000000000000000000000..ae0e12c9805d1c9bc3de59441e24fd1804eb9dc1 --- /dev/null +++ b/20250121165312/rank60.log @@ -0,0 +1,851 @@ +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:20][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:25][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 71.86 seconds, peak gpu memory 13.4G +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.246 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.9GB text_tokens: 31861.0 tgs: 210 data_time: 1.85s time: 151.43s eta: 1 day, 0:56:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.220 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31345.0 tgs: 260 data_time: 0.63s time: 120.10s eta: 19:44:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.213 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30897.0 tgs: 258 data_time: 0.81s time: 119.68s eta: 19:38:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30727.0 tgs: 256 data_time: 0.83s time: 119.65s eta: 19:36:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31993.0 tgs: 268 data_time: 0.58s time: 119.31s eta: 19:31:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.198 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32528.0 tgs: 272 data_time: 0.75s time: 119.54s eta: 19:31:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.241 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 260 data_time: 0.60s time: 122.96s eta: 20:02:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 270 data_time: 0.80s time: 119.89s eta: 19:30:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 255 data_time: 0.68s time: 120.65s eta: 19:36:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.211 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32133.0 tgs: 267 data_time: 1.07s time: 119.92s eta: 19:27:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31934.0 tgs: 265 data_time: 0.79s time: 120.46s eta: 19:30:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.255 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31332.0 tgs: 262 data_time: 0.82s time: 119.34s eta: 19:17:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32182.0 tgs: 268 data_time: 0.83s time: 119.68s eta: 19:18:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30812.0 tgs: 258 data_time: 0.57s time: 119.20s eta: 19:12:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32580.0 tgs: 273 data_time: 0.69s time: 119.21s eta: 19:10:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30706.0 tgs: 258 data_time: 0.77s time: 118.94s eta: 19:05:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 268 data_time: 0.79s time: 119.14s eta: 19:05:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32212.0 tgs: 270 data_time: 0.87s time: 119.18s eta: 19:04:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 267 data_time: 0.90s time: 119.40s eta: 19:04:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31887.0 tgs: 267 data_time: 0.82s time: 119.33s eta: 19:01:33 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31686.0 tgs: 263 data_time: 0.96s time: 120.31s eta: 19:08:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31729.0 tgs: 258 data_time: 0.71s time: 122.91s eta: 19:31:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31873.0 tgs: 266 data_time: 0.75s time: 119.48s eta: 18:57:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31838.0 tgs: 267 data_time: 0.77s time: 119.15s eta: 18:51:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 267 data_time: 0.86s time: 119.26s eta: 18:50:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31822.0 tgs: 266 data_time: 0.70s time: 119.38s eta: 18:50:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.194 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32115.0 tgs: 268 data_time: 1.12s time: 119.66s eta: 18:50:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32433.0 tgs: 271 data_time: 0.85s time: 119.38s eta: 18:46:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31453.0 tgs: 262 data_time: 0.91s time: 119.79s eta: 18:47:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32050.0 tgs: 247 data_time: 0.88s time: 129.31s eta: 20:15:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32259.0 tgs: 269 data_time: 0.66s time: 119.57s eta: 18:41:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.72s time: 119.39s eta: 18:38:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30717.0 tgs: 254 data_time: 0.75s time: 120.66s eta: 18:48:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31683.0 tgs: 265 data_time: 0.84s time: 119.39s eta: 18:34:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.279 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32425.0 tgs: 272 data_time: 1.08s time: 119.17s eta: 18:30:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31127.0 tgs: 261 data_time: 0.66s time: 119.06s eta: 18:27:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31857.0 tgs: 267 data_time: 0.77s time: 119.13s eta: 18:25:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31766.0 tgs: 263 data_time: 0.76s time: 120.39s eta: 18:35:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32158.0 tgs: 265 data_time: 0.87s time: 121.17s eta: 18:40:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32032.0 tgs: 265 data_time: 0.81s time: 120.66s eta: 18:34:05 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31101.0 tgs: 259 data_time: 0.75s time: 119.74s eta: 18:23:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30739.0 tgs: 248 data_time: 0.92s time: 123.87s eta: 18:59:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32313.0 tgs: 270 data_time: 0.83s time: 119.56s eta: 18:17:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30208.0 tgs: 253 data_time: 0.67s time: 119.36s eta: 18:14:10 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 264 data_time: 0.92s time: 119.21s eta: 18:10:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31887.0 tgs: 266 data_time: 0.90s time: 119.60s eta: 18:12:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31263.0 tgs: 262 data_time: 0.69s time: 119.28s eta: 18:07:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 268 data_time: 0.89s time: 119.00s eta: 18:02:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32230.0 tgs: 270 data_time: 1.20s time: 119.10s eta: 18:01:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31041.0 tgs: 260 data_time: 0.99s time: 119.26s eta: 18:01:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31939.0 tgs: 267 data_time: 0.65s time: 119.33s eta: 17:59:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31788.0 tgs: 264 data_time: 1.15s time: 120.02s eta: 18:04:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32224.0 tgs: 270 data_time: 0.68s time: 119.10s eta: 17:53:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32035.0 tgs: 268 data_time: 0.72s time: 119.41s eta: 17:54:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30404.0 tgs: 255 data_time: 0.74s time: 118.99s eta: 17:48:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32351.0 tgs: 271 data_time: 1.05s time: 119.02s eta: 17:47:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32058.0 tgs: 255 data_time: 0.68s time: 125.37s eta: 18:42:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.205 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 267 data_time: 0.90s time: 120.10s eta: 17:52:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 263 data_time: 0.68s time: 119.39s eta: 17:44:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31503.0 tgs: 263 data_time: 0.82s time: 119.34s eta: 17:42:06 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.142 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32460.0 tgs: 271 data_time: 0.68s time: 119.51s eta: 17:41:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.124 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 251 data_time: 0.42s time: 126.05s eta: 18:37:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 260 data_time: 0.69s time: 122.63s eta: 18:05:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31721.0 tgs: 261 data_time: 0.76s time: 121.15s eta: 17:50:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32329.0 tgs: 271 data_time: 0.70s time: 119.04s eta: 17:29:29 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31333.0 tgs: 263 data_time: 0.81s time: 118.85s eta: 17:25:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 27236.0 tgs: 219 data_time: 0.61s time: 124.00s eta: 18:09:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 258 data_time: 0.82s time: 124.56s eta: 18:11:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.150 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 271 data_time: 0.97s time: 119.40s eta: 17:24:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31521.0 tgs: 263 data_time: 0.93s time: 119.69s eta: 17:25:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31154.0 tgs: 260 data_time: 0.64s time: 119.53s eta: 17:21:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31399.0 tgs: 264 data_time: 0.77s time: 118.86s eta: 17:14:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31343.0 tgs: 261 data_time: 0.84s time: 119.85s eta: 17:20:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31053.0 tgs: 259 data_time: 0.80s time: 119.56s eta: 17:16:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32475.0 tgs: 266 data_time: 0.90s time: 121.86s eta: 17:34:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31529.0 tgs: 255 data_time: 0.73s time: 123.51s eta: 17:46:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31108.0 tgs: 261 data_time: 0.65s time: 119.02s eta: 17:05:31 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.222 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32276.0 tgs: 270 data_time: 0.83s time: 119.23s eta: 17:05:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31566.0 tgs: 264 data_time: 0.91s time: 119.14s eta: 17:02:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31630.0 tgs: 265 data_time: 0.70s time: 119.26s eta: 17:01:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.227 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32522.0 tgs: 273 data_time: 0.94s time: 119.03s eta: 16:57:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30869.0 tgs: 258 data_time: 0.74s time: 119.27s eta: 16:57:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31254.0 tgs: 262 data_time: 0.78s time: 119.16s eta: 16:54:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.213 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28967.0 tgs: 242 data_time: 0.88s time: 119.58s eta: 16:56:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.283 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.72s time: 119.13s eta: 16:50:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31381.0 tgs: 264 data_time: 0.53s time: 118.84s eta: 16:46:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31802.0 tgs: 267 data_time: 0.68s time: 119.01s eta: 16:45:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31937.0 tgs: 266 data_time: 0.80s time: 119.85s eta: 16:50:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31499.0 tgs: 263 data_time: 0.90s time: 119.37s eta: 16:44:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31043.0 tgs: 260 data_time: 0.65s time: 119.01s eta: 16:39:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30452.0 tgs: 251 data_time: 1.01s time: 120.93s eta: 16:53:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30269.0 tgs: 254 data_time: 0.60s time: 118.99s eta: 16:35:32 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.115 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 266 data_time: 0.89s time: 119.29s eta: 16:36:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31521.0 tgs: 264 data_time: 1.14s time: 119.22s eta: 16:33:31 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.208 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31815.0 tgs: 266 data_time: 1.01s time: 119.37s eta: 16:32:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30661.0 tgs: 256 data_time: 0.73s time: 119.71s eta: 16:33:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.223 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31965.0 tgs: 268 data_time: 0.91s time: 119.20s eta: 16:27:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31853.0 tgs: 265 data_time: 0.64s time: 119.76s eta: 16:30:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.257 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31960.0 tgs: 265 data_time: 0.69s time: 120.28s eta: 16:32:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 266 data_time: 0.82s time: 118.97s eta: 16:19:31 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31596.0 tgs: 248 data_time: 0.68s time: 127.33s eta: 17:26:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 265 data_time: 0.76s time: 119.23s eta: 16:17:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.107 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31940.0 tgs: 268 data_time: 0.60s time: 119.15s eta: 16:15:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31867.0 tgs: 267 data_time: 1.09s time: 119.25s eta: 16:13:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 264 data_time: 0.78s time: 119.19s eta: 16:11:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30332.0 tgs: 255 data_time: 0.91s time: 118.93s eta: 16:07:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.123 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31154.0 tgs: 262 data_time: 0.62s time: 118.86s eta: 16:04:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.214 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31244.0 tgs: 262 data_time: 0.57s time: 118.96s eta: 16:03:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 259 data_time: 0.92s time: 121.84s eta: 16:24:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.196 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.81s time: 118.99s eta: 15:59:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.197 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 268 data_time: 0.93s time: 119.23s eta: 15:59:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.113 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32297.0 tgs: 270 data_time: 0.80s time: 119.61s eta: 16:00:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.120 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31976.0 tgs: 268 data_time: 0.86s time: 119.04s eta: 15:54:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.282 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32206.0 tgs: 270 data_time: 0.97s time: 118.98s eta: 15:51:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 267 data_time: 0.68s time: 119.17s eta: 15:51:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32352.0 tgs: 270 data_time: 0.59s time: 119.64s eta: 15:53:06 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31195.0 tgs: 262 data_time: 0.69s time: 118.94s eta: 15:45:32 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32381.0 tgs: 271 data_time: 0.62s time: 119.28s eta: 15:46:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.234 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 269 data_time: 1.17s time: 119.13s eta: 15:43:06 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 267 data_time: 0.71s time: 119.00s eta: 15:40:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31534.0 tgs: 263 data_time: 0.83s time: 119.66s eta: 15:43:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30832.0 tgs: 253 data_time: 0.67s time: 121.50s eta: 15:55:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32197.0 tgs: 269 data_time: 0.62s time: 119.41s eta: 15:37:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31745.0 tgs: 244 data_time: 0.85s time: 129.96s eta: 16:58:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31717.0 tgs: 265 data_time: 1.12s time: 119.47s eta: 15:33:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.269 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31536.0 tgs: 263 data_time: 0.60s time: 119.60s eta: 15:32:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31300.0 tgs: 261 data_time: 0.91s time: 119.53s eta: 15:30:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31783.0 tgs: 260 data_time: 0.77s time: 122.14s eta: 15:48:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.235 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32422.0 tgs: 268 data_time: 0.87s time: 120.76s eta: 15:35:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30462.0 tgs: 255 data_time: 0.87s time: 119.41s eta: 15:23:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 269 data_time: 0.85s time: 119.44s eta: 15:21:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.118 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30621.0 tgs: 232 data_time: 0.96s time: 131.64s eta: 16:53:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.254 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31534.0 tgs: 264 data_time: 0.92s time: 119.29s eta: 15:16:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30926.0 tgs: 258 data_time: 0.61s time: 119.56s eta: 15:16:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 262 data_time: 0.73s time: 122.26s eta: 15:35:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30268.0 tgs: 252 data_time: 0.71s time: 119.70s eta: 15:13:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 264 data_time: 0.77s time: 119.35s eta: 15:09:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.170 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30984.0 tgs: 260 data_time: 0.91s time: 118.90s eta: 15:03:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.133 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 1.03s time: 118.99s eta: 15:02:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.126 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31140.0 tgs: 261 data_time: 0.73s time: 119.31s eta: 15:02:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31989.0 tgs: 269 data_time: 0.72s time: 118.92s eta: 14:57:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32328.0 tgs: 268 data_time: 0.59s time: 120.52s eta: 15:07:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31613.0 tgs: 264 data_time: 0.67s time: 119.31s eta: 14:56:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.189 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31698.0 tgs: 258 data_time: 0.69s time: 122.66s eta: 15:19:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31581.0 tgs: 265 data_time: 1.05s time: 119.12s eta: 14:51:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31926.0 tgs: 266 data_time: 0.71s time: 119.75s eta: 14:54:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31461.0 tgs: 263 data_time: 0.64s time: 119.46s eta: 14:50:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30414.0 tgs: 255 data_time: 0.70s time: 118.99s eta: 14:44:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32174.0 tgs: 269 data_time: 0.95s time: 119.33s eta: 14:45:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30963.0 tgs: 260 data_time: 0.57s time: 118.94s eta: 14:40:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 0.76s time: 119.05s eta: 14:39:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.219 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31750.0 tgs: 266 data_time: 0.82s time: 119.06s eta: 14:37:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32234.0 tgs: 270 data_time: 0.84s time: 119.13s eta: 14:35:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31496.0 tgs: 264 data_time: 0.72s time: 119.00s eta: 14:32:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30789.0 tgs: 258 data_time: 0.85s time: 119.08s eta: 14:31:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.261 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31168.0 tgs: 261 data_time: 0.82s time: 119.14s eta: 14:29:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31841.0 tgs: 266 data_time: 0.92s time: 119.65s eta: 14:31:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31532.0 tgs: 264 data_time: 0.71s time: 119.18s eta: 14:26:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32474.0 tgs: 272 data_time: 0.90s time: 119.20s eta: 14:24:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 265 data_time: 0.63s time: 119.39s eta: 14:23:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31173.0 tgs: 258 data_time: 1.14s time: 120.73s eta: 14:31:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32536.0 tgs: 273 data_time: 0.91s time: 118.86s eta: 14:15:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 269 data_time: 0.78s time: 119.10s eta: 14:15:31 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.124 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32038.0 tgs: 268 data_time: 0.67s time: 119.15s eta: 14:13:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 266 data_time: 0.83s time: 119.95s eta: 14:17:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31272.0 tgs: 260 data_time: 0.90s time: 119.86s eta: 14:15:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.242 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32263.0 tgs: 269 data_time: 0.67s time: 119.59s eta: 14:11:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.118 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31832.0 tgs: 267 data_time: 0.83s time: 119.22s eta: 14:06:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.201 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31974.0 tgs: 268 data_time: 0.85s time: 119.26s eta: 14:04:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31337.0 tgs: 262 data_time: 0.78s time: 119.27s eta: 14:02:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.110 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32038.0 tgs: 268 data_time: 0.92s time: 119.14s eta: 13:59:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30681.0 tgs: 257 data_time: 0.81s time: 119.12s eta: 13:57:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31452.0 tgs: 263 data_time: 0.68s time: 119.17s eta: 13:56:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31620.0 tgs: 264 data_time: 1.09s time: 119.35s eta: 13:55:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.199 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31468.0 tgs: 261 data_time: 0.73s time: 120.12s eta: 13:58:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.198 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31883.0 tgs: 260 data_time: 0.50s time: 122.55s eta: 14:13:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.125 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31475.0 tgs: 264 data_time: 0.73s time: 119.11s eta: 13:47:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 268 data_time: 0.40s time: 119.37s eta: 13:47:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32030.0 tgs: 268 data_time: 1.10s time: 119.35s eta: 13:45:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32278.0 tgs: 270 data_time: 0.96s time: 119.33s eta: 13:43:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30973.0 tgs: 260 data_time: 0.71s time: 119.11s eta: 13:39:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31679.0 tgs: 265 data_time: 0.82s time: 119.26s eta: 13:38:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.285 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 269 data_time: 0.86s time: 119.13s eta: 13:36:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31454.0 tgs: 263 data_time: 0.71s time: 119.22s eta: 13:34:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.78s time: 119.46s eta: 13:34:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 267 data_time: 0.99s time: 119.08s eta: 13:29:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.229 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31043.0 tgs: 261 data_time: 0.85s time: 118.92s eta: 13:26:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.234 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31949.0 tgs: 266 data_time: 0.73s time: 119.81s eta: 13:30:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.123 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31673.0 tgs: 265 data_time: 0.88s time: 119.25s eta: 13:24:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.105 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30259.0 tgs: 253 data_time: 0.82s time: 119.16s eta: 13:22:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.277 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31588.0 tgs: 264 data_time: 0.80s time: 119.34s eta: 13:21:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 271 data_time: 0.80s time: 119.26s eta: 13:19:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.162 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31297.0 tgs: 262 data_time: 0.74s time: 119.04s eta: 13:15:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31956.0 tgs: 267 data_time: 0.77s time: 119.37s eta: 13:15:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 269 data_time: 0.70s time: 119.19s eta: 13:12:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30991.0 tgs: 259 data_time: 1.25s time: 119.46s eta: 13:12:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31552.0 tgs: 264 data_time: 0.74s time: 119.28s eta: 13:09:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31575.0 tgs: 265 data_time: 0.84s time: 119.15s eta: 13:06:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.319 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 266 data_time: 0.66s time: 119.89s eta: 13:09:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.187 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 265 data_time: 0.88s time: 120.16s eta: 13:09:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31793.0 tgs: 266 data_time: 0.94s time: 119.50s eta: 13:02:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 267 data_time: 0.69s time: 119.51s eta: 13:00:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.125 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 264 data_time: 0.72s time: 119.57s eta: 12:59:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 257 data_time: 0.66s time: 119.79s eta: 12:58:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31551.0 tgs: 265 data_time: 0.73s time: 119.04s eta: 12:51:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.127 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 269 data_time: 0.84s time: 119.05s eta: 12:49:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31521.0 tgs: 264 data_time: 0.79s time: 119.12s eta: 12:48:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31568.0 tgs: 264 data_time: 0.75s time: 119.39s eta: 12:48:06 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31459.0 tgs: 263 data_time: 0.78s time: 119.18s eta: 12:44:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.134 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 260 data_time: 0.84s time: 119.43s eta: 12:44:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.254 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 265 data_time: 0.89s time: 119.09s eta: 12:40:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.220 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31535.0 tgs: 264 data_time: 0.71s time: 119.07s eta: 12:38:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.115 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.64s time: 119.59s eta: 12:39:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30891.0 tgs: 260 data_time: 0.86s time: 118.78s eta: 12:32:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32258.0 tgs: 271 data_time: 0.92s time: 118.88s eta: 12:30:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31925.0 tgs: 267 data_time: 0.79s time: 119.17s eta: 12:30:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31125.0 tgs: 260 data_time: 0.92s time: 119.27s eta: 12:29:23 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.326 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30968.0 tgs: 259 data_time: 1.03s time: 119.35s eta: 12:27:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.124 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31570.0 tgs: 265 data_time: 0.64s time: 119.11s eta: 12:24:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.161 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31066.0 tgs: 260 data_time: 0.82s time: 119.34s eta: 12:23:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32253.0 tgs: 270 data_time: 0.70s time: 119.26s eta: 12:21:23 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 0.78s time: 119.03s eta: 12:17:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 32034.0 tgs: 264 data_time: 0.95s time: 121.13s eta: 12:28:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.176 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31566.0 tgs: 263 data_time: 0.78s time: 119.99s eta: 12:19:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.110 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31032.0 tgs: 261 data_time: 0.78s time: 118.73s eta: 12:10:10 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.199 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30435.0 tgs: 254 data_time: 1.08s time: 119.63s eta: 12:13:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.223 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30462.0 tgs: 254 data_time: 0.83s time: 119.83s eta: 12:12:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.131 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 0.66s time: 119.14s eta: 12:06:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.195 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 266 data_time: 0.98s time: 119.17s eta: 12:04:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30964.0 tgs: 259 data_time: 0.84s time: 119.12s eta: 12:02:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32255.0 tgs: 270 data_time: 0.73s time: 119.18s eta: 12:01:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31354.0 tgs: 263 data_time: 0.80s time: 118.95s eta: 11:57:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31553.0 tgs: 263 data_time: 0.87s time: 119.70s eta: 12:00:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 0.90s time: 119.28s eta: 11:55:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.70s time: 119.27s eta: 11:53:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.255 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32032.0 tgs: 267 data_time: 0.88s time: 119.71s eta: 11:54:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31660.0 tgs: 266 data_time: 1.05s time: 118.78s eta: 11:46:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30889.0 tgs: 259 data_time: 0.85s time: 119.16s eta: 11:46:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31706.0 tgs: 265 data_time: 0.66s time: 119.61s eta: 11:47:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32249.0 tgs: 271 data_time: 0.91s time: 118.85s eta: 11:41:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 270 data_time: 0.75s time: 118.68s eta: 11:38:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.170 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32441.0 tgs: 272 data_time: 0.59s time: 119.22s eta: 11:39:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.170 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.73s time: 119.13s eta: 11:36:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.116 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32027.0 tgs: 268 data_time: 0.78s time: 119.07s eta: 11:34:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30730.0 tgs: 258 data_time: 1.14s time: 118.95s eta: 11:31:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32053.0 tgs: 269 data_time: 1.01s time: 118.91s eta: 11:29:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.190 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31359.0 tgs: 263 data_time: 0.63s time: 118.97s eta: 11:28:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.179 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32010.0 tgs: 268 data_time: 0.56s time: 119.28s eta: 11:27:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.207 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31442.0 tgs: 263 data_time: 0.69s time: 119.36s eta: 11:26:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.188 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30872.0 tgs: 259 data_time: 0.79s time: 118.87s eta: 11:21:32 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30977.0 tgs: 259 data_time: 1.01s time: 119.21s eta: 11:21:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.115 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32183.0 tgs: 268 data_time: 0.87s time: 119.71s eta: 11:22:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31874.0 tgs: 267 data_time: 0.88s time: 119.12s eta: 11:16:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.176 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31635.0 tgs: 265 data_time: 1.05s time: 119.12s eta: 11:15:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.201 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31496.0 tgs: 264 data_time: 0.92s time: 118.98s eta: 11:12:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.66s time: 119.10s eta: 11:10:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30964.0 tgs: 259 data_time: 0.57s time: 119.15s eta: 11:09:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32332.0 tgs: 271 data_time: 0.77s time: 118.98s eta: 11:06:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.191 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30372.0 tgs: 254 data_time: 1.08s time: 119.46s eta: 11:06:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31452.0 tgs: 262 data_time: 0.90s time: 119.64s eta: 11:06:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.199 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31934.0 tgs: 267 data_time: 0.78s time: 119.30s eta: 11:02:06 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32152.0 tgs: 269 data_time: 0.86s time: 119.31s eta: 11:00:10 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30536.0 tgs: 256 data_time: 0.79s time: 118.82s eta: 10:55:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.114 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31781.0 tgs: 266 data_time: 0.84s time: 119.47s eta: 10:57:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 260 data_time: 0.68s time: 119.38s eta: 10:54:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31518.0 tgs: 264 data_time: 0.74s time: 119.27s eta: 10:51:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32102.0 tgs: 269 data_time: 0.72s time: 119.26s eta: 10:49:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.119 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31697.0 tgs: 266 data_time: 0.92s time: 119.10s eta: 10:47:05 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 269 data_time: 0.86s time: 119.53s eta: 10:47:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32597.0 tgs: 273 data_time: 0.70s time: 119.34s eta: 10:44:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.269 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 264 data_time: 0.80s time: 119.49s eta: 10:43:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.63s time: 119.51s eta: 10:41:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 269 data_time: 1.03s time: 118.99s eta: 10:36:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.227 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32356.0 tgs: 271 data_time: 0.89s time: 119.22s eta: 10:35:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31688.0 tgs: 266 data_time: 0.60s time: 119.06s eta: 10:33:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.190 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32462.0 tgs: 273 data_time: 0.90s time: 118.78s eta: 10:29:31 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.225 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32143.0 tgs: 268 data_time: 0.77s time: 119.86s eta: 10:33:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31922.0 tgs: 267 data_time: 0.83s time: 119.38s eta: 10:28:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.131 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 271 data_time: 0.81s time: 118.83s eta: 10:23:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30668.0 tgs: 257 data_time: 0.88s time: 118.98s eta: 10:22:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31834.0 tgs: 266 data_time: 0.77s time: 119.28s eta: 10:22:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.122 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29624.0 tgs: 248 data_time: 0.69s time: 119.21s eta: 10:19:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.224 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30908.0 tgs: 259 data_time: 0.68s time: 119.06s eta: 10:17:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 264 data_time: 0.59s time: 119.47s eta: 10:17:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30306.0 tgs: 253 data_time: 0.69s time: 119.39s eta: 10:14:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.207 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31906.0 tgs: 267 data_time: 0.59s time: 119.45s eta: 10:13:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.131 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30612.0 tgs: 256 data_time: 0.96s time: 119.29s eta: 10:10:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32081.0 tgs: 269 data_time: 0.66s time: 119.11s eta: 10:07:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 264 data_time: 0.54s time: 119.70s eta: 10:08:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30197.0 tgs: 252 data_time: 0.73s time: 119.83s eta: 10:07:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32281.0 tgs: 265 data_time: 1.03s time: 121.55s eta: 10:13:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.229 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 266 data_time: 0.93s time: 119.25s eta: 10:00:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31464.0 tgs: 264 data_time: 0.81s time: 118.81s eta: 9:56:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.193 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30866.0 tgs: 258 data_time: 0.82s time: 119.56s eta: 9:57:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 268 data_time: 0.65s time: 119.70s eta: 9:56:31 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.114 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30739.0 tgs: 257 data_time: 0.67s time: 119.26s eta: 9:52:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.208 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31790.0 tgs: 266 data_time: 0.78s time: 119.33s eta: 9:50:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30480.0 tgs: 255 data_time: 0.77s time: 119.22s eta: 9:48:10 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31890.0 tgs: 264 data_time: 0.87s time: 120.67s eta: 9:53:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.57s time: 119.27s eta: 9:44:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31951.0 tgs: 267 data_time: 1.04s time: 119.46s eta: 9:43:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.245 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 268 data_time: 0.86s time: 119.54s eta: 9:41:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.115 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 263 data_time: 0.63s time: 119.20s eta: 9:38:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 270 data_time: 0.81s time: 119.48s eta: 9:37:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31107.0 tgs: 261 data_time: 0.58s time: 119.06s eta: 9:33:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 265 data_time: 0.75s time: 119.24s eta: 9:32:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.198 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31613.0 tgs: 265 data_time: 0.69s time: 119.16s eta: 9:29:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.198 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31659.0 tgs: 265 data_time: 0.83s time: 119.22s eta: 9:28:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.197 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31783.0 tgs: 266 data_time: 1.19s time: 119.46s eta: 9:27:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32338.0 tgs: 271 data_time: 0.51s time: 118.99s eta: 9:23:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.112 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 268 data_time: 0.72s time: 119.03s eta: 9:21:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29882.0 tgs: 249 data_time: 0.58s time: 119.53s eta: 9:21:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.296 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 266 data_time: 0.64s time: 119.59s eta: 9:20:05 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.228 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31682.0 tgs: 265 data_time: 0.83s time: 119.12s eta: 9:15:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31960.0 tgs: 268 data_time: 0.70s time: 119.06s eta: 9:13:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.221 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.83s time: 118.89s eta: 9:10:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31547.0 tgs: 264 data_time: 0.75s time: 119.32s eta: 9:10:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30915.0 tgs: 259 data_time: 0.78s time: 119.09s eta: 9:07:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32237.0 tgs: 271 data_time: 0.74s time: 118.85s eta: 9:04:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.188 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31751.0 tgs: 265 data_time: 0.86s time: 119.67s eta: 9:06:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.191 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31612.0 tgs: 264 data_time: 0.79s time: 119.43s eta: 9:03:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31026.0 tgs: 259 data_time: 0.97s time: 119.43s eta: 9:01:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31302.0 tgs: 263 data_time: 0.79s time: 118.74s eta: 8:56:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.198 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31825.0 tgs: 267 data_time: 0.71s time: 118.87s eta: 8:54:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32045.0 tgs: 268 data_time: 0.90s time: 119.34s eta: 8:55:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.116 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31346.0 tgs: 263 data_time: 0.71s time: 119.15s eta: 8:52:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.132 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 265 data_time: 0.88s time: 119.27s eta: 8:50:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.277 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 266 data_time: 1.16s time: 120.47s eta: 8:54:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 269 data_time: 0.93s time: 119.23s eta: 8:46:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.192 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31145.0 tgs: 261 data_time: 0.57s time: 119.06s eta: 8:43:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.124 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 263 data_time: 0.78s time: 120.83s eta: 8:49:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30849.0 tgs: 256 data_time: 0.61s time: 120.19s eta: 8:44:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.170 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 271 data_time: 0.60s time: 118.83s eta: 8:36:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.211 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 268 data_time: 0.95s time: 119.35s eta: 8:37:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31681.0 tgs: 265 data_time: 0.74s time: 119.33s eta: 8:35:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.256 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31682.0 tgs: 266 data_time: 0.93s time: 119.05s eta: 8:31:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.180 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31886.0 tgs: 267 data_time: 1.12s time: 119.17s eta: 8:30:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31970.0 tgs: 268 data_time: 0.73s time: 119.03s eta: 8:27:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.175 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31706.0 tgs: 266 data_time: 0.63s time: 119.12s eta: 8:26:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.111 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 267 data_time: 0.48s time: 119.24s eta: 8:24:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32274.0 tgs: 270 data_time: 0.89s time: 119.22s eta: 8:22:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31497.0 tgs: 263 data_time: 1.04s time: 119.38s eta: 8:21:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30551.0 tgs: 255 data_time: 0.79s time: 119.78s eta: 8:21:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 265 data_time: 0.72s time: 119.20s eta: 8:16:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.188 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32406.0 tgs: 272 data_time: 0.47s time: 119.02s eta: 8:13:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.143 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 259 data_time: 0.81s time: 119.89s eta: 8:15:33 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.120 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31197.0 tgs: 260 data_time: 0.81s time: 119.56s eta: 8:12:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31241.0 tgs: 261 data_time: 0.49s time: 119.59s eta: 8:10:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31534.0 tgs: 265 data_time: 0.97s time: 118.99s eta: 8:05:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32129.0 tgs: 268 data_time: 0.67s time: 119.84s eta: 8:07:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31216.0 tgs: 260 data_time: 0.74s time: 119.71s eta: 8:04:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32291.0 tgs: 270 data_time: 0.76s time: 119.40s eta: 8:01:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31509.0 tgs: 264 data_time: 0.58s time: 119.28s eta: 7:59:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31600.0 tgs: 265 data_time: 0.81s time: 119.19s eta: 7:56:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.191 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32077.0 tgs: 269 data_time: 0.91s time: 119.20s eta: 7:54:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.095 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31641.0 tgs: 266 data_time: 0.84s time: 118.94s eta: 7:51:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 264 data_time: 0.67s time: 118.97s eta: 7:49:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.85s time: 119.42s eta: 7:49:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31639.0 tgs: 265 data_time: 0.73s time: 119.04s eta: 7:46:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32063.0 tgs: 269 data_time: 0.70s time: 119.09s eta: 7:44:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.178 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31561.0 tgs: 262 data_time: 0.94s time: 120.01s eta: 7:46:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.231 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31630.0 tgs: 263 data_time: 0.80s time: 119.82s eta: 7:43:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 270 data_time: 1.12s time: 119.17s eta: 7:38:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31251.0 tgs: 262 data_time: 0.74s time: 118.96s eta: 7:36:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31497.0 tgs: 263 data_time: 0.81s time: 119.32s eta: 7:35:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.111 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 261 data_time: 0.66s time: 119.38s eta: 7:33:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.284 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31528.0 tgs: 264 data_time: 0.62s time: 119.07s eta: 7:30:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.121 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31551.0 tgs: 265 data_time: 0.67s time: 118.99s eta: 7:28:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31859.0 tgs: 267 data_time: 0.78s time: 119.24s eta: 7:27:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31645.0 tgs: 265 data_time: 0.71s time: 118.99s eta: 7:24:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.112 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31352.0 tgs: 263 data_time: 0.79s time: 118.93s eta: 7:22:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.189 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30731.0 tgs: 258 data_time: 0.81s time: 119.08s eta: 7:20:35 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.117 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30757.0 tgs: 258 data_time: 0.94s time: 118.99s eta: 7:18:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30758.0 tgs: 258 data_time: 0.62s time: 119.03s eta: 7:16:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 266 data_time: 0.63s time: 119.26s eta: 7:15:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.118 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31693.0 tgs: 265 data_time: 0.67s time: 119.37s eta: 7:13:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 271 data_time: 0.92s time: 118.94s eta: 7:10:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32484.0 tgs: 272 data_time: 0.59s time: 119.20s eta: 7:09:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31788.0 tgs: 265 data_time: 0.67s time: 119.58s eta: 7:08:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32134.0 tgs: 268 data_time: 0.85s time: 119.66s eta: 7:06:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30521.0 tgs: 256 data_time: 0.75s time: 118.95s eta: 7:02:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32126.0 tgs: 269 data_time: 0.91s time: 119.30s eta: 7:01:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31708.0 tgs: 265 data_time: 0.69s time: 119.52s eta: 7:00:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31500.0 tgs: 263 data_time: 0.54s time: 119.42s eta: 6:57:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32046.0 tgs: 269 data_time: 0.98s time: 118.91s eta: 6:54:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.232 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32042.0 tgs: 268 data_time: 0.78s time: 119.35s eta: 6:53:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32363.0 tgs: 272 data_time: 0.74s time: 118.92s eta: 6:50:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 28048.0 tgs: 235 data_time: 0.49s time: 119.27s eta: 6:49:29 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31014.0 tgs: 260 data_time: 0.72s time: 119.02s eta: 6:46:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.210 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31741.0 tgs: 266 data_time: 0.81s time: 119.24s eta: 6:45:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32336.0 tgs: 270 data_time: 0.77s time: 119.39s eta: 6:43:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 264 data_time: 0.84s time: 119.01s eta: 6:40:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.96s time: 119.01s eta: 6:38:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32075.0 tgs: 268 data_time: 0.71s time: 119.46s eta: 6:38:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 263 data_time: 0.84s time: 118.80s eta: 6:34:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.169 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 267 data_time: 1.08s time: 119.43s eta: 6:34:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.185 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 265 data_time: 0.93s time: 119.57s eta: 6:32:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31040.0 tgs: 260 data_time: 0.79s time: 119.38s eta: 6:29:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31594.0 tgs: 264 data_time: 0.83s time: 119.62s eta: 6:28:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31690.0 tgs: 265 data_time: 0.80s time: 119.37s eta: 6:25:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30959.0 tgs: 260 data_time: 0.91s time: 118.83s eta: 6:22:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 260 data_time: 0.81s time: 119.00s eta: 6:20:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32499.0 tgs: 272 data_time: 0.75s time: 119.40s eta: 6:20:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31509.0 tgs: 264 data_time: 0.81s time: 119.34s eta: 6:17:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32541.0 tgs: 273 data_time: 0.83s time: 119.05s eta: 6:15:00 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31399.0 tgs: 262 data_time: 0.76s time: 119.41s eta: 6:14:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32182.0 tgs: 269 data_time: 0.72s time: 119.39s eta: 6:12:05 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31938.0 tgs: 267 data_time: 1.10s time: 119.28s eta: 6:09:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31152.0 tgs: 261 data_time: 0.68s time: 119.10s eta: 6:07:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27695.0 tgs: 227 data_time: 0.51s time: 121.96s eta: 6:14:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31417.0 tgs: 263 data_time: 0.56s time: 119.38s eta: 6:04:07 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.188 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31945.0 tgs: 267 data_time: 1.00s time: 119.57s eta: 6:02:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31277.0 tgs: 262 data_time: 0.93s time: 119.26s eta: 5:59:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 270 data_time: 0.97s time: 119.26s eta: 5:57:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31796.0 tgs: 267 data_time: 0.64s time: 118.91s eta: 5:54:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31283.0 tgs: 262 data_time: 0.76s time: 119.02s eta: 5:53:05 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31575.0 tgs: 265 data_time: 0.75s time: 119.07s eta: 5:51:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31841.0 tgs: 266 data_time: 0.56s time: 119.29s eta: 5:49:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31425.0 tgs: 264 data_time: 0.73s time: 118.97s eta: 5:46:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.211 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30675.0 tgs: 257 data_time: 0.87s time: 119.27s eta: 5:45:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31842.0 tgs: 266 data_time: 0.83s time: 119.36s eta: 5:44:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.91s time: 119.08s eta: 5:41:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.116 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31432.0 tgs: 263 data_time: 0.61s time: 119.47s eta: 5:40:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 265 data_time: 0.61s time: 119.38s eta: 5:38:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31867.0 tgs: 266 data_time: 0.87s time: 119.57s eta: 5:36:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.116 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32481.0 tgs: 272 data_time: 0.89s time: 118.98s eta: 5:33:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.206 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31778.0 tgs: 267 data_time: 1.01s time: 118.95s eta: 5:31:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32010.0 tgs: 268 data_time: 0.61s time: 119.24s eta: 5:29:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.129 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30687.0 tgs: 257 data_time: 0.78s time: 119.40s eta: 5:28:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.123 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32287.0 tgs: 270 data_time: 0.77s time: 119.46s eta: 5:26:32 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.186 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31042.0 tgs: 260 data_time: 0.65s time: 118.95s eta: 5:23:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30746.0 tgs: 257 data_time: 0.93s time: 119.63s eta: 5:22:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 266 data_time: 0.71s time: 119.35s eta: 5:20:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30113.0 tgs: 252 data_time: 1.16s time: 119.12s eta: 5:17:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 265 data_time: 1.01s time: 119.42s eta: 5:16:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 266 data_time: 0.55s time: 119.59s eta: 5:14:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.66s time: 119.01s eta: 5:11:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31400.0 tgs: 263 data_time: 0.73s time: 118.99s eta: 5:09:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30693.0 tgs: 257 data_time: 0.66s time: 119.22s eta: 5:07:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31795.0 tgs: 265 data_time: 0.79s time: 119.69s eta: 5:07:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32314.0 tgs: 271 data_time: 0.86s time: 119.11s eta: 5:03:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 272 data_time: 0.78s time: 119.07s eta: 5:01:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 262 data_time: 0.87s time: 121.52s eta: 5:05:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 266 data_time: 0.96s time: 119.57s eta: 4:58:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32384.0 tgs: 270 data_time: 0.74s time: 119.50s eta: 4:56:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31355.0 tgs: 261 data_time: 0.69s time: 119.73s eta: 4:55:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31711.0 tgs: 265 data_time: 0.66s time: 119.52s eta: 4:52:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31718.0 tgs: 265 data_time: 0.84s time: 119.42s eta: 4:50:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31466.0 tgs: 263 data_time: 0.82s time: 119.20s eta: 4:48:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.195 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30970.0 tgs: 260 data_time: 0.99s time: 118.89s eta: 4:45:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32051.0 tgs: 269 data_time: 0.66s time: 119.08s eta: 4:43:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.106 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31215.0 tgs: 260 data_time: 0.83s time: 119.98s eta: 4:43:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.182 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31102.0 tgs: 261 data_time: 0.90s time: 119.15s eta: 4:39:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 265 data_time: 0.89s time: 119.00s eta: 4:37:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.81s time: 119.06s eta: 4:35:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.226 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31468.0 tgs: 263 data_time: 0.84s time: 119.37s eta: 4:34:33 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31405.0 tgs: 263 data_time: 0.73s time: 119.41s eta: 4:32:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31506.0 tgs: 264 data_time: 0.96s time: 118.94s eta: 4:29:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.249 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31662.0 tgs: 264 data_time: 0.96s time: 119.59s eta: 4:29:05 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32441.0 tgs: 272 data_time: 0.68s time: 119.24s eta: 4:26:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32273.0 tgs: 271 data_time: 0.63s time: 119.07s eta: 4:23:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.214 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.95s time: 118.85s eta: 4:21:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 270 data_time: 0.81s time: 118.76s eta: 4:19:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30724.0 tgs: 257 data_time: 0.68s time: 119.43s eta: 4:18:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.250 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31884.0 tgs: 267 data_time: 0.77s time: 119.40s eta: 4:16:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 265 data_time: 0.81s time: 119.36s eta: 4:14:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30653.0 tgs: 257 data_time: 0.59s time: 119.05s eta: 4:11:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31569.0 tgs: 265 data_time: 0.70s time: 119.12s eta: 4:10:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.63s time: 119.13s eta: 4:08:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.233 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31974.0 tgs: 268 data_time: 0.66s time: 118.96s eta: 4:05:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.112 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30777.0 tgs: 258 data_time: 0.86s time: 119.15s eta: 4:04:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.099 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28548.0 tgs: 238 data_time: 0.60s time: 119.50s eta: 4:02:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31850.0 tgs: 267 data_time: 0.72s time: 119.25s eta: 4:00:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 264 data_time: 0.63s time: 119.20s eta: 3:58:23 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31185.0 tgs: 261 data_time: 0.72s time: 119.35s eta: 3:56:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 269 data_time: 0.67s time: 119.37s eta: 3:54:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32087.0 tgs: 267 data_time: 0.58s time: 119.95s eta: 3:53:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.92s time: 119.32s eta: 3:50:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 270 data_time: 0.80s time: 119.63s eta: 3:49:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.117 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 267 data_time: 0.69s time: 118.87s eta: 3:45:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31697.0 tgs: 264 data_time: 0.87s time: 119.72s eta: 3:45:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32151.0 tgs: 269 data_time: 0.87s time: 119.33s eta: 3:42:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31616.0 tgs: 265 data_time: 0.94s time: 119.26s eta: 3:40:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 268 data_time: 0.72s time: 119.00s eta: 3:38:09 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31652.0 tgs: 265 data_time: 0.74s time: 119.42s eta: 3:36:57 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.83s time: 119.03s eta: 3:34:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.75s time: 118.99s eta: 3:32:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30971.0 tgs: 259 data_time: 0.70s time: 119.22s eta: 3:30:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31857.0 tgs: 265 data_time: 1.06s time: 120.11s eta: 3:30:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31749.0 tgs: 266 data_time: 0.65s time: 119.13s eta: 3:26:29 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32341.0 tgs: 271 data_time: 0.79s time: 119.32s eta: 3:24:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31570.0 tgs: 265 data_time: 0.72s time: 119.08s eta: 3:22:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32029.0 tgs: 268 data_time: 0.68s time: 119.43s eta: 3:21:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31717.0 tgs: 265 data_time: 0.52s time: 119.28s eta: 3:18:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32113.0 tgs: 269 data_time: 1.06s time: 119.22s eta: 3:16:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31484.0 tgs: 264 data_time: 0.65s time: 119.18s eta: 3:14:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32141.0 tgs: 270 data_time: 0.72s time: 119.02s eta: 3:12:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31449.0 tgs: 263 data_time: 0.61s time: 119.19s eta: 3:10:41 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32068.0 tgs: 264 data_time: 0.77s time: 121.03s eta: 3:11:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31243.0 tgs: 261 data_time: 0.68s time: 119.51s eta: 3:07:13 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31898.0 tgs: 267 data_time: 0.68s time: 119.26s eta: 3:04:51 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:43:13][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31688.0 tgs: 266 data_time: 0.71s time: 118.99s eta: 3:02:27 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 264 data_time: 0.83s time: 119.22s eta: 3:00:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.231 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31375.0 tgs: 262 data_time: 0.91s time: 119.46s eta: 2:59:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30666.0 tgs: 257 data_time: 0.85s time: 118.98s eta: 2:56:28 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30966.0 tgs: 259 data_time: 0.65s time: 119.35s eta: 2:55:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30984.0 tgs: 259 data_time: 0.76s time: 119.41s eta: 2:53:08 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 260 data_time: 0.98s time: 119.35s eta: 2:51:04 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31662.0 tgs: 260 data_time: 0.81s time: 121.46s eta: 2:52:03 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31340.0 tgs: 262 data_time: 0.93s time: 119.24s eta: 2:46:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.094 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31549.0 tgs: 263 data_time: 0.65s time: 119.57s eta: 2:45:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31700.0 tgs: 266 data_time: 0.95s time: 118.91s eta: 2:42:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31736.0 tgs: 266 data_time: 0.76s time: 119.18s eta: 2:40:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.118 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32339.0 tgs: 271 data_time: 0.75s time: 118.96s eta: 2:38:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31674.0 tgs: 265 data_time: 0.97s time: 119.23s eta: 2:36:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.206 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 267 data_time: 0.63s time: 119.10s eta: 2:34:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29915.0 tgs: 250 data_time: 0.54s time: 119.43s eta: 2:33:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31868.0 tgs: 265 data_time: 0.91s time: 120.22s eta: 2:32:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31974.0 tgs: 268 data_time: 0.90s time: 119.28s eta: 2:29:06 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31924.0 tgs: 268 data_time: 0.93s time: 118.88s eta: 2:26:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.278 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31761.0 tgs: 265 data_time: 1.06s time: 119.52s eta: 2:25:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 271 data_time: 0.77s time: 119.07s eta: 2:22:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30579.0 tgs: 256 data_time: 0.81s time: 119.14s eta: 2:20:58 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31507.0 tgs: 264 data_time: 0.61s time: 119.05s eta: 2:18:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31704.0 tgs: 264 data_time: 0.87s time: 119.84s eta: 2:17:48 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.217 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32298.0 tgs: 270 data_time: 0.73s time: 119.35s eta: 2:15:15 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30607.0 tgs: 256 data_time: 0.88s time: 119.39s eta: 2:13:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30088.0 tgs: 251 data_time: 0.87s time: 119.67s eta: 2:11:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31305.0 tgs: 261 data_time: 0.99s time: 119.49s eta: 2:09:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31969.0 tgs: 268 data_time: 0.71s time: 118.95s eta: 2:06:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 267 data_time: 0.72s time: 119.38s eta: 2:05:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30637.0 tgs: 256 data_time: 0.64s time: 119.27s eta: 2:03:14 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31282.0 tgs: 262 data_time: 0.85s time: 119.22s eta: 2:01:12 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31149.0 tgs: 260 data_time: 0.85s time: 119.50s eta: 1:59:30 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30245.0 tgs: 253 data_time: 0.71s time: 119.42s eta: 1:57:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31829.0 tgs: 266 data_time: 0.70s time: 119.54s eta: 1:55:33 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 267 data_time: 0.75s time: 119.30s eta: 1:53:20 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.206 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.86s time: 119.25s eta: 1:51:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.194 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31406.0 tgs: 264 data_time: 0.81s time: 118.93s eta: 1:49:01 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32043.0 tgs: 267 data_time: 0.54s time: 119.64s eta: 1:47:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 266 data_time: 0.97s time: 119.71s eta: 1:45:44 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31297.0 tgs: 262 data_time: 0.75s time: 119.35s eta: 1:43:26 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31855.0 tgs: 267 data_time: 0.63s time: 119.30s eta: 1:41:24 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32159.0 tgs: 269 data_time: 0.69s time: 119.28s eta: 1:39:23 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31392.0 tgs: 264 data_time: 0.73s time: 118.91s eta: 1:37:06 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29984.0 tgs: 251 data_time: 0.67s time: 119.16s eta: 1:35:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32259.0 tgs: 270 data_time: 0.80s time: 119.06s eta: 1:33:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 264 data_time: 0.91s time: 119.25s eta: 1:31:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31849.0 tgs: 266 data_time: 0.75s time: 119.72s eta: 1:29:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.220 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30912.0 tgs: 259 data_time: 0.73s time: 119.02s eta: 1:27:16 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31331.0 tgs: 263 data_time: 0.56s time: 118.86s eta: 1:25:11 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31823.0 tgs: 266 data_time: 0.63s time: 119.48s eta: 1:23:37 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.254 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.80s time: 119.48s eta: 1:21:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 268 data_time: 0.91s time: 118.95s eta: 1:19:17 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.146 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32267.0 tgs: 271 data_time: 1.01s time: 118.93s eta: 1:17:18 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32411.0 tgs: 265 data_time: 0.74s time: 122.16s eta: 1:17:22 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30612.0 tgs: 256 data_time: 0.66s time: 119.36s eta: 1:13:36 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.191 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30724.0 tgs: 258 data_time: 0.72s time: 119.05s eta: 1:11:25 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31463.0 tgs: 264 data_time: 0.64s time: 118.91s eta: 1:09:21 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32107.0 tgs: 269 data_time: 1.03s time: 119.10s eta: 1:07:29 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31192.0 tgs: 261 data_time: 0.69s time: 119.41s eta: 1:05:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.86s time: 119.34s eta: 1:03:39 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32548.0 tgs: 273 data_time: 0.85s time: 119.18s eta: 1:01:34 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32385.0 tgs: 271 data_time: 0.82s time: 119.27s eta: 0:59:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31658.0 tgs: 265 data_time: 0.84s time: 119.44s eta: 0:57:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31779.0 tgs: 266 data_time: 0.73s time: 119.40s eta: 0:55:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32641.0 tgs: 273 data_time: 0.62s time: 119.48s eta: 0:53:45 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.107 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30724.0 tgs: 257 data_time: 0.71s time: 119.36s eta: 0:51:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31732.0 tgs: 265 data_time: 0.82s time: 119.46s eta: 0:49:46 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31603.0 tgs: 265 data_time: 0.71s time: 118.90s eta: 0:47:33 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31132.0 tgs: 261 data_time: 0.72s time: 119.28s eta: 0:45:43 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31350.0 tgs: 262 data_time: 0.85s time: 119.21s eta: 0:43:42 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32332.0 tgs: 271 data_time: 0.68s time: 118.99s eta: 0:41:38 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31688.0 tgs: 266 data_time: 0.73s time: 119.03s eta: 0:39:40 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30402.0 tgs: 254 data_time: 0.70s time: 119.66s eta: 0:37:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31274.0 tgs: 262 data_time: 0.52s time: 119.28s eta: 0:35:47 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32075.0 tgs: 267 data_time: 0.65s time: 119.74s eta: 0:33:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32578.0 tgs: 268 data_time: 0.61s time: 121.23s eta: 0:32:19 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 266 data_time: 0.63s time: 119.37s eta: 0:29:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31789.0 tgs: 266 data_time: 0.79s time: 119.32s eta: 0:27:50 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 266 data_time: 0.72s time: 119.64s eta: 0:25:55 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31962.0 tgs: 268 data_time: 1.12s time: 119.10s eta: 0:23:49 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31704.0 tgs: 265 data_time: 1.03s time: 119.39s eta: 0:21:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.198 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 270 data_time: 0.97s time: 119.25s eta: 0:19:52 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31685.0 tgs: 265 data_time: 0.57s time: 119.25s eta: 0:17:53 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32089.0 tgs: 267 data_time: 0.66s time: 119.91s eta: 0:15:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30888.0 tgs: 259 data_time: 0.67s time: 119.21s eta: 0:13:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 269 data_time: 0.97s time: 119.09s eta: 0:11:54 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.206 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.75s time: 119.32s eta: 0:09:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32520.0 tgs: 272 data_time: 0.75s time: 119.16s eta: 0:07:56 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 270 data_time: 0.71s time: 119.69s eta: 0:05:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31730.0 tgs: 261 data_time: 0.78s time: 121.45s eta: 0:04:02 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31112.0 tgs: 260 data_time: 0.64s time: 119.21s eta: 0:01:59 +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 60][DP 15][SP 0][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank61.log b/20250121165312/rank61.log new file mode 100644 index 0000000000000000000000000000000000000000..75d84eccb7ebf16c98f0709c5ba68d482860ad81 --- /dev/null +++ b/20250121165312/rank61.log @@ -0,0 +1,851 @@ +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:19][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.05s +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 71.35 seconds, peak gpu memory 13.4G +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.233 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.9GB text_tokens: 31861.0 tgs: 208 data_time: 1.82s time: 152.86s eta: 1 day, 1:10:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.264 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31345.0 tgs: 260 data_time: 0.63s time: 120.21s eta: 19:46:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.230 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30897.0 tgs: 258 data_time: 0.80s time: 119.68s eta: 19:38:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30727.0 tgs: 256 data_time: 0.80s time: 119.65s eta: 19:36:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.218 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31993.0 tgs: 268 data_time: 0.55s time: 119.31s eta: 19:31:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.219 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32528.0 tgs: 272 data_time: 0.70s time: 119.53s eta: 19:31:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.224 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 260 data_time: 0.57s time: 122.97s eta: 20:03:02 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.277 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 270 data_time: 0.74s time: 119.88s eta: 19:30:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.199 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 255 data_time: 0.66s time: 120.65s eta: 19:36:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.226 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32133.0 tgs: 267 data_time: 1.07s time: 119.92s eta: 19:27:12 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31934.0 tgs: 265 data_time: 0.76s time: 120.46s eta: 19:30:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31332.0 tgs: 262 data_time: 0.81s time: 119.34s eta: 19:17:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32182.0 tgs: 268 data_time: 0.85s time: 119.68s eta: 19:18:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30812.0 tgs: 258 data_time: 0.57s time: 119.20s eta: 19:12:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32580.0 tgs: 273 data_time: 0.66s time: 119.21s eta: 19:10:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30706.0 tgs: 258 data_time: 0.76s time: 118.93s eta: 19:05:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.280 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 268 data_time: 0.78s time: 119.15s eta: 19:05:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.239 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32212.0 tgs: 270 data_time: 0.83s time: 119.17s eta: 19:04:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 267 data_time: 0.89s time: 119.40s eta: 19:04:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31887.0 tgs: 267 data_time: 0.80s time: 119.33s eta: 19:01:32 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31686.0 tgs: 263 data_time: 0.95s time: 120.31s eta: 19:08:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31729.0 tgs: 258 data_time: 0.70s time: 122.91s eta: 19:31:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31873.0 tgs: 266 data_time: 0.75s time: 119.48s eta: 18:57:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31838.0 tgs: 267 data_time: 0.75s time: 119.16s eta: 18:52:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 267 data_time: 0.84s time: 119.24s eta: 18:50:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31822.0 tgs: 266 data_time: 0.70s time: 119.39s eta: 18:50:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.226 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32115.0 tgs: 268 data_time: 1.11s time: 119.66s eta: 18:50:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32433.0 tgs: 271 data_time: 0.81s time: 119.39s eta: 18:46:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31453.0 tgs: 262 data_time: 0.88s time: 119.79s eta: 18:47:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32050.0 tgs: 247 data_time: 0.85s time: 129.31s eta: 20:15:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.251 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32259.0 tgs: 269 data_time: 0.65s time: 119.57s eta: 18:41:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.71s time: 119.39s eta: 18:38:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30717.0 tgs: 254 data_time: 0.73s time: 120.66s eta: 18:48:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.256 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31683.0 tgs: 265 data_time: 0.83s time: 119.40s eta: 18:34:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.264 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32425.0 tgs: 272 data_time: 1.04s time: 119.16s eta: 18:30:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.133 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31127.0 tgs: 261 data_time: 0.62s time: 119.07s eta: 18:27:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31857.0 tgs: 267 data_time: 0.75s time: 119.13s eta: 18:25:57 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31766.0 tgs: 263 data_time: 0.75s time: 120.39s eta: 18:35:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32158.0 tgs: 265 data_time: 0.83s time: 121.17s eta: 18:40:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32032.0 tgs: 265 data_time: 0.79s time: 120.66s eta: 18:34:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31101.0 tgs: 259 data_time: 0.71s time: 119.74s eta: 18:23:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30739.0 tgs: 248 data_time: 0.88s time: 123.87s eta: 18:59:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32313.0 tgs: 270 data_time: 0.79s time: 119.55s eta: 18:17:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30208.0 tgs: 253 data_time: 0.65s time: 119.36s eta: 18:14:07 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 264 data_time: 0.86s time: 119.22s eta: 18:10:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31887.0 tgs: 266 data_time: 0.86s time: 119.60s eta: 18:12:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31263.0 tgs: 262 data_time: 0.66s time: 119.26s eta: 18:07:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 268 data_time: 0.86s time: 119.01s eta: 18:02:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.177 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32230.0 tgs: 270 data_time: 1.18s time: 119.11s eta: 18:01:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31041.0 tgs: 260 data_time: 0.98s time: 119.27s eta: 18:01:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31939.0 tgs: 267 data_time: 0.61s time: 119.32s eta: 17:59:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31788.0 tgs: 264 data_time: 1.12s time: 120.02s eta: 18:04:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32224.0 tgs: 270 data_time: 0.68s time: 119.10s eta: 17:53:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32035.0 tgs: 268 data_time: 0.71s time: 119.42s eta: 17:54:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30404.0 tgs: 255 data_time: 0.70s time: 118.99s eta: 17:48:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32351.0 tgs: 271 data_time: 1.02s time: 119.02s eta: 17:47:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32058.0 tgs: 255 data_time: 0.64s time: 125.37s eta: 18:42:02 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 267 data_time: 0.89s time: 120.10s eta: 17:52:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 263 data_time: 0.66s time: 119.40s eta: 17:44:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31503.0 tgs: 263 data_time: 0.80s time: 119.33s eta: 17:42:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32460.0 tgs: 271 data_time: 0.69s time: 119.51s eta: 17:41:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 251 data_time: 0.41s time: 126.04s eta: 18:37:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 260 data_time: 0.68s time: 122.63s eta: 18:05:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31721.0 tgs: 261 data_time: 0.76s time: 121.16s eta: 17:50:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32329.0 tgs: 271 data_time: 0.68s time: 119.03s eta: 17:29:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31333.0 tgs: 263 data_time: 0.77s time: 118.86s eta: 17:25:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 27236.0 tgs: 219 data_time: 0.59s time: 123.99s eta: 18:09:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 258 data_time: 0.80s time: 124.57s eta: 18:12:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.181 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 271 data_time: 0.95s time: 119.40s eta: 17:24:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31521.0 tgs: 263 data_time: 0.91s time: 119.69s eta: 17:25:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.204 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31154.0 tgs: 260 data_time: 0.62s time: 119.53s eta: 17:21:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31399.0 tgs: 264 data_time: 0.75s time: 118.86s eta: 17:14:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31343.0 tgs: 261 data_time: 0.83s time: 119.84s eta: 17:20:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31053.0 tgs: 259 data_time: 0.75s time: 119.56s eta: 17:16:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32475.0 tgs: 266 data_time: 0.87s time: 121.86s eta: 17:34:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.147 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31529.0 tgs: 255 data_time: 0.74s time: 123.51s eta: 17:46:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31108.0 tgs: 261 data_time: 0.64s time: 119.01s eta: 17:05:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32276.0 tgs: 270 data_time: 0.81s time: 119.23s eta: 17:05:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31566.0 tgs: 264 data_time: 0.89s time: 119.14s eta: 17:02:36 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31630.0 tgs: 265 data_time: 0.74s time: 119.25s eta: 17:01:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32522.0 tgs: 273 data_time: 0.94s time: 119.04s eta: 16:57:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30869.0 tgs: 258 data_time: 0.73s time: 119.28s eta: 16:57:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.115 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31254.0 tgs: 262 data_time: 0.77s time: 119.16s eta: 16:54:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28967.0 tgs: 242 data_time: 0.85s time: 119.59s eta: 16:56:29 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.70s time: 119.13s eta: 16:50:34 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31381.0 tgs: 264 data_time: 0.51s time: 118.84s eta: 16:46:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.160 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31802.0 tgs: 267 data_time: 0.66s time: 119.01s eta: 16:45:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31937.0 tgs: 266 data_time: 0.78s time: 119.85s eta: 16:50:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31499.0 tgs: 263 data_time: 0.89s time: 119.37s eta: 16:44:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31043.0 tgs: 260 data_time: 0.64s time: 119.01s eta: 16:39:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30452.0 tgs: 251 data_time: 1.01s time: 120.93s eta: 16:53:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30269.0 tgs: 254 data_time: 0.60s time: 118.99s eta: 16:35:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.220 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 266 data_time: 0.87s time: 119.28s eta: 16:36:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31521.0 tgs: 264 data_time: 1.12s time: 119.23s eta: 16:33:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.194 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31815.0 tgs: 266 data_time: 0.98s time: 119.36s eta: 16:32:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30661.0 tgs: 256 data_time: 0.72s time: 119.70s eta: 16:33:31 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31965.0 tgs: 268 data_time: 0.90s time: 119.20s eta: 16:27:22 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31853.0 tgs: 265 data_time: 0.64s time: 119.77s eta: 16:30:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31960.0 tgs: 265 data_time: 0.69s time: 120.23s eta: 16:31:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 266 data_time: 0.79s time: 118.97s eta: 16:19:29 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31596.0 tgs: 248 data_time: 0.66s time: 127.33s eta: 17:26:12 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 265 data_time: 0.74s time: 119.22s eta: 16:17:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.121 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31940.0 tgs: 268 data_time: 0.60s time: 119.14s eta: 16:14:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.174 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31867.0 tgs: 267 data_time: 1.04s time: 119.24s eta: 16:13:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 264 data_time: 0.75s time: 119.18s eta: 16:11:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30332.0 tgs: 255 data_time: 0.90s time: 118.94s eta: 16:07:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31154.0 tgs: 262 data_time: 0.61s time: 118.86s eta: 16:04:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.126 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31244.0 tgs: 262 data_time: 0.56s time: 118.96s eta: 16:03:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 259 data_time: 0.91s time: 121.84s eta: 16:24:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.80s time: 118.99s eta: 15:59:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 268 data_time: 0.90s time: 119.22s eta: 15:59:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32297.0 tgs: 270 data_time: 0.78s time: 119.62s eta: 16:00:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.244 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31976.0 tgs: 268 data_time: 0.85s time: 119.04s eta: 15:54:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32206.0 tgs: 270 data_time: 0.93s time: 118.97s eta: 15:51:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.137 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 267 data_time: 0.65s time: 119.17s eta: 15:51:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32352.0 tgs: 270 data_time: 0.56s time: 119.64s eta: 15:53:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.191 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31195.0 tgs: 262 data_time: 0.68s time: 118.94s eta: 15:45:32 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32381.0 tgs: 271 data_time: 0.59s time: 119.29s eta: 15:46:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.188 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 269 data_time: 1.16s time: 119.12s eta: 15:43:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 267 data_time: 0.72s time: 119.00s eta: 15:40:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31534.0 tgs: 263 data_time: 0.83s time: 119.66s eta: 15:43:16 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.153 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30832.0 tgs: 253 data_time: 0.65s time: 121.51s eta: 15:55:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32197.0 tgs: 269 data_time: 0.61s time: 119.41s eta: 15:37:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.182 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31745.0 tgs: 244 data_time: 0.80s time: 129.97s eta: 16:58:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31717.0 tgs: 265 data_time: 1.09s time: 119.46s eta: 15:33:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.210 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31536.0 tgs: 263 data_time: 0.58s time: 119.61s eta: 15:32:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31300.0 tgs: 261 data_time: 0.89s time: 119.52s eta: 15:30:18 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31783.0 tgs: 260 data_time: 0.74s time: 122.14s eta: 15:48:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32422.0 tgs: 268 data_time: 0.86s time: 120.75s eta: 15:35:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30462.0 tgs: 255 data_time: 0.86s time: 119.42s eta: 15:23:30 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 269 data_time: 0.84s time: 119.44s eta: 15:21:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30621.0 tgs: 232 data_time: 0.93s time: 131.63s eta: 16:53:31 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31534.0 tgs: 264 data_time: 0.89s time: 119.29s eta: 15:16:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30926.0 tgs: 258 data_time: 0.60s time: 119.56s eta: 15:16:36 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 262 data_time: 0.71s time: 122.25s eta: 15:35:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30268.0 tgs: 252 data_time: 0.70s time: 119.71s eta: 15:13:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 264 data_time: 0.74s time: 119.35s eta: 15:09:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.212 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30984.0 tgs: 260 data_time: 0.91s time: 118.90s eta: 15:03:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.184 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 1.03s time: 118.99s eta: 15:02:22 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31140.0 tgs: 260 data_time: 0.71s time: 119.31s eta: 15:02:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31989.0 tgs: 269 data_time: 0.72s time: 118.91s eta: 14:57:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32328.0 tgs: 268 data_time: 0.54s time: 120.52s eta: 15:07:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.176 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31613.0 tgs: 264 data_time: 0.67s time: 119.31s eta: 14:56:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.210 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31698.0 tgs: 258 data_time: 0.67s time: 122.65s eta: 15:19:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31581.0 tgs: 265 data_time: 1.03s time: 119.12s eta: 14:51:26 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.193 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31926.0 tgs: 266 data_time: 0.70s time: 119.75s eta: 14:54:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31461.0 tgs: 263 data_time: 0.68s time: 119.47s eta: 14:50:00 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30414.0 tgs: 255 data_time: 0.69s time: 118.99s eta: 14:44:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32174.0 tgs: 269 data_time: 0.95s time: 119.33s eta: 14:45:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.112 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30963.0 tgs: 260 data_time: 0.56s time: 118.94s eta: 14:40:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 0.73s time: 119.05s eta: 14:38:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31750.0 tgs: 266 data_time: 0.80s time: 119.06s eta: 14:37:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.192 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32234.0 tgs: 270 data_time: 0.83s time: 119.12s eta: 14:35:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31496.0 tgs: 264 data_time: 0.70s time: 119.00s eta: 14:32:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30789.0 tgs: 258 data_time: 0.82s time: 119.07s eta: 14:31:12 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.154 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31168.0 tgs: 261 data_time: 0.82s time: 119.15s eta: 14:29:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.149 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31841.0 tgs: 266 data_time: 0.90s time: 119.65s eta: 14:31:26 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31532.0 tgs: 264 data_time: 0.70s time: 119.18s eta: 14:26:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32474.0 tgs: 272 data_time: 0.88s time: 119.20s eta: 14:24:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 265 data_time: 0.63s time: 119.39s eta: 14:23:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.229 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31173.0 tgs: 258 data_time: 1.10s time: 120.73s eta: 14:31:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32536.0 tgs: 273 data_time: 0.87s time: 118.85s eta: 14:15:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 269 data_time: 0.75s time: 119.10s eta: 14:15:32 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32038.0 tgs: 268 data_time: 0.66s time: 119.15s eta: 14:13:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 266 data_time: 0.77s time: 119.96s eta: 14:17:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31272.0 tgs: 260 data_time: 0.89s time: 119.86s eta: 14:15:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.122 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32263.0 tgs: 269 data_time: 0.66s time: 119.59s eta: 14:11:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31832.0 tgs: 267 data_time: 0.77s time: 119.22s eta: 14:06:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31974.0 tgs: 268 data_time: 0.83s time: 119.26s eta: 14:04:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31337.0 tgs: 262 data_time: 0.77s time: 119.27s eta: 14:02:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.181 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32038.0 tgs: 268 data_time: 0.89s time: 119.14s eta: 13:59:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.196 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30681.0 tgs: 257 data_time: 0.80s time: 119.12s eta: 13:57:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.204 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31452.0 tgs: 263 data_time: 0.66s time: 119.17s eta: 13:56:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.150 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31620.0 tgs: 264 data_time: 1.07s time: 119.35s eta: 13:55:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31468.0 tgs: 261 data_time: 0.71s time: 120.11s eta: 13:58:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31883.0 tgs: 260 data_time: 0.48s time: 122.55s eta: 14:13:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.172 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31475.0 tgs: 264 data_time: 0.71s time: 119.10s eta: 13:47:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.142 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 268 data_time: 0.38s time: 119.38s eta: 13:47:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.199 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32030.0 tgs: 268 data_time: 1.08s time: 119.34s eta: 13:45:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.124 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32278.0 tgs: 270 data_time: 0.93s time: 119.33s eta: 13:43:22 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30973.0 tgs: 260 data_time: 0.69s time: 119.11s eta: 13:39:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31679.0 tgs: 265 data_time: 0.78s time: 119.25s eta: 13:38:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 269 data_time: 0.84s time: 119.13s eta: 13:36:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.138 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31454.0 tgs: 263 data_time: 0.70s time: 119.22s eta: 13:34:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.212 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.74s time: 119.46s eta: 13:34:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 267 data_time: 0.97s time: 119.08s eta: 13:29:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31043.0 tgs: 261 data_time: 0.83s time: 118.92s eta: 13:26:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31949.0 tgs: 266 data_time: 0.68s time: 119.81s eta: 13:30:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.184 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31673.0 tgs: 265 data_time: 0.86s time: 119.25s eta: 13:24:57 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30259.0 tgs: 253 data_time: 0.79s time: 119.16s eta: 13:22:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31588.0 tgs: 264 data_time: 0.78s time: 119.36s eta: 13:21:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.221 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 271 data_time: 0.77s time: 119.25s eta: 13:18:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.114 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31297.0 tgs: 262 data_time: 0.72s time: 119.04s eta: 13:15:34 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.146 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31956.0 tgs: 267 data_time: 0.72s time: 119.37s eta: 13:15:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.191 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 269 data_time: 0.69s time: 119.19s eta: 13:12:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.144 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30991.0 tgs: 259 data_time: 1.23s time: 119.45s eta: 13:12:22 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.209 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31552.0 tgs: 264 data_time: 0.78s time: 119.29s eta: 13:09:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31575.0 tgs: 265 data_time: 0.82s time: 119.14s eta: 13:06:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 264 data_time: 0.60s time: 120.48s eta: 13:13:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 265 data_time: 0.85s time: 120.16s eta: 13:09:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.213 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31793.0 tgs: 266 data_time: 0.91s time: 119.50s eta: 13:02:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 267 data_time: 0.67s time: 119.51s eta: 13:00:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 264 data_time: 0.70s time: 119.57s eta: 12:59:12 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 257 data_time: 0.65s time: 119.79s eta: 12:58:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.233 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31551.0 tgs: 265 data_time: 0.70s time: 119.04s eta: 12:51:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.212 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 269 data_time: 0.83s time: 119.05s eta: 12:49:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31521.0 tgs: 264 data_time: 0.77s time: 119.13s eta: 12:48:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31568.0 tgs: 264 data_time: 0.75s time: 119.40s eta: 12:48:07 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31459.0 tgs: 263 data_time: 0.76s time: 119.18s eta: 12:44:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 260 data_time: 0.78s time: 119.43s eta: 12:44:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.213 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 265 data_time: 0.84s time: 119.10s eta: 12:40:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31535.0 tgs: 264 data_time: 0.74s time: 119.07s eta: 12:38:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.65s time: 119.59s eta: 12:39:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30891.0 tgs: 260 data_time: 0.84s time: 118.78s eta: 12:32:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32258.0 tgs: 271 data_time: 0.89s time: 118.89s eta: 12:30:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31925.0 tgs: 267 data_time: 0.79s time: 119.17s eta: 12:30:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31125.0 tgs: 260 data_time: 0.90s time: 119.26s eta: 12:29:22 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.117 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30968.0 tgs: 259 data_time: 1.01s time: 119.35s eta: 12:27:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31570.0 tgs: 265 data_time: 0.63s time: 119.11s eta: 12:24:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31066.0 tgs: 260 data_time: 0.82s time: 119.34s eta: 12:23:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32253.0 tgs: 270 data_time: 0.68s time: 119.26s eta: 12:21:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 0.77s time: 119.03s eta: 12:17:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 32034.0 tgs: 264 data_time: 0.95s time: 121.13s eta: 12:28:57 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31566.0 tgs: 263 data_time: 0.76s time: 119.99s eta: 12:19:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.203 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31032.0 tgs: 261 data_time: 0.76s time: 118.73s eta: 12:10:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.193 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30435.0 tgs: 254 data_time: 1.07s time: 119.63s eta: 12:13:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30462.0 tgs: 254 data_time: 0.80s time: 119.83s eta: 12:12:57 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 0.64s time: 119.14s eta: 12:06:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.190 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 266 data_time: 0.96s time: 119.17s eta: 12:04:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.141 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30964.0 tgs: 259 data_time: 0.82s time: 119.11s eta: 12:02:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.129 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32255.0 tgs: 270 data_time: 0.71s time: 119.19s eta: 12:01:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.121 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31354.0 tgs: 263 data_time: 0.77s time: 118.96s eta: 11:57:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.147 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31553.0 tgs: 263 data_time: 0.84s time: 119.70s eta: 12:00:12 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 0.87s time: 119.27s eta: 11:55:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.67s time: 119.27s eta: 11:53:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.205 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32032.0 tgs: 267 data_time: 0.86s time: 119.71s eta: 11:54:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.182 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31660.0 tgs: 266 data_time: 1.03s time: 118.78s eta: 11:46:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.207 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30889.0 tgs: 259 data_time: 0.84s time: 119.15s eta: 11:46:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31706.0 tgs: 265 data_time: 0.64s time: 119.61s eta: 11:47:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32249.0 tgs: 271 data_time: 0.89s time: 118.85s eta: 11:41:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 270 data_time: 0.73s time: 118.68s eta: 11:38:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32441.0 tgs: 272 data_time: 0.56s time: 119.21s eta: 11:39:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.70s time: 119.14s eta: 11:36:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32027.0 tgs: 268 data_time: 0.76s time: 119.08s eta: 11:34:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30730.0 tgs: 258 data_time: 1.09s time: 118.94s eta: 11:31:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32053.0 tgs: 269 data_time: 0.99s time: 118.91s eta: 11:29:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.104 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31359.0 tgs: 263 data_time: 0.59s time: 118.97s eta: 11:28:02 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32010.0 tgs: 268 data_time: 0.54s time: 119.28s eta: 11:27:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.207 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31442.0 tgs: 263 data_time: 0.68s time: 119.36s eta: 11:26:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30872.0 tgs: 259 data_time: 0.78s time: 118.88s eta: 11:21:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30977.0 tgs: 259 data_time: 0.99s time: 119.21s eta: 11:21:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.216 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32183.0 tgs: 268 data_time: 0.84s time: 119.71s eta: 11:22:22 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.214 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31874.0 tgs: 267 data_time: 0.85s time: 119.12s eta: 11:16:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.179 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31635.0 tgs: 265 data_time: 1.03s time: 119.12s eta: 11:15:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.113 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31496.0 tgs: 264 data_time: 0.89s time: 118.98s eta: 11:12:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.63s time: 119.09s eta: 11:10:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.126 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30964.0 tgs: 259 data_time: 0.55s time: 119.15s eta: 11:09:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.134 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32332.0 tgs: 271 data_time: 0.74s time: 118.97s eta: 11:06:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.215 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30372.0 tgs: 254 data_time: 1.05s time: 119.47s eta: 11:07:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31452.0 tgs: 262 data_time: 0.88s time: 119.64s eta: 11:06:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.175 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31934.0 tgs: 267 data_time: 0.77s time: 119.30s eta: 11:02:06 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32152.0 tgs: 269 data_time: 0.84s time: 119.31s eta: 11:00:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.253 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30536.0 tgs: 256 data_time: 0.77s time: 118.83s eta: 10:55:31 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31781.0 tgs: 266 data_time: 0.87s time: 119.47s eta: 10:57:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 260 data_time: 0.68s time: 119.36s eta: 10:54:31 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31518.0 tgs: 264 data_time: 0.72s time: 119.28s eta: 10:52:02 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32102.0 tgs: 269 data_time: 0.69s time: 119.25s eta: 10:49:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31697.0 tgs: 266 data_time: 0.90s time: 119.10s eta: 10:47:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.177 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 269 data_time: 0.84s time: 119.53s eta: 10:47:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.121 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32597.0 tgs: 273 data_time: 0.68s time: 119.33s eta: 10:44:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.106 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 264 data_time: 0.80s time: 119.49s eta: 10:43:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.62s time: 119.51s eta: 10:41:22 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.244 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 269 data_time: 1.01s time: 118.99s eta: 10:36:36 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.197 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32356.0 tgs: 271 data_time: 0.86s time: 119.23s eta: 10:35:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.168 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31688.0 tgs: 266 data_time: 0.59s time: 119.05s eta: 10:32:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.127 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32462.0 tgs: 273 data_time: 0.83s time: 118.78s eta: 10:29:31 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.142 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32143.0 tgs: 268 data_time: 0.76s time: 119.86s eta: 10:33:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.156 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31922.0 tgs: 267 data_time: 0.87s time: 119.38s eta: 10:28:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.159 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 271 data_time: 0.77s time: 118.84s eta: 10:23:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.186 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30668.0 tgs: 257 data_time: 0.87s time: 118.98s eta: 10:22:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.134 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31834.0 tgs: 266 data_time: 0.76s time: 119.29s eta: 10:22:16 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.105 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29624.0 tgs: 248 data_time: 0.69s time: 119.20s eta: 10:19:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30908.0 tgs: 259 data_time: 0.67s time: 119.06s eta: 10:17:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.111 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 264 data_time: 0.58s time: 119.47s eta: 10:17:16 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.133 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30306.0 tgs: 253 data_time: 0.67s time: 119.39s eta: 10:14:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.210 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31906.0 tgs: 267 data_time: 0.58s time: 119.45s eta: 10:13:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30612.0 tgs: 256 data_time: 0.92s time: 119.29s eta: 10:10:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32081.0 tgs: 269 data_time: 0.65s time: 119.11s eta: 10:07:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.193 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 264 data_time: 0.53s time: 119.70s eta: 10:08:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.157 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30197.0 tgs: 252 data_time: 0.71s time: 119.83s eta: 10:07:07 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.217 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32281.0 tgs: 265 data_time: 1.01s time: 121.55s eta: 10:13:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.252 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 266 data_time: 0.92s time: 119.25s eta: 10:00:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.211 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31464.0 tgs: 264 data_time: 0.81s time: 118.81s eta: 9:56:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30866.0 tgs: 258 data_time: 0.78s time: 119.56s eta: 9:57:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.138 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 268 data_time: 0.62s time: 119.71s eta: 9:56:32 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30739.0 tgs: 257 data_time: 0.65s time: 119.25s eta: 9:52:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.203 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31790.0 tgs: 266 data_time: 0.74s time: 119.33s eta: 9:50:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30480.0 tgs: 255 data_time: 0.74s time: 119.22s eta: 9:48:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31890.0 tgs: 265 data_time: 0.82s time: 120.10s eta: 9:50:30 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.175 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.54s time: 119.27s eta: 9:44:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.228 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31951.0 tgs: 267 data_time: 0.97s time: 119.46s eta: 9:43:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.200 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 268 data_time: 0.83s time: 119.54s eta: 9:41:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 263 data_time: 0.62s time: 119.20s eta: 9:38:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 270 data_time: 0.80s time: 119.48s eta: 9:37:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31107.0 tgs: 261 data_time: 0.55s time: 119.06s eta: 9:33:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 265 data_time: 0.74s time: 119.25s eta: 9:32:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31613.0 tgs: 265 data_time: 0.67s time: 119.15s eta: 9:29:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31659.0 tgs: 265 data_time: 0.77s time: 119.22s eta: 9:28:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31783.0 tgs: 266 data_time: 1.13s time: 119.46s eta: 9:27:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32338.0 tgs: 271 data_time: 0.49s time: 118.99s eta: 9:23:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 268 data_time: 0.70s time: 119.03s eta: 9:21:26 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29882.0 tgs: 249 data_time: 0.56s time: 119.53s eta: 9:21:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.124 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 266 data_time: 0.66s time: 119.59s eta: 9:20:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31682.0 tgs: 265 data_time: 0.81s time: 119.12s eta: 9:15:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.127 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31960.0 tgs: 268 data_time: 0.69s time: 119.06s eta: 9:13:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.160 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.81s time: 118.88s eta: 9:10:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31547.0 tgs: 264 data_time: 0.74s time: 119.32s eta: 9:10:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30915.0 tgs: 259 data_time: 0.77s time: 119.09s eta: 9:07:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32237.0 tgs: 271 data_time: 0.72s time: 118.86s eta: 9:04:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.245 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31751.0 tgs: 265 data_time: 0.83s time: 119.66s eta: 9:06:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.120 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31612.0 tgs: 264 data_time: 0.76s time: 119.43s eta: 9:03:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.120 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31026.0 tgs: 259 data_time: 0.95s time: 119.43s eta: 9:01:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31302.0 tgs: 263 data_time: 0.76s time: 118.74s eta: 8:56:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31825.0 tgs: 267 data_time: 0.68s time: 118.87s eta: 8:54:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32045.0 tgs: 268 data_time: 0.87s time: 119.33s eta: 8:55:00 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31346.0 tgs: 263 data_time: 0.68s time: 119.14s eta: 8:52:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 265 data_time: 0.85s time: 119.27s eta: 8:50:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.270 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 266 data_time: 1.15s time: 120.48s eta: 8:54:06 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 269 data_time: 0.91s time: 119.23s eta: 8:46:36 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.126 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31145.0 tgs: 261 data_time: 0.56s time: 119.06s eta: 8:43:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 263 data_time: 0.76s time: 120.83s eta: 8:49:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30849.0 tgs: 256 data_time: 0.60s time: 120.19s eta: 8:44:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 271 data_time: 0.60s time: 118.83s eta: 8:36:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 268 data_time: 0.93s time: 119.35s eta: 8:37:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.190 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31681.0 tgs: 265 data_time: 0.73s time: 119.33s eta: 8:35:06 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31682.0 tgs: 266 data_time: 0.91s time: 119.05s eta: 8:31:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.211 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31886.0 tgs: 267 data_time: 1.09s time: 119.16s eta: 8:30:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31970.0 tgs: 268 data_time: 0.71s time: 119.05s eta: 8:27:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31706.0 tgs: 266 data_time: 0.59s time: 119.12s eta: 8:26:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 267 data_time: 0.45s time: 119.24s eta: 8:24:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.177 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32274.0 tgs: 270 data_time: 0.88s time: 119.22s eta: 8:22:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.181 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31497.0 tgs: 263 data_time: 1.00s time: 119.38s eta: 8:21:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30551.0 tgs: 255 data_time: 0.77s time: 119.78s eta: 8:21:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 265 data_time: 0.70s time: 119.20s eta: 8:16:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32406.0 tgs: 272 data_time: 0.45s time: 119.02s eta: 8:13:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 259 data_time: 0.78s time: 119.89s eta: 8:15:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31197.0 tgs: 260 data_time: 0.76s time: 119.55s eta: 8:12:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31241.0 tgs: 261 data_time: 0.45s time: 119.59s eta: 8:10:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31534.0 tgs: 264 data_time: 0.97s time: 119.00s eta: 8:05:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32129.0 tgs: 268 data_time: 0.63s time: 119.84s eta: 8:07:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.186 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31216.0 tgs: 260 data_time: 0.72s time: 119.71s eta: 8:04:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32291.0 tgs: 270 data_time: 0.73s time: 119.41s eta: 8:01:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.147 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31509.0 tgs: 264 data_time: 0.56s time: 119.28s eta: 7:59:06 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.117 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31600.0 tgs: 265 data_time: 0.77s time: 119.19s eta: 7:56:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.242 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32077.0 tgs: 269 data_time: 0.90s time: 119.19s eta: 7:54:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31641.0 tgs: 266 data_time: 0.81s time: 118.94s eta: 7:51:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 264 data_time: 0.63s time: 118.97s eta: 7:49:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.83s time: 119.42s eta: 7:49:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.199 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31639.0 tgs: 265 data_time: 0.75s time: 119.03s eta: 7:46:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.124 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32063.0 tgs: 269 data_time: 0.69s time: 119.09s eta: 7:44:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.201 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31561.0 tgs: 262 data_time: 0.93s time: 120.01s eta: 7:46:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31630.0 tgs: 263 data_time: 0.77s time: 119.83s eta: 7:43:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.198 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 270 data_time: 1.11s time: 119.16s eta: 7:38:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.220 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31251.0 tgs: 262 data_time: 0.68s time: 118.97s eta: 7:36:02 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31497.0 tgs: 263 data_time: 0.78s time: 119.32s eta: 7:35:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.224 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 261 data_time: 0.63s time: 119.38s eta: 7:33:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31528.0 tgs: 264 data_time: 0.62s time: 119.07s eta: 7:30:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31551.0 tgs: 265 data_time: 0.65s time: 118.99s eta: 7:28:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31859.0 tgs: 267 data_time: 0.70s time: 119.24s eta: 7:27:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.217 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31645.0 tgs: 265 data_time: 0.69s time: 118.99s eta: 7:24:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31352.0 tgs: 263 data_time: 0.76s time: 118.93s eta: 7:22:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30731.0 tgs: 258 data_time: 0.79s time: 119.08s eta: 7:20:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30757.0 tgs: 258 data_time: 0.92s time: 118.99s eta: 7:18:16 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30758.0 tgs: 258 data_time: 0.61s time: 119.03s eta: 7:16:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.192 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 266 data_time: 0.61s time: 119.26s eta: 7:15:18 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.140 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31693.0 tgs: 265 data_time: 0.64s time: 119.37s eta: 7:13:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 271 data_time: 0.86s time: 118.94s eta: 7:10:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.120 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32484.0 tgs: 272 data_time: 0.54s time: 119.20s eta: 7:09:07 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.226 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31788.0 tgs: 265 data_time: 0.65s time: 119.59s eta: 7:08:31 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.218 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32134.0 tgs: 268 data_time: 0.83s time: 119.66s eta: 7:06:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.124 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30521.0 tgs: 256 data_time: 0.75s time: 118.95s eta: 7:02:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.133 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32126.0 tgs: 269 data_time: 0.90s time: 119.30s eta: 7:01:30 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31708.0 tgs: 265 data_time: 0.67s time: 119.52s eta: 7:00:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31500.0 tgs: 263 data_time: 0.52s time: 119.42s eta: 6:57:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32046.0 tgs: 269 data_time: 0.94s time: 118.91s eta: 6:54:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32042.0 tgs: 268 data_time: 0.75s time: 119.35s eta: 6:53:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32363.0 tgs: 272 data_time: 0.73s time: 118.93s eta: 6:50:18 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.200 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 28048.0 tgs: 235 data_time: 0.48s time: 119.27s eta: 6:49:29 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.126 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31014.0 tgs: 260 data_time: 0.68s time: 119.02s eta: 6:46:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.194 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31741.0 tgs: 266 data_time: 0.78s time: 119.24s eta: 6:45:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32336.0 tgs: 270 data_time: 0.77s time: 119.39s eta: 6:43:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 264 data_time: 0.85s time: 119.01s eta: 6:40:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.96s time: 119.01s eta: 6:38:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.181 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32075.0 tgs: 268 data_time: 0.67s time: 119.45s eta: 6:38:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 263 data_time: 0.83s time: 118.80s eta: 6:34:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 267 data_time: 1.03s time: 119.43s eta: 6:34:07 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.218 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 265 data_time: 0.90s time: 119.57s eta: 6:32:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31040.0 tgs: 260 data_time: 0.79s time: 119.38s eta: 6:29:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.178 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31594.0 tgs: 264 data_time: 0.82s time: 119.60s eta: 6:28:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.174 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31690.0 tgs: 265 data_time: 0.80s time: 119.37s eta: 6:25:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30959.0 tgs: 260 data_time: 0.90s time: 118.81s eta: 6:22:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 260 data_time: 0.80s time: 119.02s eta: 6:20:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.160 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32499.0 tgs: 272 data_time: 0.75s time: 119.39s eta: 6:20:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.207 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31509.0 tgs: 264 data_time: 0.81s time: 119.34s eta: 6:17:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32541.0 tgs: 273 data_time: 0.82s time: 119.05s eta: 6:15:00 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31399.0 tgs: 262 data_time: 0.72s time: 119.41s eta: 6:14:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.181 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32182.0 tgs: 269 data_time: 0.70s time: 119.39s eta: 6:12:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31938.0 tgs: 267 data_time: 1.07s time: 119.28s eta: 6:09:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31152.0 tgs: 261 data_time: 0.68s time: 119.10s eta: 6:07:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27695.0 tgs: 227 data_time: 0.46s time: 121.96s eta: 6:14:00 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31417.0 tgs: 263 data_time: 0.53s time: 119.38s eta: 6:04:07 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31945.0 tgs: 267 data_time: 0.97s time: 119.57s eta: 6:02:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31277.0 tgs: 262 data_time: 0.91s time: 119.26s eta: 5:59:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 270 data_time: 0.95s time: 119.26s eta: 5:57:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.163 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31796.0 tgs: 267 data_time: 0.64s time: 118.92s eta: 5:54:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31283.0 tgs: 262 data_time: 0.78s time: 119.01s eta: 5:53:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31575.0 tgs: 265 data_time: 0.69s time: 119.07s eta: 5:51:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31841.0 tgs: 266 data_time: 0.56s time: 119.29s eta: 5:49:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31425.0 tgs: 264 data_time: 0.76s time: 118.97s eta: 5:47:00 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.242 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30675.0 tgs: 257 data_time: 0.85s time: 119.27s eta: 5:45:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31842.0 tgs: 266 data_time: 0.83s time: 119.35s eta: 5:44:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.90s time: 119.05s eta: 5:41:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31432.0 tgs: 263 data_time: 0.60s time: 119.50s eta: 5:40:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 265 data_time: 0.57s time: 119.38s eta: 5:38:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31867.0 tgs: 266 data_time: 0.85s time: 119.56s eta: 5:36:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32481.0 tgs: 272 data_time: 0.86s time: 118.98s eta: 5:33:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31778.0 tgs: 267 data_time: 0.97s time: 118.95s eta: 5:31:04 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32010.0 tgs: 268 data_time: 0.60s time: 119.24s eta: 5:29:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.186 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30687.0 tgs: 257 data_time: 0.77s time: 119.40s eta: 5:28:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.133 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32287.0 tgs: 270 data_time: 0.75s time: 119.46s eta: 5:26:31 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31042.0 tgs: 260 data_time: 0.64s time: 118.95s eta: 5:23:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.136 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30746.0 tgs: 257 data_time: 0.91s time: 119.63s eta: 5:23:00 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.169 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 266 data_time: 0.69s time: 119.34s eta: 5:20:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.176 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30113.0 tgs: 252 data_time: 1.16s time: 119.12s eta: 5:17:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.117 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 265 data_time: 0.98s time: 119.42s eta: 5:16:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 266 data_time: 0.54s time: 119.59s eta: 5:14:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.64s time: 119.00s eta: 5:11:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31400.0 tgs: 263 data_time: 0.74s time: 119.00s eta: 5:09:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30693.0 tgs: 257 data_time: 0.63s time: 119.22s eta: 5:07:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31795.0 tgs: 265 data_time: 0.76s time: 119.68s eta: 5:07:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32314.0 tgs: 271 data_time: 0.87s time: 119.11s eta: 5:03:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 272 data_time: 0.76s time: 119.07s eta: 5:01:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.155 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 262 data_time: 0.83s time: 121.52s eta: 5:05:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 266 data_time: 0.94s time: 119.57s eta: 4:58:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.219 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32384.0 tgs: 271 data_time: 0.72s time: 119.49s eta: 4:56:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31355.0 tgs: 261 data_time: 0.67s time: 119.73s eta: 4:55:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.221 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31711.0 tgs: 265 data_time: 0.65s time: 119.52s eta: 4:52:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.175 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31718.0 tgs: 265 data_time: 0.77s time: 119.41s eta: 4:50:34 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.176 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31466.0 tgs: 263 data_time: 0.76s time: 119.20s eta: 4:48:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.115 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30970.0 tgs: 260 data_time: 0.96s time: 118.89s eta: 4:45:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32051.0 tgs: 269 data_time: 0.65s time: 119.08s eta: 4:43:48 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31215.0 tgs: 260 data_time: 0.82s time: 119.98s eta: 4:43:57 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31102.0 tgs: 261 data_time: 0.89s time: 119.14s eta: 4:39:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.321 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 265 data_time: 0.88s time: 119.01s eta: 4:37:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.149 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.79s time: 119.06s eta: 4:35:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31468.0 tgs: 263 data_time: 0.78s time: 119.37s eta: 4:34:32 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.114 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31405.0 tgs: 263 data_time: 0.70s time: 119.41s eta: 4:32:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31506.0 tgs: 264 data_time: 0.93s time: 118.94s eta: 4:29:35 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31662.0 tgs: 264 data_time: 0.96s time: 119.60s eta: 4:29:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32441.0 tgs: 272 data_time: 0.67s time: 119.24s eta: 4:26:18 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32273.0 tgs: 271 data_time: 0.62s time: 119.07s eta: 4:23:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.92s time: 118.85s eta: 4:21:27 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 270 data_time: 0.78s time: 118.75s eta: 4:19:16 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.186 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30724.0 tgs: 257 data_time: 0.65s time: 119.43s eta: 4:18:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.132 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31884.0 tgs: 267 data_time: 0.76s time: 119.40s eta: 4:16:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 265 data_time: 0.81s time: 119.35s eta: 4:14:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30653.0 tgs: 257 data_time: 0.59s time: 119.05s eta: 4:11:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31569.0 tgs: 265 data_time: 0.68s time: 119.12s eta: 4:10:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.62s time: 119.13s eta: 4:08:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31974.0 tgs: 268 data_time: 0.62s time: 118.97s eta: 4:05:51 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30777.0 tgs: 258 data_time: 0.86s time: 119.14s eta: 4:04:14 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28548.0 tgs: 238 data_time: 0.58s time: 119.50s eta: 4:02:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31850.0 tgs: 267 data_time: 0.77s time: 119.25s eta: 4:00:28 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 264 data_time: 0.60s time: 119.20s eta: 3:58:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31185.0 tgs: 261 data_time: 0.73s time: 119.36s eta: 3:56:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 269 data_time: 0.65s time: 119.37s eta: 3:54:45 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32087.0 tgs: 267 data_time: 0.55s time: 119.95s eta: 3:53:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.88s time: 119.32s eta: 3:50:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 270 data_time: 0.80s time: 119.63s eta: 3:49:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.199 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 267 data_time: 0.69s time: 118.86s eta: 3:45:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.132 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31697.0 tgs: 264 data_time: 0.87s time: 119.73s eta: 3:45:29 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32151.0 tgs: 269 data_time: 0.84s time: 119.32s eta: 3:42:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31616.0 tgs: 265 data_time: 0.90s time: 119.27s eta: 3:40:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 268 data_time: 0.72s time: 119.00s eta: 3:38:09 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31652.0 tgs: 265 data_time: 0.71s time: 119.42s eta: 3:36:57 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.82s time: 119.03s eta: 3:34:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.74s time: 118.99s eta: 3:32:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30971.0 tgs: 259 data_time: 0.68s time: 119.22s eta: 3:30:36 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31857.0 tgs: 265 data_time: 1.04s time: 120.11s eta: 3:30:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31749.0 tgs: 266 data_time: 0.64s time: 119.13s eta: 3:26:29 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32341.0 tgs: 271 data_time: 0.79s time: 119.31s eta: 3:24:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.181 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31570.0 tgs: 265 data_time: 0.69s time: 119.08s eta: 3:22:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32029.0 tgs: 268 data_time: 0.66s time: 119.43s eta: 3:21:02 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.186 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31717.0 tgs: 265 data_time: 0.50s time: 119.29s eta: 3:18:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.208 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32113.0 tgs: 269 data_time: 1.04s time: 119.22s eta: 3:16:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31484.0 tgs: 264 data_time: 0.62s time: 119.19s eta: 3:14:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32141.0 tgs: 270 data_time: 0.71s time: 119.01s eta: 3:12:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31449.0 tgs: 263 data_time: 0.62s time: 119.18s eta: 3:10:41 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.110 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32068.0 tgs: 264 data_time: 0.78s time: 121.05s eta: 3:11:39 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31243.0 tgs: 261 data_time: 0.68s time: 119.51s eta: 3:07:13 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31898.0 tgs: 267 data_time: 0.66s time: 119.25s eta: 3:04:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31688.0 tgs: 266 data_time: 0.70s time: 118.99s eta: 3:02:26 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 264 data_time: 0.82s time: 119.22s eta: 3:00:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31375.0 tgs: 262 data_time: 0.91s time: 119.45s eta: 2:59:10 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30666.0 tgs: 257 data_time: 0.84s time: 118.98s eta: 2:56:29 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30966.0 tgs: 259 data_time: 0.63s time: 119.35s eta: 2:55:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30984.0 tgs: 259 data_time: 0.79s time: 119.41s eta: 2:53:08 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 260 data_time: 1.00s time: 119.35s eta: 2:51:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31662.0 tgs: 260 data_time: 0.80s time: 121.46s eta: 2:52:03 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31340.0 tgs: 262 data_time: 0.93s time: 119.24s eta: 2:46:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.122 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31549.0 tgs: 263 data_time: 0.62s time: 119.57s eta: 2:45:23 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.197 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31700.0 tgs: 266 data_time: 0.94s time: 118.91s eta: 2:42:30 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31736.0 tgs: 266 data_time: 0.77s time: 119.18s eta: 2:40:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32339.0 tgs: 271 data_time: 0.74s time: 118.96s eta: 2:38:37 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31674.0 tgs: 265 data_time: 0.95s time: 119.23s eta: 2:36:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.190 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 267 data_time: 0.64s time: 119.11s eta: 2:34:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.161 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29915.0 tgs: 250 data_time: 0.55s time: 119.43s eta: 2:33:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31868.0 tgs: 265 data_time: 0.91s time: 120.23s eta: 2:32:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31974.0 tgs: 268 data_time: 0.88s time: 119.29s eta: 2:29:06 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31924.0 tgs: 268 data_time: 0.93s time: 118.87s eta: 2:26:36 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.205 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31761.0 tgs: 265 data_time: 1.05s time: 119.52s eta: 2:25:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 271 data_time: 0.76s time: 119.07s eta: 2:22:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.197 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30579.0 tgs: 256 data_time: 0.82s time: 119.14s eta: 2:20:58 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31507.0 tgs: 264 data_time: 0.60s time: 119.05s eta: 2:18:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31704.0 tgs: 264 data_time: 0.87s time: 119.84s eta: 2:17:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32298.0 tgs: 270 data_time: 0.76s time: 119.35s eta: 2:15:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.203 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30607.0 tgs: 256 data_time: 0.87s time: 119.39s eta: 2:13:18 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30088.0 tgs: 251 data_time: 0.84s time: 119.67s eta: 2:11:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31305.0 tgs: 261 data_time: 1.00s time: 119.49s eta: 2:09:26 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31969.0 tgs: 268 data_time: 0.70s time: 118.96s eta: 2:06:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 267 data_time: 0.70s time: 119.37s eta: 2:05:20 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30637.0 tgs: 256 data_time: 0.65s time: 119.28s eta: 2:03:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31282.0 tgs: 262 data_time: 0.84s time: 119.22s eta: 2:01:12 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31149.0 tgs: 260 data_time: 0.79s time: 119.50s eta: 1:59:30 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30245.0 tgs: 253 data_time: 0.69s time: 119.42s eta: 1:57:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31829.0 tgs: 266 data_time: 0.68s time: 119.54s eta: 1:55:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.198 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 268 data_time: 0.72s time: 119.30s eta: 1:53:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.81s time: 119.25s eta: 1:51:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31406.0 tgs: 264 data_time: 0.79s time: 118.93s eta: 1:49:01 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32043.0 tgs: 267 data_time: 0.53s time: 119.64s eta: 1:47:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 266 data_time: 0.97s time: 119.71s eta: 1:45:44 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.179 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31297.0 tgs: 262 data_time: 0.73s time: 119.34s eta: 1:43:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31855.0 tgs: 266 data_time: 0.61s time: 119.31s eta: 1:41:24 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32159.0 tgs: 269 data_time: 0.67s time: 119.30s eta: 1:39:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31392.0 tgs: 264 data_time: 0.71s time: 118.88s eta: 1:37:05 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29984.0 tgs: 251 data_time: 0.65s time: 119.16s eta: 1:35:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32259.0 tgs: 270 data_time: 0.78s time: 119.06s eta: 1:33:15 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 264 data_time: 0.87s time: 119.25s eta: 1:31:25 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.227 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31849.0 tgs: 266 data_time: 0.75s time: 119.73s eta: 1:29:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30912.0 tgs: 259 data_time: 0.71s time: 119.01s eta: 1:27:16 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31331.0 tgs: 263 data_time: 0.55s time: 118.86s eta: 1:25:11 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31823.0 tgs: 266 data_time: 0.59s time: 119.48s eta: 1:23:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.80s time: 119.48s eta: 1:21:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 269 data_time: 0.88s time: 118.94s eta: 1:19:17 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32267.0 tgs: 271 data_time: 0.95s time: 118.93s eta: 1:17:18 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32411.0 tgs: 265 data_time: 0.73s time: 122.16s eta: 1:17:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30612.0 tgs: 256 data_time: 0.67s time: 119.36s eta: 1:13:36 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30724.0 tgs: 258 data_time: 0.70s time: 119.06s eta: 1:11:26 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31463.0 tgs: 264 data_time: 0.62s time: 118.91s eta: 1:09:21 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32107.0 tgs: 269 data_time: 1.04s time: 119.11s eta: 1:07:29 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31192.0 tgs: 261 data_time: 0.67s time: 119.41s eta: 1:05:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.85s time: 119.34s eta: 1:03:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.222 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32548.0 tgs: 273 data_time: 0.88s time: 119.18s eta: 1:01:34 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32385.0 tgs: 271 data_time: 0.81s time: 119.27s eta: 0:59:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31658.0 tgs: 265 data_time: 0.82s time: 119.44s eta: 0:57:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31779.0 tgs: 266 data_time: 0.71s time: 119.39s eta: 0:55:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.108 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32641.0 tgs: 273 data_time: 0.63s time: 119.48s eta: 0:53:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30724.0 tgs: 257 data_time: 0.71s time: 119.36s eta: 0:51:43 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31732.0 tgs: 265 data_time: 0.80s time: 119.47s eta: 0:49:46 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31603.0 tgs: 265 data_time: 0.71s time: 118.89s eta: 0:47:33 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.183 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31132.0 tgs: 261 data_time: 0.68s time: 119.24s eta: 0:45:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31350.0 tgs: 262 data_time: 0.80s time: 119.21s eta: 0:43:42 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32332.0 tgs: 271 data_time: 0.66s time: 118.99s eta: 0:41:38 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31688.0 tgs: 266 data_time: 0.73s time: 119.04s eta: 0:39:40 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30402.0 tgs: 254 data_time: 0.68s time: 119.66s eta: 0:37:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31274.0 tgs: 262 data_time: 0.50s time: 119.29s eta: 0:35:47 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32075.0 tgs: 267 data_time: 0.64s time: 119.74s eta: 0:33:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32578.0 tgs: 268 data_time: 0.60s time: 121.23s eta: 0:32:19 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.173 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 266 data_time: 0.64s time: 119.38s eta: 0:29:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31789.0 tgs: 266 data_time: 0.77s time: 119.33s eta: 0:27:50 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 266 data_time: 0.70s time: 119.63s eta: 0:25:55 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.121 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31962.0 tgs: 268 data_time: 1.11s time: 119.10s eta: 0:23:49 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31704.0 tgs: 265 data_time: 0.98s time: 119.38s eta: 0:21:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 270 data_time: 0.99s time: 119.26s eta: 0:19:52 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31685.0 tgs: 265 data_time: 0.56s time: 119.24s eta: 0:17:53 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32089.0 tgs: 267 data_time: 0.66s time: 119.91s eta: 0:15:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30888.0 tgs: 259 data_time: 0.67s time: 119.21s eta: 0:13:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 269 data_time: 0.97s time: 119.09s eta: 0:11:54 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.195 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.72s time: 119.33s eta: 0:09:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32520.0 tgs: 272 data_time: 0.70s time: 119.16s eta: 0:07:56 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 270 data_time: 0.71s time: 119.69s eta: 0:05:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31730.0 tgs: 261 data_time: 0.76s time: 121.33s eta: 0:04:02 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31112.0 tgs: 260 data_time: 0.64s time: 119.21s eta: 0:01:59 +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 61][DP 15][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank62.log b/20250121165312/rank62.log new file mode 100644 index 0000000000000000000000000000000000000000..a8cb9dd155697c32b28057764c2924c4e362d983 --- /dev/null +++ b/20250121165312/rank62.log @@ -0,0 +1,851 @@ +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:19][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:31][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.04s +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:42][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:00:43][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 71.47 seconds, peak gpu memory 13.4G +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.276 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.9GB text_tokens: 31861.0 tgs: 208 data_time: 1.80s time: 152.98s eta: 1 day, 1:11:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.221 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31345.0 tgs: 261 data_time: 0.63s time: 120.09s eta: 19:44:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.240 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30897.0 tgs: 258 data_time: 0.80s time: 119.68s eta: 19:38:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.237 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30727.0 tgs: 256 data_time: 0.82s time: 119.65s eta: 19:36:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.188 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31993.0 tgs: 268 data_time: 0.64s time: 119.31s eta: 19:31:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.193 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32528.0 tgs: 272 data_time: 0.71s time: 119.53s eta: 19:31:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.190 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31985.0 tgs: 260 data_time: 0.59s time: 122.97s eta: 20:03:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.304 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32411.0 tgs: 270 data_time: 0.77s time: 119.88s eta: 19:30:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.189 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 255 data_time: 0.68s time: 120.65s eta: 19:36:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.234 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32133.0 tgs: 267 data_time: 1.05s time: 119.92s eta: 19:27:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.207 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31934.0 tgs: 265 data_time: 0.78s time: 120.46s eta: 19:30:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31332.0 tgs: 262 data_time: 0.83s time: 119.34s eta: 19:17:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.229 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32182.0 tgs: 268 data_time: 0.83s time: 119.68s eta: 19:18:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30812.0 tgs: 258 data_time: 0.56s time: 119.20s eta: 19:12:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32580.0 tgs: 273 data_time: 0.66s time: 119.22s eta: 19:10:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30706.0 tgs: 258 data_time: 0.79s time: 118.93s eta: 19:05:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31982.0 tgs: 268 data_time: 0.78s time: 119.15s eta: 19:05:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32212.0 tgs: 270 data_time: 0.84s time: 119.18s eta: 19:04:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31965.0 tgs: 267 data_time: 0.90s time: 119.39s eta: 19:04:10 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31887.0 tgs: 267 data_time: 0.81s time: 119.33s eta: 19:01:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.235 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31686.0 tgs: 263 data_time: 0.96s time: 120.31s eta: 19:08:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31729.0 tgs: 258 data_time: 0.70s time: 122.91s eta: 19:31:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31873.0 tgs: 266 data_time: 0.75s time: 119.48s eta: 18:57:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31838.0 tgs: 267 data_time: 0.75s time: 119.15s eta: 18:51:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 267 data_time: 0.85s time: 119.24s eta: 18:50:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31822.0 tgs: 266 data_time: 0.70s time: 119.39s eta: 18:50:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.231 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32115.0 tgs: 268 data_time: 1.11s time: 119.66s eta: 18:50:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.253 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32433.0 tgs: 271 data_time: 0.82s time: 119.38s eta: 18:46:10 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.108 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31453.0 tgs: 262 data_time: 0.88s time: 119.79s eta: 18:48:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32050.0 tgs: 247 data_time: 0.92s time: 129.30s eta: 20:15:27 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.289 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32259.0 tgs: 269 data_time: 0.66s time: 119.56s eta: 18:41:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32000.0 tgs: 268 data_time: 0.72s time: 119.39s eta: 18:38:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30717.0 tgs: 254 data_time: 0.74s time: 120.66s eta: 18:48:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31683.0 tgs: 265 data_time: 0.83s time: 119.39s eta: 18:34:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.200 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32425.0 tgs: 272 data_time: 1.06s time: 119.16s eta: 18:30:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31127.0 tgs: 261 data_time: 0.62s time: 119.07s eta: 18:27:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31857.0 tgs: 267 data_time: 0.76s time: 119.13s eta: 18:25:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31766.0 tgs: 263 data_time: 0.75s time: 120.39s eta: 18:35:36 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32158.0 tgs: 265 data_time: 0.84s time: 121.17s eta: 18:40:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32032.0 tgs: 265 data_time: 0.81s time: 120.66s eta: 18:34:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31101.0 tgs: 259 data_time: 0.71s time: 119.73s eta: 18:23:31 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30739.0 tgs: 248 data_time: 0.88s time: 123.88s eta: 18:59:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.255 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32313.0 tgs: 270 data_time: 0.80s time: 119.56s eta: 18:17:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30208.0 tgs: 253 data_time: 0.65s time: 119.35s eta: 18:14:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 264 data_time: 0.86s time: 119.22s eta: 18:10:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31887.0 tgs: 266 data_time: 0.87s time: 119.60s eta: 18:12:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31263.0 tgs: 262 data_time: 0.68s time: 119.26s eta: 18:07:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31940.0 tgs: 268 data_time: 0.86s time: 119.01s eta: 18:02:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.136 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32230.0 tgs: 270 data_time: 1.19s time: 119.10s eta: 18:01:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31041.0 tgs: 260 data_time: 0.98s time: 119.26s eta: 18:01:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31939.0 tgs: 267 data_time: 0.66s time: 119.33s eta: 17:59:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.206 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31788.0 tgs: 264 data_time: 1.13s time: 120.02s eta: 18:04:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.149 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32224.0 tgs: 270 data_time: 0.67s time: 119.11s eta: 17:53:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32035.0 tgs: 268 data_time: 0.73s time: 119.40s eta: 17:54:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.154 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30404.0 tgs: 255 data_time: 0.72s time: 118.99s eta: 17:48:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.241 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32351.0 tgs: 271 data_time: 1.03s time: 119.02s eta: 17:47:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32058.0 tgs: 255 data_time: 0.65s time: 125.36s eta: 18:41:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 267 data_time: 0.89s time: 120.10s eta: 17:52:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31472.0 tgs: 263 data_time: 0.66s time: 119.39s eta: 17:44:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31503.0 tgs: 263 data_time: 0.81s time: 119.34s eta: 17:42:06 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32460.0 tgs: 271 data_time: 0.67s time: 119.51s eta: 17:41:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 251 data_time: 0.41s time: 126.05s eta: 18:37:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 260 data_time: 0.67s time: 122.64s eta: 18:05:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31721.0 tgs: 261 data_time: 0.74s time: 121.15s eta: 17:50:08 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32329.0 tgs: 271 data_time: 0.67s time: 119.03s eta: 17:29:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31333.0 tgs: 263 data_time: 0.77s time: 118.85s eta: 17:25:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.130 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 27236.0 tgs: 219 data_time: 0.59s time: 124.00s eta: 18:09:06 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.226 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 258 data_time: 0.82s time: 124.57s eta: 18:12:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.256 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32389.0 tgs: 271 data_time: 0.95s time: 119.39s eta: 17:24:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31521.0 tgs: 263 data_time: 0.90s time: 119.70s eta: 17:25:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31154.0 tgs: 260 data_time: 0.62s time: 119.52s eta: 17:21:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31399.0 tgs: 264 data_time: 0.74s time: 118.86s eta: 17:14:03 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31343.0 tgs: 261 data_time: 0.83s time: 119.85s eta: 17:20:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.201 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31053.0 tgs: 259 data_time: 0.75s time: 119.56s eta: 17:16:10 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32475.0 tgs: 266 data_time: 0.87s time: 121.86s eta: 17:34:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.113 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31529.0 tgs: 255 data_time: 0.73s time: 123.52s eta: 17:46:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.301 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31108.0 tgs: 261 data_time: 0.64s time: 119.01s eta: 17:05:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.196 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32276.0 tgs: 270 data_time: 0.81s time: 119.22s eta: 17:05:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31566.0 tgs: 264 data_time: 0.90s time: 119.14s eta: 17:02:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31630.0 tgs: 265 data_time: 0.69s time: 119.25s eta: 17:01:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32522.0 tgs: 273 data_time: 0.93s time: 119.04s eta: 16:57:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30869.0 tgs: 258 data_time: 0.73s time: 119.28s eta: 16:57:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31254.0 tgs: 262 data_time: 0.77s time: 119.15s eta: 16:54:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 28967.0 tgs: 242 data_time: 0.85s time: 119.59s eta: 16:56:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31776.0 tgs: 266 data_time: 0.70s time: 119.12s eta: 16:50:34 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31381.0 tgs: 264 data_time: 0.51s time: 118.84s eta: 16:46:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31802.0 tgs: 267 data_time: 0.67s time: 119.01s eta: 16:45:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31937.0 tgs: 266 data_time: 0.78s time: 119.85s eta: 16:50:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31499.0 tgs: 263 data_time: 0.89s time: 119.37s eta: 16:44:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31043.0 tgs: 260 data_time: 0.64s time: 119.01s eta: 16:39:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.134 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30452.0 tgs: 251 data_time: 1.00s time: 120.93s eta: 16:53:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30269.0 tgs: 254 data_time: 0.60s time: 118.99s eta: 16:35:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 266 data_time: 0.86s time: 119.28s eta: 16:36:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31521.0 tgs: 264 data_time: 1.12s time: 119.23s eta: 16:33:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31815.0 tgs: 266 data_time: 0.98s time: 119.37s eta: 16:32:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.133 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30661.0 tgs: 256 data_time: 0.72s time: 119.71s eta: 16:33:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31965.0 tgs: 268 data_time: 0.89s time: 119.20s eta: 16:27:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31853.0 tgs: 265 data_time: 0.63s time: 119.76s eta: 16:30:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31960.0 tgs: 265 data_time: 0.67s time: 120.35s eta: 16:32:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.164 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 266 data_time: 0.79s time: 118.97s eta: 16:19:32 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31596.0 tgs: 248 data_time: 0.65s time: 127.32s eta: 17:26:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31701.0 tgs: 265 data_time: 0.73s time: 119.22s eta: 16:17:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31940.0 tgs: 268 data_time: 0.60s time: 119.14s eta: 16:14:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31867.0 tgs: 267 data_time: 1.04s time: 119.25s eta: 16:13:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.211 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 264 data_time: 0.75s time: 119.17s eta: 16:11:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30332.0 tgs: 255 data_time: 0.90s time: 118.95s eta: 16:07:27 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.163 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31154.0 tgs: 262 data_time: 0.61s time: 118.86s eta: 16:04:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31244.0 tgs: 262 data_time: 0.55s time: 118.96s eta: 16:03:36 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.249 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31671.0 tgs: 259 data_time: 0.90s time: 121.84s eta: 16:24:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.141 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32222.0 tgs: 270 data_time: 0.81s time: 119.00s eta: 15:59:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.139 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32056.0 tgs: 268 data_time: 0.90s time: 119.22s eta: 15:59:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32297.0 tgs: 270 data_time: 0.78s time: 119.61s eta: 16:00:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31976.0 tgs: 268 data_time: 0.86s time: 119.04s eta: 15:54:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32206.0 tgs: 270 data_time: 0.94s time: 118.97s eta: 15:51:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 267 data_time: 0.65s time: 119.18s eta: 15:51:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32352.0 tgs: 270 data_time: 0.56s time: 119.63s eta: 15:53:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31195.0 tgs: 262 data_time: 0.68s time: 118.93s eta: 15:45:31 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32381.0 tgs: 271 data_time: 0.59s time: 119.28s eta: 15:46:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 269 data_time: 1.16s time: 119.13s eta: 15:43:06 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 267 data_time: 0.71s time: 119.00s eta: 15:40:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.187 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31534.0 tgs: 263 data_time: 0.83s time: 119.66s eta: 15:43:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30832.0 tgs: 253 data_time: 0.65s time: 121.51s eta: 15:55:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.115 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32197.0 tgs: 269 data_time: 0.61s time: 119.41s eta: 15:37:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.169 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31745.0 tgs: 244 data_time: 0.79s time: 129.96s eta: 16:58:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31717.0 tgs: 265 data_time: 1.10s time: 119.47s eta: 15:33:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31536.0 tgs: 263 data_time: 0.58s time: 119.60s eta: 15:32:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31300.0 tgs: 261 data_time: 0.89s time: 119.53s eta: 15:30:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.221 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31783.0 tgs: 260 data_time: 0.74s time: 122.14s eta: 15:48:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.122 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32422.0 tgs: 268 data_time: 0.85s time: 120.75s eta: 15:35:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30462.0 tgs: 255 data_time: 0.85s time: 119.41s eta: 15:23:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.241 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 269 data_time: 0.84s time: 119.45s eta: 15:21:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.155 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30621.0 tgs: 232 data_time: 0.93s time: 131.63s eta: 16:53:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31534.0 tgs: 264 data_time: 0.90s time: 119.29s eta: 15:16:32 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30926.0 tgs: 258 data_time: 0.59s time: 119.56s eta: 15:16:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 262 data_time: 0.71s time: 122.25s eta: 15:35:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.129 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30268.0 tgs: 252 data_time: 0.69s time: 119.71s eta: 15:13:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 264 data_time: 0.74s time: 119.35s eta: 15:09:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.228 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30984.0 tgs: 260 data_time: 0.89s time: 118.90s eta: 15:03:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32362.0 tgs: 271 data_time: 1.03s time: 118.99s eta: 15:02:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.152 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31140.0 tgs: 260 data_time: 0.72s time: 119.31s eta: 15:02:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.116 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31989.0 tgs: 269 data_time: 0.72s time: 118.91s eta: 14:57:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.202 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32328.0 tgs: 268 data_time: 0.56s time: 120.52s eta: 15:07:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.180 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31613.0 tgs: 264 data_time: 0.67s time: 119.31s eta: 14:56:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31698.0 tgs: 258 data_time: 0.67s time: 122.65s eta: 15:19:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31581.0 tgs: 265 data_time: 1.03s time: 119.12s eta: 14:51:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31926.0 tgs: 266 data_time: 0.69s time: 119.75s eta: 14:54:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31461.0 tgs: 263 data_time: 0.64s time: 119.47s eta: 14:50:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.318 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30414.0 tgs: 255 data_time: 0.68s time: 118.98s eta: 14:44:27 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32174.0 tgs: 269 data_time: 0.95s time: 119.33s eta: 14:45:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.121 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30963.0 tgs: 260 data_time: 0.56s time: 118.94s eta: 14:40:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 271 data_time: 0.73s time: 119.05s eta: 14:39:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31750.0 tgs: 266 data_time: 0.81s time: 119.06s eta: 14:37:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.186 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32234.0 tgs: 270 data_time: 0.83s time: 119.12s eta: 14:35:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.121 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31496.0 tgs: 264 data_time: 0.71s time: 119.00s eta: 14:32:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.122 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30789.0 tgs: 258 data_time: 0.83s time: 119.07s eta: 14:31:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.165 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31168.0 tgs: 261 data_time: 0.82s time: 119.15s eta: 14:29:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31841.0 tgs: 266 data_time: 0.91s time: 119.65s eta: 14:31:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.159 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31532.0 tgs: 264 data_time: 0.70s time: 119.18s eta: 14:26:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.231 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32474.0 tgs: 272 data_time: 0.89s time: 119.20s eta: 14:24:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.199 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 265 data_time: 0.63s time: 119.39s eta: 14:23:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.214 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31173.0 tgs: 258 data_time: 1.13s time: 120.73s eta: 14:31:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.207 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32536.0 tgs: 273 data_time: 0.88s time: 118.86s eta: 14:15:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 269 data_time: 0.76s time: 119.10s eta: 14:15:32 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32038.0 tgs: 268 data_time: 0.67s time: 119.15s eta: 14:13:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.139 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 266 data_time: 0.77s time: 119.96s eta: 14:17:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31272.0 tgs: 260 data_time: 0.89s time: 119.87s eta: 14:15:03 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32263.0 tgs: 269 data_time: 0.65s time: 119.59s eta: 14:11:03 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31832.0 tgs: 267 data_time: 0.77s time: 119.22s eta: 14:06:27 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31974.0 tgs: 268 data_time: 0.83s time: 119.26s eta: 14:04:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.143 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31337.0 tgs: 262 data_time: 0.77s time: 119.27s eta: 14:02:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.171 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32038.0 tgs: 268 data_time: 0.90s time: 119.14s eta: 13:59:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.228 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30681.0 tgs: 257 data_time: 0.81s time: 119.12s eta: 13:57:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.134 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31452.0 tgs: 263 data_time: 0.67s time: 119.17s eta: 13:56:10 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.208 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31620.0 tgs: 264 data_time: 1.08s time: 119.35s eta: 13:55:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.252 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31468.0 tgs: 261 data_time: 0.72s time: 120.11s eta: 13:58:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31883.0 tgs: 260 data_time: 0.48s time: 122.55s eta: 14:13:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31475.0 tgs: 264 data_time: 0.72s time: 119.11s eta: 13:47:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32081.0 tgs: 268 data_time: 0.38s time: 119.37s eta: 13:47:36 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.217 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32030.0 tgs: 268 data_time: 1.08s time: 119.35s eta: 13:45:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.234 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32278.0 tgs: 270 data_time: 0.92s time: 119.33s eta: 13:43:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.135 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30973.0 tgs: 260 data_time: 0.69s time: 119.11s eta: 13:39:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31679.0 tgs: 265 data_time: 0.78s time: 119.26s eta: 13:38:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.207 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32123.0 tgs: 269 data_time: 0.84s time: 119.13s eta: 13:36:03 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.169 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31454.0 tgs: 263 data_time: 0.69s time: 119.22s eta: 13:34:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.73s time: 119.46s eta: 13:34:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 267 data_time: 0.97s time: 119.08s eta: 13:29:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31043.0 tgs: 261 data_time: 0.84s time: 118.92s eta: 13:26:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31949.0 tgs: 266 data_time: 0.69s time: 119.81s eta: 13:30:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31673.0 tgs: 265 data_time: 0.86s time: 119.25s eta: 13:24:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30259.0 tgs: 253 data_time: 0.77s time: 119.16s eta: 13:22:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31588.0 tgs: 264 data_time: 0.78s time: 119.34s eta: 13:21:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 271 data_time: 0.81s time: 119.26s eta: 13:19:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.126 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31297.0 tgs: 262 data_time: 0.73s time: 119.04s eta: 13:15:34 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.199 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31956.0 tgs: 267 data_time: 0.73s time: 119.37s eta: 13:15:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32084.0 tgs: 269 data_time: 0.71s time: 119.19s eta: 13:12:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.243 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30991.0 tgs: 259 data_time: 1.23s time: 119.45s eta: 13:12:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.140 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31552.0 tgs: 264 data_time: 0.73s time: 119.29s eta: 13:09:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.214 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31575.0 tgs: 265 data_time: 0.83s time: 119.14s eta: 13:06:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.125 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31903.0 tgs: 264 data_time: 0.63s time: 120.48s eta: 13:13:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 265 data_time: 0.86s time: 120.15s eta: 13:09:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.216 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31793.0 tgs: 266 data_time: 0.92s time: 119.50s eta: 13:02:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 267 data_time: 0.68s time: 119.51s eta: 13:00:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.119 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 264 data_time: 0.70s time: 119.57s eta: 12:59:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30816.0 tgs: 257 data_time: 0.65s time: 119.79s eta: 12:58:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.113 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31551.0 tgs: 265 data_time: 0.70s time: 119.04s eta: 12:51:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.179 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32111.0 tgs: 269 data_time: 0.83s time: 119.05s eta: 12:49:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31521.0 tgs: 264 data_time: 0.77s time: 119.12s eta: 12:48:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.219 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31568.0 tgs: 264 data_time: 0.72s time: 119.39s eta: 12:48:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31459.0 tgs: 263 data_time: 0.75s time: 119.18s eta: 12:44:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 260 data_time: 0.78s time: 119.43s eta: 12:44:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 265 data_time: 0.84s time: 119.10s eta: 12:40:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.218 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31535.0 tgs: 264 data_time: 0.70s time: 119.07s eta: 12:38:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.60s time: 119.60s eta: 12:39:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30891.0 tgs: 260 data_time: 0.83s time: 118.78s eta: 12:32:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32258.0 tgs: 271 data_time: 0.88s time: 118.88s eta: 12:30:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31925.0 tgs: 267 data_time: 0.78s time: 119.17s eta: 12:30:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31125.0 tgs: 260 data_time: 0.90s time: 119.27s eta: 12:29:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30968.0 tgs: 259 data_time: 1.00s time: 119.35s eta: 12:27:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31570.0 tgs: 265 data_time: 0.64s time: 119.12s eta: 12:24:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.115 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31066.0 tgs: 260 data_time: 0.82s time: 119.34s eta: 12:23:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.179 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32253.0 tgs: 270 data_time: 0.69s time: 119.26s eta: 12:21:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31752.0 tgs: 266 data_time: 0.77s time: 119.03s eta: 12:17:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 32034.0 tgs: 264 data_time: 0.94s time: 121.13s eta: 12:28:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31566.0 tgs: 263 data_time: 0.77s time: 119.99s eta: 12:19:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31032.0 tgs: 261 data_time: 0.77s time: 118.72s eta: 12:10:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.174 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30435.0 tgs: 254 data_time: 1.06s time: 119.63s eta: 12:13:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30462.0 tgs: 254 data_time: 0.81s time: 119.83s eta: 12:12:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 0.64s time: 119.14s eta: 12:06:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.126 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 266 data_time: 0.97s time: 119.17s eta: 12:04:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.142 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30964.0 tgs: 259 data_time: 0.85s time: 119.12s eta: 12:02:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32255.0 tgs: 270 data_time: 0.71s time: 119.19s eta: 12:01:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.208 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31354.0 tgs: 263 data_time: 0.78s time: 118.95s eta: 11:57:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.196 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31553.0 tgs: 263 data_time: 0.85s time: 119.70s eta: 12:00:13 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31779.0 tgs: 266 data_time: 0.89s time: 119.28s eta: 11:55:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.153 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31687.0 tgs: 265 data_time: 0.68s time: 119.27s eta: 11:53:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32032.0 tgs: 267 data_time: 0.88s time: 119.71s eta: 11:54:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.192 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31660.0 tgs: 266 data_time: 1.04s time: 118.78s eta: 11:46:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.174 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30889.0 tgs: 259 data_time: 0.85s time: 119.16s eta: 11:46:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.118 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31706.0 tgs: 265 data_time: 0.65s time: 119.61s eta: 11:47:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32249.0 tgs: 271 data_time: 0.89s time: 118.85s eta: 11:41:13 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 270 data_time: 0.74s time: 118.68s eta: 11:38:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32441.0 tgs: 272 data_time: 0.56s time: 119.21s eta: 11:39:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.70s time: 119.14s eta: 11:36:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32027.0 tgs: 268 data_time: 0.75s time: 119.07s eta: 11:34:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.182 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30730.0 tgs: 258 data_time: 1.11s time: 118.95s eta: 11:31:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32053.0 tgs: 269 data_time: 0.99s time: 118.91s eta: 11:29:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31359.0 tgs: 263 data_time: 0.60s time: 118.97s eta: 11:28:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.111 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32010.0 tgs: 268 data_time: 0.54s time: 119.29s eta: 11:27:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.126 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31442.0 tgs: 263 data_time: 0.68s time: 119.35s eta: 11:26:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30872.0 tgs: 259 data_time: 0.79s time: 118.88s eta: 11:21:34 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.163 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30977.0 tgs: 259 data_time: 0.99s time: 119.21s eta: 11:21:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32183.0 tgs: 268 data_time: 0.86s time: 119.72s eta: 11:22:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31874.0 tgs: 267 data_time: 0.87s time: 119.11s eta: 11:16:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.129 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31635.0 tgs: 265 data_time: 1.03s time: 119.12s eta: 11:15:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31496.0 tgs: 264 data_time: 0.91s time: 118.98s eta: 11:12:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31879.0 tgs: 267 data_time: 0.64s time: 119.09s eta: 11:10:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.126 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30964.0 tgs: 259 data_time: 0.56s time: 119.16s eta: 11:09:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32332.0 tgs: 271 data_time: 0.75s time: 118.97s eta: 11:06:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.112 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30372.0 tgs: 254 data_time: 1.07s time: 119.46s eta: 11:07:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31452.0 tgs: 262 data_time: 0.88s time: 119.64s eta: 11:05:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.161 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31934.0 tgs: 267 data_time: 0.78s time: 119.31s eta: 11:02:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.210 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32152.0 tgs: 269 data_time: 0.85s time: 119.30s eta: 11:00:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.162 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30536.0 tgs: 256 data_time: 0.78s time: 118.82s eta: 10:55:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.182 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31781.0 tgs: 266 data_time: 0.81s time: 119.47s eta: 10:57:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 260 data_time: 0.67s time: 119.37s eta: 10:54:32 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31518.0 tgs: 264 data_time: 0.71s time: 119.27s eta: 10:52:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32102.0 tgs: 269 data_time: 0.70s time: 119.25s eta: 10:49:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31697.0 tgs: 266 data_time: 0.90s time: 119.10s eta: 10:47:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.258 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32236.0 tgs: 269 data_time: 0.83s time: 119.53s eta: 10:47:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32597.0 tgs: 273 data_time: 0.68s time: 119.33s eta: 10:44:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31624.0 tgs: 264 data_time: 0.77s time: 119.50s eta: 10:43:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.151 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31813.0 tgs: 266 data_time: 0.61s time: 119.51s eta: 10:41:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.222 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32058.0 tgs: 269 data_time: 1.00s time: 118.99s eta: 10:36:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.192 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32356.0 tgs: 271 data_time: 0.87s time: 119.22s eta: 10:35:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31688.0 tgs: 266 data_time: 0.59s time: 119.06s eta: 10:33:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.179 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32462.0 tgs: 273 data_time: 0.83s time: 118.78s eta: 10:29:31 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32143.0 tgs: 268 data_time: 0.76s time: 119.86s eta: 10:33:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.163 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31922.0 tgs: 267 data_time: 0.81s time: 119.38s eta: 10:28:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.181 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32206.0 tgs: 270 data_time: 0.75s time: 118.84s eta: 10:23:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.188 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30668.0 tgs: 257 data_time: 0.87s time: 118.98s eta: 10:22:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.178 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31834.0 tgs: 266 data_time: 0.77s time: 119.28s eta: 10:22:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29624.0 tgs: 248 data_time: 0.69s time: 119.20s eta: 10:19:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30908.0 tgs: 259 data_time: 0.67s time: 119.06s eta: 10:17:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.167 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 264 data_time: 0.58s time: 119.47s eta: 10:17:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.129 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30306.0 tgs: 253 data_time: 0.68s time: 119.39s eta: 10:14:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.172 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31906.0 tgs: 267 data_time: 0.58s time: 119.45s eta: 10:13:10 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30612.0 tgs: 256 data_time: 0.93s time: 119.29s eta: 10:10:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.185 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32081.0 tgs: 269 data_time: 0.66s time: 119.11s eta: 10:07:27 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31641.0 tgs: 264 data_time: 0.52s time: 119.70s eta: 10:08:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30197.0 tgs: 252 data_time: 0.71s time: 119.83s eta: 10:07:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.236 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32281.0 tgs: 265 data_time: 1.03s time: 121.55s eta: 10:13:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.139 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31740.0 tgs: 266 data_time: 0.90s time: 119.25s eta: 10:00:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.122 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31464.0 tgs: 264 data_time: 0.81s time: 118.81s eta: 9:56:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.115 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30866.0 tgs: 258 data_time: 0.79s time: 119.56s eta: 9:57:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.135 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 268 data_time: 0.63s time: 119.70s eta: 9:56:31 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30739.0 tgs: 257 data_time: 0.66s time: 119.26s eta: 9:52:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31790.0 tgs: 266 data_time: 0.76s time: 119.33s eta: 9:50:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30480.0 tgs: 255 data_time: 0.74s time: 119.22s eta: 9:48:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.131 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31890.0 tgs: 263 data_time: 0.84s time: 120.88s eta: 9:54:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.184 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.55s time: 119.28s eta: 9:44:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.183 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31951.0 tgs: 267 data_time: 0.99s time: 119.45s eta: 9:43:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.211 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32135.0 tgs: 268 data_time: 0.85s time: 119.54s eta: 9:41:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31362.0 tgs: 263 data_time: 0.62s time: 119.21s eta: 9:38:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32266.0 tgs: 270 data_time: 0.81s time: 119.48s eta: 9:37:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31107.0 tgs: 261 data_time: 0.56s time: 119.04s eta: 9:33:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.113 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31653.0 tgs: 265 data_time: 0.74s time: 119.24s eta: 9:32:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31613.0 tgs: 265 data_time: 0.68s time: 119.16s eta: 9:29:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.178 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31659.0 tgs: 265 data_time: 0.78s time: 119.22s eta: 9:28:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.182 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31783.0 tgs: 266 data_time: 1.13s time: 119.46s eta: 9:27:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32338.0 tgs: 271 data_time: 0.49s time: 118.98s eta: 9:23:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 268 data_time: 0.70s time: 119.04s eta: 9:21:27 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 29882.0 tgs: 249 data_time: 0.56s time: 119.53s eta: 9:21:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.210 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 266 data_time: 0.62s time: 119.60s eta: 9:20:06 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.217 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31682.0 tgs: 265 data_time: 0.81s time: 119.12s eta: 9:15:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.197 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31960.0 tgs: 268 data_time: 0.69s time: 119.07s eta: 9:13:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 267 data_time: 0.81s time: 118.88s eta: 9:10:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31547.0 tgs: 264 data_time: 0.75s time: 119.32s eta: 9:10:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30915.0 tgs: 259 data_time: 0.77s time: 119.09s eta: 9:07:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32237.0 tgs: 271 data_time: 0.73s time: 118.86s eta: 9:04:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.265 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31751.0 tgs: 265 data_time: 0.83s time: 119.67s eta: 9:06:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31612.0 tgs: 264 data_time: 0.77s time: 119.43s eta: 9:03:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.158 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31026.0 tgs: 259 data_time: 0.95s time: 119.42s eta: 9:01:21 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31302.0 tgs: 263 data_time: 0.76s time: 118.74s eta: 8:56:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.164 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31825.0 tgs: 267 data_time: 0.68s time: 118.87s eta: 8:54:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32045.0 tgs: 268 data_time: 0.87s time: 119.33s eta: 8:55:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.140 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31346.0 tgs: 263 data_time: 0.68s time: 119.15s eta: 8:52:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.231 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31713.0 tgs: 265 data_time: 0.88s time: 119.26s eta: 8:50:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.190 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32134.0 tgs: 266 data_time: 1.17s time: 120.47s eta: 8:54:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.254 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 269 data_time: 0.92s time: 119.24s eta: 8:46:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31145.0 tgs: 261 data_time: 0.56s time: 119.07s eta: 8:43:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 263 data_time: 0.76s time: 120.82s eta: 8:49:36 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30849.0 tgs: 256 data_time: 0.61s time: 120.19s eta: 8:44:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.191 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 271 data_time: 0.60s time: 118.83s eta: 8:36:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 268 data_time: 0.93s time: 119.35s eta: 8:37:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31681.0 tgs: 265 data_time: 0.73s time: 119.33s eta: 8:35:06 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31682.0 tgs: 266 data_time: 0.92s time: 119.05s eta: 8:31:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.308 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31886.0 tgs: 267 data_time: 1.08s time: 119.16s eta: 8:30:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31970.0 tgs: 268 data_time: 0.73s time: 119.04s eta: 8:27:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31706.0 tgs: 266 data_time: 0.60s time: 119.12s eta: 8:26:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31839.0 tgs: 266 data_time: 0.47s time: 119.25s eta: 8:24:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32274.0 tgs: 270 data_time: 0.89s time: 119.22s eta: 8:22:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.185 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31497.0 tgs: 263 data_time: 1.01s time: 119.38s eta: 8:21:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.136 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30551.0 tgs: 255 data_time: 0.77s time: 119.78s eta: 8:21:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.183 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 265 data_time: 0.71s time: 119.20s eta: 8:16:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.127 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32406.0 tgs: 272 data_time: 0.47s time: 119.02s eta: 8:13:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.150 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31083.0 tgs: 259 data_time: 0.78s time: 119.89s eta: 8:15:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31197.0 tgs: 260 data_time: 0.75s time: 119.56s eta: 8:12:10 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.117 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31241.0 tgs: 261 data_time: 0.45s time: 119.59s eta: 8:10:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31534.0 tgs: 264 data_time: 0.96s time: 119.00s eta: 8:05:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32129.0 tgs: 268 data_time: 0.63s time: 119.84s eta: 8:07:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31216.0 tgs: 260 data_time: 0.72s time: 119.71s eta: 8:04:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.175 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32291.0 tgs: 270 data_time: 0.74s time: 119.40s eta: 8:01:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.156 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31509.0 tgs: 264 data_time: 0.56s time: 119.28s eta: 7:59:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.136 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31600.0 tgs: 265 data_time: 0.77s time: 119.19s eta: 7:56:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.219 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32077.0 tgs: 269 data_time: 0.90s time: 119.20s eta: 7:54:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31641.0 tgs: 266 data_time: 0.83s time: 118.94s eta: 7:51:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31504.0 tgs: 264 data_time: 0.61s time: 118.97s eta: 7:49:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.173 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32025.0 tgs: 268 data_time: 0.81s time: 119.41s eta: 7:49:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.137 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31639.0 tgs: 265 data_time: 0.71s time: 119.04s eta: 7:46:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.182 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32063.0 tgs: 269 data_time: 0.68s time: 119.09s eta: 7:44:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.157 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31561.0 tgs: 262 data_time: 0.92s time: 120.00s eta: 7:46:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.187 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31630.0 tgs: 263 data_time: 0.77s time: 119.82s eta: 7:43:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.174 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32277.0 tgs: 270 data_time: 1.11s time: 119.17s eta: 7:38:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.164 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31251.0 tgs: 262 data_time: 0.68s time: 118.97s eta: 7:36:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31497.0 tgs: 263 data_time: 0.78s time: 119.32s eta: 7:35:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31267.0 tgs: 261 data_time: 0.63s time: 119.38s eta: 7:33:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31528.0 tgs: 264 data_time: 0.62s time: 119.07s eta: 7:30:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.167 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31551.0 tgs: 265 data_time: 0.66s time: 118.99s eta: 7:28:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31859.0 tgs: 267 data_time: 0.70s time: 119.23s eta: 7:27:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31645.0 tgs: 265 data_time: 0.69s time: 118.99s eta: 7:24:13 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.126 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31352.0 tgs: 263 data_time: 0.76s time: 118.93s eta: 7:22:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.114 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30731.0 tgs: 258 data_time: 0.80s time: 119.08s eta: 7:20:36 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.148 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30757.0 tgs: 258 data_time: 0.91s time: 118.99s eta: 7:18:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30758.0 tgs: 258 data_time: 0.61s time: 119.03s eta: 7:16:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31781.0 tgs: 266 data_time: 0.59s time: 119.26s eta: 7:15:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.164 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31693.0 tgs: 265 data_time: 0.66s time: 119.37s eta: 7:13:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.167 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 271 data_time: 0.86s time: 118.94s eta: 7:10:10 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32484.0 tgs: 272 data_time: 0.55s time: 119.20s eta: 7:09:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31788.0 tgs: 265 data_time: 0.64s time: 119.58s eta: 7:08:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32134.0 tgs: 268 data_time: 0.83s time: 119.66s eta: 7:06:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30521.0 tgs: 256 data_time: 0.74s time: 118.95s eta: 7:02:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32126.0 tgs: 269 data_time: 0.89s time: 119.29s eta: 7:01:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31708.0 tgs: 265 data_time: 0.67s time: 119.52s eta: 7:00:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31500.0 tgs: 263 data_time: 0.52s time: 119.42s eta: 6:57:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.172 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32046.0 tgs: 269 data_time: 0.94s time: 118.91s eta: 6:54:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32042.0 tgs: 268 data_time: 0.76s time: 119.35s eta: 6:53:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32363.0 tgs: 272 data_time: 0.73s time: 118.94s eta: 6:50:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 28048.0 tgs: 235 data_time: 0.48s time: 119.26s eta: 6:49:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31014.0 tgs: 260 data_time: 0.67s time: 119.02s eta: 6:46:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31741.0 tgs: 266 data_time: 0.78s time: 119.23s eta: 6:45:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.257 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32336.0 tgs: 270 data_time: 0.76s time: 119.40s eta: 6:43:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.253 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 264 data_time: 0.83s time: 119.01s eta: 6:40:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32360.0 tgs: 271 data_time: 0.95s time: 119.02s eta: 6:38:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.223 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32075.0 tgs: 268 data_time: 0.67s time: 119.46s eta: 6:38:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.183 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31339.0 tgs: 263 data_time: 0.82s time: 118.80s eta: 6:34:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 267 data_time: 1.02s time: 119.43s eta: 6:34:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.190 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31798.0 tgs: 265 data_time: 0.90s time: 119.57s eta: 6:32:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31040.0 tgs: 260 data_time: 0.76s time: 119.38s eta: 6:29:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31594.0 tgs: 264 data_time: 0.80s time: 119.61s eta: 6:28:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31690.0 tgs: 265 data_time: 0.78s time: 119.36s eta: 6:25:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30959.0 tgs: 260 data_time: 0.90s time: 118.82s eta: 6:22:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.201 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31038.0 tgs: 260 data_time: 0.79s time: 119.02s eta: 6:20:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.236 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32499.0 tgs: 272 data_time: 0.74s time: 119.40s eta: 6:20:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31509.0 tgs: 264 data_time: 0.80s time: 119.34s eta: 6:17:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32541.0 tgs: 273 data_time: 0.82s time: 119.05s eta: 6:15:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31399.0 tgs: 262 data_time: 0.71s time: 119.41s eta: 6:14:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32182.0 tgs: 269 data_time: 0.69s time: 119.38s eta: 6:12:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.190 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31938.0 tgs: 267 data_time: 1.05s time: 119.28s eta: 6:09:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.178 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31152.0 tgs: 261 data_time: 0.67s time: 119.10s eta: 6:07:13 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.110 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 27695.0 tgs: 227 data_time: 0.51s time: 121.96s eta: 6:14:01 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.198 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31417.0 tgs: 263 data_time: 0.54s time: 119.39s eta: 6:04:07 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31945.0 tgs: 267 data_time: 0.97s time: 119.57s eta: 6:02:41 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.184 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31277.0 tgs: 262 data_time: 0.91s time: 119.26s eta: 5:59:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.141 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 270 data_time: 0.95s time: 119.27s eta: 5:57:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31796.0 tgs: 267 data_time: 0.64s time: 118.91s eta: 5:54:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.113 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31283.0 tgs: 262 data_time: 0.74s time: 119.02s eta: 5:53:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.124 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31575.0 tgs: 265 data_time: 0.74s time: 119.07s eta: 5:51:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31841.0 tgs: 266 data_time: 0.55s time: 119.29s eta: 5:49:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.212 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31425.0 tgs: 264 data_time: 0.73s time: 118.97s eta: 5:47:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.132 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30675.0 tgs: 257 data_time: 0.83s time: 119.27s eta: 5:45:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31842.0 tgs: 266 data_time: 0.81s time: 119.36s eta: 5:44:08 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.185 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 268 data_time: 0.88s time: 119.06s eta: 5:41:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.174 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31432.0 tgs: 263 data_time: 0.59s time: 119.49s eta: 5:40:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31746.0 tgs: 265 data_time: 0.57s time: 119.38s eta: 5:38:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.172 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31867.0 tgs: 266 data_time: 0.84s time: 119.57s eta: 5:36:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32481.0 tgs: 273 data_time: 0.86s time: 118.98s eta: 5:33:08 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31778.0 tgs: 267 data_time: 0.97s time: 118.95s eta: 5:31:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.118 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32010.0 tgs: 268 data_time: 0.59s time: 119.25s eta: 5:29:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30687.0 tgs: 257 data_time: 0.75s time: 119.39s eta: 5:28:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32287.0 tgs: 270 data_time: 0.74s time: 119.46s eta: 5:26:31 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31042.0 tgs: 260 data_time: 0.63s time: 118.95s eta: 5:23:08 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30746.0 tgs: 257 data_time: 0.89s time: 119.63s eta: 5:23:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.154 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 266 data_time: 0.68s time: 119.35s eta: 5:20:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30113.0 tgs: 252 data_time: 1.10s time: 119.12s eta: 5:17:39 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 265 data_time: 0.95s time: 119.42s eta: 5:16:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 266 data_time: 0.52s time: 119.59s eta: 5:14:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.208 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31883.0 tgs: 267 data_time: 0.64s time: 119.01s eta: 5:11:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31400.0 tgs: 263 data_time: 0.73s time: 119.00s eta: 5:09:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.123 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30693.0 tgs: 257 data_time: 0.63s time: 119.22s eta: 5:07:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31795.0 tgs: 265 data_time: 0.76s time: 119.69s eta: 5:07:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32314.0 tgs: 271 data_time: 0.86s time: 119.11s eta: 5:03:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.191 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32438.0 tgs: 272 data_time: 0.77s time: 119.07s eta: 5:01:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31933.0 tgs: 262 data_time: 0.81s time: 121.52s eta: 5:05:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.226 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 266 data_time: 0.93s time: 119.57s eta: 4:58:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32384.0 tgs: 270 data_time: 0.72s time: 119.50s eta: 4:56:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31355.0 tgs: 261 data_time: 0.66s time: 119.73s eta: 4:55:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.202 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31711.0 tgs: 265 data_time: 0.64s time: 119.52s eta: 4:52:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.131 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31718.0 tgs: 265 data_time: 0.77s time: 119.41s eta: 4:50:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31466.0 tgs: 263 data_time: 0.75s time: 119.21s eta: 4:48:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.197 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30970.0 tgs: 260 data_time: 0.95s time: 118.89s eta: 4:45:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.190 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32051.0 tgs: 269 data_time: 0.64s time: 119.08s eta: 4:43:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.181 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31215.0 tgs: 260 data_time: 0.80s time: 119.98s eta: 4:43:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.143 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31102.0 tgs: 261 data_time: 0.89s time: 119.15s eta: 4:39:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31597.0 tgs: 265 data_time: 0.88s time: 119.01s eta: 4:37:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31494.0 tgs: 264 data_time: 0.78s time: 119.06s eta: 4:35:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.164 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31468.0 tgs: 263 data_time: 0.79s time: 119.37s eta: 4:34:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31405.0 tgs: 263 data_time: 0.70s time: 119.40s eta: 4:32:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31506.0 tgs: 264 data_time: 0.92s time: 118.94s eta: 4:29:35 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31662.0 tgs: 264 data_time: 0.93s time: 119.60s eta: 4:29:05 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32441.0 tgs: 272 data_time: 0.66s time: 119.24s eta: 4:26:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.110 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32273.0 tgs: 271 data_time: 0.61s time: 119.06s eta: 4:23:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32161.0 tgs: 270 data_time: 0.92s time: 118.85s eta: 4:21:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32087.0 tgs: 270 data_time: 0.78s time: 118.76s eta: 4:19:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30724.0 tgs: 257 data_time: 0.64s time: 119.43s eta: 4:18:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.108 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31884.0 tgs: 267 data_time: 0.75s time: 119.40s eta: 4:16:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 265 data_time: 0.80s time: 119.36s eta: 4:14:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.099 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30653.0 tgs: 257 data_time: 0.58s time: 119.05s eta: 4:11:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31569.0 tgs: 265 data_time: 0.68s time: 119.12s eta: 4:10:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.130 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.61s time: 119.13s eta: 4:08:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31974.0 tgs: 268 data_time: 0.63s time: 118.95s eta: 4:05:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30777.0 tgs: 258 data_time: 0.84s time: 119.15s eta: 4:04:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 28548.0 tgs: 238 data_time: 0.58s time: 119.50s eta: 4:02:58 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31850.0 tgs: 267 data_time: 0.72s time: 119.25s eta: 4:00:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31581.0 tgs: 264 data_time: 0.60s time: 119.19s eta: 3:58:23 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31185.0 tgs: 261 data_time: 0.69s time: 119.36s eta: 3:56:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32171.0 tgs: 269 data_time: 0.65s time: 119.37s eta: 3:54:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32087.0 tgs: 267 data_time: 0.55s time: 119.96s eta: 3:53:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32135.0 tgs: 269 data_time: 0.85s time: 119.31s eta: 3:50:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 270 data_time: 0.78s time: 119.63s eta: 3:49:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 267 data_time: 0.65s time: 118.86s eta: 3:45:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31697.0 tgs: 264 data_time: 0.86s time: 119.73s eta: 3:45:28 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32151.0 tgs: 269 data_time: 0.84s time: 119.32s eta: 3:42:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31616.0 tgs: 265 data_time: 0.90s time: 119.26s eta: 3:40:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.200 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 268 data_time: 0.71s time: 119.00s eta: 3:38:09 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31652.0 tgs: 265 data_time: 0.70s time: 119.43s eta: 3:36:57 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.82s time: 119.04s eta: 3:34:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32122.0 tgs: 269 data_time: 0.73s time: 118.98s eta: 3:32:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30971.0 tgs: 259 data_time: 0.68s time: 119.22s eta: 3:30:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31857.0 tgs: 265 data_time: 1.05s time: 120.11s eta: 3:30:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.155 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31749.0 tgs: 266 data_time: 0.64s time: 119.13s eta: 3:26:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.127 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32341.0 tgs: 271 data_time: 0.78s time: 119.31s eta: 3:24:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.129 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31570.0 tgs: 265 data_time: 0.70s time: 119.08s eta: 3:22:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32029.0 tgs: 268 data_time: 0.66s time: 119.43s eta: 3:21:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31717.0 tgs: 265 data_time: 0.49s time: 119.28s eta: 3:18:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32113.0 tgs: 269 data_time: 1.02s time: 119.22s eta: 3:16:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.119 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31484.0 tgs: 264 data_time: 0.65s time: 119.19s eta: 3:14:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.207 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32141.0 tgs: 270 data_time: 0.72s time: 119.02s eta: 3:12:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31449.0 tgs: 263 data_time: 0.61s time: 119.19s eta: 3:10:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32068.0 tgs: 265 data_time: 0.76s time: 120.95s eta: 3:11:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31243.0 tgs: 261 data_time: 0.68s time: 119.50s eta: 3:07:13 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:41:14][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31898.0 tgs: 267 data_time: 0.66s time: 119.26s eta: 3:04:51 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31688.0 tgs: 266 data_time: 0.70s time: 118.98s eta: 3:02:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31515.0 tgs: 264 data_time: 0.82s time: 119.22s eta: 3:00:48 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31375.0 tgs: 262 data_time: 0.90s time: 119.46s eta: 2:59:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30666.0 tgs: 257 data_time: 0.83s time: 118.98s eta: 2:56:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30966.0 tgs: 259 data_time: 0.63s time: 119.35s eta: 2:55:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30984.0 tgs: 259 data_time: 0.73s time: 119.41s eta: 2:53:08 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 260 data_time: 0.96s time: 119.36s eta: 2:51:04 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31662.0 tgs: 260 data_time: 0.79s time: 121.46s eta: 2:52:03 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.179 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31340.0 tgs: 262 data_time: 0.92s time: 119.24s eta: 2:46:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31549.0 tgs: 263 data_time: 0.61s time: 119.57s eta: 2:45:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.184 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31700.0 tgs: 266 data_time: 0.93s time: 118.90s eta: 2:42:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31736.0 tgs: 266 data_time: 0.73s time: 119.18s eta: 2:40:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32339.0 tgs: 271 data_time: 0.73s time: 118.96s eta: 2:38:37 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31674.0 tgs: 265 data_time: 0.94s time: 119.23s eta: 2:36:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.238 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 267 data_time: 0.68s time: 119.11s eta: 2:34:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.115 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29915.0 tgs: 250 data_time: 0.54s time: 119.43s eta: 2:33:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31868.0 tgs: 265 data_time: 0.91s time: 120.23s eta: 2:32:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31974.0 tgs: 268 data_time: 0.88s time: 119.28s eta: 2:29:06 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.226 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31924.0 tgs: 268 data_time: 0.92s time: 118.88s eta: 2:26:36 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31761.0 tgs: 265 data_time: 1.03s time: 119.52s eta: 2:25:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:23:00][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32346.0 tgs: 271 data_time: 0.75s time: 119.07s eta: 2:22:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30579.0 tgs: 256 data_time: 0.79s time: 119.14s eta: 2:20:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31507.0 tgs: 264 data_time: 0.61s time: 119.05s eta: 2:18:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31704.0 tgs: 264 data_time: 0.86s time: 119.84s eta: 2:17:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32298.0 tgs: 270 data_time: 0.71s time: 119.35s eta: 2:15:15 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30607.0 tgs: 256 data_time: 0.86s time: 119.39s eta: 2:13:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30088.0 tgs: 251 data_time: 0.83s time: 119.67s eta: 2:11:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31305.0 tgs: 261 data_time: 0.97s time: 119.49s eta: 2:09:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31969.0 tgs: 268 data_time: 0.69s time: 118.96s eta: 2:06:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31990.0 tgs: 267 data_time: 0.69s time: 119.38s eta: 2:05:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30637.0 tgs: 256 data_time: 0.63s time: 119.27s eta: 2:03:14 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31282.0 tgs: 262 data_time: 0.82s time: 119.22s eta: 2:01:12 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31149.0 tgs: 260 data_time: 0.78s time: 119.51s eta: 1:59:30 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30245.0 tgs: 253 data_time: 0.68s time: 119.42s eta: 1:57:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31829.0 tgs: 266 data_time: 0.67s time: 119.53s eta: 1:55:32 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31972.0 tgs: 267 data_time: 0.71s time: 119.30s eta: 1:53:20 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 271 data_time: 0.80s time: 119.25s eta: 1:51:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31406.0 tgs: 264 data_time: 0.77s time: 118.93s eta: 1:49:00 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32043.0 tgs: 267 data_time: 0.53s time: 119.64s eta: 1:47:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31924.0 tgs: 266 data_time: 0.95s time: 119.71s eta: 1:45:44 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31297.0 tgs: 262 data_time: 0.73s time: 119.35s eta: 1:43:26 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.115 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31855.0 tgs: 267 data_time: 0.60s time: 119.30s eta: 1:41:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32159.0 tgs: 269 data_time: 0.66s time: 119.28s eta: 1:39:24 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31392.0 tgs: 264 data_time: 0.72s time: 118.90s eta: 1:37:06 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29984.0 tgs: 251 data_time: 0.66s time: 119.16s eta: 1:35:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32259.0 tgs: 270 data_time: 0.78s time: 119.06s eta: 1:33:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 264 data_time: 0.87s time: 119.26s eta: 1:31:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31849.0 tgs: 266 data_time: 0.75s time: 119.73s eta: 1:29:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.163 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30912.0 tgs: 259 data_time: 0.71s time: 119.01s eta: 1:27:16 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31331.0 tgs: 263 data_time: 0.55s time: 118.87s eta: 1:25:11 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.177 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31823.0 tgs: 266 data_time: 0.60s time: 119.48s eta: 1:23:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.143 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31554.0 tgs: 264 data_time: 0.81s time: 119.48s eta: 1:21:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31996.0 tgs: 269 data_time: 0.89s time: 118.94s eta: 1:19:17 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.211 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32267.0 tgs: 271 data_time: 1.04s time: 118.93s eta: 1:17:18 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32411.0 tgs: 265 data_time: 0.72s time: 122.16s eta: 1:17:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30612.0 tgs: 256 data_time: 0.66s time: 119.36s eta: 1:13:36 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30724.0 tgs: 258 data_time: 0.71s time: 119.05s eta: 1:11:25 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.181 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31463.0 tgs: 264 data_time: 0.62s time: 118.92s eta: 1:09:22 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32107.0 tgs: 269 data_time: 1.01s time: 119.10s eta: 1:07:29 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31192.0 tgs: 261 data_time: 0.67s time: 119.41s eta: 1:05:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.217 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 269 data_time: 0.83s time: 119.34s eta: 1:03:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32548.0 tgs: 273 data_time: 0.82s time: 119.18s eta: 1:01:34 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32385.0 tgs: 271 data_time: 0.80s time: 119.28s eta: 0:59:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31658.0 tgs: 265 data_time: 0.80s time: 119.44s eta: 0:57:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.106 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31779.0 tgs: 266 data_time: 0.71s time: 119.40s eta: 0:55:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32641.0 tgs: 273 data_time: 0.62s time: 119.48s eta: 0:53:45 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30724.0 tgs: 257 data_time: 0.70s time: 119.37s eta: 0:51:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31732.0 tgs: 265 data_time: 0.79s time: 119.47s eta: 0:49:46 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31603.0 tgs: 265 data_time: 0.70s time: 118.89s eta: 0:47:33 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31132.0 tgs: 260 data_time: 0.66s time: 119.28s eta: 0:45:43 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31350.0 tgs: 262 data_time: 0.78s time: 119.20s eta: 0:43:42 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.174 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32332.0 tgs: 271 data_time: 0.65s time: 118.99s eta: 0:41:38 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31688.0 tgs: 266 data_time: 0.72s time: 119.04s eta: 0:39:40 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30402.0 tgs: 254 data_time: 0.68s time: 119.66s eta: 0:37:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31274.0 tgs: 262 data_time: 0.49s time: 119.28s eta: 0:35:47 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32075.0 tgs: 267 data_time: 0.64s time: 119.74s eta: 0:33:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.172 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32578.0 tgs: 268 data_time: 0.60s time: 121.22s eta: 0:32:19 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 266 data_time: 0.62s time: 119.38s eta: 0:29:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.182 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31789.0 tgs: 266 data_time: 0.77s time: 119.32s eta: 0:27:50 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.192 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 266 data_time: 0.70s time: 119.64s eta: 0:25:55 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31962.0 tgs: 268 data_time: 1.12s time: 119.09s eta: 0:23:49 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31704.0 tgs: 265 data_time: 0.95s time: 119.39s eta: 0:21:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 270 data_time: 0.93s time: 119.26s eta: 0:19:52 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.132 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31685.0 tgs: 265 data_time: 0.55s time: 119.25s eta: 0:17:53 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32089.0 tgs: 267 data_time: 0.65s time: 119.92s eta: 0:15:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30888.0 tgs: 259 data_time: 0.66s time: 119.20s eta: 0:13:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 269 data_time: 0.97s time: 119.09s eta: 0:11:54 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 269 data_time: 0.71s time: 119.32s eta: 0:09:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32520.0 tgs: 272 data_time: 0.70s time: 119.17s eta: 0:07:56 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32425.0 tgs: 270 data_time: 0.70s time: 119.69s eta: 0:05:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31730.0 tgs: 261 data_time: 0.74s time: 121.40s eta: 0:04:02 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31112.0 tgs: 260 data_time: 0.62s time: 119.21s eta: 0:01:59 +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 62][DP 15][SP 2][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank7.log b/20250121165312/rank7.log new file mode 100644 index 0000000000000000000000000000000000000000..2f0ed85dc8c81893b42311cc98ca115c58c87c54 --- /dev/null +++ b/20250121165312/rank7.log @@ -0,0 +1,851 @@ +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:19][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:23][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:28][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:30][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:32][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.05s +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:00:45][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 69.27 seconds, peak gpu memory 13.4G +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.303 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 32.6GB text_tokens: 31121.0 tgs: 203 data_time: 2.13s time: 153.05s eta: 1 day, 1:12:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.273 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31200.0 tgs: 259 data_time: 0.84s time: 120.32s eta: 19:47:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.288 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30861.0 tgs: 257 data_time: 0.92s time: 119.68s eta: 19:38:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.240 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31960.0 tgs: 267 data_time: 0.94s time: 119.65s eta: 19:36:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.248 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31620.0 tgs: 265 data_time: 1.14s time: 119.31s eta: 19:31:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.284 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31899.0 tgs: 266 data_time: 0.73s time: 119.53s eta: 19:31:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.209 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32173.0 tgs: 261 data_time: 0.97s time: 122.96s eta: 20:02:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.196 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32073.0 tgs: 267 data_time: 0.74s time: 119.89s eta: 19:30:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.226 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32230.0 tgs: 267 data_time: 0.83s time: 120.65s eta: 19:36:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.209 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32211.0 tgs: 268 data_time: 0.70s time: 119.91s eta: 19:27:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31612.0 tgs: 262 data_time: 0.86s time: 120.46s eta: 19:30:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31664.0 tgs: 265 data_time: 0.66s time: 119.34s eta: 19:17:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31279.0 tgs: 261 data_time: 0.67s time: 119.68s eta: 19:18:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31574.0 tgs: 264 data_time: 0.86s time: 119.20s eta: 19:12:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31608.0 tgs: 265 data_time: 0.72s time: 119.22s eta: 19:10:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32396.0 tgs: 272 data_time: 0.85s time: 118.93s eta: 19:05:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.234 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32445.0 tgs: 272 data_time: 0.77s time: 119.14s eta: 19:05:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32483.0 tgs: 272 data_time: 0.86s time: 119.18s eta: 19:04:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30947.0 tgs: 259 data_time: 0.87s time: 119.40s eta: 19:04:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32390.0 tgs: 271 data_time: 0.77s time: 119.33s eta: 19:01:32 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.210 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31866.0 tgs: 264 data_time: 0.85s time: 120.31s eta: 19:08:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.207 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31128.0 tgs: 253 data_time: 0.82s time: 122.91s eta: 19:31:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32297.0 tgs: 270 data_time: 0.68s time: 119.48s eta: 18:57:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32495.0 tgs: 272 data_time: 1.03s time: 119.16s eta: 18:51:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31533.0 tgs: 264 data_time: 0.71s time: 119.24s eta: 18:50:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.175 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31074.0 tgs: 260 data_time: 0.74s time: 119.39s eta: 18:50:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.246 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32081.0 tgs: 268 data_time: 0.83s time: 119.67s eta: 18:50:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31457.0 tgs: 263 data_time: 0.79s time: 119.38s eta: 18:46:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31099.0 tgs: 259 data_time: 0.71s time: 119.79s eta: 18:48:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32202.0 tgs: 249 data_time: 0.74s time: 129.31s eta: 20:15:32 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.246 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30996.0 tgs: 259 data_time: 0.74s time: 119.57s eta: 18:41:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.148 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31401.0 tgs: 263 data_time: 0.93s time: 119.39s eta: 18:38:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31506.0 tgs: 261 data_time: 0.83s time: 120.66s eta: 18:48:07 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31849.0 tgs: 266 data_time: 0.79s time: 119.40s eta: 18:34:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31503.0 tgs: 264 data_time: 0.63s time: 119.16s eta: 18:30:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.215 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32241.0 tgs: 270 data_time: 0.75s time: 119.07s eta: 18:27:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.224 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31502.0 tgs: 264 data_time: 0.84s time: 119.13s eta: 18:25:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29886.0 tgs: 248 data_time: 0.94s time: 120.39s eta: 18:35:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32309.0 tgs: 266 data_time: 0.74s time: 121.17s eta: 18:40:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31388.0 tgs: 260 data_time: 0.84s time: 120.67s eta: 18:34:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.256 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31748.0 tgs: 265 data_time: 0.79s time: 119.73s eta: 18:23:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31342.0 tgs: 253 data_time: 0.67s time: 123.87s eta: 18:59:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31840.0 tgs: 266 data_time: 0.87s time: 119.56s eta: 18:17:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.172 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31789.0 tgs: 266 data_time: 0.65s time: 119.36s eta: 18:14:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.170 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31941.0 tgs: 267 data_time: 0.80s time: 119.23s eta: 18:10:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30569.0 tgs: 255 data_time: 0.75s time: 119.60s eta: 18:12:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.135 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31307.0 tgs: 262 data_time: 0.81s time: 119.26s eta: 18:07:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31568.0 tgs: 265 data_time: 0.76s time: 119.01s eta: 18:02:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.259 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31707.0 tgs: 266 data_time: 1.06s time: 119.10s eta: 18:01:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.208 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31866.0 tgs: 267 data_time: 0.88s time: 119.26s eta: 18:01:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31584.0 tgs: 264 data_time: 0.89s time: 119.33s eta: 17:59:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31870.0 tgs: 265 data_time: 0.92s time: 120.02s eta: 18:04:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.141 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30551.0 tgs: 256 data_time: 0.58s time: 119.10s eta: 17:53:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31961.0 tgs: 267 data_time: 0.80s time: 119.41s eta: 17:54:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.160 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31763.0 tgs: 266 data_time: 0.99s time: 119.00s eta: 17:48:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.199 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31917.0 tgs: 268 data_time: 1.01s time: 119.02s eta: 17:47:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31804.0 tgs: 253 data_time: 0.69s time: 125.36s eta: 18:41:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.153 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32278.0 tgs: 268 data_time: 0.96s time: 120.10s eta: 17:52:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 267 data_time: 0.78s time: 119.40s eta: 17:44:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31369.0 tgs: 262 data_time: 0.96s time: 119.34s eta: 17:42:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32214.0 tgs: 269 data_time: 0.80s time: 119.51s eta: 17:41:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31816.0 tgs: 252 data_time: 0.89s time: 126.03s eta: 18:37:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31958.0 tgs: 260 data_time: 0.61s time: 122.65s eta: 18:05:27 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31545.0 tgs: 260 data_time: 0.97s time: 121.15s eta: 17:50:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.157 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31565.0 tgs: 265 data_time: 0.74s time: 119.03s eta: 17:29:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.171 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31753.0 tgs: 267 data_time: 0.70s time: 118.85s eta: 17:25:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.195 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30827.0 tgs: 248 data_time: 0.72s time: 124.00s eta: 18:09:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31476.0 tgs: 252 data_time: 1.06s time: 124.56s eta: 18:12:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31389.0 tgs: 262 data_time: 0.42s time: 119.39s eta: 17:24:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.132 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30909.0 tgs: 258 data_time: 0.88s time: 119.70s eta: 17:25:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32110.0 tgs: 268 data_time: 0.80s time: 119.53s eta: 17:21:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31386.0 tgs: 264 data_time: 0.95s time: 118.85s eta: 17:13:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.243 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31265.0 tgs: 260 data_time: 0.76s time: 119.85s eta: 17:20:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.156 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31690.0 tgs: 265 data_time: 0.85s time: 119.56s eta: 17:16:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31628.0 tgs: 259 data_time: 0.75s time: 121.86s eta: 17:34:04 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30827.0 tgs: 249 data_time: 0.86s time: 123.51s eta: 17:46:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31817.0 tgs: 267 data_time: 0.77s time: 119.01s eta: 17:05:27 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.187 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31929.0 tgs: 267 data_time: 0.55s time: 119.23s eta: 17:05:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31847.0 tgs: 267 data_time: 0.63s time: 119.14s eta: 17:02:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.191 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30582.0 tgs: 256 data_time: 0.44s time: 119.25s eta: 17:01:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.198 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31973.0 tgs: 268 data_time: 1.01s time: 119.04s eta: 16:57:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31757.0 tgs: 266 data_time: 0.90s time: 119.28s eta: 16:57:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.143 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31980.0 tgs: 268 data_time: 0.61s time: 119.16s eta: 16:54:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.176 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31403.0 tgs: 262 data_time: 0.68s time: 119.58s eta: 16:56:27 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.212 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31893.0 tgs: 267 data_time: 0.73s time: 119.13s eta: 16:50:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.225 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31919.0 tgs: 268 data_time: 0.73s time: 118.84s eta: 16:46:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.151 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31438.0 tgs: 264 data_time: 0.62s time: 119.01s eta: 16:45:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32392.0 tgs: 270 data_time: 0.64s time: 119.85s eta: 16:50:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31675.0 tgs: 265 data_time: 0.81s time: 119.40s eta: 16:44:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.255 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31027.0 tgs: 260 data_time: 0.94s time: 119.01s eta: 16:39:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.178 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31959.0 tgs: 264 data_time: 0.72s time: 120.92s eta: 16:53:44 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31398.0 tgs: 263 data_time: 0.68s time: 118.99s eta: 16:35:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.161 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30730.0 tgs: 257 data_time: 0.56s time: 119.28s eta: 16:36:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31927.0 tgs: 267 data_time: 0.85s time: 119.23s eta: 16:33:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30756.0 tgs: 257 data_time: 0.70s time: 119.36s eta: 16:32:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.180 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31124.0 tgs: 260 data_time: 0.87s time: 119.70s eta: 16:33:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31870.0 tgs: 267 data_time: 0.82s time: 119.20s eta: 16:27:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31988.0 tgs: 267 data_time: 0.69s time: 119.76s eta: 16:30:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32040.0 tgs: 266 data_time: 0.87s time: 120.24s eta: 16:31:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.253 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31523.0 tgs: 264 data_time: 0.70s time: 118.97s eta: 16:19:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31958.0 tgs: 250 data_time: 0.80s time: 127.33s eta: 17:26:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31178.0 tgs: 261 data_time: 0.84s time: 119.22s eta: 16:17:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.202 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31326.0 tgs: 262 data_time: 0.63s time: 119.14s eta: 16:14:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.147 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31894.0 tgs: 267 data_time: 0.66s time: 119.24s eta: 16:13:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32095.0 tgs: 269 data_time: 0.71s time: 119.17s eta: 16:11:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32095.0 tgs: 269 data_time: 0.72s time: 118.95s eta: 16:07:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30958.0 tgs: 260 data_time: 0.73s time: 118.85s eta: 16:04:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.175 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32098.0 tgs: 269 data_time: 0.70s time: 118.96s eta: 16:03:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.199 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32395.0 tgs: 265 data_time: 0.71s time: 121.84s eta: 16:24:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32364.0 tgs: 271 data_time: 0.81s time: 118.99s eta: 15:59:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31865.0 tgs: 267 data_time: 0.91s time: 119.23s eta: 15:59:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32108.0 tgs: 268 data_time: 0.58s time: 119.61s eta: 16:00:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:48:49][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.143 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31679.0 tgs: 266 data_time: 0.89s time: 119.04s eta: 15:54:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31865.0 tgs: 267 data_time: 0.82s time: 118.98s eta: 15:51:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.252 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32334.0 tgs: 271 data_time: 0.53s time: 119.17s eta: 15:51:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.253 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32210.0 tgs: 269 data_time: 0.77s time: 119.63s eta: 15:53:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31761.0 tgs: 267 data_time: 1.02s time: 118.93s eta: 15:45:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31383.0 tgs: 263 data_time: 0.90s time: 119.28s eta: 15:46:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.127 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31380.0 tgs: 263 data_time: 0.66s time: 119.14s eta: 15:43:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30738.0 tgs: 258 data_time: 0.83s time: 119.01s eta: 15:40:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.186 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30780.0 tgs: 257 data_time: 0.69s time: 119.65s eta: 15:43:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31099.0 tgs: 255 data_time: 0.76s time: 121.50s eta: 15:55:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.218 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30077.0 tgs: 251 data_time: 0.90s time: 119.41s eta: 15:37:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 244 data_time: 0.55s time: 129.96s eta: 16:58:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32419.0 tgs: 271 data_time: 1.08s time: 119.46s eta: 15:33:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.131 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30440.0 tgs: 254 data_time: 0.60s time: 119.60s eta: 15:32:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.122 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32007.0 tgs: 267 data_time: 0.74s time: 119.53s eta: 15:30:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31413.0 tgs: 257 data_time: 0.77s time: 122.14s eta: 15:48:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.145 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30763.0 tgs: 254 data_time: 0.99s time: 120.75s eta: 15:35:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.107 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31478.0 tgs: 263 data_time: 1.05s time: 119.41s eta: 15:23:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.142 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31029.0 tgs: 259 data_time: 0.89s time: 119.44s eta: 15:21:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.125 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31717.0 tgs: 240 data_time: 1.00s time: 131.63s eta: 16:53:32 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31425.0 tgs: 263 data_time: 0.85s time: 119.29s eta: 15:16:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31450.0 tgs: 263 data_time: 0.60s time: 119.56s eta: 15:16:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30888.0 tgs: 252 data_time: 0.62s time: 122.25s eta: 15:35:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31977.0 tgs: 267 data_time: 0.69s time: 119.71s eta: 15:13:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.148 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31730.0 tgs: 265 data_time: 0.65s time: 119.34s eta: 15:08:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.181 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32646.0 tgs: 274 data_time: 0.82s time: 118.90s eta: 15:03:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.127 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30572.0 tgs: 256 data_time: 0.69s time: 118.99s eta: 15:02:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31909.0 tgs: 267 data_time: 0.64s time: 119.31s eta: 15:02:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31772.0 tgs: 267 data_time: 0.77s time: 118.92s eta: 14:57:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32139.0 tgs: 266 data_time: 0.62s time: 120.52s eta: 15:07:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.189 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31865.0 tgs: 267 data_time: 0.53s time: 119.31s eta: 14:56:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.162 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31565.0 tgs: 257 data_time: 0.95s time: 122.65s eta: 15:19:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.201 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32240.0 tgs: 270 data_time: 0.92s time: 119.12s eta: 14:51:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.167 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31375.0 tgs: 261 data_time: 0.53s time: 119.75s eta: 14:54:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31901.0 tgs: 267 data_time: 1.04s time: 119.46s eta: 14:50:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31560.0 tgs: 265 data_time: 1.03s time: 118.99s eta: 14:44:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31880.0 tgs: 267 data_time: 0.82s time: 119.35s eta: 14:45:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31745.0 tgs: 266 data_time: 0.54s time: 118.95s eta: 14:40:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32466.0 tgs: 272 data_time: 0.78s time: 119.05s eta: 14:38:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32075.0 tgs: 269 data_time: 0.90s time: 119.06s eta: 14:37:04 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.216 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31204.0 tgs: 261 data_time: 0.64s time: 119.12s eta: 14:35:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32130.0 tgs: 270 data_time: 0.79s time: 119.00s eta: 14:32:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31219.0 tgs: 262 data_time: 0.69s time: 119.07s eta: 14:31:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.205 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31694.0 tgs: 266 data_time: 0.67s time: 119.15s eta: 14:29:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.148 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29250.0 tgs: 244 data_time: 0.75s time: 119.65s eta: 14:31:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32365.0 tgs: 271 data_time: 0.78s time: 119.18s eta: 14:26:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31427.0 tgs: 263 data_time: 0.67s time: 119.20s eta: 14:24:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.132 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31236.0 tgs: 261 data_time: 0.70s time: 119.39s eta: 14:23:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.172 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32289.0 tgs: 267 data_time: 0.79s time: 120.73s eta: 14:31:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32092.0 tgs: 270 data_time: 0.79s time: 118.85s eta: 14:15:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.147 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31177.0 tgs: 261 data_time: 0.94s time: 119.10s eta: 14:15:30 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.096 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29651.0 tgs: 248 data_time: 0.82s time: 119.15s eta: 14:13:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.168 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31710.0 tgs: 264 data_time: 0.86s time: 119.96s eta: 14:17:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.142 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31200.0 tgs: 260 data_time: 0.83s time: 119.86s eta: 14:15:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.121 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31652.0 tgs: 264 data_time: 0.43s time: 119.58s eta: 14:11:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.151 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31725.0 tgs: 266 data_time: 0.57s time: 119.22s eta: 14:06:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31653.0 tgs: 265 data_time: 0.77s time: 119.26s eta: 14:04:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.177 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 28771.0 tgs: 241 data_time: 0.47s time: 119.27s eta: 14:02:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.217 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30512.0 tgs: 256 data_time: 0.71s time: 119.14s eta: 13:59:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.187 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31814.0 tgs: 267 data_time: 0.70s time: 119.12s eta: 13:57:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.190 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 268 data_time: 0.65s time: 119.17s eta: 13:56:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.137 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31803.0 tgs: 266 data_time: 0.81s time: 119.34s eta: 13:55:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.161 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31806.0 tgs: 264 data_time: 0.54s time: 120.11s eta: 13:58:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30861.0 tgs: 251 data_time: 0.55s time: 122.55s eta: 14:13:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:56:46][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31496.0 tgs: 264 data_time: 0.70s time: 119.10s eta: 13:47:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.134 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 268 data_time: 0.85s time: 119.37s eta: 13:47:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.193 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32122.0 tgs: 269 data_time: 1.11s time: 119.35s eta: 13:45:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.109 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31538.0 tgs: 264 data_time: 0.73s time: 119.34s eta: 13:43:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32418.0 tgs: 272 data_time: 1.08s time: 119.11s eta: 13:39:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.190 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31784.0 tgs: 266 data_time: 0.93s time: 119.25s eta: 13:38:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.152 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31232.0 tgs: 262 data_time: 0.65s time: 119.13s eta: 13:36:03 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.258 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30502.0 tgs: 255 data_time: 0.71s time: 119.22s eta: 13:34:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.189 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31234.0 tgs: 261 data_time: 0.80s time: 119.46s eta: 13:34:18 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.248 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32051.0 tgs: 269 data_time: 0.95s time: 119.08s eta: 13:29:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31706.0 tgs: 266 data_time: 0.72s time: 118.92s eta: 13:26:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31733.0 tgs: 264 data_time: 0.86s time: 119.81s eta: 13:30:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.200 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31375.0 tgs: 263 data_time: 0.97s time: 119.25s eta: 13:24:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.145 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31316.0 tgs: 262 data_time: 0.84s time: 119.16s eta: 13:22:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.173 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31743.0 tgs: 265 data_time: 0.72s time: 119.34s eta: 13:21:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32176.0 tgs: 269 data_time: 0.89s time: 119.26s eta: 13:19:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.188 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31482.0 tgs: 264 data_time: 0.85s time: 119.04s eta: 13:15:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.192 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31244.0 tgs: 261 data_time: 0.69s time: 119.37s eta: 13:15:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.206 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31400.0 tgs: 263 data_time: 0.90s time: 119.19s eta: 13:12:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32223.0 tgs: 269 data_time: 0.90s time: 119.46s eta: 13:12:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32167.0 tgs: 269 data_time: 0.84s time: 119.28s eta: 13:09:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.211 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32252.0 tgs: 270 data_time: 0.83s time: 119.14s eta: 13:06:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.177 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31860.0 tgs: 264 data_time: 0.92s time: 120.42s eta: 13:12:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.174 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32295.0 tgs: 268 data_time: 0.90s time: 120.16s eta: 13:09:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.168 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 269 data_time: 1.07s time: 119.50s eta: 13:02:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31919.0 tgs: 267 data_time: 0.65s time: 119.51s eta: 13:00:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.183 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31962.0 tgs: 267 data_time: 0.63s time: 119.57s eta: 12:59:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.154 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32469.0 tgs: 271 data_time: 0.89s time: 119.79s eta: 12:58:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31212.0 tgs: 262 data_time: 0.73s time: 119.04s eta: 12:51:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.167 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 270 data_time: 0.96s time: 119.05s eta: 12:49:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.165 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31458.0 tgs: 264 data_time: 0.78s time: 119.12s eta: 12:48:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31682.0 tgs: 265 data_time: 0.52s time: 119.39s eta: 12:48:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30946.0 tgs: 259 data_time: 0.73s time: 119.20s eta: 12:44:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.135 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30503.0 tgs: 255 data_time: 0.90s time: 119.44s eta: 12:44:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.140 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31664.0 tgs: 265 data_time: 0.75s time: 119.09s eta: 12:40:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32465.0 tgs: 272 data_time: 0.67s time: 119.07s eta: 12:38:03 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.213 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30539.0 tgs: 255 data_time: 0.64s time: 119.60s eta: 12:39:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 266 data_time: 0.50s time: 118.78s eta: 12:32:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31377.0 tgs: 263 data_time: 0.85s time: 118.88s eta: 12:30:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31891.0 tgs: 267 data_time: 0.62s time: 119.17s eta: 12:30:44 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31597.0 tgs: 264 data_time: 0.96s time: 119.27s eta: 12:29:23 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.194 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31596.0 tgs: 264 data_time: 0.61s time: 119.34s eta: 12:27:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.145 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31615.0 tgs: 265 data_time: 0.54s time: 119.12s eta: 12:24:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.201 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31693.0 tgs: 265 data_time: 1.04s time: 119.34s eta: 12:23:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.208 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31275.0 tgs: 262 data_time: 0.82s time: 119.26s eta: 12:21:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.136 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31414.0 tgs: 263 data_time: 0.71s time: 119.03s eta: 12:17:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.186 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31884.0 tgs: 263 data_time: 0.70s time: 121.13s eta: 12:28:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31301.0 tgs: 260 data_time: 0.75s time: 120.00s eta: 12:19:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.149 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32116.0 tgs: 270 data_time: 0.67s time: 118.72s eta: 12:10:07 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.188 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32076.0 tgs: 268 data_time: 0.67s time: 119.63s eta: 12:13:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 268 data_time: 0.70s time: 119.83s eta: 12:12:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.097 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31232.0 tgs: 262 data_time: 0.60s time: 119.14s eta: 12:06:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.201 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31364.0 tgs: 263 data_time: 0.97s time: 119.17s eta: 12:04:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.203 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30624.0 tgs: 257 data_time: 0.95s time: 119.11s eta: 12:02:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.255 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31739.0 tgs: 266 data_time: 0.96s time: 119.18s eta: 12:01:03 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.148 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 269 data_time: 0.87s time: 118.96s eta: 11:57:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 267 data_time: 0.70s time: 119.70s eta: 12:00:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.216 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32679.0 tgs: 273 data_time: 0.79s time: 119.28s eta: 11:55:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.122 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31804.0 tgs: 266 data_time: 0.69s time: 119.27s eta: 11:53:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.184 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32111.0 tgs: 268 data_time: 1.07s time: 119.71s eta: 11:54:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.180 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31402.0 tgs: 264 data_time: 0.88s time: 118.78s eta: 11:46:44 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.144 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31846.0 tgs: 267 data_time: 0.65s time: 119.16s eta: 11:47:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32371.0 tgs: 270 data_time: 0.78s time: 119.61s eta: 11:47:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32279.0 tgs: 271 data_time: 0.72s time: 118.87s eta: 11:41:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.196 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31716.0 tgs: 267 data_time: 0.88s time: 118.68s eta: 11:38:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31294.0 tgs: 262 data_time: 0.81s time: 119.21s eta: 11:39:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31867.0 tgs: 267 data_time: 0.66s time: 119.13s eta: 11:36:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.173 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32098.0 tgs: 269 data_time: 0.90s time: 119.08s eta: 11:34:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.130 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31651.0 tgs: 266 data_time: 0.78s time: 118.95s eta: 11:31:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31370.0 tgs: 263 data_time: 0.48s time: 118.91s eta: 11:29:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.167 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31139.0 tgs: 261 data_time: 0.91s time: 118.98s eta: 11:28:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31723.0 tgs: 265 data_time: 0.69s time: 119.28s eta: 11:27:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.132 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31960.0 tgs: 267 data_time: 0.73s time: 119.35s eta: 11:26:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.145 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31081.0 tgs: 261 data_time: 0.74s time: 118.88s eta: 11:21:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.146 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32138.0 tgs: 269 data_time: 0.67s time: 119.20s eta: 11:21:27 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31357.0 tgs: 261 data_time: 0.64s time: 119.72s eta: 11:22:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30252.0 tgs: 253 data_time: 0.56s time: 119.11s eta: 11:16:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32160.0 tgs: 269 data_time: 0.64s time: 119.12s eta: 11:15:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.169 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32206.0 tgs: 270 data_time: 0.69s time: 118.99s eta: 11:12:18 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 267 data_time: 0.83s time: 119.09s eta: 11:10:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32230.0 tgs: 270 data_time: 0.75s time: 119.15s eta: 11:09:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.166 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32335.0 tgs: 271 data_time: 0.83s time: 118.97s eta: 11:06:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31996.0 tgs: 267 data_time: 0.67s time: 119.47s eta: 11:07:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.137 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 266 data_time: 0.97s time: 119.64s eta: 11:06:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.158 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32246.0 tgs: 270 data_time: 0.60s time: 119.30s eta: 11:02:07 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.142 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31885.0 tgs: 267 data_time: 0.71s time: 119.31s eta: 11:00:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.165 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 31053.0 tgs: 261 data_time: 0.76s time: 118.82s eta: 10:55:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.239 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31841.0 tgs: 266 data_time: 0.58s time: 119.47s eta: 10:57:04 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32110.0 tgs: 268 data_time: 0.58s time: 119.37s eta: 10:54:32 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.174 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31650.0 tgs: 265 data_time: 0.56s time: 119.27s eta: 10:52:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.189 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31124.0 tgs: 260 data_time: 0.70s time: 119.25s eta: 10:49:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.187 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32085.0 tgs: 269 data_time: 0.69s time: 119.10s eta: 10:47:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31688.0 tgs: 265 data_time: 0.65s time: 119.53s eta: 10:47:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30428.0 tgs: 254 data_time: 0.64s time: 119.34s eta: 10:44:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31548.0 tgs: 264 data_time: 0.89s time: 119.49s eta: 10:43:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.197 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31922.0 tgs: 267 data_time: 0.75s time: 119.51s eta: 10:41:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30350.0 tgs: 255 data_time: 0.74s time: 118.99s eta: 10:36:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31440.0 tgs: 263 data_time: 0.61s time: 119.21s eta: 10:35:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31722.0 tgs: 266 data_time: 0.55s time: 119.06s eta: 10:32:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.120 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31123.0 tgs: 262 data_time: 0.72s time: 118.78s eta: 10:29:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32160.0 tgs: 268 data_time: 0.65s time: 119.86s eta: 10:33:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.185 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31121.0 tgs: 260 data_time: 0.79s time: 119.38s eta: 10:28:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.175 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31568.0 tgs: 265 data_time: 0.82s time: 118.83s eta: 10:23:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.173 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32076.0 tgs: 269 data_time: 0.98s time: 118.98s eta: 10:22:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 270 data_time: 0.84s time: 119.28s eta: 10:22:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.141 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31533.0 tgs: 264 data_time: 0.74s time: 119.20s eta: 10:19:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.153 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30617.0 tgs: 257 data_time: 0.76s time: 119.06s eta: 10:17:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.150 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31023.0 tgs: 259 data_time: 0.62s time: 119.47s eta: 10:17:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31981.0 tgs: 267 data_time: 0.80s time: 119.42s eta: 10:15:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.197 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32188.0 tgs: 269 data_time: 0.80s time: 119.46s eta: 10:13:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.176 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31454.0 tgs: 263 data_time: 0.69s time: 119.29s eta: 10:10:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31844.0 tgs: 267 data_time: 0.64s time: 119.11s eta: 10:07:27 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31256.0 tgs: 261 data_time: 0.84s time: 119.70s eta: 10:08:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.117 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31006.0 tgs: 258 data_time: 0.84s time: 119.82s eta: 10:07:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.144 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32260.0 tgs: 265 data_time: 0.40s time: 121.55s eta: 10:13:50 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31829.0 tgs: 266 data_time: 0.83s time: 119.25s eta: 10:00:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.195 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 263 data_time: 0.84s time: 118.80s eta: 9:56:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.128 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32025.0 tgs: 267 data_time: 0.79s time: 119.56s eta: 9:57:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.235 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32432.0 tgs: 270 data_time: 0.87s time: 119.70s eta: 9:56:30 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.091 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31904.0 tgs: 267 data_time: 0.61s time: 119.26s eta: 9:52:18 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.136 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31232.0 tgs: 261 data_time: 0.57s time: 119.33s eta: 9:50:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.224 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32155.0 tgs: 269 data_time: 0.72s time: 119.22s eta: 9:48:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.225 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32199.0 tgs: 266 data_time: 1.08s time: 120.80s eta: 9:53:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.144 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31943.0 tgs: 267 data_time: 0.76s time: 119.25s eta: 9:44:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.133 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32072.0 tgs: 268 data_time: 0.67s time: 119.46s eta: 9:43:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.179 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31197.0 tgs: 260 data_time: 0.81s time: 119.54s eta: 9:41:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31536.0 tgs: 264 data_time: 0.88s time: 119.20s eta: 9:38:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.177 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31646.0 tgs: 264 data_time: 0.84s time: 119.48s eta: 9:37:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31473.0 tgs: 264 data_time: 0.68s time: 119.05s eta: 9:33:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.193 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32219.0 tgs: 270 data_time: 0.52s time: 119.24s eta: 9:32:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30515.0 tgs: 256 data_time: 0.50s time: 119.16s eta: 9:29:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.211 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32120.0 tgs: 269 data_time: 0.64s time: 119.22s eta: 9:28:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.167 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31478.0 tgs: 263 data_time: 0.55s time: 119.46s eta: 9:27:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32223.0 tgs: 270 data_time: 0.87s time: 118.98s eta: 9:23:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.186 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30954.0 tgs: 260 data_time: 0.82s time: 119.03s eta: 9:21:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.155 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32216.0 tgs: 269 data_time: 0.65s time: 119.53s eta: 9:21:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.176 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32131.0 tgs: 268 data_time: 0.65s time: 119.59s eta: 9:20:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.190 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31640.0 tgs: 265 data_time: 0.56s time: 119.12s eta: 9:15:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.117 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31717.0 tgs: 266 data_time: 0.62s time: 119.07s eta: 9:13:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.149 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31846.0 tgs: 267 data_time: 0.98s time: 118.90s eta: 9:10:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31510.0 tgs: 264 data_time: 0.77s time: 119.32s eta: 9:10:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31930.0 tgs: 268 data_time: 0.75s time: 119.09s eta: 9:07:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31288.0 tgs: 263 data_time: 0.65s time: 118.86s eta: 9:04:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.256 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31775.0 tgs: 265 data_time: 0.64s time: 119.67s eta: 9:06:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31506.0 tgs: 263 data_time: 0.70s time: 119.43s eta: 9:03:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.157 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32383.0 tgs: 271 data_time: 0.79s time: 119.43s eta: 9:01:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.129 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31836.0 tgs: 268 data_time: 0.67s time: 118.74s eta: 8:56:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.206 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31809.0 tgs: 267 data_time: 0.80s time: 118.87s eta: 8:54:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31804.0 tgs: 266 data_time: 0.73s time: 119.33s eta: 8:55:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.161 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31819.0 tgs: 267 data_time: 0.91s time: 119.15s eta: 8:52:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.197 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31395.0 tgs: 263 data_time: 0.96s time: 119.27s eta: 8:50:44 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.186 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30913.0 tgs: 256 data_time: 0.64s time: 120.47s eta: 8:54:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.121 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30896.0 tgs: 259 data_time: 0.47s time: 119.23s eta: 8:46:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31426.0 tgs: 263 data_time: 0.62s time: 119.07s eta: 8:43:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30607.0 tgs: 253 data_time: 0.89s time: 120.83s eta: 8:49:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31666.0 tgs: 263 data_time: 1.08s time: 120.19s eta: 8:44:50 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31422.0 tgs: 264 data_time: 0.78s time: 118.82s eta: 8:36:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.112 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32065.0 tgs: 268 data_time: 0.76s time: 119.35s eta: 8:37:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29051.0 tgs: 243 data_time: 0.72s time: 119.33s eta: 8:35:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.173 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31478.0 tgs: 264 data_time: 0.88s time: 119.05s eta: 8:31:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31886.0 tgs: 267 data_time: 0.72s time: 119.16s eta: 8:30:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 266 data_time: 0.95s time: 119.04s eta: 8:27:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31903.0 tgs: 267 data_time: 0.88s time: 119.12s eta: 8:26:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.166 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31462.0 tgs: 263 data_time: 0.50s time: 119.25s eta: 8:24:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.181 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31230.0 tgs: 261 data_time: 0.88s time: 119.22s eta: 8:22:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31975.0 tgs: 267 data_time: 1.04s time: 119.38s eta: 8:21:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31603.0 tgs: 263 data_time: 0.79s time: 119.78s eta: 8:21:03 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31330.0 tgs: 262 data_time: 0.86s time: 119.20s eta: 8:16:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.184 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32314.0 tgs: 271 data_time: 0.57s time: 119.02s eta: 8:13:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.139 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30649.0 tgs: 255 data_time: 0.68s time: 119.90s eta: 8:15:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.170 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31730.0 tgs: 265 data_time: 0.76s time: 119.55s eta: 8:12:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.183 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32136.0 tgs: 268 data_time: 0.71s time: 119.59s eta: 8:10:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.169 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32233.0 tgs: 270 data_time: 0.72s time: 119.00s eta: 8:05:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.195 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32420.0 tgs: 270 data_time: 0.87s time: 119.84s eta: 8:07:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 267 data_time: 0.70s time: 119.71s eta: 8:04:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30701.0 tgs: 257 data_time: 0.62s time: 119.40s eta: 8:01:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30999.0 tgs: 259 data_time: 0.56s time: 119.28s eta: 7:59:07 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.212 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30910.0 tgs: 259 data_time: 0.74s time: 119.19s eta: 7:56:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31626.0 tgs: 265 data_time: 0.85s time: 119.19s eta: 7:54:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31205.0 tgs: 262 data_time: 0.69s time: 118.93s eta: 7:51:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.111 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31301.0 tgs: 263 data_time: 0.50s time: 118.97s eta: 7:49:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.110 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30878.0 tgs: 258 data_time: 0.76s time: 119.41s eta: 7:49:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.214 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32167.0 tgs: 270 data_time: 0.67s time: 119.03s eta: 7:46:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.163 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30569.0 tgs: 256 data_time: 0.80s time: 119.10s eta: 7:44:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.216 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31276.0 tgs: 260 data_time: 0.82s time: 120.02s eta: 7:46:04 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31148.0 tgs: 259 data_time: 0.67s time: 119.82s eta: 7:43:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31547.0 tgs: 264 data_time: 0.61s time: 119.16s eta: 7:38:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32303.0 tgs: 271 data_time: 0.77s time: 118.97s eta: 7:36:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.141 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 264 data_time: 0.58s time: 119.32s eta: 7:35:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31801.0 tgs: 266 data_time: 0.94s time: 119.38s eta: 7:33:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31606.0 tgs: 265 data_time: 0.72s time: 119.06s eta: 7:30:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.138 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 270 data_time: 0.54s time: 118.99s eta: 7:28:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32225.0 tgs: 270 data_time: 0.78s time: 119.24s eta: 7:27:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31317.0 tgs: 263 data_time: 0.77s time: 118.99s eta: 7:24:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.189 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32487.0 tgs: 273 data_time: 0.94s time: 118.93s eta: 7:22:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31473.0 tgs: 264 data_time: 0.58s time: 119.08s eta: 7:20:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.217 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31792.0 tgs: 267 data_time: 1.07s time: 118.99s eta: 7:18:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 0.81s time: 119.02s eta: 7:16:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.180 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31564.0 tgs: 264 data_time: 0.90s time: 119.26s eta: 7:15:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.141 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31494.0 tgs: 263 data_time: 1.02s time: 119.36s eta: 7:13:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32085.0 tgs: 269 data_time: 0.87s time: 118.94s eta: 7:10:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.186 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31085.0 tgs: 260 data_time: 0.75s time: 119.20s eta: 7:09:07 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30739.0 tgs: 257 data_time: 0.49s time: 119.58s eta: 7:08:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30876.0 tgs: 258 data_time: 0.91s time: 119.66s eta: 7:06:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.175 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32187.0 tgs: 270 data_time: 0.75s time: 118.95s eta: 7:02:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.198 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31679.0 tgs: 265 data_time: 0.59s time: 119.30s eta: 7:01:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.177 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31599.0 tgs: 264 data_time: 0.68s time: 119.51s eta: 7:00:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30788.0 tgs: 257 data_time: 0.69s time: 119.43s eta: 6:57:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31658.0 tgs: 266 data_time: 0.78s time: 118.90s eta: 6:54:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31288.0 tgs: 262 data_time: 0.89s time: 119.34s eta: 6:53:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 265 data_time: 0.65s time: 118.93s eta: 6:50:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 29411.0 tgs: 246 data_time: 0.76s time: 119.27s eta: 6:49:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31947.0 tgs: 268 data_time: 0.82s time: 119.02s eta: 6:46:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.127 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32296.0 tgs: 270 data_time: 0.59s time: 119.23s eta: 6:45:23 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.145 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32317.0 tgs: 270 data_time: 0.92s time: 119.42s eta: 6:44:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31893.0 tgs: 267 data_time: 0.59s time: 119.01s eta: 6:40:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.100 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31933.0 tgs: 268 data_time: 0.73s time: 119.01s eta: 6:38:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.201 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32279.0 tgs: 270 data_time: 0.63s time: 119.45s eta: 6:38:10 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32098.0 tgs: 270 data_time: 0.80s time: 118.80s eta: 6:34:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.168 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32307.0 tgs: 270 data_time: 0.83s time: 119.44s eta: 6:34:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.158 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31963.0 tgs: 267 data_time: 0.78s time: 119.57s eta: 6:32:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32047.0 tgs: 268 data_time: 0.58s time: 119.38s eta: 6:29:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32344.0 tgs: 270 data_time: 0.75s time: 119.73s eta: 6:29:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.122 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30861.0 tgs: 258 data_time: 0.62s time: 119.36s eta: 6:25:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.189 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30730.0 tgs: 258 data_time: 0.73s time: 118.82s eta: 6:22:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31433.0 tgs: 264 data_time: 0.73s time: 119.02s eta: 6:20:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31546.0 tgs: 264 data_time: 0.95s time: 119.40s eta: 6:20:04 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.208 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30937.0 tgs: 259 data_time: 1.04s time: 119.34s eta: 6:17:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31473.0 tgs: 264 data_time: 0.80s time: 119.05s eta: 6:15:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.130 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30918.0 tgs: 258 data_time: 0.67s time: 119.41s eta: 6:14:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31563.0 tgs: 264 data_time: 0.80s time: 119.39s eta: 6:12:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32363.0 tgs: 271 data_time: 0.76s time: 119.28s eta: 6:09:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30922.0 tgs: 259 data_time: 0.69s time: 119.10s eta: 6:07:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.207 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31702.0 tgs: 259 data_time: 1.02s time: 121.96s eta: 6:14:00 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29878.0 tgs: 250 data_time: 0.47s time: 119.38s eta: 6:04:07 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.139 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32043.0 tgs: 267 data_time: 0.52s time: 119.57s eta: 6:02:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.146 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31252.0 tgs: 262 data_time: 0.65s time: 119.25s eta: 5:59:44 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.150 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31754.0 tgs: 266 data_time: 0.84s time: 119.26s eta: 5:57:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.145 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31301.0 tgs: 263 data_time: 0.82s time: 118.91s eta: 5:54:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32143.0 tgs: 270 data_time: 0.52s time: 119.02s eta: 5:53:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30499.0 tgs: 256 data_time: 0.74s time: 119.07s eta: 5:51:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31429.0 tgs: 263 data_time: 0.61s time: 119.29s eta: 5:49:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.165 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31346.0 tgs: 263 data_time: 0.70s time: 118.95s eta: 5:46:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:00:05][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.159 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31986.0 tgs: 268 data_time: 0.60s time: 119.28s eta: 5:45:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31759.0 tgs: 266 data_time: 0.83s time: 119.38s eta: 5:44:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.128 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32118.0 tgs: 269 data_time: 0.76s time: 119.05s eta: 5:41:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32177.0 tgs: 269 data_time: 0.55s time: 119.49s eta: 5:40:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.127 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31619.0 tgs: 264 data_time: 0.54s time: 119.39s eta: 5:38:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31770.0 tgs: 265 data_time: 0.90s time: 119.56s eta: 5:36:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.200 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32254.0 tgs: 271 data_time: 0.74s time: 118.98s eta: 5:33:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.164 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32171.0 tgs: 270 data_time: 0.55s time: 118.95s eta: 5:31:05 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.120 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31909.0 tgs: 267 data_time: 0.66s time: 119.24s eta: 5:29:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.176 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32070.0 tgs: 268 data_time: 0.99s time: 119.40s eta: 5:28:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.162 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30860.0 tgs: 258 data_time: 0.85s time: 119.46s eta: 5:26:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.187 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31447.0 tgs: 264 data_time: 0.75s time: 118.95s eta: 5:23:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.177 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32029.0 tgs: 267 data_time: 0.54s time: 119.63s eta: 5:22:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.171 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32272.0 tgs: 270 data_time: 0.80s time: 119.35s eta: 5:20:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.109 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31746.0 tgs: 266 data_time: 0.76s time: 119.12s eta: 5:17:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31223.0 tgs: 261 data_time: 0.52s time: 119.42s eta: 5:16:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.110 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30443.0 tgs: 254 data_time: 0.68s time: 119.58s eta: 5:14:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32008.0 tgs: 268 data_time: 0.76s time: 119.00s eta: 5:11:23 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31898.0 tgs: 268 data_time: 0.82s time: 119.00s eta: 5:09:23 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.124 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30983.0 tgs: 259 data_time: 0.63s time: 119.22s eta: 5:07:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32114.0 tgs: 268 data_time: 0.57s time: 119.68s eta: 5:07:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.177 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31267.0 tgs: 262 data_time: 0.91s time: 119.11s eta: 5:03:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32116.0 tgs: 269 data_time: 0.55s time: 119.07s eta: 5:01:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31611.0 tgs: 260 data_time: 0.76s time: 121.52s eta: 5:05:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32120.0 tgs: 268 data_time: 0.82s time: 119.57s eta: 4:58:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.183 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30857.0 tgs: 258 data_time: 0.58s time: 119.50s eta: 4:56:44 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31122.0 tgs: 259 data_time: 0.64s time: 119.73s eta: 4:55:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31989.0 tgs: 267 data_time: 1.00s time: 119.52s eta: 4:52:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31299.0 tgs: 262 data_time: 0.81s time: 119.41s eta: 4:50:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31067.0 tgs: 260 data_time: 0.77s time: 119.20s eta: 4:48:04 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31110.0 tgs: 261 data_time: 0.75s time: 118.89s eta: 4:45:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30959.0 tgs: 259 data_time: 0.79s time: 119.11s eta: 4:43:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32155.0 tgs: 268 data_time: 0.80s time: 119.98s eta: 4:43:57 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30478.0 tgs: 255 data_time: 0.63s time: 119.15s eta: 4:39:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31802.0 tgs: 267 data_time: 0.84s time: 119.01s eta: 4:37:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.129 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31279.0 tgs: 262 data_time: 0.72s time: 119.05s eta: 4:35:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30979.0 tgs: 259 data_time: 0.76s time: 119.37s eta: 4:34:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.151 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31410.0 tgs: 263 data_time: 0.68s time: 119.41s eta: 4:32:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.158 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30876.0 tgs: 259 data_time: 0.68s time: 118.94s eta: 4:29:35 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31330.0 tgs: 261 data_time: 0.81s time: 119.60s eta: 4:29:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.195 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31223.0 tgs: 261 data_time: 0.74s time: 119.24s eta: 4:26:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31372.0 tgs: 263 data_time: 0.71s time: 119.06s eta: 4:23:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.154 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32058.0 tgs: 269 data_time: 0.54s time: 118.85s eta: 4:21:27 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31508.0 tgs: 265 data_time: 0.73s time: 118.76s eta: 4:19:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 265 data_time: 0.69s time: 119.43s eta: 4:18:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.146 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31940.0 tgs: 267 data_time: 0.70s time: 119.40s eta: 4:16:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31806.0 tgs: 266 data_time: 0.79s time: 119.35s eta: 4:14:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32186.0 tgs: 270 data_time: 0.75s time: 119.05s eta: 4:11:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31414.0 tgs: 263 data_time: 0.96s time: 119.12s eta: 4:10:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.127 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31737.0 tgs: 266 data_time: 0.63s time: 119.13s eta: 4:08:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.174 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31335.0 tgs: 263 data_time: 0.79s time: 118.96s eta: 4:05:50 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.223 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31978.0 tgs: 268 data_time: 0.58s time: 119.15s eta: 4:04:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32231.0 tgs: 269 data_time: 0.65s time: 119.49s eta: 4:02:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31339.0 tgs: 262 data_time: 0.85s time: 119.24s eta: 4:00:28 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32120.0 tgs: 269 data_time: 0.71s time: 119.19s eta: 3:58:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.135 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30076.0 tgs: 251 data_time: 0.48s time: 119.36s eta: 3:56:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.180 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31943.0 tgs: 267 data_time: 0.52s time: 119.37s eta: 3:54:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31315.0 tgs: 261 data_time: 0.74s time: 119.95s eta: 3:53:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.128 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31944.0 tgs: 267 data_time: 0.52s time: 119.32s eta: 3:50:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32224.0 tgs: 269 data_time: 0.91s time: 119.63s eta: 3:49:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31645.0 tgs: 266 data_time: 0.63s time: 118.86s eta: 3:45:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31541.0 tgs: 263 data_time: 0.79s time: 119.75s eta: 3:45:32 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 265 data_time: 0.72s time: 119.33s eta: 3:42:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32351.0 tgs: 271 data_time: 0.67s time: 119.26s eta: 3:40:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31994.0 tgs: 268 data_time: 0.57s time: 118.99s eta: 3:38:08 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.178 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31662.0 tgs: 265 data_time: 0.84s time: 119.44s eta: 3:36:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32004.0 tgs: 268 data_time: 0.80s time: 119.03s eta: 3:34:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 267 data_time: 0.78s time: 118.99s eta: 3:32:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32211.0 tgs: 270 data_time: 0.78s time: 119.21s eta: 3:30:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 265 data_time: 0.62s time: 120.11s eta: 3:30:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.248 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31343.0 tgs: 263 data_time: 1.02s time: 119.13s eta: 3:26:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31637.0 tgs: 265 data_time: 0.64s time: 119.32s eta: 3:24:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.174 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30575.0 tgs: 256 data_time: 0.60s time: 119.08s eta: 3:22:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32117.0 tgs: 268 data_time: 0.70s time: 119.42s eta: 3:21:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32390.0 tgs: 271 data_time: 0.91s time: 119.29s eta: 3:18:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.210 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31448.0 tgs: 263 data_time: 0.73s time: 119.22s eta: 3:16:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30578.0 tgs: 256 data_time: 0.85s time: 119.16s eta: 3:14:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.099 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30317.0 tgs: 254 data_time: 0.58s time: 119.02s eta: 3:12:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.171 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32080.0 tgs: 269 data_time: 0.90s time: 119.18s eta: 3:10:41 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.205 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32335.0 tgs: 267 data_time: 0.70s time: 121.09s eta: 3:11:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31456.0 tgs: 263 data_time: 1.05s time: 119.51s eta: 3:07:13 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32208.0 tgs: 270 data_time: 0.79s time: 119.26s eta: 3:04:51 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.113 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30833.0 tgs: 259 data_time: 0.52s time: 118.98s eta: 3:02:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32482.0 tgs: 272 data_time: 0.93s time: 119.22s eta: 3:00:48 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.238 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31234.0 tgs: 261 data_time: 0.61s time: 119.46s eta: 2:59:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.145 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30873.0 tgs: 259 data_time: 0.97s time: 118.98s eta: 2:56:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.140 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31321.0 tgs: 262 data_time: 0.72s time: 119.35s eta: 2:55:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.120 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30164.0 tgs: 252 data_time: 0.79s time: 119.42s eta: 2:53:09 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32131.0 tgs: 269 data_time: 0.51s time: 119.34s eta: 2:51:03 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.139 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31908.0 tgs: 262 data_time: 0.60s time: 121.46s eta: 2:52:04 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31358.0 tgs: 262 data_time: 0.77s time: 119.23s eta: 2:46:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31428.0 tgs: 262 data_time: 0.78s time: 119.60s eta: 2:45:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.131 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31506.0 tgs: 264 data_time: 0.65s time: 118.92s eta: 2:42:31 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.274 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31983.0 tgs: 268 data_time: 0.79s time: 119.18s eta: 2:40:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.165 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32377.0 tgs: 272 data_time: 0.58s time: 118.97s eta: 2:38:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32134.0 tgs: 269 data_time: 0.84s time: 119.22s eta: 2:36:58 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.169 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 267 data_time: 0.82s time: 119.11s eta: 2:34:50 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:13:03][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.149 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 265 data_time: 0.65s time: 119.42s eta: 2:33:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.137 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32478.0 tgs: 270 data_time: 0.71s time: 120.22s eta: 2:32:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.209 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32539.0 tgs: 272 data_time: 1.03s time: 119.28s eta: 2:29:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.163 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31644.0 tgs: 266 data_time: 0.73s time: 118.88s eta: 2:26:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31964.0 tgs: 267 data_time: 1.10s time: 119.52s eta: 2:25:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.222 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30803.0 tgs: 258 data_time: 0.79s time: 119.07s eta: 2:22:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.143 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31056.0 tgs: 260 data_time: 0.72s time: 119.14s eta: 2:20:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.202 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32306.0 tgs: 271 data_time: 1.03s time: 119.04s eta: 2:18:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31789.0 tgs: 265 data_time: 0.99s time: 119.84s eta: 2:17:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31162.0 tgs: 261 data_time: 0.50s time: 119.34s eta: 2:15:15 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.158 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32037.0 tgs: 268 data_time: 0.54s time: 119.38s eta: 2:13:18 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31794.0 tgs: 265 data_time: 0.63s time: 119.68s eta: 2:11:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.122 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31899.0 tgs: 266 data_time: 0.68s time: 119.49s eta: 2:09:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.141 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31755.0 tgs: 266 data_time: 0.64s time: 118.95s eta: 2:06:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31617.0 tgs: 264 data_time: 0.67s time: 119.37s eta: 2:05:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31404.0 tgs: 263 data_time: 0.71s time: 119.27s eta: 2:03:14 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.109 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31039.0 tgs: 260 data_time: 0.64s time: 119.22s eta: 2:01:12 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31887.0 tgs: 266 data_time: 0.74s time: 119.51s eta: 1:59:30 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31667.0 tgs: 265 data_time: 0.78s time: 119.42s eta: 1:57:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.209 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31706.0 tgs: 265 data_time: 0.84s time: 119.53s eta: 1:55:32 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32145.0 tgs: 269 data_time: 0.77s time: 119.30s eta: 1:53:20 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31837.0 tgs: 266 data_time: 1.00s time: 119.25s eta: 1:51:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31735.0 tgs: 266 data_time: 0.82s time: 118.93s eta: 1:49:01 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31939.0 tgs: 266 data_time: 0.48s time: 119.64s eta: 1:47:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.195 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31480.0 tgs: 262 data_time: 0.91s time: 119.72s eta: 1:45:45 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.178 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32280.0 tgs: 270 data_time: 0.67s time: 119.35s eta: 1:43:26 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.126 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31796.0 tgs: 266 data_time: 0.64s time: 119.31s eta: 1:41:24 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31968.0 tgs: 268 data_time: 0.87s time: 119.28s eta: 1:39:23 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32116.0 tgs: 270 data_time: 0.85s time: 118.91s eta: 1:37:06 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.188 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31277.0 tgs: 262 data_time: 0.81s time: 119.15s eta: 1:35:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31267.0 tgs: 262 data_time: 0.65s time: 119.07s eta: 1:33:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31830.0 tgs: 266 data_time: 0.60s time: 119.24s eta: 1:31:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32286.0 tgs: 269 data_time: 0.63s time: 119.72s eta: 1:29:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31342.0 tgs: 263 data_time: 0.79s time: 119.02s eta: 1:27:16 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32167.0 tgs: 270 data_time: 0.81s time: 118.87s eta: 1:25:11 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29522.0 tgs: 247 data_time: 0.55s time: 119.47s eta: 1:23:37 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30894.0 tgs: 258 data_time: 0.57s time: 119.48s eta: 1:21:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.116 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31530.0 tgs: 265 data_time: 0.62s time: 118.94s eta: 1:19:17 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31873.0 tgs: 267 data_time: 0.80s time: 118.93s eta: 1:17:18 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31926.0 tgs: 261 data_time: 0.84s time: 122.16s eta: 1:17:21 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.164 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32056.0 tgs: 268 data_time: 0.58s time: 119.37s eta: 1:13:36 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.171 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31302.0 tgs: 262 data_time: 0.60s time: 119.05s eta: 1:11:25 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31700.0 tgs: 266 data_time: 0.77s time: 118.92s eta: 1:09:22 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31840.0 tgs: 267 data_time: 0.77s time: 119.10s eta: 1:07:29 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31550.0 tgs: 264 data_time: 0.79s time: 119.41s eta: 1:05:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.155 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30968.0 tgs: 259 data_time: 0.86s time: 119.35s eta: 1:03:39 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31765.0 tgs: 266 data_time: 0.92s time: 119.17s eta: 1:01:34 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31224.0 tgs: 261 data_time: 0.64s time: 119.27s eta: 0:59:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31042.0 tgs: 259 data_time: 0.58s time: 119.44s eta: 0:57:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.214 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30602.0 tgs: 256 data_time: 0.87s time: 119.39s eta: 0:55:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.197 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32154.0 tgs: 269 data_time: 0.92s time: 119.48s eta: 0:53:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31928.0 tgs: 267 data_time: 0.75s time: 119.36s eta: 0:51:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32167.0 tgs: 269 data_time: 0.92s time: 119.47s eta: 0:49:46 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32088.0 tgs: 269 data_time: 0.79s time: 118.89s eta: 0:47:33 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.118 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31677.0 tgs: 265 data_time: 0.55s time: 119.28s eta: 0:45:43 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.201 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32432.0 tgs: 272 data_time: 0.71s time: 119.22s eta: 0:43:42 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32086.0 tgs: 269 data_time: 0.68s time: 118.99s eta: 0:41:38 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32088.0 tgs: 269 data_time: 0.63s time: 119.04s eta: 0:39:40 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31604.0 tgs: 264 data_time: 0.84s time: 119.66s eta: 0:37:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31857.0 tgs: 267 data_time: 0.51s time: 119.28s eta: 0:35:47 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32555.0 tgs: 271 data_time: 0.74s time: 119.74s eta: 0:33:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31713.0 tgs: 261 data_time: 0.90s time: 121.22s eta: 0:32:19 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.245 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31663.0 tgs: 265 data_time: 0.69s time: 119.38s eta: 0:29:50 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.130 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31575.0 tgs: 264 data_time: 0.76s time: 119.32s eta: 0:27:50 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31044.0 tgs: 259 data_time: 0.74s time: 119.63s eta: 0:25:55 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.185 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32100.0 tgs: 269 data_time: 1.01s time: 119.09s eta: 0:23:49 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.199 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31500.0 tgs: 263 data_time: 0.88s time: 119.38s eta: 0:21:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.160 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31519.0 tgs: 264 data_time: 0.61s time: 119.26s eta: 0:19:52 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31963.0 tgs: 268 data_time: 1.13s time: 119.24s eta: 0:17:53 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.147 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32063.0 tgs: 267 data_time: 0.71s time: 119.91s eta: 0:15:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32312.0 tgs: 271 data_time: 0.66s time: 119.22s eta: 0:13:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31645.0 tgs: 265 data_time: 0.76s time: 119.09s eta: 0:11:54 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30933.0 tgs: 259 data_time: 0.79s time: 119.32s eta: 0:09:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31953.0 tgs: 268 data_time: 0.72s time: 119.17s eta: 0:07:56 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.124 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32212.0 tgs: 269 data_time: 0.80s time: 119.69s eta: 0:05:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30966.0 tgs: 255 data_time: 0.55s time: 121.42s eta: 0:04:02 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31936.0 tgs: 267 data_time: 0.97s time: 119.21s eta: 0:01:59 +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 7][DP 1][SP 3][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22 diff --git a/20250121165312/rank9.log b/20250121165312/rank9.log new file mode 100644 index 0000000000000000000000000000000000000000..2a1260454d96ecb3074893db96c756403cfbdd5c --- /dev/null +++ b/20250121165312/rank9.log @@ -0,0 +1,851 @@ +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:16][INFO] Namespace(llm='/mnt/hwfile/opendatalab/panzhuoshi/huggingface/hub/models--Qwen--Qwen2.5-72B-Instruct/snapshots/d3d951150c1e5848237cd6a7ad11df4836aee842', tokenizer=None, chat_template='qwen2', use_lora=False, lora_targets=None, lora_r=64, lora_alpha=16, lora_dropout=0.1, lora_bias='none', dtype='auto', selective_recompute=1.0, shard_strategy='full', cpu_offload=False, sp_size=4, datasets=['/mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2'], dset_file_types=dict_keys(['.jsonl', '.json']), dset_sources=['local'], dset_formats=['openai'], dset_sample_ratios=[1.0], dset_cache_dir='/mnt/petrelfs/caimengzhang/cached_data/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2', dset_pack_level='soft', global_pack=True, max_length=32768, num_workers=1, file_pattern=None, group_by_length=True, mirco_batch_size=1, global_batch_size=64, lr=2e-05, lr_min=6e-06, wd=0.01, max_grad_norm=1, epochs=1, warmup_ratio=0.025, config=None, work_dir='checkpoints/qwen25_72b_inst_base50v2-new-zh-en30w-combinev9-mls-chatbeta2/20250121165312', feishu_webhook=None, gc_interval=100, checkpoint_interval=200000.0, checkpoint_max_keep=1, checkpoint_drop_optimizer=True, log_interval=1, resume=False, seed=0, debug=False) +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:16][INFO] Found 8 files in /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:18][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_1.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:19][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_2.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:21][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_3.jsonl has 5 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:22][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_4.jsonl has 6 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:24][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_5.jsonl has 2 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:26][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_6.jsonl has 4 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:27][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_7.jsonl has 3 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:29][WARNING] /mnt/petrelfs/caimengzhang/data/20b_data//base50v2-new-zh-en30w-combinev9-mls-chatbeta2/data_part_8.jsonl has 1 prompt length>32768, discard. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 16:53:38][INFO] [Dataset & Dataloader] Cost 22.05s +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch (Qwen2ForCausalLM) forward to `qwen2_casual_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.0.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.1.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.2.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.3.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.4.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.5.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.6.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.7.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.8.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.9.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.10.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.11.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.12.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.13.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.14.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.15.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.16.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.17.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.18.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.19.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.20.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.21.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.22.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.23.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.24.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.25.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.26.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.27.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.28.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.29.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.30.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.31.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.32.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.33.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.34.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.35.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.36.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.37.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.38.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.39.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.40.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.41.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.42.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.43.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.44.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.45.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.46.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.47.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.48.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.49.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.50.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.51.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.52.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.53.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.54.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.55.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.56.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.57.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.58.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.59.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.60.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.61.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.62.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.63.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.64.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.65.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.66.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.67.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.68.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.69.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.70.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.71.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.72.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.73.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.74.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.75.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.76.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.77.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.78.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.self_attn(Qwen2FlashAttention2) forward to `qwen2_attn_flash_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.input_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.layers.79.post_attention_layernorm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:00:51][DEBUG] Dispatch model.norm(Qwen2RMSNorm) forward to `rms_norm_forward` +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:01:54][SUCCESS] [Parallelize LLM] Elapsed time 62.92 seconds, peak gpu memory 13.4G +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:01:55][INFO] [Train] Begin Train Loop. The current GPU memory is 4.2GB +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:04:44][INFO] [Train] (Epoch 1) Step 1/593 lr: 0.000001 loss: 0.279 loss(reduced): 0.273 grad_norm: 2.54 if_nan_skip: 0 max_memory: 33.1GB text_tokens: 31696.0 tgs: 209 data_time: 2.09s time: 151.28s eta: 1 day, 0:55:09 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:06:44][INFO] [Train] (Epoch 1) Step 2/593 lr: 0.000003 loss: 0.282 loss(reduced): 0.257 grad_norm: 2.30 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31732.0 tgs: 264 data_time: 1.08s time: 120.12s eta: 19:45:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:08:44][INFO] [Train] (Epoch 1) Step 3/593 lr: 0.000004 loss: 0.181 loss(reduced): 0.236 grad_norm: 1.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32529.0 tgs: 271 data_time: 0.93s time: 119.69s eta: 19:38:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:10:43][INFO] [Train] (Epoch 1) Step 4/593 lr: 0.000006 loss: 0.208 loss(reduced): 0.212 grad_norm: 0.33 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31543.0 tgs: 263 data_time: 0.93s time: 119.65s eta: 19:36:32 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:12:43][INFO] [Train] (Epoch 1) Step 5/593 lr: 0.000007 loss: 0.266 loss(reduced): 0.214 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31681.0 tgs: 265 data_time: 1.00s time: 119.30s eta: 19:31:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:14:42][INFO] [Train] (Epoch 1) Step 6/593 lr: 0.000009 loss: 0.197 loss(reduced): 0.221 grad_norm: 0.44 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30335.0 tgs: 253 data_time: 0.94s time: 119.53s eta: 19:31:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:16:45][INFO] [Train] (Epoch 1) Step 7/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.205 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30779.0 tgs: 250 data_time: 0.74s time: 122.96s eta: 20:03:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:18:45][INFO] [Train] (Epoch 1) Step 8/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.232 grad_norm: 0.34 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31645.0 tgs: 263 data_time: 0.97s time: 119.89s eta: 19:30:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:20:46][INFO] [Train] (Epoch 1) Step 9/593 lr: 0.000013 loss: 0.205 loss(reduced): 0.215 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32125.0 tgs: 266 data_time: 0.71s time: 120.66s eta: 19:36:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:22:46][INFO] [Train] (Epoch 1) Step 10/593 lr: 0.000014 loss: 0.249 loss(reduced): 0.210 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31629.0 tgs: 263 data_time: 1.05s time: 119.91s eta: 19:27:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:24:46][INFO] [Train] (Epoch 1) Step 11/593 lr: 0.000016 loss: 0.236 loss(reduced): 0.204 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31480.0 tgs: 261 data_time: 0.97s time: 120.46s eta: 19:30:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:26:45][INFO] [Train] (Epoch 1) Step 12/593 lr: 0.000017 loss: 0.271 loss(reduced): 0.201 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32172.0 tgs: 269 data_time: 0.80s time: 119.34s eta: 19:17:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:28:45][INFO] [Train] (Epoch 1) Step 13/593 lr: 0.000019 loss: 0.219 loss(reduced): 0.206 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31315.0 tgs: 261 data_time: 1.28s time: 119.68s eta: 19:18:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:30:44][INFO] [Train] (Epoch 1) Step 14/593 lr: 0.000020 loss: 0.243 loss(reduced): 0.206 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32182.0 tgs: 269 data_time: 0.68s time: 119.20s eta: 19:12:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:32:43][INFO] [Train] (Epoch 1) Step 15/593 lr: 0.000020 loss: 0.259 loss(reduced): 0.198 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31862.0 tgs: 267 data_time: 0.80s time: 119.22s eta: 19:10:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:34:42][INFO] [Train] (Epoch 1) Step 16/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.191 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31743.0 tgs: 266 data_time: 0.93s time: 118.93s eta: 19:05:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:36:42][INFO] [Train] (Epoch 1) Step 17/593 lr: 0.000020 loss: 0.183 loss(reduced): 0.205 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32304.0 tgs: 271 data_time: 1.00s time: 119.14s eta: 19:05:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:38:41][INFO] [Train] (Epoch 1) Step 18/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.192 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31895.0 tgs: 267 data_time: 0.99s time: 119.18s eta: 19:04:05 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:40:40][INFO] [Train] (Epoch 1) Step 19/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.192 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30695.0 tgs: 257 data_time: 0.89s time: 119.40s eta: 19:04:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:42:39][INFO] [Train] (Epoch 1) Step 20/593 lr: 0.000020 loss: 0.151 loss(reduced): 0.198 grad_norm: 0.15 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30721.0 tgs: 257 data_time: 1.07s time: 119.32s eta: 19:01:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:44:40][INFO] [Train] (Epoch 1) Step 21/593 lr: 0.000020 loss: 0.144 loss(reduced): 0.189 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 29162.0 tgs: 242 data_time: 0.61s time: 120.31s eta: 19:08:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:46:43][INFO] [Train] (Epoch 1) Step 22/593 lr: 0.000020 loss: 0.234 loss(reduced): 0.198 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31808.0 tgs: 258 data_time: 0.87s time: 122.92s eta: 19:31:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:48:42][INFO] [Train] (Epoch 1) Step 23/593 lr: 0.000020 loss: 0.184 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32435.0 tgs: 271 data_time: 1.08s time: 119.48s eta: 18:57:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:50:41][INFO] [Train] (Epoch 1) Step 24/593 lr: 0.000020 loss: 0.174 loss(reduced): 0.190 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32461.0 tgs: 272 data_time: 0.86s time: 119.16s eta: 18:52:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:52:41][INFO] [Train] (Epoch 1) Step 25/593 lr: 0.000020 loss: 0.189 loss(reduced): 0.186 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31514.0 tgs: 264 data_time: 1.08s time: 119.24s eta: 18:50:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:54:40][INFO] [Train] (Epoch 1) Step 26/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.188 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30176.0 tgs: 252 data_time: 0.64s time: 119.39s eta: 18:50:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:56:40][INFO] [Train] (Epoch 1) Step 27/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.195 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32145.0 tgs: 268 data_time: 1.10s time: 119.66s eta: 18:50:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 17:58:39][INFO] [Train] (Epoch 1) Step 28/593 lr: 0.000020 loss: 0.152 loss(reduced): 0.184 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32029.0 tgs: 268 data_time: 0.63s time: 119.38s eta: 18:46:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:00:39][INFO] [Train] (Epoch 1) Step 29/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.183 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31383.0 tgs: 261 data_time: 0.66s time: 119.79s eta: 18:47:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:02:48][INFO] [Train] (Epoch 1) Step 30/593 lr: 0.000020 loss: 0.211 loss(reduced): 0.187 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30895.0 tgs: 238 data_time: 0.90s time: 129.33s eta: 20:15:39 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:04:48][INFO] [Train] (Epoch 1) Step 31/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.188 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 264 data_time: 1.09s time: 119.56s eta: 18:41:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:06:47][INFO] [Train] (Epoch 1) Step 32/593 lr: 0.000020 loss: 0.145 loss(reduced): 0.183 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32406.0 tgs: 271 data_time: 0.87s time: 119.39s eta: 18:38:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:08:48][INFO] [Train] (Epoch 1) Step 33/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.182 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32322.0 tgs: 267 data_time: 0.94s time: 120.65s eta: 18:48:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:10:47][INFO] [Train] (Epoch 1) Step 34/593 lr: 0.000020 loss: 0.168 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 30969.0 tgs: 259 data_time: 0.86s time: 119.40s eta: 18:34:22 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:12:46][INFO] [Train] (Epoch 1) Step 35/593 lr: 0.000020 loss: 0.169 loss(reduced): 0.182 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31334.0 tgs: 262 data_time: 0.95s time: 119.16s eta: 18:30:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:14:45][INFO] [Train] (Epoch 1) Step 36/593 lr: 0.000020 loss: 0.125 loss(reduced): 0.182 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 268 data_time: 0.91s time: 119.07s eta: 18:27:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:16:45][INFO] [Train] (Epoch 1) Step 37/593 lr: 0.000020 loss: 0.158 loss(reduced): 0.177 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32127.0 tgs: 269 data_time: 0.97s time: 119.13s eta: 18:25:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:18:45][INFO] [Train] (Epoch 1) Step 38/593 lr: 0.000020 loss: 0.140 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31858.0 tgs: 264 data_time: 0.92s time: 120.39s eta: 18:35:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:20:46][INFO] [Train] (Epoch 1) Step 39/593 lr: 0.000020 loss: 0.179 loss(reduced): 0.176 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31386.0 tgs: 259 data_time: 0.94s time: 121.17s eta: 18:40:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:22:47][INFO] [Train] (Epoch 1) Step 40/593 lr: 0.000020 loss: 0.161 loss(reduced): 0.176 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32250.0 tgs: 267 data_time: 0.75s time: 120.66s eta: 18:34:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:24:47][INFO] [Train] (Epoch 1) Step 41/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.180 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32507.0 tgs: 271 data_time: 1.15s time: 119.73s eta: 18:23:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:26:50][INFO] [Train] (Epoch 1) Step 42/593 lr: 0.000020 loss: 0.137 loss(reduced): 0.194 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31757.0 tgs: 256 data_time: 0.76s time: 123.87s eta: 18:59:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:28:50][INFO] [Train] (Epoch 1) Step 43/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.176 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 1.06s time: 119.56s eta: 18:17:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:30:49][INFO] [Train] (Epoch 1) Step 44/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31891.0 tgs: 267 data_time: 0.90s time: 119.36s eta: 18:14:09 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:32:49][INFO] [Train] (Epoch 1) Step 45/593 lr: 0.000020 loss: 0.193 loss(reduced): 0.175 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31410.0 tgs: 263 data_time: 0.75s time: 119.23s eta: 18:10:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:34:48][INFO] [Train] (Epoch 1) Step 46/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.173 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 265 data_time: 1.00s time: 119.60s eta: 18:12:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:36:48][INFO] [Train] (Epoch 1) Step 47/593 lr: 0.000020 loss: 0.165 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30226.0 tgs: 253 data_time: 1.00s time: 119.27s eta: 18:07:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:38:47][INFO] [Train] (Epoch 1) Step 48/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.171 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31914.0 tgs: 268 data_time: 0.94s time: 119.01s eta: 18:02:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:40:46][INFO] [Train] (Epoch 1) Step 49/593 lr: 0.000020 loss: 0.186 loss(reduced): 0.181 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31209.0 tgs: 262 data_time: 0.55s time: 119.10s eta: 18:01:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:42:45][INFO] [Train] (Epoch 1) Step 50/593 lr: 0.000020 loss: 0.209 loss(reduced): 0.184 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31712.0 tgs: 265 data_time: 0.77s time: 119.26s eta: 18:01:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:44:44][INFO] [Train] (Epoch 1) Step 51/593 lr: 0.000020 loss: 0.180 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31541.0 tgs: 264 data_time: 0.82s time: 119.33s eta: 17:59:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:46:44][INFO] [Train] (Epoch 1) Step 52/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.185 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32580.0 tgs: 271 data_time: 0.84s time: 120.02s eta: 18:04:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:48:43][INFO] [Train] (Epoch 1) Step 53/593 lr: 0.000020 loss: 0.155 loss(reduced): 0.173 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31793.0 tgs: 266 data_time: 0.63s time: 119.10s eta: 17:53:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:50:43][INFO] [Train] (Epoch 1) Step 54/593 lr: 0.000020 loss: 0.159 loss(reduced): 0.169 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31726.0 tgs: 265 data_time: 0.75s time: 119.40s eta: 17:54:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:52:42][INFO] [Train] (Epoch 1) Step 55/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.179 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32291.0 tgs: 271 data_time: 0.92s time: 119.00s eta: 17:48:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:54:41][INFO] [Train] (Epoch 1) Step 56/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31197.0 tgs: 262 data_time: 0.87s time: 119.02s eta: 17:47:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:56:46][INFO] [Train] (Epoch 1) Step 57/593 lr: 0.000020 loss: 0.134 loss(reduced): 0.176 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31650.0 tgs: 252 data_time: 1.02s time: 125.36s eta: 18:41:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 18:58:46][INFO] [Train] (Epoch 1) Step 58/593 lr: 0.000020 loss: 0.129 loss(reduced): 0.177 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31693.0 tgs: 263 data_time: 0.88s time: 120.10s eta: 17:52:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:00:46][INFO] [Train] (Epoch 1) Step 59/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.178 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30935.0 tgs: 259 data_time: 0.89s time: 119.37s eta: 17:44:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:02:45][INFO] [Train] (Epoch 1) Step 60/593 lr: 0.000020 loss: 0.190 loss(reduced): 0.175 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31373.0 tgs: 262 data_time: 0.81s time: 119.34s eta: 17:42:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:04:45][INFO] [Train] (Epoch 1) Step 61/593 lr: 0.000020 loss: 0.182 loss(reduced): 0.177 grad_norm: 0.17 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31466.0 tgs: 263 data_time: 0.73s time: 119.51s eta: 17:41:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:06:51][INFO] [Train] (Epoch 1) Step 62/593 lr: 0.000020 loss: 0.134 loss(reduced): 0.180 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31582.0 tgs: 250 data_time: 0.87s time: 126.05s eta: 18:37:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:08:53][INFO] [Train] (Epoch 1) Step 63/593 lr: 0.000020 loss: 0.197 loss(reduced): 0.170 grad_norm: 0.29 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31379.0 tgs: 255 data_time: 0.85s time: 122.63s eta: 18:05:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:10:54][INFO] [Train] (Epoch 1) Step 64/593 lr: 0.000020 loss: 0.167 loss(reduced): 0.180 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31834.0 tgs: 262 data_time: 0.57s time: 121.15s eta: 17:50:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:12:54][INFO] [Train] (Epoch 1) Step 65/593 lr: 0.000020 loss: 0.146 loss(reduced): 0.173 grad_norm: 0.27 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31863.0 tgs: 267 data_time: 0.67s time: 119.03s eta: 17:29:28 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:14:52][INFO] [Train] (Epoch 1) Step 66/593 lr: 0.000020 loss: 0.162 loss(reduced): 0.177 grad_norm: 0.35 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31055.0 tgs: 261 data_time: 0.72s time: 118.85s eta: 17:25:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:16:56][INFO] [Train] (Epoch 1) Step 67/593 lr: 0.000020 loss: 0.228 loss(reduced): 0.181 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31953.0 tgs: 257 data_time: 0.78s time: 124.00s eta: 18:09:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:19:01][INFO] [Train] (Epoch 1) Step 68/593 lr: 0.000020 loss: 0.138 loss(reduced): 0.173 grad_norm: 0.32 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32157.0 tgs: 258 data_time: 0.64s time: 124.57s eta: 18:12:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:21:00][INFO] [Train] (Epoch 1) Step 69/593 lr: 0.000020 loss: 0.297 loss(reduced): 0.178 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31301.0 tgs: 262 data_time: 0.83s time: 119.39s eta: 17:24:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:23:00][INFO] [Train] (Epoch 1) Step 70/593 lr: 0.000020 loss: 0.234 loss(reduced): 0.187 grad_norm: 0.24 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31423.0 tgs: 262 data_time: 0.50s time: 119.69s eta: 17:25:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:25:00][INFO] [Train] (Epoch 1) Step 71/593 lr: 0.000020 loss: 0.178 loss(reduced): 0.172 grad_norm: 0.25 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31141.0 tgs: 260 data_time: 0.58s time: 119.52s eta: 17:21:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:26:58][INFO] [Train] (Epoch 1) Step 72/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.174 grad_norm: 0.26 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31690.0 tgs: 266 data_time: 0.87s time: 118.86s eta: 17:14:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:28:58][INFO] [Train] (Epoch 1) Step 73/593 lr: 0.000020 loss: 0.173 loss(reduced): 0.174 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31998.0 tgs: 266 data_time: 1.15s time: 119.85s eta: 17:20:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:30:58][INFO] [Train] (Epoch 1) Step 74/593 lr: 0.000020 loss: 0.254 loss(reduced): 0.176 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31852.0 tgs: 266 data_time: 0.75s time: 119.55s eta: 17:16:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:33:00][INFO] [Train] (Epoch 1) Step 75/593 lr: 0.000020 loss: 0.185 loss(reduced): 0.179 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32473.0 tgs: 266 data_time: 0.97s time: 121.86s eta: 17:34:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:35:03][INFO] [Train] (Epoch 1) Step 76/593 lr: 0.000020 loss: 0.164 loss(reduced): 0.171 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32194.0 tgs: 260 data_time: 0.89s time: 123.51s eta: 17:46:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:37:02][INFO] [Train] (Epoch 1) Step 77/593 lr: 0.000020 loss: 0.282 loss(reduced): 0.190 grad_norm: 0.22 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31340.0 tgs: 263 data_time: 0.76s time: 119.01s eta: 17:05:29 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:39:02][INFO] [Train] (Epoch 1) Step 78/593 lr: 0.000020 loss: 0.214 loss(reduced): 0.172 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31330.0 tgs: 262 data_time: 1.08s time: 119.22s eta: 17:05:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:41:01][INFO] [Train] (Epoch 1) Step 79/593 lr: 0.000020 loss: 0.166 loss(reduced): 0.181 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30764.0 tgs: 258 data_time: 0.95s time: 119.14s eta: 17:02:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:43:00][INFO] [Train] (Epoch 1) Step 80/593 lr: 0.000020 loss: 0.221 loss(reduced): 0.174 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30678.0 tgs: 257 data_time: 0.80s time: 119.25s eta: 17:01:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:44:59][INFO] [Train] (Epoch 1) Step 81/593 lr: 0.000020 loss: 0.188 loss(reduced): 0.184 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32234.0 tgs: 270 data_time: 1.04s time: 119.04s eta: 16:57:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:46:58][INFO] [Train] (Epoch 1) Step 82/593 lr: 0.000020 loss: 0.139 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31845.0 tgs: 266 data_time: 0.65s time: 119.27s eta: 16:57:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:48:57][INFO] [Train] (Epoch 1) Step 83/593 lr: 0.000020 loss: 0.216 loss(reduced): 0.169 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31363.0 tgs: 263 data_time: 0.94s time: 119.16s eta: 16:54:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:50:57][INFO] [Train] (Epoch 1) Step 84/593 lr: 0.000020 loss: 0.163 loss(reduced): 0.174 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31986.0 tgs: 267 data_time: 1.13s time: 119.58s eta: 16:56:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:52:56][INFO] [Train] (Epoch 1) Step 85/593 lr: 0.000020 loss: 0.192 loss(reduced): 0.173 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32421.0 tgs: 272 data_time: 0.65s time: 119.13s eta: 16:50:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:54:55][INFO] [Train] (Epoch 1) Step 86/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.171 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31499.0 tgs: 265 data_time: 0.56s time: 118.85s eta: 16:46:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:56:54][INFO] [Train] (Epoch 1) Step 87/593 lr: 0.000019 loss: 0.205 loss(reduced): 0.173 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30877.0 tgs: 259 data_time: 0.78s time: 119.00s eta: 16:45:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 19:58:54][INFO] [Train] (Epoch 1) Step 88/593 lr: 0.000019 loss: 0.200 loss(reduced): 0.164 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 32268.0 tgs: 269 data_time: 0.75s time: 119.85s eta: 16:50:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:00:53][INFO] [Train] (Epoch 1) Step 89/593 lr: 0.000019 loss: 0.234 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31891.0 tgs: 267 data_time: 0.99s time: 119.39s eta: 16:44:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:02:52][INFO] [Train] (Epoch 1) Step 90/593 lr: 0.000019 loss: 0.177 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 266 data_time: 0.84s time: 119.01s eta: 16:39:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:04:53][INFO] [Train] (Epoch 1) Step 91/593 lr: 0.000019 loss: 0.184 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31508.0 tgs: 260 data_time: 0.70s time: 120.92s eta: 16:53:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:06:52][INFO] [Train] (Epoch 1) Step 92/593 lr: 0.000019 loss: 0.190 loss(reduced): 0.171 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.6GB text_tokens: 31771.0 tgs: 267 data_time: 0.92s time: 118.99s eta: 16:35:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:08:52][INFO] [Train] (Epoch 1) Step 93/593 lr: 0.000019 loss: 0.128 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31833.0 tgs: 266 data_time: 0.81s time: 119.28s eta: 16:36:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:10:51][INFO] [Train] (Epoch 1) Step 94/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31534.0 tgs: 264 data_time: 0.96s time: 119.23s eta: 16:33:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:12:50][INFO] [Train] (Epoch 1) Step 95/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31668.0 tgs: 265 data_time: 0.89s time: 119.36s eta: 16:32:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:14:50][INFO] [Train] (Epoch 1) Step 96/593 lr: 0.000019 loss: 0.166 loss(reduced): 0.169 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32144.0 tgs: 268 data_time: 0.81s time: 119.70s eta: 16:33:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:16:49][INFO] [Train] (Epoch 1) Step 97/593 lr: 0.000019 loss: 0.189 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32092.0 tgs: 269 data_time: 0.76s time: 119.20s eta: 16:27:22 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:18:49][INFO] [Train] (Epoch 1) Step 98/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31700.0 tgs: 264 data_time: 0.71s time: 119.77s eta: 16:30:05 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:20:51][INFO] [Train] (Epoch 1) Step 99/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.174 grad_norm: 0.12 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 29764.0 tgs: 247 data_time: 0.84s time: 120.32s eta: 16:32:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:22:50][INFO] [Train] (Epoch 1) Step 100/593 lr: 0.000019 loss: 0.185 loss(reduced): 0.174 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32195.0 tgs: 270 data_time: 0.54s time: 118.97s eta: 16:19:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:24:57][INFO] [Train] (Epoch 1) Step 101/593 lr: 0.000019 loss: 0.159 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31755.0 tgs: 249 data_time: 0.88s time: 127.33s eta: 17:26:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:26:56][INFO] [Train] (Epoch 1) Step 102/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.170 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31490.0 tgs: 264 data_time: 0.63s time: 119.22s eta: 16:17:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:28:56][INFO] [Train] (Epoch 1) Step 103/593 lr: 0.000019 loss: 0.121 loss(reduced): 0.172 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31210.0 tgs: 261 data_time: 0.65s time: 119.14s eta: 16:14:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:30:55][INFO] [Train] (Epoch 1) Step 104/593 lr: 0.000019 loss: 0.173 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32019.0 tgs: 268 data_time: 0.88s time: 119.25s eta: 16:13:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:32:54][INFO] [Train] (Epoch 1) Step 105/593 lr: 0.000019 loss: 0.204 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31720.0 tgs: 266 data_time: 1.02s time: 119.17s eta: 16:11:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:34:53][INFO] [Train] (Epoch 1) Step 106/593 lr: 0.000019 loss: 0.193 loss(reduced): 0.170 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31536.0 tgs: 265 data_time: 0.67s time: 118.95s eta: 16:07:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:36:52][INFO] [Train] (Epoch 1) Step 107/593 lr: 0.000019 loss: 0.176 loss(reduced): 0.175 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32177.0 tgs: 270 data_time: 0.88s time: 118.86s eta: 16:04:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:38:51][INFO] [Train] (Epoch 1) Step 108/593 lr: 0.000019 loss: 0.168 loss(reduced): 0.176 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31877.0 tgs: 267 data_time: 0.81s time: 118.97s eta: 16:03:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:40:53][INFO] [Train] (Epoch 1) Step 109/593 lr: 0.000019 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31758.0 tgs: 260 data_time: 0.80s time: 121.84s eta: 16:24:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:42:52][INFO] [Train] (Epoch 1) Step 110/593 lr: 0.000019 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30694.0 tgs: 257 data_time: 0.93s time: 118.99s eta: 15:59:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:44:51][INFO] [Train] (Epoch 1) Step 111/593 lr: 0.000019 loss: 0.136 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31379.0 tgs: 263 data_time: 0.69s time: 119.23s eta: 15:59:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:46:50][INFO] [Train] (Epoch 1) Step 112/593 lr: 0.000019 loss: 0.162 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31729.0 tgs: 265 data_time: 0.55s time: 119.61s eta: 16:00:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:48:50][INFO] [Train] (Epoch 1) Step 113/593 lr: 0.000019 loss: 0.167 loss(reduced): 0.170 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30787.0 tgs: 258 data_time: 0.90s time: 119.04s eta: 15:54:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:50:48][INFO] [Train] (Epoch 1) Step 114/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31745.0 tgs: 266 data_time: 0.78s time: 118.98s eta: 15:51:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:52:48][INFO] [Train] (Epoch 1) Step 115/593 lr: 0.000019 loss: 0.138 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31728.0 tgs: 266 data_time: 0.72s time: 119.17s eta: 15:51:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:54:47][INFO] [Train] (Epoch 1) Step 116/593 lr: 0.000019 loss: 0.237 loss(reduced): 0.177 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32273.0 tgs: 269 data_time: 1.10s time: 119.63s eta: 15:53:05 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:56:46][INFO] [Train] (Epoch 1) Step 117/593 lr: 0.000019 loss: 0.206 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32305.0 tgs: 271 data_time: 0.80s time: 118.94s eta: 15:45:32 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 20:58:46][INFO] [Train] (Epoch 1) Step 118/593 lr: 0.000019 loss: 0.144 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31847.0 tgs: 266 data_time: 0.90s time: 119.28s eta: 15:46:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:00:45][INFO] [Train] (Epoch 1) Step 119/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32176.0 tgs: 270 data_time: 0.74s time: 119.13s eta: 15:43:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:02:44][INFO] [Train] (Epoch 1) Step 120/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.171 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 269 data_time: 0.70s time: 119.00s eta: 15:40:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:04:43][INFO] [Train] (Epoch 1) Step 121/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.175 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 267 data_time: 1.00s time: 119.65s eta: 15:43:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:06:45][INFO] [Train] (Epoch 1) Step 122/593 lr: 0.000019 loss: 0.209 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31505.0 tgs: 259 data_time: 0.87s time: 121.50s eta: 15:55:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:08:44][INFO] [Train] (Epoch 1) Step 123/593 lr: 0.000019 loss: 0.158 loss(reduced): 0.171 grad_norm: 0.14 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30426.0 tgs: 254 data_time: 0.75s time: 119.41s eta: 15:37:22 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:10:54][INFO] [Train] (Epoch 1) Step 124/593 lr: 0.000019 loss: 0.224 loss(reduced): 0.182 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31790.0 tgs: 244 data_time: 0.92s time: 129.96s eta: 16:58:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:12:54][INFO] [Train] (Epoch 1) Step 125/593 lr: 0.000019 loss: 0.146 loss(reduced): 0.165 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30635.0 tgs: 256 data_time: 0.98s time: 119.47s eta: 15:33:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:14:53][INFO] [Train] (Epoch 1) Step 126/593 lr: 0.000019 loss: 0.195 loss(reduced): 0.168 grad_norm: 0.23 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31503.0 tgs: 263 data_time: 0.84s time: 119.61s eta: 15:32:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:16:53][INFO] [Train] (Epoch 1) Step 127/593 lr: 0.000019 loss: 0.179 loss(reduced): 0.168 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32379.0 tgs: 270 data_time: 0.97s time: 119.52s eta: 15:30:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:18:55][INFO] [Train] (Epoch 1) Step 128/593 lr: 0.000019 loss: 0.198 loss(reduced): 0.179 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30778.0 tgs: 251 data_time: 0.95s time: 122.14s eta: 15:48:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:20:56][INFO] [Train] (Epoch 1) Step 129/593 lr: 0.000019 loss: 0.149 loss(reduced): 0.182 grad_norm: 0.20 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31838.0 tgs: 263 data_time: 0.96s time: 120.75s eta: 15:35:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:22:55][INFO] [Train] (Epoch 1) Step 130/593 lr: 0.000019 loss: 0.150 loss(reduced): 0.168 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31598.0 tgs: 264 data_time: 1.08s time: 119.41s eta: 15:23:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:24:55][INFO] [Train] (Epoch 1) Step 131/593 lr: 0.000019 loss: 0.140 loss(reduced): 0.169 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32021.0 tgs: 268 data_time: 0.73s time: 119.44s eta: 15:21:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:27:06][INFO] [Train] (Epoch 1) Step 132/593 lr: 0.000019 loss: 0.152 loss(reduced): 0.168 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30189.0 tgs: 229 data_time: 0.92s time: 131.63s eta: 16:53:32 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:29:06][INFO] [Train] (Epoch 1) Step 133/593 lr: 0.000019 loss: 0.165 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30681.0 tgs: 257 data_time: 1.04s time: 119.29s eta: 15:16:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:31:05][INFO] [Train] (Epoch 1) Step 134/593 lr: 0.000019 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.21 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32013.0 tgs: 267 data_time: 0.90s time: 119.57s eta: 15:16:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:33:07][INFO] [Train] (Epoch 1) Step 135/593 lr: 0.000019 loss: 0.192 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31504.0 tgs: 257 data_time: 0.94s time: 122.25s eta: 15:35:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:35:07][INFO] [Train] (Epoch 1) Step 136/593 lr: 0.000019 loss: 0.130 loss(reduced): 0.159 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31810.0 tgs: 265 data_time: 0.63s time: 119.71s eta: 15:13:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:37:07][INFO] [Train] (Epoch 1) Step 137/593 lr: 0.000019 loss: 0.171 loss(reduced): 0.167 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31920.0 tgs: 267 data_time: 0.74s time: 119.35s eta: 15:09:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:39:05][INFO] [Train] (Epoch 1) Step 138/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.176 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30995.0 tgs: 260 data_time: 0.98s time: 118.90s eta: 15:03:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:41:04][INFO] [Train] (Epoch 1) Step 139/593 lr: 0.000018 loss: 0.124 loss(reduced): 0.167 grad_norm: 0.15 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32130.0 tgs: 270 data_time: 0.64s time: 119.00s eta: 15:02:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:43:04][INFO] [Train] (Epoch 1) Step 140/593 lr: 0.000018 loss: 0.175 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32369.0 tgs: 271 data_time: 0.52s time: 119.30s eta: 15:02:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:45:03][INFO] [Train] (Epoch 1) Step 141/593 lr: 0.000018 loss: 0.188 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31687.0 tgs: 266 data_time: 0.73s time: 118.92s eta: 14:57:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:47:03][INFO] [Train] (Epoch 1) Step 142/593 lr: 0.000018 loss: 0.203 loss(reduced): 0.172 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30954.0 tgs: 256 data_time: 0.58s time: 120.52s eta: 15:07:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:49:03][INFO] [Train] (Epoch 1) Step 143/593 lr: 0.000018 loss: 0.166 loss(reduced): 0.168 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31419.0 tgs: 263 data_time: 0.91s time: 119.32s eta: 14:56:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:51:05][INFO] [Train] (Epoch 1) Step 144/593 lr: 0.000018 loss: 0.174 loss(reduced): 0.183 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29745.0 tgs: 242 data_time: 0.85s time: 122.65s eta: 15:19:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:53:04][INFO] [Train] (Epoch 1) Step 145/593 lr: 0.000018 loss: 0.130 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31451.0 tgs: 264 data_time: 0.77s time: 119.12s eta: 14:51:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:55:04][INFO] [Train] (Epoch 1) Step 146/593 lr: 0.000018 loss: 0.179 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30525.0 tgs: 254 data_time: 0.99s time: 119.76s eta: 14:54:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:57:04][INFO] [Train] (Epoch 1) Step 147/593 lr: 0.000018 loss: 0.138 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31844.0 tgs: 266 data_time: 0.77s time: 119.46s eta: 14:50:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 21:59:03][INFO] [Train] (Epoch 1) Step 148/593 lr: 0.000018 loss: 0.178 loss(reduced): 0.177 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31736.0 tgs: 266 data_time: 0.66s time: 118.99s eta: 14:44:28 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:01:02][INFO] [Train] (Epoch 1) Step 149/593 lr: 0.000018 loss: 0.145 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31460.0 tgs: 263 data_time: 0.68s time: 119.33s eta: 14:45:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:03:01][INFO] [Train] (Epoch 1) Step 150/593 lr: 0.000018 loss: 0.206 loss(reduced): 0.162 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31348.0 tgs: 263 data_time: 0.82s time: 118.94s eta: 14:40:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:05:00][INFO] [Train] (Epoch 1) Step 151/593 lr: 0.000018 loss: 0.164 loss(reduced): 0.169 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31771.0 tgs: 266 data_time: 1.18s time: 119.06s eta: 14:39:04 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:06:59][INFO] [Train] (Epoch 1) Step 152/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31988.0 tgs: 268 data_time: 0.85s time: 119.05s eta: 14:37:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:08:58][INFO] [Train] (Epoch 1) Step 153/593 lr: 0.000018 loss: 0.135 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31563.0 tgs: 264 data_time: 0.59s time: 119.13s eta: 14:35:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:10:57][INFO] [Train] (Epoch 1) Step 154/593 lr: 0.000018 loss: 0.158 loss(reduced): 0.159 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30801.0 tgs: 258 data_time: 0.88s time: 119.00s eta: 14:32:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:12:56][INFO] [Train] (Epoch 1) Step 155/593 lr: 0.000018 loss: 0.252 loss(reduced): 0.167 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31800.0 tgs: 267 data_time: 0.93s time: 119.07s eta: 14:31:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:14:55][INFO] [Train] (Epoch 1) Step 156/593 lr: 0.000018 loss: 0.183 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31863.0 tgs: 267 data_time: 0.72s time: 119.15s eta: 14:29:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:16:55][INFO] [Train] (Epoch 1) Step 157/593 lr: 0.000018 loss: 0.250 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32314.0 tgs: 270 data_time: 1.13s time: 119.64s eta: 14:31:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:18:54][INFO] [Train] (Epoch 1) Step 158/593 lr: 0.000018 loss: 0.144 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30878.0 tgs: 259 data_time: 0.95s time: 119.18s eta: 14:26:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:20:53][INFO] [Train] (Epoch 1) Step 159/593 lr: 0.000018 loss: 0.140 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31664.0 tgs: 265 data_time: 0.87s time: 119.20s eta: 14:24:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:22:53][INFO] [Train] (Epoch 1) Step 160/593 lr: 0.000018 loss: 0.146 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32152.0 tgs: 269 data_time: 0.66s time: 119.39s eta: 14:23:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:24:54][INFO] [Train] (Epoch 1) Step 161/593 lr: 0.000018 loss: 0.160 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32029.0 tgs: 265 data_time: 0.99s time: 120.73s eta: 14:31:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:26:52][INFO] [Train] (Epoch 1) Step 162/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31643.0 tgs: 266 data_time: 0.93s time: 118.85s eta: 14:15:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:28:52][INFO] [Train] (Epoch 1) Step 163/593 lr: 0.000018 loss: 0.125 loss(reduced): 0.169 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31134.0 tgs: 261 data_time: 0.82s time: 119.10s eta: 14:15:30 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:30:51][INFO] [Train] (Epoch 1) Step 164/593 lr: 0.000018 loss: 0.163 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31687.0 tgs: 265 data_time: 0.81s time: 119.15s eta: 14:13:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:32:51][INFO] [Train] (Epoch 1) Step 165/593 lr: 0.000018 loss: 0.155 loss(reduced): 0.179 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31514.0 tgs: 262 data_time: 0.82s time: 119.96s eta: 14:17:44 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:34:51][INFO] [Train] (Epoch 1) Step 166/593 lr: 0.000018 loss: 0.191 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31467.0 tgs: 262 data_time: 1.21s time: 119.86s eta: 14:15:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:36:50][INFO] [Train] (Epoch 1) Step 167/593 lr: 0.000018 loss: 0.227 loss(reduced): 0.175 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31775.0 tgs: 265 data_time: 0.93s time: 119.59s eta: 14:11:05 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:38:49][INFO] [Train] (Epoch 1) Step 168/593 lr: 0.000018 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31887.0 tgs: 267 data_time: 0.84s time: 119.22s eta: 14:06:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:40:49][INFO] [Train] (Epoch 1) Step 169/593 lr: 0.000018 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32107.0 tgs: 269 data_time: 0.81s time: 119.26s eta: 14:04:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:42:48][INFO] [Train] (Epoch 1) Step 170/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.170 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31513.0 tgs: 264 data_time: 0.70s time: 119.26s eta: 14:02:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:44:47][INFO] [Train] (Epoch 1) Step 171/593 lr: 0.000018 loss: 0.169 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32275.0 tgs: 270 data_time: 0.89s time: 119.14s eta: 13:59:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:46:46][INFO] [Train] (Epoch 1) Step 172/593 lr: 0.000018 loss: 0.131 loss(reduced): 0.161 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31573.0 tgs: 265 data_time: 0.63s time: 119.12s eta: 13:57:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:48:45][INFO] [Train] (Epoch 1) Step 173/593 lr: 0.000018 loss: 0.174 loss(reduced): 0.164 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31903.0 tgs: 267 data_time: 0.74s time: 119.17s eta: 13:56:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:50:45][INFO] [Train] (Epoch 1) Step 174/593 lr: 0.000018 loss: 0.126 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32083.0 tgs: 268 data_time: 1.12s time: 119.35s eta: 13:55:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:52:45][INFO] [Train] (Epoch 1) Step 175/593 lr: 0.000018 loss: 0.136 loss(reduced): 0.173 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31342.0 tgs: 260 data_time: 1.07s time: 120.11s eta: 13:58:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:54:47][INFO] [Train] (Epoch 1) Step 176/593 lr: 0.000017 loss: 0.147 loss(reduced): 0.164 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32330.0 tgs: 263 data_time: 0.90s time: 122.55s eta: 14:13:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:56:47][INFO] [Train] (Epoch 1) Step 177/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31356.0 tgs: 263 data_time: 0.76s time: 119.10s eta: 13:47:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 22:58:46][INFO] [Train] (Epoch 1) Step 178/593 lr: 0.000017 loss: 0.132 loss(reduced): 0.170 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32027.0 tgs: 268 data_time: 0.91s time: 119.37s eta: 13:47:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:00:45][INFO] [Train] (Epoch 1) Step 179/593 lr: 0.000017 loss: 0.141 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31724.0 tgs: 265 data_time: 0.99s time: 119.36s eta: 13:45:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:02:45][INFO] [Train] (Epoch 1) Step 180/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31187.0 tgs: 261 data_time: 0.73s time: 119.33s eta: 13:43:22 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:04:44][INFO] [Train] (Epoch 1) Step 181/593 lr: 0.000017 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31680.0 tgs: 265 data_time: 1.04s time: 119.12s eta: 13:39:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:06:43][INFO] [Train] (Epoch 1) Step 182/593 lr: 0.000017 loss: 0.128 loss(reduced): 0.157 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30117.0 tgs: 252 data_time: 0.73s time: 119.25s eta: 13:38:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:08:42][INFO] [Train] (Epoch 1) Step 183/593 lr: 0.000017 loss: 0.235 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31411.0 tgs: 263 data_time: 0.96s time: 119.13s eta: 13:36:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:10:41][INFO] [Train] (Epoch 1) Step 184/593 lr: 0.000017 loss: 0.158 loss(reduced): 0.171 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30856.0 tgs: 258 data_time: 0.72s time: 119.22s eta: 13:34:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:12:41][INFO] [Train] (Epoch 1) Step 185/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.165 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32307.0 tgs: 270 data_time: 0.91s time: 119.46s eta: 13:34:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:14:40][INFO] [Train] (Epoch 1) Step 186/593 lr: 0.000017 loss: 0.166 loss(reduced): 0.163 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30702.0 tgs: 257 data_time: 0.67s time: 119.07s eta: 13:29:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:16:39][INFO] [Train] (Epoch 1) Step 187/593 lr: 0.000017 loss: 0.181 loss(reduced): 0.166 grad_norm: 0.18 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31557.0 tgs: 265 data_time: 0.89s time: 118.92s eta: 13:26:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:18:39][INFO] [Train] (Epoch 1) Step 188/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.163 grad_norm: 0.19 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30984.0 tgs: 258 data_time: 0.85s time: 119.81s eta: 13:30:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:20:38][INFO] [Train] (Epoch 1) Step 189/593 lr: 0.000017 loss: 0.276 loss(reduced): 0.165 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32159.0 tgs: 269 data_time: 1.02s time: 119.25s eta: 13:24:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:22:37][INFO] [Train] (Epoch 1) Step 190/593 lr: 0.000017 loss: 0.178 loss(reduced): 0.168 grad_norm: 0.19 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31857.0 tgs: 267 data_time: 0.78s time: 119.16s eta: 13:22:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:24:36][INFO] [Train] (Epoch 1) Step 191/593 lr: 0.000017 loss: 0.156 loss(reduced): 0.164 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31142.0 tgs: 260 data_time: 0.79s time: 119.34s eta: 13:21:35 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:26:36][INFO] [Train] (Epoch 1) Step 192/593 lr: 0.000017 loss: 0.153 loss(reduced): 0.157 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31976.0 tgs: 268 data_time: 0.69s time: 119.26s eta: 13:19:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:28:35][INFO] [Train] (Epoch 1) Step 193/593 lr: 0.000017 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.16 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31689.0 tgs: 266 data_time: 0.52s time: 119.04s eta: 13:15:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:30:34][INFO] [Train] (Epoch 1) Step 194/593 lr: 0.000017 loss: 0.136 loss(reduced): 0.169 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32024.0 tgs: 268 data_time: 0.90s time: 119.37s eta: 13:15:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:32:33][INFO] [Train] (Epoch 1) Step 195/593 lr: 0.000017 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.14 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31461.0 tgs: 263 data_time: 1.00s time: 119.19s eta: 13:12:35 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:34:33][INFO] [Train] (Epoch 1) Step 196/593 lr: 0.000017 loss: 0.157 loss(reduced): 0.167 grad_norm: 0.16 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30421.0 tgs: 254 data_time: 0.80s time: 119.46s eta: 13:12:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:36:32][INFO] [Train] (Epoch 1) Step 197/593 lr: 0.000017 loss: 0.150 loss(reduced): 0.166 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31602.0 tgs: 264 data_time: 0.79s time: 119.29s eta: 13:09:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:38:31][INFO] [Train] (Epoch 1) Step 198/593 lr: 0.000017 loss: 0.161 loss(reduced): 0.162 grad_norm: 0.13 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30923.0 tgs: 259 data_time: 0.77s time: 119.14s eta: 13:06:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:40:33][INFO] [Train] (Epoch 1) Step 199/593 lr: 0.000017 loss: 0.115 loss(reduced): 0.172 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31168.0 tgs: 258 data_time: 0.87s time: 120.41s eta: 13:12:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:42:33][INFO] [Train] (Epoch 1) Step 200/593 lr: 0.000017 loss: 0.245 loss(reduced): 0.160 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31210.0 tgs: 259 data_time: 0.86s time: 120.16s eta: 13:09:02 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:44:33][INFO] [Train] (Epoch 1) Step 201/593 lr: 0.000017 loss: 0.139 loss(reduced): 0.167 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31474.0 tgs: 263 data_time: 0.99s time: 119.50s eta: 13:02:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:46:32][INFO] [Train] (Epoch 1) Step 202/593 lr: 0.000017 loss: 0.133 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32426.0 tgs: 271 data_time: 0.71s time: 119.51s eta: 13:00:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:48:32][INFO] [Train] (Epoch 1) Step 203/593 lr: 0.000017 loss: 0.163 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30215.0 tgs: 252 data_time: 0.68s time: 119.58s eta: 12:59:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:50:32][INFO] [Train] (Epoch 1) Step 204/593 lr: 0.000017 loss: 0.160 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32410.0 tgs: 270 data_time: 0.76s time: 119.79s eta: 12:58:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:52:31][INFO] [Train] (Epoch 1) Step 205/593 lr: 0.000017 loss: 0.129 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30666.0 tgs: 257 data_time: 0.67s time: 119.04s eta: 12:51:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:54:30][INFO] [Train] (Epoch 1) Step 206/593 lr: 0.000017 loss: 0.213 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32326.0 tgs: 271 data_time: 0.73s time: 119.05s eta: 12:49:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:56:29][INFO] [Train] (Epoch 1) Step 207/593 lr: 0.000017 loss: 0.155 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31571.0 tgs: 265 data_time: 0.78s time: 119.12s eta: 12:48:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-21 23:58:29][INFO] [Train] (Epoch 1) Step 208/593 lr: 0.000016 loss: 0.143 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31943.0 tgs: 267 data_time: 0.60s time: 119.40s eta: 12:48:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:00:28][INFO] [Train] (Epoch 1) Step 209/593 lr: 0.000016 loss: 0.113 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31275.0 tgs: 262 data_time: 0.73s time: 119.19s eta: 12:44:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:02:27][INFO] [Train] (Epoch 1) Step 210/593 lr: 0.000016 loss: 0.183 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31952.0 tgs: 267 data_time: 1.05s time: 119.43s eta: 12:44:22 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:04:26][INFO] [Train] (Epoch 1) Step 211/593 lr: 0.000016 loss: 0.169 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31769.0 tgs: 266 data_time: 0.88s time: 119.10s eta: 12:40:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:06:25][INFO] [Train] (Epoch 1) Step 212/593 lr: 0.000016 loss: 0.205 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30931.0 tgs: 259 data_time: 0.81s time: 119.06s eta: 12:38:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:08:25][INFO] [Train] (Epoch 1) Step 213/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32408.0 tgs: 270 data_time: 0.94s time: 119.59s eta: 12:39:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:10:24][INFO] [Train] (Epoch 1) Step 214/593 lr: 0.000016 loss: 0.157 loss(reduced): 0.158 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30927.0 tgs: 260 data_time: 0.75s time: 118.78s eta: 12:32:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:12:23][INFO] [Train] (Epoch 1) Step 215/593 lr: 0.000016 loss: 0.156 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30636.0 tgs: 257 data_time: 0.86s time: 118.88s eta: 12:30:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:14:22][INFO] [Train] (Epoch 1) Step 216/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.166 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32377.0 tgs: 271 data_time: 0.81s time: 119.17s eta: 12:30:44 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:16:21][INFO] [Train] (Epoch 1) Step 217/593 lr: 0.000016 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30564.0 tgs: 256 data_time: 0.70s time: 119.27s eta: 12:29:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:18:20][INFO] [Train] (Epoch 1) Step 218/593 lr: 0.000016 loss: 0.167 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31100.0 tgs: 260 data_time: 0.80s time: 119.35s eta: 12:27:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:20:20][INFO] [Train] (Epoch 1) Step 219/593 lr: 0.000016 loss: 0.185 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31643.0 tgs: 265 data_time: 0.84s time: 119.12s eta: 12:24:29 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:22:19][INFO] [Train] (Epoch 1) Step 220/593 lr: 0.000016 loss: 0.174 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30745.0 tgs: 257 data_time: 0.88s time: 119.35s eta: 12:23:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:24:18][INFO] [Train] (Epoch 1) Step 221/593 lr: 0.000016 loss: 0.139 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 267 data_time: 0.69s time: 119.26s eta: 12:21:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:26:17][INFO] [Train] (Epoch 1) Step 222/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32037.0 tgs: 269 data_time: 0.80s time: 119.03s eta: 12:17:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:28:18][INFO] [Train] (Epoch 1) Step 223/593 lr: 0.000016 loss: 0.154 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31918.0 tgs: 263 data_time: 0.78s time: 121.12s eta: 12:28:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:30:18][INFO] [Train] (Epoch 1) Step 224/593 lr: 0.000016 loss: 0.170 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32191.0 tgs: 268 data_time: 0.88s time: 119.99s eta: 12:19:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:32:17][INFO] [Train] (Epoch 1) Step 225/593 lr: 0.000016 loss: 0.158 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32068.0 tgs: 270 data_time: 0.83s time: 118.72s eta: 12:10:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:34:17][INFO] [Train] (Epoch 1) Step 226/593 lr: 0.000016 loss: 0.133 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30805.0 tgs: 257 data_time: 0.90s time: 119.63s eta: 12:13:44 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:36:17][INFO] [Train] (Epoch 1) Step 227/593 lr: 0.000016 loss: 0.121 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 29985.0 tgs: 250 data_time: 0.62s time: 119.83s eta: 12:12:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:38:16][INFO] [Train] (Epoch 1) Step 228/593 lr: 0.000016 loss: 0.159 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31869.0 tgs: 267 data_time: 0.66s time: 119.14s eta: 12:06:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:40:15][INFO] [Train] (Epoch 1) Step 229/593 lr: 0.000016 loss: 0.128 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31337.0 tgs: 262 data_time: 1.03s time: 119.17s eta: 12:04:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:42:14][INFO] [Train] (Epoch 1) Step 230/593 lr: 0.000016 loss: 0.150 loss(reduced): 0.173 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32491.0 tgs: 272 data_time: 0.67s time: 119.11s eta: 12:02:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:44:13][INFO] [Train] (Epoch 1) Step 231/593 lr: 0.000016 loss: 0.119 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31755.0 tgs: 266 data_time: 0.75s time: 119.18s eta: 12:01:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:46:12][INFO] [Train] (Epoch 1) Step 232/593 lr: 0.000016 loss: 0.211 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31670.0 tgs: 266 data_time: 0.91s time: 118.95s eta: 11:57:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:48:12][INFO] [Train] (Epoch 1) Step 233/593 lr: 0.000016 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31798.0 tgs: 265 data_time: 0.77s time: 119.70s eta: 12:00:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:50:11][INFO] [Train] (Epoch 1) Step 234/593 lr: 0.000016 loss: 0.151 loss(reduced): 0.169 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30741.0 tgs: 257 data_time: 0.92s time: 119.28s eta: 11:55:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:52:10][INFO] [Train] (Epoch 1) Step 235/593 lr: 0.000016 loss: 0.185 loss(reduced): 0.168 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32020.0 tgs: 268 data_time: 0.70s time: 119.27s eta: 11:53:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:54:10][INFO] [Train] (Epoch 1) Step 236/593 lr: 0.000016 loss: 0.168 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31954.0 tgs: 266 data_time: 1.02s time: 119.71s eta: 11:54:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:56:09][INFO] [Train] (Epoch 1) Step 237/593 lr: 0.000016 loss: 0.234 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31112.0 tgs: 261 data_time: 0.90s time: 118.78s eta: 11:46:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 00:58:08][INFO] [Train] (Epoch 1) Step 238/593 lr: 0.000015 loss: 0.141 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31946.0 tgs: 268 data_time: 0.64s time: 119.15s eta: 11:46:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:00:08][INFO] [Train] (Epoch 1) Step 239/593 lr: 0.000015 loss: 0.152 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30980.0 tgs: 259 data_time: 0.85s time: 119.61s eta: 11:47:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:02:07][INFO] [Train] (Epoch 1) Step 240/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31681.0 tgs: 266 data_time: 0.84s time: 118.86s eta: 11:41:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:04:05][INFO] [Train] (Epoch 1) Step 241/593 lr: 0.000015 loss: 0.171 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31755.0 tgs: 267 data_time: 0.53s time: 118.68s eta: 11:38:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:06:05][INFO] [Train] (Epoch 1) Step 242/593 lr: 0.000015 loss: 0.123 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32017.0 tgs: 268 data_time: 0.77s time: 119.21s eta: 11:39:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:08:04][INFO] [Train] (Epoch 1) Step 243/593 lr: 0.000015 loss: 0.121 loss(reduced): 0.162 grad_norm: 0.38 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32488.0 tgs: 272 data_time: 0.61s time: 119.14s eta: 11:36:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:10:03][INFO] [Train] (Epoch 1) Step 244/593 lr: 0.000015 loss: 0.197 loss(reduced): 0.162 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32372.0 tgs: 271 data_time: 0.83s time: 119.08s eta: 11:34:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:12:02][INFO] [Train] (Epoch 1) Step 245/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31108.0 tgs: 261 data_time: 0.94s time: 118.95s eta: 11:31:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:14:01][INFO] [Train] (Epoch 1) Step 246/593 lr: 0.000015 loss: 0.147 loss(reduced): 0.165 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30700.0 tgs: 258 data_time: 0.71s time: 118.91s eta: 11:29:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:16:00][INFO] [Train] (Epoch 1) Step 247/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31892.0 tgs: 268 data_time: 0.81s time: 118.98s eta: 11:28:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:17:59][INFO] [Train] (Epoch 1) Step 248/593 lr: 0.000015 loss: 0.135 loss(reduced): 0.156 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31610.0 tgs: 265 data_time: 1.10s time: 119.27s eta: 11:27:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:19:58][INFO] [Train] (Epoch 1) Step 249/593 lr: 0.000015 loss: 0.187 loss(reduced): 0.158 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32159.0 tgs: 269 data_time: 1.08s time: 119.35s eta: 11:26:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:21:57][INFO] [Train] (Epoch 1) Step 250/593 lr: 0.000015 loss: 0.173 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31275.0 tgs: 263 data_time: 0.92s time: 118.88s eta: 11:21:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:23:56][INFO] [Train] (Epoch 1) Step 251/593 lr: 0.000015 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31077.0 tgs: 260 data_time: 0.74s time: 119.21s eta: 11:21:30 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:25:56][INFO] [Train] (Epoch 1) Step 252/593 lr: 0.000015 loss: 0.125 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31813.0 tgs: 265 data_time: 0.65s time: 119.71s eta: 11:22:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:27:55][INFO] [Train] (Epoch 1) Step 253/593 lr: 0.000015 loss: 0.155 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32161.0 tgs: 269 data_time: 0.90s time: 119.12s eta: 11:16:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:29:54][INFO] [Train] (Epoch 1) Step 254/593 lr: 0.000015 loss: 0.116 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30782.0 tgs: 258 data_time: 0.73s time: 119.12s eta: 11:15:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:31:53][INFO] [Train] (Epoch 1) Step 255/593 lr: 0.000015 loss: 0.136 loss(reduced): 0.165 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31593.0 tgs: 265 data_time: 0.65s time: 118.99s eta: 11:12:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:33:52][INFO] [Train] (Epoch 1) Step 256/593 lr: 0.000015 loss: 0.131 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31866.0 tgs: 267 data_time: 0.64s time: 119.09s eta: 11:10:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:35:52][INFO] [Train] (Epoch 1) Step 257/593 lr: 0.000015 loss: 0.231 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.6GB text_tokens: 30517.0 tgs: 256 data_time: 0.79s time: 119.16s eta: 11:09:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:37:51][INFO] [Train] (Epoch 1) Step 258/593 lr: 0.000015 loss: 0.148 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31824.0 tgs: 267 data_time: 0.95s time: 118.97s eta: 11:06:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:39:50][INFO] [Train] (Epoch 1) Step 259/593 lr: 0.000015 loss: 0.179 loss(reduced): 0.166 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31864.0 tgs: 266 data_time: 0.98s time: 119.47s eta: 11:07:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:41:50][INFO] [Train] (Epoch 1) Step 260/593 lr: 0.000015 loss: 0.128 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31661.0 tgs: 264 data_time: 0.97s time: 119.64s eta: 11:06:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:43:49][INFO] [Train] (Epoch 1) Step 261/593 lr: 0.000015 loss: 0.184 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31328.0 tgs: 262 data_time: 0.70s time: 119.30s eta: 11:02:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:45:48][INFO] [Train] (Epoch 1) Step 262/593 lr: 0.000015 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32509.0 tgs: 272 data_time: 1.05s time: 119.31s eta: 11:00:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:47:47][INFO] [Train] (Epoch 1) Step 263/593 lr: 0.000015 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32300.0 tgs: 271 data_time: 0.95s time: 118.82s eta: 10:55:28 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:49:47][INFO] [Train] (Epoch 1) Step 264/593 lr: 0.000015 loss: 0.154 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30530.0 tgs: 255 data_time: 0.75s time: 119.47s eta: 10:57:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:51:46][INFO] [Train] (Epoch 1) Step 265/593 lr: 0.000014 loss: 0.225 loss(reduced): 0.167 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31745.0 tgs: 265 data_time: 0.67s time: 119.36s eta: 10:54:30 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:53:45][INFO] [Train] (Epoch 1) Step 266/593 lr: 0.000014 loss: 0.135 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32592.0 tgs: 273 data_time: 0.62s time: 119.28s eta: 10:52:02 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:55:45][INFO] [Train] (Epoch 1) Step 267/593 lr: 0.000014 loss: 0.119 loss(reduced): 0.156 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30908.0 tgs: 259 data_time: 0.86s time: 119.25s eta: 10:49:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:57:44][INFO] [Train] (Epoch 1) Step 268/593 lr: 0.000014 loss: 0.166 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32199.0 tgs: 270 data_time: 1.01s time: 119.10s eta: 10:47:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 01:59:43][INFO] [Train] (Epoch 1) Step 269/593 lr: 0.000014 loss: 0.137 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 28891.0 tgs: 241 data_time: 0.48s time: 119.52s eta: 10:47:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:01:43][INFO] [Train] (Epoch 1) Step 270/593 lr: 0.000014 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 29898.0 tgs: 250 data_time: 1.06s time: 119.32s eta: 10:44:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:03:42][INFO] [Train] (Epoch 1) Step 271/593 lr: 0.000014 loss: 0.184 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31937.0 tgs: 267 data_time: 1.16s time: 119.49s eta: 10:43:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:05:42][INFO] [Train] (Epoch 1) Step 272/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31014.0 tgs: 259 data_time: 0.91s time: 119.51s eta: 10:41:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:07:41][INFO] [Train] (Epoch 1) Step 273/593 lr: 0.000014 loss: 0.170 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31620.0 tgs: 265 data_time: 0.89s time: 118.99s eta: 10:36:35 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:09:40][INFO] [Train] (Epoch 1) Step 274/593 lr: 0.000014 loss: 0.149 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31997.0 tgs: 268 data_time: 0.80s time: 119.22s eta: 10:35:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:11:39][INFO] [Train] (Epoch 1) Step 275/593 lr: 0.000014 loss: 0.189 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32163.0 tgs: 270 data_time: 0.70s time: 119.06s eta: 10:32:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:13:38][INFO] [Train] (Epoch 1) Step 276/593 lr: 0.000014 loss: 0.158 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31447.0 tgs: 264 data_time: 0.75s time: 118.78s eta: 10:29:32 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:15:38][INFO] [Train] (Epoch 1) Step 277/593 lr: 0.000014 loss: 0.219 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31725.0 tgs: 264 data_time: 0.70s time: 119.86s eta: 10:33:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:17:37][INFO] [Train] (Epoch 1) Step 278/593 lr: 0.000014 loss: 0.184 loss(reduced): 0.165 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32494.0 tgs: 272 data_time: 0.91s time: 119.38s eta: 10:28:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:19:36][INFO] [Train] (Epoch 1) Step 279/593 lr: 0.000014 loss: 0.164 loss(reduced): 0.163 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31626.0 tgs: 266 data_time: 0.62s time: 118.83s eta: 10:23:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:21:35][INFO] [Train] (Epoch 1) Step 280/593 lr: 0.000014 loss: 0.117 loss(reduced): 0.163 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31628.0 tgs: 265 data_time: 0.61s time: 118.98s eta: 10:22:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:23:34][INFO] [Train] (Epoch 1) Step 281/593 lr: 0.000014 loss: 0.179 loss(reduced): 0.163 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32482.0 tgs: 272 data_time: 0.80s time: 119.28s eta: 10:22:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:25:33][INFO] [Train] (Epoch 1) Step 282/593 lr: 0.000014 loss: 0.147 loss(reduced): 0.160 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31600.0 tgs: 265 data_time: 1.28s time: 119.20s eta: 10:19:51 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:27:32][INFO] [Train] (Epoch 1) Step 283/593 lr: 0.000014 loss: 0.146 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31925.0 tgs: 268 data_time: 0.58s time: 119.06s eta: 10:17:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:29:32][INFO] [Train] (Epoch 1) Step 284/593 lr: 0.000014 loss: 0.138 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32318.0 tgs: 270 data_time: 0.83s time: 119.47s eta: 10:17:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:31:31][INFO] [Train] (Epoch 1) Step 285/593 lr: 0.000014 loss: 0.160 loss(reduced): 0.165 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32153.0 tgs: 269 data_time: 0.67s time: 119.42s eta: 10:15:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:33:31][INFO] [Train] (Epoch 1) Step 286/593 lr: 0.000014 loss: 0.110 loss(reduced): 0.168 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31662.0 tgs: 265 data_time: 0.70s time: 119.45s eta: 10:13:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:35:30][INFO] [Train] (Epoch 1) Step 287/593 lr: 0.000014 loss: 0.171 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30437.0 tgs: 255 data_time: 1.09s time: 119.28s eta: 10:10:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:37:29][INFO] [Train] (Epoch 1) Step 288/593 lr: 0.000014 loss: 0.128 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32081.0 tgs: 269 data_time: 0.75s time: 119.11s eta: 10:07:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:39:29][INFO] [Train] (Epoch 1) Step 289/593 lr: 0.000014 loss: 0.145 loss(reduced): 0.157 grad_norm: 0.13 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31592.0 tgs: 263 data_time: 0.90s time: 119.70s eta: 10:08:28 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:41:29][INFO] [Train] (Epoch 1) Step 290/593 lr: 0.000014 loss: 0.119 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31436.0 tgs: 262 data_time: 0.58s time: 119.83s eta: 10:07:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:43:30][INFO] [Train] (Epoch 1) Step 291/593 lr: 0.000014 loss: 0.265 loss(reduced): 0.183 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31772.0 tgs: 261 data_time: 0.91s time: 121.55s eta: 10:13:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:45:29][INFO] [Train] (Epoch 1) Step 292/593 lr: 0.000013 loss: 0.148 loss(reduced): 0.158 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31699.0 tgs: 265 data_time: 0.79s time: 119.25s eta: 10:00:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:47:28][INFO] [Train] (Epoch 1) Step 293/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32257.0 tgs: 271 data_time: 0.85s time: 118.80s eta: 9:56:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:49:28][INFO] [Train] (Epoch 1) Step 294/593 lr: 0.000013 loss: 0.140 loss(reduced): 0.168 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31712.0 tgs: 265 data_time: 0.90s time: 119.56s eta: 9:57:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:51:28][INFO] [Train] (Epoch 1) Step 295/593 lr: 0.000013 loss: 0.151 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31742.0 tgs: 265 data_time: 0.86s time: 119.70s eta: 9:56:30 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:53:27][INFO] [Train] (Epoch 1) Step 296/593 lr: 0.000013 loss: 0.161 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31585.0 tgs: 264 data_time: 0.70s time: 119.26s eta: 9:52:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:55:26][INFO] [Train] (Epoch 1) Step 297/593 lr: 0.000013 loss: 0.164 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30911.0 tgs: 259 data_time: 0.67s time: 119.33s eta: 9:50:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:57:25][INFO] [Train] (Epoch 1) Step 298/593 lr: 0.000013 loss: 0.168 loss(reduced): 0.161 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31689.0 tgs: 265 data_time: 1.00s time: 119.22s eta: 9:48:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 02:59:28][INFO] [Train] (Epoch 1) Step 299/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31738.0 tgs: 262 data_time: 1.00s time: 120.87s eta: 9:54:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:01:27][INFO] [Train] (Epoch 1) Step 300/593 lr: 0.000013 loss: 0.150 loss(reduced): 0.167 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31384.0 tgs: 263 data_time: 0.82s time: 119.27s eta: 9:44:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:03:27][INFO] [Train] (Epoch 1) Step 301/593 lr: 0.000013 loss: 0.181 loss(reduced): 0.166 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32173.0 tgs: 269 data_time: 0.75s time: 119.46s eta: 9:43:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:05:26][INFO] [Train] (Epoch 1) Step 302/593 lr: 0.000013 loss: 0.152 loss(reduced): 0.166 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31151.0 tgs: 260 data_time: 0.62s time: 119.55s eta: 9:41:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:07:25][INFO] [Train] (Epoch 1) Step 303/593 lr: 0.000013 loss: 0.146 loss(reduced): 0.150 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32053.0 tgs: 268 data_time: 0.90s time: 119.20s eta: 9:38:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:09:25][INFO] [Train] (Epoch 1) Step 304/593 lr: 0.000013 loss: 0.156 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31706.0 tgs: 265 data_time: 0.69s time: 119.50s eta: 9:37:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:11:24][INFO] [Train] (Epoch 1) Step 305/593 lr: 0.000013 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30917.0 tgs: 259 data_time: 0.84s time: 119.04s eta: 9:33:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:13:23][INFO] [Train] (Epoch 1) Step 306/593 lr: 0.000013 loss: 0.111 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32313.0 tgs: 271 data_time: 0.62s time: 119.23s eta: 9:32:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:15:22][INFO] [Train] (Epoch 1) Step 307/593 lr: 0.000013 loss: 0.108 loss(reduced): 0.160 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30877.0 tgs: 259 data_time: 0.56s time: 119.16s eta: 9:29:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:17:22][INFO] [Train] (Epoch 1) Step 308/593 lr: 0.000013 loss: 0.151 loss(reduced): 0.164 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31440.0 tgs: 263 data_time: 0.85s time: 119.22s eta: 9:28:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:19:21][INFO] [Train] (Epoch 1) Step 309/593 lr: 0.000013 loss: 0.162 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32401.0 tgs: 271 data_time: 0.91s time: 119.46s eta: 9:27:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:21:20][INFO] [Train] (Epoch 1) Step 310/593 lr: 0.000013 loss: 0.163 loss(reduced): 0.155 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30842.0 tgs: 259 data_time: 0.81s time: 118.99s eta: 9:23:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:23:19][INFO] [Train] (Epoch 1) Step 311/593 lr: 0.000013 loss: 0.176 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32699.0 tgs: 274 data_time: 0.80s time: 119.03s eta: 9:21:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:25:19][INFO] [Train] (Epoch 1) Step 312/593 lr: 0.000013 loss: 0.132 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31886.0 tgs: 266 data_time: 0.77s time: 119.54s eta: 9:21:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:27:18][INFO] [Train] (Epoch 1) Step 313/593 lr: 0.000013 loss: 0.159 loss(reduced): 0.179 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32020.0 tgs: 267 data_time: 0.79s time: 119.59s eta: 9:20:04 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:29:17][INFO] [Train] (Epoch 1) Step 314/593 lr: 0.000013 loss: 0.175 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32005.0 tgs: 268 data_time: 0.72s time: 119.12s eta: 9:15:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:31:16][INFO] [Train] (Epoch 1) Step 315/593 lr: 0.000013 loss: 0.154 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31151.0 tgs: 261 data_time: 0.87s time: 119.04s eta: 9:13:32 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:33:15][INFO] [Train] (Epoch 1) Step 316/593 lr: 0.000013 loss: 0.174 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31785.0 tgs: 267 data_time: 0.91s time: 118.89s eta: 9:10:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:35:15][INFO] [Train] (Epoch 1) Step 317/593 lr: 0.000013 loss: 0.153 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.72s time: 119.32s eta: 9:10:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:37:14][INFO] [Train] (Epoch 1) Step 318/593 lr: 0.000012 loss: 0.117 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31348.0 tgs: 263 data_time: 1.07s time: 119.09s eta: 9:07:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:39:13][INFO] [Train] (Epoch 1) Step 319/593 lr: 0.000012 loss: 0.141 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31999.0 tgs: 269 data_time: 0.69s time: 118.87s eta: 9:04:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:41:12][INFO] [Train] (Epoch 1) Step 320/593 lr: 0.000012 loss: 0.247 loss(reduced): 0.174 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31957.0 tgs: 267 data_time: 0.79s time: 119.66s eta: 9:06:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:43:12][INFO] [Train] (Epoch 1) Step 321/593 lr: 0.000012 loss: 0.202 loss(reduced): 0.152 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31974.0 tgs: 267 data_time: 0.84s time: 119.43s eta: 9:03:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:45:11][INFO] [Train] (Epoch 1) Step 322/593 lr: 0.000012 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31719.0 tgs: 265 data_time: 0.90s time: 119.43s eta: 9:01:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:47:10][INFO] [Train] (Epoch 1) Step 323/593 lr: 0.000012 loss: 0.137 loss(reduced): 0.157 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32187.0 tgs: 271 data_time: 0.58s time: 118.74s eta: 8:56:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:49:09][INFO] [Train] (Epoch 1) Step 324/593 lr: 0.000012 loss: 0.139 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31292.0 tgs: 263 data_time: 0.55s time: 118.87s eta: 8:54:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:51:08][INFO] [Train] (Epoch 1) Step 325/593 lr: 0.000012 loss: 0.123 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31138.0 tgs: 260 data_time: 0.85s time: 119.33s eta: 8:54:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:53:07][INFO] [Train] (Epoch 1) Step 326/593 lr: 0.000012 loss: 0.159 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30953.0 tgs: 259 data_time: 0.82s time: 119.15s eta: 8:52:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:55:07][INFO] [Train] (Epoch 1) Step 327/593 lr: 0.000012 loss: 0.172 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31993.0 tgs: 268 data_time: 0.86s time: 119.26s eta: 8:50:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:57:07][INFO] [Train] (Epoch 1) Step 328/593 lr: 0.000012 loss: 0.145 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31220.0 tgs: 259 data_time: 0.83s time: 120.48s eta: 8:54:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 03:59:06][INFO] [Train] (Epoch 1) Step 329/593 lr: 0.000012 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31429.0 tgs: 263 data_time: 0.98s time: 119.24s eta: 8:46:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:01:05][INFO] [Train] (Epoch 1) Step 330/593 lr: 0.000012 loss: 0.169 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31862.0 tgs: 267 data_time: 0.71s time: 119.09s eta: 8:43:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:03:06][INFO] [Train] (Epoch 1) Step 331/593 lr: 0.000012 loss: 0.171 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32053.0 tgs: 265 data_time: 0.91s time: 120.82s eta: 8:49:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:05:06][INFO] [Train] (Epoch 1) Step 332/593 lr: 0.000012 loss: 0.174 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32256.0 tgs: 268 data_time: 0.65s time: 120.19s eta: 8:44:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:07:05][INFO] [Train] (Epoch 1) Step 333/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31813.0 tgs: 267 data_time: 0.94s time: 118.82s eta: 8:36:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:09:05][INFO] [Train] (Epoch 1) Step 334/593 lr: 0.000012 loss: 0.131 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31227.0 tgs: 261 data_time: 0.73s time: 119.35s eta: 8:37:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:11:04][INFO] [Train] (Epoch 1) Step 335/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32055.0 tgs: 268 data_time: 0.71s time: 119.33s eta: 8:35:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:13:03][INFO] [Train] (Epoch 1) Step 336/593 lr: 0.000012 loss: 0.135 loss(reduced): 0.164 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31599.0 tgs: 265 data_time: 0.79s time: 119.05s eta: 8:31:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:15:02][INFO] [Train] (Epoch 1) Step 337/593 lr: 0.000012 loss: 0.143 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30507.0 tgs: 255 data_time: 0.90s time: 119.17s eta: 8:30:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:17:01][INFO] [Train] (Epoch 1) Step 338/593 lr: 0.000012 loss: 0.178 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 30187.0 tgs: 253 data_time: 0.88s time: 119.03s eta: 8:27:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:19:00][INFO] [Train] (Epoch 1) Step 339/593 lr: 0.000012 loss: 0.160 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32292.0 tgs: 271 data_time: 0.75s time: 119.13s eta: 8:26:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:21:00][INFO] [Train] (Epoch 1) Step 340/593 lr: 0.000012 loss: 0.144 loss(reduced): 0.156 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31867.0 tgs: 267 data_time: 0.76s time: 119.23s eta: 8:24:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:22:59][INFO] [Train] (Epoch 1) Step 341/593 lr: 0.000012 loss: 0.128 loss(reduced): 0.162 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31197.0 tgs: 261 data_time: 0.60s time: 119.23s eta: 8:22:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:24:58][INFO] [Train] (Epoch 1) Step 342/593 lr: 0.000012 loss: 0.156 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32287.0 tgs: 270 data_time: 0.99s time: 119.37s eta: 8:21:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:26:58][INFO] [Train] (Epoch 1) Step 343/593 lr: 0.000012 loss: 0.131 loss(reduced): 0.157 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30415.0 tgs: 253 data_time: 0.72s time: 119.79s eta: 8:21:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:28:57][INFO] [Train] (Epoch 1) Step 344/593 lr: 0.000012 loss: 0.151 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31887.0 tgs: 267 data_time: 0.73s time: 119.19s eta: 8:16:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:30:56][INFO] [Train] (Epoch 1) Step 345/593 lr: 0.000011 loss: 0.149 loss(reduced): 0.154 grad_norm: 0.11 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31878.0 tgs: 267 data_time: 1.08s time: 119.02s eta: 8:13:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:32:56][INFO] [Train] (Epoch 1) Step 346/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.163 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32070.0 tgs: 267 data_time: 0.58s time: 119.89s eta: 8:15:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:34:56][INFO] [Train] (Epoch 1) Step 347/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.159 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32189.0 tgs: 269 data_time: 0.95s time: 119.56s eta: 8:12:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:36:55][INFO] [Train] (Epoch 1) Step 348/593 lr: 0.000011 loss: 0.224 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32212.0 tgs: 269 data_time: 0.91s time: 119.59s eta: 8:10:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:38:54][INFO] [Train] (Epoch 1) Step 349/593 lr: 0.000011 loss: 0.150 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31953.0 tgs: 268 data_time: 0.88s time: 119.00s eta: 8:05:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:40:54][INFO] [Train] (Epoch 1) Step 350/593 lr: 0.000011 loss: 0.171 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32381.0 tgs: 270 data_time: 0.77s time: 119.84s eta: 8:07:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:42:54][INFO] [Train] (Epoch 1) Step 351/593 lr: 0.000011 loss: 0.158 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31932.0 tgs: 266 data_time: 0.74s time: 119.71s eta: 8:04:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:44:53][INFO] [Train] (Epoch 1) Step 352/593 lr: 0.000011 loss: 0.131 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 32254.0 tgs: 270 data_time: 0.83s time: 119.40s eta: 8:01:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:46:53][INFO] [Train] (Epoch 1) Step 353/593 lr: 0.000011 loss: 0.129 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31529.0 tgs: 264 data_time: 0.76s time: 119.29s eta: 7:59:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:48:52][INFO] [Train] (Epoch 1) Step 354/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32103.0 tgs: 269 data_time: 1.13s time: 119.19s eta: 7:56:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:50:51][INFO] [Train] (Epoch 1) Step 355/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 28927.0 tgs: 242 data_time: 0.81s time: 119.19s eta: 7:54:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:52:50][INFO] [Train] (Epoch 1) Step 356/593 lr: 0.000011 loss: 0.152 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32226.0 tgs: 270 data_time: 0.85s time: 118.93s eta: 7:51:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:54:49][INFO] [Train] (Epoch 1) Step 357/593 lr: 0.000011 loss: 0.145 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30494.0 tgs: 256 data_time: 0.93s time: 118.98s eta: 7:49:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:56:48][INFO] [Train] (Epoch 1) Step 358/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31409.0 tgs: 263 data_time: 0.87s time: 119.41s eta: 7:49:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 04:58:47][INFO] [Train] (Epoch 1) Step 359/593 lr: 0.000011 loss: 0.115 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31292.0 tgs: 262 data_time: 0.65s time: 119.03s eta: 7:46:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:00:47][INFO] [Train] (Epoch 1) Step 360/593 lr: 0.000011 loss: 0.204 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31966.0 tgs: 268 data_time: 0.85s time: 119.10s eta: 7:44:29 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:02:47][INFO] [Train] (Epoch 1) Step 361/593 lr: 0.000011 loss: 0.220 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31764.0 tgs: 264 data_time: 1.07s time: 120.01s eta: 7:46:02 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:04:46][INFO] [Train] (Epoch 1) Step 362/593 lr: 0.000011 loss: 0.130 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32366.0 tgs: 270 data_time: 0.67s time: 119.82s eta: 7:43:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:06:46][INFO] [Train] (Epoch 1) Step 363/593 lr: 0.000011 loss: 0.142 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31637.0 tgs: 265 data_time: 0.80s time: 119.17s eta: 7:38:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:08:45][INFO] [Train] (Epoch 1) Step 364/593 lr: 0.000011 loss: 0.160 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32115.0 tgs: 269 data_time: 0.83s time: 118.96s eta: 7:36:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:10:44][INFO] [Train] (Epoch 1) Step 365/593 lr: 0.000011 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 30331.0 tgs: 254 data_time: 0.75s time: 119.32s eta: 7:35:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:12:43][INFO] [Train] (Epoch 1) Step 366/593 lr: 0.000011 loss: 0.195 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31872.0 tgs: 266 data_time: 0.80s time: 119.39s eta: 7:33:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:14:42][INFO] [Train] (Epoch 1) Step 367/593 lr: 0.000011 loss: 0.213 loss(reduced): 0.165 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31324.0 tgs: 263 data_time: 0.94s time: 119.06s eta: 7:30:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:16:41][INFO] [Train] (Epoch 1) Step 368/593 lr: 0.000011 loss: 0.146 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31898.0 tgs: 268 data_time: 0.88s time: 118.99s eta: 7:28:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:18:41][INFO] [Train] (Epoch 1) Step 369/593 lr: 0.000011 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31586.0 tgs: 264 data_time: 0.88s time: 119.24s eta: 7:27:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:20:40][INFO] [Train] (Epoch 1) Step 370/593 lr: 0.000011 loss: 0.180 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31144.0 tgs: 261 data_time: 0.56s time: 118.99s eta: 7:24:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:22:39][INFO] [Train] (Epoch 1) Step 371/593 lr: 0.000011 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32320.0 tgs: 271 data_time: 0.92s time: 118.93s eta: 7:22:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:24:38][INFO] [Train] (Epoch 1) Step 372/593 lr: 0.000010 loss: 0.114 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31680.0 tgs: 266 data_time: 0.79s time: 119.08s eta: 7:20:35 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:26:37][INFO] [Train] (Epoch 1) Step 373/593 lr: 0.000010 loss: 0.207 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31763.0 tgs: 266 data_time: 0.88s time: 118.99s eta: 7:18:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:28:36][INFO] [Train] (Epoch 1) Step 374/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30274.0 tgs: 254 data_time: 0.75s time: 119.02s eta: 7:16:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:30:35][INFO] [Train] (Epoch 1) Step 375/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32531.0 tgs: 272 data_time: 1.00s time: 119.27s eta: 7:15:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:32:34][INFO] [Train] (Epoch 1) Step 376/593 lr: 0.000010 loss: 0.119 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31407.0 tgs: 263 data_time: 0.70s time: 119.36s eta: 7:13:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:34:33][INFO] [Train] (Epoch 1) Step 377/593 lr: 0.000010 loss: 0.152 loss(reduced): 0.160 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31718.0 tgs: 266 data_time: 0.96s time: 118.93s eta: 7:10:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:36:32][INFO] [Train] (Epoch 1) Step 378/593 lr: 0.000010 loss: 0.159 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32091.0 tgs: 269 data_time: 0.69s time: 119.20s eta: 7:09:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:38:32][INFO] [Train] (Epoch 1) Step 379/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31913.0 tgs: 266 data_time: 0.50s time: 119.58s eta: 7:08:30 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:40:32][INFO] [Train] (Epoch 1) Step 380/593 lr: 0.000010 loss: 0.123 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31645.0 tgs: 264 data_time: 0.65s time: 119.66s eta: 7:06:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:42:31][INFO] [Train] (Epoch 1) Step 381/593 lr: 0.000010 loss: 0.208 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31979.0 tgs: 268 data_time: 0.79s time: 118.95s eta: 7:02:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:44:30][INFO] [Train] (Epoch 1) Step 382/593 lr: 0.000010 loss: 0.129 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31525.0 tgs: 264 data_time: 0.77s time: 119.30s eta: 7:01:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:46:29][INFO] [Train] (Epoch 1) Step 383/593 lr: 0.000010 loss: 0.154 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31397.0 tgs: 262 data_time: 0.86s time: 119.51s eta: 7:00:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:48:29][INFO] [Train] (Epoch 1) Step 384/593 lr: 0.000010 loss: 0.183 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31998.0 tgs: 267 data_time: 0.88s time: 119.42s eta: 6:57:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:50:28][INFO] [Train] (Epoch 1) Step 385/593 lr: 0.000010 loss: 0.165 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31762.0 tgs: 267 data_time: 0.88s time: 118.91s eta: 6:54:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:52:27][INFO] [Train] (Epoch 1) Step 386/593 lr: 0.000010 loss: 0.170 loss(reduced): 0.160 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31543.0 tgs: 264 data_time: 0.93s time: 119.35s eta: 6:53:44 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:54:26][INFO] [Train] (Epoch 1) Step 387/593 lr: 0.000010 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31681.0 tgs: 266 data_time: 0.86s time: 118.93s eta: 6:50:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:56:25][INFO] [Train] (Epoch 1) Step 388/593 lr: 0.000010 loss: 0.142 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32002.0 tgs: 268 data_time: 0.75s time: 119.26s eta: 6:49:28 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 05:58:24][INFO] [Train] (Epoch 1) Step 389/593 lr: 0.000010 loss: 0.162 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32024.0 tgs: 269 data_time: 0.66s time: 119.03s eta: 6:46:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:00:24][INFO] [Train] (Epoch 1) Step 390/593 lr: 0.000010 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31421.0 tgs: 263 data_time: 0.79s time: 119.23s eta: 6:45:22 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:02:23][INFO] [Train] (Epoch 1) Step 391/593 lr: 0.000010 loss: 0.132 loss(reduced): 0.159 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32312.0 tgs: 270 data_time: 0.89s time: 119.40s eta: 6:43:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:04:22][INFO] [Train] (Epoch 1) Step 392/593 lr: 0.000010 loss: 0.115 loss(reduced): 0.149 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31869.0 tgs: 267 data_time: 0.92s time: 119.01s eta: 6:40:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:06:21][INFO] [Train] (Epoch 1) Step 393/593 lr: 0.000010 loss: 0.139 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31229.0 tgs: 262 data_time: 0.70s time: 119.01s eta: 6:38:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:08:21][INFO] [Train] (Epoch 1) Step 394/593 lr: 0.000010 loss: 0.113 loss(reduced): 0.159 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30799.0 tgs: 257 data_time: 0.58s time: 119.45s eta: 6:38:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:10:19][INFO] [Train] (Epoch 1) Step 395/593 lr: 0.000010 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31647.0 tgs: 266 data_time: 1.12s time: 118.80s eta: 6:34:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:12:19][INFO] [Train] (Epoch 1) Step 396/593 lr: 0.000010 loss: 0.160 loss(reduced): 0.162 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31855.0 tgs: 266 data_time: 0.64s time: 119.43s eta: 6:34:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:14:18][INFO] [Train] (Epoch 1) Step 397/593 lr: 0.000010 loss: 0.134 loss(reduced): 0.166 grad_norm: 0.12 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31184.0 tgs: 260 data_time: 0.80s time: 119.57s eta: 6:32:35 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:16:18][INFO] [Train] (Epoch 1) Step 398/593 lr: 0.000010 loss: 0.161 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32529.0 tgs: 272 data_time: 0.87s time: 119.37s eta: 6:29:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:18:19][INFO] [Train] (Epoch 1) Step 399/593 lr: 0.000010 loss: 0.150 loss(reduced): 0.148 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31739.0 tgs: 265 data_time: 0.88s time: 119.68s eta: 6:28:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:20:19][INFO] [Train] (Epoch 1) Step 400/593 lr: 0.000010 loss: 0.155 loss(reduced): 0.151 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32077.0 tgs: 268 data_time: 0.97s time: 119.36s eta: 6:25:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:22:17][INFO] [Train] (Epoch 1) Step 401/593 lr: 0.000009 loss: 0.144 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31358.0 tgs: 263 data_time: 0.74s time: 118.82s eta: 6:22:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:24:16][INFO] [Train] (Epoch 1) Step 402/593 lr: 0.000009 loss: 0.158 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31827.0 tgs: 267 data_time: 0.82s time: 119.02s eta: 6:20:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:26:16][INFO] [Train] (Epoch 1) Step 403/593 lr: 0.000009 loss: 0.170 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31941.0 tgs: 267 data_time: 0.84s time: 119.40s eta: 6:20:04 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:28:15][INFO] [Train] (Epoch 1) Step 404/593 lr: 0.000009 loss: 0.119 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 29249.0 tgs: 245 data_time: 0.66s time: 119.35s eta: 6:17:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:30:14][INFO] [Train] (Epoch 1) Step 405/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30464.0 tgs: 255 data_time: 0.92s time: 119.05s eta: 6:14:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:32:14][INFO] [Train] (Epoch 1) Step 406/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31018.0 tgs: 259 data_time: 0.98s time: 119.43s eta: 6:14:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:34:13][INFO] [Train] (Epoch 1) Step 407/593 lr: 0.000009 loss: 0.151 loss(reduced): 0.149 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32262.0 tgs: 270 data_time: 0.80s time: 119.39s eta: 6:12:05 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:36:12][INFO] [Train] (Epoch 1) Step 408/593 lr: 0.000009 loss: 0.149 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31826.0 tgs: 266 data_time: 0.80s time: 119.28s eta: 6:09:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:38:11][INFO] [Train] (Epoch 1) Step 409/593 lr: 0.000009 loss: 0.194 loss(reduced): 0.150 grad_norm: 0.11 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30312.0 tgs: 254 data_time: 0.72s time: 119.10s eta: 6:07:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:40:13][INFO] [Train] (Epoch 1) Step 410/593 lr: 0.000009 loss: 0.153 loss(reduced): 0.165 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31302.0 tgs: 256 data_time: 1.08s time: 121.96s eta: 6:14:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:42:13][INFO] [Train] (Epoch 1) Step 411/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.148 grad_norm: 0.10 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31805.0 tgs: 266 data_time: 0.75s time: 119.38s eta: 6:04:07 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:44:12][INFO] [Train] (Epoch 1) Step 412/593 lr: 0.000009 loss: 0.153 loss(reduced): 0.162 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31888.0 tgs: 266 data_time: 1.06s time: 119.57s eta: 6:02:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:46:12][INFO] [Train] (Epoch 1) Step 413/593 lr: 0.000009 loss: 0.152 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30584.0 tgs: 256 data_time: 0.73s time: 119.27s eta: 5:59:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:48:11][INFO] [Train] (Epoch 1) Step 414/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31822.0 tgs: 266 data_time: 0.90s time: 119.25s eta: 5:57:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:50:10][INFO] [Train] (Epoch 1) Step 415/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31679.0 tgs: 266 data_time: 0.85s time: 118.91s eta: 5:54:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:52:09][INFO] [Train] (Epoch 1) Step 416/593 lr: 0.000009 loss: 0.131 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31785.0 tgs: 267 data_time: 0.69s time: 119.02s eta: 5:53:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:54:08][INFO] [Train] (Epoch 1) Step 417/593 lr: 0.000009 loss: 0.208 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32048.0 tgs: 269 data_time: 0.91s time: 119.06s eta: 5:51:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:56:07][INFO] [Train] (Epoch 1) Step 418/593 lr: 0.000009 loss: 0.217 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31513.0 tgs: 264 data_time: 0.63s time: 119.29s eta: 5:49:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 06:58:06][INFO] [Train] (Epoch 1) Step 419/593 lr: 0.000009 loss: 0.142 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32207.0 tgs: 270 data_time: 0.71s time: 118.97s eta: 5:46:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:00:06][INFO] [Train] (Epoch 1) Step 420/593 lr: 0.000009 loss: 0.134 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31231.0 tgs: 261 data_time: 0.97s time: 119.27s eta: 5:45:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:02:05][INFO] [Train] (Epoch 1) Step 421/593 lr: 0.000009 loss: 0.155 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32311.0 tgs: 270 data_time: 0.82s time: 119.33s eta: 5:44:04 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:04:04][INFO] [Train] (Epoch 1) Step 422/593 lr: 0.000009 loss: 0.156 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31739.0 tgs: 266 data_time: 0.61s time: 119.05s eta: 5:41:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:06:03][INFO] [Train] (Epoch 1) Step 423/593 lr: 0.000009 loss: 0.121 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31186.0 tgs: 260 data_time: 0.85s time: 119.49s eta: 5:40:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:08:03][INFO] [Train] (Epoch 1) Step 424/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.146 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32109.0 tgs: 268 data_time: 0.72s time: 119.38s eta: 5:38:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:10:02][INFO] [Train] (Epoch 1) Step 425/593 lr: 0.000009 loss: 0.140 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 30948.0 tgs: 258 data_time: 0.61s time: 119.57s eta: 5:36:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:12:01][INFO] [Train] (Epoch 1) Step 426/593 lr: 0.000009 loss: 0.231 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32614.0 tgs: 274 data_time: 0.67s time: 118.98s eta: 5:33:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:14:00][INFO] [Train] (Epoch 1) Step 427/593 lr: 0.000009 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30620.0 tgs: 257 data_time: 0.84s time: 118.95s eta: 5:31:05 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:16:00][INFO] [Train] (Epoch 1) Step 428/593 lr: 0.000009 loss: 0.117 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32093.0 tgs: 269 data_time: 0.45s time: 119.24s eta: 5:29:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:17:59][INFO] [Train] (Epoch 1) Step 429/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31923.0 tgs: 267 data_time: 0.83s time: 119.40s eta: 5:28:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:19:58][INFO] [Train] (Epoch 1) Step 430/593 lr: 0.000009 loss: 0.175 loss(reduced): 0.158 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32322.0 tgs: 270 data_time: 0.93s time: 119.46s eta: 5:26:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:21:57][INFO] [Train] (Epoch 1) Step 431/593 lr: 0.000009 loss: 0.157 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31272.0 tgs: 262 data_time: 0.59s time: 118.95s eta: 5:23:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:23:57][INFO] [Train] (Epoch 1) Step 432/593 lr: 0.000009 loss: 0.147 loss(reduced): 0.168 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31917.0 tgs: 266 data_time: 0.77s time: 119.63s eta: 5:23:00 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:25:56][INFO] [Train] (Epoch 1) Step 433/593 lr: 0.000009 loss: 0.168 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31768.0 tgs: 266 data_time: 1.18s time: 119.34s eta: 5:20:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:27:56][INFO] [Train] (Epoch 1) Step 434/593 lr: 0.000008 loss: 0.156 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32052.0 tgs: 269 data_time: 0.89s time: 119.12s eta: 5:17:39 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:29:55][INFO] [Train] (Epoch 1) Step 435/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.144 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31452.0 tgs: 263 data_time: 0.88s time: 119.42s eta: 5:16:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:31:55][INFO] [Train] (Epoch 1) Step 436/593 lr: 0.000008 loss: 0.135 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 266 data_time: 0.76s time: 119.60s eta: 5:14:57 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:33:54][INFO] [Train] (Epoch 1) Step 437/593 lr: 0.000008 loss: 0.168 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32188.0 tgs: 270 data_time: 0.94s time: 119.00s eta: 5:11:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:35:53][INFO] [Train] (Epoch 1) Step 438/593 lr: 0.000008 loss: 0.141 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 31324.0 tgs: 263 data_time: 0.48s time: 119.00s eta: 5:09:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:37:52][INFO] [Train] (Epoch 1) Step 439/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30612.0 tgs: 256 data_time: 0.74s time: 119.22s eta: 5:07:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:39:52][INFO] [Train] (Epoch 1) Step 440/593 lr: 0.000008 loss: 0.207 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31497.0 tgs: 263 data_time: 0.97s time: 119.68s eta: 5:07:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:41:51][INFO] [Train] (Epoch 1) Step 441/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31691.0 tgs: 266 data_time: 0.65s time: 119.11s eta: 5:03:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:43:50][INFO] [Train] (Epoch 1) Step 442/593 lr: 0.000008 loss: 0.134 loss(reduced): 0.164 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 29749.0 tgs: 249 data_time: 0.75s time: 119.07s eta: 5:01:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:45:51][INFO] [Train] (Epoch 1) Step 443/593 lr: 0.000008 loss: 0.144 loss(reduced): 0.163 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31551.0 tgs: 259 data_time: 1.06s time: 121.52s eta: 5:05:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:47:51][INFO] [Train] (Epoch 1) Step 444/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32444.0 tgs: 271 data_time: 1.16s time: 119.58s eta: 4:58:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:49:50][INFO] [Train] (Epoch 1) Step 445/593 lr: 0.000008 loss: 0.161 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31232.0 tgs: 261 data_time: 0.73s time: 119.50s eta: 4:56:44 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:51:50][INFO] [Train] (Epoch 1) Step 446/593 lr: 0.000008 loss: 0.140 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30928.0 tgs: 258 data_time: 0.87s time: 119.73s eta: 4:55:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:53:50][INFO] [Train] (Epoch 1) Step 447/593 lr: 0.000008 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31578.0 tgs: 264 data_time: 0.79s time: 119.52s eta: 4:52:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:55:49][INFO] [Train] (Epoch 1) Step 448/593 lr: 0.000008 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31919.0 tgs: 267 data_time: 0.62s time: 119.41s eta: 4:50:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:57:48][INFO] [Train] (Epoch 1) Step 449/593 lr: 0.000008 loss: 0.145 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31113.0 tgs: 261 data_time: 0.84s time: 119.20s eta: 4:48:04 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 07:59:47][INFO] [Train] (Epoch 1) Step 450/593 lr: 0.000008 loss: 0.120 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31562.0 tgs: 265 data_time: 0.59s time: 118.89s eta: 4:45:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:01:46][INFO] [Train] (Epoch 1) Step 451/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.164 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31492.0 tgs: 264 data_time: 0.80s time: 119.08s eta: 4:43:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:03:46][INFO] [Train] (Epoch 1) Step 452/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32289.0 tgs: 269 data_time: 0.65s time: 119.99s eta: 4:43:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:05:45][INFO] [Train] (Epoch 1) Step 453/593 lr: 0.000008 loss: 0.171 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31861.0 tgs: 267 data_time: 0.72s time: 119.15s eta: 4:39:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:07:44][INFO] [Train] (Epoch 1) Step 454/593 lr: 0.000008 loss: 0.121 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31218.0 tgs: 262 data_time: 0.77s time: 119.03s eta: 4:37:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:09:43][INFO] [Train] (Epoch 1) Step 455/593 lr: 0.000008 loss: 0.148 loss(reduced): 0.156 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31172.0 tgs: 261 data_time: 0.96s time: 119.04s eta: 4:35:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:11:43][INFO] [Train] (Epoch 1) Step 456/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.148 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32034.0 tgs: 268 data_time: 0.98s time: 119.38s eta: 4:34:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:13:42][INFO] [Train] (Epoch 1) Step 457/593 lr: 0.000008 loss: 0.122 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32099.0 tgs: 268 data_time: 0.94s time: 119.41s eta: 4:32:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:15:41][INFO] [Train] (Epoch 1) Step 458/593 lr: 0.000008 loss: 0.159 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31802.0 tgs: 267 data_time: 1.06s time: 118.95s eta: 4:29:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:17:41][INFO] [Train] (Epoch 1) Step 459/593 lr: 0.000008 loss: 0.125 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31992.0 tgs: 267 data_time: 0.67s time: 119.60s eta: 4:29:05 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:19:40][INFO] [Train] (Epoch 1) Step 460/593 lr: 0.000008 loss: 0.179 loss(reduced): 0.155 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31289.0 tgs: 262 data_time: 0.81s time: 119.24s eta: 4:26:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:21:39][INFO] [Train] (Epoch 1) Step 461/593 lr: 0.000008 loss: 0.213 loss(reduced): 0.157 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32434.0 tgs: 272 data_time: 0.67s time: 119.06s eta: 4:23:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:23:38][INFO] [Train] (Epoch 1) Step 462/593 lr: 0.000008 loss: 0.180 loss(reduced): 0.154 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31443.0 tgs: 264 data_time: 0.76s time: 118.85s eta: 4:21:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:25:37][INFO] [Train] (Epoch 1) Step 463/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31007.0 tgs: 261 data_time: 0.84s time: 118.77s eta: 4:19:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:27:36][INFO] [Train] (Epoch 1) Step 464/593 lr: 0.000008 loss: 0.150 loss(reduced): 0.160 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31973.0 tgs: 267 data_time: 0.93s time: 119.43s eta: 4:18:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:29:36][INFO] [Train] (Epoch 1) Step 465/593 lr: 0.000008 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31803.0 tgs: 266 data_time: 0.71s time: 119.38s eta: 4:16:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:31:35][INFO] [Train] (Epoch 1) Step 466/593 lr: 0.000008 loss: 0.166 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31737.0 tgs: 265 data_time: 0.95s time: 119.36s eta: 4:14:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:33:34][INFO] [Train] (Epoch 1) Step 467/593 lr: 0.000008 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32173.0 tgs: 270 data_time: 0.66s time: 119.04s eta: 4:11:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:35:33][INFO] [Train] (Epoch 1) Step 468/593 lr: 0.000008 loss: 0.209 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30968.0 tgs: 259 data_time: 0.67s time: 119.13s eta: 4:10:09 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:37:32][INFO] [Train] (Epoch 1) Step 469/593 lr: 0.000008 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32101.0 tgs: 269 data_time: 0.70s time: 119.13s eta: 4:08:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:39:31][INFO] [Train] (Epoch 1) Step 470/593 lr: 0.000008 loss: 0.170 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31927.0 tgs: 268 data_time: 0.74s time: 118.96s eta: 4:05:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:41:30][INFO] [Train] (Epoch 1) Step 471/593 lr: 0.000008 loss: 0.197 loss(reduced): 0.153 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31522.0 tgs: 264 data_time: 0.80s time: 119.15s eta: 4:04:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:43:30][INFO] [Train] (Epoch 1) Step 472/593 lr: 0.000007 loss: 0.193 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31635.0 tgs: 264 data_time: 0.88s time: 119.50s eta: 4:02:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:45:29][INFO] [Train] (Epoch 1) Step 473/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.161 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31152.0 tgs: 261 data_time: 0.86s time: 119.24s eta: 4:00:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:47:28][INFO] [Train] (Epoch 1) Step 474/593 lr: 0.000007 loss: 0.142 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30932.0 tgs: 259 data_time: 0.77s time: 119.19s eta: 3:58:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:49:28][INFO] [Train] (Epoch 1) Step 475/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.147 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31363.0 tgs: 262 data_time: 0.51s time: 119.35s eta: 3:56:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:51:27][INFO] [Train] (Epoch 1) Step 476/593 lr: 0.000007 loss: 0.208 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30430.0 tgs: 254 data_time: 0.71s time: 119.37s eta: 3:54:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:53:27][INFO] [Train] (Epoch 1) Step 477/593 lr: 0.000007 loss: 0.168 loss(reduced): 0.152 grad_norm: 0.09 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31294.0 tgs: 260 data_time: 0.72s time: 119.96s eta: 3:53:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:55:26][INFO] [Train] (Epoch 1) Step 478/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31763.0 tgs: 266 data_time: 0.74s time: 119.33s eta: 3:50:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:57:26][INFO] [Train] (Epoch 1) Step 479/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31655.0 tgs: 264 data_time: 0.53s time: 119.62s eta: 3:49:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 08:59:25][INFO] [Train] (Epoch 1) Step 480/593 lr: 0.000007 loss: 0.154 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31612.0 tgs: 265 data_time: 0.51s time: 118.86s eta: 3:45:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:01:25][INFO] [Train] (Epoch 1) Step 481/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32341.0 tgs: 270 data_time: 1.10s time: 119.75s eta: 3:45:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:03:24][INFO] [Train] (Epoch 1) Step 482/593 lr: 0.000007 loss: 0.162 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31816.0 tgs: 266 data_time: 0.85s time: 119.33s eta: 3:42:44 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:05:23][INFO] [Train] (Epoch 1) Step 483/593 lr: 0.000007 loss: 0.146 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32133.0 tgs: 269 data_time: 0.87s time: 119.26s eta: 3:40:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:07:22][INFO] [Train] (Epoch 1) Step 484/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30259.0 tgs: 254 data_time: 0.65s time: 119.00s eta: 3:38:09 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:09:22][INFO] [Train] (Epoch 1) Step 485/593 lr: 0.000007 loss: 0.121 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31645.0 tgs: 264 data_time: 0.90s time: 119.44s eta: 3:36:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:11:21][INFO] [Train] (Epoch 1) Step 486/593 lr: 0.000007 loss: 0.158 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32428.0 tgs: 272 data_time: 0.72s time: 119.03s eta: 3:34:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:13:20][INFO] [Train] (Epoch 1) Step 487/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31725.0 tgs: 266 data_time: 1.06s time: 119.01s eta: 3:32:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:15:19][INFO] [Train] (Epoch 1) Step 488/593 lr: 0.000007 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32217.0 tgs: 270 data_time: 0.88s time: 119.21s eta: 3:30:35 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:17:19][INFO] [Train] (Epoch 1) Step 489/593 lr: 0.000007 loss: 0.141 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31727.0 tgs: 264 data_time: 0.81s time: 120.10s eta: 3:30:10 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:19:18][INFO] [Train] (Epoch 1) Step 490/593 lr: 0.000007 loss: 0.116 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32137.0 tgs: 269 data_time: 0.75s time: 119.13s eta: 3:26:29 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:21:18][INFO] [Train] (Epoch 1) Step 491/593 lr: 0.000007 loss: 0.173 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32379.0 tgs: 271 data_time: 0.79s time: 119.32s eta: 3:24:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:23:17][INFO] [Train] (Epoch 1) Step 492/593 lr: 0.000007 loss: 0.151 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31484.0 tgs: 264 data_time: 0.56s time: 119.08s eta: 3:22:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:25:16][INFO] [Train] (Epoch 1) Step 493/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.159 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31724.0 tgs: 265 data_time: 1.19s time: 119.42s eta: 3:21:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:27:15][INFO] [Train] (Epoch 1) Step 494/593 lr: 0.000007 loss: 0.156 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32442.0 tgs: 271 data_time: 0.57s time: 119.29s eta: 3:18:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:29:15][INFO] [Train] (Epoch 1) Step 495/593 lr: 0.000007 loss: 0.157 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.7GB text_tokens: 30447.0 tgs: 255 data_time: 0.77s time: 119.22s eta: 3:16:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:31:14][INFO] [Train] (Epoch 1) Step 496/593 lr: 0.000007 loss: 0.164 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31863.0 tgs: 267 data_time: 0.64s time: 119.15s eta: 3:14:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:33:13][INFO] [Train] (Epoch 1) Step 497/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31138.0 tgs: 261 data_time: 0.58s time: 119.01s eta: 3:12:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:35:12][INFO] [Train] (Epoch 1) Step 498/593 lr: 0.000007 loss: 0.138 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32193.0 tgs: 270 data_time: 0.78s time: 119.18s eta: 3:10:41 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:37:15][INFO] [Train] (Epoch 1) Step 499/593 lr: 0.000007 loss: 0.150 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31020.0 tgs: 256 data_time: 0.91s time: 120.92s eta: 3:11:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:39:14][INFO] [Train] (Epoch 1) Step 500/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31011.0 tgs: 259 data_time: 0.71s time: 119.52s eta: 3:07:14 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:41:13][INFO] [Train] (Epoch 1) Step 501/593 lr: 0.000007 loss: 0.126 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30905.0 tgs: 259 data_time: 0.74s time: 119.24s eta: 3:04:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:43:12][INFO] [Train] (Epoch 1) Step 502/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.152 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31911.0 tgs: 268 data_time: 0.79s time: 119.00s eta: 3:02:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:45:12][INFO] [Train] (Epoch 1) Step 503/593 lr: 0.000007 loss: 0.108 loss(reduced): 0.147 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31956.0 tgs: 268 data_time: 0.73s time: 119.21s eta: 3:00:48 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:47:11][INFO] [Train] (Epoch 1) Step 504/593 lr: 0.000007 loss: 0.202 loss(reduced): 0.158 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31884.0 tgs: 266 data_time: 0.65s time: 119.46s eta: 2:59:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:49:10][INFO] [Train] (Epoch 1) Step 505/593 lr: 0.000007 loss: 0.166 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31683.0 tgs: 266 data_time: 0.54s time: 118.98s eta: 2:56:28 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:51:10][INFO] [Train] (Epoch 1) Step 506/593 lr: 0.000007 loss: 0.182 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32096.0 tgs: 268 data_time: 0.73s time: 119.35s eta: 2:55:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:53:09][INFO] [Train] (Epoch 1) Step 507/593 lr: 0.000007 loss: 0.194 loss(reduced): 0.146 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31574.0 tgs: 264 data_time: 0.81s time: 119.41s eta: 2:53:08 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:55:08][INFO] [Train] (Epoch 1) Step 508/593 lr: 0.000007 loss: 0.170 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32323.0 tgs: 270 data_time: 0.96s time: 119.36s eta: 2:51:04 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:57:10][INFO] [Train] (Epoch 1) Step 509/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.155 grad_norm: 0.08 if_nan_skip: 0 max_memory: 40.8GB text_tokens: 30571.0 tgs: 251 data_time: 0.64s time: 121.45s eta: 2:52:03 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 09:59:09][INFO] [Train] (Epoch 1) Step 510/593 lr: 0.000007 loss: 0.183 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31775.0 tgs: 266 data_time: 1.01s time: 119.24s eta: 2:46:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:01:09][INFO] [Train] (Epoch 1) Step 511/593 lr: 0.000007 loss: 0.177 loss(reduced): 0.148 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31908.0 tgs: 266 data_time: 1.04s time: 119.60s eta: 2:45:27 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:03:08][INFO] [Train] (Epoch 1) Step 512/593 lr: 0.000007 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31649.0 tgs: 266 data_time: 0.67s time: 118.92s eta: 2:42:31 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:05:07][INFO] [Train] (Epoch 1) Step 513/593 lr: 0.000007 loss: 0.101 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31201.0 tgs: 261 data_time: 0.46s time: 119.18s eta: 2:40:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:07:06][INFO] [Train] (Epoch 1) Step 514/593 lr: 0.000007 loss: 0.167 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32373.0 tgs: 272 data_time: 0.87s time: 118.96s eta: 2:38:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:09:05][INFO] [Train] (Epoch 1) Step 515/593 lr: 0.000007 loss: 0.152 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31901.0 tgs: 267 data_time: 1.23s time: 119.23s eta: 2:36:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:11:04][INFO] [Train] (Epoch 1) Step 516/593 lr: 0.000007 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32316.0 tgs: 271 data_time: 0.68s time: 119.11s eta: 2:34:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:13:04][INFO] [Train] (Epoch 1) Step 517/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31988.0 tgs: 267 data_time: 0.64s time: 119.42s eta: 2:33:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:15:04][INFO] [Train] (Epoch 1) Step 518/593 lr: 0.000007 loss: 0.153 loss(reduced): 0.154 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32265.0 tgs: 268 data_time: 0.83s time: 120.22s eta: 2:32:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:17:03][INFO] [Train] (Epoch 1) Step 519/593 lr: 0.000007 loss: 0.246 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31279.0 tgs: 262 data_time: 1.06s time: 119.28s eta: 2:29:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:19:02][INFO] [Train] (Epoch 1) Step 520/593 lr: 0.000007 loss: 0.147 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31860.0 tgs: 267 data_time: 0.91s time: 118.89s eta: 2:26:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:21:01][INFO] [Train] (Epoch 1) Step 521/593 lr: 0.000007 loss: 0.124 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 269 data_time: 0.70s time: 119.51s eta: 2:25:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:23:01][INFO] [Train] (Epoch 1) Step 522/593 lr: 0.000007 loss: 0.098 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32217.0 tgs: 270 data_time: 0.61s time: 119.07s eta: 2:22:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:25:00][INFO] [Train] (Epoch 1) Step 523/593 lr: 0.000007 loss: 0.144 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31118.0 tgs: 261 data_time: 0.68s time: 119.13s eta: 2:20:58 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:26:59][INFO] [Train] (Epoch 1) Step 524/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32017.0 tgs: 268 data_time: 0.72s time: 119.05s eta: 2:18:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:28:59][INFO] [Train] (Epoch 1) Step 525/593 lr: 0.000006 loss: 0.180 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32367.0 tgs: 270 data_time: 1.26s time: 119.85s eta: 2:17:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:30:58][INFO] [Train] (Epoch 1) Step 526/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.159 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31907.0 tgs: 267 data_time: 0.91s time: 119.31s eta: 2:15:13 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:32:57][INFO] [Train] (Epoch 1) Step 527/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32033.0 tgs: 268 data_time: 0.82s time: 119.38s eta: 2:13:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:34:57][INFO] [Train] (Epoch 1) Step 528/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31435.0 tgs: 262 data_time: 0.74s time: 119.69s eta: 2:11:39 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:36:56][INFO] [Train] (Epoch 1) Step 529/593 lr: 0.000006 loss: 0.137 loss(reduced): 0.147 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31094.0 tgs: 260 data_time: 0.69s time: 119.48s eta: 2:09:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:38:55][INFO] [Train] (Epoch 1) Step 530/593 lr: 0.000006 loss: 0.157 loss(reduced): 0.141 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31433.0 tgs: 264 data_time: 0.63s time: 118.96s eta: 2:06:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:40:55][INFO] [Train] (Epoch 1) Step 531/593 lr: 0.000006 loss: 0.145 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31896.0 tgs: 267 data_time: 0.72s time: 119.37s eta: 2:05:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:42:54][INFO] [Train] (Epoch 1) Step 532/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32219.0 tgs: 270 data_time: 0.78s time: 119.28s eta: 2:03:15 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:44:53][INFO] [Train] (Epoch 1) Step 533/593 lr: 0.000006 loss: 0.216 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31147.0 tgs: 261 data_time: 0.99s time: 119.22s eta: 2:01:12 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:46:53][INFO] [Train] (Epoch 1) Step 534/593 lr: 0.000006 loss: 0.159 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31791.0 tgs: 266 data_time: 1.06s time: 119.51s eta: 1:59:30 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:48:52][INFO] [Train] (Epoch 1) Step 535/593 lr: 0.000006 loss: 0.103 loss(reduced): 0.141 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30480.0 tgs: 255 data_time: 0.58s time: 119.42s eta: 1:57:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:50:52][INFO] [Train] (Epoch 1) Step 536/593 lr: 0.000006 loss: 0.101 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31941.0 tgs: 267 data_time: 0.70s time: 119.53s eta: 1:55:32 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:52:51][INFO] [Train] (Epoch 1) Step 537/593 lr: 0.000006 loss: 0.111 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 30457.0 tgs: 255 data_time: 0.84s time: 119.30s eta: 1:53:20 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:54:50][INFO] [Train] (Epoch 1) Step 538/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31861.0 tgs: 267 data_time: 0.71s time: 119.25s eta: 1:51:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:56:49][INFO] [Train] (Epoch 1) Step 539/593 lr: 0.000006 loss: 0.166 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31130.0 tgs: 261 data_time: 0.87s time: 118.93s eta: 1:49:01 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 10:58:49][INFO] [Train] (Epoch 1) Step 540/593 lr: 0.000006 loss: 0.190 loss(reduced): 0.145 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31661.0 tgs: 264 data_time: 1.05s time: 119.64s eta: 1:47:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:00:49][INFO] [Train] (Epoch 1) Step 541/593 lr: 0.000006 loss: 0.139 loss(reduced): 0.149 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30889.0 tgs: 257 data_time: 0.63s time: 119.73s eta: 1:45:45 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:02:48][INFO] [Train] (Epoch 1) Step 542/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.151 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31301.0 tgs: 262 data_time: 0.45s time: 119.35s eta: 1:43:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:04:47][INFO] [Train] (Epoch 1) Step 543/593 lr: 0.000006 loss: 0.170 loss(reduced): 0.156 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31135.0 tgs: 260 data_time: 0.77s time: 119.31s eta: 1:41:24 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:06:47][INFO] [Train] (Epoch 1) Step 544/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 31274.0 tgs: 262 data_time: 0.62s time: 119.28s eta: 1:39:23 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:08:46][INFO] [Train] (Epoch 1) Step 545/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32423.0 tgs: 272 data_time: 0.72s time: 118.91s eta: 1:37:06 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:10:45][INFO] [Train] (Epoch 1) Step 546/593 lr: 0.000006 loss: 0.138 loss(reduced): 0.155 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31206.0 tgs: 261 data_time: 0.63s time: 119.15s eta: 1:35:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:12:44][INFO] [Train] (Epoch 1) Step 547/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32018.0 tgs: 268 data_time: 0.81s time: 119.07s eta: 1:33:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:14:43][INFO] [Train] (Epoch 1) Step 548/593 lr: 0.000006 loss: 0.104 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31494.0 tgs: 264 data_time: 0.52s time: 119.26s eta: 1:31:25 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:16:43][INFO] [Train] (Epoch 1) Step 549/593 lr: 0.000006 loss: 0.161 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31682.0 tgs: 264 data_time: 0.63s time: 119.72s eta: 1:29:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:18:42][INFO] [Train] (Epoch 1) Step 550/593 lr: 0.000006 loss: 0.123 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32118.0 tgs: 269 data_time: 0.72s time: 119.01s eta: 1:27:16 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:20:41][INFO] [Train] (Epoch 1) Step 551/593 lr: 0.000006 loss: 0.168 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31034.0 tgs: 261 data_time: 0.59s time: 118.87s eta: 1:25:11 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:22:40][INFO] [Train] (Epoch 1) Step 552/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.5GB text_tokens: 30363.0 tgs: 254 data_time: 0.82s time: 119.48s eta: 1:23:37 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:24:40][INFO] [Train] (Epoch 1) Step 553/593 lr: 0.000006 loss: 0.165 loss(reduced): 0.156 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32476.0 tgs: 271 data_time: 0.78s time: 119.48s eta: 1:21:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:26:39][INFO] [Train] (Epoch 1) Step 554/593 lr: 0.000006 loss: 0.152 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31191.0 tgs: 262 data_time: 0.84s time: 118.94s eta: 1:19:17 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:28:38][INFO] [Train] (Epoch 1) Step 555/593 lr: 0.000006 loss: 0.154 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 30941.0 tgs: 260 data_time: 1.11s time: 118.93s eta: 1:17:18 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:30:40][INFO] [Train] (Epoch 1) Step 556/593 lr: 0.000006 loss: 0.117 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31316.0 tgs: 256 data_time: 0.91s time: 122.15s eta: 1:17:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:32:39][INFO] [Train] (Epoch 1) Step 557/593 lr: 0.000006 loss: 0.142 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32206.0 tgs: 269 data_time: 0.73s time: 119.36s eta: 1:13:36 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:34:38][INFO] [Train] (Epoch 1) Step 558/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31738.0 tgs: 266 data_time: 0.96s time: 119.06s eta: 1:11:26 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:36:37][INFO] [Train] (Epoch 1) Step 559/593 lr: 0.000006 loss: 0.119 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31801.0 tgs: 267 data_time: 0.66s time: 118.91s eta: 1:09:21 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:38:36][INFO] [Train] (Epoch 1) Step 560/593 lr: 0.000006 loss: 0.184 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30140.0 tgs: 253 data_time: 0.96s time: 119.11s eta: 1:07:29 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:40:36][INFO] [Train] (Epoch 1) Step 561/593 lr: 0.000006 loss: 0.224 loss(reduced): 0.154 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31946.0 tgs: 267 data_time: 0.99s time: 119.42s eta: 1:05:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:42:35][INFO] [Train] (Epoch 1) Step 562/593 lr: 0.000006 loss: 0.095 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31164.0 tgs: 261 data_time: 0.52s time: 119.34s eta: 1:03:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:44:34][INFO] [Train] (Epoch 1) Step 563/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.152 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31195.0 tgs: 261 data_time: 0.76s time: 119.18s eta: 1:01:34 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:46:33][INFO] [Train] (Epoch 1) Step 564/593 lr: 0.000006 loss: 0.148 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 40.9GB text_tokens: 30802.0 tgs: 258 data_time: 0.83s time: 119.27s eta: 0:59:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:48:33][INFO] [Train] (Epoch 1) Step 565/593 lr: 0.000006 loss: 0.105 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32180.0 tgs: 269 data_time: 0.84s time: 119.44s eta: 0:57:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:50:32][INFO] [Train] (Epoch 1) Step 566/593 lr: 0.000006 loss: 0.187 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31425.0 tgs: 263 data_time: 0.94s time: 119.39s eta: 0:55:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:52:32][INFO] [Train] (Epoch 1) Step 567/593 lr: 0.000006 loss: 0.140 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31849.0 tgs: 266 data_time: 0.71s time: 119.49s eta: 0:53:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:54:31][INFO] [Train] (Epoch 1) Step 568/593 lr: 0.000006 loss: 0.134 loss(reduced): 0.145 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31833.0 tgs: 266 data_time: 0.76s time: 119.36s eta: 0:51:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:56:31][INFO] [Train] (Epoch 1) Step 569/593 lr: 0.000006 loss: 0.176 loss(reduced): 0.161 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 30686.0 tgs: 256 data_time: 0.81s time: 119.46s eta: 0:49:46 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 11:58:30][INFO] [Train] (Epoch 1) Step 570/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.151 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31968.0 tgs: 268 data_time: 0.82s time: 118.89s eta: 0:47:33 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:00:29][INFO] [Train] (Epoch 1) Step 571/593 lr: 0.000006 loss: 0.125 loss(reduced): 0.143 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31345.0 tgs: 262 data_time: 0.70s time: 119.28s eta: 0:45:43 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:02:28][INFO] [Train] (Epoch 1) Step 572/593 lr: 0.000006 loss: 0.189 loss(reduced): 0.151 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32235.0 tgs: 270 data_time: 1.00s time: 119.21s eta: 0:43:42 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:04:27][INFO] [Train] (Epoch 1) Step 573/593 lr: 0.000006 loss: 0.151 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31559.0 tgs: 265 data_time: 0.87s time: 118.99s eta: 0:41:38 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:06:26][INFO] [Train] (Epoch 1) Step 574/593 lr: 0.000006 loss: 0.167 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31440.0 tgs: 264 data_time: 0.90s time: 119.04s eta: 0:39:40 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:08:26][INFO] [Train] (Epoch 1) Step 575/593 lr: 0.000006 loss: 0.214 loss(reduced): 0.154 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31963.0 tgs: 267 data_time: 0.83s time: 119.66s eta: 0:37:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:10:25][INFO] [Train] (Epoch 1) Step 576/593 lr: 0.000006 loss: 0.169 loss(reduced): 0.145 grad_norm: 0.08 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31929.0 tgs: 267 data_time: 0.71s time: 119.28s eta: 0:35:47 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:12:25][INFO] [Train] (Epoch 1) Step 577/593 lr: 0.000006 loss: 0.193 loss(reduced): 0.144 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32069.0 tgs: 267 data_time: 0.85s time: 119.75s eta: 0:33:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:14:26][INFO] [Train] (Epoch 1) Step 578/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.0GB text_tokens: 31757.0 tgs: 261 data_time: 0.64s time: 121.23s eta: 0:32:19 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:16:25][INFO] [Train] (Epoch 1) Step 579/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31053.0 tgs: 260 data_time: 0.99s time: 119.37s eta: 0:29:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:18:25][INFO] [Train] (Epoch 1) Step 580/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.153 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 31991.0 tgs: 268 data_time: 0.63s time: 119.32s eta: 0:27:50 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:20:24][INFO] [Train] (Epoch 1) Step 581/593 lr: 0.000006 loss: 0.096 loss(reduced): 0.143 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31171.0 tgs: 260 data_time: 0.71s time: 119.63s eta: 0:25:55 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:22:24][INFO] [Train] (Epoch 1) Step 582/593 lr: 0.000006 loss: 0.156 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 32630.0 tgs: 273 data_time: 0.90s time: 119.10s eta: 0:23:49 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:24:23][INFO] [Train] (Epoch 1) Step 583/593 lr: 0.000006 loss: 0.135 loss(reduced): 0.148 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31876.0 tgs: 267 data_time: 0.83s time: 119.38s eta: 0:21:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:26:22][INFO] [Train] (Epoch 1) Step 584/593 lr: 0.000006 loss: 0.133 loss(reduced): 0.155 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.5GB text_tokens: 31848.0 tgs: 267 data_time: 0.96s time: 119.26s eta: 0:19:52 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:28:21][INFO] [Train] (Epoch 1) Step 585/593 lr: 0.000006 loss: 0.144 loss(reduced): 0.146 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 28089.0 tgs: 235 data_time: 0.86s time: 119.24s eta: 0:17:53 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:30:21][INFO] [Train] (Epoch 1) Step 586/593 lr: 0.000006 loss: 0.149 loss(reduced): 0.144 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31984.0 tgs: 266 data_time: 1.18s time: 119.91s eta: 0:15:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:32:21][INFO] [Train] (Epoch 1) Step 587/593 lr: 0.000006 loss: 0.136 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32157.0 tgs: 269 data_time: 0.84s time: 119.21s eta: 0:13:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:34:20][INFO] [Train] (Epoch 1) Step 588/593 lr: 0.000006 loss: 0.256 loss(reduced): 0.158 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.3GB text_tokens: 32001.0 tgs: 268 data_time: 0.97s time: 119.09s eta: 0:11:54 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:36:19][INFO] [Train] (Epoch 1) Step 589/593 lr: 0.000006 loss: 0.113 loss(reduced): 0.150 grad_norm: 0.07 if_nan_skip: 0 max_memory: 41.1GB text_tokens: 31407.0 tgs: 263 data_time: 1.06s time: 119.32s eta: 0:09:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:38:18][INFO] [Train] (Epoch 1) Step 590/593 lr: 0.000006 loss: 0.131 loss(reduced): 0.149 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32062.0 tgs: 269 data_time: 0.82s time: 119.17s eta: 0:07:56 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:40:18][INFO] [Train] (Epoch 1) Step 591/593 lr: 0.000006 loss: 0.153 loss(reduced): 0.152 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 31355.0 tgs: 261 data_time: 0.52s time: 119.69s eta: 0:05:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:42:21][INFO] [Train] (Epoch 1) Step 592/593 lr: 0.000006 loss: 0.120 loss(reduced): 0.150 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.2GB text_tokens: 31391.0 tgs: 258 data_time: 0.85s time: 121.41s eta: 0:04:02 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:44:20][INFO] [Train] (Epoch 1) Step 593/593 lr: 0.000006 loss: 0.128 loss(reduced): 0.153 grad_norm: 0.06 if_nan_skip: 0 max_memory: 41.4GB text_tokens: 32391.0 tgs: 271 data_time: 0.61s time: 119.21s eta: 0:01:59 +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:56:17][SUCCESS] [HF Checkpoint] Elapsed time 717.27 seconds, peak gpu memory 17.4G +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:56:17][WARNING] The saved checkpoint cannot be resumed. If you want to save a resumable checkpoint, please remove `--checkpoint-drop-optimizer` from the command. +[XTuner][RANK 9][DP 2][SP 1][TP 0][2025-01-22 12:56:17][INFO] [Train] Cost 19:54:22